next              310 arch/alpha/include/asm/core_marvel.h 	struct io7 *next;
next               24 arch/alpha/include/asm/gct.h 	gct6_handle next;
next              124 arch/alpha/include/asm/mmu_context.h 	unsigned long next = asn + 1;
next              129 arch/alpha/include/asm/mmu_context.h 		next = (asn & ~HARDWARE_ASN_MASK) + ASN_FIRST_VERSION;
next              131 arch/alpha/include/asm/mmu_context.h 	cpu_last_asn(cpu) = next;
next              132 arch/alpha/include/asm/mmu_context.h 	return next;
next              137 arch/alpha/include/asm/mmu_context.h 	      struct task_struct *next)
next              162 arch/alpha/include/asm/mmu_context.h 	task_thread_info(next)->pcb.asn = mmc & HARDWARE_ASN_MASK;
next              167 arch/alpha/include/asm/mmu_context.h 	      struct task_struct *next)
next              182 arch/alpha/include/asm/mmu_context.h 	ev5_switch_mm(prev_mm, next_mm, next);
next               22 arch/alpha/include/asm/pci.h 	struct pci_controller *next;
next               84 arch/alpha/kernel/console.c 		for (hose = hose_head; hose; hose = hose->next) {
next               97 arch/alpha/kernel/core_marvel.c 	return (prev ? prev->next : io7_head);
next              105 arch/alpha/kernel/core_marvel.c 	for (io7 = io7_head; io7 && io7->pe != pe; io7 = io7->next)
next              142 arch/alpha/kernel/core_marvel.c 		io7->next = io7_head;
next              145 arch/alpha/kernel/core_marvel.c 		for (insp = io7_head; insp; insp = insp->next) {
next              152 arch/alpha/kernel/core_marvel.c 			if (NULL == insp->next || 
next              153 arch/alpha/kernel/core_marvel.c 			    insp->next->pe > io7->pe) { /* insert here */
next              154 arch/alpha/kernel/core_marvel.c 				io7->next = insp->next;
next              155 arch/alpha/kernel/core_marvel.c 				insp->next = io7;
next              163 arch/alpha/kernel/core_marvel.c 			io7->next = io7_head;
next              704 arch/alpha/kernel/core_marvel.c 	for (hose = hose_head; hose; hose = hose->next) {
next              440 arch/alpha/kernel/core_mcpcia.c 	for (hose = hose_head; hose; hose = hose->next)
next              551 arch/alpha/kernel/core_mcpcia.c 	for (hose = hose_head; hose; hose = hose->next, iodpp++) {
next              595 arch/alpha/kernel/core_mcpcia.c 		for (hose = hose_head; hose; hose = hose->next)
next              480 arch/alpha/kernel/core_titan.c 	for (hose = hose_head; hose; hose = hose->next)
next              754 arch/alpha/kernel/core_titan.c 	for (hose = hose_head; hose; hose = hose->next)
next              159 arch/alpha/kernel/err_common.c 	struct el_subpacket *next = NULL;
next              162 arch/alpha/kernel/err_common.c 	for (; h && h->class != header->class; h = h->next);
next              163 arch/alpha/kernel/err_common.c 	if (h) next = h->handler(header);
next              165 arch/alpha/kernel/err_common.c 	return next;
next              197 arch/alpha/kernel/err_common.c 	struct el_subpacket *next = NULL;
next              204 arch/alpha/kernel/err_common.c 		next = el_process_header_subpacket(header);
next              207 arch/alpha/kernel/err_common.c 		if (NULL == (next = el_process_subpacket_reg(header))) {
next              216 arch/alpha/kernel/err_common.c 	return next;
next              225 arch/alpha/kernel/err_common.c 	for (a = subpacket_annotation_list; a; a = a->next) {
next              286 arch/alpha/kernel/err_common.c 		for (; a->next != NULL; a = a->next) {
next              294 arch/alpha/kernel/err_common.c 		a->next = new;
next              296 arch/alpha/kernel/err_common.c 	new->next = NULL;
next              308 arch/alpha/kernel/err_common.c 		for (; h->next != NULL; h = h->next) {
next              315 arch/alpha/kernel/err_common.c 		h->next = new;
next              317 arch/alpha/kernel/err_common.c 	new->next = NULL;
next               18 arch/alpha/kernel/err_impl.h 	struct el_subpacket_annotation *next;
next               28 arch/alpha/kernel/err_impl.h 	struct el_subpacket_handler *next;
next               40 arch/alpha/kernel/gct.c 	if (node->next) 
next               41 arch/alpha/kernel/gct.c 		status |= gct6_find_nodes(GCT_NODE_PTR(node->next), search);
next               23 arch/alpha/kernel/module.c 	struct got_entry *next;
next               40 arch/alpha/kernel/module.c 	for (g = chains + r_sym; g ; g = g->next)
next               50 arch/alpha/kernel/module.c 	g->next = chains[r_sym].next;
next               54 arch/alpha/kernel/module.c 	chains[r_sym].next = g;
next              122 arch/alpha/kernel/module.c 		for (g = chains[i].next; g ; g = n) {
next              123 arch/alpha/kernel/module.c 			n = g->next;
next               42 arch/alpha/kernel/pci-noop.c 	hose_tail = &hose->next;
next               66 arch/alpha/kernel/pci-noop.c 		for (hose = hose_head; hose; hose = hose->next)
next              204 arch/alpha/kernel/pci.c 	struct pdev_srm_saved_conf *next;
next              228 arch/alpha/kernel/pci.c 	tmp->next = srm_saved_configs;
next              246 arch/alpha/kernel/pci.c 	for (tmp = srm_saved_configs; tmp; tmp = tmp->next) {
next              335 arch/alpha/kernel/pci.c 	for (next_busno = 0, hose = hose_head; hose; hose = hose->next) {
next              383 arch/alpha/kernel/pci.c 	for (hose = hose_head; hose; hose = hose->next) {
next              401 arch/alpha/kernel/pci.c 	hose_tail = &hose->next;
next              430 arch/alpha/kernel/pci.c 		for(hose = hose_head; hose; hose = hose->next) 
next              248 arch/alpha/kernel/setup.c 		for (hose = hose_head; hose; hose = hose->next)
next             1427 arch/alpha/kernel/setup.c 	.next	= c_next,
next             2342 arch/alpha/kernel/smc37c669.c int __init smcc669_open( struct FILE *fp, char *info, char *next, char *mode )
next              173 arch/alpha/kernel/sys_rawhide.c 	for (hose = hose_head; hose; hose = hose->next) {
next              129 arch/arc/include/asm/mmu_context.h static inline void switch_mm(struct mm_struct *prev, struct mm_struct *next,
next              145 arch/arc/include/asm/mmu_context.h 	cpumask_set_cpu(cpu, mm_cpumask(next));
next              149 arch/arc/include/asm/mmu_context.h 	write_aux_reg(ARC_REG_SCRATCH_DATA0, next->pgd);
next              152 arch/arc/include/asm/mmu_context.h 	get_new_mmu_context(next);
next              162 arch/arc/include/asm/mmu_context.h #define activate_mm(prev, next)		switch_mm(prev, next, NULL)
next               36 arch/arc/include/asm/switch_to.h #define switch_to(prev, next, last)	\
next               38 arch/arc/include/asm/switch_to.h 	ARC_EZNPS_DP_PREV(prev, next);	\
next               39 arch/arc/include/asm/switch_to.h 	ARC_FPU_PREV(prev, next);	\
next               40 arch/arc/include/asm/switch_to.h 	last = __switch_to(prev, next);\
next               41 arch/arc/include/asm/switch_to.h 	ARC_FPU_NEXT(next);		\
next               28 arch/arc/kernel/ctx_sw.c 	unsigned int next = (unsigned int)next_task;
next              117 arch/arc/kernel/ctx_sw.c 		: "n"(KSP_WORD_OFF), "r"(next), "r"(prev)
next               30 arch/arc/kernel/fpu.c void fpu_save_restore(struct task_struct *prev, struct task_struct *next)
next               33 arch/arc/kernel/fpu.c 	unsigned int *readfrom = &next->thread.fpu.aux_dpfp[0].l;
next              685 arch/arc/kernel/setup.c 	.next	= c_next,
next              690 arch/arc/mm/tlb.c 		pmd_huge_pte(mm, pmdp) = (pgtable_t) lh->next;
next               10 arch/arc/plat-eznps/ctop.c void dp_save_restore(struct task_struct *prev, struct task_struct *next)
next               13 arch/arc/plat-eznps/ctop.c 	struct eznps_dp *next_task_dp = &next->thread.dp;
next              140 arch/arm/include/asm/ecard.h 	struct expansion_card  *next;
next               23 arch/arm/include/asm/fiq.h 	struct fiq_handler *next;
next              105 arch/arm/include/asm/mmu_context.h #define activate_mm(prev,next)		switch_mm(prev, next, NULL)
next              128 arch/arm/include/asm/mmu_context.h switch_mm(struct mm_struct *prev, struct mm_struct *next,
next              140 arch/arm/include/asm/mmu_context.h 	    !cpumask_empty(mm_cpumask(next)) &&
next              141 arch/arm/include/asm/mmu_context.h 	    !cpumask_test_cpu(cpu, mm_cpumask(next)))
next              144 arch/arm/include/asm/mmu_context.h 	if (!cpumask_test_and_set_cpu(cpu, mm_cpumask(next)) || prev != next) {
next              145 arch/arm/include/asm/mmu_context.h 		check_and_switch_context(next, tsk);
next               73 arch/arm/include/asm/spinlock.h 	while (lockval.tickets.next != lockval.tickets.owner) {
next              116 arch/arm/include/asm/spinlock.h 	return lock.tickets.owner == lock.tickets.next;
next              127 arch/arm/include/asm/spinlock.h 	return (tickets.next - tickets.owner) > 1;
next               16 arch/arm/include/asm/spinlock_types.h 			u16 next;
next               20 arch/arm/include/asm/spinlock_types.h 			u16 next;
next               26 arch/arm/include/asm/switch_to.h #define switch_to(prev,next,last)					\
next               29 arch/arm/include/asm/switch_to.h 	last = __switch_to(prev,task_thread_info(prev), task_thread_info(next));	\
next              118 arch/arm/kernel/fiq.c 		f->next = current_fiq;
next              135 arch/arm/kernel/fiq.c 		current_fiq = current_fiq->next;
next               86 arch/arm/kernel/hibernate.c 	for (pbe = restore_pblist; pbe; pbe = pbe->next)
next             1322 arch/arm/kernel/setup.c 	.next	= c_next,
next               72 arch/arm/mach-ep93xx/timer-ep93xx.c static int ep93xx_clkevt_set_next_event(unsigned long next,
next               83 arch/arm/mach-ep93xx/timer-ep93xx.c 	writel(next, EP93XX_TIMER3_LOAD);
next              800 arch/arm/mach-omap1/clock.c 	if (clk->node.next || clk->node.prev)
next              164 arch/arm/mach-omap2/powerdomain.c 	int prev, next, state, trace_state = 0;
next              185 arch/arm/mach-omap2/powerdomain.c 		next = pwrdm_read_next_pwrst(pwrdm);
next              186 arch/arm/mach-omap2/powerdomain.c 		if (next != prev) {
next              188 arch/arm/mach-omap2/powerdomain.c 				       ((next & OMAP_POWERSTATE_MASK) << 8) |
next              499 arch/arm/mach-rpc/ecard.c 	for (ec = cards; ec; ec = ec->next) {
next              562 arch/arm/mach-rpc/ecard.c 	for (ec = cards; ec; ec = ec->next) {
next              656 arch/arm/mach-rpc/ecard.c 		ec = ec->next;
next              966 arch/arm/mach-rpc/ecard.c 	for (ecp = &cards; *ecp; ecp = &(*ecp)->next);
next               28 arch/arm/mm/idmap.c 	unsigned long next;
next               49 arch/arm/mm/idmap.c 		next = pmd_addr_end(addr, end);
next               52 arch/arm/mm/idmap.c 	} while (pmd++, addr = next, addr != end);
next               72 arch/arm/mm/idmap.c 	unsigned long next;
next               75 arch/arm/mm/idmap.c 		next = pud_addr_end(addr, end);
next               76 arch/arm/mm/idmap.c 		idmap_add_pmd(pud, addr, next, prot);
next               77 arch/arm/mm/idmap.c 	} while (pud++, addr = next, addr != end);
next               84 arch/arm/mm/idmap.c 	unsigned long next;
next               97 arch/arm/mm/idmap.c 		next = pgd_addr_end(addr, end);
next               98 arch/arm/mm/idmap.c 		idmap_add_pud(pgd, addr, next, prot);
next               99 arch/arm/mm/idmap.c 	} while (pgd++, addr = next, addr != end);
next              804 arch/arm/mm/mmu.c 	unsigned long next;
next              811 arch/arm/mm/mmu.c 		next = pmd_addr_end(addr, end);
next              818 arch/arm/mm/mmu.c 				((addr | next | phys) & ~SECTION_MASK) == 0) {
next              819 arch/arm/mm/mmu.c 			__map_init_section(pmd, addr, next, phys, type, ng);
next              821 arch/arm/mm/mmu.c 			alloc_init_pte(pmd, addr, next,
next              825 arch/arm/mm/mmu.c 		phys += next - addr;
next              827 arch/arm/mm/mmu.c 	} while (pmd++, addr = next, addr != end);
next              836 arch/arm/mm/mmu.c 	unsigned long next;
next              839 arch/arm/mm/mmu.c 		next = pud_addr_end(addr, end);
next              840 arch/arm/mm/mmu.c 		alloc_init_pmd(pud, addr, next, phys, type, alloc, ng);
next              841 arch/arm/mm/mmu.c 		phys += next - addr;
next              842 arch/arm/mm/mmu.c 	} while (pud++, addr = next, addr != end);
next              941 arch/arm/mm/mmu.c 		unsigned long next = pgd_addr_end(addr, end);
next              943 arch/arm/mm/mmu.c 		alloc_init_pud(pgd, addr, next, phys, type, alloc, ng);
next              945 arch/arm/mm/mmu.c 		phys += next - addr;
next              946 arch/arm/mm/mmu.c 		addr = next;
next             1061 arch/arm/mm/mmu.c 	unsigned long addr, next = 0;
next             1067 arch/arm/mm/mmu.c 		if (addr < next)
next             1094 arch/arm/mm/mmu.c 		next = (addr + PMD_SIZE - 1) & PMD_MASK;
next              127 arch/arm/plat-samsung/adc.c 	struct s3c_adc_client *next = adc->ts_pend;
next              129 arch/arm/plat-samsung/adc.c 	if (!next && !list_empty(&adc_pending)) {
next              130 arch/arm/plat-samsung/adc.c 		next = list_first_entry(&adc_pending,
next              132 arch/arm/plat-samsung/adc.c 		list_del(&next->pend);
next              136 arch/arm/plat-samsung/adc.c 	if (next) {
next              137 arch/arm/plat-samsung/adc.c 		adc_dbg(adc, "new client is %p\n", next);
next              138 arch/arm/plat-samsung/adc.c 		adc->cur = next;
next              139 arch/arm/plat-samsung/adc.c 		s3c_adc_select(adc, next);
next              417 arch/arm/probes/decode.c 	const struct decode_header *next;
next              445 arch/arm/probes/decode.c 	for (;; h = next) {
next              452 arch/arm/probes/decode.c 		next = (struct decode_header *)
next              465 arch/arm/probes/decode.c 			next = (struct decode_header *)d->table.table;
next               13 arch/arm64/include/asm/exec.h void uao_thread_switch(struct task_struct *next);
next               40 arch/arm64/include/asm/fpsimd.h extern void fpsimd_thread_switch(struct task_struct *next);
next              128 arch/arm64/include/asm/hw_breakpoint.h extern void hw_breakpoint_thread_switch(struct task_struct *next);
next              131 arch/arm64/include/asm/hw_breakpoint.h static inline void hw_breakpoint_thread_switch(struct task_struct *next)
next               29 arch/arm64/include/asm/mmu_context.h static inline void contextidr_thread_switch(struct task_struct *next)
next               34 arch/arm64/include/asm/mmu_context.h 	write_sysreg(task_pid_nr(next), contextidr_el1);
next              213 arch/arm64/include/asm/mmu_context.h static inline void __switch_mm(struct mm_struct *next)
next              221 arch/arm64/include/asm/mmu_context.h 	if (next == &init_mm) {
next              226 arch/arm64/include/asm/mmu_context.h 	check_and_switch_context(next, cpu);
next              230 arch/arm64/include/asm/mmu_context.h switch_mm(struct mm_struct *prev, struct mm_struct *next,
next              233 arch/arm64/include/asm/mmu_context.h 	if (prev != next)
next              234 arch/arm64/include/asm/mmu_context.h 		__switch_mm(next);
next              242 arch/arm64/include/asm/mmu_context.h 	update_saved_ttbr0(tsk, next);
next              246 arch/arm64/include/asm/mmu_context.h #define activate_mm(prev,next)	switch_mm(prev, next, current)
next              252 arch/arm64/include/asm/processor.h 					 struct task_struct *next);
next              120 arch/arm64/kernel/asm-offsets.c   DEFINE(HIBERN_PBE_NEXT,	offsetof(struct pbe, next));
next              202 arch/arm64/kernel/cpuinfo.c 	.next	= c_next,
next              983 arch/arm64/kernel/fpsimd.c void fpsimd_thread_switch(struct task_struct *next)
next             1001 arch/arm64/kernel/fpsimd.c 					&next->thread.uw.fpsimd_state;
next             1002 arch/arm64/kernel/fpsimd.c 	wrong_cpu = next->thread.fpsimd_cpu != smp_processor_id();
next             1004 arch/arm64/kernel/fpsimd.c 	update_tsk_thread_flag(next, TIF_FOREIGN_FPSTATE,
next              395 arch/arm64/kernel/hibernate.c 	unsigned long next;
next              410 arch/arm64/kernel/hibernate.c 		next = pmd_addr_end(addr, end);
next              414 arch/arm64/kernel/hibernate.c 			if (copy_pte(dst_pmdp, src_pmdp, addr, next))
next              420 arch/arm64/kernel/hibernate.c 	} while (dst_pmdp++, src_pmdp++, addr = next, addr != end);
next              430 arch/arm64/kernel/hibernate.c 	unsigned long next;
next              445 arch/arm64/kernel/hibernate.c 		next = pud_addr_end(addr, end);
next              449 arch/arm64/kernel/hibernate.c 			if (copy_pmd(dst_pudp, src_pudp, addr, next))
next              455 arch/arm64/kernel/hibernate.c 	} while (dst_pudp++, src_pudp++, addr = next, addr != end);
next              463 arch/arm64/kernel/hibernate.c 	unsigned long next;
next              469 arch/arm64/kernel/hibernate.c 		next = pgd_addr_end(addr, end);
next              472 arch/arm64/kernel/hibernate.c 		if (copy_pud(dst_pgdp, src_pgdp, addr, next))
next              474 arch/arm64/kernel/hibernate.c 	} while (dst_pgdp++, src_pgdp++, addr = next, addr != end);
next              901 arch/arm64/kernel/hw_breakpoint.c void hw_breakpoint_thread_switch(struct task_struct *next)
next              915 arch/arm64/kernel/hw_breakpoint.c 	next_debug_info = &next->thread.debug;
next              431 arch/arm64/kernel/process.c static void tls_thread_switch(struct task_struct *next)
next              435 arch/arm64/kernel/process.c 	if (is_compat_thread(task_thread_info(next)))
next              436 arch/arm64/kernel/process.c 		write_sysreg(next->thread.uw.tp_value, tpidrro_el0);
next              440 arch/arm64/kernel/process.c 	write_sysreg(*task_user_tls(next), tpidr_el0);
next              444 arch/arm64/kernel/process.c void uao_thread_switch(struct task_struct *next)
next              447 arch/arm64/kernel/process.c 		if (task_thread_info(next)->addr_limit == KERNEL_DS)
next              458 arch/arm64/kernel/process.c static void ssbs_thread_switch(struct task_struct *next)
next              460 arch/arm64/kernel/process.c 	struct pt_regs *regs = task_pt_regs(next);
next              466 arch/arm64/kernel/process.c 	if (unlikely(next->flags & PF_KTHREAD))
next              478 arch/arm64/kernel/process.c 	    test_tsk_thread_flag(next, TIF_SSBD))
next              496 arch/arm64/kernel/process.c static void entry_task_switch(struct task_struct *next)
next              498 arch/arm64/kernel/process.c 	__this_cpu_write(__entry_task, next);
next              505 arch/arm64/kernel/process.c 				struct task_struct *next)
next              509 arch/arm64/kernel/process.c 	fpsimd_thread_switch(next);
next              510 arch/arm64/kernel/process.c 	tls_thread_switch(next);
next              511 arch/arm64/kernel/process.c 	hw_breakpoint_thread_switch(next);
next              512 arch/arm64/kernel/process.c 	contextidr_thread_switch(next);
next              513 arch/arm64/kernel/process.c 	entry_task_switch(next);
next              514 arch/arm64/kernel/process.c 	uao_thread_switch(next);
next              515 arch/arm64/kernel/process.c 	ptrauth_thread_switch(next);
next              516 arch/arm64/kernel/process.c 	ssbs_thread_switch(next);
next              527 arch/arm64/kernel/process.c 	last = cpu_switch_to(prev, next);
next              629 arch/arm64/kernel/smp.c 			goto next;
next              634 arch/arm64/kernel/smp.c 			goto next;
next              647 arch/arm64/kernel/smp.c 				goto next;
next              663 arch/arm64/kernel/smp.c 			goto next;
next              669 arch/arm64/kernel/smp.c next:
next              309 arch/arm64/mm/dump.c 	unsigned long next, addr = start;
next              314 arch/arm64/mm/dump.c 		next = pmd_addr_end(addr, end);
next              320 arch/arm64/mm/dump.c 			walk_pte(st, pmdp, addr, next);
next              322 arch/arm64/mm/dump.c 	} while (pmdp++, addr = next, addr != end);
next              328 arch/arm64/mm/dump.c 	unsigned long next, addr = start;
next              333 arch/arm64/mm/dump.c 		next = pud_addr_end(addr, end);
next              339 arch/arm64/mm/dump.c 			walk_pmd(st, pudp, addr, next);
next              341 arch/arm64/mm/dump.c 	} while (pudp++, addr = next, addr != end);
next              348 arch/arm64/mm/dump.c 	unsigned long next, addr = start;
next              353 arch/arm64/mm/dump.c 		next = pgd_addr_end(addr, end);
next              359 arch/arm64/mm/dump.c 			walk_pud(st, pgdp, addr, next);
next              361 arch/arm64/mm/dump.c 	} while (pgdp++, addr = next, addr != end);
next              103 arch/arm64/mm/kasan_init.c 	unsigned long next;
next              112 arch/arm64/mm/kasan_init.c 		next = addr + PAGE_SIZE;
next              114 arch/arm64/mm/kasan_init.c 	} while (ptep++, addr = next, addr != end && pte_none(READ_ONCE(*ptep)));
next              120 arch/arm64/mm/kasan_init.c 	unsigned long next;
next              124 arch/arm64/mm/kasan_init.c 		next = pmd_addr_end(addr, end);
next              125 arch/arm64/mm/kasan_init.c 		kasan_pte_populate(pmdp, addr, next, node, early);
next              126 arch/arm64/mm/kasan_init.c 	} while (pmdp++, addr = next, addr != end && pmd_none(READ_ONCE(*pmdp)));
next              132 arch/arm64/mm/kasan_init.c 	unsigned long next;
next              136 arch/arm64/mm/kasan_init.c 		next = pud_addr_end(addr, end);
next              137 arch/arm64/mm/kasan_init.c 		kasan_pmd_populate(pudp, addr, next, node, early);
next              138 arch/arm64/mm/kasan_init.c 	} while (pudp++, addr = next, addr != end && pud_none(READ_ONCE(*pudp)));
next              144 arch/arm64/mm/kasan_init.c 	unsigned long next;
next              149 arch/arm64/mm/kasan_init.c 		next = pgd_addr_end(addr, end);
next              150 arch/arm64/mm/kasan_init.c 		kasan_pud_populate(pgdp, addr, next, node, early);
next              151 arch/arm64/mm/kasan_init.c 	} while (pgdp++, addr = next, addr != end);
next              170 arch/arm64/mm/mmu.c 	unsigned long next;
next              186 arch/arm64/mm/mmu.c 		next = pte_cont_addr_end(addr, end);
next              189 arch/arm64/mm/mmu.c 		if ((((addr | next | phys) & ~CONT_PTE_MASK) == 0) &&
next              193 arch/arm64/mm/mmu.c 		init_pte(pmdp, addr, next, phys, __prot);
next              195 arch/arm64/mm/mmu.c 		phys += next - addr;
next              196 arch/arm64/mm/mmu.c 	} while (addr = next, addr != end);
next              203 arch/arm64/mm/mmu.c 	unsigned long next;
next              210 arch/arm64/mm/mmu.c 		next = pmd_addr_end(addr, end);
next              213 arch/arm64/mm/mmu.c 		if (((addr | next | phys) & ~SECTION_MASK) == 0 &&
next              224 arch/arm64/mm/mmu.c 			alloc_init_cont_pte(pmdp, addr, next, phys, prot,
next              230 arch/arm64/mm/mmu.c 		phys += next - addr;
next              231 arch/arm64/mm/mmu.c 	} while (pmdp++, addr = next, addr != end);
next              241 arch/arm64/mm/mmu.c 	unsigned long next;
next              260 arch/arm64/mm/mmu.c 		next = pmd_cont_addr_end(addr, end);
next              263 arch/arm64/mm/mmu.c 		if ((((addr | next | phys) & ~CONT_PMD_MASK) == 0) &&
next              267 arch/arm64/mm/mmu.c 		init_pmd(pudp, addr, next, phys, __prot, pgtable_alloc, flags);
next              269 arch/arm64/mm/mmu.c 		phys += next - addr;
next              270 arch/arm64/mm/mmu.c 	} while (addr = next, addr != end);
next              273 arch/arm64/mm/mmu.c static inline bool use_1G_block(unsigned long addr, unsigned long next,
next              279 arch/arm64/mm/mmu.c 	if (((addr | next | phys) & ~PUD_MASK) != 0)
next              290 arch/arm64/mm/mmu.c 	unsigned long next;
next              307 arch/arm64/mm/mmu.c 		next = pud_addr_end(addr, end);
next              312 arch/arm64/mm/mmu.c 		if (use_1G_block(addr, next, phys) &&
next              323 arch/arm64/mm/mmu.c 			alloc_init_cont_pmd(pudp, addr, next, phys, prot,
next              329 arch/arm64/mm/mmu.c 		phys += next - addr;
next              330 arch/arm64/mm/mmu.c 	} while (pudp++, addr = next, addr != end);
next              341 arch/arm64/mm/mmu.c 	unsigned long addr, length, end, next;
next              357 arch/arm64/mm/mmu.c 		next = pgd_addr_end(addr, end);
next              358 arch/arm64/mm/mmu.c 		alloc_init_pud(pgdp, addr, next, phys, prot, pgtable_alloc,
next              360 arch/arm64/mm/mmu.c 		phys += next - addr;
next              361 arch/arm64/mm/mmu.c 	} while (pgdp++, addr = next, addr != end);
next              740 arch/arm64/mm/mmu.c 	unsigned long next;
next              746 arch/arm64/mm/mmu.c 		next = pmd_addr_end(addr, end);
next              766 arch/arm64/mm/mmu.c 			vmemmap_verify((pte_t *)pmdp, node, addr, next);
next              767 arch/arm64/mm/mmu.c 	} while (addr = next, addr != end);
next             1024 arch/arm64/mm/mmu.c 	unsigned long next, end;
next             1035 arch/arm64/mm/mmu.c 	next = addr;
next             1038 arch/arm64/mm/mmu.c 		pmd_free_pte_page(pmdp, next);
next             1039 arch/arm64/mm/mmu.c 	} while (pmdp++, next += PMD_SIZE, next != end);
next               18 arch/c6x/include/asm/switch_to.h 			     struct thread_struct *next,
next               21 arch/c6x/include/asm/switch_to.h #define switch_to(prev, next, last)				\
next               25 arch/c6x/include/asm/switch_to.h 				     &(next)->thread, (prev));	\
next               27 arch/csky/include/asm/mmu_context.h #define activate_mm(prev,next)		switch_mm(prev, next, current)
next               36 arch/csky/include/asm/mmu_context.h switch_mm(struct mm_struct *prev, struct mm_struct *next,
next               41 arch/csky/include/asm/mmu_context.h 	if (prev != next)
next               42 arch/csky/include/asm/mmu_context.h 		check_and_switch_context(next, cpu);
next               44 arch/csky/include/asm/mmu_context.h 	TLBMISS_HANDLER_SETUP_PGD(next->pgd);
next               45 arch/csky/include/asm/mmu_context.h 	write_mmu_entryhi(next->context.asid.counter);
next               31 arch/csky/include/asm/spinlock.h 	while (lockval.tickets.next != lockval.tickets.owner)
next               73 arch/csky/include/asm/spinlock.h 	return lock.tickets.owner == lock.tickets.next;
next               85 arch/csky/include/asm/spinlock.h 	return (tickets.next - tickets.owner) > 1;
next               18 arch/csky/include/asm/spinlock_types.h 			u16 next;
next               11 arch/csky/include/asm/switch_to.h 				   struct task_struct *next)
next               14 arch/csky/include/asm/switch_to.h 	restore_from_user_fp(&next->thread.user_fp);
next               18 arch/csky/include/asm/switch_to.h 				   struct task_struct *next)
next               28 arch/csky/include/asm/switch_to.h #define switch_to(prev, next, last)					\
next               31 arch/csky/include/asm/switch_to.h 		struct task_struct *__next = (next);			\
next               33 arch/csky/include/asm/switch_to.h 		((last) = __switch_to((prev), (next)));			\
next               14 arch/csky/kernel/cpu-probe.c 	unsigned int cur, next, i;
next               24 arch/csky/kernel/cpu-probe.c 		next = mfcr("cr13");
next               27 arch/csky/kernel/cpu-probe.c 		if (cur == next)
next               30 arch/csky/kernel/cpu-probe.c 		cur = next;
next               33 arch/csky/kernel/cpu-probe.c 		if (!(next >> 28)) {
next               76 arch/csky/kernel/cpu-probe.c 	.next	= c_next,
next               35 arch/h8300/include/asm/switch_to.h #define switch_to(prev, next, last) \
next               46 arch/h8300/include/asm/switch_to.h 		  "r" (&(next->thread)),		    \
next              241 arch/h8300/kernel/ptrace_h.c 	unsigned short *next;
next              243 arch/h8300/kernel/ptrace_h.c 	next = nextpc(child, (unsigned short *)h8300_get_reg(child, PT_PC));
next              244 arch/h8300/kernel/ptrace_h.c 	child->thread.breakinfo.addr = next;
next              245 arch/h8300/kernel/ptrace_h.c 	child->thread.breakinfo.inst = *next;
next              246 arch/h8300/kernel/ptrace_h.c 	*next = BREAKINST;
next              165 arch/h8300/kernel/setup.c 	.next	= c_next,
next               55 arch/hexagon/include/asm/mmu_context.h static inline void switch_mm(struct mm_struct *prev, struct mm_struct *next,
next               64 arch/hexagon/include/asm/mmu_context.h 	if (next->context.generation < prev->context.generation) {
next               66 arch/hexagon/include/asm/mmu_context.h 			next->pgd[l1] = init_mm.pgd[l1];
next               68 arch/hexagon/include/asm/mmu_context.h 		next->context.generation = prev->context.generation;
next               71 arch/hexagon/include/asm/mmu_context.h 	__vmnewmap((void *)next->context.ptbase);
next               77 arch/hexagon/include/asm/mmu_context.h static inline void activate_mm(struct mm_struct *prev, struct mm_struct *next)
next               82 arch/hexagon/include/asm/mmu_context.h 	switch_mm(prev, next, current_thread_info()->task);
next              134 arch/hexagon/kernel/setup.c 	.next   = &c_next,
next              231 arch/ia64/hp/common/sba_iommu.c 	struct ioc	*next;		/* list of IOC's in system */
next             1787 arch/ia64/hp/common/sba_iommu.c 	ioc->next = ioc_list;
next             1847 arch/ia64/hp/common/sba_iommu.c 	for (ioc = ioc_list; ioc; ioc = ioc->next)
next             1860 arch/ia64/hp/common/sba_iommu.c 	return ioc->next;
next             1912 arch/ia64/hp/common/sba_iommu.c 	.next  = ioc_next,
next             1953 arch/ia64/hp/common/sba_iommu.c 		for (ioc = ioc_list; ioc; ioc = ioc->next)
next             1987 arch/ia64/hp/common/sba_iommu.c 	ioc_found = ioc->next;
next             2039 arch/ia64/hp/common/sba_iommu.c 	ioc->next = ioc_found;
next               38 arch/ia64/include/asm/mmu_context.h 	unsigned int next;	/* next context number to use */
next               93 arch/ia64/include/asm/mmu_context.h 		if (ia64_ctx.next >= ia64_ctx.limit) {
next               94 arch/ia64/include/asm/mmu_context.h 			ia64_ctx.next = find_next_zero_bit(ia64_ctx.bitmap,
next               95 arch/ia64/include/asm/mmu_context.h 					ia64_ctx.max_ctx, ia64_ctx.next);
next               97 arch/ia64/include/asm/mmu_context.h 					ia64_ctx.max_ctx, ia64_ctx.next);
next               98 arch/ia64/include/asm/mmu_context.h 			if (ia64_ctx.next >= ia64_ctx.max_ctx)
next              101 arch/ia64/include/asm/mmu_context.h 		mm->context = context = ia64_ctx.next++;
next              187 arch/ia64/include/asm/mmu_context.h activate_mm (struct mm_struct *prev, struct mm_struct *next)
next              193 arch/ia64/include/asm/mmu_context.h 	ia64_set_kr(IA64_KR_PT_BASE, __pa(next->pgd));
next              194 arch/ia64/include/asm/mmu_context.h 	activate_context(next);
next               45 arch/ia64/include/asm/switch_to.h #define __switch_to(prev,next,last) do {							 \
next               48 arch/ia64/include/asm/switch_to.h 	if (IA64_HAS_EXTRA_STATE(next))								 \
next               49 arch/ia64/include/asm/switch_to.h 		ia64_load_extra(next);								 \
next               50 arch/ia64/include/asm/switch_to.h 	ia64_psr(task_pt_regs(next))->dfh = !ia64_is_local_fpu_owner(next);			 \
next               51 arch/ia64/include/asm/switch_to.h 	(last) = ia64_switch_to((next));							 \
next               61 arch/ia64/include/asm/switch_to.h # define switch_to(prev,next,last) do {						\
next               67 arch/ia64/include/asm/switch_to.h 	__switch_to(prev, next, last);						\
next               76 arch/ia64/include/asm/switch_to.h # define switch_to(prev,next,last)	__switch_to(prev, next, last)
next              527 arch/ia64/kernel/iosapic.c 		list_move(iosapic_intr_info[irq].rtes.next,
next              224 arch/ia64/kernel/mca.c 			unsigned long next = (mlogbuf_end + 1) % MLOGBUF_SIZE;
next              225 arch/ia64/kernel/mca.c 			if (next != mlogbuf_start) {
next              227 arch/ia64/kernel/mca.c 				mlogbuf_end = next;
next              108 arch/ia64/kernel/mca_drv.h 	(((head)->next != (head)) ? list_entry((head)->next, typeof(slidx_list_t), list) : NULL)
next              766 arch/ia64/kernel/perfmon.c 	int idx, next;
next              768 arch/ia64/kernel/perfmon.c 	next = (ctx->ctx_msgq_tail+1) % PFM_MAX_MSGS;
next              771 arch/ia64/kernel/perfmon.c 	if (next == ctx->ctx_msgq_head) return NULL;
next              774 arch/ia64/kernel/perfmon.c 	ctx->ctx_msgq_tail = next;
next             5657 arch/ia64/kernel/perfmon.c  	.next =		pfm_proc_next,
next              764 arch/ia64/kernel/setup.c 	.next =		c_next,
next              626 arch/ia64/kernel/unwind.c 	sr->curr.next = rs;
next              632 arch/ia64/kernel/unwind.c 	struct unw_reg_state *rs = sr->curr.next;
next              656 arch/ia64/kernel/unwind.c 			prev->next = copy;
next              659 arch/ia64/kernel/unwind.c 		rs = rs->next;
next              669 arch/ia64/kernel/unwind.c 	struct unw_reg_state *p, *next;
next              671 arch/ia64/kernel/unwind.c 	for (p = rs->next; p != NULL; p = next) {
next              672 arch/ia64/kernel/unwind.c 		next = p->next;
next              675 arch/ia64/kernel/unwind.c 	rs->next = NULL;
next             1037 arch/ia64/kernel/unwind.c 	for (ls = sr->labeled_states; ls; ls = ls->next) {
next             1041 arch/ia64/kernel/unwind.c 			sr->curr.next = dup_state_stack(ls->saved_state.next);
next             1060 arch/ia64/kernel/unwind.c 	ls->saved_state.next = dup_state_stack(sr->curr.next);
next             1063 arch/ia64/kernel/unwind.c 	ls->next = sr->labeled_states;
next             1531 arch/ia64/kernel/unwind.c 	struct unw_labeled_state *ls, *next;
next             1564 arch/ia64/kernel/unwind.c 	for (table = unw.tables; table; table = table->next) {
next             1575 arch/ia64/kernel/unwind.c 				prev->next = table->next;
next             1576 arch/ia64/kernel/unwind.c 				table->next = unw.tables->next;
next             1577 arch/ia64/kernel/unwind.c 				unw.tables->next = table;
next             1701 arch/ia64/kernel/unwind.c 	for (ls = sr.labeled_states; ls; ls = next) {
next             1702 arch/ia64/kernel/unwind.c 		next = ls->next;
next             2119 arch/ia64/kernel/unwind.c 		table->next = unw.tables->next;
next             2120 arch/ia64/kernel/unwind.c 		unw.tables->next = table;
next             2152 arch/ia64/kernel/unwind.c 		for (prev = (struct unw_table *) &unw.tables; prev; prev = prev->next)
next             2153 arch/ia64/kernel/unwind.c 			if (prev->next == table)
next             2161 arch/ia64/kernel/unwind.c 		prev->next = table->next;
next               50 arch/ia64/kernel/unwind_i.h 	struct unw_table *next;		/* must be first member! */
next               84 arch/ia64/kernel/unwind_i.h 	struct unw_reg_state *next;		/* next (outer) element on state stack */
next               89 arch/ia64/kernel/unwind_i.h 	struct unw_labeled_state *next;		/* next labeled state (or NULL) */
next               45 arch/ia64/mm/tlb.c 	.next =	1,
next               90 arch/ia64/mm/tlb.c 	ia64_ctx.next = find_next_zero_bit(ia64_ctx.bitmap,
next               93 arch/ia64/mm/tlb.c 				ia64_ctx.max_ctx, ia64_ctx.next);
next              185 arch/m68k/emu/nfblock.c 	struct nfhd_device *dev, *next;
next              187 arch/m68k/emu/nfblock.c 	list_for_each_entry_safe(dev, next, &nfhd_list, list) {
next              111 arch/m68k/include/asm/dvma.h 	struct Linux_SBus_DMA *next;
next              140 arch/m68k/include/asm/mac_iop.h 	struct iop_msg	*next;		/* next message in queue or NULL     */
next               75 arch/m68k/include/asm/mmu_context.h static inline void switch_mm(struct mm_struct *prev, struct mm_struct *next,
next               79 arch/m68k/include/asm/mmu_context.h 	set_context(tsk->mm->context, next->pgd);
next               95 arch/m68k/include/asm/mmu_context.h #define prepare_arch_switch(next) load_ksp_mmu(next)
next              197 arch/m68k/include/asm/mmu_context.h static inline void switch_mm(struct mm_struct *prev, struct mm_struct *next,
next              283 arch/m68k/include/asm/mmu_context.h static inline void switch_mm(struct mm_struct *prev, struct mm_struct *next, struct task_struct *tsk)
next              285 arch/m68k/include/asm/mmu_context.h 	if (prev != next) {
next              287 arch/m68k/include/asm/mmu_context.h 			switch_mm_0230(next);
next              289 arch/m68k/include/asm/mmu_context.h 			switch_mm_0460(next);
next              316 arch/m68k/include/asm/mmu_context.h static inline void switch_mm(struct mm_struct *prev, struct mm_struct *next, struct task_struct *tsk)
next               31 arch/m68k/include/asm/switch_to.h #define switch_to(prev,next,last) do { \
next               33 arch/m68k/include/asm/switch_to.h   register void *_next __asm__ ("a1") = (next); \
next              206 arch/m68k/include/asm/tlbflush.h 		     goto next;
next              213 arch/m68k/include/asm/tlbflush.h 	next:
next              494 arch/m68k/kernel/setup_mm.c 	.next	= c_next,
next              218 arch/m68k/kernel/setup_no.c 	.next	= c_next,
next              417 arch/m68k/mac/iop.c 	msg = msg->next;
next              483 arch/m68k/mac/iop.c 	msg->next = NULL;
next              494 arch/m68k/mac/iop.c 		while (q->next) q = q->next;
next              495 arch/m68k/mac/iop.c 		q->next = msg;
next               66 arch/m68k/mm/kmap.c 	for (p = &iolist; (tmp = *p) ; p = &tmp->next) {
next               77 arch/m68k/mm/kmap.c 	area->next = *p;
next               89 arch/m68k/mm/kmap.c 	for (p = &iolist ; (tmp = *p) ; p = &tmp->next) {
next               91 arch/m68k/mm/kmap.c 			*p = tmp->next;
next               62 arch/m68k/mm/memory.c 	ptable_desc *dp = ptable_list.next;
next              118 arch/m68k/mm/memory.c 	} else if (ptable_list.next != dp) {
next              145 arch/m68k/sun3/sun3dvma.c 	ret = list_entry(hole_cache.next, struct hole, list);
next              120 arch/microblaze/include/asm/mmu_context_mm.h static inline void switch_mm(struct mm_struct *prev, struct mm_struct *next,
next              123 arch/microblaze/include/asm/mmu_context_mm.h 	tsk->thread.pgdir = next->pgd;
next              124 arch/microblaze/include/asm/mmu_context_mm.h 	get_mmu_context(next);
next              125 arch/microblaze/include/asm/mmu_context_mm.h 	set_context(next->context, next->pgd);
next               16 arch/microblaze/include/asm/switch_to.h 					struct thread_info *next);
next               18 arch/microblaze/include/asm/switch_to.h #define switch_to(prev, next, last)					\
next               21 arch/microblaze/include/asm/switch_to.h 					task_thread_info(next));	\
next              155 arch/microblaze/kernel/cpu/mb.c 	.next = c_next,
next              541 arch/microblaze/pci/pci-common.c 		unsigned int next = isa_hole + 1;
next              543 arch/microblaze/pci/pci-common.c 		if (next < memno)
next              545 arch/microblaze/pci/pci-common.c 				&hose->mem_resources[next],
next              546 arch/microblaze/pci/pci-common.c 				sizeof(struct resource) * (memno - next));
next              998 arch/mips/alchemy/common/clock.c 			goto next;
next             1022 arch/mips/alchemy/common/clock.c next:
next               81 arch/mips/bcm63xx/irq.c 	unsigned int *next = &i[cpu];					\
next              102 arch/mips/bcm63xx/irq.c 		unsigned int to_call = *next;				\
next              104 arch/mips/bcm63xx/irq.c 		*next = (*next + 1) & (width - 1);			\
next              110 arch/mips/cavium-octeon/executive/cvmx-bootmem.c static void cvmx_bootmem_phy_set_next(uint64_t addr, uint64_t next)
next              112 arch/mips/cavium-octeon/executive/cvmx-bootmem.c 	cvmx_write64_uint64((addr + NEXT_OFFSET) | (1ull << 63), next);
next              161 arch/mips/emma/markeins/irq.c 	   .next = NULL,
next              114 arch/mips/include/asm/mach-cavium-octeon/kernel-entry-init.h 	# This is the variable where the next core to boot is stored
next              116 arch/mips/include/asm/mach-cavium-octeon/kernel-entry-init.h 	# Get the core id of the next to be booted
next              158 arch/mips/include/asm/mmu_context.h static inline void switch_mm(struct mm_struct *prev, struct mm_struct *next,
next              166 arch/mips/include/asm/mmu_context.h 	check_switch_mmu_context(next);
next              173 arch/mips/include/asm/mmu_context.h 	cpumask_set_cpu(cpu, mm_cpumask(next));
next              188 arch/mips/include/asm/mmu_context.h #define activate_mm(prev, next)	switch_mm(prev, next, current)
next               33 arch/mips/include/asm/switch_to.h 		struct task_struct *next, struct thread_info *next_ti);
next               62 arch/mips/include/asm/switch_to.h 	next->thread.emulated_fp = 0;					\
next               88 arch/mips/include/asm/switch_to.h # define __sanitize_fcr31(next)						\
next               90 arch/mips/include/asm/switch_to.h 	unsigned long fcr31 = mask_fcr31_x(next->thread.fpu.fcr31);	\
next               94 arch/mips/include/asm/switch_to.h 		pc = (void __user *)task_pt_regs(next)->cp0_epc;	\
next               95 arch/mips/include/asm/switch_to.h 		next->thread.fpu.fcr31 &= ~fcr31;			\
next               96 arch/mips/include/asm/switch_to.h 		force_fcr31_sig(fcr31, pc, next);			\
next              100 arch/mips/include/asm/switch_to.h # define __sanitize_fcr31(next)
next              109 arch/mips/include/asm/switch_to.h #define switch_to(prev, next, last)					\
next              113 arch/mips/include/asm/switch_to.h 	if (tsk_used_math(next))					\
next              114 arch/mips/include/asm/switch_to.h 		__sanitize_fcr31(next);					\
next              117 arch/mips/include/asm/switch_to.h 		__restore_dsp(next);					\
next              126 arch/mips/include/asm/switch_to.h 		if (KSTK_STATUS(next) & ST0_CU2 &&			\
next              128 arch/mips/include/asm/switch_to.h 			cop2_restore(next);				\
next              135 arch/mips/include/asm/switch_to.h 		write_c0_userlocal(task_thread_info(next)->tp_value);	\
next              136 arch/mips/include/asm/switch_to.h 	__restore_watch(next);						\
next              137 arch/mips/include/asm/switch_to.h 	(last) = resume(prev, next, task_thread_info(next));		\
next              151 arch/mips/include/asm/txx9/tx4939.h 	__u32 next;
next              170 arch/mips/include/asm/txx9/tx4939.h 	__u32 next;
next              327 arch/mips/kernel/asm-offsets.c 	OFFSET(PBE_NEXT, pbe, next);
next               27 arch/mips/kernel/module.c 	struct mips_hi16 *next;
next              101 arch/mips/kernel/module.c 	n->next = me->arch.r_mips_hi16_list;
next              109 arch/mips/kernel/module.c 	struct mips_hi16 *next;
next              112 arch/mips/kernel/module.c 		next = l->next;
next              114 arch/mips/kernel/module.c 		l = next;
next              136 arch/mips/kernel/module.c 			struct mips_hi16 *next;
next              164 arch/mips/kernel/module.c 			next = l->next;
next              166 arch/mips/kernel/module.c 			l = next;
next              186 arch/mips/kernel/proc.c 	.next	= c_next,
next              214 arch/mips/kernel/vpe-mt.c 	t = list_entry(v->tc.next, struct tc, tc);
next              233 arch/mips/kernel/vpe-mt.c 	t = list_entry(v->tc.next, struct tc, tc);
next              220 arch/mips/kernel/vpe.c 	struct mips_hi16 *next;
next              327 arch/mips/kernel/vpe.c 	n->next = mips_hi16_list;
next              338 arch/mips/kernel/vpe.c 	struct mips_hi16 *l, *next;
next              376 arch/mips/kernel/vpe.c 			next = l->next;
next              378 arch/mips/kernel/vpe.c 			l = next;
next              395 arch/mips/kernel/vpe.c 		next = l->next;
next              397 arch/mips/kernel/vpe.c 		l = next;
next              145 arch/mips/sibyte/common/sb_tbprof.c 	u64 next = (1ULL << 40) - tb_period;
next              178 arch/mips/sibyte/common/sb_tbprof.c 	__raw_writeq(next, IOADDR(A_SCD_PERF_CNT_1));
next               54 arch/nds32/include/asm/mmu_context.h static inline void switch_mm(struct mm_struct *prev, struct mm_struct *next,
next               59 arch/nds32/include/asm/mmu_context.h 	if (!cpumask_test_and_set_cpu(cpu, mm_cpumask(next)) || prev != next) {
next               60 arch/nds32/include/asm/mmu_context.h 		check_context(next);
next               61 arch/nds32/include/asm/mmu_context.h 		cpu_switch_mm(next);
next               66 arch/nds32/include/asm/mmu_context.h #define activate_mm(prev,next)	switch_mm(prev, next, NULL)
next              205 arch/nds32/kernel/process.c struct task_struct *_switch_fpu(struct task_struct *prev, struct task_struct *next)
next              210 arch/nds32/kernel/process.c 	if (!(next->flags & PF_KTHREAD))
next              211 arch/nds32/kernel/process.c 		clear_fpu(task_pt_regs(next));
next              379 arch/nds32/kernel/setup.c 	.next = c_next,
next               54 arch/nios2/include/asm/mmu_context.h void switch_mm(struct mm_struct *prev, struct mm_struct *next,
next               66 arch/nios2/include/asm/mmu_context.h void activate_mm(struct mm_struct *prev, struct mm_struct *next);
next               17 arch/nios2/include/asm/switch_to.h #define switch_to(prev, next, last)			\
next               26 arch/nios2/include/asm/switch_to.h 		: "r" (prev), "r" (next)		\
next              190 arch/nios2/kernel/cpuinfo.c 	.next	= cpuinfo_next,
next               80 arch/nios2/mm/mmu_context.c void switch_mm(struct mm_struct *prev, struct mm_struct *next,
next               89 arch/nios2/mm/mmu_context.c 	if (unlikely(CTX_VERSION(next->context) !=
next               91 arch/nios2/mm/mmu_context.c 		next->context = get_new_context();
next               94 arch/nios2/mm/mmu_context.c 	pgd_current = next->pgd;
next               97 arch/nios2/mm/mmu_context.c 	set_context(next->context);
next              106 arch/nios2/mm/mmu_context.c void activate_mm(struct mm_struct *prev, struct mm_struct *next)
next              108 arch/nios2/mm/mmu_context.c 	next->context = get_new_context();
next              109 arch/nios2/mm/mmu_context.c 	set_context(next->context);
next              110 arch/nios2/mm/mmu_context.c 	pgd_current = next->pgd;
next               22 arch/openrisc/include/asm/mmu_context.h extern void switch_mm(struct mm_struct *prev, struct mm_struct *next,
next               27 arch/openrisc/include/asm/mmu_context.h #define activate_mm(prev, next) switch_mm((prev), (next), NULL)
next               27 arch/openrisc/kernel/dma.c 		 unsigned long next, struct mm_walk *walk)
next               41 arch/openrisc/kernel/dma.c 	for (cl = __pa(addr); cl < __pa(next); cl += cpuinfo->dcache_block_size)
next               53 arch/openrisc/kernel/dma.c 		   unsigned long next, struct mm_walk *walk)
next              403 arch/openrisc/kernel/setup.c 	.next = c_next,
next              138 arch/openrisc/mm/tlb.c void switch_mm(struct mm_struct *prev, struct mm_struct *next,
next              147 arch/openrisc/mm/tlb.c 	current_pgd[smp_processor_id()] = next->pgd;
next              153 arch/openrisc/mm/tlb.c 	if (prev != next)
next               54 arch/parisc/include/asm/mmu_context.h 		struct mm_struct *next, struct task_struct *tsk)
next               56 arch/parisc/include/asm/mmu_context.h 	if (prev != next) {
next               57 arch/parisc/include/asm/mmu_context.h 		mtctl(__pa(next->pgd), 25);
next               58 arch/parisc/include/asm/mmu_context.h 		load_context(next->context);
next               63 arch/parisc/include/asm/mmu_context.h 		struct mm_struct *next, struct task_struct *tsk)
next               67 arch/parisc/include/asm/mmu_context.h 	if (prev == next)
next               71 arch/parisc/include/asm/mmu_context.h 	switch_mm_irqs_off(prev, next, tsk);
next               78 arch/parisc/include/asm/mmu_context.h static inline void activate_mm(struct mm_struct *prev, struct mm_struct *next)
next               88 arch/parisc/include/asm/mmu_context.h 	BUG_ON(next == &init_mm); /* Should never happen */
next               90 arch/parisc/include/asm/mmu_context.h 	if (next->context == 0)
next               91 arch/parisc/include/asm/mmu_context.h 	    next->context = alloc_sid();
next               93 arch/parisc/include/asm/mmu_context.h 	switch_mm(prev,next,current);
next               34 arch/parisc/include/asm/parisc-device.h 	struct parisc_driver *next;
next               74 arch/parisc/include/asm/ropes.h 	struct sba_device	*next;  /* list of SBA's in system */
next                9 arch/parisc/include/asm/switch_to.h #define switch_to(prev, next, last) do {			\
next               10 arch/parisc/include/asm/switch_to.h 	(last) = _switch_to(prev, next);			\
next              230 arch/parisc/kernel/irq.c 		while ((action = action->next))
next              233 arch/parisc/kernel/irq.c 		for ( ;action; action = action->next) {
next              192 arch/parisc/kernel/setup.c 	.next	= c_next,
next               29 arch/powerpc/include/asm/bootx.h 	u32	next;
next               45 arch/powerpc/include/asm/bootx.h 	u32	next;
next              185 arch/powerpc/include/asm/iommu.h 	struct list_head next;
next               43 arch/powerpc/include/asm/kvm_book3s_64.h 	struct kvm_nested_guest *next;
next              107 arch/powerpc/include/asm/kvm_book3s_64.h 			 (pos)->list.next)), true);			       \
next              188 arch/powerpc/include/asm/kvm_host.h 	struct list_head next;
next                5 arch/powerpc/include/asm/membarrier.h 					     struct mm_struct *next,
next               15 arch/powerpc/include/asm/membarrier.h 	if (likely(!(atomic_read(&next->membarrier_state) &
next               62 arch/powerpc/include/asm/mmu_context.h 				      struct mm_struct *next);
next               64 arch/powerpc/include/asm/mmu_context.h 				      struct mm_struct *next,
next               68 arch/powerpc/include/asm/mmu_context.h 		return radix__switch_mmu_context(prev, next);
next               69 arch/powerpc/include/asm/mmu_context.h 	return switch_slb(tsk, next);
next              104 arch/powerpc/include/asm/mmu_context.h extern void switch_mmu_context(struct mm_struct *prev, struct mm_struct *next,
next              129 arch/powerpc/include/asm/mmu_context.h extern void switch_cop(struct mm_struct *next);
next              196 arch/powerpc/include/asm/mmu_context.h extern void switch_mm_irqs_off(struct mm_struct *prev, struct mm_struct *next,
next              199 arch/powerpc/include/asm/mmu_context.h static inline void switch_mm(struct mm_struct *prev, struct mm_struct *next,
next              205 arch/powerpc/include/asm/mmu_context.h 	switch_mm_irqs_off(prev, next, tsk);
next              217 arch/powerpc/include/asm/mmu_context.h static inline void activate_mm(struct mm_struct *prev, struct mm_struct *next)
next              219 arch/powerpc/include/asm/mmu_context.h 	switch_mm(prev, next, current);
next              332 arch/powerpc/include/asm/mpic.h 	struct mpic		*next;
next              982 arch/powerpc/include/asm/opal-api.h 	__be64 next;
next               16 arch/powerpc/include/asm/switch_to.h #define switch_to(prev, next, last)	((last) = __switch_to((prev), (next)))
next               19 arch/powerpc/include/asm/switch_to.h 				   struct thread_struct *next);
next              371 arch/powerpc/kernel/asm-offsets.c 	OFFSET(pbe_next, pbe, next);
next               41 arch/powerpc/kernel/cacheinfo.c 	struct cache_index_dir *next; /* next index in parent directory */
next              192 arch/powerpc/kernel/cacheinfo.c 	struct cache *next = cache;
next              194 arch/powerpc/kernel/cacheinfo.c 	while (next) {
next              195 arch/powerpc/kernel/cacheinfo.c 		WARN_ONCE(cpumask_test_cpu(cpu, &next->shared_cpu_map),
next              197 arch/powerpc/kernel/cacheinfo.c 			  cpu, next->ofnode,
next              198 arch/powerpc/kernel/cacheinfo.c 			  cache_type_string(next));
next              199 arch/powerpc/kernel/cacheinfo.c 		cpumask_set_cpu(cpu, &next->shared_cpu_map);
next              200 arch/powerpc/kernel/cacheinfo.c 		next = next->next_local;
next              768 arch/powerpc/kernel/cacheinfo.c 	index_dir->next = cache_dir->index;
next              832 arch/powerpc/kernel/cacheinfo.c 		struct cache_index_dir *next;
next              834 arch/powerpc/kernel/cacheinfo.c 		next = index->next;
next              836 arch/powerpc/kernel/cacheinfo.c 		index = next;
next              855 arch/powerpc/kernel/cacheinfo.c 		struct cache *next = cache->next_local;
next              869 arch/powerpc/kernel/cacheinfo.c 		cache = next;
next               52 arch/powerpc/kernel/eeh_event.c 			event = list_entry(eeh_eventlist.next,
next              184 arch/powerpc/kernel/eeh_pe.c 	struct list_head *next = pe->child_list.next;
next              186 arch/powerpc/kernel/eeh_pe.c 	if (next == &pe->child_list) {
next              190 arch/powerpc/kernel/eeh_pe.c 			next = pe->child.next;
next              191 arch/powerpc/kernel/eeh_pe.c 			if (next != &pe->parent->child_list)
next              197 arch/powerpc/kernel/eeh_pe.c 	return list_entry(next, struct eeh_pe, child);
next              444 arch/powerpc/kernel/pci_dn.c 	struct list_head *next = pdn->child_list.next;
next              446 arch/powerpc/kernel/pci_dn.c 	if (next != &pdn->child_list)
next              447 arch/powerpc/kernel/pci_dn.c 		return list_entry(next, struct pci_dn, list);
next              453 arch/powerpc/kernel/pci_dn.c 		next = pdn->list.next;
next              454 arch/powerpc/kernel/pci_dn.c 		if (next != &pdn->parent->child_list)
next              460 arch/powerpc/kernel/pci_dn.c 	return list_entry(next, struct pci_dn, list);
next              109 arch/powerpc/kernel/rtas_flash.c 	struct flash_block_list *next;
next              171 arch/powerpc/kernel/rtas_flash.c 	for (f = flist; f; f = f->next) {
next              196 arch/powerpc/kernel/rtas_flash.c 	struct flash_block_list *next;
next              202 arch/powerpc/kernel/rtas_flash.c 		next = f->next;
next              204 arch/powerpc/kernel/rtas_flash.c 		f = next;
next              334 arch/powerpc/kernel/rtas_flash.c 	while (fl->next)
next              335 arch/powerpc/kernel/rtas_flash.c 		fl = fl->next; /* seek to last block_list for append */
next              339 arch/powerpc/kernel/rtas_flash.c 		fl->next = kmem_cache_zalloc(flash_block_cache, GFP_KERNEL);
next              340 arch/powerpc/kernel/rtas_flash.c 		if (!fl->next)
next              342 arch/powerpc/kernel/rtas_flash.c 		fl = fl->next;
next              560 arch/powerpc/kernel/rtas_flash.c 	struct flash_block_list *f, *next, *flist;
next              595 arch/powerpc/kernel/rtas_flash.c 	flist->next = rtas_firmware_flash_list;
next              607 arch/powerpc/kernel/rtas_flash.c 	for (f = flist; f; f = next) {
next              614 arch/powerpc/kernel/rtas_flash.c 		next = f->next;
next              616 arch/powerpc/kernel/rtas_flash.c 		if (f->next)
next              617 arch/powerpc/kernel/rtas_flash.c 			f->next = (struct flash_block_list *)cpu_to_be64(__pa(f->next));
next              619 arch/powerpc/kernel/rtas_flash.c 			f->next = NULL;
next              367 arch/powerpc/kernel/setup-common.c 	.next	= c_next,
next               64 arch/powerpc/kvm/book3s_64_vio.c 	list_del_rcu(&stit->next);
next               83 arch/powerpc/kvm/book3s_64_vio.c 		list_for_each_entry_safe(stit, tmp, &stt->iommu_tables, next) {
next              147 arch/powerpc/kvm/book3s_64_vio.c 	list_for_each_entry_rcu(stit, &stt->iommu_tables, next) {
next              172 arch/powerpc/kvm/book3s_64_vio.c 	list_add_rcu(&stit->next, &stt->iommu_tables);
next              248 arch/powerpc/kvm/book3s_64_vio.c 	list_for_each_entry_safe(stit, tmp, &stt->iommu_tables, next) {
next              367 arch/powerpc/kvm/book3s_64_vio.c 	list_for_each_entry_rcu(stit, &stt->iommu_tables, next) {
next              575 arch/powerpc/kvm/book3s_64_vio.c 	list_for_each_entry_lockless(stit, &stt->iommu_tables, next) {
next              671 arch/powerpc/kvm/book3s_64_vio.c 		list_for_each_entry_lockless(stit, &stt->iommu_tables, next) {
next              687 arch/powerpc/kvm/book3s_64_vio.c 	list_for_each_entry_lockless(stit, &stt->iommu_tables, next)
next              717 arch/powerpc/kvm/book3s_64_vio.c 	list_for_each_entry_lockless(stit, &stt->iommu_tables, next) {
next              739 arch/powerpc/kvm/book3s_64_vio.c 	list_for_each_entry_lockless(stit, &stt->iommu_tables, next)
next              123 arch/powerpc/kvm/book3s_64_vio_hv.c 	list_for_each_entry_lockless(stit, &stt->iommu_tables, next) {
next              420 arch/powerpc/kvm/book3s_64_vio_hv.c 	list_for_each_entry_lockless(stit, &stt->iommu_tables, next) {
next              574 arch/powerpc/kvm/book3s_64_vio_hv.c 		list_for_each_entry_lockless(stit, &stt->iommu_tables, next) {
next              590 arch/powerpc/kvm/book3s_64_vio_hv.c 	list_for_each_entry_lockless(stit, &stt->iommu_tables, next)
next              624 arch/powerpc/kvm/book3s_64_vio_hv.c 	list_for_each_entry_lockless(stit, &stt->iommu_tables, next) {
next              646 arch/powerpc/kvm/book3s_64_vio_hv.c 	list_for_each_entry_lockless(stit, &stt->iommu_tables, next)
next              661 arch/powerpc/kvm/book3s_hv_nested.c 			gp->next = freelist;
next              668 arch/powerpc/kvm/book3s_hv_nested.c 		freelist = gp->next;
next              785 arch/powerpc/kvm/book3s_hv_nested.c 		(*n_rmap)->list.next = (struct llist_node *) rmap;
next              164 arch/powerpc/kvm/book3s_hv_rm_mmu.c 	struct revmap_entry *next, *prev;
next              179 arch/powerpc/kvm/book3s_hv_rm_mmu.c 	next = real_vmalloc_addr(&kvm->arch.hpt.rev[rev->forw]);
next              181 arch/powerpc/kvm/book3s_hv_rm_mmu.c 	next->back = rev->back;
next              405 arch/powerpc/kvm/book3s_xive_template.c 				goto next;
next              408 arch/powerpc/kvm/book3s_xive_template.c 				goto next;
next              413 arch/powerpc/kvm/book3s_xive_template.c 				goto next;
next              431 arch/powerpc/kvm/book3s_xive_template.c 		next:
next              143 arch/powerpc/kvm/mpic.c 	int next;
next              295 arch/powerpc/kvm/mpic.c 	int next = -1;
next              307 arch/powerpc/kvm/mpic.c 			next = irq;
next              312 arch/powerpc/kvm/mpic.c 	q->next = next;
next              321 arch/powerpc/kvm/mpic.c 	return q->next;
next              387 arch/powerpc/kvm/mpic.c 				__func__, n_IRQ, dst->servicing.next, n_CPU);
next              390 arch/powerpc/kvm/mpic.c 				__func__, n_CPU, n_IRQ, dst->raised.next);
next              398 arch/powerpc/kvm/mpic.c 				__func__, n_IRQ, dst->raised.next,
next              549 arch/powerpc/kvm/mpic.c 		opp->dst[i].raised.next = -1;
next              551 arch/powerpc/kvm/mpic.c 		opp->dst[i].servicing.next = -1;
next             1067 arch/powerpc/kvm/mpic.c 				__func__, idx, dst->raised.next);
next               35 arch/powerpc/lib/rheap.c 	pp = (unsigned long *)&l->next;
next              133 arch/powerpc/lib/rheap.c 	blk = list_entry(info->empty_list.next, rh_block_t, list);
next              156 arch/powerpc/lib/rheap.c 	rh_block_t *next;
next              170 arch/powerpc/lib/rheap.c 	next = NULL;
next              178 arch/powerpc/lib/rheap.c 		if (next == NULL && s >= bs)
next              179 arch/powerpc/lib/rheap.c 			next = blk;
next              202 arch/powerpc/lib/rheap.c 		if (next != NULL)
next              203 arch/powerpc/lib/rheap.c 			list_add(&blkn->list, &next->list);
next               28 arch/powerpc/mm/book3s64/iommu_api.c 	struct list_head next;
next              150 arch/powerpc/mm/book3s64/iommu_api.c 	list_for_each_entry_rcu(mem2, &mm->context.iommu_group_mem_list, next) {
next              161 arch/powerpc/mm/book3s64/iommu_api.c 	list_add_rcu(&mem->next, &mm->context.iommu_group_mem_list);
next              241 arch/powerpc/mm/book3s64/iommu_api.c 	list_del_rcu(&mem->next);
next              289 arch/powerpc/mm/book3s64/iommu_api.c 	list_for_each_entry_rcu(mem, &mm->context.iommu_group_mem_list, next) {
next              308 arch/powerpc/mm/book3s64/iommu_api.c 			next) {
next              327 arch/powerpc/mm/book3s64/iommu_api.c 	list_for_each_entry_rcu(mem, &mm->context.iommu_group_mem_list, next) {
next              421 arch/powerpc/mm/book3s64/iommu_api.c 	list_for_each_entry_rcu(mem, &mm->context.iommu_group_mem_list, next) {
next              304 arch/powerpc/mm/book3s64/mmu_context.c void radix__switch_mmu_context(struct mm_struct *prev, struct mm_struct *next)
next              306 arch/powerpc/mm/book3s64/mmu_context.c 	mtspr(SPRN_PID, next->context.id);
next              719 arch/powerpc/mm/book3s64/radix_pgtable.c 	unsigned long next;
next              723 arch/powerpc/mm/book3s64/radix_pgtable.c 	for (; addr < end; addr = next, pte++) {
next              724 arch/powerpc/mm/book3s64/radix_pgtable.c 		next = (addr + PAGE_SIZE) & PAGE_MASK;
next              725 arch/powerpc/mm/book3s64/radix_pgtable.c 		if (next > end)
next              726 arch/powerpc/mm/book3s64/radix_pgtable.c 			next = end;
next              731 arch/powerpc/mm/book3s64/radix_pgtable.c 		if (!PAGE_ALIGNED(addr) || !PAGE_ALIGNED(next)) {
next              793 arch/powerpc/mm/book3s64/radix_pgtable.c 	unsigned long next;
next              798 arch/powerpc/mm/book3s64/radix_pgtable.c 	for (; addr < end; addr = next, pmd++) {
next              799 arch/powerpc/mm/book3s64/radix_pgtable.c 		next = pmd_addr_end(addr, end);
next              810 arch/powerpc/mm/book3s64/radix_pgtable.c 		remove_pte_table(pte_base, addr, next);
next              818 arch/powerpc/mm/book3s64/radix_pgtable.c 	unsigned long next;
next              823 arch/powerpc/mm/book3s64/radix_pgtable.c 	for (; addr < end; addr = next, pud++) {
next              824 arch/powerpc/mm/book3s64/radix_pgtable.c 		next = pud_addr_end(addr, end);
next              835 arch/powerpc/mm/book3s64/radix_pgtable.c 		remove_pmd_table(pmd_base, addr, next);
next              842 arch/powerpc/mm/book3s64/radix_pgtable.c 	unsigned long addr, next;
next              848 arch/powerpc/mm/book3s64/radix_pgtable.c 	for (addr = start; addr < end; addr = next) {
next              849 arch/powerpc/mm/book3s64/radix_pgtable.c 		next = pgd_addr_end(addr, end);
next              861 arch/powerpc/mm/book3s64/radix_pgtable.c 		remove_pud_table(pud_base, addr, next);
next              998 arch/powerpc/mm/book3s64/radix_pgtable.c 		pmd_huge_pte(mm, pmdp) = (pgtable_t) lh->next;
next               93 arch/powerpc/mm/book3s64/subpage_prot.c 	unsigned long next, limit;
next              104 arch/powerpc/mm/book3s64/subpage_prot.c 	for (; addr < limit; addr = next) {
next              105 arch/powerpc/mm/book3s64/subpage_prot.c 		next = pmd_addr_end(addr, limit);
next              120 arch/powerpc/mm/book3s64/subpage_prot.c 		if (addr + (nw << PAGE_SHIFT) > next)
next              121 arch/powerpc/mm/book3s64/subpage_prot.c 			nw = (next - addr) >> PAGE_SHIFT;
next              196 arch/powerpc/mm/book3s64/subpage_prot.c 	unsigned long next, limit;
next              237 arch/powerpc/mm/book3s64/subpage_prot.c 	for (limit = addr + len; addr < limit; addr = next) {
next              238 arch/powerpc/mm/book3s64/subpage_prot.c 		next = pmd_addr_end(addr, limit);
next              267 arch/powerpc/mm/book3s64/subpage_prot.c 		if (addr + (nw << PAGE_SHIFT) > next)
next              268 arch/powerpc/mm/book3s64/subpage_prot.c 			nw = (next - addr) >> PAGE_SHIFT;
next              345 arch/powerpc/mm/hugetlbpage.c 	unsigned long next;
next              353 arch/powerpc/mm/hugetlbpage.c 		next = pmd_addr_end(addr, end);
next              369 arch/powerpc/mm/hugetlbpage.c 		if (more > next)
next              370 arch/powerpc/mm/hugetlbpage.c 			next = more;
next              373 arch/powerpc/mm/hugetlbpage.c 				  addr, next, floor, ceiling);
next              374 arch/powerpc/mm/hugetlbpage.c 	} while (addr = next, addr != end);
next              398 arch/powerpc/mm/hugetlbpage.c 	unsigned long next;
next              404 arch/powerpc/mm/hugetlbpage.c 		next = pud_addr_end(addr, end);
next              408 arch/powerpc/mm/hugetlbpage.c 			hugetlb_free_pmd_range(tlb, pud, addr, next, floor,
next              419 arch/powerpc/mm/hugetlbpage.c 			if (more > next)
next              420 arch/powerpc/mm/hugetlbpage.c 				next = more;
next              423 arch/powerpc/mm/hugetlbpage.c 					  addr, next, floor, ceiling);
next              425 arch/powerpc/mm/hugetlbpage.c 	} while (addr = next, addr != end);
next              452 arch/powerpc/mm/hugetlbpage.c 	unsigned long next;
next              472 arch/powerpc/mm/hugetlbpage.c 		next = pgd_addr_end(addr, end);
next              477 arch/powerpc/mm/hugetlbpage.c 			hugetlb_free_pud_range(tlb, pgd, addr, next, floor, ceiling);
next              487 arch/powerpc/mm/hugetlbpage.c 			if (more > next)
next              488 arch/powerpc/mm/hugetlbpage.c 				next = more;
next              491 arch/powerpc/mm/hugetlbpage.c 					  addr, next, floor, ceiling);
next              493 arch/powerpc/mm/hugetlbpage.c 	} while (addr = next, addr != end);
next              117 arch/powerpc/mm/init_64.c static struct vmemmap_backing *next;
next              135 arch/powerpc/mm/init_64.c 		vmem_back = next;
next              136 arch/powerpc/mm/init_64.c 		next = next->list;
next              143 arch/powerpc/mm/init_64.c 		next = vmemmap_alloc_block(PAGE_SIZE, node);
next              144 arch/powerpc/mm/init_64.c 		if (unlikely(!next)) {
next              153 arch/powerpc/mm/init_64.c 	return next++;
next              264 arch/powerpc/mm/init_64.c 	vmem_back->list = next;
next              265 arch/powerpc/mm/init_64.c 	next = vmem_back;
next              207 arch/powerpc/mm/kasan/kasan_init_32.c 	unsigned long next;
next              215 arch/powerpc/mm/kasan/kasan_init_32.c 		next = pgd_addr_end(addr, end);
next              217 arch/powerpc/mm/kasan/kasan_init_32.c 	} while (pmd++, addr = next, addr != end);
next               34 arch/powerpc/mm/mmu_context.c void switch_mm_irqs_off(struct mm_struct *prev, struct mm_struct *next,
next               40 arch/powerpc/mm/mmu_context.c 	if (!cpumask_test_cpu(smp_processor_id(), mm_cpumask(next))) {
next               41 arch/powerpc/mm/mmu_context.c 		cpumask_set_cpu(smp_processor_id(), mm_cpumask(next));
next               42 arch/powerpc/mm/mmu_context.c 		inc_mm_active_cpus(next);
next               71 arch/powerpc/mm/mmu_context.c 	switch_mm_pgdir(tsk, next);
next               74 arch/powerpc/mm/mmu_context.c 	if (prev == next)
next               85 arch/powerpc/mm/mmu_context.c 		radix_kvm_prefetch_workaround(next);
next               87 arch/powerpc/mm/mmu_context.c 		membarrier_arch_switch_mm(prev, next, tsk);
next               93 arch/powerpc/mm/mmu_context.c 	switch_mmu_context(prev, next, tsk);
next               21 arch/powerpc/mm/nohash/book3e_hugetlbpage.c 	int this, next;
next               26 arch/powerpc/mm/nohash/book3e_hugetlbpage.c 	next = this + 1;
next               27 arch/powerpc/mm/nohash/book3e_hugetlbpage.c 	if (next >= tcd->esel_max)
next               28 arch/powerpc/mm/nohash/book3e_hugetlbpage.c 		next = tcd->esel_first;
next               30 arch/powerpc/mm/nohash/book3e_hugetlbpage.c 	tcd->esel_next = next;
next              261 arch/powerpc/mm/nohash/mmu_context.c void switch_mmu_context(struct mm_struct *prev, struct mm_struct *next,
next              274 arch/powerpc/mm/nohash/mmu_context.c 		cpu, next, next->context.active, next->context.id);
next              278 arch/powerpc/mm/nohash/mmu_context.c 	next->context.active++;
next              289 arch/powerpc/mm/nohash/mmu_context.c 	id = next->context.id;
next              292 arch/powerpc/mm/nohash/mmu_context.c 		if (context_mm[id] != next)
next              294 arch/powerpc/mm/nohash/mmu_context.c 			       next, id, id, context_mm[id]);
next              331 arch/powerpc/mm/nohash/mmu_context.c 	context_mm[id] = next;
next              332 arch/powerpc/mm/nohash/mmu_context.c 	next->context.id = id;
next              347 arch/powerpc/mm/nohash/mmu_context.c 		local_flush_tlb_mm(next);
next              360 arch/powerpc/mm/nohash/mmu_context.c 	set_context(id, next->pgd);
next             1091 arch/powerpc/mm/numa.c 	struct topology_update_data *next;
next             1268 arch/powerpc/mm/numa.c 	for (update = data; update; update = update->next) {
next             1296 arch/powerpc/mm/numa.c 	for (update = data; update; update = update->next) {
next             1368 arch/powerpc/mm/numa.c 			ud->next = &updates[i];
next             1382 arch/powerpc/mm/numa.c 		updates[i-1].next = NULL;
next             1386 arch/powerpc/mm/numa.c 		for (ud = &updates[0]; ud; ud = ud->next) {
next             1425 arch/powerpc/mm/numa.c 	for (ud = &updates[0]; ud; ud = ud->next) {
next               41 arch/powerpc/oprofile/cell/pr_util.h 	struct vma_to_fileoffset_map *next;	/* list pointer */
next               26 arch/powerpc/oprofile/cell/vma_map.c 		struct vma_to_fileoffset_map *next = map->next;
next               28 arch/powerpc/oprofile/cell/vma_map.c 		map = next;
next               46 arch/powerpc/oprofile/cell/vma_map.c 	for (; map; map = map->next) {
next               77 arch/powerpc/oprofile/cell/vma_map.c 	new->next = map;
next             1183 arch/powerpc/perf/hv-24x7.c 				 struct hv_24x7_result **next)
next             1199 arch/powerpc/perf/hv-24x7.c 		if (next)
next             1200 arch/powerpc/perf/hv-24x7.c 			*next = (struct hv_24x7_result *) res->elements;
next             1241 arch/powerpc/perf/hv-24x7.c 	if (next)
next             1242 arch/powerpc/perf/hv-24x7.c 		*next = element_data - data_offset;
next              108 arch/powerpc/platforms/83xx/mpc832x_rdb.c 		goto next;
next              113 arch/powerpc/platforms/83xx/mpc832x_rdb.c next:
next              554 arch/powerpc/platforms/cell/iommu.c 	window = list_entry(iommu->windows.next, struct iommu_window, list);
next              309 arch/powerpc/platforms/cell/spufs/inode.c 	aff_supp = !list_empty(&(list_entry(cbe_spu_info[0].spus.next,
next              336 arch/powerpc/platforms/cell/spufs/inode.c 		    !list_entry(neighbor->aff_list.next, struct spu_context,
next              387 arch/powerpc/platforms/cell/spufs/inode.c 		    || list_entry(neighbor->aff_list.next, struct spu_context,
next              382 arch/powerpc/platforms/cell/spufs/sched.c 		list_for_each_entry_reverse(spu, ref->aff_list.next, aff_list) {
next             1953 arch/powerpc/platforms/cell/spufs/switch.c static void restore_lscsa(struct spu_state *next, struct spu *spu)
next             1961 arch/powerpc/platforms/cell/spufs/switch.c 	set_watchdog_timer(next, spu);	        /* Step 26. */
next             1962 arch/powerpc/platforms/cell/spufs/switch.c 	setup_spu_status_part1(next, spu);	/* Step 27. */
next             1963 arch/powerpc/platforms/cell/spufs/switch.c 	setup_spu_status_part2(next, spu);	/* Step 28. */
next             1964 arch/powerpc/platforms/cell/spufs/switch.c 	restore_mfc_rag(next, spu);	        /* Step 29. */
next             1966 arch/powerpc/platforms/cell/spufs/switch.c 	setup_mfc_slbs(next, spu, spu_restore_code, sizeof(spu_restore_code));
next             1967 arch/powerpc/platforms/cell/spufs/switch.c 	set_spu_npc(next, spu);	                /* Step 31. */
next             1968 arch/powerpc/platforms/cell/spufs/switch.c 	set_signot1(next, spu);	                /* Step 32. */
next             1969 arch/powerpc/platforms/cell/spufs/switch.c 	set_signot2(next, spu);	                /* Step 33. */
next             1970 arch/powerpc/platforms/cell/spufs/switch.c 	setup_decr(next, spu);	                /* Step 34. */
next             1971 arch/powerpc/platforms/cell/spufs/switch.c 	setup_ppu_mb(next, spu);	        /* Step 35. */
next             1972 arch/powerpc/platforms/cell/spufs/switch.c 	setup_ppuint_mb(next, spu);	        /* Step 36. */
next             1973 arch/powerpc/platforms/cell/spufs/switch.c 	send_restore_code(next, spu);	        /* Step 37. */
next             1974 arch/powerpc/platforms/cell/spufs/switch.c 	set_ppu_querymask(next, spu);	        /* Step 38. */
next             1975 arch/powerpc/platforms/cell/spufs/switch.c 	wait_tag_complete(next, spu);	        /* Step 39. */
next             1976 arch/powerpc/platforms/cell/spufs/switch.c 	wait_spu_stopped(next, spu);	        /* Step 40. */
next             1979 arch/powerpc/platforms/cell/spufs/switch.c static void restore_csa(struct spu_state *next, struct spu *spu)
next             1986 arch/powerpc/platforms/cell/spufs/switch.c 	restore_spu_privcntl(next, spu);	/* Step 41. */
next             1987 arch/powerpc/platforms/cell/spufs/switch.c 	restore_status_part1(next, spu);	/* Step 42. */
next             1988 arch/powerpc/platforms/cell/spufs/switch.c 	restore_status_part2(next, spu);	/* Step 43. */
next             1989 arch/powerpc/platforms/cell/spufs/switch.c 	restore_ls_16kb(next, spu);	        /* Step 44. */
next             1990 arch/powerpc/platforms/cell/spufs/switch.c 	wait_tag_complete(next, spu);	        /* Step 45. */
next             1991 arch/powerpc/platforms/cell/spufs/switch.c 	suspend_mfc(next, spu);	                /* Step 46. */
next             1992 arch/powerpc/platforms/cell/spufs/switch.c 	wait_suspend_mfc_complete(next, spu);	/* Step 47. */
next             1993 arch/powerpc/platforms/cell/spufs/switch.c 	issue_mfc_tlbie(next, spu);	        /* Step 48. */
next             1994 arch/powerpc/platforms/cell/spufs/switch.c 	clear_interrupts(next, spu);	        /* Step 49. */
next             1995 arch/powerpc/platforms/cell/spufs/switch.c 	restore_mfc_queues(next, spu);	        /* Step 50. */
next             1996 arch/powerpc/platforms/cell/spufs/switch.c 	restore_ppu_querymask(next, spu);	/* Step 51. */
next             1997 arch/powerpc/platforms/cell/spufs/switch.c 	restore_ppu_querytype(next, spu);	/* Step 52. */
next             1998 arch/powerpc/platforms/cell/spufs/switch.c 	restore_mfc_csr_tsq(next, spu);	        /* Step 53. */
next             1999 arch/powerpc/platforms/cell/spufs/switch.c 	restore_mfc_csr_cmd(next, spu);	        /* Step 54. */
next             2000 arch/powerpc/platforms/cell/spufs/switch.c 	restore_mfc_csr_ato(next, spu);	        /* Step 55. */
next             2001 arch/powerpc/platforms/cell/spufs/switch.c 	restore_mfc_tclass_id(next, spu);	/* Step 56. */
next             2002 arch/powerpc/platforms/cell/spufs/switch.c 	set_llr_event(next, spu);	        /* Step 57. */
next             2003 arch/powerpc/platforms/cell/spufs/switch.c 	restore_decr_wrapped(next, spu);	/* Step 58. */
next             2004 arch/powerpc/platforms/cell/spufs/switch.c 	restore_ch_part1(next, spu);	        /* Step 59. */
next             2005 arch/powerpc/platforms/cell/spufs/switch.c 	restore_ch_part2(next, spu);	        /* Step 60. */
next             2006 arch/powerpc/platforms/cell/spufs/switch.c 	restore_spu_lslr(next, spu);	        /* Step 61. */
next             2007 arch/powerpc/platforms/cell/spufs/switch.c 	restore_spu_cfg(next, spu);	        /* Step 62. */
next             2008 arch/powerpc/platforms/cell/spufs/switch.c 	restore_pm_trace(next, spu);	        /* Step 63. */
next             2009 arch/powerpc/platforms/cell/spufs/switch.c 	restore_spu_npc(next, spu);	        /* Step 64. */
next             2010 arch/powerpc/platforms/cell/spufs/switch.c 	restore_spu_mb(next, spu);	        /* Step 65. */
next             2011 arch/powerpc/platforms/cell/spufs/switch.c 	check_ppu_mb_stat(next, spu);	        /* Step 66. */
next             2012 arch/powerpc/platforms/cell/spufs/switch.c 	check_ppuint_mb_stat(next, spu);	/* Step 67. */
next             2014 arch/powerpc/platforms/cell/spufs/switch.c 	restore_mfc_sr1(next, spu);	        /* Step 69. */
next             2015 arch/powerpc/platforms/cell/spufs/switch.c 	set_int_route(next, spu);		/* NEW      */
next             2016 arch/powerpc/platforms/cell/spufs/switch.c 	restore_other_spu_access(next, spu);	/* Step 70. */
next             2017 arch/powerpc/platforms/cell/spufs/switch.c 	restore_spu_runcntl(next, spu);	        /* Step 71. */
next             2018 arch/powerpc/platforms/cell/spufs/switch.c 	restore_mfc_cntl(next, spu);	        /* Step 72. */
next             2019 arch/powerpc/platforms/cell/spufs/switch.c 	enable_user_access(next, spu);	        /* Step 73. */
next             2020 arch/powerpc/platforms/cell/spufs/switch.c 	reset_switch_active(next, spu);	        /* Step 74. */
next             2021 arch/powerpc/platforms/cell/spufs/switch.c 	reenable_interrupts(next, spu);	        /* Step 75. */
next             2056 arch/powerpc/platforms/cell/spufs/switch.c static int __do_spu_restore(struct spu_state *next, struct spu *spu)
next             2071 arch/powerpc/platforms/cell/spufs/switch.c 	restore_lscsa(next, spu);	        /* Steps 24-39. */
next             2072 arch/powerpc/platforms/cell/spufs/switch.c 	rc = check_restore_status(next, spu);	/* Step 40.     */
next             2082 arch/powerpc/platforms/cell/spufs/switch.c 	restore_csa(next, spu);
next              104 arch/powerpc/platforms/powermac/bootx_init.c 		ppp = &pp->next;
next              256 arch/powerpc/platforms/powermac/bootx_init.c  			goto next;
next              261 arch/powerpc/platforms/powermac/bootx_init.c 	next:
next              262 arch/powerpc/platforms/powermac/bootx_init.c 		ppp = &pp->next;
next              316 arch/powerpc/platforms/powermac/bootx_init.c  			goto next;
next              319 arch/powerpc/platforms/powermac/bootx_init.c 			goto next;
next              325 arch/powerpc/platforms/powermac/bootx_init.c 	next:
next              326 arch/powerpc/platforms/powermac/bootx_init.c 		ppp = &pp->next;
next              666 arch/powerpc/platforms/powermac/pci.c 	unsigned long base, end, next = -1;
next              682 arch/powerpc/platforms/powermac/pci.c 		if (base != next) {
next              697 arch/powerpc/platforms/powermac/pci.c 		next = end + 1;
next              688 arch/powerpc/platforms/powermac/pfunc_core.c 	for (pp = dev->node->properties; pp != 0; pp = pp->next) {
next              798 arch/powerpc/platforms/powermac/pfunc_core.c 			list_entry(dev->functions.next, typeof(*func), link);
next              242 arch/powerpc/platforms/powernv/opal-async.c 		.next		= NULL,
next              276 arch/powerpc/platforms/powernv/opal-hmi.c 		msg_node = list_entry(opal_hmi_evt_list.next,
next              357 arch/powerpc/platforms/powernv/opal-hmi.c 	.next		= NULL,
next               62 arch/powerpc/platforms/powernv/opal-memory-errors.c 		 msg_node = list_entry(opal_memory_err_list.next,
next              114 arch/powerpc/platforms/powernv/opal-memory-errors.c 	.next		= NULL,
next              119 arch/powerpc/platforms/powernv/opal-power.c 	.next		= NULL,
next              126 arch/powerpc/platforms/powernv/opal-power.c 	.next		= NULL,
next              133 arch/powerpc/platforms/powernv/opal-power.c 	.next		= NULL,
next              371 arch/powerpc/platforms/powernv/opal-prd.c 	.next		= NULL,
next             1067 arch/powerpc/platforms/powernv/opal.c 			struct opal_sg_list *next;
next             1069 arch/powerpc/platforms/powernv/opal.c 			next = kzalloc(PAGE_SIZE, GFP_KERNEL);
next             1070 arch/powerpc/platforms/powernv/opal.c 			if (!next)
next             1076 arch/powerpc/platforms/powernv/opal.c 			sg->next = cpu_to_be64(__pa(next));
next             1077 arch/powerpc/platforms/powernv/opal.c 			sg = next;
next             1097 arch/powerpc/platforms/powernv/opal.c 		uint64_t next = be64_to_cpu(sg->next);
next             1101 arch/powerpc/platforms/powernv/opal.c 		if (next)
next             1102 arch/powerpc/platforms/powernv/opal.c 			sg = __va(next);
next              363 arch/powerpc/platforms/powernv/pci-ioda-tce.c 	list_for_each_entry_rcu(tgl, &tbl->it_group_list, next) {
next              365 arch/powerpc/platforms/powernv/pci-ioda-tce.c 			list_del_rcu(&tgl->next);
next              402 arch/powerpc/platforms/powernv/pci-ioda-tce.c 	list_add_rcu(&tgl->next, &tbl->it_group_list);
next             1906 arch/powerpc/platforms/powernv/pci-ioda.c 			next);
next             2050 arch/powerpc/platforms/powernv/pci-ioda.c 	list_for_each_entry_lockless(tgl, &tbl->it_group_list, next) {
next              575 arch/powerpc/platforms/ps3/mm.c 		last  = list_entry(r->chunk_list.head.next,
next              421 arch/powerpc/platforms/ps3/os-area.c next:
next              441 arch/powerpc/platforms/ps3/os-area.c 		goto next;
next              444 arch/powerpc/platforms/ps3/os-area.c 		goto next;
next               83 arch/powerpc/platforms/pseries/cmm.c 	struct cmm_page_array *next;
next              180 arch/powerpc/platforms/pseries/cmm.c 				npa->next = pa;
next              227 arch/powerpc/platforms/pseries/cmm.c 			pa = pa->next;
next              512 arch/powerpc/platforms/pseries/cmm.c 		pa = pa->next;
next              579 arch/powerpc/platforms/pseries/cmm.c 					pa_curr = pa_last->next;
next              580 arch/powerpc/platforms/pseries/cmm.c 				pa_last = pa_last->next;
next              585 arch/powerpc/platforms/pseries/cmm.c 		pa_curr = pa_curr->next;
next              608 arch/powerpc/platforms/pseries/cmm.c 				pa_last->next = npa;
next              615 arch/powerpc/platforms/pseries/cmm.c 		pa_curr = pa_curr->next;
next              105 arch/powerpc/platforms/pseries/dlpar.c 		dn->properties = prop->next;
next              210 arch/powerpc/platforms/pseries/dlpar.c 				last_property->next = property;
next               75 arch/powerpc/platforms/pseries/hvCall_inst.c         .next  = hc_next,
next               71 arch/powerpc/platforms/pseries/hvcserver.c 		element = head->next;
next              360 arch/powerpc/platforms/pseries/iommu.c 	u64 tce_size, num_tce, dma_offset, next;
next              366 arch/powerpc/platforms/pseries/iommu.c 	next = start_pfn << PAGE_SHIFT;
next              370 arch/powerpc/platforms/pseries/iommu.c 	num_tce += next & (tce_size - 1);
next              371 arch/powerpc/platforms/pseries/iommu.c 	next &= ~(tce_size - 1);
next              383 arch/powerpc/platforms/pseries/iommu.c 		dma_offset = next + be64_to_cpu(maprange->dma_base);
next              388 arch/powerpc/platforms/pseries/iommu.c 		next += limit * tce_size;
next              399 arch/powerpc/platforms/pseries/iommu.c 	u64 tce_size, num_tce, dma_offset, next, proto_tce, liobn;
next              435 arch/powerpc/platforms/pseries/iommu.c 	next = start_pfn << PAGE_SHIFT;
next              439 arch/powerpc/platforms/pseries/iommu.c 	num_tce += next & (tce_size - 1);
next              440 arch/powerpc/platforms/pseries/iommu.c 	next &= ~(tce_size - 1);
next              453 arch/powerpc/platforms/pseries/iommu.c 		dma_offset = next + be64_to_cpu(maprange->dma_base);
next              456 arch/powerpc/platforms/pseries/iommu.c 			tcep[l] = cpu_to_be64(proto_tce | next);
next              457 arch/powerpc/platforms/pseries/iommu.c 			next += tce_size;
next               92 arch/powerpc/platforms/pseries/reconfig.c 	struct property *next;
next               93 arch/powerpc/platforms/pseries/reconfig.c 	for (; prop; prop = next) {
next               94 arch/powerpc/platforms/pseries/reconfig.c 		next = prop->next;
next              184 arch/powerpc/platforms/pseries/reconfig.c 	new->next = last;
next              573 arch/powerpc/sysdev/mpic.c 			goto next;
next              577 arch/powerpc/sysdev/mpic.c 			goto next;
next              582 arch/powerpc/sysdev/mpic.c 	next:
next             1519 arch/powerpc/sysdev/mpic.c 	mpic->next = mpics;
next             1953 arch/powerpc/sysdev/mpic.c 		mpic = mpic->next;
next             1995 arch/powerpc/sysdev/mpic.c 		mpic = mpic->next;
next               32 arch/riscv/include/asm/mmu_context.h void switch_mm(struct mm_struct *prev, struct mm_struct *next,
next               36 arch/riscv/include/asm/mmu_context.h 			       struct mm_struct *next)
next               38 arch/riscv/include/asm/mmu_context.h 	switch_mm(prev, next, NULL);
next               48 arch/riscv/include/asm/switch_to.h 				   struct task_struct *next)
next               55 arch/riscv/include/asm/switch_to.h 	fstate_restore(next, task_pt_regs(next));
next               69 arch/riscv/include/asm/switch_to.h #define switch_to(prev, next, last)			\
next               72 arch/riscv/include/asm/switch_to.h 	struct task_struct *__next = (next);		\
next              151 arch/riscv/kernel/cpu.c 	.next	= c_next,
next               43 arch/riscv/mm/context.c void switch_mm(struct mm_struct *prev, struct mm_struct *next,
next               48 arch/riscv/mm/context.c 	if (unlikely(prev == next))
next               59 arch/riscv/mm/context.c 	cpumask_set_cpu(cpu, mm_cpumask(next));
next               61 arch/riscv/mm/context.c 	csr_write(CSR_SATP, virt_to_pfn(next->pgd) | SATP_MODE);
next               64 arch/riscv/mm/context.c 	flush_icache_deferred(next);
next               34 arch/s390/include/asm/debug.h 	struct debug_info *next;
next               75 arch/s390/include/asm/gmap.h 	struct gmap_rmap *next;
next               80 arch/s390/include/asm/gmap.h 	for (pos = (head); pos; pos = pos->next)
next               83 arch/s390/include/asm/gmap.h 	for (pos = (head); n = pos ? pos->next : NULL, pos; pos = n)
next              825 arch/s390/include/asm/kvm_host.h 	int next;
next               88 arch/s390/include/asm/mmu_context.h static inline void switch_mm(struct mm_struct *prev, struct mm_struct *next,
next               93 arch/s390/include/asm/mmu_context.h 	S390_lowcore.user_asce = next->context.asce;
next               94 arch/s390/include/asm/mmu_context.h 	cpumask_set_cpu(cpu, &next->context.cpu_attach_mask);
next              104 arch/s390/include/asm/mmu_context.h 	if (prev != next)
next              129 arch/s390/include/asm/mmu_context.h                                struct mm_struct *next)
next              131 arch/s390/include/asm/mmu_context.h 	switch_mm(prev, next, current);
next              132 arch/s390/include/asm/mmu_context.h 	cpumask_set_cpu(smp_processor_id(), mm_cpumask(next));
next              133 arch/s390/include/asm/mmu_context.h 	set_user_asce(next);
next               33 arch/s390/include/asm/switch_to.h #define switch_to(prev, next, last) do {				\
next               42 arch/s390/include/asm/switch_to.h 	update_cr_regs(next);						\
next               43 arch/s390/include/asm/switch_to.h 	restore_access_regs(&next->thread.acrs[0]);			\
next               44 arch/s390/include/asm/switch_to.h 	restore_ri_cb(next->thread.ri_cb, prev->thread.ri_cb);		\
next               45 arch/s390/include/asm/switch_to.h 	restore_gs_cb(next->thread.gs_cb);				\
next               46 arch/s390/include/asm/switch_to.h 	prev = __switch_to(prev, next);					\
next              341 arch/s390/kernel/debug.c 		debug_area_last->next = rc;
next              345 arch/s390/kernel/debug.c 	rc->next = NULL;
next              418 arch/s390/kernel/debug.c 			debug_area_first = db_info->next;
next              422 arch/s390/kernel/debug.c 			db_info->prev->next = db_info->next;
next              423 arch/s390/kernel/debug.c 		if (db_info->next)
next              424 arch/s390/kernel/debug.c 			db_info->next->prev = db_info->prev;
next              102 arch/s390/kernel/diag.c 	.next	= show_diag_stat_next,
next              207 arch/s390/kernel/processor.c 	.next	= c_next,
next               36 arch/s390/kernel/suspend.c 	struct page_key_data *next;
next               68 arch/s390/kernel/suspend.c 		page_key_data = pkd->next;
next               89 arch/s390/kernel/suspend.c 		pk->next = page_key_data;
next              124 arch/s390/kernel/suspend.c 	page_key_wp = page_key_wp->next;
next              148 arch/s390/kernel/suspend.c 	page_key_rp = page_key_rp->next;
next              333 arch/s390/kernel/sysinfo.c 	struct service_level *ptr, *next;
next              337 arch/s390/kernel/sysinfo.c 	list_for_each_entry_safe(ptr, next, &service_level_list, list) {
next              376 arch/s390/kernel/sysinfo.c 	.next		= service_level_next,
next               43 arch/s390/kernel/topology.c 	struct mask_info *next;
next               80 arch/s390/kernel/topology.c 			info = info->next;
next              154 arch/s390/kernel/topology.c 		info = info->next;
next              159 arch/s390/kernel/topology.c 		info = info->next;
next              164 arch/s390/kernel/topology.c 		info = info->next;
next              188 arch/s390/kernel/topology.c 			drawer = drawer->next;
next              192 arch/s390/kernel/topology.c 			book = book->next;
next              196 arch/s390/kernel/topology.c 			socket = socket->next;
next              523 arch/s390/kernel/topology.c 		mask->next = memblock_alloc(sizeof(*mask->next), 8);
next              524 arch/s390/kernel/topology.c 		if (!mask->next)
next              526 arch/s390/kernel/topology.c 			      __func__, sizeof(*mask->next), 8);
next              527 arch/s390/kernel/topology.c 		mask = mask->next;
next             1251 arch/s390/kvm/vsie.c 			page = kvm->arch.vsie.pages[kvm->arch.vsie.next];
next             1255 arch/s390/kvm/vsie.c 			kvm->arch.vsie.next++;
next             1256 arch/s390/kvm/vsie.c 			kvm->arch.vsie.next %= nr_vcpus;
next               40 arch/s390/lib/spinlock.c 	struct spin_wait *next, *prev;
next              119 arch/s390/lib/spinlock.c 	struct spin_wait *node, *next;
next              126 arch/s390/lib/spinlock.c 	node->prev = node->next = NULL;
next              157 arch/s390/lib/spinlock.c 		WRITE_ONCE(node->prev->next, node);
next              202 arch/s390/lib/spinlock.c 		while ((next = READ_ONCE(node->next)) == NULL)
next              204 arch/s390/lib/spinlock.c 		next->prev = NULL;
next               41 arch/s390/mm/cmm.c 	struct cmm_page_array *next;
next               89 arch/s390/mm/cmm.c 				npa->next = pa;
next              117 arch/s390/mm/cmm.c 			pa = pa->next;
next              780 arch/s390/mm/fault.c 	struct thread_struct *thread, *next;
next              784 arch/s390/mm/fault.c 	list_for_each_entry_safe(thread, next, &pfault_list, list) {
next              185 arch/s390/mm/gmap.c 	struct page *page, *next;
next              191 arch/s390/mm/gmap.c 	list_for_each_entry_safe(page, next, &gmap->crst_list, lru)
next              199 arch/s390/mm/gmap.c 		list_for_each_entry_safe(page, next, &gmap->pt_list, lru)
next              241 arch/s390/mm/gmap.c 	struct gmap *sg, *next;
next              247 arch/s390/mm/gmap.c 		list_for_each_entry_safe(sg, next, &gmap->children, list) {
next             1185 arch/s390/mm/gmap.c 		rmap->next = radix_tree_deref_slot_protected(slot,
next             1189 arch/s390/mm/gmap.c 		rmap->next = NULL;
next             2240 arch/s390/mm/gmap.c 	struct gmap *gmap, *sg, *next;
next             2257 arch/s390/mm/gmap.c 			list_for_each_entry_safe(sg, next,
next             2561 arch/s390/mm/gmap.c 				  unsigned long next, struct mm_walk *walk)
next             2569 arch/s390/mm/gmap.c 				      unsigned long hmask, unsigned long next,
next             2631 arch/s390/mm/gmap.c 			     unsigned long next, struct mm_walk *walk)
next              106 arch/s390/mm/page-states.c 	unsigned long next;
next              112 arch/s390/mm/page-states.c 		next = pmd_addr_end(addr, end);
next              117 arch/s390/mm/page-states.c 	} while (pmd++, addr = next, addr != end);
next              122 arch/s390/mm/page-states.c 	unsigned long next;
next              129 arch/s390/mm/page-states.c 		next = pud_addr_end(addr, end);
next              137 arch/s390/mm/page-states.c 		mark_kernel_pmd(pud, addr, next);
next              138 arch/s390/mm/page-states.c 	} while (pud++, addr = next, addr != end);
next              143 arch/s390/mm/page-states.c 	unsigned long next;
next              150 arch/s390/mm/page-states.c 		next = p4d_addr_end(addr, end);
next              158 arch/s390/mm/page-states.c 		mark_kernel_pud(p4d, addr, next);
next              159 arch/s390/mm/page-states.c 	} while (p4d++, addr = next, addr != end);
next              164 arch/s390/mm/page-states.c 	unsigned long addr, next;
next              172 arch/s390/mm/page-states.c 		next = pgd_addr_end(addr, MODULES_END);
next              180 arch/s390/mm/page-states.c 		mark_kernel_p4d(pgd, addr, next);
next              181 arch/s390/mm/page-states.c 	} while (pgd++, addr = next, addr != MODULES_END);
next              158 arch/s390/mm/pageattr.c 	unsigned long next;
next              166 arch/s390/mm/pageattr.c 		next = pmd_addr_end(addr, end);
next              168 arch/s390/mm/pageattr.c 			if (addr & ~PMD_MASK || addr + PMD_SIZE > next) {
next              176 arch/s390/mm/pageattr.c 			rc = walk_pte_level(pmdp, addr, next, flags);
next              181 arch/s390/mm/pageattr.c 		addr = next;
next              235 arch/s390/mm/pageattr.c 	unsigned long next;
next              243 arch/s390/mm/pageattr.c 		next = pud_addr_end(addr, end);
next              245 arch/s390/mm/pageattr.c 			if (addr & ~PUD_MASK || addr + PUD_SIZE > next) {
next              253 arch/s390/mm/pageattr.c 			rc = walk_pmd_level(pudp, addr, next, flags);
next              256 arch/s390/mm/pageattr.c 		addr = next;
next              265 arch/s390/mm/pageattr.c 	unsigned long next;
next              273 arch/s390/mm/pageattr.c 		next = p4d_addr_end(addr, end);
next              274 arch/s390/mm/pageattr.c 		rc = walk_pud_level(p4dp, addr, next, flags);
next              276 arch/s390/mm/pageattr.c 		addr = next;
next              287 arch/s390/mm/pageattr.c 	unsigned long next;
next              300 arch/s390/mm/pageattr.c 		next = pgd_addr_end(addr, end);
next              301 arch/s390/mm/pageattr.c 		rc = walk_p4d_level(pgdp, addr, next, flags);
next              305 arch/s390/mm/pageattr.c 	} while (pgdp++, addr = next, addr < end && !rc);
next              370 arch/s390/mm/pgalloc.c 	unsigned long next = (addr + (SIZE)) & ~((SIZE) - 1);		\
next              372 arch/s390/mm/pgalloc.c 	return (next - 1) < (end - 1) ? next : end;			\
next              394 arch/s390/mm/pgalloc.c 	unsigned long *pte, next;
next              401 arch/s390/mm/pgalloc.c 		next = base_page_addr_end(addr, end);
next              403 arch/s390/mm/pgalloc.c 	} while (pte++, addr = next, addr < end);
next              410 arch/s390/mm/pgalloc.c 	unsigned long *ste, next, table;
next              416 arch/s390/mm/pgalloc.c 		next = base_segment_addr_end(addr, end);
next              426 arch/s390/mm/pgalloc.c 		rc = base_page_walk(table, addr, next, alloc);
next              432 arch/s390/mm/pgalloc.c 	} while (ste++, addr = next, addr < end);
next              439 arch/s390/mm/pgalloc.c 	unsigned long *rtte, next, table;
next              445 arch/s390/mm/pgalloc.c 		next = base_region3_addr_end(addr, end);
next              455 arch/s390/mm/pgalloc.c 		rc = base_segment_walk(table, addr, next, alloc);
next              460 arch/s390/mm/pgalloc.c 	} while (rtte++, addr = next, addr < end);
next              467 arch/s390/mm/pgalloc.c 	unsigned long *rste, next, table;
next              473 arch/s390/mm/pgalloc.c 		next = base_region2_addr_end(addr, end);
next              483 arch/s390/mm/pgalloc.c 		rc = base_region3_walk(table, addr, next, alloc);
next              488 arch/s390/mm/pgalloc.c 	} while (rste++, addr = next, addr < end);
next              495 arch/s390/mm/pgalloc.c 	unsigned long *rfte, next, table;
next              501 arch/s390/mm/pgalloc.c 		next = base_region1_addr_end(addr, end);
next              511 arch/s390/mm/pgalloc.c 		rc = base_region2_walk(table, addr, next, alloc);
next              516 arch/s390/mm/pgalloc.c 	} while (rfte++, addr = next, addr < end);
next              546 arch/s390/mm/pgtable.c 		pmd_huge_pte(mm, pmdp) = (pgtable_t) lh->next;
next              117 arch/sh/drivers/pci/pci.c 	hose_tail = &hose->next;
next              157 arch/sh/drivers/pci/pci.c 	for (hose = hose_head; hose; hose = hose->next)
next              264 arch/sh/drivers/pci/pci.c 	for (hose = hose_head; hose; hose = hose->next) {
next              273 arch/sh/include/asm/dwarf.h 	struct dwarf_frame *prev, *next;
next              122 arch/sh/include/asm/mmu_context.h 			     struct mm_struct *next,
next              127 arch/sh/include/asm/mmu_context.h 	if (likely(prev != next)) {
next              128 arch/sh/include/asm/mmu_context.h 		cpumask_set_cpu(cpu, mm_cpumask(next));
next              129 arch/sh/include/asm/mmu_context.h 		set_TTB(next->pgd);
next              130 arch/sh/include/asm/mmu_context.h 		activate_context(next, cpu);
next              132 arch/sh/include/asm/mmu_context.h 		if (!cpumask_test_and_set_cpu(cpu, mm_cpumask(next)))
next              133 arch/sh/include/asm/mmu_context.h 			activate_context(next, cpu);
next              136 arch/sh/include/asm/mmu_context.h #define activate_mm(prev, next)		switch_mm((prev),(next),NULL)
next               18 arch/sh/include/asm/pci.h 	struct pci_channel	*next;
next               66 arch/sh/include/asm/processor.h 	unsigned long long next;
next               65 arch/sh/include/asm/switch_to_32.h 				struct task_struct *next);
next               70 arch/sh/include/asm/switch_to_32.h #define switch_to(prev, next, last)				\
next               82 arch/sh/include/asm/switch_to_32.h 	if (is_dsp_enabled(next))				\
next               83 arch/sh/include/asm/switch_to_32.h 		__restore_dsp(next);				\
next               88 arch/sh/include/asm/switch_to_32.h 	__ts5 = (u32 *)next;					\
next               89 arch/sh/include/asm/switch_to_32.h 	__ts6 = (u32 *)&next->thread.sp;			\
next               90 arch/sh/include/asm/switch_to_32.h 	__ts7 = next->thread.pc;				\
next               18 arch/sh/include/asm/switch_to_64.h 				   struct task_struct *next,
next               21 arch/sh/include/asm/switch_to_64.h #define switch_to(prev,next,last)				\
next               23 arch/sh/include/asm/switch_to_64.h 	if (last_task_used_math != next) {			\
next               24 arch/sh/include/asm/switch_to_64.h 		struct pt_regs *regs = next->thread.uregs;	\
next               27 arch/sh/include/asm/switch_to_64.h 	last = sh64_switch_to(prev, &prev->thread, next,	\
next               28 arch/sh/include/asm/switch_to_64.h 			      &next->thread);			\
next               33 arch/sh/kernel/asm-offsets.c 	DEFINE(PBE_NEXT, offsetof(struct pbe, next));
next              148 arch/sh/kernel/cpu/proc.c 	.next	= c_next,
next               34 arch/sh/kernel/cpu/sh4/sq.c 	struct sq_mapping *next;
next               78 arch/sh/kernel/cpu/sh4/sq.c 		p = &tmp->next;
next               80 arch/sh/kernel/cpu/sh4/sq.c 	map->next = tmp;
next               92 arch/sh/kernel/cpu/sh4/sq.c 	for (p = &sq_mapping_list; (tmp = *p); p = &tmp->next)
next               94 arch/sh/kernel/cpu/sh4/sq.c 			*p = tmp->next;
next              211 arch/sh/kernel/cpu/sh4/sq.c 	for (p = &sq_mapping_list; (map = *p); p = &map->next)
next              293 arch/sh/kernel/cpu/sh4/sq.c 	for (list = &sq_mapping_list; (entry = *list); list = &entry->next)
next              173 arch/sh/kernel/process_32.c __switch_to(struct task_struct *prev, struct task_struct *next)
next              175 arch/sh/kernel/process_32.c 	struct thread_struct *next_t = &next->thread;
next              178 arch/sh/kernel/process_32.c 	__stack_chk_guard = next->stack_canary;
next              184 arch/sh/kernel/process_32.c 	if (next->thread.fpu_counter > 5)
next              194 arch/sh/kernel/process_32.c 		     : "r" (task_thread_info(next)));
next              202 arch/sh/kernel/process_32.c 	if (next->thread.fpu_counter > 5)
next               34 arch/sh/kernel/unwinder.c 		.next = &unwinder_list,
next               52 arch/sh/kernel/unwinder.c 	.next = &stack_reader.list,
next               73 arch/sh/kernel/unwinder.c 	best = list_entry(unwinder_list.next, struct unwinder, list);
next               27 arch/sh/mm/tlb-sh5.c 	cpu_data->dtlb.next	= cpu_data->dtlb.first;
next               38 arch/sh/mm/tlb-sh5.c 	cpu_data->itlb.next	= cpu_data->itlb.first;
next               51 arch/sh/mm/tlb-sh5.c 	return cpu_data->dtlb.next;
next               62 arch/sh/mm/tlb-sh5.c 	cpu_data->dtlb.next  += cpu_data->dtlb.step;
next              104 arch/sh/mm/tlb-sh5.c 	cpu_data->dtlb.next	= entry;
next              191 arch/sh/mm/tlb-sh5.c 	unsigned long long next;
next              214 arch/sh/mm/tlb-sh5.c 	next = tlbp->next;
next              215 arch/sh/mm/tlb-sh5.c 	__flush_tlb_slot(next);
next              218 arch/sh/mm/tlb-sh5.c 		      :  : "r" (next), "r" (pteh), "r" (ptel) );
next              220 arch/sh/mm/tlb-sh5.c 	next += TLB_STEP;
next              221 arch/sh/mm/tlb-sh5.c 	if (next > tlbp->last)
next              222 arch/sh/mm/tlb-sh5.c 		next = tlbp->first;
next              223 arch/sh/mm/tlb-sh5.c 	tlbp->next = next;
next               25 arch/sparc/include/asm/cpudata_32.h 	int next;
next              161 arch/sparc/include/asm/leon_amba.h 	struct amba_confarea_type *next;/* next bus in chain */
next               72 arch/sparc/include/asm/mdesc.h 	struct mdesc_notifier_client	*next;
next               28 arch/sparc/include/asm/switch_to_32.h #define SWITCH_DO_LAZY_FPU(next)	/* */
next               38 arch/sparc/include/asm/switch_to_32.h #define prepare_arch_switch(next) do { \
next               58 arch/sparc/include/asm/switch_to_32.h #define switch_to(prev, next, last) do {						\
next               60 arch/sparc/include/asm/switch_to_32.h 	SWITCH_DO_LAZY_FPU(next);							\
next               61 arch/sparc/include/asm/switch_to_32.h 	cpumask_set_cpu(smp_processor_id(), mm_cpumask(next->active_mm));		\
next               93 arch/sparc/include/asm/switch_to_32.h 	  "r" (task_thread_info(next)),				\
next                7 arch/sparc/include/asm/switch_to_64.h #define prepare_arch_switch(next)		\
next               21 arch/sparc/include/asm/switch_to_64.h #define switch_to(prev, next, last)					\
next               26 arch/sparc/include/asm/switch_to_64.h 	: : "r" (task_thread_info(next)->current_ds));\
next               28 arch/sparc/include/asm/switch_to_64.h 		task_thread_info(next);					\
next               59 arch/sparc/include/asm/switch_to_64.h 	: "0" (task_thread_info(next)),					\
next              434 arch/sparc/kernel/cpu.c 	.next =	c_next,
next              216 arch/sparc/kernel/ds.c 	struct ds_info		*next;
next              752 arch/sparc/kernel/ds.c 	for (dp = ds_info_list; dp; dp = dp->next) {
next              762 arch/sparc/kernel/ds.c 		for (dp = ds_info_list; dp; dp = dp->next) {
next             1217 arch/sparc/kernel/ds.c 	dp->next = ds_info_list;
next                7 arch/sparc/kernel/irq.h         struct irq_bucket *next;
next              170 arch/sparc/kernel/irq_32.c 	p->next = irq_map[pil];
next              189 arch/sparc/kernel/irq_32.c 		pnext = &(*pnext)->next;
next              190 arch/sparc/kernel/irq_32.c 	*pnext = p->next;
next              229 arch/sparc/kernel/irq_32.c 		struct irq_bucket *next = p->next;
next              232 arch/sparc/kernel/irq_32.c 		p = next;
next              162 arch/sparc/kernel/leon_pci_grpci2.c 	unsigned int next;	/* 0x0C Next Data Descriptor in chan */
next              242 arch/sparc/kernel/leon_smp.c 			prev = &cpu_data(i).next;
next              309 arch/sparc/kernel/mdesc.c 	client->next = client_list;
next              496 arch/sparc/kernel/mdesc.c 		p = p->next;
next              446 arch/sparc/kernel/of_device_64.c 				goto next;
next              454 arch/sparc/kernel/of_device_64.c 	next:
next              453 arch/sparc/kernel/pci_fire.c 	pbm->next = pci_pbm_root;
next               62 arch/sparc/kernel/pci_impl.h 	struct pci_pbm_info		*next;
next              499 arch/sparc/kernel/pci_psycho.c 	for (pbm = pci_pbm_root; pbm; pbm = pbm->next) {
next              576 arch/sparc/kernel/pci_psycho.c 	pbm->next = pci_pbm_root;
next              571 arch/sparc/kernel/pci_sabre.c 	pbm->next = pci_pbm_root;
next             1348 arch/sparc/kernel/pci_schizo.c 	pbm->next = pci_pbm_root;
next             1408 arch/sparc/kernel/pci_schizo.c 	for (pbm = pci_pbm_root; pbm; pbm = pbm->next) {
next             1216 arch/sparc/kernel/pci_sun4v.c 	pbm->next = pci_pbm_root;
next               89 arch/sparc/kernel/prom_common.c 		prevp = &(*prevp)->next;
next               41 arch/sparc/kernel/starfire.c 	struct starfire_irqinfo *next;
next               73 arch/sparc/kernel/starfire.c 	p->next = sflist;
next               85 arch/sparc/kernel/starfire.c 	for (p = sflist; p != NULL; p = p->next)
next              136 arch/sparc/kernel/sun4d_irq.c 				struct irq_bucket *next;
next              138 arch/sparc/kernel/sun4d_irq.c 				next = p->next;
next              140 arch/sparc/kernel/sun4d_irq.c 				p = next;
next              175 arch/sparc/kernel/sun4d_irq.c 			struct irq_bucket *next;
next              177 arch/sparc/kernel/sun4d_irq.c 			next = p->next;
next              179 arch/sparc/kernel/sun4d_irq.c 			p = next;
next              173 arch/sparc/kernel/sun4d_smp.c 		prev = &cpu_data(i).next;
next              131 arch/sparc/kernel/sun4m_smp.c 		prev = &cpu_data(i).next;
next              206 arch/sparc/kernel/time_32.c 	unsigned int next = (unsigned int)delta;
next              208 arch/sparc/kernel/time_32.c 	sparc_config.load_profile_irq(cpu, next);
next              420 arch/sparc/mm/hugetlbpage.c 	unsigned long next;
next              426 arch/sparc/mm/hugetlbpage.c 		next = pmd_addr_end(addr, end);
next              433 arch/sparc/mm/hugetlbpage.c 	} while (pmd++, addr = next, addr != end);
next              457 arch/sparc/mm/hugetlbpage.c 	unsigned long next;
next              463 arch/sparc/mm/hugetlbpage.c 		next = pud_addr_end(addr, end);
next              469 arch/sparc/mm/hugetlbpage.c 			hugetlb_free_pmd_range(tlb, pud, addr, next, floor,
next              471 arch/sparc/mm/hugetlbpage.c 	} while (pud++, addr = next, addr != end);
next              495 arch/sparc/mm/hugetlbpage.c 	unsigned long next;
next              515 arch/sparc/mm/hugetlbpage.c 		next = pgd_addr_end(addr, end);
next              518 arch/sparc/mm/hugetlbpage.c 		hugetlb_free_pud_range(tlb, pgd, addr, next, floor, ceiling);
next              519 arch/sparc/mm/hugetlbpage.c 	} while (pgd++, addr = next, addr != end);
next              112 arch/sparc/mm/io-unit.c next:	j = (i & 15);
next              126 arch/sparc/mm/io-unit.c 		goto next;
next              406 arch/sparc/mm/srmmu.c 	struct ctx_list *next;
next              421 arch/sparc/mm/srmmu.c 	entry->next->prev = entry->prev;
next              422 arch/sparc/mm/srmmu.c 	entry->prev->next = entry->next;
next              427 arch/sparc/mm/srmmu.c 	entry->next = head;
next              428 arch/sparc/mm/srmmu.c 	(entry->prev = head->prev)->next = entry;
next              439 arch/sparc/mm/srmmu.c 	ctxp = ctx_free.next;
next              447 arch/sparc/mm/srmmu.c 	ctxp = ctx_used.next;
next              449 arch/sparc/mm/srmmu.c 		ctxp = ctxp->next;
next              487 arch/sparc/mm/srmmu.c 	ctx_free.next = ctx_free.prev = &ctx_free;
next              488 arch/sparc/mm/srmmu.c 	ctx_used.next = ctx_used.prev = &ctx_used;
next              293 arch/sparc/mm/tlb.c 		pmd_huge_pte(mm, pmdp) = (pgtable_t) lh->next;
next              329 arch/um/drivers/chan_kern.c 	struct list_head *ele, *next;
next              332 arch/um/drivers/chan_kern.c 	list_for_each_safe(ele, next, chans) {
next              707 arch/um/drivers/line.c 	struct list_head *ele, *next;
next              713 arch/um/drivers/line.c 	list_for_each_safe(ele, next, &winch_handlers) {
next              727 arch/um/drivers/line.c 	struct list_head *ele, *next;
next              732 arch/um/drivers/line.c 	list_for_each_safe(ele, next, &winch_handlers) {
next               65 arch/um/drivers/mconsole_kern.c 		req = list_entry(mc_requests.next, struct mconsole_entry, list);
next              322 arch/um/drivers/mconsole_kern.c 			unplugged = list_entry(unplugged_pages.next,
next              348 arch/um/drivers/mconsole_kern.c 				struct list_head *entry = unplugged_pages.next;
next              827 arch/um/drivers/mconsole_kern.c 	.next 			= NULL,
next              587 arch/um/drivers/net_kern.c 	struct list_head *ele, *next;
next              598 arch/um/drivers/net_kern.c 	list_for_each_safe(ele, next, &eth_cmd_line) {
next              238 arch/um/drivers/port_kern.c 		conn = list_entry(port->connections.next, struct connection,
next               60 arch/um/include/asm/mmu_context.h static inline void switch_mm(struct mm_struct *prev, struct mm_struct *next, 
next               65 arch/um/include/asm/mmu_context.h 	if(prev != next){
next               67 arch/um/include/asm/mmu_context.h 		cpumask_set_cpu(cpu, mm_cpumask(next));
next               68 arch/um/include/asm/mmu_context.h 		if(next != &init_mm)
next               69 arch/um/include/asm/mmu_context.h 			__switch_mm(&next->context.id);
next               13 arch/um/include/shared/irq_user.h 	struct irq_fd *next;
next               36 arch/um/include/shared/mem_user.h 	struct iomem_region *next;
next               33 arch/um/kernel/irq.c 	struct irq_entry *next;
next              147 arch/um/kernel/irq.c 		irq_entry != NULL; irq_entry = irq_entry->next) {
next              165 arch/um/kernel/irq.c 		irq_entry->next = active_fds;
next              243 arch/um/kernel/irq.c 				active_fds = walk->next;
next              245 arch/um/kernel/irq.c 				previous->next = walk->next;
next              250 arch/um/kernel/irq.c 		walk = walk->next;
next              266 arch/um/kernel/irq.c 		walk = walk->next;
next              350 arch/um/kernel/irq.c 		to_free = to_free->next;
next              402 arch/um/kernel/irq.c 		to_free = to_free->next;
next              136 arch/um/kernel/physmem.c 			region = region->next;
next              191 arch/um/kernel/physmem.c 		region = region->next;
next              216 arch/um/kernel/physmem.c 		region = region->next;
next              208 arch/um/kernel/process.c 	unsigned long long next = time_travel_time + duration;
next              219 arch/um/kernel/process.c 	    time_travel_timer_expiry < next) {
next              232 arch/um/kernel/process.c 		time_travel_set_time(next);
next              239 arch/um/kernel/process.c 			os_timer_one_shot(time_travel_timer_expiry - next);
next              263 arch/um/kernel/tlb.c 	unsigned long next;
next              268 arch/um/kernel/tlb.c 		next = pmd_addr_end(addr, end);
next              271 arch/um/kernel/tlb.c 				ret = add_munmap(addr, next - addr, hvc);
next              275 arch/um/kernel/tlb.c 		else ret = update_pte_range(pmd, addr, next, hvc);
next              276 arch/um/kernel/tlb.c 	} while (pmd++, addr = next, ((addr < end) && !ret));
next              285 arch/um/kernel/tlb.c 	unsigned long next;
next              290 arch/um/kernel/tlb.c 		next = pud_addr_end(addr, end);
next              293 arch/um/kernel/tlb.c 				ret = add_munmap(addr, next - addr, hvc);
next              297 arch/um/kernel/tlb.c 		else ret = update_pmd_range(pud, addr, next, hvc);
next              298 arch/um/kernel/tlb.c 	} while (pud++, addr = next, ((addr < end) && !ret));
next              307 arch/um/kernel/tlb.c 	unsigned long addr = start_addr, next;
next              313 arch/um/kernel/tlb.c 		next = pgd_addr_end(addr, end_addr);
next              316 arch/um/kernel/tlb.c 				ret = add_munmap(addr, next - addr, &hvc);
next              320 arch/um/kernel/tlb.c 		else ret = update_pud_range(pgd, addr, next, &hvc);
next              321 arch/um/kernel/tlb.c 	} while (pgd++, addr = next, ((addr < end_addr) && !ret));
next               95 arch/um/kernel/um_arch.c 	.next	= c_next,
next              224 arch/um/kernel/um_arch.c 	.next 			= NULL,
next              378 arch/um/os-Linux/start_up.c 	*new = ((struct iomem_region) { .next		= iomem_regions,
next               46 arch/unicore32/include/asm/mmu_context.h switch_mm(struct mm_struct *prev, struct mm_struct *next,
next               51 arch/unicore32/include/asm/mmu_context.h 	if (!cpumask_test_and_set_cpu(cpu, mm_cpumask(next)) || prev != next)
next               52 arch/unicore32/include/asm/mmu_context.h 		cpu_switch_mm(next->pgd, next);
next               56 arch/unicore32/include/asm/mmu_context.h #define activate_mm(prev, next)	switch_mm(prev, next, NULL)
next               21 arch/unicore32/include/asm/switch_to.h #define switch_to(prev, next, last)					\
next               24 arch/unicore32/include/asm/switch_to.h 					task_thread_info(next));	\
next               96 arch/unicore32/kernel/asm-offsets.c 	DEFINE(PBE_NEXT,		offsetof(struct pbe, next));
next              351 arch/unicore32/kernel/setup.c 	.next	= c_next,
next               36 arch/unicore32/kernel/time.c 	unsigned long next, oscr;
next               39 arch/unicore32/kernel/time.c 	next = readl(OST_OSCR) + delta;
next               40 arch/unicore32/kernel/time.c 	writel(next, OST_OSMR0);
next               43 arch/unicore32/kernel/time.c 	return (signed)(next - oscr) <= MIN_OSCR_DELTA ? -ETIME : 0;
next              239 arch/unicore32/mm/mmu.c 		unsigned long next = pgd_addr_end(addr, end);
next              241 arch/unicore32/mm/mmu.c 		alloc_init_section(pgd, addr, next, phys, type);
next              243 arch/unicore32/mm/mmu.c 		phys += next - addr;
next              244 arch/unicore32/mm/mmu.c 		addr = next;
next               80 arch/x86/boot/compressed/acpi.c 		pa_data = data->next;
next               84 arch/x86/boot/compressed/eboot.c 	rom->data.next	= 0;
next              164 arch/x86/boot/compressed/eboot.c 	while (data && data->next)
next              165 arch/x86/boot/compressed/eboot.c 		data = (struct setup_data *)(unsigned long)data->next;
next              184 arch/x86/boot/compressed/eboot.c 			data->next = (unsigned long)rom;
next              233 arch/x86/boot/compressed/eboot.c 	new->next = 0;
next              239 arch/x86/boot/compressed/eboot.c 		while (data->next)
next              240 arch/x86/boot/compressed/eboot.c 			data = (struct setup_data *)(unsigned long)data->next;
next              241 arch/x86/boot/compressed/eboot.c 		data->next = (unsigned long)new;
next              498 arch/x86/boot/compressed/eboot.c 	e820ext->next = 0;
next              502 arch/x86/boot/compressed/eboot.c 	while (data && data->next)
next              503 arch/x86/boot/compressed/eboot.c 		data = (struct setup_data *)(unsigned long)data->next;
next              506 arch/x86/boot/compressed/eboot.c 		data->next = (unsigned long)e820ext;
next              462 arch/x86/boot/compressed/kaslr.c 		ptr = (struct setup_data *)(unsigned long)ptr->next;
next             1815 arch/x86/events/core.c 	for (quirk = x86_pmu.quirks; quirk; quirk = quirk->next)
next              788 arch/x86/events/intel/pt.c 			buf->cur = list_entry(buf->cur->list.next, struct topa,
next              529 arch/x86/events/perf_event.h 	struct x86_pmu_quirk *next;
next              726 arch/x86/events/perf_event.h 	__quirk.next = x86_pmu.quirks;					\
next              139 arch/x86/include/asm/mce.h 	unsigned next;
next               22 arch/x86/include/asm/mmu_context.h 					struct mm_struct *next)
next              152 arch/x86/include/asm/mmu_context.h static inline void switch_ldt(struct mm_struct *prev, struct mm_struct *next)
next              173 arch/x86/include/asm/mmu_context.h 		     (unsigned long)next->context.ldt))
next              174 arch/x86/include/asm/mmu_context.h 		load_mm_ldt(next);
next              210 arch/x86/include/asm/mmu_context.h extern void switch_mm(struct mm_struct *prev, struct mm_struct *next,
next              213 arch/x86/include/asm/mmu_context.h extern void switch_mm_irqs_off(struct mm_struct *prev, struct mm_struct *next,
next              217 arch/x86/include/asm/mmu_context.h #define activate_mm(prev, next)			\
next              219 arch/x86/include/asm/mmu_context.h 	paravirt_activate_mm((prev), (next));	\
next              220 arch/x86/include/asm/mmu_context.h 	switch_mm((prev), (next), NULL);	\
next              303 arch/x86/include/asm/paravirt.h 					struct mm_struct *next)
next              305 arch/x86/include/asm/paravirt.h 	PVOP_VCALL2(mmu.activate_mm, prev, next);
next              609 arch/x86/include/asm/paravirt.h static inline void arch_end_context_switch(struct task_struct *next)
next              611 arch/x86/include/asm/paravirt.h 	PVOP_VCALL1(cpu.end_context_switch, next);
next              179 arch/x86/include/asm/paravirt_types.h 	void (*end_context_switch)(struct task_struct *next);
next              226 arch/x86/include/asm/paravirt_types.h 			    struct mm_struct *next);
next              657 arch/x86/include/asm/paravirt_types.h void paravirt_end_context_switch(struct task_struct *next);
next               10 arch/x86/include/asm/switch_to.h 				    struct task_struct *next);
next               13 arch/x86/include/asm/switch_to.h 					  struct task_struct *next);
next               16 arch/x86/include/asm/switch_to.h static inline void prepare_switch_to(struct task_struct *next)
next               32 arch/x86/include/asm/switch_to.h 	READ_ONCE(*(unsigned char *)next->thread.sp);
next               68 arch/x86/include/asm/switch_to.h #define switch_to(prev, next, last)					\
next               70 arch/x86/include/asm/switch_to.h 	prepare_switch_to(next);					\
next               72 arch/x86/include/asm/switch_to.h 	((last) = __switch_to_asm((prev), (next)));			\
next               46 arch/x86/include/uapi/asm/bootparam.h 	__u64 next;
next              485 arch/x86/kernel/alternative.c 	struct list_head next;
next              520 arch/x86/kernel/alternative.c 	list_add_tail(&smp->next, &smp_alt_modules);
next              532 arch/x86/kernel/alternative.c 	list_for_each_entry(item, &smp_alt_modules, next) {
next              535 arch/x86/kernel/alternative.c 		list_del(&item->next);
next              556 arch/x86/kernel/alternative.c 		list_for_each_entry(mod, &smp_alt_modules, next)
next              577 arch/x86/kernel/alternative.c 	list_for_each_entry(mod, &smp_alt_modules, next) {
next              333 arch/x86/kernel/apm_32.c 	struct apm_user *next;
next             1183 arch/x86/kernel/apm_32.c 	for (as = user_list; as != NULL; as = as->next) {
next             1271 arch/x86/kernel/apm_32.c 	for (as = user_list; as != NULL; as = as->next) {
next             1586 arch/x86/kernel/apm_32.c 		user_list = as->next;
next             1591 arch/x86/kernel/apm_32.c 		     (as1 != NULL) && (as1->next != as);
next             1592 arch/x86/kernel/apm_32.c 		     as1 = as1->next)
next             1597 arch/x86/kernel/apm_32.c 			as1->next = as->next;
next             1627 arch/x86/kernel/apm_32.c 	as->next = user_list;
next               49 arch/x86/kernel/asm-offsets.c 	OFFSET(pbe_next, pbe, next);
next               48 arch/x86/kernel/cpu/mce/dev-mcelog.c 	entry = mcelog.next;
next               59 arch/x86/kernel/cpu/mce/dev-mcelog.c 	mcelog.next = entry + 1;
next              204 arch/x86/kernel/cpu/mce/dev-mcelog.c 	unsigned next;
next              220 arch/x86/kernel/cpu/mce/dev-mcelog.c 	next = mcelog.next;
next              223 arch/x86/kernel/cpu/mce/dev-mcelog.c 	for (i = 0; i < next; i++) {
next              230 arch/x86/kernel/cpu/mce/dev-mcelog.c 	memset(mcelog.entry, 0, next * sizeof(struct mce));
next              231 arch/x86/kernel/cpu/mce/dev-mcelog.c 	mcelog.next = 0;
next              245 arch/x86/kernel/cpu/mce/dev-mcelog.c 	if (READ_ONCE(mcelog.next))
next              373 arch/x86/kernel/cpu/mce/severity.c 	.next	= s_next,
next              630 arch/x86/kernel/cpu/microcode/amd.c 		__list_del(p->plist.prev, p->plist.next);
next              354 arch/x86/kernel/cpu/microcode/intel.c 			goto next;
next              363 arch/x86/kernel/cpu/microcode/intel.c 				goto next;
next              372 arch/x86/kernel/cpu/microcode/intel.c 				goto next;
next              378 arch/x86/kernel/cpu/microcode/intel.c next:
next              164 arch/x86/kernel/cpu/proc.c 	.next	= c_next,
next              268 arch/x86/kernel/cpu/resctrl/ctrlmondata.c next:
next              299 arch/x86/kernel/cpu/resctrl/ctrlmondata.c 			goto next;
next              176 arch/x86/kernel/cpu/resctrl/pseudo_lock.c 	struct pseudo_lock_pm_req *pm_req, *next;
next              178 arch/x86/kernel/cpu/resctrl/pseudo_lock.c 	list_for_each_entry_safe(pm_req, next, &plr->pm_reqs, list) {
next              240 arch/x86/kernel/dumpstack.c 				goto next;
next              263 arch/x86/kernel/dumpstack.c next:
next             1002 arch/x86/kernel/e820.c 		pa_data = data->next;
next              102 arch/x86/kernel/hpet.c 		char *next = strchr(str, ',');
next              104 arch/x86/kernel/hpet.c 		if (next)
next              105 arch/x86/kernel/hpet.c 			*next++ = 0;
next              112 arch/x86/kernel/hpet.c 		str = next;
next              177 arch/x86/kernel/jailhouse.c 		pa_data = header.next;
next              115 arch/x86/kernel/kdebugfs.c 		pa_data = data->next;
next              154 arch/x86/kernel/kexec-bzimage64.c 	sd->next = params->hdr.setup_data;
next               85 arch/x86/kernel/ksysfs.c 		pa_data = data->next;
next              108 arch/x86/kernel/ksysfs.c 		pa_data = data->next;
next              257 arch/x86/kernel/ksysfs.c 		pa_data = data->next;
next              177 arch/x86/kernel/kvm.c 		struct hlist_node *p, *next;
next              180 arch/x86/kernel/kvm.c 		hlist_for_each_safe(p, next, &b->list) {
next              262 arch/x86/kernel/paravirt.c void paravirt_end_context_switch(struct task_struct *next)
next              268 arch/x86/kernel/paravirt.c 	if (test_and_clear_ti_thread_flag(task_thread_info(next), TIF_LAZY_MMU_UPDATES))
next              273 arch/x86/kernel/process.c 				    struct thread_struct *next,
next              283 arch/x86/kernel/process.c 		memcpy(tss->io_bitmap, next->io_bitmap_ptr,
next              284 arch/x86/kernel/process.c 		       max(prev->io_bitmap_max, next->io_bitmap_max));
next              492 arch/x86/kernel/process.c 	struct thread_struct *prev, *next;
next              496 arch/x86/kernel/process.c 	next = &next_p->thread;
next              500 arch/x86/kernel/process.c 	switch_to_bitmap(prev, next, tifp, tifn);
next               14 arch/x86/kernel/process.h 				   struct task_struct *next)
next               16 arch/x86/kernel/process.h 	unsigned long next_tif = task_thread_info(next)->flags;
next               38 arch/x86/kernel/process.h 		__switch_to_xtra(prev, next);
next              231 arch/x86/kernel/process_32.c 			     *next = &next_p->thread;
next              233 arch/x86/kernel/process_32.c 	struct fpu *next_fpu = &next->fpu;
next              256 arch/x86/kernel/process_32.c 	load_TLS(next, cpu);
next              264 arch/x86/kernel/process_32.c 	if (get_kernel_rpl() && unlikely(prev->iopl != next->iopl))
next              265 arch/x86/kernel/process_32.c 		set_iopl_mask(next->iopl);
next              282 arch/x86/kernel/process_32.c 	refresh_sysenter_cs(next);
next              290 arch/x86/kernel/process_32.c 	if (prev->gs | next->gs)
next              291 arch/x86/kernel/process_32.c 		lazy_load_gs(next->gs);
next              281 arch/x86/kernel/process_64.c 					      struct thread_struct *next)
next              284 arch/x86/kernel/process_64.c 			next->fsindex, next->fsbase, FS);
next              286 arch/x86/kernel/process_64.c 			next->gsindex, next->gsbase, GS);
next              507 arch/x86/kernel/process_64.c 	struct thread_struct *next = &next_p->thread;
next              509 arch/x86/kernel/process_64.c 	struct fpu *next_fpu = &next->fpu;
next              529 arch/x86/kernel/process_64.c 	load_TLS(next, cpu);
next              553 arch/x86/kernel/process_64.c 	if (unlikely(next->es | prev->es))
next              554 arch/x86/kernel/process_64.c 		loadsegment(es, next->es);
next              557 arch/x86/kernel/process_64.c 	if (unlikely(next->ds | prev->ds))
next              558 arch/x86/kernel/process_64.c 		loadsegment(ds, next->ds);
next              560 arch/x86/kernel/process_64.c 	x86_fsgsbase_load(prev, next);
next              582 arch/x86/kernel/process_64.c 		     prev->iopl != next->iopl))
next              583 arch/x86/kernel/process_64.c 		xen_set_iopl_mask(next->iopl);
next              190 arch/x86/kernel/rtc.c 		for (id = dev->id; id; id = id->next) {
next              412 arch/x86/kernel/setup.c 		pa_next = data->next;
next              441 arch/x86/kernel/setup.c 		pa_data = data->next;
next               24 arch/x86/mm/ident_map.c 	unsigned long next;
next               26 arch/x86/mm/ident_map.c 	for (; addr < end; addr = next) {
next               30 arch/x86/mm/ident_map.c 		next = (addr & PUD_MASK) + PUD_SIZE;
next               31 arch/x86/mm/ident_map.c 		if (next > end)
next               32 arch/x86/mm/ident_map.c 			next = end;
next               48 arch/x86/mm/ident_map.c 			ident_pmd_init(info, pmd, addr, next);
next               54 arch/x86/mm/ident_map.c 		ident_pmd_init(info, pmd, addr, next);
next               64 arch/x86/mm/ident_map.c 	unsigned long next;
next               66 arch/x86/mm/ident_map.c 	for (; addr < end; addr = next) {
next               70 arch/x86/mm/ident_map.c 		next = (addr & P4D_MASK) + P4D_SIZE;
next               71 arch/x86/mm/ident_map.c 		if (next > end)
next               72 arch/x86/mm/ident_map.c 			next = end;
next               76 arch/x86/mm/ident_map.c 			ident_pud_init(info, pud, addr, next);
next               82 arch/x86/mm/ident_map.c 		ident_pud_init(info, pud, addr, next);
next               94 arch/x86/mm/ident_map.c 	unsigned long next;
next              104 arch/x86/mm/ident_map.c 	for (; addr < end; addr = next) {
next              108 arch/x86/mm/ident_map.c 		next = (addr & PGDIR_MASK) + PGDIR_SIZE;
next              109 arch/x86/mm/ident_map.c 		if (next > end)
next              110 arch/x86/mm/ident_map.c 			next = end;
next              114 arch/x86/mm/ident_map.c 			result = ident_p4d_init(info, p4d, addr, next);
next              123 arch/x86/mm/ident_map.c 		result = ident_p4d_init(info, p4d, addr, next);
next              618 arch/x86/mm/init.c 	unsigned long next, start;
next              634 arch/x86/mm/init.c 			next = round_up(start + 1, step_size);
next              635 arch/x86/mm/init.c 			if (next > map_end)
next              636 arch/x86/mm/init.c 				next = map_end;
next              638 arch/x86/mm/init.c 			next = map_end;
next              641 arch/x86/mm/init.c 		mapped_ram_size += init_range_memory_mapping(start, next);
next              642 arch/x86/mm/init.c 		start = next;
next              961 arch/x86/mm/init_64.c 	unsigned long next, pages = 0;
next              967 arch/x86/mm/init_64.c 	for (; addr < end; addr = next, pte++) {
next              968 arch/x86/mm/init_64.c 		next = (addr + PAGE_SIZE) & PAGE_MASK;
next              969 arch/x86/mm/init_64.c 		if (next > end)
next              970 arch/x86/mm/init_64.c 			next = end;
next              984 arch/x86/mm/init_64.c 		if (PAGE_ALIGNED(addr) && PAGE_ALIGNED(next)) {
next             1009 arch/x86/mm/init_64.c 			memset((void *)addr, PAGE_INUSE, next - addr);
next             1032 arch/x86/mm/init_64.c 	unsigned long next, pages = 0;
next             1038 arch/x86/mm/init_64.c 	for (; addr < end; addr = next, pmd++) {
next             1039 arch/x86/mm/init_64.c 		next = pmd_addr_end(addr, end);
next             1046 arch/x86/mm/init_64.c 			    IS_ALIGNED(next, PMD_SIZE)) {
next             1057 arch/x86/mm/init_64.c 				memset((void *)addr, PAGE_INUSE, next - addr);
next             1075 arch/x86/mm/init_64.c 		remove_pte_table(pte_base, addr, next, direct);
next             1088 arch/x86/mm/init_64.c 	unsigned long next, pages = 0;
next             1094 arch/x86/mm/init_64.c 	for (; addr < end; addr = next, pud++) {
next             1095 arch/x86/mm/init_64.c 		next = pud_addr_end(addr, end);
next             1102 arch/x86/mm/init_64.c 			    IS_ALIGNED(next, PUD_SIZE)) {
next             1113 arch/x86/mm/init_64.c 				memset((void *)addr, PAGE_INUSE, next - addr);
next             1131 arch/x86/mm/init_64.c 		remove_pmd_table(pmd_base, addr, next, direct, altmap);
next             1143 arch/x86/mm/init_64.c 	unsigned long next, pages = 0;
next             1148 arch/x86/mm/init_64.c 	for (; addr < end; addr = next, p4d++) {
next             1149 arch/x86/mm/init_64.c 		next = p4d_addr_end(addr, end);
next             1157 arch/x86/mm/init_64.c 		remove_pud_table(pud_base, addr, next, altmap, direct);
next             1176 arch/x86/mm/init_64.c 	unsigned long next;
next             1181 arch/x86/mm/init_64.c 	for (addr = start; addr < end; addr = next) {
next             1182 arch/x86/mm/init_64.c 		next = pgd_addr_end(addr, end);
next             1189 arch/x86/mm/init_64.c 		remove_p4d_table(p4d, addr, next, altmap, direct);
next             1453 arch/x86/mm/init_64.c 	unsigned long next;
next             1459 arch/x86/mm/init_64.c 	for (addr = start; addr < end; addr = next) {
next             1460 arch/x86/mm/init_64.c 		next = pmd_addr_end(addr, end);
next             1505 arch/x86/mm/init_64.c 			vmemmap_verify((pte_t *)pmd, node, addr, next);
next             1508 arch/x86/mm/init_64.c 		if (vmemmap_populate_basepages(addr, next, node))
next             1540 arch/x86/mm/init_64.c 	unsigned long next;
next             1548 arch/x86/mm/init_64.c 	for (; addr < end; addr = next) {
next             1553 arch/x86/mm/init_64.c 			next = (addr + PAGE_SIZE) & PAGE_MASK;
next             1560 arch/x86/mm/init_64.c 			next = (addr + PAGE_SIZE) & PAGE_MASK;
next             1567 arch/x86/mm/init_64.c 			next = (addr + PAGE_SIZE) & PAGE_MASK;
next             1573 arch/x86/mm/init_64.c 			next = (addr + PAGE_SIZE) & PAGE_MASK;
next             1586 arch/x86/mm/init_64.c 			next = pmd_addr_end(addr, end);
next              647 arch/x86/mm/ioremap.c 		paddr_next = data->next;
next              680 arch/x86/mm/ioremap.c 		paddr_next = data->next;
next               79 arch/x86/mm/kasan_init_64.c 	unsigned long next;
next              100 arch/x86/mm/kasan_init_64.c 		next = pmd_addr_end(addr, end);
next              102 arch/x86/mm/kasan_init_64.c 			kasan_populate_pmd(pmd, addr, next, nid);
next              103 arch/x86/mm/kasan_init_64.c 	} while (pmd++, addr = next, addr != end);
next              110 arch/x86/mm/kasan_init_64.c 	unsigned long next;
next              120 arch/x86/mm/kasan_init_64.c 		next = pud_addr_end(addr, end);
next              122 arch/x86/mm/kasan_init_64.c 			kasan_populate_pud(pud, addr, next, nid);
next              123 arch/x86/mm/kasan_init_64.c 	} while (pud++, addr = next, addr != end);
next              131 arch/x86/mm/kasan_init_64.c 	unsigned long next;
next              140 arch/x86/mm/kasan_init_64.c 		next = p4d_addr_end(addr, end);
next              141 arch/x86/mm/kasan_init_64.c 		kasan_populate_p4d(p4d, addr, next, nid);
next              142 arch/x86/mm/kasan_init_64.c 	} while (p4d++, addr = next, addr != end);
next              149 arch/x86/mm/kasan_init_64.c 	unsigned long next;
next              155 arch/x86/mm/kasan_init_64.c 		next = pgd_addr_end(addr, end);
next              156 arch/x86/mm/kasan_init_64.c 		kasan_populate_pgd(pgd, addr, next, nid);
next              157 arch/x86/mm/kasan_init_64.c 	} while (pgd++, addr = next, addr != end);
next              213 arch/x86/mm/kasan_init_64.c 	unsigned long next;
next              223 arch/x86/mm/kasan_init_64.c 		next = p4d_addr_end(addr, end);
next              231 arch/x86/mm/kasan_init_64.c 	} while (p4d++, addr = next, addr != end && p4d_none(*p4d));
next              239 arch/x86/mm/kasan_init_64.c 	unsigned long next;
next              243 arch/x86/mm/kasan_init_64.c 		next = pgd_addr_end(addr, end);
next              244 arch/x86/mm/kasan_init_64.c 		kasan_early_p4d_populate(pgd, addr, next);
next              245 arch/x86/mm/kasan_init_64.c 	} while (pgd++, addr = next, addr != end);
next              486 arch/x86/mm/kmmio.c 		struct kmmio_fault_page *next = f->release_next;
next              489 arch/x86/mm/kmmio.c 		f = next;
next              777 arch/x86/mm/mpx.c 	struct vm_area_struct *next;
next              794 arch/x86/mm/mpx.c 	next = find_vma_prev(mm, start, &prev);
next              803 arch/x86/mm/mpx.c 	while (next && (next->vm_flags & VM_MPX))
next              804 arch/x86/mm/mpx.c 		next = next->vm_next;
next              814 arch/x86/mm/mpx.c 	next = find_vma_prev(mm, start, &prev);
next              816 arch/x86/mm/mpx.c 	    (!next || next->vm_start >= bta_end_vaddr)) {
next             1156 arch/x86/mm/pat.c 	.next  = memtype_seq_next,
next               76 arch/x86/mm/tlb.c static void choose_new_asid(struct mm_struct *next, u64 next_tlb_gen,
next               92 arch/x86/mm/tlb.c 		    next->context.ctx_id)
next              154 arch/x86/mm/tlb.c void switch_mm(struct mm_struct *prev, struct mm_struct *next,
next              160 arch/x86/mm/tlb.c 	switch_mm_irqs_off(prev, next, tsk);
next              192 arch/x86/mm/tlb.c static inline unsigned long mm_mangle_tif_spec_ib(struct task_struct *next)
next              194 arch/x86/mm/tlb.c 	unsigned long next_tif = task_thread_info(next)->flags;
next              197 arch/x86/mm/tlb.c 	return (unsigned long)next->mm | ibpb;
next              200 arch/x86/mm/tlb.c static void cond_ibpb(struct task_struct *next)
next              202 arch/x86/mm/tlb.c 	if (!next || !next->mm)
next              248 arch/x86/mm/tlb.c 		next_mm = mm_mangle_tif_spec_ib(next);
next              268 arch/x86/mm/tlb.c 		if (this_cpu_read(cpu_tlbstate.last_user_mm) != next->mm) {
next              270 arch/x86/mm/tlb.c 			this_cpu_write(cpu_tlbstate.last_user_mm, next->mm);
next              275 arch/x86/mm/tlb.c void switch_mm_irqs_off(struct mm_struct *prev, struct mm_struct *next,
next              333 arch/x86/mm/tlb.c 	if (real_prev == next) {
next              335 arch/x86/mm/tlb.c 			   next->context.ctx_id);
next              343 arch/x86/mm/tlb.c 				 !cpumask_test_cpu(cpu, mm_cpumask(next))))
next              344 arch/x86/mm/tlb.c 			cpumask_set_cpu(cpu, mm_cpumask(next));
next              361 arch/x86/mm/tlb.c 		next_tlb_gen = atomic64_read(&next->context.tlb_gen);
next              386 arch/x86/mm/tlb.c 			sync_current_stack_to_mm(next);
next              403 arch/x86/mm/tlb.c 		if (next != &init_mm)
next              404 arch/x86/mm/tlb.c 			cpumask_set_cpu(cpu, mm_cpumask(next));
next              405 arch/x86/mm/tlb.c 		next_tlb_gen = atomic64_read(&next->context.tlb_gen);
next              407 arch/x86/mm/tlb.c 		choose_new_asid(next, next_tlb_gen, &new_asid, &need_flush);
next              415 arch/x86/mm/tlb.c 		this_cpu_write(cpu_tlbstate.ctxs[new_asid].ctx_id, next->context.ctx_id);
next              417 arch/x86/mm/tlb.c 		load_new_mm_cr3(next->pgd, new_asid, true);
next              430 arch/x86/mm/tlb.c 		load_new_mm_cr3(next->pgd, new_asid, false);
next              439 arch/x86/mm/tlb.c 	this_cpu_write(cpu_tlbstate.loaded_mm, next);
next              442 arch/x86/mm/tlb.c 	if (next != real_prev) {
next              443 arch/x86/mm/tlb.c 		load_mm_cr4_irqsoff(next);
next              444 arch/x86/mm/tlb.c 		switch_ldt(real_prev, next);
next              697 arch/x86/pci/common.c 		pa_data = data->next;
next              120 arch/x86/pci/i386.c 	struct pcibios_fwaddrmap *entry, *next;
next              123 arch/x86/pci/i386.c 	list_for_each_entry_safe(entry, next, &pcibios_fwaddrmappings, list) {
next              334 arch/x86/pci/mmconfig-shared.c 		if (cfg->list.next == &pci_mmcfg_list)
next              337 arch/x86/pci/mmconfig-shared.c 		cfgx = list_entry(cfg->list.next, typeof(*cfg), list);
next             1656 arch/x86/platform/uv/tlb_uv.c 	.next		= ptc_seq_next,
next              331 arch/x86/xen/enlighten_pv.c static void xen_end_context_switch(struct task_struct *next)
next              334 arch/x86/xen/enlighten_pv.c 	paravirt_end_context_switch(next);
next              977 arch/x86/xen/mmu_pv.c static void xen_activate_mm(struct mm_struct *prev, struct mm_struct *next)
next              979 arch/x86/xen/mmu_pv.c 	spin_lock(&next->page_table_lock);
next              980 arch/x86/xen/mmu_pv.c 	xen_pgd_pin(next);
next              981 arch/x86/xen/mmu_pv.c 	spin_unlock(&next->page_table_lock);
next              125 arch/xtensa/include/asm/mmu_context.h static inline void switch_mm(struct mm_struct *prev, struct mm_struct *next,
next              129 arch/xtensa/include/asm/mmu_context.h 	int migrated = next->context.cpu != cpu;
next              133 arch/xtensa/include/asm/mmu_context.h 		next->context.cpu = cpu;
next              135 arch/xtensa/include/asm/mmu_context.h 	if (migrated || prev != next)
next              136 arch/xtensa/include/asm/mmu_context.h 		activate_context(next, cpu);
next              139 arch/xtensa/include/asm/mmu_context.h #define activate_mm(prev, next)	switch_mm((prev), (next), NULL)
next               23 arch/xtensa/include/asm/nommu_context.h static inline void activate_mm(struct mm_struct *prev, struct mm_struct *next)
next               27 arch/xtensa/include/asm/nommu_context.h static inline void switch_mm(struct mm_struct *prev, struct mm_struct *next,
next               35 arch/xtensa/include/asm/pci-bridge.h 	struct pci_controller *next;
next               15 arch/xtensa/include/asm/switch_to.h extern void *_switch_to(void *last, void *next);
next               17 arch/xtensa/include/asm/switch_to.h #define switch_to(prev,next,last)		\
next               19 arch/xtensa/include/asm/switch_to.h 	(last) = _switch_to(prev, next);	\
next              730 arch/xtensa/kernel/setup.c 	.next	= c_next,
next               64 arch/xtensa/kernel/time.c 	unsigned long flags, next;
next               68 arch/xtensa/kernel/time.c 	next = get_ccount() + delta;
next               69 arch/xtensa/kernel/time.c 	set_linux_timer(next);
next               70 arch/xtensa/kernel/time.c 	if (next - get_ccount() > delta)
next              666 arch/xtensa/platforms/iss/network.c 	struct list_head *ele, *next;
next              670 arch/xtensa/platforms/iss/network.c 	list_for_each_safe(ele, next, &eth_cmd_line) {
next              177 block/badblocks.c 		sector_t next = s + sectors;
next              180 block/badblocks.c 		next += (1<<bb->shift) - 1;
next              181 block/badblocks.c 		next >>= bb->shift;
next              182 block/badblocks.c 		sectors = next - s;
next              881 block/bfq-iosched.c 	rq = rq_entry_fifo(bfqq->fifo.next);
next              896 block/bfq-iosched.c 	struct request *next, *prev = NULL;
next              899 block/bfq-iosched.c 	next = bfq_check_fifo(bfqq, last);
next              900 block/bfq-iosched.c 	if (next)
next              901 block/bfq-iosched.c 		return next;
next              907 block/bfq-iosched.c 		next = rb_entry_rq(rbnext);
next              911 block/bfq-iosched.c 			next = rb_entry_rq(rbnext);
next              914 block/bfq-iosched.c 	return bfq_choose_req(bfqd, next, prev, blk_rq_pos(last));
next             2321 block/bfq-iosched.c 				struct request *next)
next             2324 block/bfq-iosched.c 		*next_bfqq = bfq_init_rq(next);
next             2339 block/bfq-iosched.c 	    !list_empty(&rq->queuelist) && !list_empty(&next->queuelist) &&
next             2340 block/bfq-iosched.c 	    next->fifo_time < rq->fifo_time) {
next             2342 block/bfq-iosched.c 		list_replace_init(&next->queuelist, &rq->queuelist);
next             2343 block/bfq-iosched.c 		rq->fifo_time = next->fifo_time;
next             2346 block/bfq-iosched.c 	if (bfqq->next_rq == next)
next             2349 block/bfq-iosched.c 	bfqg_stats_update_io_merged(bfqq_group(bfqq), next->cmd_flags);
next             4905 block/bfq-iosched.c 	struct bfq_queue *__bfqq, *next;
next             4916 block/bfq-iosched.c 		next = __bfqq->new_bfqq;
next             4918 block/bfq-iosched.c 		__bfqq = next;
next              348 block/bfq-wf2q.c 	struct rb_node *next;
next              351 block/bfq-wf2q.c 		next = rb_next(&entity->rb_node);
next              352 block/bfq-wf2q.c 		st->first_idle = bfq_entity_of(next);
next              356 block/bfq-wf2q.c 		next = rb_prev(&entity->rb_node);
next              357 block/bfq-wf2q.c 		st->last_idle = bfq_entity_of(next);
next             1718 block/bio.c    	struct bio *bio, *next;
next             1721 block/bio.c    	next = bio_dirty_list;
next             1725 block/bio.c    	while ((bio = next) != NULL) {
next             1726 block/bio.c    		next = bio->bi_private;
next              165 block/blk-integrity.c 			    struct request *next)
next              167 block/blk-integrity.c 	if (blk_integrity_rq(req) == 0 && blk_integrity_rq(next) == 0)
next              170 block/blk-integrity.c 	if (blk_integrity_rq(req) == 0 || blk_integrity_rq(next) == 0)
next              174 block/blk-integrity.c 	    bio_integrity(next->bio)->bip_flags)
next              177 block/blk-integrity.c 	if (req->nr_integrity_segments + next->nr_integrity_segments >
next              181 block/blk-integrity.c 	if (integrity_req_gap_back_merge(req, next->bio))
next              192 block/blk-integrity.c 	struct bio *next = bio->bi_next;
next              205 block/blk-integrity.c 	bio->bi_next = next;
next              218 block/blk-ioc.c 		struct io_cq *icq = list_entry(icq_list->next,
next              680 block/blk-iolatency.c 			goto next;
next              686 block/blk-iolatency.c 			goto next;
next              711 block/blk-iolatency.c next:
next               16 block/blk-merge.c 		struct request *prev_rq, struct bio *prev, struct bio *next)
next               45 block/blk-merge.c 	bio_get_first_bvec(next, &nb);
next              605 block/blk-merge.c 		struct request *next)
next              611 block/blk-merge.c 	if (blk_rq_sectors(req) + bio_sectors(next->bio) >
next              615 block/blk-merge.c 	req->nr_phys_segments = segments + blk_rq_nr_discard_segments(next);
next              623 block/blk-merge.c 				struct request *next)
next              627 block/blk-merge.c 	if (req_gap_back_merge(req, next->bio))
next              633 block/blk-merge.c 	if ((blk_rq_sectors(req) + blk_rq_sectors(next)) >
next              637 block/blk-merge.c 	total_phys_segments = req->nr_phys_segments + next->nr_phys_segments;
next              641 block/blk-merge.c 	if (blk_integrity_merge_rq(q, req, next) == false)
next              710 block/blk-merge.c 					struct request *next)
next              714 block/blk-merge.c 	else if (blk_rq_pos(req) + blk_rq_sectors(req) == blk_rq_pos(next))
next              725 block/blk-merge.c 				     struct request *req, struct request *next)
next              727 block/blk-merge.c 	if (!rq_mergeable(req) || !rq_mergeable(next))
next              730 block/blk-merge.c 	if (req_op(req) != req_op(next))
next              733 block/blk-merge.c 	if (rq_data_dir(req) != rq_data_dir(next)
next              734 block/blk-merge.c 	    || req->rq_disk != next->rq_disk)
next              738 block/blk-merge.c 	    !blk_write_same_mergeable(req->bio, next->bio))
next              745 block/blk-merge.c 	if (req->write_hint != next->write_hint)
next              748 block/blk-merge.c 	if (req->ioprio != next->ioprio)
next              759 block/blk-merge.c 	switch (blk_try_req_merge(req, next)) {
next              761 block/blk-merge.c 		if (!req_attempt_discard_merge(q, req, next))
next              765 block/blk-merge.c 		if (!ll_merge_requests_fn(q, req, next))
next              778 block/blk-merge.c 	if (((req->rq_flags | next->rq_flags) & RQF_MIXED_MERGE) ||
next              780 block/blk-merge.c 	    (next->cmd_flags & REQ_FAILFAST_MASK)) {
next              782 block/blk-merge.c 		blk_rq_set_mixed_merge(next);
next              790 block/blk-merge.c 	if (next->start_time_ns < req->start_time_ns)
next              791 block/blk-merge.c 		req->start_time_ns = next->start_time_ns;
next              793 block/blk-merge.c 	req->biotail->bi_next = next->bio;
next              794 block/blk-merge.c 	req->biotail = next->biotail;
next              796 block/blk-merge.c 	req->__data_len += blk_rq_bytes(next);
next              799 block/blk-merge.c 		elv_merge_requests(q, req, next);
next              804 block/blk-merge.c 	blk_account_io_merge(next);
next              810 block/blk-merge.c 	next->bio = NULL;
next              811 block/blk-merge.c 	return next;
next              816 block/blk-merge.c 	struct request *next = elv_latter_request(q, rq);
next              818 block/blk-merge.c 	if (next)
next              819 block/blk-merge.c 		return attempt_merge(q, rq, next);
next              835 block/blk-merge.c 			  struct request *next)
next              839 block/blk-merge.c 	free = attempt_merge(q, rq, next);
next               70 block/blk-mq-debugfs.c 	.next	= queue_requeue_list_next,
next              383 block/blk-mq-debugfs.c 	.next	= hctx_dispatch_next,
next              653 block/blk-mq-debugfs.c 	.next	= ctx_##name##_rq_list_next,				\
next              852 block/blk-mq-debugfs.c 			rqos = rqos->next;
next              746 block/blk-mq.c 	struct request *rq, *next;
next              752 block/blk-mq.c 	list_for_each_entry_safe(rq, next, &rq_list, queuelist) {
next              770 block/blk-mq.c 		rq = list_entry(rq_list.next, struct request, queuelist);
next              869 block/blk-mq.c static bool blk_mq_req_expired(struct request *rq, unsigned long *next)
next              882 block/blk-mq.c 	if (*next == 0)
next              883 block/blk-mq.c 		*next = deadline;
next              884 block/blk-mq.c 	else if (time_after(*next, deadline))
next              885 block/blk-mq.c 		*next = deadline;
next              892 block/blk-mq.c 	unsigned long *next = priv;
next              898 block/blk-mq.c 	if (!blk_mq_req_expired(rq, next))
next              919 block/blk-mq.c 	if (blk_mq_req_expired(rq, next))
next              934 block/blk-mq.c 	unsigned long next = 0;
next              954 block/blk-mq.c 	blk_mq_queue_tag_busy_iter(q, blk_mq_check_expired, &next);
next              956 block/blk-mq.c 	if (next != 0) {
next              957 block/blk-mq.c 		mod_timer(&q->timeout, next);
next             1023 block/blk-mq.c 		dispatch_data->rq = list_entry_rq(ctx->rq_lists[type].next);
next             1751 block/blk-mq.c 		rq = list_entry_rq(list.next);
next             2683 block/blk-mq.c 	struct blk_mq_hw_ctx *hctx, *next;
next             2690 block/blk-mq.c 	list_for_each_entry_safe(hctx, next, &q->unused_hctx_list, hctx_list) {
next               37 block/blk-rq-qos.c 		rqos = rqos->next;
next               46 block/blk-rq-qos.c 		rqos = rqos->next;
next               55 block/blk-rq-qos.c 		rqos = rqos->next;
next               64 block/blk-rq-qos.c 		rqos = rqos->next;
next               73 block/blk-rq-qos.c 		rqos = rqos->next;
next               82 block/blk-rq-qos.c 		rqos = rqos->next;
next               91 block/blk-rq-qos.c 		rqos = rqos->next;
next              100 block/blk-rq-qos.c 		rqos = rqos->next;
next              109 block/blk-rq-qos.c 		rqos = rqos->next;
next              301 block/blk-rq-qos.c 		q->rq_qos = rqos->next;
next               30 block/blk-rq-qos.h 	struct rq_qos *next;
next               64 block/blk-rq-qos.h 	for (rqos = q->rq_qos; rqos; rqos = rqos->next) {
next              102 block/blk-rq-qos.h 	rqos->next = q->rq_qos;
next              113 block/blk-rq-qos.h 	for (cur = &q->rq_qos; *cur; cur = &(*cur)->next) {
next              115 block/blk-rq-qos.h 			*cur = rqos->next;
next               35 block/blk-softirq.c 		rq = list_entry(local_list.next, struct request, ipi_list);
next               52 block/blk-softirq.c 	if (list->next == &rq->ipi_list)
next              139 block/blk-softirq.c 		if (list->next == &req->ipi_list)
next              134 block/blk.h    		struct bio *next)
next              137 block/blk.h    	struct bio_integrity_payload *bip_next = bio_integrity(next);
next              154 block/blk.h    		struct bio *next)
next              238 block/blk.h    				struct request *next);
next               43 block/cmdline-parser.c 		char *next = strchr(++partdef, ')');
next               45 block/cmdline-parser.c 		if (!next) {
next               51 block/cmdline-parser.c 		length = min_t(int, next - partdef,
next               56 block/cmdline-parser.c 		partdef = ++next;
next               93 block/cmdline-parser.c 	char *next;
next              105 block/cmdline-parser.c 	next = strchr(bdevdef, ':');
next              106 block/cmdline-parser.c 	if (!next) {
next              111 block/cmdline-parser.c 	length = min_t(int, next - bdevdef, sizeof(newparts->name) - 1);
next              118 block/cmdline-parser.c 	while (next && *(++next)) {
next              119 block/cmdline-parser.c 		bdevdef = next;
next              120 block/cmdline-parser.c 		next = strchr(bdevdef, ',');
next              122 block/cmdline-parser.c 		length = (!next) ? (sizeof(buf) - 1) :
next              123 block/cmdline-parser.c 			min_t(int, next - bdevdef, sizeof(buf) - 1);
next              169 block/cmdline-parser.c 	char *next;
next              174 block/cmdline-parser.c 	next = pbuf = buf = kstrdup(cmdline, GFP_KERNEL);
next              180 block/cmdline-parser.c 	while (next && *pbuf) {
next              181 block/cmdline-parser.c 		next = strchr(pbuf, ';');
next              182 block/cmdline-parser.c 		if (next)
next              183 block/cmdline-parser.c 			*next = '\0';
next              189 block/cmdline-parser.c 		if (next)
next              190 block/cmdline-parser.c 			pbuf = ++next;
next              233 block/elevator.c 	struct hlist_node *next;
next              236 block/elevator.c 	hash_for_each_possible_safe(e->hash, rq, next, hash, offset) {
next              405 block/elevator.c 			     struct request *next)
next              410 block/elevator.c 		e->type->ops.requests_merged(q, rq, next);
next              310 block/genhd.c  	struct blk_major_name *next;
next              327 block/genhd.c  	for (dp = major_names[major_to_index(offset)]; dp; dp = dp->next)
next              394 block/genhd.c  	p->next = NULL;
next              397 block/genhd.c  	for (n = &major_names[index]; *n; n = &(*n)->next) {
next              425 block/genhd.c  	for (n = &major_names[index]; *n; n = &(*n)->next)
next              432 block/genhd.c  		*n = p->next;
next             1079 block/genhd.c  	.next	= disk_seqf_next,
next             1415 block/genhd.c  	.next	= disk_seqf_next,
next              591 block/kyber-iosched.c 	struct request *rq, *next;
next              593 block/kyber-iosched.c 	list_for_each_entry_safe(rq, next, rq_list, queuelist) {
next              936 block/kyber-iosched.c 	.next	= kyber_##name##_rqs_next,				\
next              141 block/mq-deadline.c 			       struct request *next)
next              147 block/mq-deadline.c 	if (!list_empty(&req->queuelist) && !list_empty(&next->queuelist)) {
next              148 block/mq-deadline.c 		if (time_before((unsigned long)next->fifo_time,
next              150 block/mq-deadline.c 			list_move(&req->queuelist, &next->queuelist);
next              151 block/mq-deadline.c 			req->fifo_time = next->fifo_time;
next              158 block/mq-deadline.c 	deadline_remove_request(q, next);
next              185 block/mq-deadline.c 	struct request *rq = rq_entry_fifo(dd->fifo_list[ddir].next);
next              212 block/mq-deadline.c 	rq = rq_entry_fifo(dd->fifo_list[data_dir].next);
next              688 block/mq-deadline.c 	.next	= deadline_##name##_fifo_next,				\
next              755 block/mq-deadline.c 	.next	= deadline_dispatch_next,
next              530 block/partitions/acorn.c 		sector_t next;
next              535 block/partitions/acorn.c 		next = le32_to_cpu(p->start);
next              537 block/partitions/acorn.c 			put_partition(state, slot++, start, next - start);
next              538 block/partitions/acorn.c 		start = next;
next               20 block/partitions/check.h 	int next;
next              140 block/partitions/msdos.c 		if (state->next == state->limit)
next              164 block/partitions/msdos.c 			sector_t offs, size, next;
next              173 block/partitions/msdos.c 			next = this_sector + offs;
next              177 block/partitions/msdos.c 				if (next < first_sector)
next              179 block/partitions/msdos.c 				if (next + size > first_sector + first_size)
next              183 block/partitions/msdos.c 			put_partition(state, state->next, next, size);
next              184 block/partitions/msdos.c 			set_info(state, state->next, disksig);
next              186 block/partitions/msdos.c 				state->parts[state->next].flags = ADDPART_FLAG_RAID;
next              188 block/partitions/msdos.c 			if (++state->next == state->limit)
next              249 block/partitions/msdos.c 	for (i = 0; i < max_nparts && state->next < state->limit; i++) {
next              259 block/partitions/msdos.c 		put_partition(state, state->next++,
next              298 block/partitions/msdos.c 		if (state->next == state->limit)
next              315 block/partitions/msdos.c 		put_partition(state, state->next++, bsd_start, bsd_size);
next              381 block/partitions/msdos.c 		if (state->next == state->limit)
next              385 block/partitions/msdos.c 			put_partition(state, state->next++,
next              425 block/partitions/msdos.c 			if (state->next == state->limit)
next              429 block/partitions/msdos.c 				put_partition(state, state->next++,
next              532 block/partitions/msdos.c 	state->next = 5;
next             2006 block/sed-opal.c 	struct opal_suspend_data *suspend, *next;
next             2009 block/sed-opal.c 	list_for_each_entry_safe(suspend, next, &dev->unlk_lst, node) {
next               67 block/t10-pi.c 				goto next;
next               80 block/t10-pi.c 				goto next;
next               93 block/t10-pi.c next:
next               99 crypto/algapi.c 		n->alg = (n->list.next == stack) ? alg :
next              175 crypto/algapi.c 			if (spawns->next == NULL)
next              937 crypto/algapi.c 		queue->backlog = queue->backlog->next;
next              939 crypto/algapi.c 	request = queue->list.next;
next              479 crypto/asymmetric_keys/asymmetric_type.c 	char *next;
next              494 crypto/asymmetric_keys/asymmetric_type.c 	next = parse_buf;
next              495 crypto/asymmetric_keys/asymmetric_type.c 	restrict_method = strsep(&next, ":");
next              497 crypto/asymmetric_keys/asymmetric_type.c 	if ((strcmp(restrict_method, "key_or_keyring") == 0) && next) {
next              505 crypto/asymmetric_keys/asymmetric_type.c 		key_text = strsep(&next, ":");
next              507 crypto/asymmetric_keys/asymmetric_type.c 		if (next) {
next              508 crypto/asymmetric_keys/asymmetric_type.c 			if (strcmp(next, "chain") != 0)
next               65 crypto/asymmetric_keys/pkcs7_parser.c 			pkcs7->certs = cert->next;
next               70 crypto/asymmetric_keys/pkcs7_parser.c 			pkcs7->crl = cert->next;
next               75 crypto/asymmetric_keys/pkcs7_parser.c 			pkcs7->signed_infos = sinfo->next;
next              100 crypto/asymmetric_keys/pkcs7_parser.c 	for (sinfo = sinfo->next; sinfo; sinfo = sinfo->next)
next              158 crypto/asymmetric_keys/pkcs7_parser.c 		ctx->certs = cert->next;
next              409 crypto/asymmetric_keys/pkcs7_parser.c 	ctx->ppcerts = &x509->next;
next              673 crypto/asymmetric_keys/pkcs7_parser.c 	ctx->ppsinfo = &sinfo->next;
next               18 crypto/asymmetric_keys/pkcs7_parser.h 	struct pkcs7_signed_info *next;
next              166 crypto/asymmetric_keys/pkcs7_trust.c 	for (p = pkcs7->certs; p; p = p->next)
next              169 crypto/asymmetric_keys/pkcs7_trust.c 	for (sinfo = pkcs7->signed_infos; sinfo; sinfo = sinfo->next) {
next              134 crypto/asymmetric_keys/pkcs7_verify.c 	if (sinfo == NULL || sinfo->next != NULL)
next              167 crypto/asymmetric_keys/pkcs7_verify.c 	for (x509 = pkcs7->certs; x509; x509 = x509->next, certix++) {
next              210 crypto/asymmetric_keys/pkcs7_verify.c 	for (p = pkcs7->certs; p; p = p->next)
next              261 crypto/asymmetric_keys/pkcs7_verify.c 			for (p = pkcs7->certs; p; p = p->next) {
next              270 crypto/asymmetric_keys/pkcs7_verify.c 			for (p = pkcs7->certs; p; p = p->next) {
next              451 crypto/asymmetric_keys/pkcs7_verify.c 	for (sinfo = pkcs7->signed_infos; sinfo; sinfo = sinfo->next) {
next               13 crypto/asymmetric_keys/x509_parser.h 	struct x509_certificate *next;
next               87 crypto/proc.c  	.next		= c_next,
next              162 crypto/rsa-pkcs1pad.c 		struct scatterlist *next)
next              164 crypto/rsa-pkcs1pad.c 	int nsegs = next ? 2 : 1;
next              169 crypto/rsa-pkcs1pad.c 	if (next)
next              170 crypto/rsa-pkcs1pad.c 		sg_chain(sg, nsegs, next);
next               66 crypto/xor.c   	tmpl->next = template_list;
next              134 crypto/xor.c   	for (f = fastest; f; f = f->next)
next             2004 drivers/acpi/acpi_video.c 	struct acpi_video_device *dev, *next;
next             2007 drivers/acpi/acpi_video.c 	list_for_each_entry_safe(dev, next, &video->video_device_list, entry) {
next              245 drivers/acpi/acpica/acdebug.h 			     char **next, acpi_object_type *return_type);
next              193 drivers/acpi/acpica/aclocal.h 	struct acpi_new_table_desc *next;
next              412 drivers/acpi/acpica/aclocal.h 	struct acpi_sci_handler_info *next;
next              431 drivers/acpi/acpica/aclocal.h 	struct acpi_gpe_notify_info *next;
next              476 drivers/acpi/acpica/aclocal.h 	struct acpi_gpe_block_info *next;
next              492 drivers/acpi/acpica/aclocal.h 	struct acpi_gpe_xrupt_info *next;
next              552 drivers/acpi/acpica/aclocal.h 	void                            *next; \
next              680 drivers/acpi/acpica/aclocal.h 	struct acpi_address_range *next;
next              736 drivers/acpi/acpica/aclocal.h 	union acpi_parse_object         *next;              /* Next op */\
next              795 drivers/acpi/acpica/aclocal.h 	struct acpi_comment_node *next;
next              800 drivers/acpi/acpica/aclocal.h 	struct acpi_comment_addr_node *next;
next              812 drivers/acpi/acpica/aclocal.h 	struct acpi_file_node *next;
next             1040 drivers/acpi/acpica/aclocal.h 	struct acpi_interface_info *next;
next             1137 drivers/acpi/acpica/aclocal.h 	struct acpi_external_list *next;
next             1155 drivers/acpi/acpica/aclocal.h 	struct acpi_external_file *next;
next             1160 drivers/acpi/acpica/aclocal.h 	struct acpi_parse_object_list *next;
next             1231 drivers/acpi/acpica/aclocal.h 	struct acpi_debug_mem_block     *next; \
next              129 drivers/acpi/acpica/acobject.h 	union acpi_operand_object *next;	/* Link for list of acquired mutexes */
next              138 drivers/acpi/acpica/acobject.h 	union acpi_operand_object *next;
next              277 drivers/acpi/acpica/acobject.h 	union acpi_operand_object *next[2];	/* Device and System handler lists */
next              288 drivers/acpi/acpica/acobject.h 	union acpi_operand_object *next;
next              357 drivers/acpi/acpica/acobject.h 	ACPI_OBJECT_COMMON_HEADER union acpi_operand_object *next;	/* Link for object cache and internal lists */
next               38 drivers/acpi/acpica/acstruct.h 	struct acpi_walk_state *next;	/* Next walk_state in list */
next               86 drivers/acpi/acpica/amlresrc.h 	struct asl_resource_node *next;
next              429 drivers/acpi/acpica/dbcmds.c 			next_interface = next_interface->next;
next              164 drivers/acpi/acpica/dbconvert.c 	char *next;
next              176 drivers/acpi/acpica/dbconvert.c 		this = acpi_db_get_next_token(this, &next, &type);
next              190 drivers/acpi/acpica/dbconvert.c 		this = next;
next              122 drivers/acpi/acpica/dbdisply.c 	acpi_os_printf("%20.20s : %p\n", "NextOp", op->common.next);
next              537 drivers/acpi/acpica/dbdisply.c 			frame = frame->results.next;
next              575 drivers/acpi/acpica/dbdisply.c 		walk_state = walk_state->next;
next              883 drivers/acpi/acpica/dbdisply.c 							notify = notify->next;
next              910 drivers/acpi/acpica/dbdisply.c 			gpe_block = gpe_block->next;
next              913 drivers/acpi/acpica/dbdisply.c 		gpe_xrupt_info = gpe_xrupt_info->next;
next              992 drivers/acpi/acpica/dbdisply.c 			handler_obj = handler_obj->address_space.next;
next             1091 drivers/acpi/acpica/dbdisply.c 		handler_obj = handler_obj->address_space.next;
next              114 drivers/acpi/acpica/dbfileio.c 		table_list_head = table_list_head->next;
next              389 drivers/acpi/acpica/dbinput.c 	const struct acpi_db_command_help *next;
next              392 drivers/acpi/acpica/dbinput.c 	next = acpi_gbl_db_command_help;
next              393 drivers/acpi/acpica/dbinput.c 	while (next->invocation) {
next              394 drivers/acpi/acpica/dbinput.c 		matched = acpi_db_match_command_help(command, next);
next              399 drivers/acpi/acpica/dbinput.c 		next++;
next              419 drivers/acpi/acpica/dbinput.c 	const struct acpi_db_command_help *next = acpi_gbl_db_command_help;
next              427 drivers/acpi/acpica/dbinput.c 		while (next->invocation) {
next              428 drivers/acpi/acpica/dbinput.c 			acpi_os_printf("%-38s%s", next->invocation,
next              429 drivers/acpi/acpica/dbinput.c 				       next->description);
next              430 drivers/acpi/acpica/dbinput.c 			next++;
next              455 drivers/acpi/acpica/dbinput.c 			     char **next, acpi_object_type *return_type)
next              559 drivers/acpi/acpica/dbinput.c 		*next = NULL;
next              562 drivers/acpi/acpica/dbinput.c 		*next = string + 1;
next              586 drivers/acpi/acpica/dbinput.c 	char *next;
next              600 drivers/acpi/acpica/dbinput.c 		acpi_gbl_db_args[i] = acpi_db_get_next_token(this, &next,
next              607 drivers/acpi/acpica/dbinput.c 		this = next;
next              370 drivers/acpi/acpica/dbutils.c 			start_op = (method->common.value.arg)->common.next;
next              205 drivers/acpi/acpica/dbxface.c 	union acpi_parse_object *next;
next              291 drivers/acpi/acpica/dbxface.c 		next = op->common.next;
next              292 drivers/acpi/acpica/dbxface.c 		op->common.next = NULL;
next              325 drivers/acpi/acpica/dbxface.c 		op->common.next = next;
next               93 drivers/acpi/acpica/dsdebug.c 	union acpi_parse_object *next;
next              161 drivers/acpi/acpica/dsdebug.c 				next = op->common.next;
next              162 drivers/acpi/acpica/dsdebug.c 				op->common.next = NULL;
next              175 drivers/acpi/acpica/dsdebug.c 				op->common.next = next;
next              189 drivers/acpi/acpica/dsdebug.c 		next_walk_state = next_walk_state->next;
next              453 drivers/acpi/acpica/dsfield.c 		arg = arg->common.next;
next              511 drivers/acpi/acpica/dsfield.c 	arg = arg->common.next;
next              520 drivers/acpi/acpica/dsfield.c 	status = acpi_ds_get_field_names(&info, walk_state, arg->common.next);
next              648 drivers/acpi/acpica/dsfield.c 		arg = arg->common.next;
next              703 drivers/acpi/acpica/dsfield.c 	arg = arg->common.next;
next              720 drivers/acpi/acpica/dsfield.c 	arg = arg->common.next;
next              724 drivers/acpi/acpica/dsfield.c 	arg = arg->common.next;
next              743 drivers/acpi/acpica/dsfield.c 	status = acpi_ds_get_field_names(&info, walk_state, arg->common.next);
next              788 drivers/acpi/acpica/dsfield.c 	arg = arg->common.next;
next              802 drivers/acpi/acpica/dsfield.c 	arg = arg->common.next;
next              810 drivers/acpi/acpica/dsfield.c 	status = acpi_ds_get_field_names(&info, walk_state, arg->common.next);
next              187 drivers/acpi/acpica/dsobject.c 	byte_list = arg->named.next;
next              376 drivers/acpi/acpica/dsopcode.c 	next_op = next_op->common.next;
next              701 drivers/acpi/acpica/dsopcode.c 	next_op = next_op->common.next;
next              705 drivers/acpi/acpica/dsopcode.c 	next_op = next_op->common.next;
next              754 drivers/acpi/acpica/dsopcode.c 		arg = arg->common.next;
next              130 drivers/acpi/acpica/dspkginit.c 	arg = arg->common.next;
next              246 drivers/acpi/acpica/dspkginit.c 		arg = arg->common.next;
next              280 drivers/acpi/acpica/dspkginit.c 			arg = arg->common.next;
next              688 drivers/acpi/acpica/dsutils.c 		arg = arg->common.next;
next              498 drivers/acpi/acpica/dswexec.c 			next_op = next_op->common.next;
next              517 drivers/acpi/acpica/dswload.c 		    (u8)op->common.value.arg->common.next->common.value.integer;
next               40 drivers/acpi/acpica/dswscope.c 		walk_state->scope_info = scope_info->scope.next;
next              446 drivers/acpi/acpica/dswstate.c 	walk_state->next = thread->walk_state_list;
next              478 drivers/acpi/acpica/dswstate.c 		thread->walk_state_list = walk_state->next;
next              698 drivers/acpi/acpica/dswstate.c 		walk_state->control_state = state->common.next;
next              707 drivers/acpi/acpica/dswstate.c 		walk_state->scope_info = state->common.next;
next              716 drivers/acpi/acpica/dswstate.c 		walk_state->results = state->common.next;
next              432 drivers/acpi/acpica/evgpe.c 		gpe_block = gpe_block->next;
next              485 drivers/acpi/acpica/evgpe.c 			notify = notify->next;
next               67 drivers/acpi/acpica/evgpeblk.c 		while (next_gpe_block->next) {
next               68 drivers/acpi/acpica/evgpeblk.c 			next_gpe_block = next_gpe_block->next;
next               71 drivers/acpi/acpica/evgpeblk.c 		next_gpe_block->next = gpe_block;
next              114 drivers/acpi/acpica/evgpeblk.c 	if (!gpe_block->previous && !gpe_block->next) {
next              127 drivers/acpi/acpica/evgpeblk.c 			gpe_block->previous->next = gpe_block->next;
next              130 drivers/acpi/acpica/evgpeblk.c 			    gpe_block->next;
next              133 drivers/acpi/acpica/evgpeblk.c 		if (gpe_block->next) {
next              134 drivers/acpi/acpica/evgpeblk.c 			gpe_block->next->previous = gpe_block->previous;
next              243 drivers/acpi/acpica/evgpeinit.c 			gpe_block = gpe_block->next;
next              246 drivers/acpi/acpica/evgpeinit.c 		gpe_xrupt_info = gpe_xrupt_info->next;
next               64 drivers/acpi/acpica/evgpeutil.c 			gpe_block = gpe_block->next;
next               67 drivers/acpi/acpica/evgpeutil.c 		gpe_xrupt_info = gpe_xrupt_info->next;
next              150 drivers/acpi/acpica/evgpeutil.c 		next_gpe_xrupt = next_gpe_xrupt->next;
next              167 drivers/acpi/acpica/evgpeutil.c 		while (next_gpe_xrupt->next) {
next              168 drivers/acpi/acpica/evgpeutil.c 			next_gpe_xrupt = next_gpe_xrupt->next;
next              171 drivers/acpi/acpica/evgpeutil.c 		next_gpe_xrupt->next = gpe_xrupt;
next              237 drivers/acpi/acpica/evgpeutil.c 		gpe_xrupt->previous->next = gpe_xrupt->next;
next              241 drivers/acpi/acpica/evgpeutil.c 		acpi_gbl_gpe_xrupt_list_head = gpe_xrupt->next;
next              244 drivers/acpi/acpica/evgpeutil.c 	if (gpe_xrupt->next) {
next              245 drivers/acpi/acpica/evgpeutil.c 		gpe_xrupt->next->previous = gpe_xrupt->previous;
next              276 drivers/acpi/acpica/evgpeutil.c 	struct acpi_gpe_notify_info *next;
next              311 drivers/acpi/acpica/evgpeutil.c 					next = notify->next;
next              313 drivers/acpi/acpica/evgpeutil.c 					notify = next;
next              140 drivers/acpi/acpica/evhandler.c 			handler_obj = handler_obj->address_space.next;
next              304 drivers/acpi/acpica/evhandler.c 		handler_obj = handler_obj->address_space.next;
next              502 drivers/acpi/acpica/evhandler.c 	handler_obj->address_space.next = obj_desc->common_notify.handler;
next              185 drivers/acpi/acpica/evmisc.c 		    handler_obj->notify.next[info->notify.handler_list_id];
next              353 drivers/acpi/acpica/evregion.c 			*last_obj_ptr = obj_desc->region.next;
next              354 drivers/acpi/acpica/evregion.c 			obj_desc->region.next = NULL;	/* Must clear field */
next              435 drivers/acpi/acpica/evregion.c 		last_obj_ptr = &obj_desc->region.next;
next              436 drivers/acpi/acpica/evregion.c 		obj_desc = obj_desc->region.next;
next              495 drivers/acpi/acpica/evregion.c 	region_obj->region.next = handler_obj->address_space.region_list;
next               56 drivers/acpi/acpica/evsci.c 		sci_handler = sci_handler->next;
next              206 drivers/acpi/acpica/evsci.c 		acpi_gbl_sci_handler_list = sci_handler->next;
next              152 drivers/acpi/acpica/evxface.c 				handler_obj = handler_obj->notify.next[i];
next              174 drivers/acpi/acpica/evxface.c 			handler_obj->notify.next[i] =
next              296 drivers/acpi/acpica/evxface.c 				handler_obj = handler_obj->notify.next[i];
next              307 drivers/acpi/acpica/evxface.c 				previous_handler_obj->notify.next[i] =
next              308 drivers/acpi/acpica/evxface.c 				    handler_obj->notify.next[i];
next              312 drivers/acpi/acpica/evxface.c 				    handler_obj->notify.next[i];
next              430 drivers/acpi/acpica/evxface.c 		sci_handler = sci_handler->next;
next              435 drivers/acpi/acpica/evxface.c 	new_sci_handler->next = acpi_gbl_sci_handler_list;
next              493 drivers/acpi/acpica/evxface.c 				prev_sci_handler->next = next_sci_handler->next;
next              496 drivers/acpi/acpica/evxface.c 				    next_sci_handler->next;
next              505 drivers/acpi/acpica/evxface.c 		next_sci_handler = next_sci_handler->next;
next              448 drivers/acpi/acpica/evxfgpe.c 			notify = notify->next;
next              454 drivers/acpi/acpica/evxfgpe.c 		new_notify->next = gpe_event_info->dispatch.notify_list;
next              200 drivers/acpi/acpica/evxfregn.c 			*last_obj_ptr = handler_obj->address_space.next;
next              210 drivers/acpi/acpica/evxfregn.c 		last_obj_ptr = &handler_obj->address_space.next;
next              211 drivers/acpi/acpica/evxfregn.c 		handler_obj = handler_obj->address_space.next;
next              122 drivers/acpi/acpica/exdump.c 	{ACPI_EXD_POINTER, ACPI_EXD_OFFSET(region.next), "Next"}
next              210 drivers/acpi/acpica/exdump.c 	{ACPI_EXD_HDLR_LIST, ACPI_EXD_OFFSET(address_space.next), "Next"},
next              223 drivers/acpi/acpica/exdump.c 	{ACPI_EXD_POINTER, ACPI_EXD_OFFSET(notify.next[0]),
next              225 drivers/acpi/acpica/exdump.c 	{ACPI_EXD_POINTER, ACPI_EXD_OFFSET(notify.next[1]), "Next Device Notify"}
next              338 drivers/acpi/acpica/exdump.c 	union acpi_operand_object *next;
next              440 drivers/acpi/acpica/exdump.c 			next = start;
next              442 drivers/acpi/acpica/exdump.c 			acpi_os_printf("%20s : %p ", name, next);
next              443 drivers/acpi/acpica/exdump.c 			if (next) {
next              446 drivers/acpi/acpica/exdump.c 					       (next), next->common.type);
next              448 drivers/acpi/acpica/exdump.c 				while (next->common.next_object) {
next              449 drivers/acpi/acpica/exdump.c 					if ((next->common.type ==
next              451 drivers/acpi/acpica/exdump.c 						data = next;
next              454 drivers/acpi/acpica/exdump.c 					next = next->common.next_object;
next              455 drivers/acpi/acpica/exdump.c 					acpi_os_printf("->%p(%s %2.2X)", next,
next              457 drivers/acpi/acpica/exdump.c 						       (next),
next              458 drivers/acpi/acpica/exdump.c 						       next->common.type);
next              460 drivers/acpi/acpica/exdump.c 					if ((next == start) || (next == data)) {
next              476 drivers/acpi/acpica/exdump.c 			next = start;
next              478 drivers/acpi/acpica/exdump.c 			acpi_os_printf("%20s : %p", name, next);
next              479 drivers/acpi/acpica/exdump.c 			if (next) {
next              482 drivers/acpi/acpica/exdump.c 					       (next),
next              483 drivers/acpi/acpica/exdump.c 					       next->address_space.space_id);
next              485 drivers/acpi/acpica/exdump.c 				while (next->address_space.next) {
next              486 drivers/acpi/acpica/exdump.c 					if ((next->common.type ==
next              488 drivers/acpi/acpica/exdump.c 						data = next;
next              491 drivers/acpi/acpica/exdump.c 					next = next->address_space.next;
next              492 drivers/acpi/acpica/exdump.c 					acpi_os_printf("->%p(%s %2.2X)", next,
next              494 drivers/acpi/acpica/exdump.c 						       (next),
next              495 drivers/acpi/acpica/exdump.c 						       next->address_space.
next              498 drivers/acpi/acpica/exdump.c 					if ((next == start) || (next == data)) {
next              512 drivers/acpi/acpica/exdump.c 			next = start;
next              514 drivers/acpi/acpica/exdump.c 			acpi_os_printf("%20s : %p", name, next);
next              515 drivers/acpi/acpica/exdump.c 			if (next) {
next              518 drivers/acpi/acpica/exdump.c 					       (next), next->common.type);
next              520 drivers/acpi/acpica/exdump.c 				while (next->region.next) {
next              521 drivers/acpi/acpica/exdump.c 					if ((next->common.type ==
next              523 drivers/acpi/acpica/exdump.c 						data = next;
next              526 drivers/acpi/acpica/exdump.c 					next = next->region.next;
next              527 drivers/acpi/acpica/exdump.c 					acpi_os_printf("->%p(%s %2.2X)", next,
next              529 drivers/acpi/acpica/exdump.c 						       (next),
next              530 drivers/acpi/acpica/exdump.c 						       next->common.type);
next              532 drivers/acpi/acpica/exdump.c 					if ((next == start) || (next == data)) {
next               45 drivers/acpi/acpica/exmutex.c 	if (obj_desc->mutex.next) {
next               46 drivers/acpi/acpica/exmutex.c 		(obj_desc->mutex.next)->mutex.prev = obj_desc->mutex.prev;
next               50 drivers/acpi/acpica/exmutex.c 		(obj_desc->mutex.prev)->mutex.next = obj_desc->mutex.next;
next               61 drivers/acpi/acpica/exmutex.c 		thread->acquired_mutex_list = obj_desc->mutex.next;
next               89 drivers/acpi/acpica/exmutex.c 	obj_desc->mutex.next = list_head;
next              463 drivers/acpi/acpica/exmutex.c 	union acpi_operand_object *next = thread->acquired_mutex_list;
next              470 drivers/acpi/acpica/exmutex.c 	while (next) {
next              471 drivers/acpi/acpica/exmutex.c 		obj_desc = next;
next              496 drivers/acpi/acpica/exmutex.c 		next = obj_desc->mutex.next;
next              499 drivers/acpi/acpica/exmutex.c 		obj_desc->mutex.next = NULL;
next               24 drivers/acpi/acpica/hwpci.c 	struct acpi_pci_device *next;
next              179 drivers/acpi/acpica/hwpci.c 		list_element->next = *return_list_head;
next              237 drivers/acpi/acpica/hwpci.c 		info = info->next;
next              264 drivers/acpi/acpica/hwpci.c 	struct acpi_pci_device *next;
next              267 drivers/acpi/acpica/hwpci.c 	next = list_head;
next              268 drivers/acpi/acpica/hwpci.c 	while (next) {
next              269 drivers/acpi/acpica/hwpci.c 		previous = next;
next              270 drivers/acpi/acpica/hwpci.c 		next = previous->next;
next              758 drivers/acpi/acpica/psargs.c 					prev->common.next = field;
next              104 drivers/acpi/acpica/psparse.c 	union acpi_parse_object *next;
next              235 drivers/acpi/acpica/psparse.c 				replacement_op->common.next = op->common.next;
next              238 drivers/acpi/acpica/psparse.c 				    op->common.next;
next              249 drivers/acpi/acpica/psparse.c 				next = prev->common.next;
next              250 drivers/acpi/acpica/psparse.c 				if (next == op) {
next              258 drivers/acpi/acpica/psparse.c 						prev->common.next =
next              260 drivers/acpi/acpica/psparse.c 						replacement_op->common.next =
next              261 drivers/acpi/acpica/psparse.c 						    op->common.next;
next              262 drivers/acpi/acpica/psparse.c 						next = NULL;
next              264 drivers/acpi/acpica/psparse.c 						prev->common.next =
next              265 drivers/acpi/acpica/psparse.c 						    op->common.next;
next              266 drivers/acpi/acpica/psparse.c 						next = NULL;
next              269 drivers/acpi/acpica/psparse.c 				prev = next;
next              174 drivers/acpi/acpica/psscope.c 	if (scope->common.next) {
next               74 drivers/acpi/acpica/pstree.c 		arg = arg->common.next;
next              133 drivers/acpi/acpica/pstree.c 		while (prev_arg->common.next) {
next              134 drivers/acpi/acpica/pstree.c 			prev_arg = prev_arg->common.next;
next              136 drivers/acpi/acpica/pstree.c 		prev_arg->common.next = arg;
next              147 drivers/acpi/acpica/pstree.c 		arg = arg->common.next;
next              172 drivers/acpi/acpica/pstree.c 	union acpi_parse_object *next = NULL;
next              184 drivers/acpi/acpica/pstree.c 	next = acpi_ps_get_arg(op, 0);
next              185 drivers/acpi/acpica/pstree.c 	if (next) {
next              186 drivers/acpi/acpica/pstree.c 		ASL_CV_LABEL_FILENODE(next);
next              187 drivers/acpi/acpica/pstree.c 		return (next);
next              192 drivers/acpi/acpica/pstree.c 	next = op->common.next;
next              193 drivers/acpi/acpica/pstree.c 	if (next) {
next              194 drivers/acpi/acpica/pstree.c 		ASL_CV_LABEL_FILENODE(next);
next              195 drivers/acpi/acpica/pstree.c 		return (next);
next              207 drivers/acpi/acpica/pstree.c 			arg = arg->common.next;
next              217 drivers/acpi/acpica/pstree.c 		if (parent->common.next) {
next              221 drivers/acpi/acpica/pstree.c 			ASL_CV_LABEL_FILENODE(parent->common.next);
next              222 drivers/acpi/acpica/pstree.c 			return (parent->common.next);
next              229 drivers/acpi/acpica/pstree.c 	ASL_CV_LABEL_FILENODE(next);
next              230 drivers/acpi/acpica/pstree.c 	return (next);
next               32 drivers/acpi/acpica/pswalk.c 	union acpi_parse_object *next = NULL;
next               72 drivers/acpi/acpica/pswalk.c 			next = acpi_ps_get_arg(op, 0);
next               73 drivers/acpi/acpica/pswalk.c 			if (next) {
next               77 drivers/acpi/acpica/pswalk.c 				op = next;
next               85 drivers/acpi/acpica/pswalk.c 		next = op->common.next;
next               96 drivers/acpi/acpica/pswalk.c 		if (next) {
next               97 drivers/acpi/acpica/pswalk.c 			op = next;
next               65 drivers/acpi/acpica/utaddress.c 	range_info->next = acpi_gbl_address_range_list[space_id];
next              115 drivers/acpi/acpica/utaddress.c 				    range_info->next;
next              117 drivers/acpi/acpica/utaddress.c 				prev->next = range_info->next;
next              134 drivers/acpi/acpica/utaddress.c 		range_info = range_info->next;
next              218 drivers/acpi/acpica/utaddress.c 		range_info = range_info->next;
next              239 drivers/acpi/acpica/utaddress.c 	struct acpi_address_range *next;
next              246 drivers/acpi/acpica/utaddress.c 		next = acpi_gbl_address_range_list[i];
next              248 drivers/acpi/acpica/utaddress.c 		while (next) {
next              249 drivers/acpi/acpica/utaddress.c 			range_info = next;
next              250 drivers/acpi/acpica/utaddress.c 			next = range_info->next;
next               76 drivers/acpi/acpica/utcache.c 	void *next;
next               96 drivers/acpi/acpica/utcache.c 		next = ACPI_GET_DESCRIPTOR_PTR(cache->list_head);
next               99 drivers/acpi/acpica/utcache.c 		cache->list_head = next;
next              124 drivers/acpi/acpica/utdelete.c 			next_desc = handler_desc->address_space.next;
next              216 drivers/acpi/acpica/utdelete.c 						    next_desc->region.next;
next              222 drivers/acpi/acpica/utdelete.c 					last_obj_ptr = &next_desc->region.next;
next              223 drivers/acpi/acpica/utdelete.c 					next_desc = next_desc->region.next;
next              512 drivers/acpi/acpica/utdelete.c 					    prev_object->notify.next[i];
next               57 drivers/acpi/acpica/utinit.c 			next_gpe_block = gpe_block->next;
next               64 drivers/acpi/acpica/utinit.c 		next_gpe_xrupt_info = gpe_xrupt_info->next;
next              124 drivers/acpi/acpica/utosi.c 		acpi_default_supported_interfaces[i].next =
next              157 drivers/acpi/acpica/utosi.c 		acpi_gbl_supported_interfaces = next_interface->next;
next              217 drivers/acpi/acpica/utosi.c 	interface_info->next = acpi_gbl_supported_interfaces;
next              254 drivers/acpi/acpica/utosi.c 					    next_interface->next;
next              256 drivers/acpi/acpica/utosi.c 					previous_interface->next =
next              257 drivers/acpi/acpica/utosi.c 					    next_interface->next;
next              278 drivers/acpi/acpica/utosi.c 		next_interface = next_interface->next;
next              323 drivers/acpi/acpica/utosi.c 		next_interface = next_interface->next;
next              352 drivers/acpi/acpica/utosi.c 		next_interface = next_interface->next;
next               34 drivers/acpi/acpica/utstate.c 	state->common.next = *list_head;
next               65 drivers/acpi/acpica/utstate.c 		*list_head = state->common.next;
next              299 drivers/acpi/acpica/uttrack.c 		if (!element->next) {
next              303 drivers/acpi/acpica/uttrack.c 		element = element->next;
next              383 drivers/acpi/acpica/uttrack.c 		allocation->next = mem_list->list_head;
next              390 drivers/acpi/acpica/uttrack.c 		allocation->next = element->next;
next              393 drivers/acpi/acpica/uttrack.c 		if (element->next) {
next              394 drivers/acpi/acpica/uttrack.c 			(element->next)->previous = allocation;
next              397 drivers/acpi/acpica/uttrack.c 		element->next = allocation;
next              452 drivers/acpi/acpica/uttrack.c 		(allocation->previous)->next = allocation->next;
next              454 drivers/acpi/acpica/uttrack.c 		mem_list->list_head = allocation->next;
next              457 drivers/acpi/acpica/uttrack.c 	if (allocation->next) {
next              458 drivers/acpi/acpica/uttrack.c 		(allocation->next)->previous = allocation->previous;
next              687 drivers/acpi/acpica/uttrack.c 		element = element->next;
next              337 drivers/acpi/apei/apei-base.c 	res1 = list_entry(res_list1->next, struct apei_res, list);
next              338 drivers/acpi/apei/apei-base.c 	resn1 = list_entry(res1->list.next, struct apei_res, list);
next              367 drivers/acpi/apei/apei-base.c 		resn1 = list_entry(resn1->list.next, struct apei_res, list);
next              826 drivers/acpi/apei/ghes.c 	struct llist_node *llnode, *next;
next              839 drivers/acpi/apei/ghes.c 		next = llnode->next;
next              853 drivers/acpi/apei/ghes.c 		llnode = next;
next              876 drivers/acpi/apei/ghes.c 		llnode = llnode->next;
next              166 drivers/acpi/evged.c 	struct acpi_ged_event *event, *next;
next              168 drivers/acpi/evged.c 	list_for_each_entry_safe(event, next, &geddev->event_list, node) {
next               95 drivers/acpi/nvs.c 	struct nvs_page *entry, *next;
next              118 drivers/acpi/nvs.c 	list_for_each_entry_safe(entry, next, &nvs_list, node) {
next              693 drivers/acpi/pci_root.c 			goto next;
next              701 drivers/acpi/pci_root.c 			goto next;
next              725 drivers/acpi/pci_root.c 				goto next;
next              729 drivers/acpi/pci_root.c next:
next              438 drivers/acpi/pptt.c 	struct acpi_pptt_processor *next;
next              446 drivers/acpi/pptt.c 		next = fetch_pptt_node(table_hdr, cpu->parent);
next              447 drivers/acpi/pptt.c 		if (!(next && next->flags & ACPI_PPTT_ACPI_IDENTICAL))
next               25 drivers/acpi/proc.c 	struct list_head *node, *next;
next               30 drivers/acpi/proc.c 	list_for_each_safe(node, next, &acpi_wakeup_device_list) {
next               58 drivers/acpi/proc.c 						dev->physical_node_list.next)
next               99 drivers/acpi/proc.c 	struct list_head *node, *next;
next              112 drivers/acpi/proc.c 	list_for_each_safe(node, next, &acpi_wakeup_device_list) {
next              435 drivers/acpi/property.c 	struct acpi_data_node *dn, *next;
next              440 drivers/acpi/property.c 	list_for_each_entry_safe_reverse(dn, next, list, sibling) {
next             1019 drivers/acpi/property.c 	struct list_head *next;
next             1034 drivers/acpi/property.c 			next = adev->node.next;
next             1035 drivers/acpi/property.c 			if (next == head) {
next             1039 drivers/acpi/property.c 			child_adev = list_entry(next, struct acpi_device, node);
next             1072 drivers/acpi/property.c 			next = dn->sibling.next;
next             1073 drivers/acpi/property.c 			if (next == head)
next             1076 drivers/acpi/property.c 			dn = list_entry(next, struct acpi_data_node, sibling);
next               57 drivers/acpi/sleep.c 	.next		= NULL,
next               42 drivers/acpi/wakeup.c 	struct list_head *node, *next;
next               44 drivers/acpi/wakeup.c 	list_for_each_safe(node, next, &acpi_wakeup_device_list) {
next               69 drivers/acpi/wakeup.c 	struct list_head *node, *next;
next               71 drivers/acpi/wakeup.c 	list_for_each_safe(node, next, &acpi_wakeup_device_list) {
next               91 drivers/acpi/wakeup.c 	struct list_head *node, *next;
next               94 drivers/acpi/wakeup.c 	list_for_each_safe(node, next, &acpi_wakeup_device_list) {
next              800 drivers/android/binder.c 	BUG_ON(work->entry.next && !list_empty(&work->entry));
next             1943 drivers/android/binder.c 	struct binder_transaction *next;
next             1979 drivers/android/binder.c 		next = t->from_parent;
next             1986 drivers/android/binder.c 		if (next == NULL) {
next             1991 drivers/android/binder.c 		t = next;
next               52 drivers/android/binder_alloc.c 	return list_entry(buffer->entry.next, struct binder_buffer, entry);
next              549 drivers/android/binder_alloc.c 	struct binder_buffer *prev, *next = NULL;
next              551 drivers/android/binder_alloc.c 	BUG_ON(alloc->buffers.next == &buffer->entry);
next              563 drivers/android/binder_alloc.c 		next = binder_buffer_next(buffer);
next              564 drivers/android/binder_alloc.c 		if (buffer_start_page(next) == buffer_start_page(buffer)) {
next              570 drivers/android/binder_alloc.c 					   next->user_data);
next              586 drivers/android/binder_alloc.c 				   next ? next->user_data : NULL);
next              631 drivers/android/binder_alloc.c 		struct binder_buffer *next = binder_buffer_next(buffer);
next              633 drivers/android/binder_alloc.c 		if (next->free) {
next              634 drivers/android/binder_alloc.c 			rb_erase(&next->rb_node, &alloc->free_buffers);
next              635 drivers/android/binder_alloc.c 			binder_delete_free_buffer(alloc, next);
next              638 drivers/android/binder_alloc.c 	if (alloc->buffers.next != &buffer->entry) {
next              259 drivers/ata/libata-core.c  next:
next              277 drivers/ata/libata-core.c 		goto next;
next             7033 drivers/ata/libata-core.c 	char *p, *cur, *next;
next             7048 drivers/ata/libata-core.c 	for (cur = ata_force_param_buf; *cur != '\0'; cur = next) {
next             7052 drivers/ata/libata-core.c 		next = cur;
next             7053 drivers/ata/libata-core.c 		if (ata_parse_force_one(&next, &te, &reason)) {
next              190 drivers/atm/adummy.c 	struct adummy_dev *adummy_dev, *next;
next              192 drivers/atm/adummy.c 	list_for_each_entry_safe(adummy_dev, next, &adummy_devs, entry) {
next              666 drivers/atm/eni.c 		eni_dev->fast = ENI_VCC(curr)->next;
next              667 drivers/atm/eni.c 		ENI_VCC(curr)->next = ENI_VCC_NOS;
next              674 drivers/atm/eni.c 		eni_dev->slow = ENI_VCC(curr)->next;
next              675 drivers/atm/eni.c 		ENI_VCC(curr)->next = ENI_VCC_NOS;
next              701 drivers/atm/eni.c 		if (ENI_VCC(vcc)->next != ENI_VCC_NOS) {
next              707 drivers/atm/eni.c 		ENI_VCC(vcc)->next = NULL;
next              710 drivers/atm/eni.c 				ENI_VCC(eni_dev->last_fast)->next = vcc;
next              716 drivers/atm/eni.c 				ENI_VCC(eni_dev->last_slow)->next = vcc;
next              805 drivers/atm/eni.c 	eni_vcc->next = ENI_VCC_NOS;
next               64 drivers/atm/eni.h 	struct atm_vcc *next;		/* next pending RX */
next              724 drivers/atm/firestream.c 				    td->flags, td->next, td->bsa, td->aal_bufsize, td->skb );
next              784 drivers/atm/firestream.c 			    pe->flags, pe->next, pe->bsa, pe->aal_bufsize, 
next             1186 drivers/atm/firestream.c 	td->next = 0;
next             1195 drivers/atm/firestream.c 	dq[qd].next  = td->next;
next             1494 drivers/atm/firestream.c 		ne->next  = virt_to_bus (NULL);
next             1510 drivers/atm/firestream.c 			qe->next = virt_to_bus(ne);
next             1915 drivers/atm/firestream.c 	fs_dev->next = fs_boards;
next             1977 drivers/atm/firestream.c 				nxt = bus_to_virt (fp->next);
next             2012 drivers/atm/firestream.c 		nxtdev = dev->next;
next               39 drivers/atm/firestream.h 	u32 next;
next              455 drivers/atm/firestream.h 	struct fs_dev *next;		/* other FS devices */
next              891 drivers/atm/fore200e.c 	buffer = buffer->next;
next              936 drivers/atm/fore200e.c 		    bsq->freebuf = buffer->next;
next             1077 drivers/atm/fore200e.c 	buffer->next = bsq->freebuf;
next             2057 drivers/atm/fore200e.c 		buffer[ i ].next = bsq->freebuf;
next              579 drivers/atm/fore200e.h     struct buffer*       next;        /* next receive buffer     */
next              394 drivers/atm/he.c 	he_dev->next = NULL;
next              396 drivers/atm/he.c 		he_dev->next = he_devs;
next              772 drivers/atm/he.c 	struct he_buff *heb, *next;
next              894 drivers/atm/he.c 	list_for_each_entry_safe(heb, next, &he_dev->rbpl_outstanding, entry)
next             1529 drivers/atm/he.c 	struct he_buff *heb, *next;
next             1573 drivers/atm/he.c 		list_for_each_entry_safe(heb, next, &he_dev->rbpl_outstanding, entry)
next             1651 drivers/atm/he.c 	struct he_buff *heb, *next;
next             1777 drivers/atm/he.c 		list_for_each_entry_safe(heb, next, &he_vcc->buffers, entry)
next              314 drivers/atm/he.h 	struct he_dev *next;
next             1864 drivers/atm/horizon.c     wr_mem (dev, &buf->next, BUFF_STATUS_EMPTY);
next             1873 drivers/atm/horizon.c   wr_mem (dev, &memmap->txfreebufstart.next, BUF_PTR(tx_desc) | BUFF_STATUS_EMPTY);
next             1876 drivers/atm/horizon.c     wr_mem (dev, &tx_desc->next, BUF_PTR(tx_desc+1) | BUFF_STATUS_EMPTY);
next             1880 drivers/atm/horizon.c   wr_mem (dev, &tx_desc->next, BUF_PTR(&memmap->txfreebufend) | BUFF_STATUS_EMPTY);
next             1902 drivers/atm/horizon.c   wr_mem (dev, &memmap->rxfreebufstart.next, BUF_PTR(rx_desc) | BUFF_STATUS_EMPTY);
next             1905 drivers/atm/horizon.c     wr_mem (dev, &rx_desc->next, BUF_PTR(rx_desc+1) | BUFF_STATUS_EMPTY);
next             1910 drivers/atm/horizon.c   wr_mem (dev, &rx_desc->next, BUF_PTR(&memmap->rxfreebufend) | BUFF_STATUS_EMPTY);
next              299 drivers/atm/horizon.h   HDW next;
next               37 drivers/atm/idt77105.c 	struct idt77105_priv *next;
next               91 drivers/atm/idt77105.c 	for (walk = idt77105_all; walk; walk = walk->next) {
next              120 drivers/atm/idt77105.c 	for (walk = idt77105_all; walk; walk = walk->next) {
next              269 drivers/atm/idt77105.c 	PRIV(dev)->next = idt77105_all;
next              333 drivers/atm/idt77105.c              prev = walk, walk = walk->next) {
next              336 drivers/atm/idt77105.c                     prev->next = walk->next;
next              338 drivers/atm/idt77105.c                     idt77105_all = walk->next;
next              651 drivers/atm/idt77252.c 	scq->next = scq->base;
next              662 drivers/atm/idt77252.c 		 scq->base, scq->next, scq->last, (unsigned long long)scq->paddr);
next              710 drivers/atm/idt77252.c 	TXPRINTK("%s: SCQ: next 0x%p\n", card->name, scq->next);
next              745 drivers/atm/idt77252.c 	scq->next->word_1 = cpu_to_le32(tbd->word_1 |
next              747 drivers/atm/idt77252.c 	scq->next->word_2 = cpu_to_le32(tbd->word_2);
next              748 drivers/atm/idt77252.c 	scq->next->word_3 = cpu_to_le32(tbd->word_3);
next              749 drivers/atm/idt77252.c 	scq->next->word_4 = cpu_to_le32(tbd->word_4);
next              751 drivers/atm/idt77252.c 	if (scq->next == scq->last)
next              752 drivers/atm/idt77252.c 		scq->next = scq->base;
next              754 drivers/atm/idt77252.c 		scq->next++;
next              758 drivers/atm/idt77252.c 		   (u32)((unsigned long)scq->next - (unsigned long)scq->base));
next              772 drivers/atm/idt77252.c 		read_sram(card, scq->scd + 1), scq->next);
next              798 drivers/atm/idt77252.c 		 card->name, atomic_read(&scq->used), scq->next);
next              982 drivers/atm/idt77252.c 	card->rsq.next = card->rsq.last;
next             1211 drivers/atm/idt77252.c 	if (card->rsq.next == card->rsq.last)
next             1214 drivers/atm/idt77252.c 		rsqe = card->rsq.next + 1;
next             1224 drivers/atm/idt77252.c 		card->rsq.next = rsqe;
next             1225 drivers/atm/idt77252.c 		if (card->rsq.next == card->rsq.last)
next             1228 drivers/atm/idt77252.c 			rsqe = card->rsq.next + 1;
next             1231 drivers/atm/idt77252.c 	writel((unsigned long) card->rsq.next - (unsigned long) card->rsq.base,
next             1337 drivers/atm/idt77252.c 			struct sk_buff *next;
next             1343 drivers/atm/idt77252.c 			next = sb_pool_skb(card, handle);
next             1346 drivers/atm/idt77252.c 			if (next) {
next             1347 drivers/atm/idt77252.c 				card->raw_cell_head = next;
next             1384 drivers/atm/idt77252.c 	card->tsq.next = card->tsq.last;
next             1389 drivers/atm/idt77252.c 	writel((unsigned long) card->tsq.next - (unsigned long) card->tsq.base,
next             1410 drivers/atm/idt77252.c 	if (card->tsq.next == card->tsq.last)
next             1413 drivers/atm/idt77252.c 		tsqe = card->tsq.next + 1;
next             1416 drivers/atm/idt77252.c 		 card->tsq.base, card->tsq.next, card->tsq.last);
next             1507 drivers/atm/idt77252.c 		card->tsq.next = tsqe;
next             1508 drivers/atm/idt77252.c 		if (card->tsq.next == card->tsq.last)
next             1511 drivers/atm/idt77252.c 			tsqe = card->tsq.next + 1;
next             1514 drivers/atm/idt77252.c 			 card->tsq.base, card->tsq.next, card->tsq.last);
next             1520 drivers/atm/idt77252.c 	writel((unsigned long)card->tsq.next - (unsigned long)card->tsq.base,
next             1525 drivers/atm/idt77252.c 		readl(SAR_REG_TSQT), card->tsq.next);
next             3703 drivers/atm/idt77252.c 	last = &card->next;
next             3768 drivers/atm/idt77252.c 		idt77252_chain = card->next;
next              163 drivers/atm/idt77252.h 	struct scqe		*next;
next              272 drivers/atm/idt77252.h 	struct rsq_entry	*next;
next              307 drivers/atm/idt77252.h 	struct tsq_entry	*next;
next              413 drivers/atm/idt77252.h 	struct idt77252_dev	*next;
next               98 drivers/atm/iphase.c    que->next = NULL; 
next              104 drivers/atm/iphase.c    data->next = NULL;
next              105 drivers/atm/iphase.c    if (que->next == NULL) 
next              106 drivers/atm/iphase.c       que->next = que->tail = data;
next              108 drivers/atm/iphase.c       data->next = que->next;
next              109 drivers/atm/iphase.c       que->next = data;
next              119 drivers/atm/iphase.c    entry->next = NULL;
next              120 drivers/atm/iphase.c    if (que->next == NULL) 
next              121 drivers/atm/iphase.c       que->next = que->tail = entry;
next              123 drivers/atm/iphase.c       que->tail->next = entry;
next              124 drivers/atm/iphase.c       que->tail = que->tail->next;
next              131 drivers/atm/iphase.c    if (que->next == NULL)
next              133 drivers/atm/iphase.c    tmpdata = que->next;
next              134 drivers/atm/iphase.c    if ( que->next == que->tail)  
next              135 drivers/atm/iphase.c       que->next = que->tail = NULL;
next              137 drivers/atm/iphase.c       que->next = que->next->next;
next              301 drivers/atm/iphase.h 	struct free_desc_q *next;  
next              888 drivers/atm/iphase.h    struct ia_rtn_q *next, *tail;
next              539 drivers/atm/nicstar.c 	card->tsq.next = card->tsq.base;
next              558 drivers/atm/nicstar.c 	card->rsq.next = card->rsq.base;
next              885 drivers/atm/nicstar.c 	scq->next = scq->base;
next             1473 drivers/atm/nicstar.c 			scqep = scq->next;
next             1484 drivers/atm/nicstar.c 			if (!ns_scqe_is_tsr(scqep) && scq->tail != scq->next) {
next             1492 drivers/atm/nicstar.c 				scqi = scq->next - scq->base;
next             1496 drivers/atm/nicstar.c 				*scq->next = tsr;
next             1499 drivers/atm/nicstar.c 				if (scq->next == scq->last)
next             1500 drivers/atm/nicstar.c 					scq->next = scq->base;
next             1502 drivers/atm/nicstar.c 					scq->next++;
next             1503 drivers/atm/nicstar.c 				data = scq_virt_to_bus(scq, scq->next);
next             1556 drivers/atm/nicstar.c 		     card->tsq.base, card->tsq.next,
next             1560 drivers/atm/nicstar.c 		     card->rsq.base, card->rsq.next,
next             1728 drivers/atm/nicstar.c 	while (scq->tail == scq->next) {
next             1737 drivers/atm/nicstar.c 							  scq->tail != scq->next,
next             1748 drivers/atm/nicstar.c 	*scq->next = *tbd;
next             1749 drivers/atm/nicstar.c 	index = (int)(scq->next - scq->base);
next             1756 drivers/atm/nicstar.c 		scq->next);
next             1757 drivers/atm/nicstar.c 	if (scq->next == scq->last)
next             1758 drivers/atm/nicstar.c 		scq->next = scq->base;
next             1760 drivers/atm/nicstar.c 		scq->next++;
next             1773 drivers/atm/nicstar.c 		while (scq->tail == scq->next) {
next             1775 drivers/atm/nicstar.c 				data = scq_virt_to_bus(scq, scq->next);
next             1787 drivers/atm/nicstar.c 								  scq->tail != scq->next,
next             1798 drivers/atm/nicstar.c 			scqi = scq->next - scq->base;
next             1803 drivers/atm/nicstar.c 			*scq->next = tsr;
next             1810 drivers/atm/nicstar.c 			     le32_to_cpu(tsr.word_4), scq->next);
next             1811 drivers/atm/nicstar.c 			if (scq->next == scq->last)
next             1812 drivers/atm/nicstar.c 				scq->next = scq->base;
next             1814 drivers/atm/nicstar.c 				scq->next++;
next             1821 drivers/atm/nicstar.c 	data = scq_virt_to_bus(scq, scq->next);
next             1838 drivers/atm/nicstar.c 	if (card->tsq.next == card->tsq.last)
next             1841 drivers/atm/nicstar.c 		one_ahead = card->tsq.next + 1;
next             1848 drivers/atm/nicstar.c 	while (!ns_tsi_isempty(card->tsq.next) || !ns_tsi_isempty(one_ahead) ||
next             1855 drivers/atm/nicstar.c 		while (ns_tsi_isempty(card->tsq.next)) {
next             1856 drivers/atm/nicstar.c 			if (card->tsq.next == card->tsq.last)
next             1857 drivers/atm/nicstar.c 				card->tsq.next = card->tsq.base;
next             1859 drivers/atm/nicstar.c 				card->tsq.next++;
next             1862 drivers/atm/nicstar.c 		if (!ns_tsi_tmrof(card->tsq.next)) {
next             1863 drivers/atm/nicstar.c 			scdi = ns_tsi_getscdindex(card->tsq.next);
next             1871 drivers/atm/nicstar.c 					ns_tsi_init(card->tsq.next);
next             1876 drivers/atm/nicstar.c 			drain_scq(card, scq, ns_tsi_getscqpos(card->tsq.next));
next             1881 drivers/atm/nicstar.c 		ns_tsi_init(card->tsq.next);
next             1882 drivers/atm/nicstar.c 		previous = card->tsq.next;
next             1883 drivers/atm/nicstar.c 		if (card->tsq.next == card->tsq.last)
next             1884 drivers/atm/nicstar.c 			card->tsq.next = card->tsq.base;
next             1886 drivers/atm/nicstar.c 			card->tsq.next++;
next             1888 drivers/atm/nicstar.c 		if (card->tsq.next == card->tsq.last)
next             1891 drivers/atm/nicstar.c 			one_ahead = card->tsq.next + 1;
next             1950 drivers/atm/nicstar.c 	if (!ns_rsqe_valid(card->rsq.next))
next             1953 drivers/atm/nicstar.c 		dequeue_rx(card, card->rsq.next);
next             1954 drivers/atm/nicstar.c 		ns_rsqe_init(card->rsq.next);
next             1955 drivers/atm/nicstar.c 		previous = card->rsq.next;
next             1956 drivers/atm/nicstar.c 		if (card->rsq.next == card->rsq.last)
next             1957 drivers/atm/nicstar.c 			card->rsq.next = card->rsq.base;
next             1959 drivers/atm/nicstar.c 			card->rsq.next++;
next             1960 drivers/atm/nicstar.c 	} while (ns_rsqe_valid(card->rsq.next));
next              658 drivers/atm/nicstar.h 	ns_tsi *next;
next              667 drivers/atm/nicstar.h 	ns_scqe *next;
next              684 drivers/atm/nicstar.h 	ns_rsqe *next;
next               63 drivers/atm/suni.c 	for (walk = sunis; walk; walk = walk->next) {
next              315 drivers/atm/suni.c 	PRIV(dev)->next = sunis;
next              333 drivers/atm/suni.c     (unsigned long) poll_timer.list.next);
next              349 drivers/atm/suni.c 	    walk = &PRIV((*walk)->dev)->next);
next              350 drivers/atm/suni.c 	*walk = PRIV((*walk)->dev)->next;
next              236 drivers/atm/suni.h 	struct suni_priv *next;			/* next SUNI */
next              437 drivers/base/devres.c 		cur = cur->next;
next              447 drivers/base/devres.c 				first = first->next;
next              468 drivers/base/devres.c 		cur = cur->next;
next              526 drivers/base/devres.c 	if (WARN_ON(dev->devres_head.next == NULL))
next              529 drivers/base/devres.c 	return release_nodes(dev, dev->devres_head.next, &dev->devres_head,
next              676 drivers/base/devres.c 			end = grp->node[1].entry.next;
next               42 drivers/base/devtmpfs.c 	struct req *next;
next              123 drivers/base/devtmpfs.c 	req.next = requests;
next              153 drivers/base/devtmpfs.c 	req.next = requests;
next              410 drivers/base/devtmpfs.c 				struct req *next = req->next;
next              414 drivers/base/devtmpfs.c 				req = next;
next              110 drivers/base/firmware_loader/fallback.c 	struct fw_priv *next;
next              113 drivers/base/firmware_loader/fallback.c 	list_for_each_entry_safe(fw_priv, next, &pending_fw_head,
next             1257 drivers/base/firmware_loader/main.c 		fce = list_entry(fwc->fw_names.next,
next               20 drivers/base/isa.c 	struct device *next;
next              107 drivers/base/isa.c 		struct device *tmp = to_isa_dev(dev)->next;
next              155 drivers/base/isa.c 			isa_dev->next = isa_driver->devices;
next               21 drivers/base/map.c 		struct probe *next;
next               60 drivers/base/map.c 			s = &(*s)->next;
next               61 drivers/base/map.c 		p->next = *s;
next               81 drivers/base/map.c 		for (s = &domain->probes[index % 255]; *s; s = &(*s)->next) {
next               84 drivers/base/map.c 				*s = p->next;
next              103 drivers/base/map.c 	for (p = domain->probes[MAJOR(dev) % 255]; p; p = p->next) {
next              339 drivers/base/node.c 	struct node_cache_info *info, *next;
next              344 drivers/base/node.c 	list_for_each_entry_safe(info, next, &node->cache_attrs, node) {
next             1326 drivers/base/platform.c 	if (!epdrv->list.next) {
next             1383 drivers/base/platform.c 		if (!dev->devres_head.next) {
next              766 drivers/base/power/main.c 		dev = to_device(dpm_noirq_list.next);
next              916 drivers/base/power/main.c 		dev = to_device(dpm_late_early_list.next);
next             1077 drivers/base/power/main.c 		dev = to_device(dpm_suspended_list.next);
next             2008 drivers/base/power/main.c 		struct device *dev = to_device(dpm_list.next);
next             1102 drivers/base/power/wakeup.c 	.next  = wakeup_sources_stats_seq_next,
next              215 drivers/base/regmap/regcache-rbtree.c 	struct rb_node *next;
next              225 drivers/base/regmap/regcache-rbtree.c 	next = rb_first(&rbtree_ctx->root);
next              226 drivers/base/regmap/regcache-rbtree.c 	while (next) {
next              227 drivers/base/regmap/regcache-rbtree.c 		rbtree_node = rb_entry(next, struct regcache_rbtree_node, node);
next              228 drivers/base/regmap/regcache-rbtree.c 		next = rb_next(&rbtree_node->node);
next              537 drivers/base/regmap/regmap-debugfs.c 	struct rb_node *next;
next              623 drivers/base/regmap/regmap-debugfs.c 	next = rb_first(&map->range_tree);
next              624 drivers/base/regmap/regmap-debugfs.c 	while (next) {
next              625 drivers/base/regmap/regmap-debugfs.c 		range_node = rb_entry(next, struct regmap_range_node, node);
next              632 drivers/base/regmap/regmap-debugfs.c 		next = rb_next(&range_node->node);
next              578 drivers/base/regmap/regmap.c 	struct rb_node *next;
next              581 drivers/base/regmap/regmap.c 	next = rb_first(&map->range_tree);
next              582 drivers/base/regmap/regmap.c 	while (next) {
next              583 drivers/base/regmap/regmap.c 		range_node = rb_entry(next, struct regmap_range_node, node);
next              584 drivers/base/regmap/regmap.c 		next = rb_next(&range_node->node);
next              157 drivers/block/aoe/aoe.h 	struct aoedev *next;
next              137 drivers/block/aoe/aoeblk.c 			list_empty(&(*t)->ffree) ? NULL : (*t)->ffree.next);
next              221 drivers/block/aoe/aoecmd.c 		pos = t->ffree.next;
next              769 drivers/block/aoe/aoecmd.c 		pos = flist.next;
next             1205 drivers/block/aoe/aoecmd.c 		pos = iocq[id].head.next;
next             1665 drivers/block/aoe/aoecmd.c 		pos = flist.next;
next              329 drivers/block/aoe/aoedev.c 	for (d = devlist; d; d = d->next) {
next              360 drivers/block/aoe/aoedev.c 	for (d = devlist; d; d = d->next) {
next              378 drivers/block/aoe/aoedev.c 			*dd = d->next;
next              381 drivers/block/aoe/aoedev.c 			dd = &d->next;
next              449 drivers/block/aoe/aoedev.c 	for (d=devlist; d; d=d->next)
next              491 drivers/block/aoe/aoedev.c 	d->next = devlist;
next              494 drivers/block/brd.c 	struct brd_device *brd, *next;
next              542 drivers/block/brd.c 	list_for_each_entry_safe(brd, next, &brd_devices, brd_list) {
next              554 drivers/block/brd.c 	struct brd_device *brd, *next;
next              556 drivers/block/brd.c 	list_for_each_entry_safe(brd, next, &brd_devices, brd_list)
next              742 drivers/block/drbd/drbd_actlog.c 	int next = (device->rs_last_mark + 1) % DRBD_SYNC_MARKS;
next              747 drivers/block/drbd/drbd_actlog.c 			device->rs_mark_time[next] = now;
next              748 drivers/block/drbd/drbd_actlog.c 			device->rs_mark_left[next] = still_to_go;
next              749 drivers/block/drbd/drbd_actlog.c 			device->rs_last_mark = next;
next             3772 drivers/block/drbd/drbd_nl.c 	if (list_empty(connections) || connections->next->next != connections)
next             3976 drivers/block/drbd/drbd_nl.c 			pos = list_entry_rcu(resource->resources.next,
next             4942 drivers/block/drbd/drbd_nl.c 		goto next;
next             4948 drivers/block/drbd/drbd_nl.c 		goto next;
next             4954 drivers/block/drbd/drbd_nl.c 		goto next;
next             4960 drivers/block/drbd/drbd_nl.c 		goto next;
next             4963 drivers/block/drbd/drbd_nl.c next:
next             4966 drivers/block/drbd/drbd_nl.c 			list_entry(state_change->list.next,
next             5015 drivers/block/drbd/drbd_nl.c 			list_entry(head.next, struct drbd_state_change, list);
next             1398 drivers/block/drbd/drbd_receiver.c 				next_epoch = list_entry(epoch->list.next, struct drbd_epoch, list);
next             3048 drivers/block/floppy.c 		if (raw_cmd->next &&
next             3053 drivers/block/floppy.c 			raw_cmd = raw_cmd->next;
next             3074 drivers/block/floppy.c 		cmd.next = NULL;
next             3090 drivers/block/floppy.c 		ptr = ptr->next;
next             3098 drivers/block/floppy.c 	struct floppy_raw_cmd *next;
next             3109 drivers/block/floppy.c 		next = this->next;
next             3111 drivers/block/floppy.c 		this = next;
next             3130 drivers/block/floppy.c 	ptr->next = NULL;
next             3167 drivers/block/floppy.c 		rcmd = &(ptr->next);
next              607 drivers/block/nbd.c 		struct bio *next = bio->bi_next;
next              612 drivers/block/nbd.c 			bool is_last = !next && bio_iter_last(bvec, iter);
next              652 drivers/block/nbd.c 		bio = next;
next              986 drivers/block/null_blk_main.c 			goto next;
next              991 drivers/block/null_blk_main.c next:
next             1490 drivers/block/null_blk_main.c 		cmd->ll_list.next = NULL;
next             1828 drivers/block/null_blk_main.c 		nullb = list_entry(nullb_list.next, struct nullb, list);
next             1854 drivers/block/null_blk_main.c 		nullb = list_entry(nullb_list.next, struct nullb, list);
next              593 drivers/block/pktcdvd.c 	struct packet_data *pkt, *next;
next              597 drivers/block/pktcdvd.c 	list_for_each_entry_safe(pkt, next, &pd->cdrw.pkt_free_list, list) {
next              645 drivers/block/pktcdvd.c 	struct rb_node *next;
next              656 drivers/block/pktcdvd.c 			next = n->rb_left;
next              658 drivers/block/pktcdvd.c 			next = n->rb_right;
next              659 drivers/block/pktcdvd.c 		if (!next)
next              661 drivers/block/pktcdvd.c 		n = next;
next             1065 drivers/block/pktcdvd.c 		if (pkt->sector == zone || pkt->list.next == &pd->cdrw.pkt_free_list) {
next             1381 drivers/block/pktcdvd.c 	struct packet_data *pkt, *next;
next             1397 drivers/block/pktcdvd.c 	list_for_each_entry_safe(pkt, next, &pd->cdrw.pkt_active_list, list) {
next              547 drivers/block/ps3vram.c 	struct bio *next;
next              580 drivers/block/ps3vram.c 	next = bio_list_peek(&priv->list);
next              585 drivers/block/ps3vram.c 	return next;
next              409 drivers/block/rsxx/dma.c 		dma = list_entry(ctrl->queue.next, struct rsxx_dma, list);
next              477 drivers/block/xen-blkback/blkback.c 		req = list_entry(ring->pending_free.next, struct pending_req,
next              880 drivers/block/xen-blkback/blkback.c 				goto next;
next              900 drivers/block/xen-blkback/blkback.c 				goto next;
next              909 drivers/block/xen-blkback/blkback.c 				goto next;
next              915 drivers/block/xen-blkback/blkback.c 			goto next;
next              926 drivers/block/xen-blkback/blkback.c next:
next              682 drivers/block/zram/zram_drv.c 			goto next;
next              687 drivers/block/zram/zram_drv.c 			goto next;
next              691 drivers/block/zram/zram_drv.c 			goto next;
next              694 drivers/block/zram/zram_drv.c 			goto next;
next              746 drivers/block/zram/zram_drv.c 			goto next;
next              759 drivers/block/zram/zram_drv.c next:
next              887 drivers/block/zram/zram_drv.c 			goto next;
next              905 drivers/block/zram/zram_drv.c next:
next              278 drivers/bluetooth/btsdio.c 		tuple = tuple->next;
next               92 drivers/bus/brcmstb_gisb.c 	struct list_head next;
next              232 drivers/bus/brcmstb_gisb.c 	list_for_each_entry(gdev, &brcmstb_gisb_arb_device_list, next) {
next              286 drivers/bus/brcmstb_gisb.c 	list_for_each_entry(gdev, &brcmstb_gisb_arb_device_list, next)
next              333 drivers/bus/brcmstb_gisb.c 	INIT_LIST_HEAD(&gdev->next);
next              390 drivers/bus/brcmstb_gisb.c 	list_add_tail(&gdev->next, &brcmstb_gisb_arb_device_list);
next              555 drivers/bus/fsl-mc/fsl-mc-allocator.c 	struct fsl_mc_resource *next;
next              561 drivers/bus/fsl-mc/fsl-mc-allocator.c 	list_for_each_entry_safe(resource, next, &res_pool->free_list, node) {
next             3255 drivers/cdrom/cdrom.c 	long next = 0;
next             3257 drivers/cdrom/cdrom.c 	ret = cdrom_get_next_writable(cdi, &next);
next             3260 drivers/cdrom/cdrom.c 	if (copy_to_user((long __user *)arg, &next, sizeof(next)))
next               59 drivers/char/agp/frontend.c 		curr = curr->next;
next               69 drivers/char/agp/frontend.c 	struct agp_memory *next;
next               75 drivers/char/agp/frontend.c 		next = temp->next;
next               79 drivers/char/agp/frontend.c 			prev->next = next;
next               80 drivers/char/agp/frontend.c 			if (next != NULL)
next               81 drivers/char/agp/frontend.c 				next->prev = prev;
next               85 drivers/char/agp/frontend.c 			if (next != NULL)
next               86 drivers/char/agp/frontend.c 				next->prev = NULL;
next               88 drivers/char/agp/frontend.c 			agp_fe.current_controller->pool = next;
next              207 drivers/char/agp/frontend.c 		temp->next = prev;
next              224 drivers/char/agp/frontend.c 		curr = curr->next;
next              238 drivers/char/agp/frontend.c 	priv->next = prev;
next              244 drivers/char/agp/frontend.c 	struct agp_file_private *next;
next              247 drivers/char/agp/frontend.c 	next = priv->next;
next              251 drivers/char/agp/frontend.c 		prev->next = next;
next              253 drivers/char/agp/frontend.c 		if (next != NULL)
next              254 drivers/char/agp/frontend.c 			next->prev = prev;
next              257 drivers/char/agp/frontend.c 		if (next != NULL)
next              258 drivers/char/agp/frontend.c 			next->prev = NULL;
next              260 drivers/char/agp/frontend.c 		agp_fe.file_priv_list = next;
next              302 drivers/char/agp/frontend.c 		controller = controller->next;
next              325 drivers/char/agp/frontend.c 	controller->next = prev_controller;
next              353 drivers/char/agp/frontend.c 		client = client->next;
next              367 drivers/char/agp/frontend.c 		memory = memory->next;
next              378 drivers/char/agp/frontend.c 	next_controller = controller->next;
next              381 drivers/char/agp/frontend.c 		prev_controller->next = next_controller;
next              419 drivers/char/agp/frontend.c 		clients = clients->next;
next              441 drivers/char/agp/frontend.c 		clients = clients->next;
next              467 drivers/char/agp/frontend.c 		client = client->next;
next              482 drivers/char/agp/frontend.c 		controller = controller->next;
next              504 drivers/char/agp/frontend.c 	client->next = prev_client;
next              542 drivers/char/agp/frontend.c 	next_client = client->next;
next              545 drivers/char/agp/frontend.c 		prev_client->next = next_client;
next               40 drivers/char/agp/isoch.c 	struct list_head *pos, *tmp, *head = &list->list, *start = head->next;
next               53 drivers/char/agp/isoch.c 		pos = pos->next;
next              304 drivers/char/agp/isoch.c 	for (pos=head->next; cdev<ndevs; cdev++, pos=pos->next) {
next              462 drivers/char/agp/isoch.c 	for (pos=head->next; pos!=head; ) {
next              465 drivers/char/agp/isoch.c 		pos = pos->next;
next              305 drivers/char/hw_random/core.c 		new_rng = list_entry(rng_list.next, struct hwrng, list);
next              126 drivers/char/ipmi/ipmi_devintf.c 	struct ipmi_recv_msg *msg, *next;
next              132 drivers/char/ipmi/ipmi_devintf.c 	list_for_each_entry_safe(msg, next, &priv->recv_msgs, link)
next              226 drivers/char/ipmi/ipmi_devintf.c 	entry = priv->recv_msgs.next;
next               28 drivers/char/ipmi/ipmi_dmi.c 	struct ipmi_dmi_info *next;
next               85 drivers/char/ipmi/ipmi_dmi.c 		info->next = ipmi_dmi_infos;
next              110 drivers/char/ipmi/ipmi_dmi.c 		info = info->next;
next              255 drivers/char/ipmi/ipmi_msghandler.c 	struct cmd_rcvr *next;
next             1346 drivers/char/ipmi/ipmi_msghandler.c 			rcvr->next = rcvrs;
next             1354 drivers/char/ipmi/ipmi_msghandler.c 		rcvrs = rcvr->next;
next             1715 drivers/char/ipmi/ipmi_msghandler.c 				rcvr->next = rcvrs;
next             1726 drivers/char/ipmi/ipmi_msghandler.c 		rcvrs = rcvr->next;
next             3569 drivers/char/ipmi/ipmi_msghandler.c 		entry = tmplist.next;
next             4437 drivers/char/ipmi/ipmi_msghandler.c 		smi_msg = list_entry(intf->waiting_rcv_msgs.next,
next             4508 drivers/char/ipmi/ipmi_msghandler.c 			entry = intf->hp_xmit_msgs.next;
next             4510 drivers/char/ipmi/ipmi_msghandler.c 			entry = intf->xmit_msgs.next;
next             5137 drivers/char/ipmi/ipmi_msghandler.c 	.next		= NULL,
next              188 drivers/char/ipmi/ipmi_si_hotmod.c 	char *str = kstrdup(val, GFP_KERNEL), *curr, *next;
next              205 drivers/char/ipmi/ipmi_si_hotmod.c 	for (curr = str; curr; curr = next) {
next              208 drivers/char/ipmi/ipmi_si_hotmod.c 		next = strchr(curr, ':');
next              209 drivers/char/ipmi/ipmi_si_hotmod.c 		if (next) {
next              210 drivers/char/ipmi/ipmi_si_hotmod.c 			*next = '\0';
next              211 drivers/char/ipmi/ipmi_si_hotmod.c 			next++;
next             1166 drivers/char/ipmi/ipmi_watchdog.c 	.next		= NULL,
next               94 drivers/char/misc.c 	.next  = misc_seq_next,
next              286 drivers/char/tpm/eventlog/tpm1.c 	.next = tpm1_bios_measurements_next,
next              293 drivers/char/tpm/eventlog/tpm1.c 	.next = tpm1_bios_measurements_next,
next              158 drivers/char/tpm/eventlog/tpm2.c 	.next = tpm2_bios_measurements_next,
next             4851 drivers/clk/clk.c 	struct clock_provider *clk_provider, *next;
next             4868 drivers/clk/clk.c 			list_for_each_entry_safe(clk_provider, next,
next             4885 drivers/clk/clk.c 		list_for_each_entry_safe(clk_provider, next,
next              116 drivers/clocksource/em_sti.c static u64 em_sti_set_next(struct em_sti_priv *p, u64 next)
next              126 drivers/clocksource/em_sti.c 	em_sti_write(p, STI_COMPA_H, next >> 32);
next              127 drivers/clocksource/em_sti.c 	em_sti_write(p, STI_COMPA_L, next & 0xffffffff);
next              137 drivers/clocksource/em_sti.c 	return next;
next              253 drivers/clocksource/em_sti.c 	u64 next;
next              256 drivers/clocksource/em_sti.c 	next = em_sti_set_next(p, em_sti_count(p) + delta);
next              257 drivers/clocksource/em_sti.c 	safe = em_sti_count(p) < (next - 1);
next               69 drivers/clocksource/ingenic-timer.c static int ingenic_tcu_cevt_set_next(unsigned long next,
next               74 drivers/clocksource/ingenic-timer.c 	if (next > 0xffff)
next               77 drivers/clocksource/ingenic-timer.c 	regmap_write(tcu->map, TCU_REG_TDFRc(tcu->timer_channel), next);
next               62 drivers/clocksource/mps2-timer.c static int mps2_timer_set_next_event(unsigned long next, struct clock_event_device *ce)
next               64 drivers/clocksource/mps2-timer.c 	clockevent_mps2_writel(next, ce, TIMER_VALUE);
next              340 drivers/clocksource/timer-cadence-ttc.c 	ttccs->ttc.clk_rate_change_nb.next = NULL;
next              421 drivers/clocksource/timer-cadence-ttc.c 	ttcce->ttc.clk_rate_change_nb.next = NULL;
next               60 drivers/clocksource/timer-imx-sysctr.c 	u64 next;
next               64 drivers/clocksource/timer-imx-sysctr.c 	next = sysctr_read_counter();
next               66 drivers/clocksource/timer-imx-sysctr.c 	next += delta;
next               68 drivers/clocksource/timer-imx-sysctr.c 	cmp_hi = (next >> 32) & 0x00fffff;
next               69 drivers/clocksource/timer-imx-sysctr.c 	cmp_lo = next & 0xffffffff;
next               86 drivers/clocksource/timer-imx-tpm.c 	unsigned long next, now;
next               88 drivers/clocksource/timer-imx-tpm.c 	next = tpm_read_counter();
next               89 drivers/clocksource/timer-imx-tpm.c 	next += delta;
next               90 drivers/clocksource/timer-imx-tpm.c 	writel(next, timer_base + TPM_C0V);
next               99 drivers/clocksource/timer-imx-tpm.c 	return (int)(next - now) <= 0 ? -ETIME : 0;
next              103 drivers/clocksource/timer-integrator-ap.c static int clkevt_set_next_event(unsigned long next, struct clock_event_device *evt)
next              108 drivers/clocksource/timer-integrator-ap.c 	writel(next, clkevt_base + TIMER_LOAD);
next               91 drivers/clocksource/timer-prima2.c 	unsigned long now, next;
next               96 drivers/clocksource/timer-prima2.c 	next = now + delta;
next               97 drivers/clocksource/timer-prima2.c 	writel_relaxed(next, sirfsoc_timer_base + SIRFSOC_TIMER_MATCH_0);
next              102 drivers/clocksource/timer-prima2.c 	return next - now > delta ? -ETIME : 0;
next               81 drivers/clocksource/timer-pxa.c 	unsigned long next, oscr;
next               84 drivers/clocksource/timer-pxa.c 	next = timer_readl(OSCR) + delta;
next               85 drivers/clocksource/timer-pxa.c 	timer_writel(next, OSMR0);
next               88 drivers/clocksource/timer-pxa.c 	return (signed)(next - oscr) <= MIN_OSCR_DELTA ? -ETIME : 0;
next              147 drivers/clocksource/timer-sp804.c static int sp804_set_next_event(unsigned long next,
next              153 drivers/clocksource/timer-sp804.c 	writel(next, clkevt_base + TIMER_LOAD);
next              130 drivers/clocksource/timer-stm32.c 	unsigned long now, next;
next              132 drivers/clocksource/timer-stm32.c 	next = readl_relaxed(timer_of_base(to) + TIM_CNT) + evt;
next              133 drivers/clocksource/timer-stm32.c 	writel_relaxed(next, timer_of_base(to) + TIM_CCR1);
next              136 drivers/clocksource/timer-stm32.c 	if ((next - now) > evt)
next              214 drivers/clocksource/timer-sun5i.c 	cs->timer.clk_rate_cb.next = NULL;
next              293 drivers/clocksource/timer-sun5i.c 	ce->timer.clk_rate_cb.next = NULL;
next              466 drivers/cpufreq/cpufreq.c 	for (nb = cpufreq_transition_notifier_list.head; nb; nb = nb->next)
next              264 drivers/cpufreq/powernv-cpufreq.c 		goto next;
next              270 drivers/cpufreq/powernv-cpufreq.c 		goto next;
next              278 drivers/cpufreq/powernv-cpufreq.c next:
next              569 drivers/cpufreq/powernv-cpufreq.c 			goto next;
next              590 drivers/cpufreq/powernv-cpufreq.c next:
next              999 drivers/cpufreq/powernv-cpufreq.c 	.next		= NULL,
next              158 drivers/cpufreq/pxa3xx-cpufreq.c 	struct pxa3xx_freq_info *next;
next              164 drivers/cpufreq/pxa3xx-cpufreq.c 	next = &pxa3xx_freqs[index];
next              167 drivers/cpufreq/pxa3xx-cpufreq.c 	__update_core_freq(next);
next              168 drivers/cpufreq/pxa3xx-cpufreq.c 	__update_bus_freq(next);
next              432 drivers/crypto/axis/artpec6_crypto.c 	struct artpec6_crypto_bounce_buffer *next;
next              434 drivers/crypto/axis/artpec6_crypto.c 	list_for_each_entry_safe(b, next, &dma->bounce_buffers, list) {
next             1063 drivers/crypto/axis/artpec6_crypto.c 	struct artpec6_crypto_bounce_buffer *next;
next             1065 drivers/crypto/axis/artpec6_crypto.c 	list_for_each_entry_safe(b, next, &dma->bounce_buffers, list) {
next             5369 drivers/crypto/caam/caamalg_qi2.c 	if (hash_list.next) {
next             1928 drivers/crypto/caam/caamhash.c 	if (!hash_list.next)
next              124 drivers/crypto/ccp/ccp-crypto-main.c 			req_queue.backlog = crypto_cmd->entry.next;
next              128 drivers/crypto/ccp/ccp-crypto-main.c 		req_queue.backlog = req_queue.backlog->next;
next              132 drivers/crypto/ccp/ccp-crypto-main.c 			req_queue.backlog = crypto_cmd->entry.next;
next              147 drivers/crypto/ccp/ccp-crypto-main.c 	struct ccp_crypto_cmd *held, *next, *backlog;
next              198 drivers/crypto/ccp/ccp-crypto-main.c 		next = ccp_crypto_cmd_complete(held, &backlog);
next              205 drivers/crypto/ccp/ccp-crypto-main.c 		held = next;
next               70 drivers/crypto/chelsio/chtls/chtls.h 	struct listen_info *next;  /* Link to next entry */
next              226 drivers/crypto/chelsio/chtls/chtls.h 		struct sk_buff *next;
next              291 drivers/crypto/chelsio/chtls/chtls_cm.c 	int next = (int)new_state[sk->sk_state];
next              293 drivers/crypto/chelsio/chtls/chtls_cm.c 	tcp_set_state(sk, next & TCP_STATE_MASK);
next              294 drivers/crypto/chelsio/chtls/chtls_cm.c 	return next & TCP_ACTION_FIN;
next              512 drivers/crypto/chelsio/chtls/chtls_cm.c 		p->next = cdev->listen_hash_tab[key];
next              529 drivers/crypto/chelsio/chtls/chtls_cm.c 	for (p = cdev->listen_hash_tab[key]; p; p = p->next)
next              549 drivers/crypto/chelsio/chtls/chtls_cm.c 	for (p = *prev; p; prev = &p->next, p = p->next)
next              552 drivers/crypto/chelsio/chtls/chtls_cm.c 			*prev = p->next;
next              164 drivers/crypto/hisilicon/sec/sec_algs.c 		sgl_next = sgl_current->next;
next              205 drivers/crypto/hisilicon/sec/sec_algs.c 				sgl_current->next = sgl_next;
next              388 drivers/crypto/hisilicon/sec/sec_drv.h 	struct sec_hw_sgl *next;
next               54 drivers/crypto/hisilicon/sgl.c 	struct hisi_acc_hw_sgl *next;
next              105 drivers/crypto/ixp4xx_crypto.c 	struct buffer_desc *next;
next              330 drivers/crypto/ixp4xx_crypto.c 		buf1 = buf->next;
next              643 drivers/crypto/ixp4xx_crypto.c 	buf->next = 0;
next              799 drivers/crypto/ixp4xx_crypto.c 		buf->next = next_buf;
next              807 drivers/crypto/ixp4xx_crypto.c 	buf->next = NULL;
next              917 drivers/crypto/ixp4xx_crypto.c 		req_ctx->dst = dst_hook.next;
next              927 drivers/crypto/ixp4xx_crypto.c 	req_ctx->src = src_hook.next;
next             1027 drivers/crypto/ixp4xx_crypto.c 	req_ctx->src = src_hook.next;
next             1047 drivers/crypto/ixp4xx_crypto.c 		req_ctx->dst = dst_hook.next;
next              309 drivers/crypto/marvell/cesa.h 	struct mv_cesa_tdma_desc *next;
next               70 drivers/crypto/marvell/tdma.c 		tdma = tdma->next;
next               84 drivers/crypto/marvell/tdma.c 	for (tdma = dreq->chain.first; tdma; tdma = tdma->next) {
next              106 drivers/crypto/marvell/tdma.c 		last->next = dreq->chain.first;
next              124 drivers/crypto/marvell/tdma.c 	struct mv_cesa_tdma_desc *tdma = NULL, *next = NULL;
next              130 drivers/crypto/marvell/tdma.c 	for (tdma = engine->chain.first; tdma; tdma = next) {
next              132 drivers/crypto/marvell/tdma.c 		next = tdma->next;
next              152 drivers/crypto/marvell/tdma.c 			engine->chain.first = tdma->next;
next              153 drivers/crypto/marvell/tdma.c 			tdma->next = NULL;
next              203 drivers/crypto/marvell/tdma.c 		chain->last->next = new_tdma;
next              227 drivers/crypto/marvell/tdma.c 	for (op_desc = chain->first; op_desc; op_desc = op_desc->next) {
next             1782 drivers/crypto/picoxcell_crypto.c 	struct spacc_alg *alg, *next;
next             1793 drivers/crypto/picoxcell_crypto.c 	list_for_each_entry_safe(alg, next, &engine->registered_algs, entry) {
next               94 drivers/crypto/qat/qat_common/adf_cfg.c 	.next = qat_dev_cfg_next,
next               57 drivers/crypto/qat/qat_common/adf_cfg_user.h 		struct adf_user_cfg_key_val *next;
next               70 drivers/crypto/qat/qat_common/adf_cfg_user.h 		struct adf_user_cfg_section *next;
next              214 drivers/crypto/qat/qat_common/adf_ctl_drv.c 			params_head = key_val.next;
next              216 drivers/crypto/qat/qat_common/adf_ctl_drv.c 		section_head = section.next;
next              235 drivers/crypto/qat/qat_common/adf_dev_mgr.c 			struct vf_id_map *next;
next              241 drivers/crypto/qat/qat_common/adf_dev_mgr.c 			next = list_next_entry(map, list);
next              242 drivers/crypto/qat/qat_common/adf_dev_mgr.c 			while (next && &next->list != &vfs_table) {
next              243 drivers/crypto/qat/qat_common/adf_dev_mgr.c 				next->fake_id++;
next              244 drivers/crypto/qat/qat_common/adf_dev_mgr.c 				next = list_next_entry(next, list);
next              299 drivers/crypto/qat/qat_common/adf_dev_mgr.c 		struct vf_id_map *map, *next;
next              308 drivers/crypto/qat/qat_common/adf_dev_mgr.c 		next = list_next_entry(map, list);
next              309 drivers/crypto/qat/qat_common/adf_dev_mgr.c 		while (next && &next->list != &vfs_table) {
next              310 drivers/crypto/qat/qat_common/adf_dev_mgr.c 			next->fake_id--;
next              311 drivers/crypto/qat/qat_common/adf_dev_mgr.c 			next = list_next_entry(next, list);
next              126 drivers/crypto/qat/qat_common/adf_transport_debug.c 	.next = adf_ring_next,
next              237 drivers/crypto/qat/qat_common/adf_transport_debug.c 	.next = adf_bank_next,
next              403 drivers/crypto/qat/qat_common/icp_qat_uclo.h 	struct icp_qat_uof_batch_init *next;
next             1149 drivers/crypto/qat/qat_common/qat_hal.c 	plm_init = lm_init_header->next;
next             1168 drivers/crypto/qat/qat_common/qat_hal.c 		plm_init = plm_init->next;
next              210 drivers/crypto/qat/qat_common/qat_uclo.c 	umem_init = umem_init_header->next;
next              219 drivers/crypto/qat/qat_common/qat_uclo.c 		umem_init = umem_init->next;
next              234 drivers/crypto/qat/qat_common/qat_uclo.c 		umem_init = umem_init->next;
next              316 drivers/crypto/qat/qat_common/qat_uclo.c 	while (tail_old->next)
next              317 drivers/crypto/qat/qat_common/qat_uclo.c 		tail_old = tail_old->next;
next              327 drivers/crypto/qat/qat_common/qat_uclo.c 		mem_init->next = NULL;
next              328 drivers/crypto/qat/qat_common/qat_uclo.c 		tail->next = mem_init;
next              336 drivers/crypto/qat/qat_common/qat_uclo.c 		mem_init = tail_old->next;
next              134 drivers/crypto/sahara.c 	u32	next;
next              140 drivers/crypto/sahara.c 	u32	next;
next              418 drivers/crypto/sahara.c 			dev->hw_desc[i]->next);
next              436 drivers/crypto/sahara.c 			dev->hw_link[i]->next);
next              463 drivers/crypto/sahara.c 		dev->hw_desc[idx]->next = dev->hw_phys_desc[1];
next              506 drivers/crypto/sahara.c 			dev->hw_link[i]->next = 0;
next              508 drivers/crypto/sahara.c 			dev->hw_link[i]->next = dev->hw_phys_link[i + 1];
next              520 drivers/crypto/sahara.c 			dev->hw_link[j]->next = 0;
next              522 drivers/crypto/sahara.c 			dev->hw_link[j]->next = dev->hw_phys_link[j + 1];
next              531 drivers/crypto/sahara.c 	dev->hw_desc[idx]->next = 0;
next              824 drivers/crypto/sahara.c 			dev->hw_link[i]->next = 0;
next              826 drivers/crypto/sahara.c 			dev->hw_link[i]->next = dev->hw_phys_link[i + 1];
next              873 drivers/crypto/sahara.c 	dev->hw_link[i]->next = 0;
next              901 drivers/crypto/sahara.c 	dev->hw_link[index]->next = 0;
next             1011 drivers/crypto/sahara.c 		dev->hw_desc[0]->next = 0;
next             1017 drivers/crypto/sahara.c 		dev->hw_desc[0]->next = dev->hw_phys_desc[1];
next             1019 drivers/crypto/sahara.c 		dev->hw_desc[1]->next = 0;
next             1111 drivers/crypto/talitos.c 			goto next;
next             1135 drivers/crypto/talitos.c next:
next              194 drivers/dma-buf/sw_sync.c 	struct sync_pt *pt, *next;
next              202 drivers/dma-buf/sw_sync.c 	list_for_each_entry_safe(pt, next, &obj->pt_list, link) {
next              312 drivers/dma-buf/sw_sync.c 	struct sync_pt *pt, *next;
next              316 drivers/dma-buf/sw_sync.c 	list_for_each_entry_safe(pt, next, &obj->pt_list, link) {
next              231 drivers/dma/altera-msgdma.c 	struct msgdma_sw_desc *child, *next;
next              235 drivers/dma/altera-msgdma.c 	list_for_each_entry_safe(child, next, &desc->tx_list, node) {
next              249 drivers/dma/altera-msgdma.c 	struct msgdma_sw_desc *desc, *next;
next              251 drivers/dma/altera-msgdma.c 	list_for_each_entry_safe(desc, next, list, node)
next              534 drivers/dma/altera-msgdma.c 	struct msgdma_sw_desc *sdesc, *next;
next              538 drivers/dma/altera-msgdma.c 	list_for_each_entry_safe(sdesc, next, &desc->tx_list, node)
next              582 drivers/dma/altera-msgdma.c 	struct msgdma_sw_desc *desc, *next;
next              584 drivers/dma/altera-msgdma.c 	list_for_each_entry_safe(desc, next, &mdev->done_list, node) {
next              906 drivers/dma/amba-pl08x.c 	struct pl08x_dma_chan *p, *next;
next              909 drivers/dma/amba-pl08x.c 	next = NULL;
next              920 drivers/dma/amba-pl08x.c 			next = p;
next              924 drivers/dma/amba-pl08x.c 	if (!next && pl08x->has_slave) {
next              928 drivers/dma/amba-pl08x.c 				next = p;
next              936 drivers/dma/amba-pl08x.c 	if (next) {
next              943 drivers/dma/amba-pl08x.c 		spin_lock(&next->vc.lock);
next              945 drivers/dma/amba-pl08x.c 		success = next->state == PL08X_CHAN_WAITING;
next              947 drivers/dma/amba-pl08x.c 			pl08x_phy_reassign_start(plchan->phychan, next);
next              948 drivers/dma/amba-pl08x.c 		spin_unlock(&next->vc.lock);
next             2430 drivers/dma/amba-pl08x.c 	struct pl08x_dma_chan *next;
next             2433 drivers/dma/amba-pl08x.c 				 next, &dmadev->channels, vc.chan.device_node) {
next              117 drivers/dma/bcm-sba-raid.c 	struct list_head next;
next              232 drivers/dma/bcm-sba-raid.c 	INIT_LIST_HEAD(&req->next);
next              303 drivers/dma/bcm-sba-raid.c 	list_for_each_entry(nreq, &req->next, next)
next              317 drivers/dma/bcm-sba-raid.c 	list_add_tail(&req->next, &first->next);
next              442 drivers/dma/bcm-sba-raid.c 		list_for_each_entry(nreq, &first->next, next)
next              444 drivers/dma/bcm-sba-raid.c 		INIT_LIST_HEAD(&first->next);
next              540 drivers/dma/bcm-sba-raid.c 	list_for_each_entry(nreq, &req->next, next)
next             1509 drivers/dma/bcm-sba-raid.c 		INIT_LIST_HEAD(&req->next);
next               61 drivers/dma/bcm2835-dma.c 	uint32_t next;
next              344 drivers/dma/bcm2835-dma.c 		control_block->next = 0;
next              359 drivers/dma/bcm2835-dma.c 			d->cb_list[frame - 1].cb->next = cb_entry->paddr;
next              775 drivers/dma/bcm2835-dma.c 	d->cb_list[d->frames - 1].cb->next = d->cb_list[0].paddr;
next              849 drivers/dma/bcm2835-dma.c 	struct bcm2835_chan *c, *next;
next              851 drivers/dma/bcm2835-dma.c 	list_for_each_entry_safe(c, next, &od->ddev.channels,
next              116 drivers/dma/coh901318_lli.c 	struct coh901318_lli *next;
next              129 drivers/dma/coh901318_lli.c 		next = l->virt_link_addr;
next              132 drivers/dma/coh901318_lli.c 		l = next;
next               53 drivers/dma/dw/core.c 	return to_dw_desc(dwc->active_list.next);
next              175 drivers/dma/dw/core.c 	dwc->tx_node_active = dwc->tx_node_active->next;
next              231 drivers/dma/dw/core.c 	list_move(dwc->queue.next, &dwc->active_list);
next              333 drivers/dma/dw/core.c 				if (active == head->next)
next              445 drivers/dma/dw/core.c 	list_move(dwc->queue.next, dwc->active_list.prev);
next              120 drivers/dma/fsldma.c 			  struct fsl_dma_ld_hw *hw, dma_addr_t next)
next              126 drivers/dma/fsldma.c 	hw->next_ln_addr = CPU_TO_DMA(chan, snoop_bits | next, 64);
next              263 drivers/dma/img-mdc-dma.c 	struct mdc_hw_list_desc *curr, *next;
next              269 drivers/dma/img-mdc-dma.c 		next = curr->next_desc;
next              272 drivers/dma/img-mdc-dma.c 		curr = next;
next             1026 drivers/dma/img-mdc-dma.c 	struct mdc_chan *mchan, *next;
next             1031 drivers/dma/img-mdc-dma.c 	list_for_each_entry_safe(mchan, next, &mdma->dma_dev.channels,
next              647 drivers/dma/imx-dma.c 	list_move_tail(imxdmac->ld_active.next, &imxdmac->ld_free);
next              652 drivers/dma/imx-dma.c 		list_move_tail(imxdmac->ld_queue.next, &imxdmac->ld_active);
next              752 drivers/dma/imx-dma.c 	list_move_tail(imxdmac->ld_free.next, &imxdmac->ld_queue);
next             1017 drivers/dma/imx-dma.c 			list_move_tail(imxdmac->ld_queue.next,
next              423 drivers/dma/ioat/dma.c 		struct ioat_ring_ent *next = ring[i+1];
next              426 drivers/dma/ioat/dma.c 		hw->next = next->txd.phys;
next              428 drivers/dma/ioat/dma.c 	ring[i]->hw->next = ring[0]->txd.phys;
next              225 drivers/dma/ioat/dma.h 		(unsigned long long) hw->next, tx->cookie, tx->flags,
next               94 drivers/dma/ioat/hw.h 	uint64_t	next;
next              127 drivers/dma/ioat/hw.h 	uint64_t	next;
next              138 drivers/dma/ioat/hw.h 	uint64_t	next;
next              181 drivers/dma/ioat/hw.h 	uint64_t	next;
next              195 drivers/dma/ioat/hw.h 	uint64_t	next;
next              225 drivers/dma/ioat/hw.h 	uint64_t	next;
next              330 drivers/dma/ioat/init.c 	dma_chan = container_of(dma->channels.next, struct dma_chan,
next              843 drivers/dma/ioat/init.c 	dma_chan = container_of(dma->channels.next, struct dma_chan,
next              290 drivers/dma/ioat/prep.c 		(unsigned long long) (pq_ex ? pq_ex->next : pq->next),
next              300 drivers/dma/ioat/prep.c 	dev_dbg(dev, "\tNEXT: %#llx\n", pq->next);
next              323 drivers/dma/ioat/prep.c 		(unsigned long long) pq->next,
next               45 drivers/dma/iop-adma.c 		slot = list_entry(slot->slot_node.next,
next               91 drivers/dma/iop-adma.c 	if (desc->chain_node.next == &iop_chan->chain)
next              325 drivers/dma/iop-adma.c 					iter = list_entry(iter->slot_node.next,
next              460 drivers/dma/iop-adma.c 		iop_chan->last_used = list_entry(iop_chan->all_slots.next,
next              858 drivers/dma/iop-adma.c 	dma_chan = container_of(device->common.channels.next,
next              956 drivers/dma/iop-adma.c 	dma_chan = container_of(device->common.channels.next,
next             1118 drivers/dma/iop-adma.c 	dma_chan = container_of(device->common.channels.next,
next             1132 drivers/dma/ipu/ipu_idmac.c 	if ((*desc)->list.next == &ichan->queue)
next             1137 drivers/dma/ipu/ipu_idmac.c 	*desc = list_entry((*desc)->list.next, struct idmac_tx_desc, list);
next             1229 drivers/dma/ipu/ipu_idmac.c 	desc = list_entry(ichan->queue.next, struct idmac_tx_desc, list);
next             1357 drivers/dma/ipu/ipu_idmac.c 		desc = list_entry(ichan->free_list.next,
next              530 drivers/dma/mediatek/mtk-cqdma.c 		tx->next = NULL;
next              535 drivers/dma/mediatek/mtk-cqdma.c 			prev_tx->next = tx;
next              543 drivers/dma/mediatek/mtk-hsdma.c 	u16 next;
next              560 drivers/dma/mediatek/mtk-hsdma.c 		next = MTK_HSDMA_NEXT_DESP_IDX(pc->ring.cur_rptr,
next              562 drivers/dma/mediatek/mtk-hsdma.c 		rxd = &pc->ring.rxd[next];
next              572 drivers/dma/mediatek/mtk-hsdma.c 		cb = &pc->ring.cb[next];
next              612 drivers/dma/mediatek/mtk-hsdma.c 		pc->ring.cur_rptr = next;
next              311 drivers/dma/mv_xor.c 			iter = list_entry(mv_chan->chain.next,
next              321 drivers/dma/mv_xor.c 				iter = list_entry(iter->node.next,
next               97 drivers/dma/mxs-dma.c 	u32		next;
next              527 drivers/dma/mxs-dma.c 		ccw->next = mxs_chan->ccw_phys + sizeof(*ccw) * idx;
next              563 drivers/dma/mxs-dma.c 			ccw->next = mxs_chan->ccw_phys + sizeof(*ccw) * idx;
next              628 drivers/dma/mxs-dma.c 			ccw->next = mxs_chan->ccw_phys;
next              630 drivers/dma/mxs-dma.c 			ccw->next = mxs_chan->ccw_phys + sizeof(*ccw) * (i + 1);
next              128 drivers/dma/nbpfaxi.c 	u32	next;
next              817 drivers/dma/nbpfaxi.c 				prev->hwdesc->next = (u32)ldesc->hwdesc_dma_addr;
next              826 drivers/dma/nbpfaxi.c 	prev->hwdesc->next = 0;
next              345 drivers/dma/owl-dma.c 					   struct owl_dma_lli *next,
next              349 drivers/dma/owl-dma.c 		list_add_tail(&next->node, &txd->lli_list);
next              352 drivers/dma/owl-dma.c 		prev->hw.next_lli = next->phys;
next              356 drivers/dma/owl-dma.c 	return next;
next             1012 drivers/dma/owl-dma.c 	struct owl_dma_vchan *next;
next             1015 drivers/dma/owl-dma.c 				 next, &od->dma.channels, vc.chan.device_node) {
next               68 drivers/dma/pch_dma.c 	u32	next;
next              334 drivers/dma/pch_dma.c 		pd_chan->chan.chan_id, desc->regs.next);
next              340 drivers/dma/pch_dma.c 		channel_writel(pd_chan, NEXT, desc->regs.next);
next              604 drivers/dma/pch_dma.c 		desc->regs.next = DMA_DESC_FOLLOW_WITHOUT_IRQ;
next              629 drivers/dma/pch_dma.c 			prev->regs.next |= desc->txd.phys;
next              637 drivers/dma/pch_dma.c 		desc->regs.next = DMA_DESC_END_WITH_IRQ;
next              639 drivers/dma/pch_dma.c 		desc->regs.next = DMA_DESC_END_WITHOUT_IRQ;
next              756 drivers/dma/pch_dma.c 		pd->ch_regs[i].next = channel_readl(pd_chan, NEXT);
next              779 drivers/dma/pch_dma.c 		channel_writel(pd_chan, NEXT, pd->ch_regs[i].next);
next             2500 drivers/dma/pl330.c 		desc = list_entry(last->node.next, struct dma_pl330_desc, node);
next             2563 drivers/dma/pl330.c 		desc = list_entry(pool->next,
next             2702 drivers/dma/pl330.c 				desc = list_entry(first->node.next,
next             2818 drivers/dma/pl330.c 		desc = list_entry(first->node.next,
next              343 drivers/dma/ppc4xx/adma.c 			iter->hw_next = list_entry(iter->chain_node.next,
next              447 drivers/dma/ppc4xx/adma.c 	iter = list_entry(iter->chain_node.next,
next              451 drivers/dma/ppc4xx/adma.c 		iter = list_entry(iter->chain_node.next,
next              478 drivers/dma/ppc4xx/adma.c 			iter->hw_next = list_entry(iter->chain_node.next,
next             1453 drivers/dma/ppc4xx/adma.c 		slot = list_entry(slot->slot_node.next,
next             1742 drivers/dma/ppc4xx/adma.c 					iter = list_entry(iter->slot_node.next,
next             1812 drivers/dma/ppc4xx/adma.c 			list_entry(ppc440spe_chan->all_slots.next,
next             2118 drivers/dma/ppc4xx/adma.c 		iter->hw_next = list_entry(iter->chain_node.next,
next             2201 drivers/dma/ppc4xx/adma.c 		iter->hw_next = list_entry(iter->chain_node.next,
next             2224 drivers/dma/ppc4xx/adma.c 		iter->hw_next = list_entry(iter->chain_node.next,
next             2469 drivers/dma/ppc4xx/adma.c 					list_entry(iter->chain_node.next,
next             2642 drivers/dma/ppc4xx/adma.c 			iter->hw_next = list_entry(iter->chain_node.next,
next             2668 drivers/dma/ppc4xx/adma.c 				iter = list_entry(iter->chain_node.next,
next             2674 drivers/dma/ppc4xx/adma.c 			iter->hw_next = list_entry(iter->chain_node.next,
next             2995 drivers/dma/ppc4xx/adma.c 			iter = list_entry(iter->chain_node.next,
next             3007 drivers/dma/ppc4xx/adma.c 				iter = list_entry(iter->chain_node.next,
next             3076 drivers/dma/ppc4xx/adma.c 		end = list_entry(end->chain_node.next,
next             3251 drivers/dma/ppc4xx/adma.c 		desc = list_entry(hdesc->chain_node.next,
next             3372 drivers/dma/ppc4xx/adma.c 			desc = list_entry(desc->chain_node.next,
next             3413 drivers/dma/ppc4xx/adma.c 			desc = list_entry(desc->chain_node.next,
next              121 drivers/dma/qcom/hidma.c 	struct hidma_desc *next;
next              133 drivers/dma/qcom/hidma.c 	list_for_each_entry_safe(mdesc, next, &list, node) {
next              181 drivers/dma/qcom/hidma.c 	if (mdesc->node.next) {
next              240 drivers/dma/qcom/hidma.c 	struct hidma_desc *qdesc, *next;
next              244 drivers/dma/qcom/hidma.c 	list_for_each_entry_safe(qdesc, next, &mchan->queued, node) {
next              518 drivers/dma/s3c24xx-dma.c 	txd->at = txd->dsg_list.next;
next              584 drivers/dma/s3c24xx-dma.c 	struct s3c24xx_dma_chan *p, *next;
next              587 drivers/dma/s3c24xx-dma.c 	next = NULL;
next              592 drivers/dma/s3c24xx-dma.c 			next = p;
next              596 drivers/dma/s3c24xx-dma.c 	if (!next) {
next              601 drivers/dma/s3c24xx-dma.c 				next = p;
next              609 drivers/dma/s3c24xx-dma.c 	if (next) {
next              616 drivers/dma/s3c24xx-dma.c 		spin_lock(&next->vc.lock);
next              618 drivers/dma/s3c24xx-dma.c 		success = next->state == S3C24XX_DMA_CHAN_WAITING;
next              620 drivers/dma/s3c24xx-dma.c 			s3c24xx_dma_phy_reassign_start(s3cchan->phy, next);
next              621 drivers/dma/s3c24xx-dma.c 		spin_unlock(&next->vc.lock);
next              675 drivers/dma/s3c24xx-dma.c 			txd->at = txd->at->next;
next              695 drivers/dma/s3c24xx-dma.c 			txd->at = txd->dsg_list.next;
next             1135 drivers/dma/s3c24xx-dma.c 	struct s3c24xx_dma_chan *next;
next             1138 drivers/dma/s3c24xx-dma.c 				 next, &dmadev->channels, vc.chan.device_node) {
next              330 drivers/dma/st_fdma.c 	hw_node->next = 0;
next              485 drivers/dma/st_fdma.c 		hw_node->next = fdesc->node[(i + 1) % sg_len].pdesc;
next              534 drivers/dma/st_fdma.c 		hw_node->next = fdesc->node[(i + 1) % sg_len].pdesc;
next               50 drivers/dma/st_fdma.h 	u32 next;
next              223 drivers/dma/ste_dma40_ll.c 	dma_addr_t next = lli_phys;
next              243 drivers/dma/ste_dma40_ll.c 			next = cyclic ? first_phys : 0;
next              245 drivers/dma/ste_dma40_ll.c 			next = ALIGN(next + sizeof(struct d40_phy_lli),
next              248 drivers/dma/ste_dma40_ll.c 		err = d40_phy_fill_lli(lli, addr, size_seg, next,
next              312 drivers/dma/ste_dma40_ll.c 			     int next, unsigned int flags)
next              318 drivers/dma/ste_dma40_ll.c 	if (next != -EINVAL) {
next              319 drivers/dma/ste_dma40_ll.c 		slos = next * 2;
next              320 drivers/dma/ste_dma40_ll.c 		dlos = next * 2 + 1;
next              338 drivers/dma/ste_dma40_ll.c 			   int next, unsigned int flags)
next              340 drivers/dma/ste_dma40_ll.c 	d40_log_lli_link(lli_dst, lli_src, next, flags);
next              351 drivers/dma/ste_dma40_ll.c 			   int next, unsigned int flags)
next              353 drivers/dma/ste_dma40_ll.c 	d40_log_lli_link(lli_dst, lli_src, next, flags);
next              463 drivers/dma/ste_dma40_ll.h 			    int next, unsigned int flags);
next              468 drivers/dma/ste_dma40_ll.h 			    int next, unsigned int flags);
next              698 drivers/dma/stm32-mdma.c 	u32 next = count + 1;
next              720 drivers/dma/stm32-mdma.c 		hwdesc->clar = desc->node[next].hwdesc_phys;
next              366 drivers/dma/sun6i-dma.c 			       struct sun6i_dma_lli *next,
next              370 drivers/dma/sun6i-dma.c 	if ((!prev && !txd) || !next)
next              375 drivers/dma/sun6i-dma.c 		txd->v_lli = next;
next              378 drivers/dma/sun6i-dma.c 		prev->v_lli_next = next;
next              381 drivers/dma/sun6i-dma.c 	next->p_lli_next = LLI_LAST_ITEM;
next              382 drivers/dma/sun6i-dma.c 	next->v_lli_next = NULL;
next              384 drivers/dma/sun6i-dma.c 	return next;
next              188 drivers/dma/timb_dma.c 	td_desc = list_entry(td_chan->active_list.next, struct timb_dma_desc,
next              229 drivers/dma/timb_dma.c 	td_desc = list_entry(td_chan->active_list.next, struct timb_dma_desc,
next              267 drivers/dma/timb_dma.c 				list_entry(td_chan->active_list.next,
next              284 drivers/dma/timb_dma.c 	td_desc = list_entry(td_chan->queue.next, struct timb_dma_desc,
next              165 drivers/dma/txx9dmac.c 	return list_entry(dc->active_list.next,
next              177 drivers/dma/txx9dmac.c 	return list_entry(dc->queue.next, struct txx9dmac_desc, desc_node);
next              948 drivers/dma/txx9dmac.c 	desc = list_entry(list.next, struct txx9dmac_desc, desc_node);
next              694 drivers/dma/xilinx/xilinx_dma.c 	struct xilinx_vdma_tx_segment *segment, *next;
next              702 drivers/dma/xilinx/xilinx_dma.c 		list_for_each_entry_safe(segment, next, &desc->segments, node) {
next              733 drivers/dma/xilinx/xilinx_dma.c 	struct xilinx_dma_tx_descriptor *desc, *next;
next              735 drivers/dma/xilinx/xilinx_dma.c 	list_for_each_entry_safe(desc, next, list, node) {
next              820 drivers/dma/xilinx/xilinx_dma.c 	struct xilinx_dma_tx_descriptor *desc, *next;
next              825 drivers/dma/xilinx/xilinx_dma.c 	list_for_each_entry_safe(desc, next, &chan->done_list, node) {
next             1395 drivers/dma/xilinx/xilinx_dma.c 	struct xilinx_dma_tx_descriptor *desc, *next;
next             1401 drivers/dma/xilinx/xilinx_dma.c 	list_for_each_entry_safe(desc, next, &chan->active_list, node) {
next              432 drivers/dma/xilinx/zynqmp_dma.c 	struct zynqmp_dma_desc_sw *child, *next;
next              436 drivers/dma/xilinx/zynqmp_dma.c 	list_for_each_entry_safe(child, next, &sdesc->tx_list, node) {
next              450 drivers/dma/xilinx/zynqmp_dma.c 	struct zynqmp_dma_desc_sw *desc, *next;
next              452 drivers/dma/xilinx/zynqmp_dma.c 	list_for_each_entry_safe(desc, next, list, node)
next              599 drivers/dma/xilinx/zynqmp_dma.c 	struct zynqmp_dma_desc_sw *desc, *next;
next              601 drivers/dma/xilinx/zynqmp_dma.c 	list_for_each_entry_safe(desc, next, &chan->done_list, node) {
next             1658 drivers/edac/altera_edac.c 	list_add(&altdev->next, &altdev->edac->a10_ecc_devices);
next             2067 drivers/edac/altera_edac.c 	list_add(&altdev->next, &edac->a10_ecc_devices);
next             2144 drivers/edac/altera_edac.c 					next);
next              374 drivers/edac/altera_edac.h 	struct list_head next;
next              747 drivers/edac/sb_edac.c 	sbridge_dev = list_entry(prev ? prev->list.next
next              748 drivers/edac/sb_edac.c 				      : sbridge_edac_list.next, struct sbridge_dev, list);
next             1713 drivers/edac/sb_edac.c 				goto next;
next             1728 drivers/edac/sb_edac.c next:
next              132 drivers/edac/xgene_edac.c 	struct list_head	next;
next              421 drivers/edac/xgene_edac.c 	list_add(&ctx->next, &edac->mcus);
next              509 drivers/edac/xgene_edac.c 	struct list_head	next;
next              963 drivers/edac/xgene_edac.c 	list_add(&ctx->next, &edac->pmds);
next             1017 drivers/edac/xgene_edac.c 	struct list_head	next;
next             1246 drivers/edac/xgene_edac.c 	list_add(&ctx->next, &edac->l3s);
next             1787 drivers/edac/xgene_edac.c 	list_add(&ctx->next, &edac->socs);
next             1829 drivers/edac/xgene_edac.c 		list_for_each_entry(mcu, &ctx->mcus, next)
next             1833 drivers/edac/xgene_edac.c 	list_for_each_entry(pmd, &ctx->pmds, next) {
next             1838 drivers/edac/xgene_edac.c 	list_for_each_entry(node, &ctx->l3s, next)
next             1841 drivers/edac/xgene_edac.c 	list_for_each_entry(node, &ctx->socs, next)
next             1976 drivers/edac/xgene_edac.c 	list_for_each_entry_safe(mcu, temp_mcu, &edac->mcus, next)
next             1979 drivers/edac/xgene_edac.c 	list_for_each_entry_safe(pmd, temp_pmd, &edac->pmds, next)
next             1982 drivers/edac/xgene_edac.c 	list_for_each_entry_safe(node, temp_node, &edac->l3s, next)
next             1985 drivers/edac/xgene_edac.c 	list_for_each_entry_safe(node, temp_node, &edac->socs, next)
next             1050 drivers/firewire/core-cdev.c 	struct fw_cdev_iso_packet __user *p, *end, *next;
next             1119 drivers/firewire/core-cdev.c 		next = (struct fw_cdev_iso_packet __user *)
next             1121 drivers/firewire/core-cdev.c 		if (next > end)
next             1136 drivers/firewire/core-cdev.c 		p = next;
next              261 drivers/firewire/core-topology.c 				child = fw_node(child->link.next);
next              316 drivers/firewire/core-topology.c 	struct fw_node *node, *next, *child, *parent;
next              342 drivers/firewire/core-topology.c 	list_for_each_entry_safe(node, next, &list, link)
next              420 drivers/firewire/core-topology.c 	node0 = fw_node(list0.next);
next              421 drivers/firewire/core-topology.c 	node1 = fw_node(list1.next);
next              483 drivers/firewire/core-topology.c 		node0 = fw_node(node0->link.next);
next              484 drivers/firewire/core-topology.c 		next1 = fw_node(node1->link.next);
next              309 drivers/firewire/net.c 			fi2 = list_entry(fi->fi_link.next,
next              445 drivers/firewire/net.c 	fi = list_entry(pd->fi_list.next, struct fwnet_fragment_info, fi_link);
next               41 drivers/firewire/nosy.c 	__le32 next;
next              575 drivers/firewire/nosy.c 	lynx->rcv_start_pcl->next	= cpu_to_le32(lynx->rcv_pcl_bus);
next              576 drivers/firewire/nosy.c 	lynx->rcv_pcl->next		= cpu_to_le32(PCL_NEXT_INVALID);
next              898 drivers/firewire/ohci.c 	void *next;
next              901 drivers/firewire/ohci.c 		next = handle_ar_packet(ctx, p);
next              902 drivers/firewire/ohci.c 		if (!next)
next              904 drivers/firewire/ohci.c 		p = next;
next             1060 drivers/firewire/ohci.c 	desc = list_entry(ctx->buffer_list.next,
next             1074 drivers/firewire/ohci.c 			desc = list_entry(desc->list.next,
next             1145 drivers/firewire/ohci.c 	ctx->buffer_tail = list_entry(ctx->buffer_list.next,
next             1192 drivers/firewire/ohci.c 		if (desc->list.next == &ctx->buffer_list) {
next             1198 drivers/firewire/ohci.c 		desc = list_entry(desc->list.next,
next              307 drivers/firewire/sbp2.c 		struct sbp2_pointer next;
next              514 drivers/firewire/sbp2.c 	struct sbp2_orb *orb, *next;
next              523 drivers/firewire/sbp2.c 	list_for_each_entry_safe(orb, next, &list, link) {
next             1214 drivers/firewire/sbp2.c 	struct sbp2_logical_unit *lu, *next;
next             1222 drivers/firewire/sbp2.c 	list_for_each_entry_safe(lu, next, &tgt->lu_list, link) {
next             1457 drivers/firewire/sbp2.c 	orb->request.next.high = cpu_to_be32(SBP2_ORB_NULL);
next              641 drivers/firmware/dmi-sysfs.c 	struct dmi_sysfs_entry *entry, *next;
next              644 drivers/firmware/dmi-sysfs.c 	list_for_each_entry_safe(entry, next, &entry_list, list) {
next              951 drivers/firmware/dmi_scan.c 	for (d = head->next; d != &dmi_devices; d = d->next) {
next              193 drivers/firmware/efi/apple-properties.c 			pa_data = data->next;
next              110 drivers/firmware/efi/efi-pstore.c 					struct efivar_entry *next,
next              114 drivers/firmware/efi/efi-pstore.c 	if (&next->list != head)
next              115 drivers/firmware/efi/efi-pstore.c 		next->scanning = true;
next              146 drivers/firmware/efi/efi-pstore.c 				       struct efivar_entry *next,
next              155 drivers/firmware/efi/efi-pstore.c 		ret = __efi_pstore_scan_sysfs_exit(next, &next->list != head);
next              617 drivers/firmware/efi/efi.c 			prsv = rsv->next;
next             1007 drivers/firmware/efi/efi.c 	for (prsv = efi_memreserve_root->next; prsv; prsv = rsv->next) {
next             1043 drivers/firmware/efi/efi.c 	rsv->next = efi_memreserve_root->next;
next             1044 drivers/firmware/efi/efi.c 	efi_memreserve_root->next = __pa(rsv);
next              369 drivers/firmware/efi/esrt.c 	struct esre_entry *entry, *next;
next              371 drivers/firmware/efi/esrt.c 	list_for_each_entry_safe(entry, next, &entry_list, list) {
next               82 drivers/firmware/efi/libstub/arm-stub.c 	rsv->next = 0;
next              387 drivers/firmware/qemu_fw_cfg.c 	struct fw_cfg_sysfs_entry *entry, *next;
next              389 drivers/firmware/qemu_fw_cfg.c 	list_for_each_entry_safe(entry, next, &fw_cfg_entry_cache, list) {
next              564 drivers/firmware/qemu_fw_cfg.c 	struct kobject *k, *next;
next              566 drivers/firmware/qemu_fw_cfg.c 	list_for_each_entry_safe(k, next, &kset->list, entry)
next              211 drivers/fpga/fpga-bridge.c 	struct fpga_bridge *bridge, *next;
next              214 drivers/fpga/fpga-bridge.c 	list_for_each_entry_safe(bridge, next, bridge_list, node) {
next              257 drivers/gpio/gpiolib.c 	struct gpio_device *prev, *next;
next              265 drivers/gpio/gpiolib.c 	next = list_entry(gpio_devices.next, struct gpio_device, list);
next              266 drivers/gpio/gpiolib.c 	if (gdev->base + gdev->ngpio <= next->base) {
next              279 drivers/gpio/gpiolib.c 	list_for_each_entry_safe(prev, next, &gpio_devices, list) {
next              281 drivers/gpio/gpiolib.c 		if (&next->list == &gpio_devices)
next              286 drivers/gpio/gpiolib.c 				&& gdev->base + gdev->ngpio <= next->base) {
next             5019 drivers/gpio/gpiolib.c 		ret = list_entry(gdev->list.next, struct gpio_device, list);
next             5068 drivers/gpio/gpiolib.c 	.next = gpiolib_seq_next,
next             3394 drivers/gpu/drm/amd/amdgpu/amdgpu_device.c 	struct dma_fence *fence = NULL, *next = NULL;
next             3413 drivers/gpu/drm/amd/amdgpu/amdgpu_device.c 		r = amdgpu_bo_restore_shadow(shadow, &next);
next             3420 drivers/gpu/drm/amd/amdgpu/amdgpu_device.c 			fence = next;
next             3429 drivers/gpu/drm/amd/amdgpu/amdgpu_device.c 			fence = next;
next              714 drivers/gpu/drm/amd/amdgpu/amdgpu_gem.c 		for (base = robj->vm_bo; base; base = base->next)
next              156 drivers/gpu/drm/amd/amdgpu/amdgpu_gfx.c 		char *next;
next              171 drivers/gpu/drm/amd/amdgpu/amdgpu_gfx.c 		next = strchr(p, ',');
next              172 drivers/gpu/drm/amd/amdgpu/amdgpu_gfx.c 		if (!next)
next              174 drivers/gpu/drm/amd/amdgpu/amdgpu_gfx.c 		p = next + 1;
next              291 drivers/gpu/drm/amd/amdgpu/amdgpu_gmc.c 		fault = &gmc->fault_ring[fault->next];
next              304 drivers/gpu/drm/amd/amdgpu/amdgpu_gmc.c 	fault->next = gmc->fault_hash[hash].idx;
next               70 drivers/gpu/drm/amd/amdgpu/amdgpu_gmc.h 	uint64_t	next:AMDGPU_GMC_FAULT_RING_ORDER;
next              451 drivers/gpu/drm/amd/amdgpu/amdgpu_mn.c 	head = bo->mn_list.next;
next              118 drivers/gpu/drm/amd/amdgpu/amdgpu_sa.c 	if (sa_manager->hole->next == &sa_manager->olist)
next              121 drivers/gpu/drm/amd/amdgpu/amdgpu_sa.c 	sa_bo = list_entry(sa_manager->hole->next, struct amdgpu_sa_bo, olist);
next              145 drivers/gpu/drm/amd/amdgpu/amdgpu_sa.c 	if (hole->next != &sa_manager->olist) {
next              146 drivers/gpu/drm/amd/amdgpu/amdgpu_sa.c 		return list_entry(hole->next, struct amdgpu_sa_bo, olist)->soffset;
next              214 drivers/gpu/drm/amd/amdgpu/amdgpu_sa.c 	if (sa_manager->hole->next == &sa_manager->olist) {
next              341 drivers/gpu/drm/amd/amdgpu/amdgpu_ttm.c 		struct dma_fence *next;
next              379 drivers/gpu/drm/amd/amdgpu/amdgpu_ttm.c 				       resv, &next, false, true);
next              384 drivers/gpu/drm/amd/amdgpu/amdgpu_ttm.c 		fence = next;
next              297 drivers/gpu/drm/amd/amdgpu/amdgpu_vm.c 	base->next = NULL;
next              302 drivers/gpu/drm/amd/amdgpu/amdgpu_vm.c 	base->next = bo->vm_bo;
next              583 drivers/gpu/drm/amd/amdgpu/amdgpu_vm.c 	for (bo_base = abo->vm_bo; bo_base; bo_base = bo_base->next) {
next             1145 drivers/gpu/drm/amd/amdgpu/amdgpu_vm.c 	for (base = bo->vm_bo; base; base = base->next) {
next             2297 drivers/gpu/drm/amd/amdgpu/amdgpu_vm.c 	struct amdgpu_bo_va_mapping *before, *after, *tmp, *next;
next             2349 drivers/gpu/drm/amd/amdgpu/amdgpu_vm.c 	list_for_each_entry_safe(tmp, next, &removed, list) {
next             2445 drivers/gpu/drm/amd/amdgpu/amdgpu_vm.c 	struct amdgpu_bo_va_mapping *mapping, *next;
next             2455 drivers/gpu/drm/amd/amdgpu/amdgpu_vm.c 		     base = &(*base)->next) {
next             2459 drivers/gpu/drm/amd/amdgpu/amdgpu_vm.c 			*base = bo_va->base.next;
next             2468 drivers/gpu/drm/amd/amdgpu/amdgpu_vm.c 	list_for_each_entry_safe(mapping, next, &bo_va->valids, list) {
next             2475 drivers/gpu/drm/amd/amdgpu/amdgpu_vm.c 	list_for_each_entry_safe(mapping, next, &bo_va->invalids, list) {
next             2512 drivers/gpu/drm/amd/amdgpu/amdgpu_vm.c 	for (bo_base = bo->vm_bo; bo_base; bo_base = bo_base->next) {
next              138 drivers/gpu/drm/amd/amdgpu/amdgpu_vm.h 	struct amdgpu_vm_bo_base	*next;
next              809 drivers/gpu/drm/amd/amdkfd/kfd_device_queue_manager.c 	struct device_process_node *cur, *next;
next              817 drivers/gpu/drm/amd/amdkfd/kfd_device_queue_manager.c 	list_for_each_entry_safe(cur, next, &dqm->queues, list) {
next             1498 drivers/gpu/drm/amd/amdkfd/kfd_device_queue_manager.c 	struct queue *q, *next;
next             1506 drivers/gpu/drm/amd/amdkfd/kfd_device_queue_manager.c 	list_for_each_entry_safe(q, next, &qpd->queues_list, list) {
next             1572 drivers/gpu/drm/amd/amdkfd/kfd_device_queue_manager.c 	struct queue *q, *next;
next             1638 drivers/gpu/drm/amd/amdkfd/kfd_device_queue_manager.c 	list_for_each_entry_safe(q, next, &qpd->queues_list, list) {
next              150 drivers/gpu/drm/amd/amdkfd/kfd_process_queue_manager.c 	struct process_queue_node *pqn, *next;
next              152 drivers/gpu/drm/amd/amdkfd/kfd_process_queue_manager.c 	list_for_each_entry_safe(pqn, next, &pqm->queues, process_queue_list) {
next              143 drivers/gpu/drm/amd/amdkfd/kfd_topology.c 	while (dev->mem_props.next != &dev->mem_props) {
next              144 drivers/gpu/drm/amd/amdkfd/kfd_topology.c 		mem = container_of(dev->mem_props.next,
next              150 drivers/gpu/drm/amd/amdkfd/kfd_topology.c 	while (dev->cache_props.next != &dev->cache_props) {
next              151 drivers/gpu/drm/amd/amdkfd/kfd_topology.c 		cache = container_of(dev->cache_props.next,
next              157 drivers/gpu/drm/amd/amdkfd/kfd_topology.c 	while (dev->io_link_props.next != &dev->io_link_props) {
next              158 drivers/gpu/drm/amd/amdkfd/kfd_topology.c 		iolink = container_of(dev->io_link_props.next,
next              164 drivers/gpu/drm/amd/amdkfd/kfd_topology.c 	while (dev->perf_props.next != &dev->perf_props) {
next              165 drivers/gpu/drm/amd/amdkfd/kfd_topology.c 		perf = container_of(dev->perf_props.next,
next              836 drivers/gpu/drm/amd/amdkfd/kfd_topology.c 		list_move_tail(temp_list->next, master_list);
next               41 drivers/gpu/drm/amd/powerplay/inc/power_state.h 	struct pp_power_state *next;
next              116 drivers/gpu/drm/amd/powerplay/inc/smu_ucode_xfer_cz.h     uint16_t next;
next              340 drivers/gpu/drm/amd/powerplay/smumgr/smu8_smumgr.c 	task->next = is_last ? END_OF_TASK_LIST : smu8_smu->toc_entry_used_count;
next              377 drivers/gpu/drm/amd/powerplay/smumgr/smu8_smumgr.c 	task->next = is_last ? END_OF_TASK_LIST : smu8_smu->toc_entry_used_count;
next              428 drivers/gpu/drm/arm/display/komeda/komeda_private_obj.c 	struct drm_private_obj *obj, *next;
next              430 drivers/gpu/drm/arm/display/komeda/komeda_private_obj.c 	list_for_each_entry_safe(obj, next, &config->privobj_list, head)
next              197 drivers/gpu/drm/atmel-hlcdc/atmel_hlcdc_dc.h 	dma_addr_t next;
next              856 drivers/gpu/drm/atmel-hlcdc/atmel_hlcdc_plane.c 		dscr->next = dscr_dma;
next               35 drivers/gpu/drm/bridge/thc63lvd1024.c 	struct drm_bridge *next;
next               49 drivers/gpu/drm/bridge/thc63lvd1024.c 	return drm_bridge_attach(bridge->encoder, thc63->next, bridge);
next              147 drivers/gpu/drm/bridge/thc63lvd1024.c 	thc63->next = of_drm_find_bridge(remote);
next              149 drivers/gpu/drm/bridge/thc63lvd1024.c 	if (!thc63->next)
next              141 drivers/gpu/drm/drm_bridge.c 		previous->next = bridge;
next              201 drivers/gpu/drm/drm_bridge.c 	ret = ret && drm_bridge_mode_fixup(bridge->next, mode, adjusted_mode);
next              236 drivers/gpu/drm/drm_bridge.c 	return drm_bridge_mode_valid(bridge->next, mode);
next              255 drivers/gpu/drm/drm_bridge.c 	drm_bridge_disable(bridge->next);
next              280 drivers/gpu/drm/drm_bridge.c 	drm_bridge_post_disable(bridge->next);
next              306 drivers/gpu/drm/drm_bridge.c 	drm_bridge_mode_set(bridge->next, mode, adjusted_mode);
next              326 drivers/gpu/drm/drm_bridge.c 	drm_bridge_pre_enable(bridge->next);
next              351 drivers/gpu/drm/drm_bridge.c 	drm_bridge_enable(bridge->next);
next              373 drivers/gpu/drm/drm_bridge.c 	drm_atomic_bridge_disable(bridge->next, state);
next              406 drivers/gpu/drm/drm_bridge.c 	drm_atomic_bridge_post_disable(bridge->next, state);
next              429 drivers/gpu/drm/drm_bridge.c 	drm_atomic_bridge_pre_enable(bridge->next, state);
next              461 drivers/gpu/drm/drm_bridge.c 	drm_atomic_bridge_enable(bridge->next, state);
next              812 drivers/gpu/drm/drm_bufs.c 		buf->next = NULL;
next             1011 drivers/gpu/drm/drm_bufs.c 			buf->next = NULL;
next             1175 drivers/gpu/drm/drm_bufs.c 		buf->next = NULL;
next              683 drivers/gpu/drm/drm_connector.c 		if (lhead->next == &config->connector_list) {
next              688 drivers/gpu/drm/drm_connector.c 		lhead = lhead->next;
next              893 drivers/gpu/drm/drm_dp_mst_topology.c 			list_del(&txmsg->next);
next             1120 drivers/gpu/drm/drm_dp_mst_topology.c 	list_for_each_entry_safe(port, tmp, &mstb->ports, next) {
next             1121 drivers/gpu/drm/drm_dp_mst_topology.c 		list_del(&port->next);
next             1259 drivers/gpu/drm/drm_dp_mst_topology.c 			list_add(&port->next, &mgr->destroy_connector_list);
next             1354 drivers/gpu/drm/drm_dp_mst_topology.c 	list_for_each_entry(port, &mstb->ports, next) {
next             1389 drivers/gpu/drm/drm_dp_mst_topology.c 	list_for_each_entry(port, &mstb->ports, next) {
next             1426 drivers/gpu/drm/drm_dp_mst_topology.c 	list_for_each_entry(port, &mstb->ports, next) {
next             1670 drivers/gpu/drm/drm_dp_mst_topology.c 		list_add(&port->next, &mstb->ports);
next             1704 drivers/gpu/drm/drm_dp_mst_topology.c 			list_del(&port->next);
next             1781 drivers/gpu/drm/drm_dp_mst_topology.c 		list_for_each_entry(port, &mstb->ports, next) {
next             1812 drivers/gpu/drm/drm_dp_mst_topology.c 	list_for_each_entry(port, &mstb->ports, next) {
next             1854 drivers/gpu/drm/drm_dp_mst_topology.c 	list_for_each_entry(port, &mstb->ports, next) {
next             2062 drivers/gpu/drm/drm_dp_mst_topology.c 	txmsg = list_first_entry(&mgr->tx_msg_downq, struct drm_dp_sideband_msg_tx, next);
next             2066 drivers/gpu/drm/drm_dp_mst_topology.c 		list_del(&txmsg->next);
next             2069 drivers/gpu/drm/drm_dp_mst_topology.c 		list_del(&txmsg->next);
next             2100 drivers/gpu/drm/drm_dp_mst_topology.c 	list_add_tail(&txmsg->next, &mgr->tx_msg_downq);
next             3254 drivers/gpu/drm/drm_dp_mst_topology.c 	list_for_each_entry(pos, &topology_state->vcpis, next) {
next             3290 drivers/gpu/drm/drm_dp_mst_topology.c 		list_add(&vcpi->next, &topology_state->vcpis);
next             3337 drivers/gpu/drm/drm_dp_mst_topology.c 	list_for_each_entry(pos, &topology_state->vcpis, next) {
next             3620 drivers/gpu/drm/drm_dp_mst_topology.c 	list_for_each_entry(port, &mstb->ports, next) {
next             3751 drivers/gpu/drm/drm_dp_mst_topology.c 		port = list_first_entry_or_null(&mgr->destroy_connector_list, struct drm_dp_mst_port, next);
next             3756 drivers/gpu/drm/drm_dp_mst_topology.c 		list_del(&port->next);
next             3759 drivers/gpu/drm/drm_dp_mst_topology.c 		INIT_LIST_HEAD(&port->next);
next             3788 drivers/gpu/drm/drm_dp_mst_topology.c 	list_for_each_entry(pos, &old_state->vcpis, next) {
next             3798 drivers/gpu/drm/drm_dp_mst_topology.c 		list_add(&vcpi->next, &state->vcpis);
next             3804 drivers/gpu/drm/drm_dp_mst_topology.c 	list_for_each_entry_safe(pos, vcpi, &state->vcpis, next) {
next             3820 drivers/gpu/drm/drm_dp_mst_topology.c 	list_for_each_entry_safe(pos, tmp, &mst_state->vcpis, next) {
next             3837 drivers/gpu/drm/drm_dp_mst_topology.c 	list_for_each_entry(vcpi, &mst_state->vcpis, next) {
next              170 drivers/gpu/drm/drm_encoder.c 		struct drm_bridge *next;
next              173 drivers/gpu/drm/drm_encoder.c 			next = bridge->next;
next              175 drivers/gpu/drm/drm_encoder.c 			bridge = next;
next              344 drivers/gpu/drm/drm_file.c 			struct pci_bus *b = list_entry(pci_root_buses.next,
next              206 drivers/gpu/drm/drm_syncobj.c 	if (!wait->node.next)
next              370 drivers/gpu/drm/drm_syncobj.c 	if (wait.node.next)
next              112 drivers/gpu/drm/exynos/exynos_dp.c 			bridge->next = NULL;
next              508 drivers/gpu/drm/gma500/mmu.c 	unsigned long next;
next              531 drivers/gpu/drm/gma500/mmu.c 			next = psb_pd_addr_end(addr, end);
next              538 drivers/gpu/drm/gma500/mmu.c 				 (addr & clflush_mask) < next);
next              541 drivers/gpu/drm/gma500/mmu.c 		} while (addr = next, next != end);
next              561 drivers/gpu/drm/gma500/mmu.c 	unsigned long next;
next              570 drivers/gpu/drm/gma500/mmu.c 		next = psb_pd_addr_end(addr, end);
next              577 drivers/gpu/drm/gma500/mmu.c 		} while (addr += PAGE_SIZE, addr < next);
next              580 drivers/gpu/drm/gma500/mmu.c 	} while (addr = next, next != end);
next              603 drivers/gpu/drm/gma500/mmu.c 	unsigned long next;
next              626 drivers/gpu/drm/gma500/mmu.c 			next = psb_pd_addr_end(addr, end);
next              634 drivers/gpu/drm/gma500/mmu.c 			} while (addr += PAGE_SIZE, addr < next);
next              637 drivers/gpu/drm/gma500/mmu.c 		} while (addr = next, next != end);
next              658 drivers/gpu/drm/gma500/mmu.c 	unsigned long next;
next              668 drivers/gpu/drm/gma500/mmu.c 		next = psb_pd_addr_end(addr, end);
next              678 drivers/gpu/drm/gma500/mmu.c 		} while (addr += PAGE_SIZE, addr < next);
next              681 drivers/gpu/drm/gma500/mmu.c 	} while (addr = next, next != end);
next              707 drivers/gpu/drm/gma500/mmu.c 	unsigned long next;
next              732 drivers/gpu/drm/gma500/mmu.c 			next = psb_pd_addr_end(addr, end);
next              741 drivers/gpu/drm/gma500/mmu.c 			} while (addr += PAGE_SIZE, addr < next);
next              744 drivers/gpu/drm/gma500/mmu.c 		} while (addr = next, next != end);
next             13864 drivers/gpu/drm/i915/display/intel_display.c 	struct intel_atomic_state *state, *next;
next             13868 drivers/gpu/drm/i915/display/intel_display.c 	llist_for_each_entry_safe(state, next, freed, freed)
next              499 drivers/gpu/drm/i915/display/intel_dsi_vbt.c 			const u8 *next = data + operation_size;
next              504 drivers/gpu/drm/i915/display/intel_dsi_vbt.c 			if (operation_size && data != next) {
next             1859 drivers/gpu/drm/i915/gem/i915_gem_execbuffer.c 				capture->next = eb->request->capture_list;
next               16 drivers/gpu/drm/i915/gem/i915_gem_pm.c 	struct llist_node *node, *next;
next               18 drivers/gpu/drm/i915/gem/i915_gem_pm.c 	llist_for_each_safe(node, next, llist_del_all(&engine->barrier_tasks)) {
next              426 drivers/gpu/drm/i915/gem/i915_gem_shrinker.c 	struct i915_vma *vma, *next;
next              442 drivers/gpu/drm/i915/gem/i915_gem_shrinker.c 	list_for_each_entry_safe(vma, next,
next              136 drivers/gpu/drm/i915/gt/intel_breadcrumbs.c 	struct list_head *pos, *next;
next              147 drivers/gpu/drm/i915/gt/intel_breadcrumbs.c 		list_for_each_safe(pos, next, &ce->signals) {
next              187 drivers/gpu/drm/i915/gt/intel_breadcrumbs.c 	list_for_each_safe(pos, next, &signal) {
next              254 drivers/gpu/drm/i915/gt/intel_engine.h 				       u32 next, u32 prev)
next              256 drivers/gpu/drm/i915/gt/intel_engine.h 	typecheck(typeof(ring->size), next);
next              258 drivers/gpu/drm/i915/gt/intel_engine.h 	return (next - prev) << ring->wrap;
next               79 drivers/gpu/drm/i915/gt/intel_engine_user.c 	struct llist_node *pos, *next;
next               81 drivers/gpu/drm/i915/gt/intel_engine_user.c 	llist_for_each_safe(pos, next, get_engines(i915)) {
next              197 drivers/gpu/drm/i915/gt/intel_engine_user.c 	struct list_head *it, *next;
next              205 drivers/gpu/drm/i915/gt/intel_engine_user.c 	list_for_each_safe(it, next, &engines) {
next              360 drivers/gpu/drm/i915/gt/intel_lrc.c 			struct i915_request *next;
next              363 drivers/gpu/drm/i915/gt/intel_lrc.c 			next = READ_ONCE(ve->request);
next              364 drivers/gpu/drm/i915/gt/intel_lrc.c 			if (next)
next              365 drivers/gpu/drm/i915/gt/intel_lrc.c 				preempt = rq_prio(next) > last_prio;
next              388 drivers/gpu/drm/i915/gt/intel_lrc.c 		      const struct i915_request *next)
next              400 drivers/gpu/drm/i915/gt/intel_lrc.c 	return rq_prio(prev) >= rq_prio(next);
next              597 drivers/gpu/drm/i915/gt/intel_lrc.c 	struct i915_request *next = READ_ONCE(ve->request);
next              599 drivers/gpu/drm/i915/gt/intel_lrc.c 	if (next && next->execution_mask & ~rq->execution_mask)
next              804 drivers/gpu/drm/i915/gt/intel_lrc.c 			  const struct intel_context *next)
next              806 drivers/gpu/drm/i915/gt/intel_lrc.c 	if (prev != next)
next              816 drivers/gpu/drm/i915/gt/intel_lrc.c 			 const struct i915_request *next)
next              818 drivers/gpu/drm/i915/gt/intel_lrc.c 	GEM_BUG_ON(prev == next);
next              819 drivers/gpu/drm/i915/gt/intel_lrc.c 	GEM_BUG_ON(!assert_priority_queue(prev, next));
next              829 drivers/gpu/drm/i915/gt/intel_lrc.c 	if (i915_request_completed(next))
next              832 drivers/gpu/drm/i915/gt/intel_lrc.c 	if (!can_merge_ctx(prev->hw_context, next->hw_context))
next               94 drivers/gpu/drm/i915/gvt/debugfs.c 	struct diff_mmio *node, *next;
next              113 drivers/gpu/drm/i915/gvt/debugfs.c 	list_for_each_entry_safe(node, next, &param.diff_mmio_list, node) {
next              406 drivers/gpu/drm/i915/gvt/execlist.c 	struct list_head *next = workload_q_head(vgpu, ring_id)->next;
next              419 drivers/gpu/drm/i915/gvt/execlist.c 		next_workload = container_of(next,
next             1531 drivers/gpu/drm/i915/gvt/gtt.c 		oos_page = container_of(gtt->oos_page_use_list_head.next,
next             1540 drivers/gpu/drm/i915/gvt/gtt.c 		oos_page = container_of(gtt->oos_page_free_list_head.next,
next             2501 drivers/gpu/drm/i915/gvt/gtt.c 	struct intel_gvt_partial_pte *pos, *next;
next             2503 drivers/gpu/drm/i915/gvt/gtt.c 	list_for_each_entry_safe(pos, next,
next              622 drivers/gpu/drm/i915/gvt/handlers.c 	unsigned int next, unsigned int end, i915_reg_t i915_end)
next              624 drivers/gpu/drm/i915/gvt/handlers.c 	unsigned int range = next - start;
next              391 drivers/gpu/drm/i915/gvt/mmio_context.c static void switch_mocs(struct intel_vgpu *pre, struct intel_vgpu *next,
next              407 drivers/gpu/drm/i915/gvt/mmio_context.c 	dev_priv = pre ? pre->gvt->dev_priv : next->gvt->dev_priv;
next              423 drivers/gpu/drm/i915/gvt/mmio_context.c 		if (next)
next              424 drivers/gpu/drm/i915/gvt/mmio_context.c 			new_v = vgpu_vreg_t(next, offset);
next              441 drivers/gpu/drm/i915/gvt/mmio_context.c 			if (next)
next              442 drivers/gpu/drm/i915/gvt/mmio_context.c 				new_v = vgpu_vreg_t(next, l3_offset);
next              468 drivers/gpu/drm/i915/gvt/mmio_context.c 			struct intel_vgpu *next,
next              476 drivers/gpu/drm/i915/gvt/mmio_context.c 	dev_priv = pre ? pre->gvt->dev_priv : next->gvt->dev_priv;
next              478 drivers/gpu/drm/i915/gvt/mmio_context.c 		switch_mocs(pre, next, ring_id);
next              503 drivers/gpu/drm/i915/gvt/mmio_context.c 		if (next) {
next              504 drivers/gpu/drm/i915/gvt/mmio_context.c 			s = &next->submission;
next              515 drivers/gpu/drm/i915/gvt/mmio_context.c 				new_v = vgpu_vreg_t(next, mmio->reg) |
next              518 drivers/gpu/drm/i915/gvt/mmio_context.c 				new_v = vgpu_vreg_t(next, mmio->reg);
next              531 drivers/gpu/drm/i915/gvt/mmio_context.c 				  next ? next->id : 0,
next              537 drivers/gpu/drm/i915/gvt/mmio_context.c 	if (next)
next              538 drivers/gpu/drm/i915/gvt/mmio_context.c 		handle_tlb_pending_event(next, ring_id);
next              551 drivers/gpu/drm/i915/gvt/mmio_context.c 			   struct intel_vgpu *next, int ring_id)
next              555 drivers/gpu/drm/i915/gvt/mmio_context.c 	if (WARN_ON(!pre && !next))
next              559 drivers/gpu/drm/i915/gvt/mmio_context.c 		       pre ? "vGPU" : "host", next ? "vGPU" : "HOST");
next              561 drivers/gpu/drm/i915/gvt/mmio_context.c 	dev_priv = pre ? pre->gvt->dev_priv : next->gvt->dev_priv;
next              569 drivers/gpu/drm/i915/gvt/mmio_context.c 	switch_mmio(pre, next, ring_id);
next               48 drivers/gpu/drm/i915/gvt/mmio_context.h 			   struct intel_vgpu *next, int ring_id);
next              785 drivers/gpu/drm/i915/gvt/scheduler.c 			workload_q_head(scheduler->current_vgpu, ring_id)->next,
next              265 drivers/gpu/drm/i915/i915_active.c 	struct llist_node *pos, *next;
next              284 drivers/gpu/drm/i915/i915_active.c 	llist_for_each_safe(pos, next, llist_del_all(&engine->barrier_tasks)) {
next              290 drivers/gpu/drm/i915/i915_active.c 		pos->next = head;
next              590 drivers/gpu/drm/i915/i915_active.c 	struct llist_node *pos, *next;
next              646 drivers/gpu/drm/i915/i915_active.c 	llist_for_each_safe(pos, next, take_preallocated_barriers(ref)) {
next              659 drivers/gpu/drm/i915/i915_active.c 	struct llist_node *pos, *next;
next              670 drivers/gpu/drm/i915/i915_active.c 	llist_for_each_safe(pos, next, take_preallocated_barriers(ref)) {
next              700 drivers/gpu/drm/i915/i915_active.c 	struct llist_node *node, *next;
next              710 drivers/gpu/drm/i915/i915_active.c 	llist_for_each_safe(node, next, llist_del_all(&engine->barrier_tasks)) {
next              101 drivers/gpu/drm/i915/i915_gem_evict.c 	struct i915_vma *vma, *next;
next              142 drivers/gpu/drm/i915/i915_gem_evict.c 	list_for_each_entry_safe(vma, next, &vm->bound_list, vm_link) {
next              180 drivers/gpu/drm/i915/i915_gem_evict.c 	list_for_each_entry_safe(vma, next, &eviction_list, evict_link) {
next              226 drivers/gpu/drm/i915/i915_gem_evict.c 	list_for_each_entry_safe(vma, next, &eviction_list, evict_link) {
next              235 drivers/gpu/drm/i915/i915_gem_evict.c 	list_for_each_entry_safe(vma, next, &eviction_list, evict_link) {
next              268 drivers/gpu/drm/i915/i915_gem_evict.c 	struct i915_vma *vma, *next;
next              351 drivers/gpu/drm/i915/i915_gem_evict.c 	list_for_each_entry_safe(vma, next, &eviction_list, evict_link) {
next              375 drivers/gpu/drm/i915/i915_gem_evict.c 	struct i915_vma *vma, *next;
next              404 drivers/gpu/drm/i915/i915_gem_evict.c 	list_for_each_entry_safe(vma, next, &eviction_list, evict_link) {
next               57 drivers/gpu/drm/i915/i915_globals.c 	struct i915_global *global, *next;
next               59 drivers/gpu/drm/i915/i915_globals.c 	list_for_each_entry_safe_reverse(global, next, &globals, link)
next              684 drivers/gpu/drm/i915/i915_gpu_error.c 	for (ee = error->engine; ee; ee = ee->next)
next              737 drivers/gpu/drm/i915/i915_gpu_error.c 	for (ee = error->engine; ee; ee = ee->next)
next              740 drivers/gpu/drm/i915/i915_gpu_error.c 	for (ee = error->engine; ee; ee = ee->next) {
next              929 drivers/gpu/drm/i915/i915_gpu_error.c 		error->engine = ee->next;
next             1275 drivers/gpu/drm/i915/i915_gpu_error.c 	struct capture_vma *next;
next             1280 drivers/gpu/drm/i915/i915_gpu_error.c capture_vma(struct capture_vma *next,
next             1288 drivers/gpu/drm/i915/i915_gpu_error.c 		return next;
next             1292 drivers/gpu/drm/i915/i915_gpu_error.c 		return next;
next             1296 drivers/gpu/drm/i915/i915_gpu_error.c 		return next;
next             1302 drivers/gpu/drm/i915/i915_gpu_error.c 	c->next = next;
next             1316 drivers/gpu/drm/i915/i915_gpu_error.c 	for (c = request->capture_list; c; c = c->next)
next             1331 drivers/gpu/drm/i915/i915_gpu_error.c 	for (c = request->capture_list; c; c = c->next) {
next             1437 drivers/gpu/drm/i915/i915_gpu_error.c 			capture = this->next;
next             1456 drivers/gpu/drm/i915/i915_gpu_error.c 		ee->next = error->engine;
next             1655 drivers/gpu/drm/i915/i915_gpu_error.c 	for (ee = error->engine; ee; ee = ee->next) {
next              163 drivers/gpu/drm/i915/i915_gpu_error.h 		struct drm_i915_error_engine *next;
next              190 drivers/gpu/drm/i915/i915_request.c 		struct i915_capture_list *next = capture->next;
next              193 drivers/gpu/drm/i915/i915_request.c 		capture = next;
next              220 drivers/gpu/drm/i915/i915_request.c 	struct i915_active_request *active, *next;
next              256 drivers/gpu/drm/i915/i915_request.c 	list_for_each_entry_safe(active, next, &rq->active_list, link) {
next              267 drivers/gpu/drm/i915/i915_request.c 		prefetchw(next);
next               49 drivers/gpu/drm/i915/i915_request.h 	struct i915_capture_list *next;
next              305 drivers/gpu/drm/i915/i915_scheduler.c 		if (stack.dfs_link.next == stack.dfs_link.prev)
next              135 drivers/gpu/drm/i915/i915_sw_fence.c 	wait_queue_entry_t *pos, *next;
next              150 drivers/gpu/drm/i915/i915_sw_fence.c 		list_for_each_entry_safe(pos, next, &x->head, entry) {
next              160 drivers/gpu/drm/i915/i915_sw_fence.c 			list_for_each_entry_safe(pos, next, &x->head, entry) {
next              279 drivers/gpu/drm/i915/i915_syncmap.c 		struct i915_syncmap *next;
next              285 drivers/gpu/drm/i915/i915_syncmap.c 			next = kzalloc(sizeof(*next) + KSYNCMAP * sizeof(next),
next              287 drivers/gpu/drm/i915/i915_syncmap.c 			if (unlikely(!next))
next              293 drivers/gpu/drm/i915/i915_syncmap.c 			next->height = above + p->height;
next              294 drivers/gpu/drm/i915/i915_syncmap.c 			next->prefix = __sync_branch_prefix(next, id);
next              299 drivers/gpu/drm/i915/i915_syncmap.c 				__sync_child(p->parent)[idx] = next;
next              302 drivers/gpu/drm/i915/i915_syncmap.c 			next->parent = p->parent;
next              306 drivers/gpu/drm/i915/i915_syncmap.c 			__sync_set_child(next, idx, p);
next              307 drivers/gpu/drm/i915/i915_syncmap.c 			p->parent = next;
next              310 drivers/gpu/drm/i915/i915_syncmap.c 			p = next;
next              319 drivers/gpu/drm/i915/i915_syncmap.c 		next = __sync_child(p)[idx];
next              320 drivers/gpu/drm/i915/i915_syncmap.c 		if (!next) {
next              321 drivers/gpu/drm/i915/i915_syncmap.c 			next = __sync_alloc_leaf(p, id);
next              322 drivers/gpu/drm/i915/i915_syncmap.c 			if (unlikely(!next))
next              325 drivers/gpu/drm/i915/i915_syncmap.c 			__sync_set_child(p, idx, next);
next              326 drivers/gpu/drm/i915/i915_syncmap.c 			p = next;
next              330 drivers/gpu/drm/i915/i915_syncmap.c 		p = next;
next               26 drivers/gpu/drm/i915/i915_user_extensions.c 		u64 next;
next               53 drivers/gpu/drm/i915/i915_user_extensions.c 		if (get_user(next, &ext->next_extension) ||
next               54 drivers/gpu/drm/i915/i915_user_extensions.c 		    overflows_type(next, ext))
next               57 drivers/gpu/drm/i915/i915_user_extensions.c 		ext = u64_to_user_ptr(next);
next              245 drivers/gpu/drm/i915/i915_utils.h 	WRITE_ONCE(head->next, first);
next              837 drivers/gpu/drm/i915/i915_vma.c 	struct i915_vma *vma, *next;
next              840 drivers/gpu/drm/i915/i915_vma.c 	list_for_each_entry_safe(vma, next, &i915->gt.closed_vma, closed_link) {
next              390 drivers/gpu/drm/i915/selftests/i915_gem_evict.c 		struct reserved *next;
next              442 drivers/gpu/drm/i915/selftests/i915_gem_evict.c 		r->next = reserved;
next              516 drivers/gpu/drm/i915/selftests/i915_gem_evict.c 		struct reserved *next = reserved->next;
next              521 drivers/gpu/drm/i915/selftests/i915_gem_evict.c 		reserved = next;
next              147 drivers/gpu/drm/ingenic/ingenic-drm.c 	u32 next;
next              353 drivers/gpu/drm/ingenic/ingenic-drm.c 		regmap_write(priv->map, JZ_REG_LCD_DA0, priv->dma_hwdesc->next);
next              696 drivers/gpu/drm/ingenic/ingenic-drm.c 	priv->dma_hwdesc->next = priv->dma_hwdesc_phys;
next              198 drivers/gpu/drm/mediatek/mtk_drm_ddp.c 				    enum mtk_ddp_comp_id next,
next              203 drivers/gpu/drm/mediatek/mtk_drm_ddp.c 	if (cur == DDP_COMPONENT_OVL0 && next == DDP_COMPONENT_COLOR0) {
next              206 drivers/gpu/drm/mediatek/mtk_drm_ddp.c 	} else if (cur == DDP_COMPONENT_OVL0 && next == DDP_COMPONENT_RDMA0) {
next              209 drivers/gpu/drm/mediatek/mtk_drm_ddp.c 	} else if (cur == DDP_COMPONENT_OD0 && next == DDP_COMPONENT_RDMA0) {
next              212 drivers/gpu/drm/mediatek/mtk_drm_ddp.c 	} else if (cur == DDP_COMPONENT_UFOE && next == DDP_COMPONENT_DSI0) {
next              215 drivers/gpu/drm/mediatek/mtk_drm_ddp.c 	} else if (cur == DDP_COMPONENT_OVL1 && next == DDP_COMPONENT_COLOR1) {
next              218 drivers/gpu/drm/mediatek/mtk_drm_ddp.c 	} else if (cur == DDP_COMPONENT_GAMMA && next == DDP_COMPONENT_RDMA1) {
next              221 drivers/gpu/drm/mediatek/mtk_drm_ddp.c 	} else if (cur == DDP_COMPONENT_OD1 && next == DDP_COMPONENT_RDMA1) {
next              224 drivers/gpu/drm/mediatek/mtk_drm_ddp.c 	} else if (cur == DDP_COMPONENT_RDMA0 && next == DDP_COMPONENT_DPI0) {
next              227 drivers/gpu/drm/mediatek/mtk_drm_ddp.c 	} else if (cur == DDP_COMPONENT_RDMA0 && next == DDP_COMPONENT_DPI1) {
next              230 drivers/gpu/drm/mediatek/mtk_drm_ddp.c 	} else if (cur == DDP_COMPONENT_RDMA0 && next == DDP_COMPONENT_DSI1) {
next              233 drivers/gpu/drm/mediatek/mtk_drm_ddp.c 	} else if (cur == DDP_COMPONENT_RDMA0 && next == DDP_COMPONENT_DSI2) {
next              236 drivers/gpu/drm/mediatek/mtk_drm_ddp.c 	} else if (cur == DDP_COMPONENT_RDMA0 && next == DDP_COMPONENT_DSI3) {
next              239 drivers/gpu/drm/mediatek/mtk_drm_ddp.c 	} else if (cur == DDP_COMPONENT_RDMA1 && next == DDP_COMPONENT_DSI1) {
next              242 drivers/gpu/drm/mediatek/mtk_drm_ddp.c 	} else if (cur == DDP_COMPONENT_RDMA1 && next == DDP_COMPONENT_DSI2) {
next              245 drivers/gpu/drm/mediatek/mtk_drm_ddp.c 	} else if (cur == DDP_COMPONENT_RDMA1 && next == DDP_COMPONENT_DSI3) {
next              248 drivers/gpu/drm/mediatek/mtk_drm_ddp.c 	} else if (cur == DDP_COMPONENT_RDMA1 && next == DDP_COMPONENT_DPI0) {
next              251 drivers/gpu/drm/mediatek/mtk_drm_ddp.c 	} else if (cur == DDP_COMPONENT_RDMA1 && next == DDP_COMPONENT_DPI1) {
next              254 drivers/gpu/drm/mediatek/mtk_drm_ddp.c 	} else if (cur == DDP_COMPONENT_RDMA2 && next == DDP_COMPONENT_DPI0) {
next              257 drivers/gpu/drm/mediatek/mtk_drm_ddp.c 	} else if (cur == DDP_COMPONENT_RDMA2 && next == DDP_COMPONENT_DPI1) {
next              260 drivers/gpu/drm/mediatek/mtk_drm_ddp.c 	} else if (cur == DDP_COMPONENT_RDMA2 && next == DDP_COMPONENT_DSI1) {
next              263 drivers/gpu/drm/mediatek/mtk_drm_ddp.c 	} else if (cur == DDP_COMPONENT_RDMA2 && next == DDP_COMPONENT_DSI2) {
next              266 drivers/gpu/drm/mediatek/mtk_drm_ddp.c 	} else if (cur == DDP_COMPONENT_RDMA2 && next == DDP_COMPONENT_DSI3) {
next              277 drivers/gpu/drm/mediatek/mtk_drm_ddp.c 				   enum mtk_ddp_comp_id next,
next              282 drivers/gpu/drm/mediatek/mtk_drm_ddp.c 	if (cur == DDP_COMPONENT_OVL0 && next == DDP_COMPONENT_COLOR0) {
next              285 drivers/gpu/drm/mediatek/mtk_drm_ddp.c 	} else if (cur == DDP_COMPONENT_RDMA1 && next == DDP_COMPONENT_DPI0) {
next              288 drivers/gpu/drm/mediatek/mtk_drm_ddp.c 	} else if (cur == DDP_COMPONENT_RDMA1 && next == DDP_COMPONENT_DPI1) {
next              291 drivers/gpu/drm/mediatek/mtk_drm_ddp.c 	} else if (cur == DDP_COMPONENT_RDMA1 && next == DDP_COMPONENT_DSI0) {
next              294 drivers/gpu/drm/mediatek/mtk_drm_ddp.c 	} else if (cur == DDP_COMPONENT_RDMA1 && next == DDP_COMPONENT_DSI1) {
next              297 drivers/gpu/drm/mediatek/mtk_drm_ddp.c 	} else if (cur == DDP_COMPONENT_RDMA1 && next == DDP_COMPONENT_DSI2) {
next              300 drivers/gpu/drm/mediatek/mtk_drm_ddp.c 	} else if (cur == DDP_COMPONENT_RDMA1 && next == DDP_COMPONENT_DSI3) {
next              303 drivers/gpu/drm/mediatek/mtk_drm_ddp.c 	} else if (cur == DDP_COMPONENT_RDMA2 && next == DDP_COMPONENT_DPI0) {
next              306 drivers/gpu/drm/mediatek/mtk_drm_ddp.c 	} else if (cur == DDP_COMPONENT_RDMA2 && next == DDP_COMPONENT_DPI1) {
next              309 drivers/gpu/drm/mediatek/mtk_drm_ddp.c 	} else if (cur == DDP_COMPONENT_RDMA2 && next == DDP_COMPONENT_DSI0) {
next              312 drivers/gpu/drm/mediatek/mtk_drm_ddp.c 	} else if (cur == DDP_COMPONENT_RDMA2 && next == DDP_COMPONENT_DSI1) {
next              315 drivers/gpu/drm/mediatek/mtk_drm_ddp.c 	} else if (cur == DDP_COMPONENT_RDMA2 && next == DDP_COMPONENT_DSI2) {
next              318 drivers/gpu/drm/mediatek/mtk_drm_ddp.c 	} else if (cur == DDP_COMPONENT_RDMA2 && next == DDP_COMPONENT_DSI3) {
next              321 drivers/gpu/drm/mediatek/mtk_drm_ddp.c 	} else if (cur == DDP_COMPONENT_OVL1 && next == DDP_COMPONENT_COLOR1) {
next              324 drivers/gpu/drm/mediatek/mtk_drm_ddp.c 	} else if (cur == DDP_COMPONENT_BLS && next == DDP_COMPONENT_DSI0) {
next              336 drivers/gpu/drm/mediatek/mtk_drm_ddp.c 			     enum mtk_ddp_comp_id next)
next              338 drivers/gpu/drm/mediatek/mtk_drm_ddp.c 	if (cur == DDP_COMPONENT_BLS && next == DDP_COMPONENT_DSI0) {
next              341 drivers/gpu/drm/mediatek/mtk_drm_ddp.c 	} else if (cur == DDP_COMPONENT_BLS && next == DDP_COMPONENT_DPI0) {
next              353 drivers/gpu/drm/mediatek/mtk_drm_ddp.c 			      enum mtk_ddp_comp_id next)
next              357 drivers/gpu/drm/mediatek/mtk_drm_ddp.c 	value = mtk_ddp_mout_en(cur, next, &addr);
next              363 drivers/gpu/drm/mediatek/mtk_drm_ddp.c 	mtk_ddp_sout_sel(config_regs, cur, next);
next              365 drivers/gpu/drm/mediatek/mtk_drm_ddp.c 	value = mtk_ddp_sel_in(cur, next, &addr);
next              374 drivers/gpu/drm/mediatek/mtk_drm_ddp.c 				   enum mtk_ddp_comp_id next)
next              378 drivers/gpu/drm/mediatek/mtk_drm_ddp.c 	value = mtk_ddp_mout_en(cur, next, &addr);
next              384 drivers/gpu/drm/mediatek/mtk_drm_ddp.c 	value = mtk_ddp_sel_in(cur, next, &addr);
next               17 drivers/gpu/drm/mediatek/mtk_drm_ddp.h 			      enum mtk_ddp_comp_id next);
next               20 drivers/gpu/drm/mediatek/mtk_drm_ddp.h 				   enum mtk_ddp_comp_id next);
next             1245 drivers/gpu/drm/mediatek/mtk_hdmi.c 	if (hdmi->bridge.next) {
next             1249 drivers/gpu/drm/mediatek/mtk_hdmi.c 		if (!drm_bridge_mode_fixup(hdmi->bridge.next, mode,
next              235 drivers/gpu/drm/mga/mga_dma.c 	for (entry = dev_priv->head->next; entry; entry = entry->next) {
next              267 drivers/gpu/drm/mga/mga_dma.c 		entry->next = dev_priv->head->next;
next              272 drivers/gpu/drm/mga/mga_dma.c 		if (dev_priv->head->next != NULL)
next              273 drivers/gpu/drm/mga/mga_dma.c 			dev_priv->head->next->prev = entry;
next              274 drivers/gpu/drm/mga/mga_dma.c 		if (entry->next == NULL)
next              281 drivers/gpu/drm/mga/mga_dma.c 		dev_priv->head->next = entry;
next              291 drivers/gpu/drm/mga/mga_dma.c 	drm_mga_freelist_t *next;
next              296 drivers/gpu/drm/mga/mga_dma.c 		next = entry->next;
next              298 drivers/gpu/drm/mga/mga_dma.c 		entry = next;
next              325 drivers/gpu/drm/mga/mga_dma.c 	drm_mga_freelist_t *next;
next              343 drivers/gpu/drm/mga/mga_dma.c 		next = dev_priv->tail;
next              344 drivers/gpu/drm/mga/mga_dma.c 		prev->next = NULL;
next              345 drivers/gpu/drm/mga/mga_dma.c 		next->prev = next->next = NULL;
next              347 drivers/gpu/drm/mga/mga_dma.c 		SET_AGE(&next->age, MGA_BUFFER_USED, 0);
next              348 drivers/gpu/drm/mga/mga_dma.c 		return next->buf;
next              372 drivers/gpu/drm/mga/mga_dma.c 		prev->next = entry;
next              374 drivers/gpu/drm/mga/mga_dma.c 		entry->next = NULL;
next              376 drivers/gpu/drm/mga/mga_dma.c 		prev = head->next;
next              377 drivers/gpu/drm/mga/mga_dma.c 		head->next = entry;
next              380 drivers/gpu/drm/mga/mga_dma.c 		entry->next = prev;
next               80 drivers/gpu/drm/mga/mga_drv.h 	struct drm_mga_freelist *next;
next               31 drivers/gpu/drm/msm/adreno/a5xx_gpu.c 	ring->cur = ring->next;
next               60 drivers/gpu/drm/msm/adreno/a6xx_gpu.c 	ring->cur = ring->next;
next              350 drivers/gpu/drm/msm/adreno/adreno_gpu.c 		ring->next = ring->start;
next              492 drivers/gpu/drm/msm/adreno/adreno_gpu.c 	ring->cur = ring->next;
next              807 drivers/gpu/drm/msm/adreno/adreno_gpu.c 	uint32_t wptr = ring->next - ring->start;
next              937 drivers/gpu/drm/msm/msm_gem.c 	struct msm_gem_object *msm_obj, *next;
next              943 drivers/gpu/drm/msm/msm_gem.c 		llist_for_each_entry_safe(msm_obj, next,
next               41 drivers/gpu/drm/msm/msm_ringbuffer.c 	ring->next  = ring->start;
next               40 drivers/gpu/drm/msm/msm_ringbuffer.h 	uint32_t *start, *end, *cur, *next;
next               64 drivers/gpu/drm/msm/msm_ringbuffer.h 	if (ring->next == ring->end)
next               65 drivers/gpu/drm/msm/msm_ringbuffer.h 		ring->next = ring->start;
next               66 drivers/gpu/drm/msm/msm_ringbuffer.h 	*(ring->next++) = data;
next              111 drivers/gpu/drm/nouveau/include/nvif/list.h     struct list_head *next, *prev;
next              130 drivers/gpu/drm/nouveau/include/nvif/list.h     list->next = list->prev = list;
next              135 drivers/gpu/drm/nouveau/include/nvif/list.h                 struct list_head *prev, struct list_head *next)
next              137 drivers/gpu/drm/nouveau/include/nvif/list.h     next->prev = entry;
next              138 drivers/gpu/drm/nouveau/include/nvif/list.h     entry->next = next;
next              140 drivers/gpu/drm/nouveau/include/nvif/list.h     prev->next = entry;
next              161 drivers/gpu/drm/nouveau/include/nvif/list.h     __list_add(entry, head, head->next);
next              186 drivers/gpu/drm/nouveau/include/nvif/list.h __list_del(struct list_head *prev, struct list_head *next)
next              188 drivers/gpu/drm/nouveau/include/nvif/list.h     next->prev = prev;
next              189 drivers/gpu/drm/nouveau/include/nvif/list.h     prev->next = next;
next              209 drivers/gpu/drm/nouveau/include/nvif/list.h     __list_del(entry->prev, entry->next);
next              215 drivers/gpu/drm/nouveau/include/nvif/list.h     __list_del(entry->prev, entry->next);
next              222 drivers/gpu/drm/nouveau/include/nvif/list.h 	__list_del(list->prev, list->next);
next              237 drivers/gpu/drm/nouveau/include/nvif/list.h     return head->next == head;
next              277 drivers/gpu/drm/nouveau/include/nvif/list.h     list_entry((ptr)->next, type, member)
next              315 drivers/gpu/drm/nouveau/include/nvif/list.h     for (pos = __container_of((head)->next, pos, member);		\
next              317 drivers/gpu/drm/nouveau/include/nvif/list.h 	 pos = __container_of(pos->member.next, pos, member))
next              327 drivers/gpu/drm/nouveau/include/nvif/list.h     for (pos = __container_of((head)->next, pos, member),		\
next              328 drivers/gpu/drm/nouveau/include/nvif/list.h 	 tmp = __container_of(pos->member.next, pos, member);		\
next              330 drivers/gpu/drm/nouveau/include/nvif/list.h 	 pos = tmp, tmp = __container_of(pos->member.next, tmp, member))
next              339 drivers/gpu/drm/nouveau/include/nvif/list.h 	for (pos = __container_of(pos->member.next, pos, member);	\
next              341 drivers/gpu/drm/nouveau/include/nvif/list.h 	     pos = __container_of(pos->member.next, pos, member))
next              351 drivers/gpu/drm/nouveau/include/nvif/list.h 	     pos = __container_of(pos->member.next, pos, member))
next                9 drivers/gpu/drm/nouveau/include/nvkm/core/mm.h 	struct nvkm_mm_node *next;
next               58 drivers/gpu/drm/nouveau/include/nvkm/core/mm.h 	return !node->next;
next               75 drivers/gpu/drm/nouveau/include/nvkm/core/mm.h 	} while ((node = node->next));
next              136 drivers/gpu/drm/nouveau/include/nvkm/subdev/fb.h 	struct nvkm_ram_data *next;
next               75 drivers/gpu/drm/nouveau/include/nvkm/subdev/mmu.h 	u64 next;
next               96 drivers/gpu/drm/nouveau/nouveau_fence.c 		fence = list_entry(fctx->pending.next, typeof(*fence), head);
next              133 drivers/gpu/drm/nouveau/nouveau_fence.c 		fence = list_entry(fctx->pending.next, typeof(*fence), head);
next              157 drivers/gpu/drm/nouveau/nouveau_fence.c 		fence = list_entry(fctx->pending.next, typeof(*fence), head);
next              338 drivers/gpu/drm/nouveau/nouveau_gem.c 		nvbo = list_entry(op->list.next, struct nouveau_bo, entry);
next              183 drivers/gpu/drm/nouveau/nouveau_svm.c 		unsigned long next;
next              190 drivers/gpu/drm/nouveau/nouveau_svm.c 		next = min(vma->vm_end, end);
next              192 drivers/gpu/drm/nouveau/nouveau_svm.c 		nouveau_dmem_migrate_vma(cli->drm, vma, addr, next);
next              193 drivers/gpu/drm/nouveau/nouveau_svm.c 		addr = next;
next               54 drivers/gpu/drm/nouveau/nvkm/core/mm.c 		struct nvkm_mm_node *next = node(this, next);
next               62 drivers/gpu/drm/nouveau/nvkm/core/mm.c 		if (next && next->type == NVKM_MM_TYPE_NONE) {
next               63 drivers/gpu/drm/nouveau/nvkm/core/mm.c 			next->offset  = this->offset;
next               64 drivers/gpu/drm/nouveau/nvkm/core/mm.c 			next->length += this->length;
next              114 drivers/gpu/drm/nouveau/nvkm/core/mm.c 	struct nvkm_mm_node *prev, *this, *next;
next              133 drivers/gpu/drm/nouveau/nvkm/core/mm.c 		next = node(this, next);
next              134 drivers/gpu/drm/nouveau/nvkm/core/mm.c 		if (next && next->type != type)
next              150 drivers/gpu/drm/nouveau/nvkm/core/mm.c 		this->next = NULL;
next              189 drivers/gpu/drm/nouveau/nvkm/core/mm.c 	struct nvkm_mm_node *prev, *this, *next;
next              207 drivers/gpu/drm/nouveau/nvkm/core/mm.c 		next = node(this, next);
next              208 drivers/gpu/drm/nouveau/nvkm/core/mm.c 		if (next && next->type != type) {
next              210 drivers/gpu/drm/nouveau/nvkm/core/mm.c 			c = next->offset - e;
next              229 drivers/gpu/drm/nouveau/nvkm/core/mm.c 		this->next = NULL;
next              243 drivers/gpu/drm/nouveau/nvkm/core/mm.c 	u32 next;
next              247 drivers/gpu/drm/nouveau/nvkm/core/mm.c 		next = prev->offset + prev->length;
next              248 drivers/gpu/drm/nouveau/nvkm/core/mm.c 		if (next != offset) {
next              249 drivers/gpu/drm/nouveau/nvkm/core/mm.c 			BUG_ON(next > offset);
next              253 drivers/gpu/drm/nouveau/nvkm/core/mm.c 			node->offset = next;
next              254 drivers/gpu/drm/nouveau/nvkm/core/mm.c 			node->length = offset - next;
next               47 drivers/gpu/drm/nouveau/nvkm/engine/disp/channv50.c 			u32 next = nvkm_rd32(device, list->data[i].addr + base + 0);
next               53 drivers/gpu/drm/nouveau/nvkm/engine/disp/channv50.c 			if (prev != next)
next               54 drivers/gpu/drm/nouveau/nvkm/engine/disp/channv50.c 				snprintf(mods, sizeof(mods), "-> %08x", next);
next               48 drivers/gpu/drm/nouveau/nvkm/engine/fifo/gk104.c 	} prev, next, *chan;
next               62 drivers/gpu/drm/nouveau/nvkm/engine/fifo/gk104.c 	status->next.tsg = !!(stat & 0x10000000);
next               63 drivers/gpu/drm/nouveau/nvkm/engine/fifo/gk104.c 	status->next.id  =   (stat & 0x0fff0000) >> 16;
next               74 drivers/gpu/drm/nouveau/nvkm/engine/fifo/gk104.c 				status->chan = &status->next;
next               79 drivers/gpu/drm/nouveau/nvkm/engine/fifo/gk104.c 			status->chan = &status->next;
next               94 drivers/gpu/drm/nouveau/nvkm/engine/fifo/gk104.c 		   status->next.tsg ? "tsg" : "ch", status->next.id,
next               95 drivers/gpu/drm/nouveau/nvkm/engine/fifo/gk104.c 		   status->chan == &status->next ? "*" : " ");
next              189 drivers/gpu/drm/nouveau/nvkm/engine/fifo/gk104.c 	mem = fifo->runlist[runl].mem[fifo->runlist[runl].next];
next              190 drivers/gpu/drm/nouveau/nvkm/engine/fifo/gk104.c 	fifo->runlist[runl].next = !fifo->runlist[runl].next;
next               33 drivers/gpu/drm/nouveau/nvkm/engine/fifo/gk104.h 		int next;
next             1027 drivers/gpu/drm/nouveau/nvkm/engine/gr/gf100.c 		u32 next = init->addr + init->count * init->pitch;
next             1029 drivers/gpu/drm/nouveau/nvkm/engine/gr/gf100.c 		while (addr < next) {
next             1047 drivers/gpu/drm/nouveau/nvkm/engine/gr/gf100.c 		u32 next = init->addr + init->count * init->pitch;
next             1055 drivers/gpu/drm/nouveau/nvkm/engine/gr/gf100.c 		while (addr < next) {
next             1084 drivers/gpu/drm/nouveau/nvkm/engine/gr/gf100.c 		u32 next = init->addr + init->count * init->pitch;
next             1092 drivers/gpu/drm/nouveau/nvkm/engine/gr/gf100.c 		while (addr < next) {
next             1118 drivers/gpu/drm/nouveau/nvkm/engine/gr/nv04.c 	struct nv04_gr_chan *next = NULL;
next             1130 drivers/gpu/drm/nouveau/nvkm/engine/gr/nv04.c 	next = gr->chan[chid];
next             1131 drivers/gpu/drm/nouveau/nvkm/engine/gr/nv04.c 	if (next)
next             1132 drivers/gpu/drm/nouveau/nvkm/engine/gr/nv04.c 		nv04_gr_load_context(next, chid);
next              936 drivers/gpu/drm/nouveau/nvkm/engine/gr/nv10.c 	struct nv10_gr_chan *next = NULL;
next              948 drivers/gpu/drm/nouveau/nvkm/engine/gr/nv10.c 	next = gr->chan[chid];
next              949 drivers/gpu/drm/nouveau/nvkm/engine/gr/nv10.c 	if (next)
next              950 drivers/gpu/drm/nouveau/nvkm/engine/gr/nv10.c 		nv10_gr_load_context(next, chid);
next              233 drivers/gpu/drm/nouveau/nvkm/engine/pm/base.c 	dom->func->next(pm, dom);
next              256 drivers/gpu/drm/nouveau/nvkm/engine/pm/base.c 		dom->func->next(pm, dom);
next              320 drivers/gpu/drm/nouveau/nvkm/engine/pm/base.c 			if (ctr->head.next)
next              172 drivers/gpu/drm/nouveau/nvkm/engine/pm/gf100.c 	.next = gf100_perfctr_next,
next               74 drivers/gpu/drm/nouveau/nvkm/engine/pm/nv40.c 	.next = nv40_perfctr_next,
next               93 drivers/gpu/drm/nouveau/nvkm/engine/pm/priv.h 	void (*next)(struct nvkm_pm *, struct nvkm_perfdom *);
next               75 drivers/gpu/drm/nouveau/nvkm/subdev/fb/gddr3.c 	switch (ram->next->bios.timing_ver) {
next               77 drivers/gpu/drm/nouveau/nvkm/subdev/fb/gddr3.c 		CWL = ram->next->bios.timing_10_CWL;
next               78 drivers/gpu/drm/nouveau/nvkm/subdev/fb/gddr3.c 		CL  = ram->next->bios.timing_10_CL;
next               79 drivers/gpu/drm/nouveau/nvkm/subdev/fb/gddr3.c 		WR  = ram->next->bios.timing_10_WR;
next               80 drivers/gpu/drm/nouveau/nvkm/subdev/fb/gddr3.c 		DLL = !ram->next->bios.ramcfg_DLLoff;
next               81 drivers/gpu/drm/nouveau/nvkm/subdev/fb/gddr3.c 		ODT = ram->next->bios.timing_10_ODT;
next               82 drivers/gpu/drm/nouveau/nvkm/subdev/fb/gddr3.c 		RON = ram->next->bios.ramcfg_RON;
next               85 drivers/gpu/drm/nouveau/nvkm/subdev/fb/gddr3.c 		CWL = (ram->next->bios.timing[1] & 0x00000f80) >> 7;
next               86 drivers/gpu/drm/nouveau/nvkm/subdev/fb/gddr3.c 		CL  = (ram->next->bios.timing[1] & 0x0000001f) >> 0;
next               87 drivers/gpu/drm/nouveau/nvkm/subdev/fb/gddr3.c 		WR  = (ram->next->bios.timing[2] & 0x007f0000) >> 16;
next               96 drivers/gpu/drm/nouveau/nvkm/subdev/fb/gddr3.c 	if (ram->next->bios.timing_ver == 0x20 ||
next               97 drivers/gpu/drm/nouveau/nvkm/subdev/fb/gddr3.c 	    ram->next->bios.ramcfg_timing == 0xff) {
next               41 drivers/gpu/drm/nouveau/nvkm/subdev/fb/gddr5.c 	xd = !ram->next->bios.ramcfg_DLLoff;
next               43 drivers/gpu/drm/nouveau/nvkm/subdev/fb/gddr5.c 	switch (ram->next->bios.ramcfg_ver) {
next               45 drivers/gpu/drm/nouveau/nvkm/subdev/fb/gddr5.c 		pd =  ram->next->bios.ramcfg_11_01_80;
next               46 drivers/gpu/drm/nouveau/nvkm/subdev/fb/gddr5.c 		lf =  ram->next->bios.ramcfg_11_01_40;
next               47 drivers/gpu/drm/nouveau/nvkm/subdev/fb/gddr5.c 		vh =  ram->next->bios.ramcfg_11_02_10;
next               48 drivers/gpu/drm/nouveau/nvkm/subdev/fb/gddr5.c 		vr =  ram->next->bios.ramcfg_11_02_04;
next               49 drivers/gpu/drm/nouveau/nvkm/subdev/fb/gddr5.c 		vo =  ram->next->bios.ramcfg_11_06;
next               50 drivers/gpu/drm/nouveau/nvkm/subdev/fb/gddr5.c 		l3 = !ram->next->bios.ramcfg_11_07_02;
next               56 drivers/gpu/drm/nouveau/nvkm/subdev/fb/gddr5.c 	switch (ram->next->bios.timing_ver) {
next               58 drivers/gpu/drm/nouveau/nvkm/subdev/fb/gddr5.c 		WL = (ram->next->bios.timing[1] & 0x00000f80) >> 7;
next               59 drivers/gpu/drm/nouveau/nvkm/subdev/fb/gddr5.c 		CL = (ram->next->bios.timing[1] & 0x0000001f);
next               60 drivers/gpu/drm/nouveau/nvkm/subdev/fb/gddr5.c 		WR = (ram->next->bios.timing[2] & 0x007f0000) >> 16;
next               61 drivers/gpu/drm/nouveau/nvkm/subdev/fb/gddr5.c 		at[0] = ram->next->bios.timing_20_2e_c0;
next               62 drivers/gpu/drm/nouveau/nvkm/subdev/fb/gddr5.c 		at[1] = ram->next->bios.timing_20_2e_30;
next               63 drivers/gpu/drm/nouveau/nvkm/subdev/fb/gddr5.c 		dt =  ram->next->bios.timing_20_2e_03;
next               64 drivers/gpu/drm/nouveau/nvkm/subdev/fb/gddr5.c 		ds =  ram->next->bios.timing_20_2f_03;
next               82 drivers/gpu/drm/nouveau/nvkm/subdev/fb/ram.c 	struct nvkm_mm_node *next = vram->mn;
next               85 drivers/gpu/drm/nouveau/nvkm/subdev/fb/ram.c 	while ((node = next)) {
next               86 drivers/gpu/drm/nouveau/nvkm/subdev/fb/ram.c 		next = node->next;
next              143 drivers/gpu/drm/nouveau/nvkm/subdev/fb/ram.c 		node = &r->next;
next              210 drivers/gpu/drm/nouveau/nvkm/subdev/fb/ramgk104.c 	struct nvkm_ram_data *next = ram->base.next;
next              211 drivers/gpu/drm/nouveau/nvkm/subdev/fb/ramgk104.c 	u8 v0 = next->bios.ramcfg_11_03_c0;
next              212 drivers/gpu/drm/nouveau/nvkm/subdev/fb/ramgk104.c 	u8 v1 = next->bios.ramcfg_11_03_30;
next              243 drivers/gpu/drm/nouveau/nvkm/subdev/fb/ramgk104.c 			u32 next = (prev & ~mask) | data;
next              244 drivers/gpu/drm/nouveau/nvkm/subdev/fb/ramgk104.c 			nvkm_memx_wr32(fuc->memx, addr, next);
next              255 drivers/gpu/drm/nouveau/nvkm/subdev/fb/ramgk104.c 	struct nvkm_ram_data *next = ram->base.next;
next              256 drivers/gpu/drm/nouveau/nvkm/subdev/fb/ramgk104.c 	int vc = !next->bios.ramcfg_11_02_08;
next              257 drivers/gpu/drm/nouveau/nvkm/subdev/fb/ramgk104.c 	int mv = !next->bios.ramcfg_11_02_04;
next              311 drivers/gpu/drm/nouveau/nvkm/subdev/fb/ramgk104.c 		switch (next->bios.ramcfg_11_03_c0) {
next              318 drivers/gpu/drm/nouveau/nvkm/subdev/fb/ramgk104.c 		switch (next->bios.ramcfg_11_03_30) {
next              326 drivers/gpu/drm/nouveau/nvkm/subdev/fb/ramgk104.c 	if (next->bios.ramcfg_11_02_80)
next              328 drivers/gpu/drm/nouveau/nvkm/subdev/fb/ramgk104.c 	if (next->bios.ramcfg_11_02_40)
next              330 drivers/gpu/drm/nouveau/nvkm/subdev/fb/ramgk104.c 	if (next->bios.ramcfg_11_07_10)
next              332 drivers/gpu/drm/nouveau/nvkm/subdev/fb/ramgk104.c 	if (next->bios.ramcfg_11_07_08)
next              367 drivers/gpu/drm/nouveau/nvkm/subdev/fb/ramgk104.c 	if (next->bios.ramcfg_11_02_40 ||
next              368 drivers/gpu/drm/nouveau/nvkm/subdev/fb/ramgk104.c 	    next->bios.ramcfg_11_07_10) {
next              391 drivers/gpu/drm/nouveau/nvkm/subdev/fb/ramgk104.c 		if (next->bios.ramcfg_11_07_40)
next              395 drivers/gpu/drm/nouveau/nvkm/subdev/fb/ramgk104.c 	ram_wr32(fuc, 0x10f65c, 0x00000011 * next->bios.rammap_11_11_0c);
next              396 drivers/gpu/drm/nouveau/nvkm/subdev/fb/ramgk104.c 	ram_wr32(fuc, 0x10f6b8, 0x01010101 * next->bios.ramcfg_11_09);
next              397 drivers/gpu/drm/nouveau/nvkm/subdev/fb/ramgk104.c 	ram_wr32(fuc, 0x10f6bc, 0x01010101 * next->bios.ramcfg_11_09);
next              399 drivers/gpu/drm/nouveau/nvkm/subdev/fb/ramgk104.c 	if (!next->bios.ramcfg_11_07_08 && !next->bios.ramcfg_11_07_04) {
next              400 drivers/gpu/drm/nouveau/nvkm/subdev/fb/ramgk104.c 		ram_wr32(fuc, 0x10f698, 0x01010101 * next->bios.ramcfg_11_04);
next              401 drivers/gpu/drm/nouveau/nvkm/subdev/fb/ramgk104.c 		ram_wr32(fuc, 0x10f69c, 0x01010101 * next->bios.ramcfg_11_04);
next              403 drivers/gpu/drm/nouveau/nvkm/subdev/fb/ramgk104.c 	if (!next->bios.ramcfg_11_07_08) {
next              409 drivers/gpu/drm/nouveau/nvkm/subdev/fb/ramgk104.c 		u32 data = 0x01000100 * next->bios.ramcfg_11_04;
next              414 drivers/gpu/drm/nouveau/nvkm/subdev/fb/ramgk104.c 	if (ram->mode == 2 && next->bios.ramcfg_11_08_10)
next              422 drivers/gpu/drm/nouveau/nvkm/subdev/fb/ramgk104.c 	if (!next->bios.ramcfg_11_02_80)
next              424 drivers/gpu/drm/nouveau/nvkm/subdev/fb/ramgk104.c 	if (!next->bios.ramcfg_11_02_40)
next              426 drivers/gpu/drm/nouveau/nvkm/subdev/fb/ramgk104.c 	if (!next->bios.ramcfg_11_07_10)
next              428 drivers/gpu/drm/nouveau/nvkm/subdev/fb/ramgk104.c 	if (!next->bios.ramcfg_11_07_08)
next              434 drivers/gpu/drm/nouveau/nvkm/subdev/fb/ramgk104.c 	if (next->bios.ramcfg_11_01_08)
next              445 drivers/gpu/drm/nouveau/nvkm/subdev/fb/ramgk104.c 	if (next->bios.ramcfg_11_08_01)
next              452 drivers/gpu/drm/nouveau/nvkm/subdev/fb/ramgk104.c 	if (next->bios.ramcfg_11_08_08)
next              454 drivers/gpu/drm/nouveau/nvkm/subdev/fb/ramgk104.c 	if (next->bios.ramcfg_11_08_04)
next              456 drivers/gpu/drm/nouveau/nvkm/subdev/fb/ramgk104.c 	if (next->bios.ramcfg_11_08_02)
next              461 drivers/gpu/drm/nouveau/nvkm/subdev/fb/ramgk104.c 	ram_mask(fuc, 0x10f248, 0xffffffff, next->bios.timing[10]);
next              462 drivers/gpu/drm/nouveau/nvkm/subdev/fb/ramgk104.c 	ram_mask(fuc, 0x10f290, 0xffffffff, next->bios.timing[0]);
next              463 drivers/gpu/drm/nouveau/nvkm/subdev/fb/ramgk104.c 	ram_mask(fuc, 0x10f294, 0xffffffff, next->bios.timing[1]);
next              464 drivers/gpu/drm/nouveau/nvkm/subdev/fb/ramgk104.c 	ram_mask(fuc, 0x10f298, 0xffffffff, next->bios.timing[2]);
next              465 drivers/gpu/drm/nouveau/nvkm/subdev/fb/ramgk104.c 	ram_mask(fuc, 0x10f29c, 0xffffffff, next->bios.timing[3]);
next              466 drivers/gpu/drm/nouveau/nvkm/subdev/fb/ramgk104.c 	ram_mask(fuc, 0x10f2a0, 0xffffffff, next->bios.timing[4]);
next              467 drivers/gpu/drm/nouveau/nvkm/subdev/fb/ramgk104.c 	ram_mask(fuc, 0x10f2a4, 0xffffffff, next->bios.timing[5]);
next              468 drivers/gpu/drm/nouveau/nvkm/subdev/fb/ramgk104.c 	ram_mask(fuc, 0x10f2a8, 0xffffffff, next->bios.timing[6]);
next              469 drivers/gpu/drm/nouveau/nvkm/subdev/fb/ramgk104.c 	ram_mask(fuc, 0x10f2ac, 0xffffffff, next->bios.timing[7]);
next              470 drivers/gpu/drm/nouveau/nvkm/subdev/fb/ramgk104.c 	ram_mask(fuc, 0x10f2cc, 0xffffffff, next->bios.timing[8]);
next              471 drivers/gpu/drm/nouveau/nvkm/subdev/fb/ramgk104.c 	ram_mask(fuc, 0x10f2e8, 0xffffffff, next->bios.timing[9]);
next              475 drivers/gpu/drm/nouveau/nvkm/subdev/fb/ramgk104.c 		if (next->bios.ramcfg_11_08_20)
next              483 drivers/gpu/drm/nouveau/nvkm/subdev/fb/ramgk104.c 		data |= next->bios.ramcfg_11_02_03 << 8;
next              487 drivers/gpu/drm/nouveau/nvkm/subdev/fb/ramgk104.c 		if (next->bios.ramcfg_11_01_10)
next              495 drivers/gpu/drm/nouveau/nvkm/subdev/fb/ramgk104.c 		data |= next->bios.timing_20_30_07 << 28;
next              499 drivers/gpu/drm/nouveau/nvkm/subdev/fb/ramgk104.c 		if (next->bios.ramcfg_11_01_01)
next              507 drivers/gpu/drm/nouveau/nvkm/subdev/fb/ramgk104.c 		data |= next->bios.timing_20_30_07 << 28;
next              511 drivers/gpu/drm/nouveau/nvkm/subdev/fb/ramgk104.c 		if (next->bios.ramcfg_11_01_02)
next              519 drivers/gpu/drm/nouveau/nvkm/subdev/fb/ramgk104.c 	if (!next->bios.ramcfg_11_01_04)
next              521 drivers/gpu/drm/nouveau/nvkm/subdev/fb/ramgk104.c 	if (!next->bios.ramcfg_11_07_80)
next              526 drivers/gpu/drm/nouveau/nvkm/subdev/fb/ramgk104.c 	if (next->bios.ramcfg_11_03_f0) {
next              527 drivers/gpu/drm/nouveau/nvkm/subdev/fb/ramgk104.c 		if (next->bios.rammap_11_08_0c) {
next              528 drivers/gpu/drm/nouveau/nvkm/subdev/fb/ramgk104.c 			if (!next->bios.ramcfg_11_07_80)
next              541 drivers/gpu/drm/nouveau/nvkm/subdev/fb/ramgk104.c 	ram_wr32(fuc, 0x10f870, 0x11111111 * next->bios.ramcfg_11_03_0f);
next              545 drivers/gpu/drm/nouveau/nvkm/subdev/fb/ramgk104.c 		data |= next->bios.ramcfg_11_02_03;
next              549 drivers/gpu/drm/nouveau/nvkm/subdev/fb/ramgk104.c 		if (next->bios.ramcfg_11_01_10)
next              560 drivers/gpu/drm/nouveau/nvkm/subdev/fb/ramgk104.c 	data = next->bios.timing_20_30_07 << 8;
next              561 drivers/gpu/drm/nouveau/nvkm/subdev/fb/ramgk104.c 	if (next->bios.ramcfg_11_01_01)
next              565 drivers/gpu/drm/nouveau/nvkm/subdev/fb/ramgk104.c 	ram_mask(fuc, 0x10f250, 0x000003f0, next->bios.timing_20_2c_003f << 4);
next              566 drivers/gpu/drm/nouveau/nvkm/subdev/fb/ramgk104.c 	data = (next->bios.timing[10] & 0x7f000000) >> 24;
next              567 drivers/gpu/drm/nouveau/nvkm/subdev/fb/ramgk104.c 	if (data < next->bios.timing_20_2c_1fc0)
next              568 drivers/gpu/drm/nouveau/nvkm/subdev/fb/ramgk104.c 		data = next->bios.timing_20_2c_1fc0;
next              570 drivers/gpu/drm/nouveau/nvkm/subdev/fb/ramgk104.c 	ram_mask(fuc, 0x10f224, 0x001f0000, next->bios.timing_20_30_f8 << 16);
next              572 drivers/gpu/drm/nouveau/nvkm/subdev/fb/ramgk104.c 	ram_mask(fuc, 0x10fec4, 0x041e0f07, next->bios.timing_20_31_0800 << 26 |
next              573 drivers/gpu/drm/nouveau/nvkm/subdev/fb/ramgk104.c 					    next->bios.timing_20_31_0780 << 17 |
next              574 drivers/gpu/drm/nouveau/nvkm/subdev/fb/ramgk104.c 					    next->bios.timing_20_31_0078 << 8 |
next              575 drivers/gpu/drm/nouveau/nvkm/subdev/fb/ramgk104.c 					    next->bios.timing_20_31_0007);
next              576 drivers/gpu/drm/nouveau/nvkm/subdev/fb/ramgk104.c 	ram_mask(fuc, 0x10fec8, 0x00000027, next->bios.timing_20_31_8000 << 5 |
next              577 drivers/gpu/drm/nouveau/nvkm/subdev/fb/ramgk104.c 					    next->bios.timing_20_31_7000);
next              585 drivers/gpu/drm/nouveau/nvkm/subdev/fb/ramgk104.c 	if (next->bios.ramcfg_11_08_10 && (ram->mode == 2) /*XXX*/) {
next              618 drivers/gpu/drm/nouveau/nvkm/subdev/fb/ramgk104.c 	if (!next->bios.ramcfg_11_07_08) {
next              619 drivers/gpu/drm/nouveau/nvkm/subdev/fb/ramgk104.c 		if (!next->bios.ramcfg_11_07_04)
next              633 drivers/gpu/drm/nouveau/nvkm/subdev/fb/ramgk104.c 	if (!next->bios.ramcfg_11_07_08) {
next              635 drivers/gpu/drm/nouveau/nvkm/subdev/fb/ramgk104.c 		if ( next->bios.ramcfg_11_07_04)
next              637 drivers/gpu/drm/nouveau/nvkm/subdev/fb/ramgk104.c 		if (!next->bios.rammap_11_08_10)
next              659 drivers/gpu/drm/nouveau/nvkm/subdev/fb/ramgk104.c 	if (next->bios.ramcfg_11_07_02)
next              667 drivers/gpu/drm/nouveau/nvkm/subdev/fb/ramgk104.c 	if (next->bios.rammap_11_08_01)
next              706 drivers/gpu/drm/nouveau/nvkm/subdev/fb/ramgk104.c 	struct nvkm_ram_data *next = ram->base.next;
next              707 drivers/gpu/drm/nouveau/nvkm/subdev/fb/ramgk104.c 	int vc = !next->bios.ramcfg_11_02_08;
next              708 drivers/gpu/drm/nouveau/nvkm/subdev/fb/ramgk104.c 	int mv = !next->bios.ramcfg_11_02_04;
next              726 drivers/gpu/drm/nouveau/nvkm/subdev/fb/ramgk104.c 	if (next->bios.ramcfg_11_03_f0)
next              731 drivers/gpu/drm/nouveau/nvkm/subdev/fb/ramgk104.c 	if (next->bios.ramcfg_DLLoff)
next              755 drivers/gpu/drm/nouveau/nvkm/subdev/fb/ramgk104.c 		switch (next->bios.ramcfg_11_03_c0) {
next              762 drivers/gpu/drm/nouveau/nvkm/subdev/fb/ramgk104.c 		switch (next->bios.ramcfg_11_03_30) {
next              770 drivers/gpu/drm/nouveau/nvkm/subdev/fb/ramgk104.c 	if (next->bios.ramcfg_11_02_80)
next              772 drivers/gpu/drm/nouveau/nvkm/subdev/fb/ramgk104.c 	if (next->bios.ramcfg_11_02_40)
next              774 drivers/gpu/drm/nouveau/nvkm/subdev/fb/ramgk104.c 	if (next->bios.ramcfg_11_07_10)
next              776 drivers/gpu/drm/nouveau/nvkm/subdev/fb/ramgk104.c 	if (next->bios.ramcfg_11_07_08)
next              786 drivers/gpu/drm/nouveau/nvkm/subdev/fb/ramgk104.c 	data |= next->bios.ramcfg_11_03_30 << 16;
next              818 drivers/gpu/drm/nouveau/nvkm/subdev/fb/ramgk104.c 	if (next->bios.ramcfg_11_02_40 ||
next              819 drivers/gpu/drm/nouveau/nvkm/subdev/fb/ramgk104.c 	    next->bios.ramcfg_11_07_10) {
next              825 drivers/gpu/drm/nouveau/nvkm/subdev/fb/ramgk104.c 		if (next->bios.ramcfg_11_07_40)
next              829 drivers/gpu/drm/nouveau/nvkm/subdev/fb/ramgk104.c 	ram_wr32(fuc, 0x10f65c, 0x00000011 * next->bios.rammap_11_11_0c);
next              830 drivers/gpu/drm/nouveau/nvkm/subdev/fb/ramgk104.c 	ram_wr32(fuc, 0x10f6b8, 0x01010101 * next->bios.ramcfg_11_09);
next              831 drivers/gpu/drm/nouveau/nvkm/subdev/fb/ramgk104.c 	ram_wr32(fuc, 0x10f6bc, 0x01010101 * next->bios.ramcfg_11_09);
next              835 drivers/gpu/drm/nouveau/nvkm/subdev/fb/ramgk104.c 	if (!next->bios.ramcfg_11_02_80)
next              837 drivers/gpu/drm/nouveau/nvkm/subdev/fb/ramgk104.c 	if (!next->bios.ramcfg_11_02_40)
next              839 drivers/gpu/drm/nouveau/nvkm/subdev/fb/ramgk104.c 	if (!next->bios.ramcfg_11_07_10)
next              841 drivers/gpu/drm/nouveau/nvkm/subdev/fb/ramgk104.c 	if (!next->bios.ramcfg_11_07_08)
next              848 drivers/gpu/drm/nouveau/nvkm/subdev/fb/ramgk104.c 	if (next->bios.ramcfg_11_08_01)
next              855 drivers/gpu/drm/nouveau/nvkm/subdev/fb/ramgk104.c 	ram_mask(fuc, 0x10f248, 0xffffffff, next->bios.timing[10]);
next              856 drivers/gpu/drm/nouveau/nvkm/subdev/fb/ramgk104.c 	ram_mask(fuc, 0x10f290, 0xffffffff, next->bios.timing[0]);
next              857 drivers/gpu/drm/nouveau/nvkm/subdev/fb/ramgk104.c 	ram_mask(fuc, 0x10f294, 0xffffffff, next->bios.timing[1]);
next              858 drivers/gpu/drm/nouveau/nvkm/subdev/fb/ramgk104.c 	ram_mask(fuc, 0x10f298, 0xffffffff, next->bios.timing[2]);
next              859 drivers/gpu/drm/nouveau/nvkm/subdev/fb/ramgk104.c 	ram_mask(fuc, 0x10f29c, 0xffffffff, next->bios.timing[3]);
next              860 drivers/gpu/drm/nouveau/nvkm/subdev/fb/ramgk104.c 	ram_mask(fuc, 0x10f2a0, 0xffffffff, next->bios.timing[4]);
next              861 drivers/gpu/drm/nouveau/nvkm/subdev/fb/ramgk104.c 	ram_mask(fuc, 0x10f2a4, 0xffffffff, next->bios.timing[5]);
next              862 drivers/gpu/drm/nouveau/nvkm/subdev/fb/ramgk104.c 	ram_mask(fuc, 0x10f2a8, 0xffffffff, next->bios.timing[6]);
next              863 drivers/gpu/drm/nouveau/nvkm/subdev/fb/ramgk104.c 	ram_mask(fuc, 0x10f2ac, 0xffffffff, next->bios.timing[7]);
next              864 drivers/gpu/drm/nouveau/nvkm/subdev/fb/ramgk104.c 	ram_mask(fuc, 0x10f2cc, 0xffffffff, next->bios.timing[8]);
next              865 drivers/gpu/drm/nouveau/nvkm/subdev/fb/ramgk104.c 	ram_mask(fuc, 0x10f2e8, 0xffffffff, next->bios.timing[9]);
next              869 drivers/gpu/drm/nouveau/nvkm/subdev/fb/ramgk104.c 	if (!next->bios.ramcfg_11_01_04)
next              871 drivers/gpu/drm/nouveau/nvkm/subdev/fb/ramgk104.c 	if (!next->bios.ramcfg_11_07_80)
next              876 drivers/gpu/drm/nouveau/nvkm/subdev/fb/ramgk104.c 	if (next->bios.ramcfg_11_03_f0) {
next              877 drivers/gpu/drm/nouveau/nvkm/subdev/fb/ramgk104.c 		if (next->bios.rammap_11_08_0c) {
next              878 drivers/gpu/drm/nouveau/nvkm/subdev/fb/ramgk104.c 			if (!next->bios.ramcfg_11_07_80)
next              892 drivers/gpu/drm/nouveau/nvkm/subdev/fb/ramgk104.c 	ram_wr32(fuc, 0x10f870, 0x11111111 * next->bios.ramcfg_11_03_0f);
next              894 drivers/gpu/drm/nouveau/nvkm/subdev/fb/ramgk104.c 	ram_mask(fuc, 0x10f250, 0x000003f0, next->bios.timing_20_2c_003f << 4);
next              896 drivers/gpu/drm/nouveau/nvkm/subdev/fb/ramgk104.c 	data = (next->bios.timing[10] & 0x7f000000) >> 24;
next              897 drivers/gpu/drm/nouveau/nvkm/subdev/fb/ramgk104.c 	if (data < next->bios.timing_20_2c_1fc0)
next              898 drivers/gpu/drm/nouveau/nvkm/subdev/fb/ramgk104.c 		data = next->bios.timing_20_2c_1fc0;
next              901 drivers/gpu/drm/nouveau/nvkm/subdev/fb/ramgk104.c 	ram_mask(fuc, 0x10f224, 0x001f0000, next->bios.timing_20_30_f8 << 16);
next              911 drivers/gpu/drm/nouveau/nvkm/subdev/fb/ramgk104.c 	if (!next->bios.ramcfg_DLLoff) {
next              921 drivers/gpu/drm/nouveau/nvkm/subdev/fb/ramgk104.c 	if (!next->bios.ramcfg_DLLoff) {
next              949 drivers/gpu/drm/nouveau/nvkm/subdev/fb/ramgk104.c 	if (next->bios.rammap_11_08_01)
next             1041 drivers/gpu/drm/nouveau/nvkm/subdev/fb/ramgk104.c gk104_ram_calc_xits(struct gk104_ram *ram, struct nvkm_ram_data *next)
next             1052 drivers/gpu/drm/nouveau/nvkm/subdev/fb/ramgk104.c 	ram->mode = (next->freq > fuc->refpll.vco1.max_freq) ? 2 : 1;
next             1063 drivers/gpu/drm/nouveau/nvkm/subdev/fb/ramgk104.c 	refclk = next->freq;
next             1065 drivers/gpu/drm/nouveau/nvkm/subdev/fb/ramgk104.c 		ret = gk104_pll_calc_hiclk(next->freq, subdev->device->crystal,
next             1074 drivers/gpu/drm/nouveau/nvkm/subdev/fb/ramgk104.c 				" (refclock: %i kHz)\n", next->freq, ret);
next             1090 drivers/gpu/drm/nouveau/nvkm/subdev/fb/ramgk104.c 	ram->base.freq = next->freq;
next             1096 drivers/gpu/drm/nouveau/nvkm/subdev/fb/ramgk104.c 			ret = gk104_ram_calc_sddr3(ram, next->freq);
next             1101 drivers/gpu/drm/nouveau/nvkm/subdev/fb/ramgk104.c 			ret = gk104_ram_calc_gddr5(ram, next->freq);
next             1120 drivers/gpu/drm/nouveau/nvkm/subdev/fb/ramgk104.c 	if (ram->base.next == NULL) {
next             1143 drivers/gpu/drm/nouveau/nvkm/subdev/fb/ramgk104.c 		ram->base.next = &ram->base.target;
next             1145 drivers/gpu/drm/nouveau/nvkm/subdev/fb/ramgk104.c 			ram->base.next = &ram->base.xition;
next             1147 drivers/gpu/drm/nouveau/nvkm/subdev/fb/ramgk104.c 		BUG_ON(ram->base.next != &ram->base.xition);
next             1148 drivers/gpu/drm/nouveau/nvkm/subdev/fb/ramgk104.c 		ram->base.next = &ram->base.target;
next             1151 drivers/gpu/drm/nouveau/nvkm/subdev/fb/ramgk104.c 	return gk104_ram_calc_xits(ram, ram->base.next);
next             1236 drivers/gpu/drm/nouveau/nvkm/subdev/fb/ramgk104.c 	struct nvkm_ram_data *next = ram->base.next;
next             1240 drivers/gpu/drm/nouveau/nvkm/subdev/fb/ramgk104.c 		return (ram->base.next == &ram->base.xition);
next             1245 drivers/gpu/drm/nouveau/nvkm/subdev/fb/ramgk104.c 	gk104_ram_prog_0(ram, next->freq);
next             1247 drivers/gpu/drm/nouveau/nvkm/subdev/fb/ramgk104.c 	return (ram->base.next == &ram->base.xition);
next             1254 drivers/gpu/drm/nouveau/nvkm/subdev/fb/ramgk104.c 	ram->base.next = NULL;
next              502 drivers/gpu/drm/nouveau/nvkm/subdev/fb/ramgt215.c 	struct nvkm_ram_data *next;
next              511 drivers/gpu/drm/nouveau/nvkm/subdev/fb/ramgt215.c 	next = &ram->base.target;
next              512 drivers/gpu/drm/nouveau/nvkm/subdev/fb/ramgt215.c 	next->freq = freq;
next              513 drivers/gpu/drm/nouveau/nvkm/subdev/fb/ramgt215.c 	ram->base.next = next;
next              520 drivers/gpu/drm/nouveau/nvkm/subdev/fb/ramgt215.c 			       &next->bios);
next              534 drivers/gpu/drm/nouveau/nvkm/subdev/fb/ramgt215.c 			       &ver, &hdr, &next->bios);
next              541 drivers/gpu/drm/nouveau/nvkm/subdev/fb/ramgt215.c 	if (next->bios.ramcfg_timing != 0xff) {
next              542 drivers/gpu/drm/nouveau/nvkm/subdev/fb/ramgt215.c 		data = nvbios_timingEp(bios, next->bios.ramcfg_timing,
next              544 drivers/gpu/drm/nouveau/nvkm/subdev/fb/ramgt215.c 				       &next->bios);
next              597 drivers/gpu/drm/nouveau/nvkm/subdev/fb/ramgt215.c 	if (!next->bios.ramcfg_DLLoff)
next              605 drivers/gpu/drm/nouveau/nvkm/subdev/fb/ramgt215.c 	if (next->bios.ramcfg_10_02_10) {
next              629 drivers/gpu/drm/nouveau/nvkm/subdev/fb/ramgt215.c 	if (!next->bios.ramcfg_10_02_10) {
next              637 drivers/gpu/drm/nouveau/nvkm/subdev/fb/ramgt215.c 	switch (next->bios.ramcfg_DLLoff * ram->base.type) {
next              646 drivers/gpu/drm/nouveau/nvkm/subdev/fb/ramgt215.c 	if (next->bios.timing_10_ODT)
next              663 drivers/gpu/drm/nouveau/nvkm/subdev/fb/ramgt215.c 			next->bios.ramcfg_FBVDDQ) {
next              671 drivers/gpu/drm/nouveau/nvkm/subdev/fb/ramgt215.c 		gt215_ram_gpio(fuc, 0x18, !next->bios.ramcfg_FBVDDQ);
next              707 drivers/gpu/drm/nouveau/nvkm/subdev/fb/ramgt215.c 	if (next->bios.rammap_10_04_08) {
next              708 drivers/gpu/drm/nouveau/nvkm/subdev/fb/ramgt215.c 		ram_wr32(fuc, 0x1005a0, next->bios.ramcfg_10_06 << 16 |
next              709 drivers/gpu/drm/nouveau/nvkm/subdev/fb/ramgt215.c 					next->bios.ramcfg_10_05 << 8 |
next              710 drivers/gpu/drm/nouveau/nvkm/subdev/fb/ramgt215.c 					next->bios.ramcfg_10_05);
next              711 drivers/gpu/drm/nouveau/nvkm/subdev/fb/ramgt215.c 		ram_wr32(fuc, 0x1005a4, next->bios.ramcfg_10_08 << 8 |
next              712 drivers/gpu/drm/nouveau/nvkm/subdev/fb/ramgt215.c 					next->bios.ramcfg_10_07);
next              713 drivers/gpu/drm/nouveau/nvkm/subdev/fb/ramgt215.c 		ram_wr32(fuc, 0x10f804, next->bios.ramcfg_10_09_f0 << 20 |
next              714 drivers/gpu/drm/nouveau/nvkm/subdev/fb/ramgt215.c 					next->bios.ramcfg_10_03_0f << 16 |
next              715 drivers/gpu/drm/nouveau/nvkm/subdev/fb/ramgt215.c 					next->bios.ramcfg_10_09_0f |
next              766 drivers/gpu/drm/nouveau/nvkm/subdev/fb/ramgt215.c 	ram_mask(fuc, 0x100200, 0x00001000, !next->bios.ramcfg_10_02_08 << 12);
next              777 drivers/gpu/drm/nouveau/nvkm/subdev/fb/ramgt215.c 		if (!next->bios.ramcfg_10_02_04)
next              780 drivers/gpu/drm/nouveau/nvkm/subdev/fb/ramgt215.c 		if (next->bios.ramcfg_10_02_04) {
next              785 drivers/gpu/drm/nouveau/nvkm/subdev/fb/ramgt215.c 				if (next->bios.ramcfg_10_02_10)
next              800 drivers/gpu/drm/nouveau/nvkm/subdev/fb/ramgt215.c 				if (next->bios.ramcfg_10_02_10)
next              814 drivers/gpu/drm/nouveau/nvkm/subdev/fb/ramgt215.c 	unk714 |= (next->bios.ramcfg_10_04_01) << 8;
next              816 drivers/gpu/drm/nouveau/nvkm/subdev/fb/ramgt215.c 	if (next->bios.ramcfg_10_02_20)
next              818 drivers/gpu/drm/nouveau/nvkm/subdev/fb/ramgt215.c 	if (next->bios.ramcfg_10_02_02)
next              820 drivers/gpu/drm/nouveau/nvkm/subdev/fb/ramgt215.c 	if (next->bios.ramcfg_10_02_01)
next              822 drivers/gpu/drm/nouveau/nvkm/subdev/fb/ramgt215.c 	if (next->bios.timing_10_24 != 0xff) {
next              824 drivers/gpu/drm/nouveau/nvkm/subdev/fb/ramgt215.c 		unk718 |= next->bios.timing_10_24 << 28;
next              826 drivers/gpu/drm/nouveau/nvkm/subdev/fb/ramgt215.c 	if (next->bios.ramcfg_10_02_10)
next              834 drivers/gpu/drm/nouveau/nvkm/subdev/fb/ramgt215.c 	if (!next->bios.timing_10_ODT)
next              838 drivers/gpu/drm/nouveau/nvkm/subdev/fb/ramgt215.c 	if (!next->bios.ramcfg_DLLoff)
next              861 drivers/gpu/drm/nouveau/nvkm/subdev/fb/ramgt215.c 	if (next->bios.rammap_10_04_02)
next              863 drivers/gpu/drm/nouveau/nvkm/subdev/fb/ramgt215.c 	if (next->bios.ramcfg_10_02_10) {
next               97 drivers/gpu/drm/nouveau/nvkm/subdev/fb/ramnv50.c 		unkt3b = 0x19 + ram->base.next->bios.rammap_00_16_40;
next               99 drivers/gpu/drm/nouveau/nvkm/subdev/fb/ramnv50.c 				ram->base.next->bios.rammap_00_16_40) << 16 |
next              227 drivers/gpu/drm/nouveau/nvkm/subdev/fb/ramnv50.c 	struct nvkm_ram_data *next;
next              235 drivers/gpu/drm/nouveau/nvkm/subdev/fb/ramnv50.c 	next = &ram->base.target;
next              236 drivers/gpu/drm/nouveau/nvkm/subdev/fb/ramnv50.c 	next->freq = freq;
next              237 drivers/gpu/drm/nouveau/nvkm/subdev/fb/ramnv50.c 	ram->base.next = next;
next              251 drivers/gpu/drm/nouveau/nvkm/subdev/fb/ramnv50.c 	nvbios_rammapEp_from_perf(bios, data, hdr, &next->bios);
next              261 drivers/gpu/drm/nouveau/nvkm/subdev/fb/ramnv50.c 			&next->bios);
next              268 drivers/gpu/drm/nouveau/nvkm/subdev/fb/ramnv50.c 	if (next->bios.ramcfg_timing != 0xff) {
next              269 drivers/gpu/drm/nouveau/nvkm/subdev/fb/ramnv50.c 		data = nvbios_timingEp(bios, next->bios.ramcfg_timing,
next              270 drivers/gpu/drm/nouveau/nvkm/subdev/fb/ramnv50.c 					&ver, &hdr, &cnt, &len, &next->bios);
next              305 drivers/gpu/drm/nouveau/nvkm/subdev/fb/ramnv50.c 	if (subdev->device->chipset <= 0x96 && !next->bios.ramcfg_00_03_02)
next              319 drivers/gpu/drm/nouveau/nvkm/subdev/fb/ramnv50.c 	if (next->bios.timing_10_ODT)
next              355 drivers/gpu/drm/nouveau/nvkm/subdev/fb/ramnv50.c 			next->bios.rammap_00_16_40 << 14);
next              363 drivers/gpu/drm/nouveau/nvkm/subdev/fb/ramnv50.c 	nv50_ram_gpio(hwsq, 0x18, !next->bios.ramcfg_FBVDDQ);
next              400 drivers/gpu/drm/nouveau/nvkm/subdev/fb/ramnv50.c 	if (!next->bios.ramcfg_00_03_02)
next              402 drivers/gpu/drm/nouveau/nvkm/subdev/fb/ramnv50.c 	ram_mask(hwsq, 0x100200, 0x00001000, !next->bios.ramcfg_00_04_02 << 12);
next              413 drivers/gpu/drm/nouveau/nvkm/subdev/fb/ramnv50.c 		if (!next->bios.ramcfg_00_03_08)
next              415 drivers/gpu/drm/nouveau/nvkm/subdev/fb/ramnv50.c 		if (!next->bios.ramcfg_FBVDDQ)
next              417 drivers/gpu/drm/nouveau/nvkm/subdev/fb/ramnv50.c 		if ( next->bios.ramcfg_00_04_04)
next              422 drivers/gpu/drm/nouveau/nvkm/subdev/fb/ramnv50.c 		if (!next->bios.ramcfg_00_03_08)
next              426 drivers/gpu/drm/nouveau/nvkm/subdev/fb/ramnv50.c 	if ( next->bios.ramcfg_00_03_01)
next              428 drivers/gpu/drm/nouveau/nvkm/subdev/fb/ramnv50.c 	if ( next->bios.ramcfg_00_03_02)
next              430 drivers/gpu/drm/nouveau/nvkm/subdev/fb/ramnv50.c 	if (!next->bios.ramcfg_00_03_08)
next              432 drivers/gpu/drm/nouveau/nvkm/subdev/fb/ramnv50.c 	if ( next->bios.ramcfg_00_04_04)
next              434 drivers/gpu/drm/nouveau/nvkm/subdev/fb/ramnv50.c 	if ( next->bios.ramcfg_00_04_20)
next              444 drivers/gpu/drm/nouveau/nvkm/subdev/fb/ramnv50.c 	if (next->bios.rammap_00_16_20) {
next              445 drivers/gpu/drm/nouveau/nvkm/subdev/fb/ramnv50.c 		ram_wr32(hwsq, 0x1005a0, next->bios.ramcfg_00_07 << 16 |
next              446 drivers/gpu/drm/nouveau/nvkm/subdev/fb/ramnv50.c 					 next->bios.ramcfg_00_06 << 8 |
next              447 drivers/gpu/drm/nouveau/nvkm/subdev/fb/ramnv50.c 					 next->bios.ramcfg_00_05);
next              448 drivers/gpu/drm/nouveau/nvkm/subdev/fb/ramnv50.c 		ram_wr32(hwsq, 0x1005a4, next->bios.ramcfg_00_09 << 8 |
next              449 drivers/gpu/drm/nouveau/nvkm/subdev/fb/ramnv50.c 					 next->bios.ramcfg_00_08);
next              456 drivers/gpu/drm/nouveau/nvkm/subdev/fb/ramnv50.c 	if (!next->bios.timing_10_ODT)
next              460 drivers/gpu/drm/nouveau/nvkm/subdev/fb/ramnv50.c 	if (!next->bios.ramcfg_DLLoff)
next              468 drivers/gpu/drm/nouveau/nvkm/subdev/fb/ramnv50.c 	if (next->bios.rammap_00_17_02)
next              470 drivers/gpu/drm/nouveau/nvkm/subdev/fb/ramnv50.c 	if (!next->bios.rammap_00_16_40)
next              472 drivers/gpu/drm/nouveau/nvkm/subdev/fb/ramnv50.c 	if (next->bios.ramcfg_00_03_02)
next              474 drivers/gpu/drm/nouveau/nvkm/subdev/fb/ramnv50.c 	if (subdev->device->chipset <= 0x96 && next->bios.ramcfg_00_03_02)
next               65 drivers/gpu/drm/nouveau/nvkm/subdev/fb/sddr2.c 	switch (ram->next->bios.timing_ver) {
next               67 drivers/gpu/drm/nouveau/nvkm/subdev/fb/sddr2.c 		CL  = ram->next->bios.timing_10_CL;
next               68 drivers/gpu/drm/nouveau/nvkm/subdev/fb/sddr2.c 		WR  = ram->next->bios.timing_10_WR;
next               69 drivers/gpu/drm/nouveau/nvkm/subdev/fb/sddr2.c 		DLL = !ram->next->bios.ramcfg_DLLoff;
next               70 drivers/gpu/drm/nouveau/nvkm/subdev/fb/sddr2.c 		ODT = ram->next->bios.timing_10_ODT & 3;
next               73 drivers/gpu/drm/nouveau/nvkm/subdev/fb/sddr2.c 		CL  = (ram->next->bios.timing[1] & 0x0000001f);
next               74 drivers/gpu/drm/nouveau/nvkm/subdev/fb/sddr2.c 		WR  = (ram->next->bios.timing[2] & 0x007f0000) >> 16;
next               80 drivers/gpu/drm/nouveau/nvkm/subdev/fb/sddr2.c 	if (ram->next->bios.timing_ver == 0x20 ||
next               81 drivers/gpu/drm/nouveau/nvkm/subdev/fb/sddr2.c 	    ram->next->bios.ramcfg_timing == 0xff) {
next               74 drivers/gpu/drm/nouveau/nvkm/subdev/fb/sddr3.c 	DLL = !ram->next->bios.ramcfg_DLLoff;
next               76 drivers/gpu/drm/nouveau/nvkm/subdev/fb/sddr3.c 	switch (ram->next->bios.timing_ver) {
next               78 drivers/gpu/drm/nouveau/nvkm/subdev/fb/sddr3.c 		if (ram->next->bios.timing_hdr < 0x17) {
next               82 drivers/gpu/drm/nouveau/nvkm/subdev/fb/sddr3.c 		CWL = ram->next->bios.timing_10_CWL;
next               83 drivers/gpu/drm/nouveau/nvkm/subdev/fb/sddr3.c 		CL  = ram->next->bios.timing_10_CL;
next               84 drivers/gpu/drm/nouveau/nvkm/subdev/fb/sddr3.c 		WR  = ram->next->bios.timing_10_WR;
next               85 drivers/gpu/drm/nouveau/nvkm/subdev/fb/sddr3.c 		ODT = ram->next->bios.timing_10_ODT;
next               88 drivers/gpu/drm/nouveau/nvkm/subdev/fb/sddr3.c 		CWL = (ram->next->bios.timing[1] & 0x00000f80) >> 7;
next               89 drivers/gpu/drm/nouveau/nvkm/subdev/fb/sddr3.c 		CL  = (ram->next->bios.timing[1] & 0x0000001f) >> 0;
next               90 drivers/gpu/drm/nouveau/nvkm/subdev/fb/sddr3.c 		WR  = (ram->next->bios.timing[2] & 0x007f0000) >> 16;
next              204 drivers/gpu/drm/nouveau/nvkm/subdev/instmem/nv50.c 		if (likely(iobj->lru.next) && iobj->map) {
next              245 drivers/gpu/drm/nouveau/nvkm/subdev/instmem/nv50.c 		if (likely(iobj->lru.next))
next              269 drivers/gpu/drm/nouveau/nvkm/subdev/instmem/nv50.c 	if (likely(iobj->lru.next)) {
next              271 drivers/gpu/drm/nouveau/nvkm/subdev/instmem/nv50.c 		iobj->lru.next = NULL;
next              297 drivers/gpu/drm/nouveau/nvkm/subdev/instmem/nv50.c 		iobj->lru.next = NULL; /* Exclude from eviction. */
next              319 drivers/gpu/drm/nouveau/nvkm/subdev/instmem/nv50.c 	if (likely(iobj->lru.next))
next              450 drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmm.c 				bool next = pgt->pte[ptei] & NVKM_VMM_PTE_SPTES;
next              451 drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmm.c 				if (spte != next)
next              650 drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmm.c 			u64 next = 1ULL << page[i - 1].shift;
next              651 drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmm.c 			u64 part = ALIGN(addr, next) - addr;
next              652 drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmm.c 			if (size - part >= next)
next              884 drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmm.c 		    struct nvkm_vma *vma, struct nvkm_vma *next, u64 size)
next              886 drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmm.c 	if (next) {
next              888 drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmm.c 			vma->size += next->size;
next              889 drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmm.c 			nvkm_vmm_node_delete(vmm, next);
next              899 drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmm.c 		nvkm_vmm_node_remove(vmm, next);
next              901 drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmm.c 		next->addr -= size;
next              902 drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmm.c 		next->size += size;
next              903 drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmm.c 		nvkm_vmm_node_insert(vmm, next);
next              904 drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmm.c 		return next;
next             1155 drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmm.c 	struct nvkm_vma *next = NULL;
next             1162 drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmm.c 	if (vma->addr + vma->size == addr + size && (next = node(vma, next))) {
next             1163 drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmm.c 		if (!next->part ||
next             1164 drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmm.c 		    next->memory || next->mapped != map)
next             1165 drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmm.c 			next = NULL;
next             1168 drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmm.c 	if (prev || next)
next             1169 drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmm.c 		return nvkm_vmm_node_merge(vmm, prev, vma, next, size);
next             1177 drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmm.c 	struct nvkm_vma *next;
next             1193 drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmm.c 		next = nvkm_vmm_pfn_split_merge(vmm, vma, start, size, 0, false);
next             1194 drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmm.c 		if (!WARN_ON(!next)) {
next             1195 drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmm.c 			vma = next;
next             1199 drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmm.c 	} while ((vma = node(vma, next)) && (start = vma->addr) < limit);
next             1260 drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmm.c 			goto next;
next             1280 drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmm.c 				goto next;
next             1311 drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmm.c next:
next             1314 drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmm.c 			vma = node(vma, next);
next             1337 drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmm.c 	struct nvkm_vma *next;
next             1345 drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmm.c 	if ((next = node(vma, next)) && (!next->part || next->mapped))
next             1346 drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmm.c 		next = NULL;
next             1347 drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmm.c 	nvkm_vmm_node_merge(vmm, prev, vma, next, vma->size);
next             1468 drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmm.c 		for (; map->off; map->mem = map->mem->next) {
next             1524 drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmm.c 	struct nvkm_vma *prev, *next;
next             1532 drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmm.c 	if ((next = node(vma, next)) && !next->used) {
next             1533 drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmm.c 		vma->size += next->size;
next             1534 drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmm.c 		nvkm_vmm_free_delete(vmm, next);
next             1544 drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmm.c 	struct nvkm_vma *next = vma;
next             1550 drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmm.c 			const bool mem = next->memory != NULL;
next             1551 drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmm.c 			const bool map = next->mapped;
next             1552 drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmm.c 			const u8  refd = next->refd;
next             1553 drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmm.c 			const u64 addr = next->addr;
next             1554 drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmm.c 			u64 size = next->size;
next             1557 drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmm.c 			while ((next = node(next, next)) && next->part &&
next             1558 drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmm.c 			       (next->mapped == map) &&
next             1559 drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmm.c 			       (next->memory != NULL) == mem &&
next             1560 drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmm.c 			       (next->refd == refd))
next             1561 drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmm.c 				size += next->size;
next             1576 drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmm.c 		} while (next && next->part);
next             1583 drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmm.c 	next = vma;
next             1585 drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmm.c 		if (next->mapped)
next             1586 drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmm.c 			nvkm_vmm_unmap_region(vmm, next);
next             1587 drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmm.c 	} while ((next = node(vma, next)) && next->part);
next             1707 drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmm.c 		struct nvkm_vma *next = node(this, next);
next             1716 drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmm.c 		if (vmm->func->page_block && next && next->page != p)
next              315 drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmm.h 		     (MAP->mem = MAP->mem->next))
next               38 drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmmgf100.c 	if (map->ctag && !(map->next & (1ULL << 44))) {
next               45 drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmmgf100.c 			base += map->next;
next               52 drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmmgf100.c 			data += map->next;
next              254 drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmmgf100.c 	map->next = (1 << page->shift) >> 8;
next              304 drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmmgf100.c 				map->next |= 1ULL << 44;
next              118 drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmmgp100.c 		data += map->next;
next              201 drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmmgp100.c 		data += map->next;
next              327 drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmmgp100.c 	map->next = (1ULL << page->shift) >> 4;
next              375 drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmmgp100.c 			map->next |= map->ctag;
next               35 drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmmnv50.c 	u64 next = addr + map->type, data;
next               48 drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmmnv50.c 		data  = next | (log2blk << 7);
next               49 drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmmnv50.c 		next += pten * map->next;
next              243 drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmmnv50.c 	map->next = 1 << page->shift;
next              312 drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmmnv50.c 			map->next |= map->ctag;
next               38 drivers/gpu/drm/nouveau/nvkm/subdev/therm/gf100.c 	u32 next, addr;
next               41 drivers/gpu/drm/nouveau/nvkm/subdev/therm/gf100.c 		next = init->addr + init->count * 8;
next               46 drivers/gpu/drm/nouveau/nvkm/subdev/therm/gf100.c 		while (addr < next) {
next               32 drivers/gpu/drm/omapdrm/displays/encoder-opa362.c 	return omapdss_device_connect(dst->dss, dst, dst->next);
next               38 drivers/gpu/drm/omapdrm/displays/encoder-opa362.c 	omapdss_device_disconnect(dst, dst->next);
next               91 drivers/gpu/drm/omapdrm/displays/encoder-opa362.c 	dssdev->next = omapdss_of_find_connected_device(pdev->dev.of_node, 1);
next               92 drivers/gpu/drm/omapdrm/displays/encoder-opa362.c 	if (IS_ERR(dssdev->next)) {
next               93 drivers/gpu/drm/omapdrm/displays/encoder-opa362.c 		if (PTR_ERR(dssdev->next) != -EPROBE_DEFER)
next               95 drivers/gpu/drm/omapdrm/displays/encoder-opa362.c 		return PTR_ERR(dssdev->next);
next              108 drivers/gpu/drm/omapdrm/displays/encoder-opa362.c 	if (dssdev->next)
next              109 drivers/gpu/drm/omapdrm/displays/encoder-opa362.c 		omapdss_device_put(dssdev->next);
next               38 drivers/gpu/drm/omapdrm/displays/encoder-tpd12s015.c 	r = omapdss_device_connect(dst->dss, dst, dst->next);
next               59 drivers/gpu/drm/omapdrm/displays/encoder-tpd12s015.c 	omapdss_device_disconnect(dst, dst->next);
next              172 drivers/gpu/drm/omapdrm/displays/encoder-tpd12s015.c 	dssdev->next = omapdss_of_find_connected_device(pdev->dev.of_node, 1);
next              173 drivers/gpu/drm/omapdrm/displays/encoder-tpd12s015.c 	if (IS_ERR(dssdev->next)) {
next              174 drivers/gpu/drm/omapdrm/displays/encoder-tpd12s015.c 		if (PTR_ERR(dssdev->next) != -EPROBE_DEFER)
next              176 drivers/gpu/drm/omapdrm/displays/encoder-tpd12s015.c 		return PTR_ERR(dssdev->next);
next              189 drivers/gpu/drm/omapdrm/displays/encoder-tpd12s015.c 	if (dssdev->next)
next              190 drivers/gpu/drm/omapdrm/displays/encoder-tpd12s015.c 		omapdss_device_put(dssdev->next);
next              153 drivers/gpu/drm/omapdrm/dss/base.c 		    (dssdev->next || dssdev->bridge || dssdev->panel))
next              241 drivers/gpu/drm/omapdrm/dss/base.c 	omapdss_device_pre_enable(dssdev->next);
next              256 drivers/gpu/drm/omapdrm/dss/base.c 	omapdss_device_enable(dssdev->next);
next              267 drivers/gpu/drm/omapdrm/dss/base.c 	omapdss_device_disable(dssdev->next);
next              282 drivers/gpu/drm/omapdrm/dss/base.c 	omapdss_device_post_disable(dssdev->next);
next               45 drivers/gpu/drm/omapdrm/dss/display.c 	while (output->next)
next               46 drivers/gpu/drm/omapdrm/dss/display.c 		output = output->next;
next              582 drivers/gpu/drm/omapdrm/dss/dpi.c 	return omapdss_device_connect(dst->dss, dst, dst->next);
next              588 drivers/gpu/drm/omapdrm/dss/dpi.c 	omapdss_device_disconnect(dst, dst->next);
next             4874 drivers/gpu/drm/omapdrm/dss/dsi.c 	return omapdss_device_connect(dst->dss, dst, dst->next);
next             4880 drivers/gpu/drm/omapdrm/dss/dsi.c 	omapdss_device_disconnect(dst, dst->next);
next              399 drivers/gpu/drm/omapdrm/dss/hdmi4.c 	return omapdss_device_connect(dst->dss, dst, dst->next);
next              405 drivers/gpu/drm/omapdrm/dss/hdmi4.c 	omapdss_device_disconnect(dst, dst->next);
next              404 drivers/gpu/drm/omapdrm/dss/hdmi5.c 	return omapdss_device_connect(dst->dss, dst, dst->next);
next              410 drivers/gpu/drm/omapdrm/dss/hdmi5.c 	omapdss_device_disconnect(dst, dst->next);
next              401 drivers/gpu/drm/omapdrm/dss/omapdss.h 	struct omap_dss_device *next;
next               31 drivers/gpu/drm/omapdrm/dss/output.c 	out->next = omapdss_find_device_by_node(remote_node);
next               39 drivers/gpu/drm/omapdrm/dss/output.c 	if (out->next && out->type != out->next->type) {
next               41 drivers/gpu/drm/omapdrm/dss/output.c 		omapdss_device_put(out->next);
next               42 drivers/gpu/drm/omapdrm/dss/output.c 		out->next = NULL;
next               46 drivers/gpu/drm/omapdrm/dss/output.c 	return out->next || out->bridge || out->panel ? 0 : -EPROBE_DEFER;
next               52 drivers/gpu/drm/omapdrm/dss/output.c 	if (out->next)
next               53 drivers/gpu/drm/omapdrm/dss/output.c 		omapdss_device_put(out->next);
next              237 drivers/gpu/drm/omapdrm/dss/sdi.c 	return omapdss_device_connect(dst->dss, dst, dst->next);
next              243 drivers/gpu/drm/omapdrm/dss/sdi.c 	omapdss_device_disconnect(dst, dst->next);
next              679 drivers/gpu/drm/omapdrm/dss/venc.c 	return omapdss_device_connect(dst->dss, dst, dst->next);
next              685 drivers/gpu/drm/omapdrm/dss/venc.c 	omapdss_device_disconnect(dst, dst->next);
next               41 drivers/gpu/drm/omapdrm/omap_connector.c 	for (dssdev = omap_connector->output; dssdev; dssdev = dssdev->next) {
next              102 drivers/gpu/drm/omapdrm/omap_connector.c 	for (d = omap_connector->output; d; d = d->next) {
next              250 drivers/gpu/drm/omapdrm/omap_connector.c 	for (; dssdev; dssdev = dssdev->next) {
next              366 drivers/gpu/drm/omapdrm/omap_crtc.c 	struct omap_dss_device *dssdev = omap_crtc->pipe->output->next;
next              504 drivers/gpu/drm/omapdrm/omap_crtc.c 	if (omap_crtc->pipe->output->next == NULL ||
next              505 drivers/gpu/drm/omapdrm/omap_crtc.c 	    omap_crtc->pipe->output->next->type != OMAP_DISPLAY_TYPE_DSI) {
next              557 drivers/gpu/drm/omapdrm/omap_crtc.c 	struct omap_dss_device *display = omap_crtc->pipe->output->next;
next              331 drivers/gpu/drm/omapdrm/omap_dmm_tiler.c 		engine = list_entry(dmm->idle_head.next, struct refill_engine,
next              210 drivers/gpu/drm/omapdrm/omap_drv.c 	if (output->next) {
next              219 drivers/gpu/drm/omapdrm/omap_drv.c 		while (bridge->next)
next              220 drivers/gpu/drm/omapdrm/omap_drv.c 			bridge = bridge->next;
next              125 drivers/gpu/drm/omapdrm/omap_encoder.c 	for (dssdev = output; dssdev; dssdev = dssdev->next)
next              128 drivers/gpu/drm/omapdrm/omap_encoder.c 	for (bridge = output->bridge; bridge; bridge = bridge->next) {
next              142 drivers/gpu/drm/omapdrm/omap_encoder.c 	for (dssdev = output; dssdev; dssdev = dssdev->next) {
next              170 drivers/gpu/drm/omapdrm/omap_encoder.c 	omapdss_device_disable(dssdev->next);
next              186 drivers/gpu/drm/omapdrm/omap_encoder.c 	omapdss_device_post_disable(dssdev->next);
next              198 drivers/gpu/drm/omapdrm/omap_encoder.c 	omapdss_device_pre_enable(dssdev->next);
next              214 drivers/gpu/drm/omapdrm/omap_encoder.c 	omapdss_device_enable(dssdev->next);
next              235 drivers/gpu/drm/qxl/qxl_cmd.c 			next_id = info->next;
next              375 drivers/gpu/drm/qxl/qxl_dev.h 	uint64_t next;	  /* out */
next              166 drivers/gpu/drm/qxl/qxl_release.c 		entry = container_of(release->bos.next,
next              799 drivers/gpu/drm/r128/r128_cce.c 		entry->next = dev_priv->head->next;
next              800 drivers/gpu/drm/r128/r128_cce.c 		if (!entry->next)
next              807 drivers/gpu/drm/r128/r128_cce.c 		dev_priv->head->next = entry;
next              809 drivers/gpu/drm/r128/r128_cce.c 		if (dev_priv->head->next)
next              810 drivers/gpu/drm/r128/r128_cce.c 			dev_priv->head->next->prev = entry;
next               70 drivers/gpu/drm/r128/r128_drv.h 	struct drm_r128_freelist *next;
next               43 drivers/gpu/drm/radeon/mkregtable.c 	struct list_head *next, *prev;
next               49 drivers/gpu/drm/radeon/mkregtable.c 	list->next = list;
next               61 drivers/gpu/drm/radeon/mkregtable.c 			      struct list_head *prev, struct list_head *next)
next               63 drivers/gpu/drm/radeon/mkregtable.c 	next->prev = new;
next               64 drivers/gpu/drm/radeon/mkregtable.c 	new->next = next;
next               66 drivers/gpu/drm/radeon/mkregtable.c 	prev->next = new;
next               70 drivers/gpu/drm/radeon/mkregtable.c 		       struct list_head *prev, struct list_head *next);
next              102 drivers/gpu/drm/radeon/mkregtable.c 	for (pos = list_entry((head)->next, typeof(*pos), member);	\
next              104 drivers/gpu/drm/radeon/mkregtable.c 	     pos = list_entry(pos->member.next, typeof(*pos), member))
next              242 drivers/gpu/drm/radeon/radeon_mn.c 	head = bo->mn_list.next;
next              158 drivers/gpu/drm/radeon/radeon_sa.c 	if (sa_manager->hole->next == &sa_manager->olist)
next              161 drivers/gpu/drm/radeon/radeon_sa.c 	sa_bo = list_entry(sa_manager->hole->next, struct radeon_sa_bo, olist);
next              184 drivers/gpu/drm/radeon/radeon_sa.c 	if (hole->next != &sa_manager->olist) {
next              185 drivers/gpu/drm/radeon/radeon_sa.c 		return list_entry(hole->next, struct radeon_sa_bo, olist)->soffset;
next              255 drivers/gpu/drm/radeon/radeon_sa.c 	if (sa_manager->hole->next == &sa_manager->olist) {
next              197 drivers/gpu/drm/radeon/rv6xx_dpm.c 	struct rv6xx_sclk_stepping next;
next              199 drivers/gpu/drm/radeon/rv6xx_dpm.c 	next.post_divider = cur->post_divider;
next              202 drivers/gpu/drm/radeon/rv6xx_dpm.c 		next.vco_frequency = (cur->vco_frequency * (100 + step_size)) / 100;
next              204 drivers/gpu/drm/radeon/rv6xx_dpm.c 		next.vco_frequency = (cur->vco_frequency * 100 + 99 + step_size) / (100 + step_size);
next              206 drivers/gpu/drm/radeon/rv6xx_dpm.c 	return next;
next              222 drivers/gpu/drm/radeon/rv6xx_dpm.c 	struct rv6xx_sclk_stepping next = *cur;
next              224 drivers/gpu/drm/radeon/rv6xx_dpm.c 	while (rv6xx_can_step_post_div(rdev, &next, target))
next              225 drivers/gpu/drm/radeon/rv6xx_dpm.c 		next.post_divider--;
next              227 drivers/gpu/drm/radeon/rv6xx_dpm.c 	return next;
next              259 drivers/gpu/drm/radeon/rv6xx_dpm.c 		struct rv6xx_sclk_stepping next;
next              262 drivers/gpu/drm/radeon/rv6xx_dpm.c 			next = rv6xx_next_post_div_step(rdev, &cur, &target);
next              264 drivers/gpu/drm/radeon/rv6xx_dpm.c 			next = rv6xx_next_vco_step(rdev, &cur, increasing_vco, R600_VCOSTEPPCT_DFLT);
next              266 drivers/gpu/drm/radeon/rv6xx_dpm.c 		if (rv6xx_reached_stepping_target(rdev, &next, &target, increasing_vco)) {
next              269 drivers/gpu/drm/radeon/rv6xx_dpm.c 			tiny.post_divider = next.post_divider;
next              274 drivers/gpu/drm/radeon/rv6xx_dpm.c 			if ((next.post_divider != target.post_divider) &&
next              275 drivers/gpu/drm/radeon/rv6xx_dpm.c 			    (next.vco_frequency != target.vco_frequency)) {
next              279 drivers/gpu/drm/radeon/rv6xx_dpm.c 				final_vco.post_divider = next.post_divider;
next              287 drivers/gpu/drm/radeon/rv6xx_dpm.c 			rv6xx_output_stepping(rdev, step_index++, &next);
next              289 drivers/gpu/drm/radeon/rv6xx_dpm.c 		cur = next;
next              224 drivers/gpu/drm/savage/savage_bci.c 	dev_priv->head.next = &dev_priv->tail;
next              228 drivers/gpu/drm/savage/savage_bci.c 	dev_priv->tail.next = NULL;
next              239 drivers/gpu/drm/savage/savage_bci.c 		entry->next = dev_priv->head.next;
next              241 drivers/gpu/drm/savage/savage_bci.c 		dev_priv->head.next->prev = entry;
next              242 drivers/gpu/drm/savage/savage_bci.c 		dev_priv->head.next = entry;
next              269 drivers/gpu/drm/savage/savage_bci.c 		drm_savage_buf_priv_t *next = tail->next;
next              271 drivers/gpu/drm/savage/savage_bci.c 		prev->next = next;
next              272 drivers/gpu/drm/savage/savage_bci.c 		next->prev = prev;
next              273 drivers/gpu/drm/savage/savage_bci.c 		tail->next = tail->prev = NULL;
next              284 drivers/gpu/drm/savage/savage_bci.c 	drm_savage_buf_priv_t *entry = buf->dev_private, *prev, *next;
next              288 drivers/gpu/drm/savage/savage_bci.c 	if (entry->next != NULL || entry->prev != NULL) {
next              294 drivers/gpu/drm/savage/savage_bci.c 	next = prev->next;
next              295 drivers/gpu/drm/savage/savage_bci.c 	prev->next = entry;
next              296 drivers/gpu/drm/savage/savage_bci.c 	next->prev = entry;
next              298 drivers/gpu/drm/savage/savage_bci.c 	entry->next = next;
next             1071 drivers/gpu/drm/savage/savage_bci.c 		    buf_priv->next == NULL && buf_priv->prev == NULL) {
next               64 drivers/gpu/drm/savage/savage_drv.h 	struct drm_savage_buf_priv *next;
next              375 drivers/gpu/drm/selftests/test-drm_mm.c 	struct drm_mm_node tmp, *nodes, *node, *next;
next              489 drivers/gpu/drm/selftests/test-drm_mm.c 	drm_mm_for_each_node_safe(node, next, &mm)
next              573 drivers/gpu/drm/selftests/test-drm_mm.c 	struct drm_mm_node *nodes, *node, *next;
next              683 drivers/gpu/drm/selftests/test-drm_mm.c 		drm_mm_for_each_node_safe(node, next, &mm)
next              692 drivers/gpu/drm/selftests/test-drm_mm.c 	drm_mm_for_each_node_safe(node, next, &mm)
next              880 drivers/gpu/drm/selftests/test-drm_mm.c 	struct drm_mm_node *nodes, *node, *next;
next              947 drivers/gpu/drm/selftests/test-drm_mm.c 		drm_mm_for_each_node_safe(node, next, &mm)
next              956 drivers/gpu/drm/selftests/test-drm_mm.c 	drm_mm_for_each_node_safe(node, next, &mm)
next             1041 drivers/gpu/drm/selftests/test-drm_mm.c 	struct drm_mm_node *nodes, *node, *next;
next             1073 drivers/gpu/drm/selftests/test-drm_mm.c 		drm_mm_for_each_node_safe(node, next, &mm)
next             1082 drivers/gpu/drm/selftests/test-drm_mm.c 	drm_mm_for_each_node_safe(node, next, &mm)
next             1093 drivers/gpu/drm/selftests/test-drm_mm.c 	struct drm_mm_node *node, *next;
next             1125 drivers/gpu/drm/selftests/test-drm_mm.c 	drm_mm_for_each_node_safe(node, next, &mm) {
next             1156 drivers/gpu/drm/selftests/test-drm_mm.c 		struct drm_mm_node *next = list_next_entry(hole, node_list);
next             1164 drivers/gpu/drm/selftests/test-drm_mm.c 		if (next->allocated)
next             1167 drivers/gpu/drm/selftests/test-drm_mm.c 					  next->start, next->size, next->color);
next             1261 drivers/gpu/drm/selftests/test-drm_mm.c 		e->link.next = NULL;
next             1266 drivers/gpu/drm/selftests/test-drm_mm.c 		e->link.next = &e->link;
next             1272 drivers/gpu/drm/selftests/test-drm_mm.c 		if (!e->link.next) {
next             1408 drivers/gpu/drm/selftests/test-drm_mm.c 	struct drm_mm_node *node, *next;
next             1501 drivers/gpu/drm/selftests/test-drm_mm.c 	drm_mm_for_each_node_safe(node, next, &mm)
next             1521 drivers/gpu/drm/selftests/test-drm_mm.c 	struct drm_mm_node *node, *next;
next             1598 drivers/gpu/drm/selftests/test-drm_mm.c 	drm_mm_for_each_node_safe(node, next, &mm)
next             1621 drivers/gpu/drm/selftests/test-drm_mm.c 	struct drm_mm_node *nodes, *node, *next;
next             1707 drivers/gpu/drm/selftests/test-drm_mm.c 		drm_mm_for_each_node_safe(node, next, &mm)
next             1715 drivers/gpu/drm/selftests/test-drm_mm.c 	drm_mm_for_each_node_safe(node, next, &mm)
next             1735 drivers/gpu/drm/selftests/test-drm_mm.c 	struct drm_mm_node *nodes, *node, *next;
next             1807 drivers/gpu/drm/selftests/test-drm_mm.c 		drm_mm_for_each_node_safe(node, next, &mm)
next             1815 drivers/gpu/drm/selftests/test-drm_mm.c 	drm_mm_for_each_node_safe(node, next, &mm)
next             2160 drivers/gpu/drm/selftests/test-drm_mm.c 	struct drm_mm_node *node, *next;
next             2242 drivers/gpu/drm/selftests/test-drm_mm.c 	drm_mm_for_each_node_safe(node, next, &mm)
next             2263 drivers/gpu/drm/selftests/test-drm_mm.c 	struct drm_mm_node *node, *next;
next             2343 drivers/gpu/drm/selftests/test-drm_mm.c 	drm_mm_for_each_node_safe(node, next, &mm)
next              318 drivers/gpu/drm/sis/sis_mm.c 	struct sis_memblock *entry, *next;
next              336 drivers/gpu/drm/sis/sis_mm.c 	list_for_each_entry_safe(entry, next, &file_priv->obj_list,
next              997 drivers/gpu/drm/ttm/ttm_page_alloc_dma.c 	struct dma_page *d_page, *next;
next             1009 drivers/gpu/drm/ttm/ttm_page_alloc_dma.c 		list_for_each_entry_safe(d_page, next, &ttm_dma->pages_list,
next             1039 drivers/gpu/drm/ttm/ttm_page_alloc_dma.c 	list_for_each_entry_safe(d_page, next, &ttm_dma->pages_list,
next              182 drivers/gpu/drm/udl/udl_main.c 		node = udl->urbs.list.next; /* have reserved one with sem */
next              283 drivers/gpu/drm/udl/udl_main.c 	entry = udl->urbs.list.next;
next              168 drivers/gpu/drm/vboxvideo/vbva_base.c 	u32 next;
next              176 drivers/gpu/drm/vboxvideo/vbva_base.c 	next = (vbva_ctx->vbva->record_free_index + 1) % VBVA_MAX_RECORDS;
next              179 drivers/gpu/drm/vboxvideo/vbva_base.c 	if (next == vbva_ctx->vbva->record_first_index)
next              183 drivers/gpu/drm/vboxvideo/vbva_base.c 	if (next == vbva_ctx->vbva->record_first_index)
next              188 drivers/gpu/drm/vboxvideo/vbva_base.c 	vbva_ctx->vbva->record_free_index = next;
next              490 drivers/gpu/drm/vc4/vc4_gem.c 		struct vc4_exec_info *next;
next              493 drivers/gpu/drm/vc4/vc4_gem.c 		next = vc4_first_bin_job(vc4);
next              500 drivers/gpu/drm/vc4/vc4_gem.c 		if (next && next->perfmon == exec->perfmon)
next              117 drivers/gpu/drm/vc4/vc4_irq.c 	struct vc4_exec_info *next, *exec = vc4_first_bin_job(vc4);
next              123 drivers/gpu/drm/vc4/vc4_irq.c 	next = vc4_first_bin_job(vc4);
next              129 drivers/gpu/drm/vc4/vc4_irq.c 	if (next && next->perfmon == exec->perfmon)
next               56 drivers/gpu/drm/via/via_dmablit.c 	uint32_t next;
next               74 drivers/gpu/drm/via/via_dmablit.c 	dma_addr_t next = vsg->chain_start;
next               83 drivers/gpu/drm/via/via_dmablit.c 		dma_unmap_single(&pdev->dev, next, sizeof(*desc_ptr), DMA_TO_DEVICE);
next               85 drivers/gpu/drm/via/via_dmablit.c 		next = (dma_addr_t) desc_ptr->next;
next              114 drivers/gpu/drm/via/via_dmablit.c 	dma_addr_t next = 0 | VIA_DMA_DPR_EC;
next              141 drivers/gpu/drm/via/via_dmablit.c 				desc_ptr->next = (uint32_t) next;
next              142 drivers/gpu/drm/via/via_dmablit.c 				next = dma_map_single(&pdev->dev, desc_ptr, sizeof(*desc_ptr),
next              161 drivers/gpu/drm/via/via_dmablit.c 		vsg->chain_start = next;
next              214 drivers/gpu/drm/via/via_mm.c 	struct via_memblock *entry, *next;
next              231 drivers/gpu/drm/via/via_mm.c 	list_for_each_entry_safe(entry, next, &file_priv->obj_list,
next              479 drivers/gpu/drm/vmwgfx/ttm_object.c 		list = tfile->ref_list.next;
next              353 drivers/gpu/drm/vmwgfx/vmwgfx_binding.c 	struct vmw_ctx_bindinfo *entry, *next;
next              356 drivers/gpu/drm/vmwgfx/vmwgfx_binding.c 	list_for_each_entry_safe(entry, next, &cbs->list, ctx_list)
next              395 drivers/gpu/drm/vmwgfx/vmwgfx_binding.c 	struct vmw_ctx_bindinfo *entry, *next;
next              398 drivers/gpu/drm/vmwgfx/vmwgfx_binding.c 	list_for_each_entry_safe(entry, next, head, res_list)
next              447 drivers/gpu/drm/vmwgfx/vmwgfx_binding.c 	struct vmw_ctx_bindinfo *entry, *next;
next              449 drivers/gpu/drm/vmwgfx/vmwgfx_binding.c 	list_for_each_entry_safe(entry, next, &from->list, ctx_list) {
next             1238 drivers/gpu/drm/vmwgfx/vmwgfx_binding.c 	struct vmw_ctx_bindinfo *entry, *next;
next             1240 drivers/gpu/drm/vmwgfx/vmwgfx_binding.c 	list_for_each_entry_safe(entry, next, &cbs->list, ctx_list)
next              381 drivers/gpu/drm/vmwgfx/vmwgfx_cmdbuf.c 	struct vmw_cmdbuf_header *entry, *next;
next              385 drivers/gpu/drm/vmwgfx/vmwgfx_cmdbuf.c 	list_for_each_entry_safe(entry, next, &ctx->hw_submitted, list) {
next              516 drivers/gpu/drm/vmwgfx/vmwgfx_cmdbuf.c 	struct vmw_cmdbuf_header *entry, *next;
next              529 drivers/gpu/drm/vmwgfx/vmwgfx_cmdbuf.c 	list_for_each_entry_safe(entry, next, &man->error, list) {
next              125 drivers/gpu/drm/vmwgfx/vmwgfx_cmdbuf_res.c 	struct vmw_cmdbuf_res *entry, *next;
next              127 drivers/gpu/drm/vmwgfx/vmwgfx_cmdbuf_res.c 	list_for_each_entry_safe(entry, next, list, head) {
next              162 drivers/gpu/drm/vmwgfx/vmwgfx_cmdbuf_res.c 	struct vmw_cmdbuf_res *entry, *next;
next              165 drivers/gpu/drm/vmwgfx/vmwgfx_cmdbuf_res.c 	list_for_each_entry_safe(entry, next, list, head) {
next              320 drivers/gpu/drm/vmwgfx/vmwgfx_cmdbuf_res.c 	struct vmw_cmdbuf_res *entry, *next;
next              322 drivers/gpu/drm/vmwgfx/vmwgfx_cmdbuf_res.c 	list_for_each_entry_safe(entry, next, &man->list, head)
next              324 drivers/gpu/drm/vmwgfx/vmwgfx_drv.h 	bool (*next)(struct vmw_piter *);
next              953 drivers/gpu/drm/vmwgfx/vmwgfx_drv.h 	return viter->next(viter);
next              108 drivers/gpu/drm/vmwgfx/vmwgfx_ldu.c 		entry = list_entry(lds->active.next, typeof(*entry), active);
next               47 drivers/gpu/drm/vmwgfx/vmwgfx_marker.c 	struct vmw_marker *marker, *next;
next               50 drivers/gpu/drm/vmwgfx/vmwgfx_marker.c 	list_for_each_entry_safe(marker, next, &queue->head, head) {
next               76 drivers/gpu/drm/vmwgfx/vmwgfx_marker.c 	struct vmw_marker *marker, *next;
next               90 drivers/gpu/drm/vmwgfx/vmwgfx_marker.c 	list_for_each_entry_safe(marker, next, &queue->head, head) {
next              716 drivers/gpu/drm/vmwgfx/vmwgfx_resource.c 	struct vmw_resource *res, *next;
next              723 drivers/gpu/drm/vmwgfx/vmwgfx_resource.c 	list_for_each_entry_safe(res, next, &vbo->res_list, mob_head) {
next              550 drivers/gpu/drm/vmwgfx/vmwgfx_shader.c 	struct vmw_dx_shader *entry, *next;
next              554 drivers/gpu/drm/vmwgfx/vmwgfx_shader.c 	list_for_each_entry_safe(entry, next, list, cotable_head) {
next              436 drivers/gpu/drm/vmwgfx/vmwgfx_so.c 	struct vmw_view *entry, *next;
next              440 drivers/gpu/drm/vmwgfx/vmwgfx_so.c 	list_for_each_entry_safe(entry, next, list, cotable_head)
next              456 drivers/gpu/drm/vmwgfx/vmwgfx_so.c 	struct vmw_view *entry, *next;
next              460 drivers/gpu/drm/vmwgfx/vmwgfx_so.c 	list_for_each_entry_safe(entry, next, list, srf_head)
next              333 drivers/gpu/drm/vmwgfx/vmwgfx_ttm_buffer.c 		viter->next = &__vmw_piter_non_sg_next;
next              337 drivers/gpu/drm/vmwgfx/vmwgfx_ttm_buffer.c 		viter->next = &__vmw_piter_non_sg_next;
next              343 drivers/gpu/drm/vmwgfx/vmwgfx_ttm_buffer.c 		viter->next = &__vmw_piter_sg_next;
next              149 drivers/gpu/drm/vmwgfx/vmwgfx_validation.c 	struct page *entry, *next;
next              151 drivers/gpu/drm/vmwgfx/vmwgfx_validation.c 	list_for_each_entry_safe(entry, next, &ctx->page_list, lru) {
next               60 drivers/gpu/host1x/intr.c 	struct host1x_waitlist *waiter, *next, *prev;
next               62 drivers/gpu/host1x/intr.c 	list_for_each_entry_safe(waiter, next, head, list) {
next              140 drivers/gpu/host1x/intr.c 		struct host1x_waitlist *waiter, *next;
next              142 drivers/gpu/host1x/intr.c 		list_for_each_entry_safe(waiter, next, head, list) {
next              316 drivers/gpu/host1x/intr.c 		struct host1x_waitlist *waiter, *next;
next              318 drivers/gpu/host1x/intr.c 		list_for_each_entry_safe(waiter, next,
next             1541 drivers/gpu/ipu-v3/ipu-image-convert.c 		run = list_entry(chan->done_q.next,
next             1184 drivers/greybus/interface.c 	struct gb_bundle *next;
next             1197 drivers/greybus/interface.c 	list_for_each_entry_safe(bundle, next, &intf->bundles, links)
next               54 drivers/greybus/manifest.c 	struct manifest_desc *next;
next               56 drivers/greybus/manifest.c 	list_for_each_entry_safe(descriptor, next, &intf->manifest_descs, links)
next               79 drivers/greybus/manifest.c 	struct manifest_desc *next;
next               81 drivers/greybus/manifest.c 	list_for_each_entry_safe(descriptor, next, &intf->manifest_descs, links)
next              232 drivers/greybus/manifest.c 	struct manifest_desc *desc, *next, *tmp;
next              240 drivers/greybus/manifest.c 	list_for_each_entry_safe(desc, next, &intf->manifest_descs, links) {
next              286 drivers/greybus/manifest.c 	list_for_each_entry_safe(desc, next, &list, links) {
next              347 drivers/hid/hid-bigbenff.c 	bigben->report = list_entry(report_list->next,
next              985 drivers/hid/hid-core.c 				hid->report_enum[type].report_list.next,
next             1191 drivers/hid/hid-core.c 	__u8 *next;
next             1245 drivers/hid/hid-core.c 	while ((next = fetch_item(start, end, &item)) != NULL) {
next             1246 drivers/hid/hid-core.c 		start = next;
next              198 drivers/hid/hid-cougar.c 	struct hid_input *next, *hidinput = NULL;
next              234 drivers/hid/hid-cougar.c 		list_for_each_entry_safe(hidinput, next, &hdev->inputs, list) {
next              630 drivers/hid/hid-debug.c 		list = report_enum->report_list.next;
next              644 drivers/hid/hid-debug.c 			list = list->next;
next               78 drivers/hid/hid-gaff.c 	hidinput = list_entry(hid->inputs.next, struct hid_input, list);
next               86 drivers/hid/hid-gaff.c 	report_ptr = report_ptr->next;
next              137 drivers/hid/hid-holtekff.c 	hidinput = list_entry(hid->inputs.next, struct hid_input, list);
next              145 drivers/hid/hid-holtekff.c 	report = list_entry(report_list->next, struct hid_report, list);
next             1860 drivers/hid/hid-input.c 	struct hid_input *next, *hidinput = NULL;
next             1924 drivers/hid/hid-input.c 	list_for_each_entry_safe(hidinput, next, &hid->inputs, list) {
next             1961 drivers/hid/hid-input.c 	struct hid_input *hidinput, *next;
next             1965 drivers/hid/hid-input.c 	list_for_each_entry_safe(hidinput, next, &hid->inputs, list) {
next               61 drivers/hid/hid-lg2ff.c 	hidinput = list_entry(hid->inputs.next, struct hid_input, list);
next               53 drivers/hid/hid-lg3ff.c 	struct hid_report *report = list_entry(report_list->next, struct hid_report, list);
next               91 drivers/hid/hid-lg3ff.c 	struct hid_report *report = list_entry(report_list->next, struct hid_report, list);
next              130 drivers/hid/hid-lg3ff.c 	hidinput = list_entry(hid->inputs.next, struct hid_input, list);
next             1259 drivers/hid/hid-lg4ff.c 	struct hid_report *report = list_entry(report_list->next, struct hid_report, list);
next             1273 drivers/hid/hid-lg4ff.c 	hidinput = list_entry(hid->inputs.next, struct hid_input, list);
next               61 drivers/hid/hid-lgff.c 	struct hid_report *report = list_entry(report_list->next, struct hid_report, list);
next              103 drivers/hid/hid-lgff.c 	struct hid_report *report = list_entry(report_list->next, struct hid_report, list);
next              128 drivers/hid/hid-lgff.c 	hidinput = list_entry(hid->inputs.next, struct hid_input, list);
next             2053 drivers/hid/hid-logitech-hidpp.c 	struct hid_input *hidinput = list_entry(hid->inputs.next, struct hid_input, list);
next             2063 drivers/hid/hid-logitech-hidpp.c 	struct hid_input *hidinput = list_entry(hid->inputs.next, struct hid_input, list);
next             2109 drivers/hid/hid-logitech-hidpp.c 	hidinput = list_entry(hid->inputs.next, struct hid_input, list);
next               88 drivers/hid/hid-mf.c 		input_ptr = input_ptr->next;
next              339 drivers/hid/hid-microsoft.c 	hidinput = list_entry(hdev->inputs.next, struct hid_input, list);
next              103 drivers/hid/hid-pl.c 		report_ptr = report_ptr->next;
next               68 drivers/hid/hid-sjoy.c 		report_ptr = report_ptr->next;
next              942 drivers/hid/hid-sony.c 	struct hid_input *hidinput = list_entry(sc->hdev->inputs.next,
next             1828 drivers/hid/hid-sony.c 	struct hid_report *report = list_entry(report_list->next,
next             2281 drivers/hid/hid-sony.c 	hidinput = list_entry(sc->hdev->inputs.next, struct hid_input, list);
next             2430 drivers/hid/hid-sony.c 	if (sc->list_node.next) {
next              112 drivers/hid/hid-steelseries.c 	struct hid_report *report = list_entry(report_list->next, struct hid_report, list);
next              136 drivers/hid/hid-tmff.c 	hidinput = list_entry(hid->inputs.next, struct hid_input, list);
next               65 drivers/hid/hid-zpff.c 	hidinput = list_entry(hid->inputs.next, struct hid_input, list);
next              105 drivers/hid/intel-ish-hid/ishtp/client-buffers.c 		rb = list_entry(cl->free_rb_list.list.next, struct ishtp_cl_rb,
next              115 drivers/hid/intel-ish-hid/ishtp/client-buffers.c 		rb = list_entry(cl->in_process_list.list.next,
next              138 drivers/hid/intel-ish-hid/ishtp/client-buffers.c 		tx_buf = list_entry(cl->tx_free_list.list.next,
next              150 drivers/hid/intel-ish-hid/ishtp/client-buffers.c 		tx_buf = list_entry(cl->tx_list.list.next,
next               44 drivers/hid/intel-ish-hid/ishtp/client.c 	struct ishtp_cl_rb *next;
next               48 drivers/hid/intel-ish-hid/ishtp/client.c 	list_for_each_entry_safe(rb, next, &cl->dev->read_list.list, list)
next              481 drivers/hid/intel-ish-hid/ishtp/client.c 	rb = list_entry(cl->free_rb_list.list.next, struct ishtp_cl_rb, list);
next              676 drivers/hid/intel-ish-hid/ishtp/client.c 	cl_msg = list_entry(cl->tx_list.list.next, struct ishtp_cl_tx_ring,
next              754 drivers/hid/intel-ish-hid/ishtp/client.c 	cl_msg = list_entry(cl->tx_list.list.next, struct ishtp_cl_tx_ring,
next              896 drivers/hid/intel-ish-hid/ishtp/client.c 				new_rb = list_entry(cl->free_rb_list.list.next,
next             1016 drivers/hid/intel-ish-hid/ishtp/client.c 			new_rb = list_entry(cl->free_rb_list.list.next,
next             1229 drivers/hid/usbhid/hid-pidff.c 	struct hid_input *hidinput = list_entry(hid->inputs.next,
next               79 drivers/hid/usbhid/hiddev.c 		list = report_enum->report_list.next;
next               89 drivers/hid/usbhid/hiddev.c 		list = report->list.next;
next             1475 drivers/hid/wacom_sys.c 	int group, next;
next             1481 drivers/hid/wacom_sys.c 	next = cur->id;
next             1484 drivers/hid/wacom_sys.c 		next_led = wacom_led_find(wacom, group, ++next);
next              167 drivers/hsi/clients/cmt_speech.c 	entry = list_entry(head->next, struct char_queue, list);
next             1323 drivers/hsi/clients/cmt_speech.c 	struct list_head *cursor, *next;
next             1326 drivers/hsi/clients/cmt_speech.c 		list_for_each_safe(cursor, next, head) {
next               74 drivers/hv/ring_buffer.c 	u32 next = ring_info->ring_buffer->write_index;
next               76 drivers/hv/ring_buffer.c 	return next;
next              250 drivers/hwmon/ibmpex.c 	struct ibmpex_bmc_data *p, *next;
next              252 drivers/hwmon/ibmpex.c 	list_for_each_entry_safe(p, next, &driver_data.bmc_data, list)
next              582 drivers/hwmon/ibmpex.c 	struct ibmpex_bmc_data *p, *next;
next              585 drivers/hwmon/ibmpex.c 	list_for_each_entry_safe(p, next, &driver_data.bmc_data, list)
next               48 drivers/hwmon/pmbus/pmbus_core.c 	struct pmbus_sensor *next;
next              407 drivers/hwmon/pmbus/pmbus_core.c 	for (sensor = data->sensors; sensor; sensor = sensor->next) {
next              593 drivers/hwmon/pmbus/pmbus_core.c 		for (sensor = data->sensors; sensor; sensor = sensor->next) {
next             1103 drivers/hwmon/pmbus/pmbus_core.c 	sensor->next = data->sensors;
next              699 drivers/hwtracing/coresight/coresight.c 	struct coresight_node *nd, *next;
next              701 drivers/hwtracing/coresight/coresight.c 	list_for_each_entry_safe(nd, next, path, link) {
next              302 drivers/hwtracing/intel_th/msu.c 	return win->entry.next == &win->msc->win_list;
next              506 drivers/hwtracing/stm/policy.c next:
next              526 drivers/hwtracing/stm/policy.c 				goto next;
next               64 drivers/i2c/busses/scx200_acb.c 	struct scx200_acb_iface *next;
next              461 drivers/i2c/busses/scx200_acb.c 		iface->next = scx200_acb_list;
next              589 drivers/i2c/busses/scx200_acb.c 		scx200_acb_list = iface->next;
next             1131 drivers/i2c/i2c-core-base.c 	struct i2c_client *client, *next;
next             1151 drivers/i2c/i2c-core-base.c 	list_for_each_entry_safe(client, next, &adap->userspace_clients,
next             1553 drivers/i2c/i2c-core-base.c 	struct i2c_client *client, *next;
next             1574 drivers/i2c/i2c-core-base.c 	list_for_each_entry_safe(client, next, &adap->userspace_clients,
next              775 drivers/iio/adc/ina2xx-adc.c 	struct timespec64 next, now, delta;
next              785 drivers/iio/adc/ina2xx-adc.c 	ktime_get_ts64(&next);
next              798 drivers/iio/adc/ina2xx-adc.c 				ktime_get_ts64(&next);
next              816 drivers/iio/adc/ina2xx-adc.c 			timespec64_add_ns(&next, 1000 * sampling_us);
next              817 drivers/iio/adc/ina2xx-adc.c 			delta = timespec64_sub(next, now);
next               61 drivers/iio/inkern.c 	struct iio_map_internal *mapi, *next;
next               64 drivers/iio/inkern.c 	list_for_each_entry_safe(mapi, next, &iio_map_list, l) {
next              897 drivers/infiniband/core/cm.c 	work = list_entry(cm_id_priv->work_list.next, struct cm_work, list);
next             1753 drivers/infiniband/core/cma.c 		dev_id_priv = list_entry(id_priv->listen_list.next,
next             1822 drivers/infiniband/core/cma.c 		mc = container_of(id_priv->mc_list.next,
next             3439 drivers/infiniband/core/cma.c 	if (bind_list->owners.first->next)
next             4652 drivers/infiniband/core/cma.c 		id_priv = list_entry(cma_dev->id_list.next,
next              237 drivers/infiniband/core/counters.c 			goto next;
next              241 drivers/infiniband/core/counters.c next:
next              377 drivers/infiniband/core/counters.c 			goto next;
next              381 drivers/infiniband/core/counters.c next:
next              359 drivers/infiniband/core/fmr_pool.c 	struct ib_pool_fmr *fmr, *next;
next              368 drivers/infiniband/core/fmr_pool.c 	list_for_each_entry_safe(fmr, next, &pool->free_list, list) {
next              427 drivers/infiniband/core/fmr_pool.c 	fmr = list_entry(pool->free_list.next, struct ib_pool_fmr, list);
next              147 drivers/infiniband/core/iwcm.c 	work = list_entry(cm_id_priv->work_free_list.next, struct iwcm_work,
next             1028 drivers/infiniband/core/iwcm.c 		work = list_entry(cm_id_priv->work_list.next,
next             1037 drivers/infiniband/core/mad.c 	send_wr->cur_seg = container_of(send_wr->rmpp_list.next,
next             1319 drivers/infiniband/core/mad.c 		next_send_buf = send_buf->next;
next             2398 drivers/infiniband/core/mad.c 		mad_send_wr = list_entry(mad_agent_priv->wait_list.next,
next             2442 drivers/infiniband/core/mad.c 	if (mad_agent_priv->wait_list.next == &mad_send_wr->agent_list)
next             2548 drivers/infiniband/core/mad.c 		mad_list = container_of(qp_info->overflow_list.next,
next             2761 drivers/infiniband/core/mad.c 		local = list_entry(mad_agent_priv->local_list.next,
next             2887 drivers/infiniband/core/mad.c 		mad_send_wr = list_entry(mad_agent_priv->wait_list.next,
next             2939 drivers/infiniband/core/mad.c 	recv_wr.next = NULL;
next             3010 drivers/infiniband/core/mad.c 		mad_list = list_entry(qp_info->recv_queue.list.next,
next              397 drivers/infiniband/core/mad_rmpp.c 	if (seg->list.next == rmpp_list)
next              400 drivers/infiniband/core/mad_rmpp.c 	return container_of(seg->list.next, struct ib_mad_recv_buf, list);
next              402 drivers/infiniband/core/multicast.c 		member = list_entry(group->active_list.next,
next              444 drivers/infiniband/core/multicast.c 		member = list_entry(group->pending_list.next,
next              497 drivers/infiniband/core/multicast.c 	member = list_entry(group->pending_list.next,
next             1345 drivers/infiniband/core/nldev.c 			goto next;
next             1371 drivers/infiniband/core/nldev.c next:		idx++;
next               75 drivers/infiniband/core/rw.c 		reg->inv_wr.next = &reg->reg_wr.wr;
next               78 drivers/infiniband/core/rw.c 		reg->inv_wr.next = NULL;
next              146 drivers/infiniband/core/rw.c 				prev->wr.wr.next = &reg->inv_wr;
next              148 drivers/infiniband/core/rw.c 				prev->wr.wr.next = &reg->reg_wr.wr;
next              151 drivers/infiniband/core/rw.c 		reg->reg_wr.wr.next = &reg->wr.wr;
next              176 drivers/infiniband/core/rw.c 		prev->wr.wr.next = NULL;
next              232 drivers/infiniband/core/rw.c 		rdma_wr->wr.next = i + 1 < ctx->nr_ops ?
next              453 drivers/infiniband/core/rw.c 	ctx->reg->reg_wr.wr.next = &rdma_wr->wr;
next              515 drivers/infiniband/core/rw.c 		if (ctx->reg[0].inv_wr.next)
next              534 drivers/infiniband/core/rw.c 		last_wr->next = chain_wr;
next              881 drivers/infiniband/core/sa_query.c 	if (ib_nl_request_list.next == &query->list)
next              985 drivers/infiniband/core/sa_query.c 		query = list_entry(ib_nl_request_list.next,
next              427 drivers/infiniband/core/ucma.c 	uevent = list_entry(file->event_list.next, struct ucma_event, list);
next              776 drivers/infiniband/core/umem_odp.c 	struct interval_tree_node *node, *next;
next              783 drivers/infiniband/core/umem_odp.c 			node; node = next) {
next              787 drivers/infiniband/core/umem_odp.c 		next = interval_tree_iter_next(node, start, last - 1);
next              334 drivers/infiniband/core/user_mad.c 			recv_buf = container_of(recv_buf->list.next,
next              395 drivers/infiniband/core/user_mad.c 	packet = list_entry(file->recv_list.next, struct ib_umad_packet, list);
next             2004 drivers/infiniband/core/uverbs_cmd.c 	struct ib_send_wr              *wr = NULL, *last, *next;
next             2080 drivers/infiniband/core/uverbs_cmd.c 			next = &ud->wr;
next             2096 drivers/infiniband/core/uverbs_cmd.c 			next = &rdma->wr;
next             2113 drivers/infiniband/core/uverbs_cmd.c 			next = &atomic->wr;
next             2117 drivers/infiniband/core/uverbs_cmd.c 			next_size = sizeof(*next);
next             2118 drivers/infiniband/core/uverbs_cmd.c 			next = alloc_wr(next_size, user_wr->num_sge);
next             2119 drivers/infiniband/core/uverbs_cmd.c 			if (!next) {
next             2130 drivers/infiniband/core/uverbs_cmd.c 			next->ex.imm_data =
next             2133 drivers/infiniband/core/uverbs_cmd.c 			next->ex.invalidate_rkey = user_wr->ex.invalidate_rkey;
next             2137 drivers/infiniband/core/uverbs_cmd.c 			wr = next;
next             2139 drivers/infiniband/core/uverbs_cmd.c 			last->next = next;
next             2140 drivers/infiniband/core/uverbs_cmd.c 		last = next;
next             2142 drivers/infiniband/core/uverbs_cmd.c 		next->next       = NULL;
next             2143 drivers/infiniband/core/uverbs_cmd.c 		next->wr_id      = user_wr->wr_id;
next             2144 drivers/infiniband/core/uverbs_cmd.c 		next->num_sge    = user_wr->num_sge;
next             2145 drivers/infiniband/core/uverbs_cmd.c 		next->opcode     = user_wr->opcode;
next             2146 drivers/infiniband/core/uverbs_cmd.c 		next->send_flags = user_wr->send_flags;
next             2148 drivers/infiniband/core/uverbs_cmd.c 		if (next->num_sge) {
next             2149 drivers/infiniband/core/uverbs_cmd.c 			next->sg_list = (void *) next +
next             2151 drivers/infiniband/core/uverbs_cmd.c 			if (copy_from_user(next->sg_list, sgls + sg_ind,
next             2152 drivers/infiniband/core/uverbs_cmd.c 					   next->num_sge *
next             2157 drivers/infiniband/core/uverbs_cmd.c 			sg_ind += next->num_sge;
next             2159 drivers/infiniband/core/uverbs_cmd.c 			next->sg_list = NULL;
next             2165 drivers/infiniband/core/uverbs_cmd.c 		for (next = wr; next; next = next->next) {
next             2167 drivers/infiniband/core/uverbs_cmd.c 			if (next == bad_wr)
next             2181 drivers/infiniband/core/uverbs_cmd.c 		next = wr->next;
next             2183 drivers/infiniband/core/uverbs_cmd.c 		wr = next;
next             2197 drivers/infiniband/core/uverbs_cmd.c 	struct ib_recv_wr        *wr = NULL, *last, *next;
next             2237 drivers/infiniband/core/uverbs_cmd.c 		    (U32_MAX - ALIGN(sizeof *next, sizeof (struct ib_sge))) /
next             2243 drivers/infiniband/core/uverbs_cmd.c 		next = kmalloc(ALIGN(sizeof *next, sizeof (struct ib_sge)) +
next             2246 drivers/infiniband/core/uverbs_cmd.c 		if (!next) {
next             2252 drivers/infiniband/core/uverbs_cmd.c 			wr = next;
next             2254 drivers/infiniband/core/uverbs_cmd.c 			last->next = next;
next             2255 drivers/infiniband/core/uverbs_cmd.c 		last = next;
next             2257 drivers/infiniband/core/uverbs_cmd.c 		next->next       = NULL;
next             2258 drivers/infiniband/core/uverbs_cmd.c 		next->wr_id      = user_wr->wr_id;
next             2259 drivers/infiniband/core/uverbs_cmd.c 		next->num_sge    = user_wr->num_sge;
next             2261 drivers/infiniband/core/uverbs_cmd.c 		if (next->num_sge) {
next             2262 drivers/infiniband/core/uverbs_cmd.c 			next->sg_list = (void *) next +
next             2263 drivers/infiniband/core/uverbs_cmd.c 				ALIGN(sizeof *next, sizeof (struct ib_sge));
next             2264 drivers/infiniband/core/uverbs_cmd.c 			if (copy_from_user(next->sg_list, sgls + sg_ind,
next             2265 drivers/infiniband/core/uverbs_cmd.c 					   next->num_sge *
next             2270 drivers/infiniband/core/uverbs_cmd.c 			sg_ind += next->num_sge;
next             2272 drivers/infiniband/core/uverbs_cmd.c 			next->sg_list = NULL;
next             2282 drivers/infiniband/core/uverbs_cmd.c 		next = wr->next;
next             2284 drivers/infiniband/core/uverbs_cmd.c 		wr = next;
next             2294 drivers/infiniband/core/uverbs_cmd.c 	struct ib_recv_wr              *wr, *next;
next             2320 drivers/infiniband/core/uverbs_cmd.c 		for (next = wr; next; next = next->next) {
next             2322 drivers/infiniband/core/uverbs_cmd.c 			if (next == bad_wr)
next             2332 drivers/infiniband/core/uverbs_cmd.c 		next = wr->next;
next             2334 drivers/infiniband/core/uverbs_cmd.c 		wr = next;
next             2344 drivers/infiniband/core/uverbs_cmd.c 	struct ib_recv_wr                  *wr, *next;
next             2371 drivers/infiniband/core/uverbs_cmd.c 		for (next = wr; next; next = next->next) {
next             2373 drivers/infiniband/core/uverbs_cmd.c 			if (next == bad_wr)
next             2383 drivers/infiniband/core/uverbs_cmd.c 		next = wr->next;
next             2385 drivers/infiniband/core/uverbs_cmd.c 		wr = next;
next               39 drivers/infiniband/core/uverbs_ioctl.c 	struct bundle_alloc_head *next;
next              122 drivers/infiniband/core/uverbs_ioctl.c 		buf->next = pbundle->allocated_mem;
next              548 drivers/infiniband/core/uverbs_ioctl.c 		memblock = memblock->next;
next              586 drivers/infiniband/core/uverbs_ioctl.c 		pbundle->alloc_head.next = NULL;
next              252 drivers/infiniband/core/uverbs_main.c 	event = list_entry(ev_queue->event_list.next, struct ib_uverbs_event, list);
next              258 drivers/infiniband/core/uverbs_main.c 		list_del(ev_queue->event_list.next);
next             2124 drivers/infiniband/core/verbs.c 	fmr = list_entry(fmr_list->next, struct ib_fmr, list);
next             2248 drivers/infiniband/core/verbs.c 		qp = list_entry(xrcd->tgt_qp_list.next, struct ib_qp, xrcd_list);
next             2637 drivers/infiniband/core/verbs.c 			.next = NULL,
next             1496 drivers/infiniband/hw/bnxt_re/ib_verbs.c 		wr = wr->next;
next             2316 drivers/infiniband/hw/bnxt_re/ib_verbs.c 		wr = wr->next;
next             2413 drivers/infiniband/hw/bnxt_re/ib_verbs.c 		wr = wr->next;
next             2450 drivers/infiniband/hw/bnxt_re/ib_verbs.c 		wr = wr->next;
next             2503 drivers/infiniband/hw/bnxt_re/ib_verbs.c 		wr = wr->next;
next             2943 drivers/infiniband/hw/bnxt_re/ib_verbs.c 	rwr.next = NULL;
next             2956 drivers/infiniband/hw/bnxt_re/ib_verbs.c 	swr->next = NULL;
next             1709 drivers/infiniband/hw/bnxt_re/main.c 	struct bnxt_re_dev *rdev, *next;
next             1721 drivers/infiniband/hw/bnxt_re/main.c 	list_for_each_entry_safe_reverse(rdev, next, &to_be_deleted, list) {
next              667 drivers/infiniband/hw/bnxt_re/qplib_fp.c 	int i, rc = 0, next;
next              677 drivers/infiniband/hw/bnxt_re/qplib_fp.c 	next = srq->start_idx;
next              678 drivers/infiniband/hw/bnxt_re/qplib_fp.c 	srq->start_idx = srq->swq[next].next_idx;
next              696 drivers/infiniband/hw/bnxt_re/qplib_fp.c 	srqe->wr_id[0] = cpu_to_le32((u32)next);
next              697 drivers/infiniband/hw/bnxt_re/qplib_fp.c 	srq->swq[next].wr_id = wqe->wr_id;
next              197 drivers/infiniband/hw/cxgb3/cxio_hal.c 		entry = list_entry(uctx->qpids.next, struct cxio_qpid_list,
next              937 drivers/infiniband/hw/cxgb3/iwch_provider.c 	char *cp, *next;
next              942 drivers/infiniband/hw/cxgb3/iwch_provider.c 	next = info.fw_version + 1;
next              943 drivers/infiniband/hw/cxgb3/iwch_provider.c 	cp = strsep(&next, ".");
next              945 drivers/infiniband/hw/cxgb3/iwch_provider.c 	cp = strsep(&next, ".");
next              947 drivers/infiniband/hw/cxgb3/iwch_provider.c 	cp = strsep(&next, ".");
next              451 drivers/infiniband/hw/cxgb3/iwch_qp.c 		wr = wr->next;
next              516 drivers/infiniband/hw/cxgb3/iwch_qp.c 		wr = wr->next;
next             4278 drivers/infiniband/hw/cxgb4/cm.c 		tmp = timeout_list.next;
next             4280 drivers/infiniband/hw/cxgb4/cm.c 		tmp->next = NULL;
next             4329 drivers/infiniband/hw/cxgb4/cm.c 		if (!ep->entry.next) {
next              635 drivers/infiniband/hw/cxgb4/qp.c 	if (wr->next->opcode == IB_WR_SEND)
next              638 drivers/infiniband/hw/cxgb4/qp.c 		wcwr->stag_inv = cpu_to_be32(wr->next->ex.invalidate_rkey);
next              643 drivers/infiniband/hw/cxgb4/qp.c 	if (wr->next->send_flags & IB_SEND_INLINE)
next              644 drivers/infiniband/hw/cxgb4/qp.c 		build_immd_cmpl(sq, &wcwr->u_cmpl.immd_src, wr->next);
next              647 drivers/infiniband/hw/cxgb4/qp.c 			   &wcwr->u_cmpl.isgl_src, wr->next->sg_list, 1, NULL);
next              691 drivers/infiniband/hw/cxgb4/qp.c 	bool send_signaled = (wr->next->send_flags & IB_SEND_SIGNALED) ||
next              732 drivers/infiniband/hw/cxgb4/qp.c 	if (wr->next->opcode == IB_WR_SEND)
next              740 drivers/infiniband/hw/cxgb4/qp.c 	swsqe->wr_id = wr->next->wr_id;
next             1029 drivers/infiniband/hw/cxgb4/qp.c 		wr = wr->next;
next             1072 drivers/infiniband/hw/cxgb4/qp.c 		wr = wr->next;
next             1123 drivers/infiniband/hw/cxgb4/qp.c 	    wr && wr->next && !wr->next->next &&
next             1126 drivers/infiniband/hw/cxgb4/qp.c 	    (wr->next->opcode == IB_WR_SEND ||
next             1127 drivers/infiniband/hw/cxgb4/qp.c 	    wr->next->opcode == IB_WR_SEND_WITH_INV) &&
next             1128 drivers/infiniband/hw/cxgb4/qp.c 	    wr->next->sg_list[0].length == T4_WRITE_CMPL_MAX_CQE &&
next             1129 drivers/infiniband/hw/cxgb4/qp.c 	    wr->next->num_sge == 1 && num_wrs >= 2) {
next             1245 drivers/infiniband/hw/cxgb4/qp.c 		wr = wr->next;
next             1327 drivers/infiniband/hw/cxgb4/qp.c 		wr = wr->next;
next             1414 drivers/infiniband/hw/cxgb4/qp.c 		wr = wr->next;
next              116 drivers/infiniband/hw/cxgb4/resource.c 		entry = list_entry(uctx->cqids.next, struct c4iw_qid_list,
next              186 drivers/infiniband/hw/cxgb4/resource.c 		entry = list_entry(uctx->qpids.next, struct c4iw_qid_list,
next               55 drivers/infiniband/hw/hfi1/debugfs.h 	.next  = _##name##_seq_next, \
next              637 drivers/infiniband/hw/hfi1/driver.c 			goto next;
next              640 drivers/infiniband/hw/hfi1/driver.c 			goto next;
next              653 drivers/infiniband/hw/hfi1/driver.c 			goto next; /* just in case */
next              657 drivers/infiniband/hw/hfi1/driver.c 			goto next;
next              666 drivers/infiniband/hw/hfi1/driver.c 			goto next;
next              675 drivers/infiniband/hw/hfi1/driver.c next:
next             1776 drivers/infiniband/hw/hfi1/driver.c 			goto next;
next             1779 drivers/infiniband/hw/hfi1/driver.c 			goto next;
next             1791 drivers/infiniband/hw/hfi1/driver.c 			goto next; /* just in case */
next             1799 drivers/infiniband/hw/hfi1/driver.c next:
next             1458 drivers/infiniband/hw/hfi1/pio.c 	u32 head, next;
next             1518 drivers/infiniband/hw/hfi1/pio.c 	next = head + 1;
next             1519 drivers/infiniband/hw/hfi1/pio.c 	if (next >= sc->sr_size)
next             1520 drivers/infiniband/hw/hfi1/pio.c 		next = 0;
next             1526 drivers/infiniband/hw/hfi1/pio.c 	sc->sr_head = next;
next              123 drivers/infiniband/hw/hfi1/rc.c 	u8 next = qp->s_tail_ack_queue;
next              151 drivers/infiniband/hw/hfi1/rc.c 		if (++next > rvt_size_atomic(&dev->rdi))
next              152 drivers/infiniband/hw/hfi1/rc.c 			next = 0;
next              160 drivers/infiniband/hw/hfi1/rc.c 			qp->s_acked_ack_queue = next;
next              161 drivers/infiniband/hw/hfi1/rc.c 		qp->s_tail_ack_queue = next;
next             3060 drivers/infiniband/hw/hfi1/rc.c 		u8 next;
next             3064 drivers/infiniband/hw/hfi1/rc.c 		next = qp->r_head_ack_queue + 1;
next             3066 drivers/infiniband/hw/hfi1/rc.c 		if (next > rvt_size_atomic(ib_to_rvt(qp->ibqp.device)))
next             3067 drivers/infiniband/hw/hfi1/rc.c 			next = 0;
next             3069 drivers/infiniband/hw/hfi1/rc.c 		if (unlikely(next == qp->s_acked_ack_queue)) {
next             3070 drivers/infiniband/hw/hfi1/rc.c 			if (!qp->s_ack_queue[next].sent)
next             3072 drivers/infiniband/hw/hfi1/rc.c 			update_ack_queue(qp, next);
next             3112 drivers/infiniband/hw/hfi1/rc.c 		qp->r_head_ack_queue = next;
next             3135 drivers/infiniband/hw/hfi1/rc.c 		u8 next;
next             3140 drivers/infiniband/hw/hfi1/rc.c 		next = qp->r_head_ack_queue + 1;
next             3141 drivers/infiniband/hw/hfi1/rc.c 		if (next > rvt_size_atomic(ib_to_rvt(qp->ibqp.device)))
next             3142 drivers/infiniband/hw/hfi1/rc.c 			next = 0;
next             3144 drivers/infiniband/hw/hfi1/rc.c 		if (unlikely(next == qp->s_acked_ack_queue)) {
next             3145 drivers/infiniband/hw/hfi1/rc.c 			if (!qp->s_ack_queue[next].sent)
next             3147 drivers/infiniband/hw/hfi1/rc.c 			update_ack_queue(qp, next);
next             3183 drivers/infiniband/hw/hfi1/rc.c 		qp->r_head_ack_queue = next;
next               15 drivers/infiniband/hw/hfi1/rc.h 	unsigned int next;
next               17 drivers/infiniband/hw/hfi1/rc.h 	next = n + 1;
next               18 drivers/infiniband/hw/hfi1/rc.h 	if (next > rvt_size_atomic(ib_to_rvt(qp->ibqp.device)))
next               19 drivers/infiniband/hw/hfi1/rc.h 		next = 0;
next               20 drivers/infiniband/hw/hfi1/rc.h 	qp->s_tail_ack_queue = next;
next               21 drivers/infiniband/hw/hfi1/rc.h 	qp->s_acked_ack_queue = next;
next             2233 drivers/infiniband/hw/hfi1/tid_rdma.c 	u8 next;
next             2266 drivers/infiniband/hw/hfi1/tid_rdma.c 	next = qp->r_head_ack_queue + 1;
next             2267 drivers/infiniband/hw/hfi1/tid_rdma.c 	if (next > rvt_size_atomic(ib_to_rvt(qp->ibqp.device)))
next             2268 drivers/infiniband/hw/hfi1/tid_rdma.c 		next = 0;
next             2270 drivers/infiniband/hw/hfi1/tid_rdma.c 	if (unlikely(next == qp->s_tail_ack_queue)) {
next             2271 drivers/infiniband/hw/hfi1/tid_rdma.c 		if (!qp->s_ack_queue[next].sent) {
next             2275 drivers/infiniband/hw/hfi1/tid_rdma.c 		update_ack_queue(qp, next);
next             2302 drivers/infiniband/hw/hfi1/tid_rdma.c 	qp->r_head_ack_queue = next;
next             3657 drivers/infiniband/hw/hfi1/tid_rdma.c 	u8 next;
next             3697 drivers/infiniband/hw/hfi1/tid_rdma.c 	next = qp->r_head_ack_queue + 1;
next             3698 drivers/infiniband/hw/hfi1/tid_rdma.c 	if (next > rvt_size_atomic(ib_to_rvt(qp->ibqp.device)))
next             3699 drivers/infiniband/hw/hfi1/tid_rdma.c 		next = 0;
next             3701 drivers/infiniband/hw/hfi1/tid_rdma.c 	if (unlikely(next == qp->s_acked_ack_queue)) {
next             3702 drivers/infiniband/hw/hfi1/tid_rdma.c 		if (!qp->s_ack_queue[next].sent)
next             3704 drivers/infiniband/hw/hfi1/tid_rdma.c 		update_ack_queue(qp, next);
next             3791 drivers/infiniband/hw/hfi1/tid_rdma.c 	qp->r_head_ack_queue = next;
next             4267 drivers/infiniband/hw/hfi1/tid_rdma.c 	u32 psn, next;
next             4355 drivers/infiniband/hw/hfi1/tid_rdma.c 		for (next = priv->r_tid_tail + 1; ; next++) {
next             4356 drivers/infiniband/hw/hfi1/tid_rdma.c 			if (next > rvt_size_atomic(&dev->rdi))
next             4357 drivers/infiniband/hw/hfi1/tid_rdma.c 				next = 0;
next             4358 drivers/infiniband/hw/hfi1/tid_rdma.c 			if (next == priv->r_tid_head)
next             4360 drivers/infiniband/hw/hfi1/tid_rdma.c 			e = &qp->s_ack_queue[next];
next             4364 drivers/infiniband/hw/hfi1/tid_rdma.c 		priv->r_tid_tail = next;
next             4932 drivers/infiniband/hw/hfi1/tid_rdma.c 				u32 next;
next             4937 drivers/infiniband/hw/hfi1/tid_rdma.c 				next = flow->flow_state.r_next_psn;
next             4938 drivers/infiniband/hw/hfi1/tid_rdma.c 				flow->npkts = delta_psn(lpsn, next - 1);
next             5180 drivers/infiniband/hw/hfi1/tid_rdma.c 	u32 hwords, next;
next             5253 drivers/infiniband/hw/hfi1/tid_rdma.c 		next = qpriv->r_tid_ack + 1;
next             5254 drivers/infiniband/hw/hfi1/tid_rdma.c 		if (next > rvt_size_atomic(&dev->rdi))
next             5255 drivers/infiniband/hw/hfi1/tid_rdma.c 			next = 0;
next             5256 drivers/infiniband/hw/hfi1/tid_rdma.c 		qpriv->r_tid_ack = next;
next             5257 drivers/infiniband/hw/hfi1/tid_rdma.c 		if (qp->s_ack_queue[next].opcode != TID_OP(WRITE_REQ))
next             5259 drivers/infiniband/hw/hfi1/tid_rdma.c 		nreq = ack_to_tid_req(&qp->s_ack_queue[next]);
next              318 drivers/infiniband/hw/hfi1/user_sdma.c 	static u8 next;
next              328 drivers/infiniband/hw/hfi1/user_sdma.c 		mapping[hash] = next;
next              329 drivers/infiniband/hw/hfi1/user_sdma.c 		next = (next + 1) & 0x7F;
next              122 drivers/infiniband/hw/hns/hns_roce_cmd.c 	cmd->free_head = context->next;
next              153 drivers/infiniband/hw/hns/hns_roce_cmd.c 	context->next = cmd->free_head;
next              242 drivers/infiniband/hw/hns/hns_roce_cmd.c 		hr_cmd->context[i].next = i + 1;
next              245 drivers/infiniband/hw/hns/hns_roce_cmd.c 	hr_cmd->context[hr_cmd->max_cmds - 1].next = -1;
next              597 drivers/infiniband/hw/hns/hns_roce_device.h 	int			next;
next              155 drivers/infiniband/hw/hns/hns_roce_hem.h 				 list_entry(hem->chunk_list.next,
next              168 drivers/infiniband/hw/hns/hns_roce_hem.h 		if (iter->chunk->list.next == &iter->hem->chunk_list) {
next              173 drivers/infiniband/hw/hns/hns_roce_hem.h 		iter->chunk = list_entry(iter->chunk->list.next,
next               92 drivers/infiniband/hw/hns/hns_roce_hw_v1.c 	for (nreq = 0; wr; ++nreq, wr = wr->next) {
next              365 drivers/infiniband/hw/hns/hns_roce_hw_v1.c 	for (nreq = 0; wr; ++nreq, wr = wr->next) {
next             1011 drivers/infiniband/hw/hns/hns_roce_hw_v1.c 	send_wr.next	= NULL;
next              279 drivers/infiniband/hw/hns/hns_roce_hw_v2.c 	for (nreq = 0; wr; ++nreq, wr = wr->next) {
next              643 drivers/infiniband/hw/hns/hns_roce_hw_v2.c 	for (nreq = 0; wr; ++nreq, wr = wr->next) {
next             6249 drivers/infiniband/hw/hns/hns_roce_hw_v2.c 	for (nreq = 0; wr; ++nreq, wr = wr->next) {
next              434 drivers/infiniband/hw/i40iw/i40iw.h 				       u32 *next)
next              440 drivers/infiniband/hw/i40iw/i40iw.h 	resource_num = find_next_zero_bit(resource_array, max_resources, *next);
next              449 drivers/infiniband/hw/i40iw/i40iw.h 	*next = resource_num + 1;
next              450 drivers/infiniband/hw/i40iw/i40iw.h 	if (*next == max_resources)
next              451 drivers/infiniband/hw/i40iw/i40iw.h 		*next = 0;
next              328 drivers/infiniband/hw/i40iw/i40iw_ctrl.c 		entry = head->next;
next              331 drivers/infiniband/hw/i40iw/i40iw_ctrl.c 		entry = (lastentry != head) ? lastentry->next : NULL;
next               60 drivers/infiniband/hw/i40iw/i40iw_puda.c 		buf = (struct i40iw_puda_buf *)list->next;
next              837 drivers/infiniband/hw/i40iw/i40iw_puda.c 		nextbuf = buf->next;
next              870 drivers/infiniband/hw/i40iw/i40iw_puda.c 			buf->next = nextbuf;
next             1417 drivers/infiniband/hw/i40iw/i40iw_puda.c 		tmpbuf = (struct i40iw_puda_buf *)rxlist->next;
next             1422 drivers/infiniband/hw/i40iw/i40iw_puda.c 			tmpbuf = (struct i40iw_puda_buf *)plist->next;
next               82 drivers/infiniband/hw/i40iw/i40iw_puda.h 	struct i40iw_puda_buf *next;	/* for alloclist in rsrc struct */
next              369 drivers/infiniband/hw/i40iw/i40iw_utils.c 		cqp_request = list_entry(cqp->cqp_avail_reqs.next,
next              739 drivers/infiniband/hw/i40iw/i40iw_utils.c 	entry = (void *)list->next;
next             2273 drivers/infiniband/hw/i40iw/i40iw_verbs.c 		ib_wr = ib_wr->next;
next             2329 drivers/infiniband/hw/i40iw/i40iw_verbs.c 		ib_wr = ib_wr->next;
next              811 drivers/infiniband/hw/mlx4/alias_GUID.c 			cb_ctx = list_entry(det->cb_list.next,
next              644 drivers/infiniband/hw/mlx4/mad.c 	wr.wr.next = NULL;
next             1326 drivers/infiniband/hw/mlx4/mad.c 	recv_wr.next = NULL;
next             1444 drivers/infiniband/hw/mlx4/mad.c 	wr.wr.next = NULL;
next             3583 drivers/infiniband/hw/mlx4/qp.c 	for (nreq = 0; wr; ++nreq, wr = wr->next) {
next             3821 drivers/infiniband/hw/mlx4/qp.c 		if (wr->next)
next             3881 drivers/infiniband/hw/mlx4/qp.c 	for (nreq = 0; wr; ++nreq, wr = wr->next) {
next             4485 drivers/infiniband/hw/mlx4/qp.c 			.next = NULL,
next               80 drivers/infiniband/hw/mlx4/srq.c 	struct mlx4_wqe_srq_next_seg *next;
next              147 drivers/infiniband/hw/mlx4/srq.c 			next = get_wqe(srq, i);
next              148 drivers/infiniband/hw/mlx4/srq.c 			next->next_wqe_index =
next              151 drivers/infiniband/hw/mlx4/srq.c 			for (scatter = (void *) (next + 1);
next              152 drivers/infiniband/hw/mlx4/srq.c 			     (void *) scatter < (void *) next + desc_size;
next              288 drivers/infiniband/hw/mlx4/srq.c 	struct mlx4_wqe_srq_next_seg *next;
next              293 drivers/infiniband/hw/mlx4/srq.c 	next = get_wqe(srq, srq->tail);
next              294 drivers/infiniband/hw/mlx4/srq.c 	next->next_wqe_index = cpu_to_be16(wqe_index);
next              304 drivers/infiniband/hw/mlx4/srq.c 	struct mlx4_wqe_srq_next_seg *next;
next              320 drivers/infiniband/hw/mlx4/srq.c 	for (nreq = 0; wr; ++nreq, wr = wr->next) {
next              335 drivers/infiniband/hw/mlx4/srq.c 		next      = get_wqe(srq, srq->head);
next              336 drivers/infiniband/hw/mlx4/srq.c 		srq->head = be16_to_cpu(next->next_wqe_index);
next              337 drivers/infiniband/hw/mlx4/srq.c 		scat      = (struct mlx4_wqe_data_seg *) (next + 1);
next              343 drivers/infiniband/hw/mlx5/cq.c 		tail = qp->sq.w_list[idx].next;
next              345 drivers/infiniband/hw/mlx5/cq.c 	tail = qp->sq.w_list[idx].next;
next              411 drivers/infiniband/hw/mlx5/cq.c 			wq->last_poll = wq->w_list[idx].next;
next              574 drivers/infiniband/hw/mlx5/cq.c 	struct mlx5_ib_wc *soft_wc, *next;
next              577 drivers/infiniband/hw/mlx5/cq.c 	list_for_each_entry_safe(soft_wc, next, &cq->wc_list, list) {
next             2476 drivers/infiniband/hw/mlx5/devx.c 	event = list_entry(ev_queue->event_list.next,
next             2486 drivers/infiniband/hw/mlx5/devx.c 	list_del(ev_queue->event_list.next);
next              488 drivers/infiniband/hw/mlx5/gsi.c 	for (; wr; wr = wr->next) {
next              491 drivers/infiniband/hw/mlx5/gsi.c 		cur_wr.wr.next = NULL;
next              259 drivers/infiniband/hw/mlx5/mlx5_ib.h 	u16	next;
next              698 drivers/infiniband/hw/mlx5/odp.c 		struct ib_umem_odp *next;
next              701 drivers/infiniband/hw/mlx5/odp.c 		next = odp_next(odp);
next              702 drivers/infiniband/hw/mlx5/odp.c 		if (unlikely(!next || ib_umem_start(next) != io_virt)) {
next              704 drivers/infiniband/hw/mlx5/odp.c 				    io_virt, next);
next              707 drivers/infiniband/hw/mlx5/odp.c 		odp = next;
next              732 drivers/infiniband/hw/mlx5/odp.c 	struct pf_frame *next;
next              896 drivers/infiniband/hw/mlx5/odp.c 			frame->next = head;
next              912 drivers/infiniband/hw/mlx5/odp.c 		head = frame->next;
next              926 drivers/infiniband/hw/mlx5/odp.c 		head = frame->next;
next             4950 drivers/infiniband/hw/mlx5/qp.c 	qp->sq.w_list[idx].next = qp->sq.cur_post;
next             5002 drivers/infiniband/hw/mlx5/qp.c 	for (nreq = 0; wr; nreq++, wr = wr->next) {
next             5382 drivers/infiniband/hw/mlx5/qp.c 	for (nreq = 0; wr; nreq++, wr = wr->next) {
next             6447 drivers/infiniband/hw/mlx5/qp.c 			.next = NULL,
next              136 drivers/infiniband/hw/mlx5/srq.c 	struct mlx5_wqe_srq_next_seg *next;
next              159 drivers/infiniband/hw/mlx5/srq.c 		next = get_wqe(srq, i);
next              160 drivers/infiniband/hw/mlx5/srq.c 		next->next_wqe_index =
next              408 drivers/infiniband/hw/mlx5/srq.c 	struct mlx5_wqe_srq_next_seg *next;
next              413 drivers/infiniband/hw/mlx5/srq.c 	next = get_wqe(srq, srq->tail);
next              414 drivers/infiniband/hw/mlx5/srq.c 	next->next_wqe_index = cpu_to_be16(wqe_index);
next              424 drivers/infiniband/hw/mlx5/srq.c 	struct mlx5_wqe_srq_next_seg *next;
next              441 drivers/infiniband/hw/mlx5/srq.c 	for (nreq = 0; wr; nreq++, wr = wr->next) {
next              456 drivers/infiniband/hw/mlx5/srq.c 		next      = get_wqe(srq, srq->head);
next              457 drivers/infiniband/hw/mlx5/srq.c 		srq->head = be16_to_cpu(next->next_wqe_index);
next              458 drivers/infiniband/hw/mlx5/srq.c 		scat      = (struct mlx5_wqe_data_seg *)(next + 1);
next              181 drivers/infiniband/hw/mthca/mthca_cmd.c 	int               next;
next              424 drivers/infiniband/hw/mthca/mthca_cmd.c 	dev->cmd.free_head = context->next;
next              460 drivers/infiniband/hw/mthca/mthca_cmd.c 	context->next = dev->cmd.free_head;
next              570 drivers/infiniband/hw/mthca/mthca_cmd.c 		dev->cmd.context[i].next = i + 1;
next              573 drivers/infiniband/hw/mthca/mthca_cmd.c 	dev->cmd.context[dev->cmd.max_cmds - 1].next = -1;
next             1728 drivers/infiniband/hw/mthca/mthca_cmd.c 		    enum ib_qp_state next, u32 num, int is_ee,
next             1775 drivers/infiniband/hw/mthca/mthca_cmd.c 	if (op[cur][next] == CMD_ERR2RST_QPEE) {
next             1790 drivers/infiniband/hw/mthca/mthca_cmd.c 				    op[cur][next], CMD_TIME_CLASS_C);
next             1824 drivers/infiniband/hw/mthca/mthca_cmd.c 				op_mod, op[cur][next], CMD_TIME_CLASS_C);
next              309 drivers/infiniband/hw/mthca/mthca_cmd.h 		    enum ib_qp_state next, u32 num, int is_ee,
next              104 drivers/infiniband/hw/mthca/mthca_memfree.h 		NULL : list_entry(icm->chunk_list.next,
next              117 drivers/infiniband/hw/mthca/mthca_memfree.h 		if (iter->chunk->list.next == &iter->icm->chunk_list) {
next              122 drivers/infiniband/hw/mthca/mthca_memfree.h 		iter->chunk = list_entry(iter->chunk->list.next,
next             1169 drivers/infiniband/hw/mthca/mthca_qp.c 	struct mthca_next_seg *next;
next             1217 drivers/infiniband/hw/mthca/mthca_qp.c 			next = get_recv_wqe(qp, i);
next             1218 drivers/infiniband/hw/mthca/mthca_qp.c 			next->nda_op = cpu_to_be32(((i + 1) & (qp->rq.max - 1)) <<
next             1220 drivers/infiniband/hw/mthca/mthca_qp.c 			next->ee_nds = cpu_to_be32(size);
next             1222 drivers/infiniband/hw/mthca/mthca_qp.c 			for (scatter = (void *) (next + 1);
next             1223 drivers/infiniband/hw/mthca/mthca_qp.c 			     (void *) scatter < (void *) next + (1 << qp->rq.wqe_shift);
next             1229 drivers/infiniband/hw/mthca/mthca_qp.c 			next = get_send_wqe(qp, i);
next             1230 drivers/infiniband/hw/mthca/mthca_qp.c 			next->nda_op = cpu_to_be32((((i + 1) & (qp->sq.max - 1)) <<
next             1236 drivers/infiniband/hw/mthca/mthca_qp.c 			next = get_recv_wqe(qp, i);
next             1237 drivers/infiniband/hw/mthca/mthca_qp.c 			next->nda_op = htonl((((i + 1) % qp->rq.max) <<
next             1653 drivers/infiniband/hw/mthca/mthca_qp.c 	for (nreq = 0; wr; ++nreq, wr = wr->next) {
next             1849 drivers/infiniband/hw/mthca/mthca_qp.c 	for (nreq = 0; wr; wr = wr->next) {
next             1957 drivers/infiniband/hw/mthca/mthca_qp.c 	for (nreq = 0; wr; ++nreq, wr = wr->next) {
next             2178 drivers/infiniband/hw/mthca/mthca_qp.c 	for (nreq = 0; wr; ++nreq, wr = wr->next) {
next             2234 drivers/infiniband/hw/mthca/mthca_qp.c 	struct mthca_next_seg *next;
next             2246 drivers/infiniband/hw/mthca/mthca_qp.c 		next = get_send_wqe(qp, index);
next             2248 drivers/infiniband/hw/mthca/mthca_qp.c 		next = get_recv_wqe(qp, index);
next             2250 drivers/infiniband/hw/mthca/mthca_qp.c 	*dbd = !!(next->ee_nds & cpu_to_be32(MTHCA_NEXT_DBD));
next             2251 drivers/infiniband/hw/mthca/mthca_qp.c 	if (next->ee_nds & cpu_to_be32(0x3f))
next             2252 drivers/infiniband/hw/mthca/mthca_qp.c 		*new_wqe = (next->nda_op & cpu_to_be32(~0x3f)) |
next             2253 drivers/infiniband/hw/mthca/mthca_qp.c 			(next->ee_nds & cpu_to_be32(0x3f));
next              183 drivers/infiniband/hw/mthca/mthca_srq.c 		struct mthca_next_seg *next;
next              185 drivers/infiniband/hw/mthca/mthca_srq.c 		next = wqe = get_wqe(srq, i);
next              189 drivers/infiniband/hw/mthca/mthca_srq.c 			next->nda_op = htonl(((i + 1) << srq->wqe_shift) | 1);
next              192 drivers/infiniband/hw/mthca/mthca_srq.c 			next->nda_op = 0;
next              502 drivers/infiniband/hw/mthca/mthca_srq.c 	for (nreq = 0; wr; wr = wr->next) {
next              592 drivers/infiniband/hw/mthca/mthca_srq.c 	for (nreq = 0; wr; ++nreq, wr = wr->next) {
next             2200 drivers/infiniband/hw/ocrdma/ocrdma_verbs.c 		wr = wr->next;
next             2267 drivers/infiniband/hw/ocrdma/ocrdma_verbs.c 		wr = wr->next;
next             2334 drivers/infiniband/hw/ocrdma/ocrdma_verbs.c 		wr = wr->next;
next              599 drivers/infiniband/hw/qedr/qedr_roce_cm.c 	if (wr->next) {
next              602 drivers/infiniband/hw/qedr/qedr_roce_cm.c 		*bad_wr = wr->next;
next              660 drivers/infiniband/hw/qedr/qedr_roce_cm.c 		wr = wr->next;
next             3431 drivers/infiniband/hw/qedr/verbs.c 		wr = wr->next;
next             3541 drivers/infiniband/hw/qedr/verbs.c 		wr = wr->next;
next             3644 drivers/infiniband/hw/qedr/verbs.c 		wr = wr->next;
next               46 drivers/infiniband/hw/qib/qib_debugfs.c 	.next  = _##name##_seq_next, \
next               68 drivers/infiniband/hw/qib/qib_diag.c 	struct qib_diag_client *next;
next               85 drivers/infiniband/hw/qib/qib_diag.c 		client_pool = dc->next;
next               91 drivers/infiniband/hw/qib/qib_diag.c 		dc->next = NULL;
next              109 drivers/infiniband/hw/qib/qib_diag.c 		dd->diag_client = dc->next;
next              114 drivers/infiniband/hw/qib/qib_diag.c 			if (dc == tdc->next) {
next              115 drivers/infiniband/hw/qib/qib_diag.c 				tdc->next = dc->next;
next              119 drivers/infiniband/hw/qib/qib_diag.c 			tdc = tdc->next;
next              126 drivers/infiniband/hw/qib/qib_diag.c 		rdc->next = client_pool;
next              202 drivers/infiniband/hw/qib/qib_diag.c 		client_pool = dc->next;
next              530 drivers/infiniband/hw/qib/qib_diag.c 	dc->next = dd->diag_client;
next              690 drivers/infiniband/hw/qib/qib_diag.c 	struct diag_observer_list_elt *next;
next              708 drivers/infiniband/hw/qib/qib_diag.c 	olp->next = dd->diag_observer_list;
next              725 drivers/infiniband/hw/qib/qib_diag.c 		dd->diag_observer_list = olp->next;
next              751 drivers/infiniband/hw/qib/qib_diag.c 		olp = olp->next;
next             1189 drivers/infiniband/hw/qib/qib_init.c 	.next           = NULL,
next              253 drivers/infiniband/hw/qib/qib_qsfp.c static int qsfp_cks(struct qib_pportdata *ppd, int first, int next)
next              260 drivers/infiniband/hw/qib/qib_qsfp.c 	while (first < next) {
next             1689 drivers/infiniband/hw/qib/qib_rc.c 	unsigned next;
next             1691 drivers/infiniband/hw/qib/qib_rc.c 	next = n + 1;
next             1692 drivers/infiniband/hw/qib/qib_rc.c 	if (next > QIB_MAX_RDMA_ATOMIC)
next             1693 drivers/infiniband/hw/qib/qib_rc.c 		next = 0;
next             1694 drivers/infiniband/hw/qib/qib_rc.c 	qp->s_tail_ack_queue = next;
next             1939 drivers/infiniband/hw/qib/qib_rc.c 		u8 next;
next             1943 drivers/infiniband/hw/qib/qib_rc.c 		next = qp->r_head_ack_queue + 1;
next             1945 drivers/infiniband/hw/qib/qib_rc.c 		if (next > QIB_MAX_RDMA_ATOMIC)
next             1946 drivers/infiniband/hw/qib/qib_rc.c 			next = 0;
next             1948 drivers/infiniband/hw/qib/qib_rc.c 		if (unlikely(next == qp->s_tail_ack_queue)) {
next             1949 drivers/infiniband/hw/qib/qib_rc.c 			if (!qp->s_ack_queue[next].sent)
next             1951 drivers/infiniband/hw/qib/qib_rc.c 			qib_update_ack_queue(qp, next);
next             1994 drivers/infiniband/hw/qib/qib_rc.c 		qp->r_head_ack_queue = next;
next             2011 drivers/infiniband/hw/qib/qib_rc.c 		u8 next;
next             2015 drivers/infiniband/hw/qib/qib_rc.c 		next = qp->r_head_ack_queue + 1;
next             2016 drivers/infiniband/hw/qib/qib_rc.c 		if (next > QIB_MAX_RDMA_ATOMIC)
next             2017 drivers/infiniband/hw/qib/qib_rc.c 			next = 0;
next             2019 drivers/infiniband/hw/qib/qib_rc.c 		if (unlikely(next == qp->s_tail_ack_queue)) {
next             2020 drivers/infiniband/hw/qib/qib_rc.c 			if (!qp->s_ack_queue[next].sent)
next             2022 drivers/infiniband/hw/qib/qib_rc.c 			qib_update_ack_queue(qp, next);
next             2057 drivers/infiniband/hw/qib/qib_rc.c 		qp->r_head_ack_queue = next;
next              342 drivers/infiniband/hw/qib/qib_sdma.c 		lp = ppd->sdma_activelist.next;
next              373 drivers/infiniband/hw/qib/qib_sdma.c 				lp = ppd->sdma_activelist.next;
next             1253 drivers/infiniband/hw/qib/qib_user_sdma.c 			list_entry(pktlist->next, struct qib_user_sdma_pkt,
next              373 drivers/infiniband/hw/qib/qib_verbs.c 		priv = list_entry(list->next, struct qib_qp_priv, iowait);
next              571 drivers/infiniband/hw/qib/qib_verbs.c 		struct list_head *l = dev->txreq_free.next;
next              601 drivers/infiniband/hw/qib/qib_verbs.c 		struct list_head *l = dev->txreq_free.next;
next              643 drivers/infiniband/hw/qib/qib_verbs.c 		priv = list_entry(dev->txwait.next, struct qib_qp_priv,
next             1167 drivers/infiniband/hw/qib/qib_verbs.c 		priv = list_entry(list->next, struct qib_qp_priv, iowait);
next             1629 drivers/infiniband/hw/qib/qib_verbs.c 		struct list_head *l = dev->txreq_free.next;
next             1665 drivers/infiniband/hw/qib/qib_verbs.c 		struct list_head *l = dev->txreq_free.next;
next              806 drivers/infiniband/hw/vmw_pvrdma/pvrdma_qp.c 		wr = wr->next;
next              897 drivers/infiniband/hw/vmw_pvrdma/pvrdma_qp.c 		wr = wr->next;
next               75 drivers/infiniband/sw/rdmavt/cq.c 	u32 next;
next               98 drivers/infiniband/sw/rdmavt/cq.c 		next = 0;
next              100 drivers/infiniband/sw/rdmavt/cq.c 		next = head + 1;
next              103 drivers/infiniband/sw/rdmavt/cq.c 	if (unlikely(next == tail || cq->cq_full)) {
next              137 drivers/infiniband/sw/rdmavt/cq.c 		RDMA_WRITE_UAPI_ATOMIC(u_wc->head, next);
next              140 drivers/infiniband/sw/rdmavt/cq.c 		k_wc->head = next;
next              205 drivers/infiniband/sw/rdmavt/mr.c 	r = rkt->next;
next              214 drivers/infiniband/sw/rdmavt/mr.c 	rkt->next = (r + 1) & (rkt->max - 1);
next              791 drivers/infiniband/sw/rdmavt/qp.c 			qpp = &q->next) {
next              794 drivers/infiniband/sw/rdmavt/qp.c 				     rcu_dereference_protected(qp->next,
next             1134 drivers/infiniband/sw/rdmavt/qp.c 		RCU_INIT_POINTER(qp->next, NULL);
next             1429 drivers/infiniband/sw/rdmavt/qp.c 		qp->next = rdi->qp_dev->qp_table[n];
next             1836 drivers/infiniband/sw/rdmavt/qp.c 	for (; wr; wr = wr->next) {
next             1838 drivers/infiniband/sw/rdmavt/qp.c 		u32 next;
next             1847 drivers/infiniband/sw/rdmavt/qp.c 		next = wq->head + 1;
next             1848 drivers/infiniband/sw/rdmavt/qp.c 		if (next >= qp->r_rq.size)
next             1849 drivers/infiniband/sw/rdmavt/qp.c 			next = 0;
next             1850 drivers/infiniband/sw/rdmavt/qp.c 		if (next == READ_ONCE(wq->tail)) {
next             1877 drivers/infiniband/sw/rdmavt/qp.c 			smp_store_release(&wq->head, next);
next             2004 drivers/infiniband/sw/rdmavt/qp.c 	u32 next;
next             2071 drivers/infiniband/sw/rdmavt/qp.c 	next = qp->s_head + 1;
next             2072 drivers/infiniband/sw/rdmavt/qp.c 	if (next >= qp->s_size)
next             2073 drivers/infiniband/sw/rdmavt/qp.c 		next = 0;
next             2155 drivers/infiniband/sw/rdmavt/qp.c 	qp->s_head = next;
next             2208 drivers/infiniband/sw/rdmavt/qp.c 	call_send = qp->s_head == READ_ONCE(qp->s_last) && !wr->next;
next             2210 drivers/infiniband/sw/rdmavt/qp.c 	for (; wr; wr = wr->next) {
next             2250 drivers/infiniband/sw/rdmavt/qp.c 	for (; wr; wr = wr->next) {
next             2252 drivers/infiniband/sw/rdmavt/qp.c 		u32 next;
next             2262 drivers/infiniband/sw/rdmavt/qp.c 		next = wq->head + 1;
next             2263 drivers/infiniband/sw/rdmavt/qp.c 		if (next >= srq->rq.size)
next             2264 drivers/infiniband/sw/rdmavt/qp.c 			next = 0;
next             2265 drivers/infiniband/sw/rdmavt/qp.c 		if (next == READ_ONCE(wq->tail)) {
next             2280 drivers/infiniband/sw/rdmavt/qp.c 		smp_store_release(&wq->head, next);
next             2768 drivers/infiniband/sw/rdmavt/qp.c 			qp = rcu_dereference(pqp->next);
next               71 drivers/infiniband/sw/rxe/rxe.h 			    u32 crc, void *next, size_t len)
next               80 drivers/infiniband/sw/rxe/rxe.h 	err = crypto_shash_update(shash, next, len);
next               83 drivers/infiniband/sw/rxe/rxe.h 		return crc32_le(crc, next, len);
next              315 drivers/infiniband/sw/rxe/rxe_recv.c 		if (mce->qp_list.next != &mcg->qp_list)
next              400 drivers/infiniband/sw/rxe/rxe_verbs.c 		wr = wr->next;
next              708 drivers/infiniband/sw/rxe/rxe_verbs.c 		wr = wr->next;
next              769 drivers/infiniband/sw/rxe/rxe_verbs.c 		wr = wr->next;
next              278 drivers/infiniband/sw/siw/siw_cm.c 		work = list_entry(cep->work_freelist.next, struct siw_cm_work,
next               24 drivers/infiniband/sw/siw/siw_mem.c 	u32 id, next;
next               26 drivers/infiniband/sw/siw/siw_mem.c 	get_random_bytes(&next, 4);
next               27 drivers/infiniband/sw/siw/siw_mem.c 	next &= 0x00ffffff;
next               29 drivers/infiniband/sw/siw/siw_mem.c 	if (xa_alloc_cyclic(&sdev->mem_xa, &id, m, limit, &next,
next               95 drivers/infiniband/sw/siw/siw_mem.c 	u32 id, next;
next              111 drivers/infiniband/sw/siw/siw_mem.c 	get_random_bytes(&next, 4);
next              112 drivers/infiniband/sw/siw/siw_mem.c 	next &= 0x00ffffff;
next              114 drivers/infiniband/sw/siw/siw_mem.c 	if (xa_alloc_cyclic(&sdev->mem_xa, &id, mem, limit, &next,
next             1227 drivers/infiniband/sw/siw/siw_qp_tx.c 			tmp->next = fifo_list;
next             1233 drivers/infiniband/sw/siw/siw_qp_tx.c 			qp->tx_list.next = NULL;
next             1241 drivers/infiniband/sw/siw/siw_qp_tx.c 			qp->tx_list.next = NULL;
next              704 drivers/infiniband/sw/siw/siw_verbs.c 		wr = wr->next;
next              724 drivers/infiniband/sw/siw/siw_verbs.c 		wr = wr->next;
next              924 drivers/infiniband/sw/siw/siw_verbs.c 		wr = wr->next;
next             1068 drivers/infiniband/sw/siw/siw_verbs.c 		wr = wr->next;
next             1789 drivers/infiniband/sw/siw/siw_verbs.c 		wr = wr->next;
next              225 drivers/infiniband/ulp/ipoib/ipoib_cm.c 	p = list_entry(priv->cm.rx_flush_list.next, typeof(*p), list);
next              342 drivers/infiniband/ulp/ipoib/ipoib_cm.c 	wr->next    = NULL;
next              942 drivers/infiniband/ulp/ipoib/ipoib_cm.c 		p = list_entry(priv->cm.passive_ids.next, typeof(*p), list);
next             1359 drivers/infiniband/ulp/ipoib/ipoib_cm.c 		p = list_entry(priv->cm.start_list.next, typeof(*p), list);
next             1413 drivers/infiniband/ulp/ipoib/ipoib_cm.c 		p = list_entry(priv->cm.reap_list.next, typeof(*p), list);
next              129 drivers/infiniband/ulp/ipoib/ipoib_fs.c 	.next  = ipoib_mcg_seq_next,
next              234 drivers/infiniband/ulp/ipoib/ipoib_fs.c 	.next  = ipoib_path_seq_next,
next              230 drivers/infiniband/ulp/ipoib/ipoib_verbs.c 	priv->rx_wr.next = NULL;
next              208 drivers/infiniband/ulp/iser/iscsi_iser.c 	tx_desc->inv_wr.next = NULL;
next              209 drivers/infiniband/ulp/iser/iscsi_iser.c 	tx_desc->reg_wr.wr.next = NULL;
next              376 drivers/infiniband/ulp/iser/iser_memory.c 	inv_wr->next = next_wr;
next              414 drivers/infiniband/ulp/iser/iser_memory.c 	wr->wr.next = &tx_desc->send_wr;
next              461 drivers/infiniband/ulp/iser/iser_memory.c 	wr->wr.next = &tx_desc->send_wr;
next              987 drivers/infiniband/ulp/iser/iser_verbs.c 	wr.next = NULL;
next             1013 drivers/infiniband/ulp/iser/iser_verbs.c 		wr->next = wr + 1;
next             1018 drivers/infiniband/ulp/iser/iser_verbs.c 	wr->next = NULL; /* mark end of work requests list */
next             1048 drivers/infiniband/ulp/iser/iser_verbs.c 	wr->next = NULL;
next             1055 drivers/infiniband/ulp/iser/iser_verbs.c 	if (tx_desc->inv_wr.next)
next             1057 drivers/infiniband/ulp/iser/iser_verbs.c 	else if (tx_desc->reg_wr.wr.next)
next              815 drivers/infiniband/ulp/isert/ib_isert.c 		rx_wr->next = rx_wr + 1;
next              819 drivers/infiniband/ulp/isert/ib_isert.c 	rx_wr->next = NULL; /* mark end of work requests list */
next              846 drivers/infiniband/ulp/isert/ib_isert.c 	rx_wr.next = NULL;
next              867 drivers/infiniband/ulp/isert/ib_isert.c 	send_wr.next	= NULL;
next             1231 drivers/infiniband/ulp/srp/ib_srp.c 		.next		    = NULL,
next             1473 drivers/infiniband/ulp/srp/ib_srp.c 	if (state->fmr.next >= state->fmr.end) {
next             1496 drivers/infiniband/ulp/srp/ib_srp.c 	*state->fmr.next++ = fmr;
next             1532 drivers/infiniband/ulp/srp/ib_srp.c 	if (state->fr.next >= state->fr.end) {
next             1573 drivers/infiniband/ulp/srp/ib_srp.c 	wr.wr.next = NULL;
next             1584 drivers/infiniband/ulp/srp/ib_srp.c 	*state->fr.next++ = desc;
next             1651 drivers/infiniband/ulp/srp/ib_srp.c 	state->fmr.next = req->fmr_list;
next             1673 drivers/infiniband/ulp/srp/ib_srp.c 	state->fr.next = req->fr_list;
next             1732 drivers/infiniband/ulp/srp/ib_srp.c 	state.gen.next = next_mr;
next             1951 drivers/infiniband/ulp/srp/ib_srp.c 		ret = srp_map_idb(ch, req, state.gen.next, state.gen.end,
next             2085 drivers/infiniband/ulp/srp/ib_srp.c 	wr.next       = NULL;
next             2107 drivers/infiniband/ulp/srp/ib_srp.c 	wr.next     = NULL;
next              328 drivers/infiniband/ulp/srp/ib_srp.h 			struct ib_pool_fmr **next;
next              332 drivers/infiniband/ulp/srp/ib_srp.h 			struct srp_fr_desc **next;
next              336 drivers/infiniband/ulp/srp/ib_srp.h 			void		   **next;
next              834 drivers/infiniband/ulp/srpt/ib_srpt.c 	wr.next = NULL;
next              857 drivers/infiniband/ulp/srpt/ib_srpt.c 			.next		= NULL,
next             2863 drivers/infiniband/ulp/srpt/ib_srpt.c 	send_wr.next = NULL;
next              294 drivers/input/gameport/gameport.c 	struct gameport_event *e, *next;
next              299 drivers/input/gameport/gameport.c 	list_for_each_entry_safe(e, next, &gameport_event_list, node) {
next              409 drivers/input/gameport/gameport.c 	struct gameport_event *event, *next;
next              414 drivers/input/gameport/gameport.c 	list_for_each_entry_safe(event, next, &gameport_event_list, node) {
next             1210 drivers/input/input.c 	.next	= input_devices_seq_next,
next             1274 drivers/input/input.c 	.next	= input_handlers_seq_next,
next             2086 drivers/input/input.c 	struct input_handle *handle, *next;
next             2092 drivers/input/input.c 	list_for_each_entry_safe(handle, next, &dev->h_list, d_node)
next             2321 drivers/input/input.c 	struct input_handle *handle, *next;
next             2325 drivers/input/input.c 	list_for_each_entry_safe(handle, next, &handler->h_list, h_node)
next              926 drivers/input/serio/i8042-x86ia64io.h 		id = id->next;
next              176 drivers/input/serio/serio.c 	struct serio_event *e, *next;
next              181 drivers/input/serio/serio.c 	list_for_each_entry_safe(e, next, &serio_event_list, node) {
next              297 drivers/input/serio/serio.c 	struct serio_event *event, *next;
next              302 drivers/input/serio/serio.c 	list_for_each_entry_safe(event, next, &serio_event_list, node) {
next              630 drivers/input/serio/serio.c 				s = list_entry(s->child_node.next,
next              719 drivers/input/serio/serio.c 	struct serio *s, *next;
next              722 drivers/input/serio/serio.c 	list_for_each_entry_safe(s, next, &serio->children, child_node) {
next              769 drivers/input/touchscreen/cyttsp4_core.c 	int next;
next              771 drivers/input/touchscreen/cyttsp4_core.c 	for (nbyte = 0, *axis = 0, next = 0; nbyte < size; nbyte++) {
next              775 drivers/input/touchscreen/cyttsp4_core.c 			__func__, *axis, *axis, size, max, xy_data, next,
next              776 drivers/input/touchscreen/cyttsp4_core.c 			xy_data[next], xy_data[next], bofs);
next              777 drivers/input/touchscreen/cyttsp4_core.c 		*axis = (*axis * 256) + (xy_data[next] >> bofs);
next              778 drivers/input/touchscreen/cyttsp4_core.c 		next++;
next              786 drivers/input/touchscreen/cyttsp4_core.c 		__func__, *axis, *axis, size, max, xy_data, next,
next              787 drivers/input/touchscreen/cyttsp4_core.c 		xy_data[next], xy_data[next]);
next              546 drivers/input/touchscreen/sur40.c 	new_buf = list_entry(sur40->buf_list.next, struct sur40_buffer, list);
next              220 drivers/interconnect/core.c 	struct icc_node *next, *prev = NULL;
next              225 drivers/interconnect/core.c 		next = path->reqs[i].node;
next              231 drivers/interconnect/core.c 		if (!prev || next->provider != prev->provider) {
next              232 drivers/interconnect/core.c 			prev = next;
next              237 drivers/interconnect/core.c 		ret = next->provider->set(prev, next);
next              241 drivers/interconnect/core.c 		prev = next;
next             3231 drivers/iommu/amd_iommu.c 	struct iommu_resv_region *entry, *next;
next             3233 drivers/iommu/amd_iommu.c 	list_for_each_entry_safe(entry, next, head, list)
next             1418 drivers/iommu/amd_iommu_init.c 	struct amd_iommu *iommu, *next;
next             1420 drivers/iommu/amd_iommu_init.c 	for_each_iommu_safe(iommu, next) {
next             2070 drivers/iommu/amd_iommu_init.c 	struct unity_map_entry *entry, *next;
next             2072 drivers/iommu/amd_iommu_init.c 	list_for_each_entry_safe(entry, next, &amd_iommu_unity_map, list) {
next              431 drivers/iommu/amd_iommu_types.h #define for_each_iommu_safe(iommu, next) \
next              432 drivers/iommu/amd_iommu_types.h 	list_for_each_entry_safe((iommu), (next), &amd_iommu_list, list)
next             2718 drivers/iommu/arm-smmu-v3.c 	struct iommu_resv_region *entry, *next;
next             2720 drivers/iommu/arm-smmu-v3.c 	list_for_each_entry_safe(entry, next, head, list)
next             1549 drivers/iommu/arm-smmu.c 	struct iommu_resv_region *entry, *next;
next             1551 drivers/iommu/arm-smmu.c 	list_for_each_entry_safe(entry, next, head, list)
next              227 drivers/iommu/dma-iommu.c 		if (window->node.next == &bridge->dma_ranges &&
next              559 drivers/iommu/dmar.c 	struct acpi_dmar_header *iter, *next;
next              562 drivers/iommu/dmar.c 	for (iter = start; iter < end; iter = next) {
next              563 drivers/iommu/dmar.c 		next = (void *)iter + iter->length;
next              568 drivers/iommu/dmar.c 		} else if (next > end) {
next              797 drivers/iommu/exynos-iommu.c 	struct sysmmu_drvdata *data, *next;
next              805 drivers/iommu/exynos-iommu.c 	list_for_each_entry_safe(data, next, &domain->clients, domain_node) {
next              843 drivers/iommu/exynos-iommu.c 	struct sysmmu_drvdata *data, *next;
next              859 drivers/iommu/exynos-iommu.c 	list_for_each_entry_safe(data, next, &domain->clients, domain_node) {
next             1012 drivers/iommu/intel-iommu.c 			goto next;
next             1033 drivers/iommu/intel-iommu.c next:
next             1109 drivers/iommu/intel-iommu.c 			goto next;
next             1132 drivers/iommu/intel-iommu.c next:
next             5726 drivers/iommu/intel-iommu.c 	struct iommu_resv_region *entry, *next;
next             5728 drivers/iommu/intel-iommu.c 	list_for_each_entry_safe(entry, next, head, list)
next              384 drivers/iommu/iommu.c 	struct iommu_resv_region *region, *next;
next              391 drivers/iommu/iommu.c 	list_for_each_entry_safe(region, next, &group_resv_regions, list) {
next              733 drivers/iommu/iova.c 	struct iova *prev = NULL, *next = NULL;
next              742 drivers/iommu/iova.c 		next = alloc_and_init_iova(pfn_hi + 1, iova->pfn_hi);
next              743 drivers/iommu/iova.c 		if (next == NULL)
next              754 drivers/iommu/iova.c 	if (next) {
next              755 drivers/iommu/iova.c 		iova_insert_rbtree(&iovad->rbroot, next, NULL);
next              353 drivers/iommu/virtio-iommu.c 	struct interval_tree_node *node, *next;
next              356 drivers/iommu/virtio-iommu.c 	next = interval_tree_iter_first(&vdomain->mappings, iova, last);
next              357 drivers/iommu/virtio-iommu.c 	while (next) {
next              358 drivers/iommu/virtio-iommu.c 		node = next;
next              360 drivers/iommu/virtio-iommu.c 		next = interval_tree_iter_next(node, iova, last);
next              845 drivers/iommu/virtio-iommu.c 	struct iommu_resv_region *entry, *next;
next              847 drivers/iommu/virtio-iommu.c 	list_for_each_entry_safe(entry, next, head, list)
next               13 drivers/isdn/capi/capilib.c 	struct capilib_msgidqueue *next;
next               41 drivers/isdn/capi/capilib.c 		np->msgidpool[i].next = np->msgidfree;
next               51 drivers/isdn/capi/capilib.c 	np->msgidfree = mq->next;
next               53 drivers/isdn/capi/capilib.c 	mq->next = NULL;
next               55 drivers/isdn/capi/capilib.c 		np->msgidlast->next = mq;
next               66 drivers/isdn/capi/capilib.c 	for (pp = &np->msgidqueue; *pp; pp = &(*pp)->next) {
next               69 drivers/isdn/capi/capilib.c 			*pp = mq->next;
next               72 drivers/isdn/capi/capilib.c 			mq->next = np->msgidfree;
next               99 drivers/isdn/capi/kcapi_proc.c 	.next	= controller_next,
next              106 drivers/isdn/capi/kcapi_proc.c 	.next	= controller_next,
next              181 drivers/isdn/capi/kcapi_proc.c 	.next	= applications_next,
next              188 drivers/isdn/capi/kcapi_proc.c 	.next	= applications_next,
next              223 drivers/isdn/capi/kcapi_proc.c 	.next	= capi_driver_next,
next              914 drivers/isdn/hardware/mISDN/hfcmulti.c 	struct hfc_multi *hc, *next, *pcmmaster = NULL;
next              931 drivers/isdn/hardware/mISDN/hfcmulti.c 		list_for_each_entry_safe(hc, next, &HFClist, list) {
next              942 drivers/isdn/hardware/mISDN/hfcmulti.c 	list_for_each_entry_safe(hc, next, &HFClist, list) {
next             1117 drivers/isdn/hardware/mISDN/hfcmulti.c 	struct hfc_multi	*pos, *next, *plx_last_hc;
next             1219 drivers/isdn/hardware/mISDN/hfcmulti.c 		list_for_each_entry_safe(pos, next, &HFClist, list) {
next             5464 drivers/isdn/hardware/mISDN/hfcmulti.c 	struct hfc_multi *card, *next;
next             5467 drivers/isdn/hardware/mISDN/hfcmulti.c 	list_for_each_entry_safe(card, next, &HFClist, list)
next             2122 drivers/isdn/hardware/mISDN/hfcsusb.c 	struct hfcsusb *next;
next             2130 drivers/isdn/hardware/mISDN/hfcsusb.c 	list_for_each_entry_safe(hw, next, &HFClist, list)
next              513 drivers/isdn/mISDN/dsp_cmx.c 	member = list_entry(conf->mlist.next, struct dsp_conf_member, list);
next              672 drivers/isdn/mISDN/dsp_cmx.c 		member = list_entry(conf->mlist.next, struct dsp_conf_member,
next              686 drivers/isdn/mISDN/dsp_cmx.c 		member = list_entry(conf->mlist.next, struct dsp_conf_member,
next              688 drivers/isdn/mISDN/dsp_cmx.c 		nextm = list_entry(member->list.next, struct dsp_conf_member,
next             1121 drivers/isdn/mISDN/dsp_cmx.c 		member = list_entry(conf->mlist.next, struct dsp_conf_member,
next             1460 drivers/isdn/mISDN/dsp_cmx.c 		other = (list_entry(conf->mlist.next,
next             1705 drivers/isdn/mISDN/dsp_cmx.c 			member = list_entry(conf->mlist.next,
next             1272 drivers/isdn/mISDN/l1oip_core.c 	struct l1oip *hc, *next;
next             1274 drivers/isdn/mISDN/l1oip_core.c 	list_for_each_entry_safe(hc, next, &l1oip_ilist, list)
next              283 drivers/isdn/mISDN/socket.c 	struct mISDNchannel	*bchan, *next;
next              297 drivers/isdn/mISDN/socket.c 			list_for_each_entry_safe(bchan, next,
next               68 drivers/isdn/mISDN/timerdev.c 	struct mISDNtimer	*timer, *next;
next               85 drivers/isdn/mISDN/timerdev.c 	list_for_each_entry_safe(timer, next, &dev->expired, list) {
next               28 drivers/leds/trigger/ledtrig-pattern.c 	struct led_pattern *next;
next               41 drivers/leds/trigger/ledtrig-pattern.c 	data->curr = data->next;
next               45 drivers/leds/trigger/ledtrig-pattern.c 	if (data->next == data->patterns + data->npatterns - 1)
next               46 drivers/leds/trigger/ledtrig-pattern.c 		data->next = data->patterns;
next               48 drivers/leds/trigger/ledtrig-pattern.c 		data->next++;
next               65 drivers/leds/trigger/ledtrig-pattern.c 	step_brightness = abs(data->next->brightness - data->curr->brightness);
next               68 drivers/leds/trigger/ledtrig-pattern.c 	if (data->next->brightness > data->curr->brightness)
next               82 drivers/leds/trigger/ledtrig-pattern.c 		if (data->curr->brightness == data->next->brightness) {
next               88 drivers/leds/trigger/ledtrig-pattern.c 			if (!data->next->delta_t) {
next              138 drivers/leds/trigger/ledtrig-pattern.c 	data->next = data->patterns + 1;
next              765 drivers/lightnvm/pblk-recovery.c 			goto next;
next              770 drivers/lightnvm/pblk-recovery.c 			goto next;
next              775 drivers/lightnvm/pblk-recovery.c 			goto next;
next              786 drivers/lightnvm/pblk-recovery.c next:
next               68 drivers/macintosh/adb-iop.c 	current_req = req->next;
next              245 drivers/macintosh/adb-iop.c 	req->next = NULL;
next              251 drivers/macintosh/adb-iop.c 		last_req->next = req;
next              635 drivers/macintosh/adb.c 			ap = &(*ap)->next;
next              636 drivers/macintosh/adb.c 		req->next = NULL;
next              736 drivers/macintosh/adb.c 			state->completed = req->next;
next              178 drivers/macintosh/macio-adb.c 	req->next = NULL;
next              185 drivers/macintosh/macio-adb.c 		last_req->next = req;
next              224 drivers/macintosh/macio-adb.c 				current_req = req->next;
next              244 drivers/macintosh/macio-adb.c 			current_req = req->next;
next              119 drivers/macintosh/smu.c 	cmd = list_entry(smu->cmd_list.next, struct smu_cmd, link);
next              738 drivers/macintosh/smu.c 		newcmd = list_entry(smu->cmd_i2c_list.next,
next              988 drivers/macintosh/smu.c 	prop->next = NULL;
next              514 drivers/macintosh/via-cuda.c     req->next = NULL;
next              521 drivers/macintosh/via-cuda.c 	last_req->next = req;
next              645 drivers/macintosh/via-cuda.c 		current_req = req->next;
next              699 drivers/macintosh/via-cuda.c 	    current_req = req->next;
next              208 drivers/macintosh/via-macii.c 	req.next = current_req;
next              244 drivers/macintosh/via-macii.c 	req->next = NULL;
next              252 drivers/macintosh/via-macii.c 		last_req->next = req;
next              426 drivers/macintosh/via-macii.c 				current_req = req->next;
next              505 drivers/macintosh/via-macii.c 			current_req = req->next;
next             1174 drivers/macintosh/via-pmu.c 	req->next = NULL;
next             1180 drivers/macintosh/via-pmu.c 		last_req->next = req;
next             1384 drivers/macintosh/via-pmu.c next:
next             1476 drivers/macintosh/via-pmu.c 	goto next;
next             1517 drivers/macintosh/via-pmu.c 			current_req = req->next;
next             1567 drivers/macintosh/via-pmu.c 			current_req = req->next;
next             2138 drivers/macintosh/via-pmu.c 	for (list = &all_pmu_pvt; (list = list->next) != &all_pmu_pvt; ) {
next               89 drivers/macintosh/windfarm_core.c 	unsigned long next, delay;
next               91 drivers/macintosh/windfarm_core.c 	next = jiffies;
next               99 drivers/macintosh/windfarm_core.c 		if (time_after_eq(jiffies, next)) {
next              114 drivers/macintosh/windfarm_core.c 			next += HZ;
next              117 drivers/macintosh/windfarm_core.c 		delay = next - jiffies;
next              314 drivers/macintosh/windfarm_smu_controls.c 		fct = list_entry(smu_fans.next, struct smu_fan_control, link);
next              467 drivers/macintosh/windfarm_smu_sensors.c 		ads = list_entry(smu_ads.next, struct smu_ad_sensor, link);
next              989 drivers/mailbox/bcm-flexrm-mailbox.c 	void *next;
next             1052 drivers/mailbox/bcm-flexrm-mailbox.c 	next = flexrm_write_descs(msg, nhcnt, reqid,
next             1056 drivers/mailbox/bcm-flexrm-mailbox.c 	if (IS_ERR(next)) {
next             1057 drivers/mailbox/bcm-flexrm-mailbox.c 		ret = PTR_ERR(next);
next             1063 drivers/mailbox/bcm-flexrm-mailbox.c 	ring->bd_write_offset = (unsigned long)(next - ring->bd_base);
next               90 drivers/md/bcache/alloc.c 	unsigned int next = c->nbuckets * c->sb.bucket_size / 1024;
next              101 drivers/md/bcache/alloc.c 	} while (atomic_cmpxchg(&c->rescale, r, r + next) != r);
next               23 drivers/md/bcache/bset.c 	struct bkey *k, *next;
next               25 drivers/md/bcache/bset.c 	for (k = i->start; k < bset_bkey_last(i); k = next) {
next               26 drivers/md/bcache/bset.c 		next = bkey_next(k);
next               36 drivers/md/bcache/bset.c 		if (next < bset_bkey_last(i) &&
next               38 drivers/md/bcache/bset.c 			     &START_KEY(next) : next) > 0)
next              113 drivers/md/bcache/bset.c 	struct bkey *k = iter->data->k, *next = bkey_next(k);
next              115 drivers/md/bcache/bset.c 	if (next < iter->data->end &&
next              117 drivers/md/bcache/bset.c 		     &START_KEY(next) : next) > 0) {
next              787 drivers/md/bcache/btree.c 	if (c->shrink.list.next)
next              678 drivers/md/bcache/journal.c 		unsigned int next = (ja->cur_idx + 1) % ca->sb.njournal_buckets;
next              681 drivers/md/bcache/journal.c 		if (next == ja->discard_idx)
next              684 drivers/md/bcache/journal.c 		ja->cur_idx = next;
next              228 drivers/md/bcache/sysfs.c 		next_io = wb ? div64_s64(dc->writeback_rate.next-local_clock(),
next              205 drivers/md/bcache/util.c 	d->next += div_u64(done * NSEC_PER_SEC, atomic_long_read(&d->rate));
next              213 drivers/md/bcache/util.c 	if (time_before64(now + NSEC_PER_SEC * 5LLU / 2LLU, d->next))
next              214 drivers/md/bcache/util.c 		d->next = now + NSEC_PER_SEC * 5LLU / 2LLU;
next              216 drivers/md/bcache/util.c 	if (time_after64(now - NSEC_PER_SEC * 2, d->next))
next              217 drivers/md/bcache/util.c 		d->next = now - NSEC_PER_SEC * 2;
next              219 drivers/md/bcache/util.c 	return time_after64(d->next, now)
next              220 drivers/md/bcache/util.c 		? div_u64(d->next - now, NSEC_PER_SEC / HZ)
next              439 drivers/md/bcache/util.h 	uint64_t		next;
next              450 drivers/md/bcache/util.h 	d->next = local_clock();
next               93 drivers/md/bcache/writeback.c 			 dc->writeback_rate.next + NSEC_PER_MSEC))) {
next              396 drivers/md/bcache/writeback.c 	struct keybuf_key *next, *keys[MAX_WRITEBACKS_IN_PASS], *w;
next              412 drivers/md/bcache/writeback.c 	next = bch_keybuf_next(&dc->writeback_keys);
next              416 drivers/md/bcache/writeback.c 	       next) {
next              421 drivers/md/bcache/writeback.c 			BUG_ON(ptr_stale(dc->disk.c, &next->key, 0));
next              447 drivers/md/bcache/writeback.c 						&START_KEY(&next->key)))
next              450 drivers/md/bcache/writeback.c 			size += KEY_SIZE(&next->key);
next              451 drivers/md/bcache/writeback.c 			keys[nk++] = next;
next              452 drivers/md/bcache/writeback.c 		} while ((next = bch_keybuf_next(&dc->writeback_keys)));
next              585 drivers/md/bcache/writeback.c 			goto next;
next              602 drivers/md/bcache/writeback.c next:
next              734 drivers/md/dm-bufio.c 			list_entry(write_list->next, struct dm_buffer, write_list);
next              866 drivers/md/dm-bufio.c 			b = list_entry(c->reserved_buffers.next,
next             1708 drivers/md/dm-bufio.c 		struct dm_buffer *b = list_entry(c->reserved_buffers.next,
next             1748 drivers/md/dm-bufio.c 		struct dm_buffer *b = list_entry(c->reserved_buffers.next,
next               41 drivers/md/dm-cache-policy-smq.c 	unsigned next:28;
next              134 drivers/md/dm-cache-policy-smq.c 	return to_entry(es, e->next);
next              151 drivers/md/dm-cache-policy-smq.c 	e->next = l->head;
next              167 drivers/md/dm-cache-policy-smq.c 	e->next = INDEXER_NULL;
next              171 drivers/md/dm-cache-policy-smq.c 		tail->next = l->tail = to_index(es, e);
next              189 drivers/md/dm-cache-policy-smq.c 		e->next = to_index(es, old);
next              190 drivers/md/dm-cache-policy-smq.c 		prev->next = old->prev = to_index(es, e);
next              200 drivers/md/dm-cache-policy-smq.c 	struct entry *next = l_next(es, e);
next              203 drivers/md/dm-cache-policy-smq.c 		prev->next = e->next;
next              205 drivers/md/dm-cache-policy-smq.c 		l->head = e->next;
next              207 drivers/md/dm-cache-policy-smq.c 	if (next)
next              208 drivers/md/dm-cache-policy-smq.c 		next->prev = e->prev;
next              719 drivers/md/dm-cache-policy-smq.c 	e->next = INDEXER_NULL;
next               82 drivers/md/dm-delay.c 	struct dm_delay_info *delayed, *next;
next               88 drivers/md/dm-delay.c 	list_for_each_entry_safe(delayed, next, &dc->delayed_bios, list) {
next              113 drivers/md/dm-init.c 	char *next;
next              123 drivers/md/dm-init.c 	next = str_field_delimit(&field[i], ',');
next              148 drivers/md/dm-init.c 	return next;
next              192 drivers/md/dm-init.c 	char *next;
next              202 drivers/md/dm-init.c 	next = str_field_delimit(&field[i], ';');
next              223 drivers/md/dm-init.c 	return next;
next             1245 drivers/md/dm-integrity.c 	struct rb_node *next;
next             1255 drivers/md/dm-integrity.c 	next = rb_next(&node->node);
next             1256 drivers/md/dm-integrity.c 	if (unlikely(!next))
next             1259 drivers/md/dm-integrity.c 	next_node = container_of(next, struct journal_node, node);
next             1265 drivers/md/dm-integrity.c 	struct rb_node *next;
next             1271 drivers/md/dm-integrity.c 	next = rb_next(&node->node);
next             1272 drivers/md/dm-integrity.c 	if (unlikely(!next))
next             1275 drivers/md/dm-integrity.c 	next_node = container_of(next, struct journal_node, node);
next             2125 drivers/md/dm-integrity.c 		struct bio *next = flushes->bi_next;
next             2128 drivers/md/dm-integrity.c 		flushes = next;
next             2906 drivers/md/dm-integrity.c 	ic->reboot_notifier.next = NULL;
next             3172 drivers/md/dm-integrity.c 			pl[i - 1].next = &pl[i];
next             3175 drivers/md/dm-integrity.c 	pl[i].next = NULL;
next              193 drivers/md/dm-io.c 	dp->context_ptr = pl->next;
next              546 drivers/md/dm-ioctl.c 				old_nl->next = (uint32_t) ((void *) nl -
next              550 drivers/md/dm-ioctl.c 			nl->next = 0;
next              592 drivers/md/dm-ioctl.c 	info->old_vers->next = (uint32_t) ((void *)info->vers -
next              597 drivers/md/dm-ioctl.c     info->vers->next = 0;
next             1198 drivers/md/dm-ioctl.c 		spec->next = outptr - outbuf;
next             1270 drivers/md/dm-ioctl.c static int next_target(struct dm_target_spec *last, uint32_t next, void *end,
next             1273 drivers/md/dm-ioctl.c 	*spec = (struct dm_target_spec *) ((unsigned char *) last + next);
next             1288 drivers/md/dm-ioctl.c 	uint32_t next = param->data_start;
next             1299 drivers/md/dm-ioctl.c 		r = next_target(spec, next, end, &spec, &target_params);
next             1314 drivers/md/dm-ioctl.c 		next = spec->next;
next              243 drivers/md/dm-kcopyd.c 	struct page_list *next;
next              246 drivers/md/dm-kcopyd.c 		next = pl->next;
next              251 drivers/md/dm-kcopyd.c 			pl->next = kc->pages;
next              256 drivers/md/dm-kcopyd.c 		pl = next;
next              274 drivers/md/dm-kcopyd.c 			kc->pages = pl->next;
next              277 drivers/md/dm-kcopyd.c 		pl->next = *pages;
next              294 drivers/md/dm-kcopyd.c 	struct page_list *next;
next              297 drivers/md/dm-kcopyd.c 		next = pl->next;
next              299 drivers/md/dm-kcopyd.c 		pl = next;
next              309 drivers/md/dm-kcopyd.c 	struct page_list *pl = NULL, *next;
next              312 drivers/md/dm-kcopyd.c 		next = alloc_pl(GFP_KERNEL);
next              313 drivers/md/dm-kcopyd.c 		if (!next) {
next              318 drivers/md/dm-kcopyd.c 		next->next = pl;
next              319 drivers/md/dm-kcopyd.c 		pl = next;
next              395 drivers/md/dm-kcopyd.c 	zero_page_list.next = &zero_page_list;
next              448 drivers/md/dm-kcopyd.c 			job = list_entry(jobs->next, struct kcopyd_job, list);
next              469 drivers/md/dm-log-writes.c 				goto next;
next              486 drivers/md/dm-log-writes.c 				goto next;
next              495 drivers/md/dm-log-writes.c next:
next               68 drivers/md/dm-queue-length.c 	struct path_info *pi, *next;
next               70 drivers/md/dm-queue-length.c 	list_for_each_entry_safe(pi, next, paths, list) {
next              442 drivers/md/dm-region-hash.c 	struct dm_region *reg, *next;
next              483 drivers/md/dm-region-hash.c 	list_for_each_entry_safe(reg, next, &recovered, list) {
next              489 drivers/md/dm-region-hash.c 	list_for_each_entry_safe(reg, next, &failed_recovered, list) {
next              494 drivers/md/dm-region-hash.c 	list_for_each_entry_safe(reg, next, &clean, list) {
next              651 drivers/md/dm-region-hash.c 		reg = list_entry(rh->quiesced_regions.next,
next               34 drivers/md/dm-round-robin.c 	struct path_info *pi, *next;
next               36 drivers/md/dm-round-robin.c 	list_for_each_entry_safe(pi, next, paths, list) {
next              189 drivers/md/dm-round-robin.c 		pi = list_entry(s->valid_paths.next, struct path_info, list);
next               64 drivers/md/dm-service-time.c 	struct path_info *pi, *next;
next               66 drivers/md/dm-service-time.c 	list_for_each_entry_safe(pi, next, paths, list) {
next             1745 drivers/md/dm-snap.c 		struct rb_node *next;
next             1750 drivers/md/dm-snap.c 		next = rb_first(&s->out_of_order_tree);
next             1751 drivers/md/dm-snap.c 		while (next) {
next             1752 drivers/md/dm-snap.c 			pe = rb_entry(next, struct dm_snap_pending_exception,
next             1756 drivers/md/dm-snap.c 			next = rb_next(next);
next              213 drivers/md/dm-stats.c 		s = container_of(stats->list.next, struct dm_stat, list_entry);
next              219 drivers/md/dm-table.c 	struct list_head *tmp, *next;
next              221 drivers/md/dm-table.c 	list_for_each_safe(tmp, next, devices) {
next             2327 drivers/md/dm-thin.c 		tc = list_entry_rcu(pool->active_thins.next, struct thin_c, list);
next              124 drivers/md/dm-uevent.c 	struct dm_uevent *event, *next;
next              126 drivers/md/dm-uevent.c 	list_for_each_entry_safe(event, next, events, elist) {
next              639 drivers/md/dm-writecache.c 		struct rb_node *next;
next              643 drivers/md/dm-writecache.c 		next = rb_next(&e->rb_node);
next              645 drivers/md/dm-writecache.c 		if (unlikely(!next))
next              646 drivers/md/dm-writecache.c 			next = rb_first(&wc->freetree);
next              647 drivers/md/dm-writecache.c 		wc->current_free = next ? container_of(next, struct wc_entry, rb_node) : NULL;
next              651 drivers/md/dm-writecache.c 		e = container_of(wc->freelist.next, struct wc_entry, lru);
next              688 drivers/md/dm-writecache.c 	wc->lru.next = LIST_POISON1;
next              690 drivers/md/dm-writecache.c 	wc->freelist.next = LIST_POISON1;
next              717 drivers/md/dm-writecache.c 	e = container_of(wc->lru.next, struct wc_entry, lru);
next              728 drivers/md/dm-writecache.c 		if (unlikely(e->lru.next == &wc->lru))
next              730 drivers/md/dm-writecache.c 		e2 = container_of(e->lru.next, struct wc_entry, lru);
next             1365 drivers/md/dm-writecache.c 		wb = list_entry(list->next, struct writeback_struct, endio_entry);
next             1402 drivers/md/dm-writecache.c 		c = list_entry(list->next, struct copy_struct, endio_entry);
next             1448 drivers/md/dm-writecache.c 		list.next->prev = list.prev->next = &list;
next             1381 drivers/md/dm-zoned-metadata.c 			goto next;
next             1402 drivers/md/dm-zoned-metadata.c 			goto next;
next             1424 drivers/md/dm-zoned-metadata.c next:
next             2341 drivers/md/dm-zoned-metadata.c 	struct dmz_mblock *mblk, *next;
next             2381 drivers/md/dm-zoned-metadata.c 	rbtree_postorder_for_each_entry_safe(mblk, next, root, node) {
next              847 drivers/md/dm.c 	struct list_head *tmp, *next;
next              849 drivers/md/dm.c 	list_for_each_safe(tmp, next, devices) {
next              344 drivers/md/md-bitmap.c 		struct buffer_head *next = bh->b_this_page;
next              346 drivers/md/md-bitmap.c 		bh = next;
next              307 drivers/md/md.c 		_tmp = all_mddevs.next;					\
next              316 drivers/md/md.c 		_tmp = _tmp->next;})					\
next             4461 drivers/md/md.c 				= list_entry(mddev->disks.next,
next             6280 drivers/md/md.c 		rdev0 = list_entry(pending_raid_disks.next,
next             6521 drivers/md/md.c 				= list_entry(mddev->disks.next,
next             7976 drivers/md/md.c 		tmp = all_mddevs.next;
next             7978 drivers/md/md.c 		tmp = mddev->all_mddevs.next;
next             8101 drivers/md/md.c 	.next   = md_seq_next,
next             8688 drivers/md/md.c 			int next = (last_mark+1) % SYNC_MARKS;
next             8690 drivers/md/md.c 			mddev->resync_mark = mark[next];
next             8691 drivers/md/md.c 			mddev->resync_mark_cnt = mark_cnt[next];
next             8692 drivers/md/md.c 			mark[next] = jiffies;
next             8693 drivers/md/md.c 			mark_cnt[next] = io_sectors - atomic_read(&mddev->recovery_active);
next             8694 drivers/md/md.c 			last_mark = next;
next             9305 drivers/md/md.c 	.next		= NULL,
next             9571 drivers/md/md.c 		node_detected_dev = list_entry(all_detected_devices.next,
next              122 drivers/md/persistent-data/dm-space-map-metadata.c 	unsigned next = brb_next(brb, brb->end);
next              128 drivers/md/persistent-data/dm-space-map-metadata.c 	if (next == brb->begin)
next              135 drivers/md/persistent-data/dm-space-map-metadata.c 	brb->end = next;
next              809 drivers/md/raid1.c 		struct bio *next = bio->bi_next;
next              821 drivers/md/raid1.c 		bio = next;
next              909 drivers/md/raid10.c 			struct bio *next = bio->bi_next;
next              921 drivers/md/raid10.c 			bio = next;
next             1094 drivers/md/raid10.c 		struct bio *next = bio->bi_next;
next             1106 drivers/md/raid10.c 		bio = next;
next             4435 drivers/md/raid10.c 	sector_t next, safe, last;
next             4471 drivers/md/raid10.c 		next = first_dev_address(conf->reshape_progress - 1,
next             4480 drivers/md/raid10.c 		if (next + conf->offset_diff < safe)
next             4492 drivers/md/raid10.c 		next = last_dev_address(conf->reshape_progress, &conf->geo);
next             4502 drivers/md/raid10.c 		if (next > safe + conf->offset_diff)
next              517 drivers/md/raid5-cache.c 	struct stripe_head *sh, *next;
next              519 drivers/md/raid5-cache.c 	list_for_each_entry_safe(sh, next, &io->stripe_list, log_list) {
next              531 drivers/md/raid5-cache.c 	struct r5l_io_unit *io, *next;
next              535 drivers/md/raid5-cache.c 	list_for_each_entry_safe(io, next, &log->running_ios, log_sibling) {
next              547 drivers/md/raid5-cache.c 	struct r5l_io_unit *io, *next;
next              551 drivers/md/raid5-cache.c 	list_for_each_entry_safe(io, next, &log->running_ios, log_sibling) {
next             1202 drivers/md/raid5-cache.c 	struct r5l_io_unit *io, *next;
next             1207 drivers/md/raid5-cache.c 	list_for_each_entry_safe(io, next, &log->finished_ios, log_sibling) {
next             1402 drivers/md/raid5-cache.c 	struct stripe_head *sh, *next;
next             1409 drivers/md/raid5-cache.c 	list_for_each_entry_safe(sh, next, &conf->r5c_full_stripe_list, lru) {
next             1416 drivers/md/raid5-cache.c 	list_for_each_entry_safe(sh, next,
next             1960 drivers/md/raid5-cache.c 	struct stripe_head *sh, *next;
next             1962 drivers/md/raid5-cache.c 	list_for_each_entry_safe(sh, next, cached_stripe_list, lru) {
next             1973 drivers/md/raid5-cache.c 	struct stripe_head *sh, *next;
next             1975 drivers/md/raid5-cache.c 	list_for_each_entry_safe(sh, next, cached_stripe_list, lru)
next             2432 drivers/md/raid5-cache.c 	struct stripe_head *sh, *next;
next             2439 drivers/md/raid5-cache.c 	list_for_each_entry_safe(sh, next, &ctx->cached_list, lru) {
next              402 drivers/md/raid5-ppl.c 	struct stripe_head *sh, *next;
next              409 drivers/md/raid5-ppl.c 	list_for_each_entry_safe(sh, next, &io->stripe_list, log_list) {
next              439 drivers/md/raid5.c 	first = (conf->inactive_list + hash)->next;
next              896 drivers/md/raid5.c 	struct list_head *first, *next = NULL;
next              904 drivers/md/raid5.c 	first = conf->pending_list.next;
next              915 drivers/md/raid5.c 			first = data->sibling.next;
next              916 drivers/md/raid5.c 		next = data->sibling.next;
next              927 drivers/md/raid5.c 	if (next != &conf->pending_list)
next              928 drivers/md/raid5.c 		conf->next_pending_data = list_entry(next,
next             2340 drivers/md/raid5.c 			nsh = list_entry(newstripes.next, struct stripe_head, lru);
next             2413 drivers/md/raid5.c 		nsh = list_entry(newstripes.next, struct stripe_head, lru);
next             4608 drivers/md/raid5.c 	struct stripe_head *sh, *next;
next             4612 drivers/md/raid5.c 	list_for_each_entry_safe(sh, next, &head_sh->batch_list, batch_list) {
next             5065 drivers/md/raid5.c 			struct list_head *l = conf->delayed_list.next;
next             5086 drivers/md/raid5.c 		struct stripe_head *sh = list_entry(head.next, struct stripe_head, lru);
next             5363 drivers/md/raid5.c 		sh = list_entry(handle_list->next, typeof(*sh), lru);
next             5368 drivers/md/raid5.c 			if (conf->hold_list.next == conf->last_hold)
next             5371 drivers/md/raid5.c 				conf->last_hold = conf->hold_list.next;
next             5433 drivers/md/raid5.c 	if (cb->list.next && !list_empty(&cb->list)) {
next             5477 drivers/md/raid5.c 	if (cb->list.next == NULL) {
next             5975 drivers/md/raid5.c 		sh = list_entry(stripes.next, struct stripe_head, lru);
next              118 drivers/media/common/saa7146/saa7146_fops.c 	struct saa7146_buf *buf,*next = NULL;
next              127 drivers/media/common/saa7146/saa7146_fops.c 		buf = list_entry(q->queue.next,struct saa7146_buf,vb.queue);
next              130 drivers/media/common/saa7146/saa7146_fops.c 			next = list_entry(q->queue.next,struct saa7146_buf, vb.queue);
next              133 drivers/media/common/saa7146/saa7146_fops.c 			buf, q->queue.prev, q->queue.next);
next              134 drivers/media/common/saa7146/saa7146_fops.c 		buf->activate(dev,buf,next);
next              346 drivers/media/common/saa7146/saa7146_fops.c 		buf = list_entry(q->stream.next, struct videobuf_buffer, stream);
next              312 drivers/media/common/saa7146/saa7146_hlp.c 	int cur = 0, next = 0;
next              328 drivers/media/common/saa7146/saa7146_hlp.c 	for( cur = 0, next = 1; next < *count; next++) {
next              329 drivers/media/common/saa7146/saa7146_hlp.c 		if( values[cur] != values[next])
next              330 drivers/media/common/saa7146/saa7146_hlp.c 			values[++cur] = values[next];
next              997 drivers/media/common/saa7146/saa7146_hlp.c void saa7146_set_capture(struct saa7146_dev *dev, struct saa7146_buf *buf, struct saa7146_buf *next)
next             1003 drivers/media/common/saa7146/saa7146_hlp.c 	DEB_CAP("buf:%p, next:%p\n", buf, next);
next              135 drivers/media/common/saa7146/saa7146_vbi.c static void saa7146_set_vbi_capture(struct saa7146_dev *dev, struct saa7146_buf *buf, struct saa7146_buf *next)
next              207 drivers/media/common/saa7146/saa7146_vbi.c 			   struct saa7146_buf *next)
next              212 drivers/media/common/saa7146/saa7146_vbi.c 	DEB_VBI("dev:%p, buf:%p, next:%p\n", dev, buf, next);
next              213 drivers/media/common/saa7146/saa7146_vbi.c 	saa7146_set_vbi_capture(dev,buf,next);
next             1020 drivers/media/common/saa7146/saa7146_video.c 		     struct saa7146_buf *next)
next             1025 drivers/media/common/saa7146/saa7146_video.c 	saa7146_set_capture(dev,buf,next);
next              430 drivers/media/common/siano/smscoreapi.c 	struct list_head *next;
next              433 drivers/media/common/siano/smscoreapi.c 	for (next = g_smscore_registry.next;
next              434 drivers/media/common/siano/smscoreapi.c 	     next != &g_smscore_registry;
next              435 drivers/media/common/siano/smscoreapi.c 	     next = next->next) {
next              436 drivers/media/common/siano/smscoreapi.c 		entry = (struct smscore_registry_entry_t *) next;
next              527 drivers/media/common/siano/smscoreapi.c 	struct list_head *next, *first;
next              535 drivers/media/common/siano/smscoreapi.c 		for (next = first->next;
next              536 drivers/media/common/siano/smscoreapi.c 		     next != first && !rc;
next              537 drivers/media/common/siano/smscoreapi.c 		     next = next->next) {
next              539 drivers/media/common/siano/smscoreapi.c 				(struct smscore_device_t *) next;
next              565 drivers/media/common/siano/smscoreapi.c 	struct list_head *next, *first;
next              571 drivers/media/common/siano/smscoreapi.c 	for (next = first->next; next != first;) {
next              573 drivers/media/common/siano/smscoreapi.c 			(struct smscore_device_notifyee_t *) next;
next              574 drivers/media/common/siano/smscoreapi.c 		next = next->next;
next              592 drivers/media/common/siano/smscoreapi.c 		client = (struct smscore_client_t *) coredev->clients.next;
next             1213 drivers/media/common/siano/smscoreapi.c 			cb = (struct smscore_buffer_t *) coredev->buffers.next;
next             1639 drivers/media/common/siano/smscoreapi.c 		cb = (struct smscore_buffer_t *) coredev->buffers.next;
next             1765 drivers/media/common/siano/smscoreapi.c 			(struct smscore_idlist_t *) client->idlist.next;
next             2140 drivers/media/common/siano/smscoreapi.c 				g_smscore_notifyees.next;
next             2151 drivers/media/common/siano/smscoreapi.c 				g_smscore_registry.next;
next             1219 drivers/media/common/siano/smsdvb-main.c 		smsdvb_unregister_client((struct smsdvb_client_t *)g_smsdvb_clients.next);
next              499 drivers/media/dvb-core/dmxdev.c 		list_for_each_entry(feed, &dmxdevfilter->feed.ts, next)
next              520 drivers/media/dvb-core/dmxdev.c 		list_for_each_entry(feed, &filter->feed.ts, next) {
next              580 drivers/media/dvb-core/dmxdev.c 		list_for_each_entry(feed, &dmxdevfilter->feed.ts, next) {
next              600 drivers/media/dvb-core/dmxdev.c 	list_for_each_entry_safe(feed, tmp, &dmxdevfilter->feed.ts, next) {
next              601 drivers/media/dvb-core/dmxdev.c 		list_del(&feed->next);
next              774 drivers/media/dvb-core/dmxdev.c 		list_for_each_entry(feed, &filter->feed.ts, next) {
next              891 drivers/media/dvb-core/dmxdev.c 	list_add(&feed->next, &filter->feed.ts);
next              908 drivers/media/dvb-core/dmxdev.c 	list_for_each_entry_safe(feed, tmp, &filter->feed.ts, next) {
next              913 drivers/media/dvb-core/dmxdev.c 			list_del(&feed->next);
next              191 drivers/media/dvb-core/dvb_demux.c 	} while ((f = f->next) && sec->is_filtering);
next              899 drivers/media/dvb-core/dvb_demux.c 	dvbdmxfilter->next = dvbdmxfeed->filter;
next              948 drivers/media/dvb-core/dvb_demux.c 	} while ((f = f->next));
next             1046 drivers/media/dvb-core/dvb_demux.c 		dvbdmxfeed->filter = dvbdmxfilter->next;
next             1048 drivers/media/dvb-core/dvb_demux.c 		while (f->next != dvbdmxfilter)
next             1049 drivers/media/dvb-core/dvb_demux.c 			f = f->next;
next             1050 drivers/media/dvb-core/dvb_demux.c 		f->next = f->next->next;
next              102 drivers/media/dvb-core/dvb_vb2.c 		buf = list_entry(ctx->dvb_q.next,
next              286 drivers/media/dvb-core/dvb_vb2.c 			ctx->buf = list_entry(ctx->dvb_q.next,
next             1217 drivers/media/dvb-frontends/drx39xyj/drx_driver.h 	struct drx_version_list *next;
next              163 drivers/media/dvb-frontends/rtl2832_sdr.c 	buf = list_entry(dev->queued_bufs.next,
next              410 drivers/media/dvb-frontends/rtl2832_sdr.c 		buf = list_entry(dev->queued_bufs.next,
next              268 drivers/media/i2c/et8ek8/et8ek8_driver.c 	const struct et8ek8_reg *next;
next              277 drivers/media/i2c/et8ek8/et8ek8_driver.c 	next = regs;
next              284 drivers/media/i2c/et8ek8/et8ek8_driver.c 		while (next->type != ET8EK8_REG_TERM &&
next              285 drivers/media/i2c/et8ek8/et8ek8_driver.c 		       next->type != ET8EK8_REG_DELAY) {
next              290 drivers/media/i2c/et8ek8/et8ek8_driver.c 			if (WARN(next->type != ET8EK8_REG_8BIT &&
next              291 drivers/media/i2c/et8ek8/et8ek8_driver.c 				 next->type != ET8EK8_REG_16BIT,
next              292 drivers/media/i2c/et8ek8/et8ek8_driver.c 				 "Invalid type = %d", next->type)) {
next              300 drivers/media/i2c/et8ek8/et8ek8_driver.c 			next++;
next              316 drivers/media/i2c/et8ek8/et8ek8_driver.c 		if (next->type == ET8EK8_REG_DELAY) {
next              317 drivers/media/i2c/et8ek8/et8ek8_driver.c 			msleep(next->val);
next              322 drivers/media/i2c/et8ek8/et8ek8_driver.c 			next++;
next              323 drivers/media/i2c/et8ek8/et8ek8_driver.c 			regs = next;
next              326 drivers/media/i2c/et8ek8/et8ek8_driver.c 	} while (next->type != ET8EK8_REG_TERM);
next              363 drivers/media/i2c/et8ek8/et8ek8_driver.c 	struct et8ek8_reglist **next = &meta->reglist[0].ptr;
next              365 drivers/media/i2c/et8ek8/et8ek8_driver.c 	while (*next) {
next              366 drivers/media/i2c/et8ek8/et8ek8_driver.c 		if ((*next)->type == type)
next              367 drivers/media/i2c/et8ek8/et8ek8_driver.c 			return *next;
next              369 drivers/media/i2c/et8ek8/et8ek8_driver.c 		next++;
next              571 drivers/media/i2c/imx274.c 	const struct reg_8 *next;
next              579 drivers/media/i2c/imx274.c 	for (next = table;; next++) {
next              580 drivers/media/i2c/imx274.c 		if ((next->addr != range_start + range_count) ||
next              581 drivers/media/i2c/imx274.c 		    (next->addr == IMX274_TABLE_END) ||
next              582 drivers/media/i2c/imx274.c 		    (next->addr == IMX274_TABLE_WAIT_MS) ||
next              601 drivers/media/i2c/imx274.c 			if (next->addr == IMX274_TABLE_END)
next              604 drivers/media/i2c/imx274.c 			if (next->addr == IMX274_TABLE_WAIT_MS) {
next              605 drivers/media/i2c/imx274.c 				msleep_range(next->val);
next              610 drivers/media/i2c/imx274.c 		val = next->val;
next              613 drivers/media/i2c/imx274.c 			range_start = next->addr;
next              357 drivers/media/i2c/tvp514x.c 	const struct tvp514x_reg *next = reglist;
next              359 drivers/media/i2c/tvp514x.c 	for (; next->token != TOK_TERM; next++) {
next              360 drivers/media/i2c/tvp514x.c 		if (next->token == TOK_DELAY) {
next              361 drivers/media/i2c/tvp514x.c 			msleep(next->val);
next              365 drivers/media/i2c/tvp514x.c 		if (next->token == TOK_SKIP)
next              368 drivers/media/i2c/tvp514x.c 		err = tvp514x_write_reg(sd, next->reg, (u8) next->val);
next               84 drivers/media/mc/mc-device.c 	int next = id & MEDIA_ENT_ID_FLAG_NEXT;
next               89 drivers/media/mc/mc-device.c 		if (((media_entity_id(entity) == id) && !next) ||
next               90 drivers/media/mc/mc-device.c 		    ((media_entity_id(entity) > id) && next)) {
next              586 drivers/media/mc/mc-device.c 	struct media_entity_notify *notify, *next;
next              621 drivers/media/mc/mc-device.c 	list_for_each_entry_safe(notify, next, &mdev->entity_notify, list)
next              802 drivers/media/mc/mc-device.c 	struct media_entity *next;
next              821 drivers/media/mc/mc-device.c 	list_for_each_entry_safe(entity, next, &mdev->entities, graph_obj.list)
next              249 drivers/media/mc/mc-entity.c 	graph->stack[graph->top].link = entity->links.next;
next              312 drivers/media/mc/mc-entity.c 	struct media_entity *next;
next              318 drivers/media/mc/mc-entity.c 		link_top(graph) = link_top(graph)->next;
next              327 drivers/media/mc/mc-entity.c 	next = media_entity_other(entity, link);
next              330 drivers/media/mc/mc-entity.c 	if (media_entity_enum_test_and_set(&graph->ent_enum, next)) {
next              331 drivers/media/mc/mc-entity.c 		link_top(graph) = link_top(graph)->next;
next              334 drivers/media/mc/mc-entity.c 			next->name);
next              339 drivers/media/mc/mc-entity.c 	link_top(graph) = link_top(graph)->next;
next              340 drivers/media/mc/mc-entity.c 	stack_push(graph, next);
next              342 drivers/media/mc/mc-entity.c 		next->name);
next             2928 drivers/media/pci/bt8xx/bttv-driver.c 		buf = list_entry(fh->cap.stream.next,struct bttv_buffer,vb.stream);
next             3473 drivers/media/pci/bt8xx/bttv-driver.c 		item = list_entry(btv->capture.next, struct bttv_buffer, vb.queue);
next             3481 drivers/media/pci/bt8xx/bttv-driver.c 		    (item->vb.queue.next != &btv->capture)) {
next             3482 drivers/media/pci/bt8xx/bttv-driver.c 			item = list_entry(item->vb.queue.next, struct bttv_buffer, vb.queue);
next             3632 drivers/media/pci/bt8xx/bttv-driver.c 		item = list_entry(btv->capture.next, struct bttv_buffer, vb.queue);
next             3638 drivers/media/pci/bt8xx/bttv-driver.c 		item = list_entry(btv->vcapture.next, struct bttv_buffer, vb.queue);
next             3726 drivers/media/pci/bt8xx/bttv-driver.c 		new = list_entry(btv->vcapture.next, struct bttv_buffer, vb.queue);
next              548 drivers/media/pci/bt8xx/bttv-risc.c 	unsigned long next = btv->main.dma + ((slot+2) << 2);
next              552 drivers/media/pci/bt8xx/bttv-risc.c 		btv->main.cpu[slot+1] = cpu_to_le32(next);
next              564 drivers/media/pci/bt8xx/bttv-risc.c 		risc->jmp[1] = cpu_to_le32(next);
next              644 drivers/media/pci/bt8xx/bttv-risc.c 			if (set->top->vb.queue.next)
next              649 drivers/media/pci/bt8xx/bttv-risc.c 			if (set->top->vb.queue.next)
next              651 drivers/media/pci/bt8xx/bttv-risc.c 			if (set->bottom->vb.queue.next)
next              666 drivers/media/pci/bt8xx/bttv-risc.c 		if (set->top->vb.queue.next)
next              677 drivers/media/pci/bt8xx/bttv-risc.c 		if (set->bottom->vb.queue.next)
next              155 drivers/media/pci/cobalt/cobalt-omnitek.c 	dma_addr_t next = desc->bus;
next              164 drivers/media/pci/cobalt/cobalt-omnitek.c 	WARN_ON(next & 3);
next              209 drivers/media/pci/cobalt/cobalt-omnitek.c 				next += sizeof(struct sg_dma_descriptor);
next              210 drivers/media/pci/cobalt/cobalt-omnitek.c 				d->next_h = (u32)((u64)next >> 32);
next              211 drivers/media/pci/cobalt/cobalt-omnitek.c 				d->next_l = (u32)next |
next              253 drivers/media/pci/cobalt/cobalt-omnitek.c 		next += sizeof(struct sg_dma_descriptor);
next              263 drivers/media/pci/cobalt/cobalt-omnitek.c 			d->next_h = (u32)((u64)next >> 32);
next              264 drivers/media/pci/cobalt/cobalt-omnitek.c 			d->next_l = (u32)next | (to_pci ? WRITE_TO_PCI : 0);
next              272 drivers/media/pci/cobalt/cobalt-omnitek.c 			   struct sg_dma_desc_info *next)
next              277 drivers/media/pci/cobalt/cobalt-omnitek.c 	if (next == NULL) {
next              281 drivers/media/pci/cobalt/cobalt-omnitek.c 		d->next_h = (u32)((u64)next->bus >> 32);
next              282 drivers/media/pci/cobalt/cobalt-omnitek.c 		d->next_l = (u32)next->bus | direction | INTERRUPT_ENABLE;
next               40 drivers/media/pci/cobalt/cobalt-omnitek.h 			   struct sg_dma_desc_info *next);
next              689 drivers/media/pci/cx18/cx18-fileops.c 		buf = list_entry(s->vb_capture.next,
next             1176 drivers/media/pci/cx23885/cx23885-417.c 		struct cx23885_buffer *buf = list_entry(dmaq->active.next,
next             1184 drivers/media/pci/cx23885/cx23885-417.c 		struct cx23885_buffer *buf = list_entry(dmaq->active.next,
next              434 drivers/media/pci/cx23885/cx23885-core.c 		buf = list_entry(q->active.next,
next             1651 drivers/media/pci/cx23885/cx23885-core.c 		buf = list_entry(q->active.next, struct cx23885_buffer,
next              154 drivers/media/pci/cx23885/cx23885-dvb.c 	struct cx23885_buffer *buf = list_entry(dmaq->active.next,
next              221 drivers/media/pci/cx23885/cx23885-vbi.c 	struct cx23885_buffer *buf = list_entry(dmaq->active.next,
next              237 drivers/media/pci/cx23885/cx23885-vbi.c 		struct cx23885_buffer *buf = list_entry(dmaq->active.next,
next               95 drivers/media/pci/cx23885/cx23885-video.c 	buf = list_entry(q->active.next,
next              488 drivers/media/pci/cx23885/cx23885-video.c 	struct cx23885_buffer *buf = list_entry(dmaq->active.next,
next              504 drivers/media/pci/cx23885/cx23885-video.c 		struct cx23885_buffer *buf = list_entry(dmaq->active.next,
next              113 drivers/media/pci/cx25821/cx25821-video.c 			buf = list_entry(dmaq->active.next,
next              266 drivers/media/pci/cx25821/cx25821-video.c 	struct cx25821_buffer *buf = list_entry(dmaq->active.next,
next              284 drivers/media/pci/cx25821/cx25821-video.c 		struct cx25821_buffer *buf = list_entry(dmaq->active.next,
next              733 drivers/media/pci/cx88/cx88-blackbird.c 		buf = list_entry(dmaq->active.next, struct cx88_buffer, list);
next              741 drivers/media/pci/cx88/cx88-blackbird.c 		struct cx88_buffer *buf = list_entry(dmaq->active.next,
next              769 drivers/media/pci/cx88/cx88-blackbird.c 		struct cx88_buffer *buf = list_entry(dmaq->active.next,
next              523 drivers/media/pci/cx88/cx88-core.c 	buf = list_entry(q->active.next,
next              125 drivers/media/pci/cx88/cx88-dvb.c 	buf = list_entry(dmaq->active.next, struct cx88_buffer, list);
next              140 drivers/media/pci/cx88/cx88-dvb.c 		struct cx88_buffer *buf = list_entry(dmaq->active.next,
next              204 drivers/media/pci/cx88/cx88-mpeg.c 	buf = list_entry(q->active.next, struct cx88_buffer, list);
next              277 drivers/media/pci/cx88/cx88-mpeg.c 		buf = list_entry(q->active.next, struct cx88_buffer, list);
next              106 drivers/media/pci/cx88/cx88-vbi.c 	buf = list_entry(q->active.next, struct cx88_buffer, list);
next              198 drivers/media/pci/cx88/cx88-vbi.c 	struct cx88_buffer *buf = list_entry(dmaq->active.next,
next              217 drivers/media/pci/cx88/cx88-vbi.c 		struct cx88_buffer *buf = list_entry(dmaq->active.next,
next              411 drivers/media/pci/cx88/cx88-video.c 		buf = list_entry(q->active.next, struct cx88_buffer, list);
next              533 drivers/media/pci/cx88/cx88-video.c 	struct cx88_buffer *buf = list_entry(dmaq->active.next,
next              551 drivers/media/pci/cx88/cx88-video.c 		struct cx88_buffer *buf = list_entry(dmaq->active.next,
next              906 drivers/media/pci/intel/ipu3/ipu3-cio2.c 	unsigned int i, j, next = q->bufs_next;
next              938 drivers/media/pci/intel/ipu3/ipu3-cio2.c 	if (bufs_queued <= 1 || fbpt_rp == next)
next              940 drivers/media/pci/intel/ipu3/ipu3-cio2.c 		next = (fbpt_rp + 1) % CIO2_MAX_BUFFERS;
next              949 drivers/media/pci/intel/ipu3/ipu3-cio2.c 		if (!q->bufs[next]) {
next              950 drivers/media/pci/intel/ipu3/ipu3-cio2.c 			q->bufs[next] = b;
next              951 drivers/media/pci/intel/ipu3/ipu3-cio2.c 			entry = &q->fbpt[next * CIO2_MAX_LOPS];
next              954 drivers/media/pci/intel/ipu3/ipu3-cio2.c 			q->bufs_next = (next + 1) % CIO2_MAX_BUFFERS;
next              961 drivers/media/pci/intel/ipu3/ipu3-cio2.c 		dev_dbg(&cio2->pci_dev->dev, "entry %i was full!\n", next);
next              962 drivers/media/pci/intel/ipu3/ipu3-cio2.c 		next = (next + 1) % CIO2_MAX_BUFFERS;
next              217 drivers/media/pci/ivtv/ivtv-irq.c 	buf = list_entry(s->q_predma.list.next, struct ivtv_buffer, list);
next               66 drivers/media/pci/ivtv/ivtv-queue.c 		buf = list_entry(q->list.next, struct ivtv_buffer, list);
next               67 drivers/media/pci/ivtv/ivtv-queue.c 		list_del_init(q->list.next);
next               79 drivers/media/pci/ivtv/ivtv-queue.c 	struct ivtv_buffer *buf = list_entry(from->list.next, struct ivtv_buffer, list);
next               81 drivers/media/pci/ivtv/ivtv-queue.c 	list_move_tail(from->list.next, &to->list);
next              267 drivers/media/pci/saa7134/saa7134-core.c 	struct saa7134_buf *next = NULL;
next              279 drivers/media/pci/saa7134/saa7134-core.c 			next = list_entry(q->queue.next, struct saa7134_buf,
next              282 drivers/media/pci/saa7134/saa7134-core.c 			buf->activate(dev, buf, next);
next              307 drivers/media/pci/saa7134/saa7134-core.c 	struct saa7134_buf *buf,*next = NULL;
next              314 drivers/media/pci/saa7134/saa7134-core.c 		buf = list_entry(q->queue.next, struct saa7134_buf, entry);
next              316 drivers/media/pci/saa7134/saa7134-core.c 			buf, q->queue.prev, q->queue.next);
next              319 drivers/media/pci/saa7134/saa7134-core.c 			next = list_entry(q->queue.next, struct saa7134_buf, entry);
next              321 drivers/media/pci/saa7134/saa7134-core.c 		buf->activate(dev, buf, next);
next              323 drivers/media/pci/saa7134/saa7134-core.c 			q->queue.prev, q->queue.next);
next             1190 drivers/media/pci/saa7134/saa7134-core.c 	list_for_each_entry(mops, &mops_list, next)
next             1342 drivers/media/pci/saa7134/saa7134-core.c 	list_for_each_entry(mops, &mops_list, next)
next             1379 drivers/media/pci/saa7134/saa7134-core.c 	struct saa7134_buf *buf, *next;
next             1384 drivers/media/pci/saa7134/saa7134-core.c 	next = buf;
next             1393 drivers/media/pci/saa7134/saa7134-core.c 		next = list_entry(q->queue.next, struct saa7134_buf,
next             1395 drivers/media/pci/saa7134/saa7134-core.c 	buf->activate(dev, buf, next);
next             1504 drivers/media/pci/saa7134/saa7134-core.c 	list_add_tail(&ops->next,&mops_list);
next             1514 drivers/media/pci/saa7134/saa7134-core.c 	list_del(&ops->next);
next               33 drivers/media/pci/saa7134/saa7134-ts.c 			   struct saa7134_buf *next)
next               42 drivers/media/pci/saa7134/saa7134-ts.c 	if (NULL == next)
next               43 drivers/media/pci/saa7134/saa7134-ts.c 		next = buf;
next               45 drivers/media/pci/saa7134/saa7134-ts.c 		ts_dbg("- [top]     buf=%p next=%p\n", buf, next);
next               47 drivers/media/pci/saa7134/saa7134-ts.c 		saa_writel(SAA7134_RS_BA2(5),saa7134_buffer_base(next));
next               50 drivers/media/pci/saa7134/saa7134-ts.c 		ts_dbg("- [bottom]  buf=%p next=%p\n", buf, next);
next               51 drivers/media/pci/saa7134/saa7134-ts.c 		saa_writel(SAA7134_RS_BA1(5),saa7134_buffer_base(next));
next               71 drivers/media/pci/saa7134/saa7134-vbi.c 			   struct saa7134_buf *next)
next              822 drivers/media/pci/saa7134/saa7134-video.c 			   struct saa7134_buf *next)
next              471 drivers/media/pci/saa7134/saa7134.h 			struct saa7134_buf *next);
next              530 drivers/media/pci/saa7134/saa7134.h 	struct list_head           next;
next               83 drivers/media/pci/solo6x10/solo6x10-g723.c 	for (ss = pstr->substream; ss != NULL; ss = ss->next) {
next              351 drivers/media/pci/solo6x10/solo6x10-g723.c 	     ss; ss = ss->next, i++)
next              724 drivers/media/pci/solo6x10/solo6x10-v4l2-enc.c 				solo_enc->vidq_active.next,
next              339 drivers/media/pci/solo6x10/solo6x10-v4l2.c 				solo_dev->vidq_active.next,
next              207 drivers/media/pci/tw5864/tw5864-core.c 			goto next;
next              234 drivers/media/pci/tw5864/tw5864-core.c next:
next              437 drivers/media/pci/tw5864/tw5864-video.c 		struct tw5864_buf *buf = list_entry(input->active.next,
next              459 drivers/media/pci/tw5864/tw5864-video.c 		struct tw5864_buf *buf = list_entry(input->active.next,
next              403 drivers/media/pci/tw68/tw68-core.c 	buf = container_of(dev->active.next, struct tw68_buf, list);
next              495 drivers/media/pci/tw68/tw68-video.c 		container_of(dev->active.next, struct tw68_buf, list);
next              510 drivers/media/pci/tw68/tw68-video.c 			container_of(dev->active.next, struct tw68_buf, list);
next              987 drivers/media/pci/tw68/tw68-video.c 		buf = list_entry(dev->active.next, struct tw68_buf, list);
next               40 drivers/media/pci/tw686x/tw686x-audio.c 		struct tw686x_audio_buf *next = NULL;
next               54 drivers/media/pci/tw686x/tw686x-audio.c 			next = list_first_entry(&ac->buf_list,
next               56 drivers/media/pci/tw686x/tw686x-audio.c 			list_move_tail(&next->list, &ac->buf_list);
next               58 drivers/media/pci/tw686x/tw686x-audio.c 			ac->curr_bufs[pb] = next;
next               62 drivers/media/pci/tw686x/tw686x-audio.c 		if (!done || !next)
next               74 drivers/media/pci/tw686x/tw686x-audio.c 			reg_write(dev, reg, next->dma);
next              298 drivers/media/pci/tw686x/tw686x-audio.c 	     ss; ss = ss->next, i++)
next             1245 drivers/media/platform/am437x/am437x-vpfe.c 	vpfe->next_frm = list_entry(vpfe->dma_queue.next,
next             1993 drivers/media/platform/am437x/am437x-vpfe.c 	vpfe->next_frm = list_entry(vpfe->dma_queue.next,
next             2061 drivers/media/platform/am437x/am437x-vpfe.c 		vpfe->next_frm = list_entry(vpfe->dma_queue.next,
next              191 drivers/media/platform/atmel/atmel-isi.c 		isi->active = list_entry(isi->video_buffer_list.next,
next              330 drivers/media/platform/atmel/atmel-isi.c 			desc = list_entry(isi->dma_desc_head.next,
next               92 drivers/media/platform/davinci/vpbe_display.c 	layer->next_frm = list_entry(layer->dma_queue.next,
next              279 drivers/media/platform/davinci/vpbe_display.c 	layer->next_frm = layer->cur_frm = list_entry(layer->dma_queue.next,
next              340 drivers/media/platform/davinci/vpbe_display.c 		layer->next_frm = list_entry(layer->dma_queue.next,
next              461 drivers/media/platform/davinci/vpfe_capture.c 	vpfe_dev->next_frm = list_entry(vpfe_dev->dma_queue.next,
next             1434 drivers/media/platform/davinci/vpfe_capture.c 	vpfe_dev->next_frm = list_entry(vpfe_dev->dma_queue.next,
next              206 drivers/media/platform/davinci/vpif_capture.c 	common->cur_frm = common->next_frm = list_entry(common->dma_queue.next,
next              297 drivers/media/platform/davinci/vpif_capture.c 		common->next_frm = list_entry(common->dma_queue.next,
next              345 drivers/media/platform/davinci/vpif_capture.c 	common->next_frm = list_entry(common->dma_queue.next,
next              199 drivers/media/platform/davinci/vpif_display.c 			    list_entry(common->dma_queue.next,
next              284 drivers/media/platform/davinci/vpif_display.c 		common->next_frm = list_entry(common->dma_queue.next,
next              309 drivers/media/platform/davinci/vpif_display.c 	common->next_frm = list_entry(common->dma_queue.next,
next             1464 drivers/media/platform/exynos4-is/fimc-capture.c 			buf = list_entry(fimc->vid_cap.active_buf_q.next,
next              692 drivers/media/platform/exynos4-is/fimc-core.h 	buf = list_entry(vid_cap->active_buf_q.next,
next              718 drivers/media/platform/exynos4-is/fimc-core.h 	buf = list_entry(vid_cap->pending_buf_q.next,
next              201 drivers/media/platform/exynos4-is/fimc-lite.h 	struct flite_buffer *buf = list_entry(dev->active_buf_q.next,
next              216 drivers/media/platform/exynos4-is/fimc-lite.h 	struct flite_buffer *buf = list_entry(dev->pending_buf_q.next,
next              376 drivers/media/platform/exynos4-is/media-dev.c 		p = list_entry(fmd->pipelines.next, typeof(*p), list);
next              296 drivers/media/platform/fsl-viu.c 		buf = list_entry(vidq->active.next, struct viu_buf, vb.queue);
next              316 drivers/media/platform/fsl-viu.c 		buf = list_entry(vidq->queued.next, struct viu_buf, vb.queue);
next              350 drivers/media/platform/fsl-viu.c 		buf = list_entry(vidq->active.next, struct viu_buf, vb.queue);
next              505 drivers/media/platform/fsl-viu.c 			dev, &vidq->queued, vidq->queued.next,
next              963 drivers/media/platform/fsl-viu.c 		buf = list_entry(vidq->active.next, struct viu_buf,
next              968 drivers/media/platform/fsl-viu.c 		buf = list_entry(vidq->queued.next, struct viu_buf,
next             1074 drivers/media/platform/fsl-viu.c 		buf = list_entry(vidq->active.next,
next             1545 drivers/media/platform/fsl-viu.c 	struct v4l2_subdev *sdev = list_entry(v4l2_dev->subdevs.next,
next              542 drivers/media/platform/omap/omap_vout.c 	vout->next_frm = list_entry(vout->dma_queue.next,
next             1002 drivers/media/platform/omap/omap_vout.c 	vout->next_frm = vout->cur_frm = list_entry(vout->dma_queue.next,
next             1060 drivers/media/platform/omap3isp/isp.c 	struct isp_reg *next = reg_list;
next             1062 drivers/media/platform/omap3isp/isp.c 	for (; next->reg != ISP_TOK_TERM; next++)
next             1063 drivers/media/platform/omap3isp/isp.c 		next->val = isp_reg_readl(isp, next->mmio_range, next->reg);
next             1075 drivers/media/platform/omap3isp/isp.c 	struct isp_reg *next = reg_list;
next             1077 drivers/media/platform/omap3isp/isp.c 	for (; next->reg != ISP_TOK_TERM; next++)
next             1078 drivers/media/platform/omap3isp/isp.c 		isp_reg_writel(isp, next->val, next->mmio_range, next->reg);
next             1001 drivers/media/platform/pxa_camera.c 	pcdev->active = list_entry(pcdev->capture.next,
next              851 drivers/media/platform/rcar-vin/rcar-dma.c 		buf = list_entry(vin->buf_list.next, struct rvin_buffer, list);
next              548 drivers/media/platform/rcar_fdp1.c 	struct fdp1_field_buffer	*next;
next              934 drivers/media/platform/rcar_fdp1.c 	if (job->next)
next              935 drivers/media/platform/rcar_fdp1.c 		fdp1_write(fdp1, job->next->addrs[0], FD1_RPF2_ADDR_Y);
next             1233 drivers/media/platform/rcar_fdp1.c 		job->next = fdp1_peek_queued_field(ctx);
next              136 drivers/media/platform/s5p-mfc/s5p_mfc.c 		b = list_entry(lh->next, struct s5p_mfc_buf, list);
next              227 drivers/media/platform/s5p-mfc/s5p_mfc.c 		dst_buf = list_entry(ctx->dst_queue.next,
next              265 drivers/media/platform/s5p-mfc/s5p_mfc.c 	src_buf = list_entry(ctx->src_queue.next, struct s5p_mfc_buf, list);
next              423 drivers/media/platform/s5p-mfc/s5p_mfc.c 		src_buf = list_entry(ctx->src_queue.next, struct s5p_mfc_buf,
next              538 drivers/media/platform/s5p-mfc/s5p_mfc.c 			src_buf = list_entry(ctx->src_queue.next,
next              577 drivers/media/platform/s5p-mfc/s5p_mfc.c 				src_buf = list_entry(ctx->src_queue.next,
next              612 drivers/media/platform/s5p-mfc/s5p_mfc.c 		mb_entry = list_entry(ctx->dst_queue.next, struct s5p_mfc_buf,
next             1114 drivers/media/platform/s5p-mfc/s5p_mfc_enc.c 		mb_entry = list_entry((&ctx->ref_queue)->next,
next             1134 drivers/media/platform/s5p-mfc/s5p_mfc_enc.c 	dst_mb = list_entry(ctx->dst_queue.next, struct s5p_mfc_buf, list);
next             1151 drivers/media/platform/s5p-mfc/s5p_mfc_enc.c 			dst_mb = list_entry(ctx->dst_queue.next,
next             1192 drivers/media/platform/s5p-mfc/s5p_mfc_enc.c 	src_mb = list_entry(ctx->src_queue.next, struct s5p_mfc_buf, list);
next             1198 drivers/media/platform/s5p-mfc/s5p_mfc_enc.c 	dst_mb = list_entry(ctx->dst_queue.next, struct s5p_mfc_buf, list);
next             1255 drivers/media/platform/s5p-mfc/s5p_mfc_enc.c 		mb_entry = list_entry(ctx->src_queue.next, struct s5p_mfc_buf,
next             1267 drivers/media/platform/s5p-mfc/s5p_mfc_enc.c 		mb_entry = list_entry(ctx->dst_queue.next, struct s5p_mfc_buf,
next             1179 drivers/media/platform/s5p-mfc/s5p_mfc_opr_v5.c 	temp_vb = list_entry(ctx->src_queue.next, struct s5p_mfc_buf, list);
next             1216 drivers/media/platform/s5p-mfc/s5p_mfc_opr_v5.c 		src_mb = list_entry(ctx->src_queue.next, struct s5p_mfc_buf,
next             1236 drivers/media/platform/s5p-mfc/s5p_mfc_opr_v5.c 	dst_mb = list_entry(ctx->dst_queue.next, struct s5p_mfc_buf, list);
next             1255 drivers/media/platform/s5p-mfc/s5p_mfc_opr_v5.c 	temp_vb = list_entry(ctx->src_queue.next, struct s5p_mfc_buf, list);
next             1274 drivers/media/platform/s5p-mfc/s5p_mfc_opr_v5.c 	dst_mb = list_entry(ctx->dst_queue.next, struct s5p_mfc_buf, list);
next             1300 drivers/media/platform/s5p-mfc/s5p_mfc_opr_v5.c 	temp_vb = list_entry(ctx->src_queue.next, struct s5p_mfc_buf, list);
next             1812 drivers/media/platform/s5p-mfc/s5p_mfc_opr_v6.c 	temp_vb = list_entry(ctx->src_queue.next, struct s5p_mfc_buf, list);
next             1856 drivers/media/platform/s5p-mfc/s5p_mfc_opr_v6.c 		src_mb = list_entry(ctx->src_queue.next, struct s5p_mfc_buf, list);
next             1874 drivers/media/platform/s5p-mfc/s5p_mfc_opr_v6.c 	dst_mb = list_entry(ctx->dst_queue.next, struct s5p_mfc_buf, list);
next             1894 drivers/media/platform/s5p-mfc/s5p_mfc_opr_v6.c 	temp_vb = list_entry(ctx->src_queue.next, struct s5p_mfc_buf, list);
next             1910 drivers/media/platform/s5p-mfc/s5p_mfc_opr_v6.c 	dst_mb = list_entry(ctx->dst_queue.next, struct s5p_mfc_buf, list);
next              306 drivers/media/platform/sh_vou.c 	buf = list_entry(vou_dev->buf_list.next, struct sh_vou_buffer, list);
next              316 drivers/media/platform/sh_vou.c 	buf = list_entry(buf->list.next, struct sh_vou_buffer, list);
next             1067 drivers/media/platform/sh_vou.c 	vou_dev->active = list_entry(vou_dev->buf_list.next,
next             1074 drivers/media/platform/sh_vou.c 		struct sh_vou_buffer *new = list_entry(vou_dev->active->list.next,
next             1347 drivers/media/platform/sh_vou.c 	struct v4l2_subdev *sd = list_entry(v4l2_dev->subdevs.next,
next              204 drivers/media/platform/sti/delta/delta-v4l2.c 	struct delta_dts *next;
next              212 drivers/media/platform/sti/delta/delta-v4l2.c 	list_for_each_entry_safe(dts, next, &ctx->dts, list)
next              250 drivers/media/platform/stm32/stm32-dcmi.c 	buf = list_entry(dcmi->buffers.next, struct dcmi_buf, list);
next              809 drivers/media/platform/stm32/stm32-dcmi.c 	buf = list_entry(dcmi->buffers.next, struct dcmi_buf, list);
next              181 drivers/media/platform/sunxi/sun4i-csi/sun4i_dma.c 	unsigned int next;
next              184 drivers/media/platform/sunxi/sun4i-csi/sun4i_dma.c 	next = !(reg & CSI_BUF_CTRL_DBS);
next              187 drivers/media/platform/sunxi/sun4i-csi/sun4i_dma.c 	sun4i_csi_buffer_mark_done(csi, next, sequence);
next              190 drivers/media/platform/sunxi/sun4i-csi/sun4i_dma.c 	return sun4i_csi_buffer_fill_slot(csi, next);
next              816 drivers/media/platform/ti-vpe/cal.c 	buf = list_entry(dma_q->active.next, struct cal_buffer, list);
next             1287 drivers/media/platform/ti-vpe/cal.c 	buf = list_entry(dma_q->active.next, struct cal_buffer, list);
next              484 drivers/media/platform/ti-vpe/vpdma.c 	list->next = list->buf.addr;
next              498 drivers/media/platform/ti-vpe/vpdma.c 	list->next = list->buf.addr;
next              510 drivers/media/platform/ti-vpe/vpdma.c 	list->next = NULL;
next              533 drivers/media/platform/ti-vpe/vpdma.c 	list_size = (list->next - list->buf.addr) >> 4;
next              636 drivers/media/platform/ti-vpe/vpdma.c 	cfd = list->next;
next              645 drivers/media/platform/ti-vpe/vpdma.c 	list->next = cfd + 1;
next              665 drivers/media/platform/ti-vpe/vpdma.c 	cfd = list->next;
next              674 drivers/media/platform/ti-vpe/vpdma.c 	list->next = cfd + 1;
next              703 drivers/media/platform/ti-vpe/vpdma.c 	ctd = list->next;
next              712 drivers/media/platform/ti-vpe/vpdma.c 	list->next = ctd + 1;
next              728 drivers/media/platform/ti-vpe/vpdma.c 	ctd = list->next;
next              737 drivers/media/platform/ti-vpe/vpdma.c 	list->next = ctd + 1;
next              836 drivers/media/platform/ti-vpe/vpdma.c 	dtd = list->next;
next              855 drivers/media/platform/ti-vpe/vpdma.c 	list->next = dtd + 1;
next              904 drivers/media/platform/ti-vpe/vpdma.c 	dtd = list->next;
next              926 drivers/media/platform/ti-vpe/vpdma.c 	list->next = dtd + 1;
next               27 drivers/media/platform/ti-vpe/vpdma.h 	void *next;
next              336 drivers/media/platform/via-camera.c 	return list_entry(cam->buffer_queue.next, struct via_buffer, queue);
next              270 drivers/media/platform/vivid/vivid-kthread-cap.c 		vid_out_buf = list_entry(dev->vid_out_active.next,
next              696 drivers/media/platform/vivid/vivid-kthread-cap.c 		vid_cap_buf = list_entry(dev->vid_cap_active.next, struct vivid_buffer, list);
next              702 drivers/media/platform/vivid/vivid-kthread-cap.c 			vbi_cap_buf = list_entry(dev->vbi_cap_active.next,
next              932 drivers/media/platform/vivid/vivid-kthread-cap.c 			buf = list_entry(dev->vid_cap_active.next,
next              947 drivers/media/platform/vivid/vivid-kthread-cap.c 			buf = list_entry(dev->vbi_cap_active.next,
next               61 drivers/media/platform/vivid/vivid-kthread-out.c 		vid_out_buf = list_entry(dev->vid_out_active.next,
next               68 drivers/media/platform/vivid/vivid-kthread-out.c 		vbi_out_buf = list_entry(dev->vbi_out_active.next,
next              277 drivers/media/platform/vivid/vivid-kthread-out.c 			buf = list_entry(dev->vid_out_active.next,
next              292 drivers/media/platform/vivid/vivid-kthread-out.c 			buf = list_entry(dev->vbi_out_active.next,
next               97 drivers/media/platform/vivid/vivid-sdr-cap.c 		sdr_cap_buf = list_entry(dev->sdr_cap_active.next,
next              301 drivers/media/platform/vivid/vivid-sdr-cap.c 		buf = list_entry(dev->sdr_cap_active.next,
next              801 drivers/media/platform/vsp1/vsp1_dl.c 		struct vsp1_dl_list *next = list_next_entry(dl, chain);
next              803 drivers/media/platform/vsp1/vsp1_dl.c 		dl->header->next_header = next->dma;
next             1155 drivers/media/platform/vsp1/vsp1_dl.c 	struct vsp1_dl_list *dl, *next;
next             1160 drivers/media/platform/vsp1/vsp1_dl.c 	list_for_each_entry_safe(dl, next, &dlm->free, list) {
next               67 drivers/media/platform/vsp1/vsp1_drm.c 			      struct vsp1_entity *next, unsigned int next_pad)
next               77 drivers/media/platform/vsp1/vsp1_drm.c 		prev->sink = next;
next              108 drivers/media/platform/vsp1/vsp1_drm.c 	uif->sink = next;
next              539 drivers/media/platform/vsp1/vsp1_drm.c 	struct vsp1_entity *next;
next              552 drivers/media/platform/vsp1/vsp1_drm.c 	list_for_each_entry_safe(entity, next, &pipe->entities, list_pipe) {
next              316 drivers/media/platform/vsp1/vsp1_video.c 	struct vsp1_vb2_buffer *next = NULL;
next              334 drivers/media/platform/vsp1/vsp1_video.c 		next = list_first_entry(&video->irqqueue,
next              346 drivers/media/platform/vsp1/vsp1_video.c 	return next;
next              223 drivers/media/usb/airspy/airspy.c 	buf = list_entry(s->queued_bufs.next,
next              448 drivers/media/usb/airspy/airspy.c 		buf = list_entry(s->queued_bufs.next,
next              403 drivers/media/usb/au0828/au0828-video.c 	*buf = list_entry(dma_q->active.next, struct au0828_buffer, list);
next              478 drivers/media/usb/au0828/au0828-video.c 	*buf = list_entry(dma_q->active.next, struct au0828_buffer, list);
next              875 drivers/media/usb/au0828/au0828-video.c 		buf = list_entry(vidq->active.next, struct au0828_buffer, list);
next              905 drivers/media/usb/au0828/au0828-video.c 		buf = list_entry(vbiq->active.next, struct au0828_buffer, list);
next              351 drivers/media/usb/cpia2/cpia2.h 	struct framebuf *next;
next             2262 drivers/media/usb/cpia2/cpia2_core.c 		cam->buffers[i].next = &cam->buffers[i+1];
next             2269 drivers/media/usb/cpia2/cpia2_core.c 	cam->buffers[i].next = cam->buffers;
next             2276 drivers/media/usb/cpia2/cpia2_core.c 	cam->workbuff = cam->curbuff->next;
next              139 drivers/media/usb/cpia2/cpia2_usb.c 		cam->workbuff = cam->workbuff->next;
next              233 drivers/media/usb/cpia2/cpia2_usb.c 			for (ptr = cam->workbuff->next;
next              235 drivers/media/usb/cpia2/cpia2_usb.c 			     ptr = ptr->next)
next              739 drivers/media/usb/cpia2/cpia2_usb.c 		cam->workbuff = cam->curbuff->next;
next             1278 drivers/media/usb/cx231xx/cx231xx-417.c 		buf = list_entry(dma_q->active.next,
next             1334 drivers/media/usb/cx231xx/cx231xx-417.c 	buf = list_entry(dma_q->active.next,
next               88 drivers/media/usb/cx231xx/cx231xx-core.c 	list_add_tail(&ops->next, &cx231xx_extension_devlist);
next              108 drivers/media/usb/cx231xx/cx231xx-core.c 	list_del(&ops->next);
next              119 drivers/media/usb/cx231xx/cx231xx-core.c 		list_for_each_entry(ops, &cx231xx_extension_devlist, next) {
next              133 drivers/media/usb/cx231xx/cx231xx-core.c 		list_for_each_entry(ops, &cx231xx_extension_devlist, next) {
next              614 drivers/media/usb/cx231xx/cx231xx-vbi.c 	*buf = list_entry(dma_q->active.next, struct cx231xx_buffer, vb.queue);
next              244 drivers/media/usb/cx231xx/cx231xx-video.c 	*buf = list_entry(dma_q->active.next, struct cx231xx_buffer, vb.queue);
next              739 drivers/media/usb/cx231xx/cx231xx.h 	struct list_head next;
next             1088 drivers/media/usb/em28xx/em28xx-core.c 	list_add_tail(&ops->next, &em28xx_extension_devlist);
next             1114 drivers/media/usb/em28xx/em28xx-core.c 	list_del(&ops->next);
next             1126 drivers/media/usb/em28xx/em28xx-core.c 	list_for_each_entry(ops, &em28xx_extension_devlist, next) {
next             1141 drivers/media/usb/em28xx/em28xx-core.c 	list_for_each_entry(ops, &em28xx_extension_devlist, next) {
next             1158 drivers/media/usb/em28xx/em28xx-core.c 	list_for_each_entry(ops, &em28xx_extension_devlist, next) {
next             1174 drivers/media/usb/em28xx/em28xx-core.c 	list_for_each_entry(ops, &em28xx_extension_devlist, next) {
next              608 drivers/media/usb/em28xx/em28xx-video.c 	buf = list_entry(dma_q->active.next, struct em28xx_buffer, list);
next             1175 drivers/media/usb/em28xx/em28xx-video.c 		buf = list_entry(vidq->active.next, struct em28xx_buffer, list);
next             1210 drivers/media/usb/em28xx/em28xx-video.c 		buf = list_entry(vbiq->active.next, struct em28xx_buffer, list);
next              787 drivers/media/usb/em28xx/em28xx.h 	struct list_head next;
next              447 drivers/media/usb/hackrf/hackrf.c 	buffer = list_entry(buffer_list->next, struct hackrf_buffer, list);
next              104 drivers/media/usb/hdpvr/hdpvr-video.c 	for (p = q->next; p != q;) {
next              111 drivers/media/usb/hdpvr/hdpvr-video.c 		tmp = p->next;
next              196 drivers/media/usb/hdpvr/hdpvr-video.c 		buf = list_entry(dev->free_buff_list.next, struct hdpvr_buffer,
next              237 drivers/media/usb/hdpvr/hdpvr-video.c 	buf = list_entry(dev->rec_buff_list.next, struct hdpvr_buffer,
next              153 drivers/media/usb/msi2500/msi2500.c 	buf = list_entry(dev->queued_bufs.next, struct msi2500_frame_buf, list);
next              554 drivers/media/usb/msi2500/msi2500.c 		buf = list_entry(dev->queued_bufs.next,
next              382 drivers/media/usb/pvrusb2/pvrusb2-io.c 	while ((lp = sp->queued_list.next) != &sp->queued_list) {
next              549 drivers/media/usb/pvrusb2/pvrusb2-io.c 	struct list_head *lp = sp->idle_list.next;
next              556 drivers/media/usb/pvrusb2/pvrusb2-io.c 	struct list_head *lp = sp->ready_list.next;
next              185 drivers/media/usb/pwc/pwc-if.c 	buf = list_entry(pdev->queued_bufs.next, struct pwc_frame_buf, list);
next              551 drivers/media/usb/pwc/pwc-if.c 		buf = list_entry(pdev->queued_bufs.next, struct pwc_frame_buf,
next              555 drivers/media/usb/s2255/s2255drv.c 	buf = list_entry(vc->buf_list.next,
next              384 drivers/media/usb/stkwebcam/stk-webcam.c 					list_move_tail(dev->sio_avail.next,
next              857 drivers/media/usb/tm6000/tm6000-core.c 		list_for_each_entry(ops, &tm6000_extension_devlist, next) {
next              871 drivers/media/usb/tm6000/tm6000-core.c 	list_add_tail(&ops->next, &tm6000_extension_devlist);
next              891 drivers/media/usb/tm6000/tm6000-core.c 	list_del(&ops->next);
next              902 drivers/media/usb/tm6000/tm6000-core.c 		list_for_each_entry(ops, &tm6000_extension_devlist, next) {
next              916 drivers/media/usb/tm6000/tm6000-core.c 		list_for_each_entry(ops, &tm6000_extension_devlist, next) {
next               91 drivers/media/usb/tm6000/tm6000-video.c 	*buf = list_entry(dma_q->active.next,
next             1435 drivers/media/usb/tm6000/tm6000-video.c 		buf = list_entry(fh->vb_vidq.stream.next, struct tm6000_buffer, vb.stream);
next              272 drivers/media/usb/tm6000/tm6000.h 	struct list_head	next;
next              570 drivers/media/usb/ttusb-dec/ttusb_dec.c 	for (item = dec->filter_info_list.next; item != &dec->filter_info_list;
next              571 drivers/media/usb/ttusb-dec/ttusb_dec.c 	     item = item->next) {
next              780 drivers/media/usb/ttusb-dec/ttusb_dec.c 		if ((item = dec->urb_frame_list.next) != &dec->urb_frame_list) {
next             1606 drivers/media/usb/ttusb-dec/ttusb_dec.c 	while ((item = dec->urb_frame_list.next) != &dec->urb_frame_list) {
next             1624 drivers/media/usb/ttusb-dec/ttusb_dec.c 	while ((item = dec->filter_info_list.next) != &dec->filter_info_list) {
next             1299 drivers/media/usb/usbvision/usbvision-core.c 				(*f) = list_entry(usbvision->inqueue.next,
next              754 drivers/media/usb/usbvision/usbvision-video.c 	f = list_entry(usbvision->outqueue.next,
next              756 drivers/media/usb/usbvision/usbvision-video.c 	list_del(usbvision->outqueue.next);
next              950 drivers/media/usb/usbvision/usbvision-video.c 	frame = list_entry(usbvision->outqueue.next,
next              952 drivers/media/usb/usbvision/usbvision-video.c 	list_del(usbvision->outqueue.next);
next              858 drivers/media/usb/uvc/uvc_ctrl.c 	int next)
next              873 drivers/media/usb/uvc/uvc_ctrl.c 			if ((map->id == v4l2_id) && !next) {
next              880 drivers/media/usb/uvc/uvc_ctrl.c 			    (map->id > v4l2_id) && next) {
next              893 drivers/media/usb/uvc/uvc_ctrl.c 	int next = v4l2_id & V4L2_CTRL_FLAG_NEXT_CTRL;
next              902 drivers/media/usb/uvc/uvc_ctrl.c 		__uvc_find_control(entity, v4l2_id, mapping, &ctrl, next);
next              903 drivers/media/usb/uvc/uvc_ctrl.c 		if (ctrl && !next)
next              907 drivers/media/usb/uvc/uvc_ctrl.c 	if (ctrl == NULL && !next)
next             1512 drivers/media/usb/uvc/uvc_driver.c 		if (forward->chain.next || forward->chain.prev) {
next             1598 drivers/media/usb/uvc/uvc_driver.c 			if (term->chain.next || term->chain.prev) {
next             1657 drivers/media/usb/uvc/uvc_driver.c 		if (entity->chain.next || entity->chain.prev) {
next             1843 drivers/media/usb/uvc/uvc_driver.c 		if (term->chain.next || term->chain.prev)
next              529 drivers/media/usb/zr364xx/zr364xx.c 	buf = list_entry(dma_q->active.next,
next              537 drivers/media/v4l2-core/v4l2-async.c 	if (!notifier || !notifier->asd_list.next)
next              140 drivers/media/v4l2-core/v4l2-compat-ioctl32.c 	compat_caddr_t		next;
next              187 drivers/media/v4l2-core/v4l2-compat-ioctl32.c 		if (put_user(clipcount ? kclips + 1 : NULL, &kclips->next))
next               40 drivers/media/v4l2-core/v4l2-ctrls.c 	u32 next;
next             2234 drivers/media/v4l2-core/v4l2-ctrls.c 		ref = ref->next;
next             2330 drivers/media/v4l2-core/v4l2-ctrls.c 	new_ref->next = hdl->buckets[bucket];
next             3370 drivers/media/v4l2-core/v4l2-ctrls.c 		h->next = 0;
next             3396 drivers/media/v4l2-core/v4l2-ctrls.c 			mref->helper->next = i;
next             3492 drivers/media/v4l2-core/v4l2-ctrls.c 				idx = helpers[idx].next;
next             3859 drivers/media/v4l2-core/v4l2-ctrls.c 				tmp_idx = helpers[tmp_idx].next;
next             3875 drivers/media/v4l2-core/v4l2-ctrls.c 			idx = helpers[idx].next;
next             3896 drivers/media/v4l2-core/v4l2-ctrls.c 				idx = helpers[idx].next;
next               90 drivers/media/v4l2-core/v4l2-device.c 	struct v4l2_subdev *sd, *next;
next               99 drivers/media/v4l2-core/v4l2-device.c 	list_for_each_entry_safe(sd, next, &v4l2_dev->subdevs, list) {
next              684 drivers/media/v4l2-core/videobuf-core.c 	buf = list_entry(q->stream.next, struct videobuf_buffer, stream);
next             1069 drivers/media/v4l2-core/videobuf-core.c 			q->read_buf = list_entry(q->stream.next,
next             1129 drivers/media/v4l2-core/videobuf-core.c 			buf = list_entry(q->stream.next,
next             1137 drivers/media/v4l2-core/videobuf-core.c 			q->read_buf = list_entry(q->stream.next,
next              911 drivers/message/fusion/mptbase.c 		mf = list_entry(ioc->FreeQ.next, MPT_FRAME_HDR,
next             1130 drivers/message/fusion/mptbase.c mpt_add_chain(void *pAddr, u8 next, u16 length, dma_addr_t dma_addr)
next             1136 drivers/message/fusion/mptbase.c 	pChain->NextChainOffset = next;
next             1150 drivers/message/fusion/mptbase.c mpt_add_chain_64bit(void *pAddr, u8 next, u16 length, dma_addr_t dma_addr)
next             1159 drivers/message/fusion/mptbase.c 	pChain->NextChainOffset = next;
next              595 drivers/message/fusion/mptbase.h typedef void (*MPT_ADD_CHAIN)(void *pAddr, u8 next, u16 length,
next              812 drivers/message/fusion/mptbase.h 	void				*next;
next              336 drivers/message/fusion/mptsas.c 	struct fw_event_work *fw_event, *next;
next              357 drivers/message/fusion/mptsas.c 	list_for_each_entry_safe(fw_event, next, &ioc->fw_event_list, list) {
next              557 drivers/message/fusion/mptsas.c 	struct mptsas_device_info	*sas_info, *next;
next              566 drivers/message/fusion/mptsas.c 	list_for_each_entry_safe(sas_info, next, &ioc->sas_device_info_list,
next              660 drivers/message/fusion/mptsas.c 	struct mptsas_device_info	*sas_info, *next;
next              722 drivers/message/fusion/mptsas.c 	list_for_each_entry_safe(sas_info, next, &ioc->sas_device_info_list,
next              792 drivers/message/fusion/mptsas.c 	struct mptsas_device_info	*sas_info, *next;
next              797 drivers/message/fusion/mptsas.c 	list_for_each_entry_safe(sas_info, next, &ioc->sas_device_info_list,
next              812 drivers/message/fusion/mptsas.c 	struct mptsas_device_info	*sas_info, *next;
next              815 drivers/message/fusion/mptsas.c 	list_for_each_entry_safe(sas_info, next, &ioc->sas_device_info_list,
next             1171 drivers/message/fusion/mptsas.c 	target_reset_list = list_entry(head->next,
next             1244 drivers/message/fusion/mptsas.c 	target_reset_list = list_entry(head->next,
next              147 drivers/message/fusion/mptscsih.c 		chainBuf = list_entry(ioc->FreeChainQ.next, MPT_FRAME_HDR,
next             1454 drivers/message/fusion/mptscsih.c 	int next;
next             1465 drivers/message/fusion/mptscsih.c 		next = ioc->ChainToChain[chain_idx];
next             1483 drivers/message/fusion/mptscsih.c 		chain_idx = next;
next             1539 drivers/mfd/ab8500-debugfs.c 			while ((action = action->next) != NULL)
next              102 drivers/mfd/ipaq-micro.c 				micro->msg = list_entry(micro->queue.next,
next               55 drivers/mfd/sta2x11-mfd.c 		return list_entry(sta2x11_mfd_list.next,
next              200 drivers/mfd/twl4030-power.c 					   u8 delay, u8 next)
next              214 drivers/mfd/twl4030-power.c 	err = twl4030_write_script_byte(address++, next);
next              116 drivers/misc/altera-stapl/altera.c 	struct altera_procinfo	*next;
next             2382 drivers/misc/altera-stapl/altera.c 			procptr->next = NULL;
next             2389 drivers/misc/altera-stapl/altera.c 				while (tmpptr->next != NULL)
next             2390 drivers/misc/altera-stapl/altera.c 					tmpptr = tmpptr->next;
next             2391 drivers/misc/altera-stapl/altera.c 				tmpptr->next = procptr;
next             2484 drivers/misc/altera-stapl/altera.c 				proc_list = procptr->next;
next              191 drivers/misc/genwqe/card_dev.c 	struct list_head *node, *next;
next              196 drivers/misc/genwqe/card_dev.c 	list_for_each_safe(node, next, &cfile->map_list) {
next              229 drivers/misc/genwqe/card_dev.c 	struct list_head *node, *next;
next              233 drivers/misc/genwqe/card_dev.c 	list_for_each_safe(node, next, &cfile->pin_list) {
next              212 drivers/misc/habanalabs/command_submission.c 			struct hl_cs *next;
next              220 drivers/misc/habanalabs/command_submission.c 			next = list_first_entry_or_null(
next              224 drivers/misc/habanalabs/command_submission.c 			if ((next) && (!next->tdr_active)) {
next              225 drivers/misc/habanalabs/command_submission.c 				next->tdr_active = true;
next              226 drivers/misc/habanalabs/command_submission.c 				schedule_delayed_work(&next->work_tdr,
next              407 drivers/misc/habanalabs/memory.c 	struct hl_vm_va_block *prev, *next;
next              418 drivers/misc/habanalabs/memory.c 	next = list_next_entry(va_block, node);
next              419 drivers/misc/habanalabs/memory.c 	if (&next->node != va_list && va_block->end + 1 == next->start) {
next              420 drivers/misc/habanalabs/memory.c 		next->start = va_block->start;
next              421 drivers/misc/habanalabs/memory.c 		next->size = next->end - next->start;
next               71 drivers/misc/ibmasm/command.c 	struct list_head *next;
next               76 drivers/misc/ibmasm/command.c 	next = sp->command_queue.next;
next               77 drivers/misc/ibmasm/command.c 	list_del_init(next);
next               78 drivers/misc/ibmasm/command.c 	cmd = list_entry(next, struct command, queue_node);
next              203 drivers/misc/lkdtm/bugs.c 	test_head.next = redirection;
next              228 drivers/misc/lkdtm/bugs.c 	item.node.next = redirection;
next             1021 drivers/misc/mei/bus.c 	struct mei_cl_device *cldev, *next;
next             1024 drivers/misc/mei/bus.c 	list_for_each_entry_safe(cldev, next, &bus->device_list, bus_list)
next              304 drivers/misc/mei/client.c 	struct mei_me_client *me_cl, *next;
next              307 drivers/misc/mei/client.c 	list_for_each_entry_safe(me_cl, next, &dev->me_clients, list)
next              393 drivers/misc/mei/client.c 	struct mei_cl_cb *cb, *next;
next              395 drivers/misc/mei/client.c 	list_for_each_entry_safe(cb, next, head, list) {
next              413 drivers/misc/mei/client.c 	struct mei_cl_cb *cb, *next;
next              415 drivers/misc/mei/client.c 	list_for_each_entry_safe(cb, next, head, list) {
next              429 drivers/misc/mei/client.c 	struct mei_cl_cb *cb, *next;
next              431 drivers/misc/mei/client.c 	list_for_each_entry_safe(cb, next, head, list)
next              861 drivers/misc/mei/hbm.c 	struct mei_cl_cb *cb, *next;
next              864 drivers/misc/mei/hbm.c 	list_for_each_entry_safe(cb, next, &dev->ctrl_rd_list, list) {
next               31 drivers/misc/mei/interrupt.c 	struct mei_cl_cb *cb, *next;
next               34 drivers/misc/mei/interrupt.c 	list_for_each_entry_safe(cb, next, cmpl_list, list) {
next              372 drivers/misc/mei/interrupt.c 	struct mei_cl_cb *cb, *next;
next              390 drivers/misc/mei/interrupt.c 	list_for_each_entry_safe(cb, next, &dev->write_waiting_list, list) {
next              401 drivers/misc/mei/interrupt.c 	list_for_each_entry_safe(cb, next, &dev->ctrl_wr_list, list) {
next              445 drivers/misc/mei/interrupt.c 	list_for_each_entry_safe(cb, next, &dev->write_list, list) {
next              135 drivers/misc/mic/vop/vop_debugfs.c 					   desc->flags, desc->next);
next               64 drivers/misc/sgi-gru/grufile.c 	struct list_head *entry, *next;
next               73 drivers/misc/sgi-gru/grufile.c 	list_for_each_safe(entry, next, &vdata->vd_head) {
next              118 drivers/misc/sgi-gru/grukservices.c 	void 			*next __gru_cacheline_aligned__;/* CL 1 */
next              554 drivers/misc/sgi-gru/grukservices.c 	mq->next = &mq->data;
next              839 drivers/misc/sgi-gru/grukservices.c 	struct message_header *mhdr = mq->next;
next              840 drivers/misc/sgi-gru/grukservices.c 	void *next, *pnext;
next              848 drivers/misc/sgi-gru/grukservices.c 	pnext = mq->next;
next              849 drivers/misc/sgi-gru/grukservices.c 	next = pnext + GRU_CACHE_LINE_BYTES * lines;
next              850 drivers/misc/sgi-gru/grukservices.c 	if (next == mq->limit) {
next              851 drivers/misc/sgi-gru/grukservices.c 		next = mq->start;
next              853 drivers/misc/sgi-gru/grukservices.c 	} else if (pnext < mq->start2 && next >= mq->start2) {
next              859 drivers/misc/sgi-gru/grukservices.c 	mq->next = next;
next              871 drivers/misc/sgi-gru/grukservices.c 	struct message_header *mhdr = mq->next;
next              877 drivers/misc/sgi-gru/grukservices.c 		mhdr = mq->next;
next              233 drivers/misc/sgi-gru/gruprocfs.c 	.next	= seq_next,
next              240 drivers/misc/sgi-gru/gruprocfs.c 	.next	= seq_next,
next              245 drivers/misc/sgi-xp/xpc.h 	struct xpc_fifo_entry_uv *next;
next              268 drivers/misc/sgi-xp/xpc.h 		struct xpc_fifo_entry_uv next;	/* FOR XPC INTERNAL USE ONLY */
next              290 drivers/misc/sgi-xp/xpc.h 	struct xpc_fifo_entry_uv next;
next              976 drivers/misc/sgi-xp/xpc_uv.c 		head->first = first->next;
next              983 drivers/misc/sgi-xp/xpc_uv.c 		first->next = NULL;
next              995 drivers/misc/sgi-xp/xpc_uv.c 	last->next = NULL;
next              998 drivers/misc/sgi-xp/xpc_uv.c 		head->last->next = last;
next             1106 drivers/misc/sgi-xp/xpc_uv.c 					      &msg_slot->next);
next             1352 drivers/misc/sgi-xp/xpc_uv.c 	msg_slot = container_of(entry, struct xpc_send_msg_slot_uv, next);
next             1361 drivers/misc/sgi-xp/xpc_uv.c 	xpc_put_fifo_entry_uv(&ch->sn.uv.msg_slot_free_list, &msg_slot->next);
next             1462 drivers/misc/sgi-xp/xpc_uv.c 	xpc_put_fifo_entry_uv(&ch_uv->recv_msg_list, &msg_slot->hdr.u.next);
next             1657 drivers/misc/sgi-xp/xpc_uv.c 					   hdr.u.next);
next              525 drivers/misc/vmw_vmci/vmci_context.c 	list_item = context->datagram_queue.next;
next              551 drivers/misc/vmw_vmci/vmci_context.c 		list_item = context->datagram_queue.next;
next             1936 drivers/mmc/core/mmc_test.c 	unsigned int next;
next             1947 drivers/mmc/core/mmc_test.c 			next = rnd_next;
next             1951 drivers/mmc/core/mmc_test.c 			rnd_next = next;
next             1959 drivers/mmc/core/mmc_test.c 		next = rnd_next;
next             1963 drivers/mmc/core/mmc_test.c 		rnd_next = next;
next              311 drivers/mmc/core/sdio_cis.c 			this->next = NULL;
next              315 drivers/mmc/core/sdio_cis.c 			prev = &this->next;
next              362 drivers/mmc/core/sdio_cis.c 		tuple = tuple->next;
next              403 drivers/mmc/core/sdio_cis.c 		tuple = tuple->next;
next             1057 drivers/mmc/core/sdio_uart.c 		for (tpl = func->tuples; tpl; tpl = tpl->next) {
next             1574 drivers/mmc/host/atmel-mci.c 		slot = list_entry(host->queue.next,
next             1802 drivers/mmc/host/dw_mmc.c 		slot = list_entry(host->queue.next,
next              465 drivers/mmc/host/mmci.c int mmci_prep_data(struct mmci_host *host, struct mmc_data *data, bool next)
next              472 drivers/mmc/host/mmci.c 	err = host->ops->prep_data(host, data, next);
next              474 drivers/mmc/host/mmci.c 	if (next && !err)
next              853 drivers/mmc/host/mmci.c 			bool next)
next              861 drivers/mmc/host/mmci.c 	if (next)
next              888 drivers/mmc/host/mmci.c 	struct mmci_dmae_next *next = &dmae->next_data;
next              893 drivers/mmc/host/mmci.c 	WARN_ON(!data->host_cookie && (next->desc || next->chan));
next              895 drivers/mmc/host/mmci.c 	dmae->desc_current = next->desc;
next              896 drivers/mmc/host/mmci.c 	dmae->cur = next->chan;
next              897 drivers/mmc/host/mmci.c 	next->desc = NULL;
next              898 drivers/mmc/host/mmci.c 	next->chan = NULL;
next              913 drivers/mmc/host/mmci.c 		struct mmci_dmae_next *next = &dmae->next_data;
next              921 drivers/mmc/host/mmci.c 		if (dmae->desc_current == next->desc)
next              924 drivers/mmc/host/mmci.c 		if (dmae->cur == next->chan) {
next              929 drivers/mmc/host/mmci.c 		next->desc = NULL;
next              930 drivers/mmc/host/mmci.c 		next->chan = NULL;
next              357 drivers/mmc/host/mmci.h 			 bool next);
next              435 drivers/mmc/host/mmci.h 			bool next);
next               77 drivers/mmc/host/mmci_stm32_sdmmc.c 				struct mmc_data *data, bool next)
next               80 drivers/mmc/host/mmci_stm32_sdmmc.c 	if (!next && data->host_cookie == host->next_cookie)
next              311 drivers/mmc/host/mtk-sd.c 	u32 next;
next              329 drivers/mmc/host/mtk-sd.c 	u32 next;
next             1637 drivers/mmc/host/mtk-sd.c 	gpd->next = lower_32_bits(dma_addr);
next             1649 drivers/mmc/host/mtk-sd.c 		bd[i].next = lower_32_bits(dma_addr);
next             1200 drivers/mmc/host/omap_hsmmc.c 				       struct omap_hsmmc_next *next,
next             1205 drivers/mmc/host/omap_hsmmc.c 	if (!next && data->host_cookie &&
next             1214 drivers/mmc/host/omap_hsmmc.c 	if (next || data->host_cookie != host->next_data.cookie) {
next             1227 drivers/mmc/host/omap_hsmmc.c 	if (next) {
next             1228 drivers/mmc/host/omap_hsmmc.c 		next->dma_len = dma_len;
next             1229 drivers/mmc/host/omap_hsmmc.c 		data->host_cookie = ++next->cookie < 0 ? 1 : next->cookie;
next              486 drivers/mmc/host/usdhi6rol0.c 		struct scatterlist *next = sg_next(host->sg);
next              490 drivers/mmc/host/usdhi6rol0.c 		if (!next)
next              492 drivers/mmc/host/usdhi6rol0.c 		host->sg = next;
next              494 drivers/mmc/host/usdhi6rol0.c 		if (WARN(next && sg_dma_len(next) % data->blksz,
next              496 drivers/mmc/host/usdhi6rol0.c 			 sg_dma_len(next), data->blksz))
next              477 drivers/mtd/devices/block2mtd.c 	struct list_head *pos, *next;
next              480 drivers/mtd/devices/block2mtd.c 	list_for_each_safe(pos, next, &blkmtd_device_list) {
next              220 drivers/mtd/devices/ms02-nv.c 	mp->next = root_ms02nv_mtd;
next              251 drivers/mtd/devices/ms02-nv.c 	root_ms02nv_mtd = mp->next;
next               91 drivers/mtd/devices/ms02-nv.h 	struct mtd_info *next;
next               62 drivers/mtd/devices/slram.c 	struct slram_mtd_list *next;
next              135 drivers/mtd/devices/slram.c 		curmtd = &(*curmtd)->next;
next              144 drivers/mtd/devices/slram.c 	(*curmtd)->next = NULL;
next              204 drivers/mtd/devices/slram.c 		nextitem = slram_mtdlist->next;
next              524 drivers/mtd/mtd_blkdevs.c 	struct mtd_blktrans_dev *dev, *next;
next              527 drivers/mtd/mtd_blkdevs.c 		list_for_each_entry_safe(dev, next, &tr->devs, list)
next              556 drivers/mtd/mtd_blkdevs.c 	if (!blktrans_notifier.list.next)
next              588 drivers/mtd/mtd_blkdevs.c 	struct mtd_blktrans_dev *dev, *next;
next              595 drivers/mtd/mtd_blkdevs.c 	list_for_each_entry_safe(dev, next, &tr->devs, list)
next              609 drivers/mtd/mtd_blkdevs.c 	if (blktrans_notifier.list.next)
next              639 drivers/mtd/mtdpart.c 	struct mtd_part *child, *next;
next              642 drivers/mtd/mtdpart.c 	list_for_each_entry_safe(child, next, &mtd_partitions, list) {
next              668 drivers/mtd/mtdpart.c 	struct mtd_part *slave, *next;
next              672 drivers/mtd/mtdpart.c 	list_for_each_entry_safe(slave, next, &mtd_partitions, list)
next              685 drivers/mtd/mtdpart.c 	struct mtd_part *slave, *next;
next              689 drivers/mtd/mtdpart.c 	list_for_each_entry_safe(slave, next, &mtd_partitions, list)
next               60 drivers/mtd/parsers/cmdlinepart.c 	struct cmdline_mtd_partition *next;
next              267 drivers/mtd/parsers/cmdlinepart.c 		this_mtd->next = partitions;
next              316 drivers/mtd/parsers/cmdlinepart.c 	for (part = partitions; part; part = part->next) {
next               33 drivers/mtd/parsers/redboot.c 	struct fis_list *next;
next              218 drivers/mtd/parsers/redboot.c 			prev = &(*prev)->next;
next              219 drivers/mtd/parsers/redboot.c 		new_fl->next = *prev;
next              230 drivers/mtd/parsers/redboot.c 	for (tmp_fl = fl; tmp_fl->next; tmp_fl = tmp_fl->next) {
next              231 drivers/mtd/parsers/redboot.c 		if (tmp_fl->img->flash_base + tmp_fl->img->size + master->erasesize <= tmp_fl->next->img->flash_base) {
next              278 drivers/mtd/parsers/redboot.c 		if(fl->next && fl->img->flash_base + fl->img->size + master->erasesize <= fl->next->img->flash_base) {
next              281 drivers/mtd/parsers/redboot.c 			parts[i].size = fl->next->img->flash_base - parts[i].offset;
next              286 drivers/mtd/parsers/redboot.c 		fl = fl->next;
next              294 drivers/mtd/parsers/redboot.c 		fl = fl->next;
next             1111 drivers/mtd/spi-nor/spi-nor.c 	struct spi_nor_erase_command *cmd, *next;
next             1113 drivers/mtd/spi-nor/spi-nor.c 	list_for_each_entry_safe(cmd, next, erase_list, list) {
next             1201 drivers/mtd/spi-nor/spi-nor.c 	struct spi_nor_erase_command *cmd, *next;
next             1208 drivers/mtd/spi-nor/spi-nor.c 	list_for_each_entry_safe(cmd, next, &erase_list, list) {
next              825 drivers/mtd/ubi/attach.c 		aeb = list_entry(ai->free.next, struct ubi_ainf_peb, u.list);
next              657 drivers/mtd/ubi/block.c 	struct ubiblock *next;
next              661 drivers/mtd/ubi/block.c 	list_for_each_entry_safe(dev, next, &ubiblock_devices, list) {
next              454 drivers/mtd/ubi/debug.c 	.next = eraseblk_count_seq_next,
next              209 drivers/mtd/ubi/wl.c 	wrk = list_entry(ubi->works.next, struct ubi_work, list);
next             1679 drivers/mtd/ubi/wl.c 		wrk = list_entry(ubi->works.next, struct ubi_work, list);
next              133 drivers/net/appletalk/ipddp.c         for(rt = ipddp_route_list; rt != NULL; rt = rt->next)
next              204 drivers/net/appletalk/ipddp.c         rt->next = NULL;
next              217 drivers/net/appletalk/ipddp.c         rt->next = ipddp_route_list;
next              241 drivers/net/appletalk/ipddp.c                         *r = tmp->next;
next              246 drivers/net/appletalk/ipddp.c                 r = &tmp->next;
next              260 drivers/net/appletalk/ipddp.c         for(f = ipddp_route_list; f != NULL; f = f->next)
next              333 drivers/net/appletalk/ipddp.c                 p = ipddp_route_list->next;
next               21 drivers/net/appletalk/ipddp.h         struct ipddp_route *next;
next              269 drivers/net/appletalk/ltpc.c 	struct xmitQel *next;
next              287 drivers/net/appletalk/ltpc.c 	qel->next = NULL;
next              291 drivers/net/appletalk/ltpc.c 		xmQtl->next = qel;
next              311 drivers/net/appletalk/ltpc.c 		xmQhd = qel->next;
next               83 drivers/net/bonding/bond_alb.c 	entry->next = TLB_NULL_INDEX;
next              106 drivers/net/bonding/bond_alb.c 			u32 next_index = tx_hash_table[index].next;
next              208 drivers/net/bonding/bond_alb.c 			hash_table[hash_index].next = next_index;
next              808 drivers/net/bonding/bond_alb.c 	u32 next;
next              811 drivers/net/bonding/bond_alb.c 	next = bond_info->rx_hashtbl[ip_src_hash].src_first;
next              812 drivers/net/bonding/bond_alb.c 	bond_info->rx_hashtbl[ip_dst_hash].src_next = next;
next              813 drivers/net/bonding/bond_alb.c 	if (next != RLB_NULL_INDEX)
next              814 drivers/net/bonding/bond_alb.c 		bond_info->rx_hashtbl[next].src_prev = ip_dst_hash;
next             3442 drivers/net/bonding/bond_main.c 	struct net_device *ldev, *next, *now, *dev_stack[MAX_NEST_DEV + 1];
next             3450 drivers/net/bonding/bond_main.c 		next = NULL;
next             3456 drivers/net/bonding/bond_main.c 			next = ldev;
next             3465 drivers/net/bonding/bond_main.c 		if (!next) {
next             3468 drivers/net/bonding/bond_main.c 			next = dev_stack[--cur];
next             3472 drivers/net/bonding/bond_main.c 		now = next;
next              257 drivers/net/bonding/bond_procfs.c 	.next  = bond_info_seq_next,
next               23 drivers/net/dsa/bcm_sf2_cfp.c 	struct list_head next;
next              547 drivers/net/dsa/bcm_sf2_cfp.c 	list_for_each_entry(rule, &priv->cfp.rules_list, next) {
next              565 drivers/net/dsa/bcm_sf2_cfp.c 	list_for_each_entry(rule, &priv->cfp.rules_list, next) {
next              909 drivers/net/dsa/bcm_sf2_cfp.c 	list_add_tail(&rule->next, &priv->cfp.rules_list);
next              989 drivers/net/dsa/bcm_sf2_cfp.c 	list_del(&rule->next);
next             1164 drivers/net/dsa/bcm_sf2_cfp.c 	list_for_each_entry_safe_reverse(rule, n, &priv->cfp.rules_list, next)
next             1186 drivers/net/dsa/bcm_sf2_cfp.c 	list_for_each_entry(rule, &priv->cfp.rules_list, next) {
next              281 drivers/net/ethernet/3com/3c515.c 	u32 next;
next              294 drivers/net/ethernet/3com/3c515.c 	u32 next;
next              339 drivers/net/ethernet/3com/3c515.c 		next:8;			/* The media type to try next. */
next              732 drivers/net/ethernet/3com/3c515.c 			dev->if_port = media_tbl[dev->if_port].next;
next              823 drivers/net/ethernet/3com/3c515.c 				vp->rx_ring[i].next =
next              826 drivers/net/ethernet/3com/3c515.c 				vp->rx_ring[i].next = 0;
next              837 drivers/net/ethernet/3com/3c515.c 			vp->rx_ring[i - 1].next =
next              923 drivers/net/ethernet/3com/3c515.c 				    media_tbl[dev->if_port].next;
next             1029 drivers/net/ethernet/3com/3c515.c 		vp->tx_ring[entry].next = 0;
next             1041 drivers/net/ethernet/3com/3c515.c 			prev_entry->next = isa_virt_to_bus(&vp->tx_ring[entry]);
next             1572 drivers/net/ethernet/3com/3c515.c 		vp = list_entry(root_corkscrew_dev.next,
next              544 drivers/net/ethernet/3com/3c59x.c 	__le32 next;					/* Last entry points to 0.   */
next              564 drivers/net/ethernet/3com/3c59x.c 	__le32 next;					/* Last entry points to 0.   */
next              724 drivers/net/ethernet/3com/3c59x.c 		next:8;						/* The media type to try next. */
next             1582 drivers/net/ethernet/3com/3c59x.c 				dev->if_port = media_tbl[dev->if_port].next;
next             1736 drivers/net/ethernet/3com/3c59x.c 			vp->rx_ring[i].next = cpu_to_le32(vp->rx_ring_dma + sizeof(struct boom_rx_desc) * (i+1));
next             1759 drivers/net/ethernet/3com/3c59x.c 		vp->rx_ring[i-1].next = cpu_to_le32(vp->rx_ring_dma);
next             1839 drivers/net/ethernet/3com/3c59x.c 			dev->if_port = media_tbl[dev->if_port].next;
next             2145 drivers/net/ethernet/3com/3c59x.c 	vp->tx_ring[entry].next = 0;
next             2214 drivers/net/ethernet/3com/3c59x.c 	prev_entry->next = cpu_to_le32(vp->tx_ring_dma + entry * sizeof(struct boom_tx_desc));
next               20 drivers/net/ethernet/8390/8390.h   unsigned char next;   /* pointer to next packet. */
next             1381 drivers/net/ethernet/8390/axnet_cs.c 				  rx_frame.next);
next             1415 drivers/net/ethernet/8390/axnet_cs.c 				  rx_frame.status, rx_frame.next,
next             1422 drivers/net/ethernet/8390/axnet_cs.c 		next_frame = rx_frame.next;
next              703 drivers/net/ethernet/8390/lib8390.c 		if (rx_frame.next != next_frame &&
next              704 drivers/net/ethernet/8390/lib8390.c 		    rx_frame.next != next_frame + 1 &&
next              705 drivers/net/ethernet/8390/lib8390.c 		    rx_frame.next != next_frame - num_rx_pages &&
next              706 drivers/net/ethernet/8390/lib8390.c 		    rx_frame.next != next_frame + 1 - num_rx_pages) {
next              717 drivers/net/ethernet/8390/lib8390.c 				  rx_frame.next);
next              745 drivers/net/ethernet/8390/lib8390.c 				  rx_frame.status, rx_frame.next,
next              752 drivers/net/ethernet/8390/lib8390.c 		next_frame = rx_frame.next;
next              202 drivers/net/ethernet/8390/stnic.c   hdr->next = buf[0] >> 8;
next              211 drivers/net/ethernet/8390/stnic.c 	    ring_page, hdr->status, hdr->next, hdr->count);
next              350 drivers/net/ethernet/agere/et131x.c 	struct tcb *next;	/* Next entry in ring */
next             1764 drivers/net/ethernet/agere/et131x.c 		tcb->next = tcb + 1;
next             1770 drivers/net/ethernet/agere/et131x.c 	tcb->next = NULL;
next             2020 drivers/net/ethernet/agere/et131x.c 		rfd = list_entry(rx_ring->recv_list.next,
next             2142 drivers/net/ethernet/agere/et131x.c 		struct fbr_desc *next;
next             2149 drivers/net/ethernet/agere/et131x.c 		next = (struct fbr_desc *)(fbr->ring_virtaddr) +
next             2156 drivers/net/ethernet/agere/et131x.c 		next->addr_hi = fbr->bus_high[buff_index];
next             2157 drivers/net/ethernet/agere/et131x.c 		next->addr_lo = fbr->bus_low[buff_index];
next             2158 drivers/net/ethernet/agere/et131x.c 		next->word2 = buff_index;
next             2248 drivers/net/ethernet/agere/et131x.c 	element = rx_local->recv_list.next;
next             2563 drivers/net/ethernet/agere/et131x.c 		tx_ring->send_tail->next = tcb;
next             2569 drivers/net/ethernet/agere/et131x.c 	WARN_ON(tcb->next != NULL);
next             2608 drivers/net/ethernet/agere/et131x.c 	tx_ring->tcb_qhead = tcb->next;
next             2616 drivers/net/ethernet/agere/et131x.c 	tcb->next = NULL;
next             2624 drivers/net/ethernet/agere/et131x.c 			tx_ring->tcb_qtail->next = tcb;
next             2684 drivers/net/ethernet/agere/et131x.c 		tx_ring->tcb_qtail->next = tcb;
next             2708 drivers/net/ethernet/agere/et131x.c 		struct tcb *next = tcb->next;
next             2710 drivers/net/ethernet/agere/et131x.c 		tx_ring->send_head = next;
next             2712 drivers/net/ethernet/agere/et131x.c 		if (next == NULL)
next             2761 drivers/net/ethernet/agere/et131x.c 		tx_ring->send_head = tcb->next;
next             2762 drivers/net/ethernet/agere/et131x.c 		if (tcb->next == NULL)
next             2776 drivers/net/ethernet/agere/et131x.c 		tx_ring->send_head = tcb->next;
next             2777 drivers/net/ethernet/agere/et131x.c 		if (tcb->next == NULL)
next              683 drivers/net/ethernet/alteon/acenic.h 	struct net_device	*next;
next              320 drivers/net/ethernet/altera/altera_sgdma.c 	csrwr32(lower_32_bits(ndesc_phys), desc, sgdma_descroffs(next));
next              427 drivers/net/ethernet/altera/altera_sgdma.c 			entry = list_entry((list)->next, type, member);	\
next              436 drivers/net/ethernet/altera/altera_sgdma.c 			entry = list_entry((list)->next, type, member);	\
next               15 drivers/net/ethernet/altera/altera_sgdmahw.h 	u32	next;
next              769 drivers/net/ethernet/amd/amd8111e.h 	struct net_device *next;
next              251 drivers/net/ethernet/amd/declance.c 	struct net_device *next;
next             1048 drivers/net/ethernet/amd/declance.c 			dev = lp->next;
next             1266 drivers/net/ethernet/amd/declance.c 		lp->next = root_lance_dev;
next             1309 drivers/net/ethernet/amd/declance.c 		root_lance_dev = lp->next;
next              298 drivers/net/ethernet/amd/pcnet32.c 	struct net_device	*next;
next             1991 drivers/net/ethernet/amd/pcnet32.c 		lp->next = pcnet32_dev;
next             3036 drivers/net/ethernet/amd/pcnet32.c 		next_dev = lp->next;
next              818 drivers/net/ethernet/apple/bmac.c 	unsigned int counter, cur = curval, next = nxtval;
next              822 drivers/net/ethernet/apple/bmac.c 	next = ((next & 0x00FF) << 8) | (next >> 8);
next              832 drivers/net/ethernet/apple/bmac.c 		if ((next & 0x0001) == 0) low_data_set = 0;
next              835 drivers/net/ethernet/apple/bmac.c 		next = next >> 1;
next              535 drivers/net/ethernet/apple/mace.c     int fill, next, len;
next              540 drivers/net/ethernet/apple/mace.c     next = fill + 1;
next              541 drivers/net/ethernet/apple/mace.c     if (next >= N_TX_RING)
next              542 drivers/net/ethernet/apple/mace.c 	next = 0;
next              543 drivers/net/ethernet/apple/mace.c     if (next == mp->tx_empty) {
next              562 drivers/net/ethernet/apple/mace.c     np = mp->tx_cmds + NCMDS_TX * next;
next              567 drivers/net/ethernet/apple/mace.c     mp->tx_fill = next;
next              575 drivers/net/ethernet/apple/mace.c     if (++next >= N_TX_RING)
next              576 drivers/net/ethernet/apple/mace.c 	next = 0;
next              577 drivers/net/ethernet/apple/mace.c     if (next == mp->tx_empty)
next              880 drivers/net/ethernet/apple/mace.c     int i, nb, stat, next;
next              892 drivers/net/ethernet/apple/mace.c 	    next = i + 1;
next              893 drivers/net/ethernet/apple/mace.c 	    if (next >= N_RX_RING)
next              894 drivers/net/ethernet/apple/mace.c 		next = 0;
next              895 drivers/net/ethernet/apple/mace.c 	    np = mp->rx_cmds + next;
next              896 drivers/net/ethernet/apple/mace.c 	    if (next != mp->rx_fill &&
next              949 drivers/net/ethernet/apple/mace.c 	next = i + 1;
next              950 drivers/net/ethernet/apple/mace.c 	if (next >= N_RX_RING)
next              951 drivers/net/ethernet/apple/mace.c 	    next = 0;
next              952 drivers/net/ethernet/apple/mace.c 	if (next == mp->rx_empty)
next              975 drivers/net/ethernet/apple/mace.c 	i = next;
next              308 drivers/net/ethernet/aquantia/atlantic/aq_ring.c 				next_ = buff_->next,
next              330 drivers/net/ethernet/aquantia/atlantic/aq_ring.c 					next_ = buff_->next,
next              390 drivers/net/ethernet/aquantia/atlantic/aq_ring.c 					next_ = buff_->next,
next               43 drivers/net/ethernet/aquantia/atlantic/aq_ring.h 			u16 next;
next              705 drivers/net/ethernet/aquantia/atlantic/hw_atl/hw_atl_a0.c 				buff->next = 0U;
next              709 drivers/net/ethernet/aquantia/atlantic/hw_atl/hw_atl_a0.c 				buff->next = aq_ring_next_dx(ring,
next              759 drivers/net/ethernet/aquantia/atlantic/hw_atl/hw_atl_b0.c 			buff->next = 0U;
next              769 drivers/net/ethernet/aquantia/atlantic/hw_atl/hw_atl_b0.c 				buff->next = rxd_wb->next_desc_ptr;
next              773 drivers/net/ethernet/aquantia/atlantic/hw_atl/hw_atl_b0.c 				buff->next =
next              243 drivers/net/ethernet/atheros/ag71xx.c 	u32 next;
next              997 drivers/net/ethernet/atheros/ag71xx.c 		desc->next = (u32)(ring->descs_dma +
next             1070 drivers/net/ethernet/atheros/ag71xx.c 		desc->next = (u32)(ring->descs_dma +
next             1074 drivers/net/ethernet/atheros/ag71xx.c 			  desc, desc->next);
next             1440 drivers/net/ethernet/atheros/ag71xx.c 	struct sk_buff *next, *skb;
next             1483 drivers/net/ethernet/atheros/ag71xx.c 			goto next;
next             1498 drivers/net/ethernet/atheros/ag71xx.c next:
next             1507 drivers/net/ethernet/atheros/ag71xx.c 	list_for_each_entry_safe(skb, next, &rx_list, list)
next             1737 drivers/net/ethernet/atheros/ag71xx.c 	ag->stop_desc->next = (u32)ag->stop_desc_dma;
next               78 drivers/net/ethernet/atheros/alx/main.c 	u16 cur, next, count = 0;
next               80 drivers/net/ethernet/atheros/alx/main.c 	next = cur = rxq->write_idx;
next               81 drivers/net/ethernet/atheros/alx/main.c 	if (++next == alx->rx_ringsz)
next               82 drivers/net/ethernet/atheros/alx/main.c 		next = 0;
next               85 drivers/net/ethernet/atheros/alx/main.c 	while (!cur_buf->skb && next != rxq->read_idx) {
next              125 drivers/net/ethernet/atheros/alx/main.c 		cur = next;
next              126 drivers/net/ethernet/atheros/alx/main.c 		if (++next == alx->rx_ringsz)
next              127 drivers/net/ethernet/atheros/alx/main.c 			next = 0;
next             1853 drivers/net/ethernet/atheros/atlx/atl1.c 			goto next;
next             1878 drivers/net/ethernet/atheros/atlx/atl1.c next:
next              290 drivers/net/ethernet/aurora/nb8800.c 	unsigned int next;
next              299 drivers/net/ethernet/aurora/nb8800.c 		next = (last + 1) % RX_DESC_COUNT;
next              301 drivers/net/ethernet/aurora/nb8800.c 		rxd = &priv->rx_descs[next];
next              311 drivers/net/ethernet/aurora/nb8800.c 			nb8800_receive(dev, next, len);
next              320 drivers/net/ethernet/aurora/nb8800.c 		last = next;
next              339 drivers/net/ethernet/aurora/nb8800.c 		if (priv->rx_descs[next].report)
next              396 drivers/net/ethernet/aurora/nb8800.c 	unsigned int next;
next              423 drivers/net/ethernet/aurora/nb8800.c 	next = priv->tx_next;
next              424 drivers/net/ethernet/aurora/nb8800.c 	txb = &priv->tx_bufs[next];
next              425 drivers/net/ethernet/aurora/nb8800.c 	txd = &priv->tx_descs[next];
next              428 drivers/net/ethernet/aurora/nb8800.c 	next = (next + 1) % TX_DESC_COUNT;
next              442 drivers/net/ethernet/aurora/nb8800.c 	desc->n_addr = priv->tx_bufs[next].dma_desc;
next              461 drivers/net/ethernet/aurora/nb8800.c 	priv->tx_next = next;
next              767 drivers/net/ethernet/broadcom/bcmsysport.c 			goto next;
next              786 drivers/net/ethernet/broadcom/bcmsysport.c 			goto next;
next              794 drivers/net/ethernet/broadcom/bcmsysport.c 			goto next;
next              804 drivers/net/ethernet/broadcom/bcmsysport.c 			goto next;
next              836 drivers/net/ethernet/broadcom/bcmsysport.c next:
next              534 drivers/net/ethernet/broadcom/bnx2x/bnx2x_init.h 	u64 next;
next              921 drivers/net/ethernet/broadcom/bnx2x/bnx2x_init_ops.h 		t2[i].next = (u64)(t2_mapping +
next              610 drivers/net/ethernet/broadcom/bnx2x/bnx2x_sp.c 	u8 *next = base;
next              622 drivers/net/ethernet/broadcom/bnx2x/bnx2x_sp.c 			memcpy(next, &pos->u, size);
next              625 drivers/net/ethernet/broadcom/bnx2x/bnx2x_sp.c 			   counter, next);
next              626 drivers/net/ethernet/broadcom/bnx2x/bnx2x_sp.c 			next += stride + size;
next              780 drivers/net/ethernet/broadcom/bnxt/bnxt.c 	u16 next, max = rxr->rx_agg_bmap_size;
next              782 drivers/net/ethernet/broadcom/bnxt/bnxt.c 	next = find_next_zero_bit(rxr->rx_agg_bmap, max, idx);
next              783 drivers/net/ethernet/broadcom/bnxt/bnxt.c 	if (next >= max)
next              784 drivers/net/ethernet/broadcom/bnxt/bnxt.c 		next = find_first_zero_bit(rxr->rx_agg_bmap, max);
next              785 drivers/net/ethernet/broadcom/bnxt/bnxt.c 	return next;
next              657 drivers/net/ethernet/broadcom/cnic.c 			    u32 next)
next              661 drivers/net/ethernet/broadcom/cnic.c 	id_tbl->next = next;
next              699 drivers/net/ethernet/broadcom/cnic.c 	id = find_next_zero_bit(id_tbl->table, id_tbl->max, id_tbl->next);
next              702 drivers/net/ethernet/broadcom/cnic.c 		if (id_tbl->next != 0) {
next              703 drivers/net/ethernet/broadcom/cnic.c 			id = find_first_zero_bit(id_tbl->table, id_tbl->next);
next              704 drivers/net/ethernet/broadcom/cnic.c 			if (id >= id_tbl->next)
next              711 drivers/net/ethernet/broadcom/cnic.c 		id_tbl->next = (id + 1) & (id_tbl->max - 1);
next             1009 drivers/net/ethernet/broadcom/cnic.c 		struct bnx2x_bd_chain_next *next =
next             1015 drivers/net/ethernet/broadcom/cnic.c 		next->addr_hi = (u64) info->dma.pg_map_arr[j] >> 32;
next             1016 drivers/net/ethernet/broadcom/cnic.c 		next->addr_lo = info->dma.pg_map_arr[j] & 0xffffffff;
next             5757 drivers/net/ethernet/broadcom/cnic.c 		udev = list_entry(cnic_udev_list.next, struct cnic_uio_dev,
next              144 drivers/net/ethernet/broadcom/cnic.h 	u32		next;
next             1793 drivers/net/ethernet/broadcom/genet/bcmgenet.c 			goto next;
next             1822 drivers/net/ethernet/broadcom/genet/bcmgenet.c 			goto next;
next             1843 drivers/net/ethernet/broadcom/genet/bcmgenet.c 			goto next;
next             1880 drivers/net/ethernet/broadcom/genet/bcmgenet.c next:
next             7903 drivers/net/ethernet/broadcom/tg3.c 		segs = segs->next;
next             7904 drivers/net/ethernet/broadcom/tg3.c 		nskb->next = NULL;
next              686 drivers/net/ethernet/brocade/bna/bnad.c 			goto next;
next              719 drivers/net/ethernet/brocade/bna/bnad.c next:
next             1045 drivers/net/ethernet/cavium/liquidio/octeon_device.c 		while (dispatch->next != dispatch) {
next             1046 drivers/net/ethernet/cavium/liquidio/octeon_device.c 			temp = dispatch->next;
next              615 drivers/net/ethernet/cavium/liquidio/octeon_network.h 	if (root->prev == root && root->next == root)
next              618 drivers/net/ethernet/cavium/liquidio/octeon_network.h 		node = root->next;
next               97 drivers/net/ethernet/cavium/thunder/nicvf_queues.c 	struct pgcache *pgcache, *next;
next              166 drivers/net/ethernet/cavium/thunder/nicvf_queues.c 	next = &rbdr->pgcache[rbdr->pgidx];
next              167 drivers/net/ethernet/cavium/thunder/nicvf_queues.c 	page = next->page;
next              595 drivers/net/ethernet/cavium/thunder/nicvf_queues.c 			goto next;
next              599 drivers/net/ethernet/cavium/thunder/nicvf_queues.c 			goto next;
next              617 drivers/net/ethernet/cavium/thunder/nicvf_queues.c next:
next               83 drivers/net/ethernet/chelsio/cxgb3/cxgb3_defs.h 	if ((void *)e->next >= (void *)t->tid_tab &&
next               84 drivers/net/ethernet/chelsio/cxgb3/cxgb3_defs.h 	    (void *)e->next < (void *)&t->atid_tab[t->natids])
next              102 drivers/net/ethernet/chelsio/cxgb3/cxgb3_defs.h 	if ((void *)e->next >= (void *)t->tid_tab &&
next              103 drivers/net/ethernet/chelsio/cxgb3/cxgb3_defs.h 	    (void *)e->next < (void *)&t->atid_tab[t->natids])
next              508 drivers/net/ethernet/chelsio/cxgb3/cxgb3_offload.c 	p->next = t->afree;
next              527 drivers/net/ethernet/chelsio/cxgb3/cxgb3_offload.c 	p->next = t->sfree;
next              663 drivers/net/ethernet/chelsio/cxgb3/cxgb3_offload.c 		t->afree = p->next;
next              685 drivers/net/ethernet/chelsio/cxgb3/cxgb3_offload.c 		t->sfree = p->next;
next             1187 drivers/net/ethernet/chelsio/cxgb3/cxgb3_offload.c 			t->stid_tab[nstids - 1].next = &t->stid_tab[nstids];
next             1192 drivers/net/ethernet/chelsio/cxgb3/cxgb3_offload.c 			t->atid_tab[natids - 1].next = &t->atid_tab[natids];
next              144 drivers/net/ethernet/chelsio/cxgb3/cxgb3_offload.h 	union listen_entry *next;
next              149 drivers/net/ethernet/chelsio/cxgb3/cxgb3_offload.h 	union active_open_entry *next;
next              243 drivers/net/ethernet/chelsio/cxgb3/l2t.c 		for (p = &d->l2tab[hash].first; *p; p = &(*p)->next)
next              245 drivers/net/ethernet/chelsio/cxgb3/l2t.c 				*p = e->next;
next              334 drivers/net/ethernet/chelsio/cxgb3/l2t.c 	for (e = d->l2tab[hash].first; e; e = e->next)
next              347 drivers/net/ethernet/chelsio/cxgb3/l2t.c 		e->next = d->l2tab[hash].first;
next              409 drivers/net/ethernet/chelsio/cxgb3/l2t.c 	for (e = d->l2tab[hash].first; e; e = e->next)
next               66 drivers/net/ethernet/chelsio/cxgb3/l2t.h 	struct l2t_entry *next;	/* next l2t_entry on chain */
next               90 drivers/net/ethernet/chelsio/cxgb4/cxgb4_debugfs.c 	.next  = seq_tab_next,
next             1072 drivers/net/ethernet/chelsio/cxgb4/cxgb4_debugfs.c 	.next  = devlog_next,
next             1216 drivers/net/ethernet/chelsio/cxgb4/cxgb4_debugfs.c 	.next  = mboxlog_next,
next             1920 drivers/net/ethernet/chelsio/cxgb4/cxgb4_debugfs.c 	.next  = mps_tcam_next,
next             2552 drivers/net/ethernet/chelsio/cxgb4/cxgb4_debugfs.c 	.next  = dcb_info_next,
next             3038 drivers/net/ethernet/chelsio/cxgb4/cxgb4_debugfs.c 	.next  = sge_queue_next,
next             1224 drivers/net/ethernet/chelsio/cxgb4/cxgb4_main.c 		t->afree = p->next;
next             1241 drivers/net/ethernet/chelsio/cxgb4/cxgb4_main.c 	p->next = t->afree;
next             1498 drivers/net/ethernet/chelsio/cxgb4/cxgb4_main.c 			t->atid_tab[natids - 1].next = &t->atid_tab[natids];
next              216 drivers/net/ethernet/chelsio/cxgb4/cxgb4_tc_u32.c 		const struct cxgb4_next_header *next;
next              229 drivers/net/ethernet/chelsio/cxgb4/cxgb4_tc_u32.c 		next = is_ipv6 ? cxgb4_ipv6_jumps : cxgb4_ipv4_jumps;
next              232 drivers/net/ethernet/chelsio/cxgb4/cxgb4_tc_u32.c 		for (i = 0; next[i].jump; i++) {
next              233 drivers/net/ethernet/chelsio/cxgb4/cxgb4_tc_u32.c 			if (next[i].offoff != cls->knode.sel->offoff ||
next              234 drivers/net/ethernet/chelsio/cxgb4/cxgb4_tc_u32.c 			    next[i].shift != cls->knode.sel->offshift ||
next              235 drivers/net/ethernet/chelsio/cxgb4/cxgb4_tc_u32.c 			    next[i].mask != cls->knode.sel->offmask ||
next              236 drivers/net/ethernet/chelsio/cxgb4/cxgb4_tc_u32.c 			    next[i].offset != cls->knode.sel->off)
next              248 drivers/net/ethernet/chelsio/cxgb4/cxgb4_tc_u32.c 				if (next[i].match_off == off &&
next              249 drivers/net/ethernet/chelsio/cxgb4/cxgb4_tc_u32.c 				    next[i].match_val == val &&
next              250 drivers/net/ethernet/chelsio/cxgb4/cxgb4_tc_u32.c 				    next[i].match_mask == mask) {
next              270 drivers/net/ethernet/chelsio/cxgb4/cxgb4_tc_u32.c 			link->match_field = next[i].jump;
next               89 drivers/net/ethernet/chelsio/cxgb4/cxgb4_uld.h 	union aopen_entry *next;
next              287 drivers/net/ethernet/chelsio/cxgb4/l2t.c 		for (p = &d->l2tab[e->hash].first; *p; p = &(*p)->next)
next              289 drivers/net/ethernet/chelsio/cxgb4/l2t.c 				*p = e->next;
next              290 drivers/net/ethernet/chelsio/cxgb4/l2t.c 				e->next = NULL;
next              329 drivers/net/ethernet/chelsio/cxgb4/l2t.c 		for (p = &d->l2tab[e->hash].first; *p; p = &(*p)->next)
next              331 drivers/net/ethernet/chelsio/cxgb4/l2t.c 				*p = e->next;
next              332 drivers/net/ethernet/chelsio/cxgb4/l2t.c 				e->next = NULL;
next              444 drivers/net/ethernet/chelsio/cxgb4/l2t.c 	for (e = d->l2tab[hash].first; e; e = e->next)
next              468 drivers/net/ethernet/chelsio/cxgb4/l2t.c 		e->next = d->l2tab[hash].first;
next              545 drivers/net/ethernet/chelsio/cxgb4/l2t.c 	for (e = d->l2tab[hash].first; e; e = e->next)
next              734 drivers/net/ethernet/chelsio/cxgb4/l2t.c 	.next = l2t_seq_next,
next               80 drivers/net/ethernet/chelsio/cxgb4/l2t.h 	struct l2t_entry *next;     /* next l2t_entry on chain */
next             2023 drivers/net/ethernet/chelsio/cxgb4vf/cxgb4vf_main.c 	.next  = mboxlog_next,
next             2177 drivers/net/ethernet/chelsio/cxgb4vf/cxgb4vf_main.c 	.next  = sge_queue_next,
next             2323 drivers/net/ethernet/chelsio/cxgb4vf/cxgb4vf_main.c 	.next  = sge_qstats_next,
next             2421 drivers/net/ethernet/chelsio/cxgb4vf/cxgb4vf_main.c 	.next  = interfaces_next,
next              135 drivers/net/ethernet/chelsio/libcxgb/libcxgb_ppm.c 				    pool->next, count, 0);
next              137 drivers/net/ethernet/chelsio/libcxgb/libcxgb_ppm.c 		pool->next = 0;
next              142 drivers/net/ethernet/chelsio/libcxgb/libcxgb_ppm.c 	pool->next = i + count;
next              143 drivers/net/ethernet/chelsio/libcxgb/libcxgb_ppm.c 	if (pool->next >= ppm->pool_index_max)
next              144 drivers/net/ethernet/chelsio/libcxgb/libcxgb_ppm.c 		pool->next = 0;
next              150 drivers/net/ethernet/chelsio/libcxgb/libcxgb_ppm.c 		pool->next);
next              165 drivers/net/ethernet/chelsio/libcxgb/libcxgb_ppm.c 				    ppm->next, count, 0);
next              167 drivers/net/ethernet/chelsio/libcxgb/libcxgb_ppm.c 		ppm->next = 0;
next              174 drivers/net/ethernet/chelsio/libcxgb/libcxgb_ppm.c 	ppm->next = i + count;
next              175 drivers/net/ethernet/chelsio/libcxgb/libcxgb_ppm.c 	if (ppm->max_index_in_edram && (ppm->next >= ppm->max_index_in_edram))
next              176 drivers/net/ethernet/chelsio/libcxgb/libcxgb_ppm.c 		ppm->next = 0;
next              177 drivers/net/ethernet/chelsio/libcxgb/libcxgb_ppm.c 	else if (ppm->next >= ppm->bmap_index_max)
next              178 drivers/net/ethernet/chelsio/libcxgb/libcxgb_ppm.c 		ppm->next = 0;
next              183 drivers/net/ethernet/chelsio/libcxgb/libcxgb_ppm.c 		 __func__, i, count, i + ppm->pool_rsvd, ppm->next,
next              207 drivers/net/ethernet/chelsio/libcxgb/libcxgb_ppm.c 		if (i < pool->next)
next              208 drivers/net/ethernet/chelsio/libcxgb/libcxgb_ppm.c 			pool->next = i;
next              212 drivers/net/ethernet/chelsio/libcxgb/libcxgb_ppm.c 			 __func__, cpu, i, pool->next);
next              219 drivers/net/ethernet/chelsio/libcxgb/libcxgb_ppm.c 		if (i < ppm->next)
next              220 drivers/net/ethernet/chelsio/libcxgb/libcxgb_ppm.c 			ppm->next = i;
next              223 drivers/net/ethernet/chelsio/libcxgb/libcxgb_ppm.c 		pr_debug("%s: idx %d, next %u.\n", __func__, i, ppm->next);
next              271 drivers/net/ethernet/chelsio/libcxgb/libcxgb_ppm.c 			 nr_pages, npods, ppm->next, caller_data);
next              479 drivers/net/ethernet/chelsio/libcxgb/libcxgb_ppm.c 	ppm->next = 0;
next              123 drivers/net/ethernet/chelsio/libcxgb/libcxgb_ppm.h 	unsigned int next;		/* next possible free index */
next              145 drivers/net/ethernet/chelsio/libcxgb/libcxgb_ppm.h 	unsigned int next;
next              844 drivers/net/ethernet/cisco/enic/enic_main.c 		while (!buf->os_buf && (buf->next != wq->to_clean)) {
next              849 drivers/net/ethernet/cisco/enic/enic_main.c 		wq->to_use = buf->next;
next               50 drivers/net/ethernet/cisco/enic/vnic_rq.c 				buf->next = rq->bufs[0];
next               53 drivers/net/ethernet/cisco/enic/vnic_rq.c 				buf->next = rq->bufs[i + 1];
next               55 drivers/net/ethernet/cisco/enic/vnic_rq.c 				buf->next = buf + 1;
next              196 drivers/net/ethernet/cisco/enic/vnic_rq.c 		buf = buf->next;
next               69 drivers/net/ethernet/cisco/enic/vnic_rq.h 	struct vnic_rq_buf *next;
next              132 drivers/net/ethernet/cisco/enic/vnic_rq.h 	buf = buf->next;
next              183 drivers/net/ethernet/cisco/enic/vnic_rq.h 		rq->to_clean = buf->next;
next               50 drivers/net/ethernet/cisco/enic/vnic_wq.c 				buf->next = wq->bufs[0];
next               51 drivers/net/ethernet/cisco/enic/vnic_wq.c 				buf->next->prev = buf;
next               54 drivers/net/ethernet/cisco/enic/vnic_wq.c 				buf->next = wq->bufs[i + 1];
next               55 drivers/net/ethernet/cisco/enic/vnic_wq.c 				buf->next->prev = buf;
next               57 drivers/net/ethernet/cisco/enic/vnic_wq.c 				buf->next = buf + 1;
next               58 drivers/net/ethernet/cisco/enic/vnic_wq.c 				buf->next->prev = buf;
next              206 drivers/net/ethernet/cisco/enic/vnic_wq.c 		buf = wq->to_clean = buf->next;
next               54 drivers/net/ethernet/cisco/enic/vnic_wq.h 	struct vnic_wq_buf *next;
next              148 drivers/net/ethernet/cisco/enic/vnic_wq.h 	buf = buf->next;
next              169 drivers/net/ethernet/cisco/enic/vnic_wq.h 		wq->to_clean = buf->next;
next              557 drivers/net/ethernet/dec/tulip/de2104x.c 			goto next;
next              562 drivers/net/ethernet/dec/tulip/de2104x.c 			goto next;
next              591 drivers/net/ethernet/dec/tulip/de2104x.c next:
next              762 drivers/net/ethernet/dec/tulip/de4x5.c     __le32 next;
next             1189 drivers/net/ethernet/dec/tulip/de4x5.c 	    lp->rx_ring[i].next = 0;
next             1207 drivers/net/ethernet/dec/tulip/de4x5.c 	    		lp->rx_ring[i].next = 0;
next              389 drivers/net/ethernet/dec/tulip/tulip.h 	struct mediainfo *next;
next              341 drivers/net/ethernet/dlink/dl2k.c 	int cid, next;
next              374 drivers/net/ethernet/dlink/dl2k.c 		next = psib[i++];
next              375 drivers/net/ethernet/dlink/dl2k.c 		if ((cid == 0 && next == 0) || (cid == 0xff && next == 0xff)) {
next              392 drivers/net/ethernet/dlink/dl2k.c 			j = (next - i > 255) ? 255 : next - i;
next              404 drivers/net/ethernet/dlink/dl2k.c 		i = next;
next               76 drivers/net/ethernet/ec_bhf.c 	__le32 next;
next              379 drivers/net/ethernet/ec_bhf.c 		u32 next;
next              382 drivers/net/ethernet/ec_bhf.c 			next = (u8 *)(desc + 1) - priv->rx_buf.buf;
next              384 drivers/net/ethernet/ec_bhf.c 			next = 0;
next              385 drivers/net/ethernet/ec_bhf.c 		next |= RXHDR_NEXT_VALID;
next              386 drivers/net/ethernet/ec_bhf.c 		desc->header.next = cpu_to_le32(next);
next             1320 drivers/net/ethernet/fealnx.c 		struct fealnx_desc *next;
next             1330 drivers/net/ethernet/fealnx.c 		next = np->cur_tx_copy->next_desc_logical;
next             1331 drivers/net/ethernet/fealnx.c 		next->skbuff = skb;
next             1332 drivers/net/ethernet/fealnx.c 		next->control = TXIC | TXLD | CRCEnable | PADEnable;
next             1333 drivers/net/ethernet/fealnx.c 		next->control |= (skb->len << PKTSShift);	/* pkt size */
next             1334 drivers/net/ethernet/fealnx.c 		next->control |= ((skb->len - BPT) << TBSShift);	/* buf size */
next             1338 drivers/net/ethernet/fealnx.c 		next->buffer = pci_map_single(ep->pci_dev, skb->data + BPT,
next             1341 drivers/net/ethernet/fealnx.c 		next->status = TXOWN;
next             1344 drivers/net/ethernet/fealnx.c 		np->cur_tx_copy = next->next_desc_logical;
next             1493 drivers/net/ethernet/fealnx.c 				struct fealnx_desc *next;
next             1495 drivers/net/ethernet/fealnx.c 				next = np->cur_tx->next_desc_logical;
next             1496 drivers/net/ethernet/fealnx.c 				tx_status = next->status;
next             1497 drivers/net/ethernet/fealnx.c 				tx_control = next->control;
next              218 drivers/net/ethernet/freescale/fman/fman_mac.h 		hash_entry = ETH_HASH_ENTRY_OBJ(addr_lst->next);
next             2188 drivers/net/ethernet/freescale/gianfar.c 	struct txbd8 *bdp, *next = NULL;
next             2232 drivers/net/ethernet/freescale/gianfar.c 			next = next_txbd(bdp, base, tx_ring_size);
next             2233 drivers/net/ethernet/freescale/gianfar.c 			buflen = be16_to_cpu(next->length) +
next             2251 drivers/net/ethernet/freescale/gianfar.c 			bdp = next;
next              196 drivers/net/ethernet/freescale/ucc_geth.c 		struct list_head *node = lh->next;
next              280 drivers/net/ethernet/hisilicon/hisi_femac.c 			goto next;
next              287 drivers/net/ethernet/hisilicon/hisi_femac.c next:
next              547 drivers/net/ethernet/hisilicon/hix5hd2_gmac.c 			goto next;
next              554 drivers/net/ethernet/hisilicon/hix5hd2_gmac.c next:
next              125 drivers/net/ethernet/hisilicon/hns/hns_dsaf_ppe.c 		ppe_cb->next = NULL;
next               74 drivers/net/ethernet/hisilicon/hns/hns_dsaf_ppe.h 	struct hns_ppe_cb *next;	/* pointer to next ppe device */
next              205 drivers/net/ethernet/hisilicon/hns3/hnae3.h 	struct hnae3_ring_chain_node *next;
next              438 drivers/net/ethernet/hisilicon/hns3/hns3_enet.c 			ring = ring->next;
next             2726 drivers/net/ethernet/hisilicon/hns3/hns3_enet.c 				ring->tail_skb->next = new_skb;
next             3206 drivers/net/ethernet/hisilicon/hns3/hns3_enet.c 		cur_chain->next = NULL;
next             3208 drivers/net/ethernet/hisilicon/hns3/hns3_enet.c 		while (tx_ring->next) {
next             3209 drivers/net/ethernet/hisilicon/hns3/hns3_enet.c 			tx_ring = tx_ring->next;
next             3216 drivers/net/ethernet/hisilicon/hns3/hns3_enet.c 			cur_chain->next = chain;
next             3231 drivers/net/ethernet/hisilicon/hns3/hns3_enet.c 		cur_chain->next = NULL;
next             3238 drivers/net/ethernet/hisilicon/hns3/hns3_enet.c 		rx_ring = rx_ring->next;
next             3246 drivers/net/ethernet/hisilicon/hns3/hns3_enet.c 		cur_chain->next = chain;
next             3255 drivers/net/ethernet/hisilicon/hns3/hns3_enet.c 		rx_ring = rx_ring->next;
next             3261 drivers/net/ethernet/hisilicon/hns3/hns3_enet.c 	cur_chain = head->next;
next             3263 drivers/net/ethernet/hisilicon/hns3/hns3_enet.c 		chain = cur_chain->next;
next             3267 drivers/net/ethernet/hisilicon/hns3/hns3_enet.c 	head->next = NULL;
next             3278 drivers/net/ethernet/hisilicon/hns3/hns3_enet.c 	chain = head->next;
next             3281 drivers/net/ethernet/hisilicon/hns3/hns3_enet.c 		chain_tmp = chain->next;
next             3290 drivers/net/ethernet/hisilicon/hns3/hns3_enet.c 	ring->next = group->ring;
next              405 drivers/net/ethernet/hisilicon/hns3/hns3_enet.h 	struct hns3_enet_ring *next;
next              636 drivers/net/ethernet/hisilicon/hns3/hns3_enet.h 	for (pos = (head).ring; pos; pos = pos->next)
next             4508 drivers/net/ethernet/hisilicon/hns3/hns3pf/hclge_main.c 	for (node = ring_chain; node; node = node->next) {
next              113 drivers/net/ethernet/hisilicon/hns3/hns3pf/hclge_mbx.c 	chain = head->next;
next              116 drivers/net/ethernet/hisilicon/hns3/hns3pf/hclge_mbx.c 		chain_tmp = chain->next;
next              176 drivers/net/ethernet/hisilicon/hns3/hns3pf/hclge_mbx.c 		cur_chain->next = new_chain;
next             1008 drivers/net/ethernet/hisilicon/hns3/hns3vf/hclgevf_main.c 	for (node = ring_chain; node; node = node->next) {
next             1031 drivers/net/ethernet/hisilicon/hns3/hns3vf/hclgevf_main.c 		    !node->next) {
next             1174 drivers/net/ethernet/hp/hp100.c 			lp->rxring[i].next = ringptr;
next             1183 drivers/net/ethernet/hp/hp100.c 			lp->txring[i].next = ringptr;
next             1359 drivers/net/ethernet/hp/hp100.c 		lp->rxrtail = ringptr->next;
next             1492 drivers/net/ethernet/hp/hp100.c 	if (lp->txrtail->next == lp->txrhead) {
next             1544 drivers/net/ethernet/hp/hp100.c 	lp->txrtail = ringptr->next;
next             1617 drivers/net/ethernet/hp/hp100.c 		lp->txrhead = lp->txrhead->next;
next             1945 drivers/net/ethernet/hp/hp100.c 		lp->rxrhead = lp->rxrhead->next;
next             1956 drivers/net/ethernet/hp/hp100.c 			lp->rxrtail = lp->rxrtail->next;
next              544 drivers/net/ethernet/hp/hp100.h 	struct hp100_ring *next;
next              307 drivers/net/ethernet/huawei/hinic/hinic_rx.c 			curr_skb->next = skb;
next              212 drivers/net/ethernet/i825xx/82596.c 	struct i596_tbd *next;
next             1080 drivers/net/ethernet/i825xx/82596.c 		tbd->next = I596_NULL;
next              446 drivers/net/ethernet/i825xx/ether1.c 	int i, status, addr, next, next2;
next              479 drivers/net/ethernet/i825xx/ether1.c 		next = addr + RFD_SIZE + RBD_SIZE + ETH_FRAME_LEN + 10;
next              480 drivers/net/ethernet/i825xx/ether1.c 		next2 = next + RFD_SIZE + RBD_SIZE + ETH_FRAME_LEN + 10;
next              483 drivers/net/ethernet/i825xx/ether1.c 			next = RX_AREA_START;
next              492 drivers/net/ethernet/i825xx/ether1.c 		init_rfd.rfd_link = next;
next              493 drivers/net/ethernet/i825xx/ether1.c 		init_rbd.rbd_link = next + RFD_SIZE;
next              498 drivers/net/ethernet/i825xx/ether1.c 		addr = next;
next              178 drivers/net/ethernet/i825xx/lib82596.c 	u32            next;
next             1002 drivers/net/ethernet/i825xx/lib82596.c 		tbd->next = I596_NULL;
next              598 drivers/net/ethernet/i825xx/sun3_82586.c 		p->xmit_buffs[i]->next = 0xffff;
next              650 drivers/net/ethernet/i825xx/sun3_82586.c 		rfd[i].next = make16(rfd + (i+1) % (p->num_recv_buffs+rfdadd) );
next              665 drivers/net/ethernet/i825xx/sun3_82586.c 		rbd[i].next = make16((rbd + (i+1) % p->num_recv_buffs));
next              808 drivers/net/ethernet/i825xx/sun3_82586.c 						rbd = (struct rbd_struct *) make32(rbd->next);
next              826 drivers/net/ethernet/i825xx/sun3_82586.c 		p->rfd_top = (struct rfd_struct *) make32(p->rfd_top->next); /* step to next RFD */
next              866 drivers/net/ethernet/i825xx/sun3_82586.c 			rfds = (struct rfd_struct *) make32(rfds->next);
next              134 drivers/net/ethernet/i825xx/sun3_82586.h   unsigned short next;		/* linkoffset to next RFD */
next              165 drivers/net/ethernet/i825xx/sun3_82586.h   unsigned short next;		/* pointeroffset to next RBD */
next              310 drivers/net/ethernet/i825xx/sun3_82586.h   unsigned short next;          /* pointeroffset to next TBD */
next             1779 drivers/net/ethernet/ibm/emac/core.c 			goto next;
next             1785 drivers/net/ethernet/ibm/emac/core.c 			goto next;
next             1810 drivers/net/ethernet/ibm/emac/core.c 	next:
next             1850 drivers/net/ethernet/ibm/emac/core.c 		goto next;
next             2224 drivers/net/ethernet/ibm/ibmvnic.c 		union sub_crq *next;
next             2238 drivers/net/ethernet/ibm/ibmvnic.c 		next = ibmvnic_next_scrq(adapter, adapter->rx_scrq[scrq_num]);
next             2240 drivers/net/ethernet/ibm/ibmvnic.c 		    (struct ibmvnic_rx_buff *)be64_to_cpu(next->
next             2243 drivers/net/ethernet/ibm/ibmvnic.c 		if (next->rx_comp.rc) {
next             2245 drivers/net/ethernet/ibm/ibmvnic.c 				   be16_to_cpu(next->rx_comp.rc));
next             2247 drivers/net/ethernet/ibm/ibmvnic.c 			next->rx_comp.first = 0;
next             2253 drivers/net/ethernet/ibm/ibmvnic.c 			next->rx_comp.first = 0;
next             2258 drivers/net/ethernet/ibm/ibmvnic.c 		length = be32_to_cpu(next->rx_comp.len);
next             2259 drivers/net/ethernet/ibm/ibmvnic.c 		offset = be16_to_cpu(next->rx_comp.off_frame_data);
next             2260 drivers/net/ethernet/ibm/ibmvnic.c 		flags = next->rx_comp.flags;
next             2271 drivers/net/ethernet/ibm/ibmvnic.c 					       ntohs(next->rx_comp.vlan_tci));
next             2274 drivers/net/ethernet/ibm/ibmvnic.c 		next->rx_comp.first = 0;
next             2907 drivers/net/ethernet/ibm/ibmvnic.c 	union sub_crq *next;
next             2916 drivers/net/ethernet/ibm/ibmvnic.c 		next = ibmvnic_next_scrq(adapter, scrq);
next             2917 drivers/net/ethernet/ibm/ibmvnic.c 		for (i = 0; i < next->tx_comp.num_comps; i++) {
next             2918 drivers/net/ethernet/ibm/ibmvnic.c 			if (next->tx_comp.rcs[i]) {
next             2920 drivers/net/ethernet/ibm/ibmvnic.c 					next->tx_comp.rcs[i]);
next             2923 drivers/net/ethernet/ibm/ibmvnic.c 			index = be32_to_cpu(next->tx_comp.correlators[i]);
next             2953 drivers/net/ethernet/ibm/ibmvnic.c 		next->tx_comp.first = 0;
next              420 drivers/net/ethernet/intel/e100.c 	struct rx *next, *prev;
next              499 drivers/net/ethernet/intel/e100.c 	struct cb *next, *prev;
next              862 drivers/net/ethernet/intel/e100.c 	nic->cb_to_use = cb->next;
next              895 drivers/net/ethernet/intel/e100.c 			nic->cb_to_send = nic->cb_to_send->next;
next             1823 drivers/net/ethernet/intel/e100.c 	    cb = nic->cb_to_clean = cb->next) {
next             1867 drivers/net/ethernet/intel/e100.c 			nic->cb_to_clean = nic->cb_to_clean->next;
next             1894 drivers/net/ethernet/intel/e100.c 		cb->next = (i + 1 < count) ? cb + 1 : nic->cbs;
next             2062 drivers/net/ethernet/intel/e100.c 	for (rx = nic->rx_to_clean; rx->skb; rx = nic->rx_to_clean = rx->next) {
next             2083 drivers/net/ethernet/intel/e100.c 	for (rx = nic->rx_to_use; !rx->skb; rx = nic->rx_to_use = rx->next) {
next             2165 drivers/net/ethernet/intel/e100.c 		rx->next = (i + 1 < count) ? rx + 1 : nic->rxs;
next              101 drivers/net/ethernet/intel/fm10k/fm10k_debugfs.c 	.next  = fm10k_dbg_desc_seq_next,
next              108 drivers/net/ethernet/intel/fm10k/fm10k_debugfs.c 	.next  = fm10k_dbg_desc_seq_next,
next             2010 drivers/net/ethernet/intel/i40e/i40e_main.c struct i40e_new_mac_filter *i40e_next_filter(struct i40e_new_mac_filter *next)
next             2012 drivers/net/ethernet/intel/i40e/i40e_main.c 	hlist_for_each_entry_continue(next, hlist) {
next             2013 drivers/net/ethernet/intel/i40e/i40e_main.c 		if (!is_broadcast_ether_addr(next->f->macaddr))
next             2014 drivers/net/ethernet/intel/i40e/i40e_main.c 			return next;
next             4168 drivers/net/ethernet/intel/i40e/i40e_main.c 	tx_ring->next = q_vector->tx.ring;
next             4177 drivers/net/ethernet/intel/i40e/i40e_main.c 		xdp_ring->next = q_vector->tx.ring;
next             4183 drivers/net/ethernet/intel/i40e/i40e_main.c 	rx_ring->next = q_vector->rx.ring;
next             4649 drivers/net/ethernet/intel/i40e/i40e_main.c 				u32 next;
next             4665 drivers/net/ethernet/intel/i40e/i40e_main.c 				next = (val & I40E_QINT_TQCTL_NEXTQ_INDX_MASK)
next             4677 drivers/net/ethernet/intel/i40e/i40e_main.c 				qp = next;
next              353 drivers/net/ethernet/intel/i40e/i40e_txrx.h 	struct i40e_ring *next;		/* pointer to next ring in q_vector */
next              469 drivers/net/ethernet/intel/i40e/i40e_txrx.h 	for (pos = (head).ring; pos != NULL; pos = pos->next)
next              309 drivers/net/ethernet/intel/iavf/iavf_main.c 	rx_ring->next = q_vector->rx.ring;
next              335 drivers/net/ethernet/intel/iavf/iavf_main.c 	tx_ring->next = q_vector->tx.ring;
next              327 drivers/net/ethernet/intel/iavf/iavf_txrx.h 	struct iavf_ring *next;		/* pointer to next ring in q_vector */
next              426 drivers/net/ethernet/intel/iavf/iavf_txrx.h 	for (pos = (head).ring; pos != NULL; pos = pos->next)
next              397 drivers/net/ethernet/intel/ice/ice_lib.c 	int next;
next              400 drivers/net/ethernet/intel/ice/ice_lib.c 		next = curr + 1;
next              407 drivers/net/ethernet/intel/ice/ice_lib.c 			next = ICE_NO_VSI;
next              409 drivers/net/ethernet/intel/ice/ice_lib.c 			next = i;
next              411 drivers/net/ethernet/intel/ice/ice_lib.c 	return next;
next             1380 drivers/net/ethernet/intel/ice/ice_lib.c 			tx_ring->next = q_vector->tx.ring;
next             1396 drivers/net/ethernet/intel/ice/ice_lib.c 			rx_ring->next = q_vector->rx.ring;
next              167 drivers/net/ethernet/intel/ice/ice_txrx.h 	struct ice_ring *next;		/* pointer to next ring in q_vector */
next              231 drivers/net/ethernet/intel/ice/ice_txrx.h 	for (pos = (head).ring; pos; pos = pos->next)
next              204 drivers/net/ethernet/intel/igb/igb_main.c 	.next		= NULL,
next              309 drivers/net/ethernet/intel/ixgbe/ixgbe.h 	struct ixgbe_ring *next;	/* pointer to next ring in q_vector */
next              438 drivers/net/ethernet/intel/ixgbe/ixgbe.h 	for (pos = (head).ring; pos != NULL; pos = pos->next)
next              809 drivers/net/ethernet/intel/ixgbe/ixgbe_lib.c 	ring->next = head->ring;
next              143 drivers/net/ethernet/intel/ixgbe/ixgbe_main.c 	.next          = NULL,
next               91 drivers/net/ethernet/intel/ixgbevf/ixgbevf.h 	struct ixgbevf_ring *next;
next              226 drivers/net/ethernet/intel/ixgbevf/ixgbevf.h 	for (pos = (head).ring; pos != NULL; pos = pos->next)
next             2676 drivers/net/ethernet/intel/ixgbevf/ixgbevf_main.c 	ring->next = head->ring;
next              375 drivers/net/ethernet/korina.c 			goto next;
next              392 drivers/net/ethernet/korina.c 			goto next;
next              424 drivers/net/ethernet/korina.c next:
next             1623 drivers/net/ethernet/marvell/octeontx2/af/rvu_nix.c 			 u16 pcifunc, int next, bool eol)
next             1638 drivers/net/ethernet/marvell/octeontx2/af/rvu_nix.c 	aq_req.mce.next = next;
next             1746 drivers/net/ethernet/marvell/octeontx2/af/rvu_nix.c 			next_mce = hlist_entry(mce->node.next,
next             1346 drivers/net/ethernet/marvell/octeontx2/af/rvu_npc.c 	u16 index, next, end;
next             1356 drivers/net/ethernet/marvell/octeontx2/af/rvu_npc.c 	next = find_next_bit(map, end, index);
next             1357 drivers/net/ethernet/marvell/octeontx2/af/rvu_npc.c 	if (*max_area < (next - index)) {
next             1358 drivers/net/ethernet/marvell/octeontx2/af/rvu_npc.c 		*max_area = next - index;
next             1362 drivers/net/ethernet/marvell/octeontx2/af/rvu_npc.c 	if (next < end) {
next             1363 drivers/net/ethernet/marvell/octeontx2/af/rvu_npc.c 		start = next + 1;
next             1375 drivers/net/ethernet/marvell/octeontx2/af/rvu_npc.c 	u16 index, next;
next             1386 drivers/net/ethernet/marvell/octeontx2/af/rvu_npc.c 	next = find_next_bit(map, end, index);
next             1387 drivers/net/ethernet/marvell/octeontx2/af/rvu_npc.c 	if (next <= end) {
next             1388 drivers/net/ethernet/marvell/octeontx2/af/rvu_npc.c 		fcnt += next - index;
next             1389 drivers/net/ethernet/marvell/octeontx2/af/rvu_npc.c 		start = next + 1;
next              829 drivers/net/ethernet/marvell/octeontx2/af/rvu_struct.h 	uint64_t next       : 16;
next              843 drivers/net/ethernet/marvell/octeontx2/af/rvu_struct.h 	uint64_t next       : 16;
next              922 drivers/net/ethernet/marvell/skge.c 			e->next = ring->start;
next              925 drivers/net/ethernet/marvell/skge.c 			e->next = e + 1;
next              999 drivers/net/ethernet/marvell/skge.c 	} while ((e = e->next) != ring->start);
next             1026 drivers/net/ethernet/marvell/skge.c 	} while ((e = e->next) != ring->start);
next             2793 drivers/net/ethernet/marvell/skge.c 			e = e->next;
next             2820 drivers/net/ethernet/marvell/skge.c 	skge->tx_ring.to_use = e->next;
next             2837 drivers/net/ethernet/marvell/skge.c 		e = e->next;
next             2873 drivers/net/ethernet/marvell/skge.c 	for (e = skge->tx_ring.to_clean; e != skge->tx_ring.to_use; e = e->next) {
next             3155 drivers/net/ethernet/marvell/skge.c 	for (e = ring->to_clean; e != ring->to_use; e = e->next) {
next             3205 drivers/net/ethernet/marvell/skge.c 	for (e = ring->to_clean; prefetch(e->next), work_done < budget; e = e->next) {
next             3702 drivers/net/ethernet/marvell/skge.c 	for (e = skge->tx_ring.to_clean; e != skge->tx_ring.to_use; e = e->next) {
next             3710 drivers/net/ethernet/marvell/skge.c 	for (e = skge->rx_ring.to_clean; ; e = e->next) {
next             2395 drivers/net/ethernet/marvell/skge.h 	struct skge_element	*next;
next             1549 drivers/net/ethernet/mediatek/mtk_eth_soc.c 		int next = (i + 1) % MTK_DMA_SIZE;
next             1550 drivers/net/ethernet/mediatek/mtk_eth_soc.c 		u32 next_ptr = ring->phys + next * sz;
next              292 drivers/net/ethernet/mellanox/mlx4/alloc.c 			struct mlx4_zone_entry *next = list_first_entry(&entry->list,
next              293 drivers/net/ethernet/mellanox/mlx4/alloc.c 									typeof(*next),
next              296 drivers/net/ethernet/mellanox/mlx4/alloc.c 			if (next->priority == entry->priority)
next              297 drivers/net/ethernet/mellanox/mlx4/alloc.c 				list_add_tail(&next->prio_list, &entry->prio_list);
next              130 drivers/net/ethernet/mellanox/mlx4/cmd.c 	int			next;
next              352 drivers/net/ethernet/mellanox/mlx4/cmd.c 	cmd->free_head = context->next;
next              399 drivers/net/ethernet/mellanox/mlx4/cmd.c 	context->next = cmd->free_head;
next              692 drivers/net/ethernet/mellanox/mlx4/cmd.c 	cmd->free_head = context->next;
next              766 drivers/net/ethernet/mellanox/mlx4/cmd.c 	context->next = cmd->free_head;
next             2640 drivers/net/ethernet/mellanox/mlx4/cmd.c 		priv->cmd.context[i].next  = i + 1;
next             2648 drivers/net/ethernet/mellanox/mlx4/cmd.c 	priv->cmd.context[priv->cmd.max_cmds - 1].next = -1;
next              160 drivers/net/ethernet/mellanox/mlx4/en_netdev.c 	struct list_head next;
next              312 drivers/net/ethernet/mellanox/mlx4/en_netdev.c 	list_add_tail(&filter->next, &priv->filters);
next              325 drivers/net/ethernet/mellanox/mlx4/en_netdev.c 	list_del(&filter->next);
next              425 drivers/net/ethernet/mellanox/mlx4/en_netdev.c 	list_for_each_entry_safe(filter, tmp, &priv->filters, next) {
next              426 drivers/net/ethernet/mellanox/mlx4/en_netdev.c 		list_move(&filter->next, &del_list);
next              431 drivers/net/ethernet/mellanox/mlx4/en_netdev.c 	list_for_each_entry_safe(filter, tmp, &del_list, next) {
next              444 drivers/net/ethernet/mellanox/mlx4/en_netdev.c 	list_for_each_entry_safe(filter, tmp, &priv->filters, next) {
next              453 drivers/net/ethernet/mellanox/mlx4/en_netdev.c 			list_move(&filter->next, &del_list);
next              461 drivers/net/ethernet/mellanox/mlx4/en_netdev.c 	if (last_filter && (&last_filter->next != priv->filters.next))
next              462 drivers/net/ethernet/mellanox/mlx4/en_netdev.c 		list_move(&priv->filters, &last_filter->next);
next              466 drivers/net/ethernet/mellanox/mlx4/en_netdev.c 	list_for_each_entry_safe(filter, tmp, &del_list, next)
next              135 drivers/net/ethernet/mellanox/mlx4/en_port.c static unsigned long en_stats_adder(__be64 *start, __be64 *next, int num)
next              140 drivers/net/ethernet/mellanox/mlx4/en_port.c 	int offset = next - start;
next              719 drivers/net/ethernet/mellanox/mlx4/en_rx.c 			goto next;
next              723 drivers/net/ethernet/mellanox/mlx4/en_rx.c 			goto next;
next              750 drivers/net/ethernet/mellanox/mlx4/en_rx.c 						goto next;
next              757 drivers/net/ethernet/mellanox/mlx4/en_rx.c 			goto next;
next              802 drivers/net/ethernet/mellanox/mlx4/en_rx.c 					goto next;
next              815 drivers/net/ethernet/mellanox/mlx4/en_rx.c 				goto next;
next              824 drivers/net/ethernet/mellanox/mlx4/en_rx.c 			goto next;
next              896 drivers/net/ethernet/mellanox/mlx4/en_rx.c next:
next              101 drivers/net/ethernet/mellanox/mlx4/icm.h 		NULL : list_entry(icm->chunk_list.next,
next              114 drivers/net/ethernet/mellanox/mlx4/icm.h 		if (iter->chunk->list.next == &iter->icm->chunk_list) {
next              119 drivers/net/ethernet/mellanox/mlx4/icm.h 		iter->chunk = list_entry(iter->chunk->list.next,
next              184 drivers/net/ethernet/mellanox/mlx4/pd.c 		uar = list_entry(priv->bf_list.next, struct mlx4_uar, bf_list);
next              186 drivers/net/ethernet/mellanox/mlx5/core/cmd.c 	struct mlx5_cmd_mailbox *next = msg->next;
next              190 drivers/net/ethernet/mellanox/mlx5/core/cmd.c 	for (i = 0; i < n && next; i++)  {
next              191 drivers/net/ethernet/mellanox/mlx5/core/cmd.c 		calc_block_sig(next->buf);
next              192 drivers/net/ethernet/mellanox/mlx5/core/cmd.c 		next = next->next;
next              229 drivers/net/ethernet/mellanox/mlx5/core/cmd.c 	struct mlx5_cmd_mailbox *next = ent->out->next;
next              239 drivers/net/ethernet/mellanox/mlx5/core/cmd.c 	for (i = 0; i < n && next; i++) {
next              240 drivers/net/ethernet/mellanox/mlx5/core/cmd.c 		err = verify_block_sig(next->buf);
next              244 drivers/net/ethernet/mellanox/mlx5/core/cmd.c 		next = next->next;
next              774 drivers/net/ethernet/mellanox/mlx5/core/cmd.c 	struct mlx5_cmd_mailbox *next = msg->next;
next              806 drivers/net/ethernet/mellanox/mlx5/core/cmd.c 	for (i = 0; i < n && next; i++)  {
next              809 drivers/net/ethernet/mellanox/mlx5/core/cmd.c 			dump_buf(next->buf, dump_len, 1, offset);
next              813 drivers/net/ethernet/mellanox/mlx5/core/cmd.c 			dump_buf(next->buf, sizeof(struct mlx5_cmd_prot_block), 0, offset);
next              816 drivers/net/ethernet/mellanox/mlx5/core/cmd.c 		next = next->next;
next              897 drivers/net/ethernet/mellanox/mlx5/core/cmd.c 	if (ent->in->next)
next              898 drivers/net/ethernet/mellanox/mlx5/core/cmd.c 		lay->in_ptr = cpu_to_be64(ent->in->next->dma);
next              900 drivers/net/ethernet/mellanox/mlx5/core/cmd.c 	if (ent->out->next)
next              901 drivers/net/ethernet/mellanox/mlx5/core/cmd.c 		lay->out_ptr = cpu_to_be64(ent->out->next->dma);
next             1120 drivers/net/ethernet/mellanox/mlx5/core/cmd.c 	struct mlx5_cmd_mailbox *next;
next             1131 drivers/net/ethernet/mellanox/mlx5/core/cmd.c 	next = to->next;
next             1133 drivers/net/ethernet/mellanox/mlx5/core/cmd.c 		if (!next) {
next             1139 drivers/net/ethernet/mellanox/mlx5/core/cmd.c 		block = next->buf;
next             1144 drivers/net/ethernet/mellanox/mlx5/core/cmd.c 		next = next->next;
next             1153 drivers/net/ethernet/mellanox/mlx5/core/cmd.c 	struct mlx5_cmd_mailbox *next;
next             1164 drivers/net/ethernet/mellanox/mlx5/core/cmd.c 	next = from->next;
next             1166 drivers/net/ethernet/mellanox/mlx5/core/cmd.c 		if (!next) {
next             1172 drivers/net/ethernet/mellanox/mlx5/core/cmd.c 		block = next->buf;
next             1177 drivers/net/ethernet/mellanox/mlx5/core/cmd.c 		next = next->next;
next             1199 drivers/net/ethernet/mellanox/mlx5/core/cmd.c 	mailbox->next = NULL;
next             1238 drivers/net/ethernet/mellanox/mlx5/core/cmd.c 		tmp->next = head;
next             1239 drivers/net/ethernet/mellanox/mlx5/core/cmd.c 		block->next = cpu_to_be64(tmp->next ? tmp->next->dma : 0);
next             1244 drivers/net/ethernet/mellanox/mlx5/core/cmd.c 	msg->next = head;
next             1249 drivers/net/ethernet/mellanox/mlx5/core/cmd.c 		tmp = head->next;
next             1261 drivers/net/ethernet/mellanox/mlx5/core/cmd.c 	struct mlx5_cmd_mailbox *head = msg->next;
next             1262 drivers/net/ethernet/mellanox/mlx5/core/cmd.c 	struct mlx5_cmd_mailbox *next;
next             1265 drivers/net/ethernet/mellanox/mlx5/core/cmd.c 		next = head->next;
next             1267 drivers/net/ethernet/mellanox/mlx5/core/cmd.c 		head = next;
next             1650 drivers/net/ethernet/mellanox/mlx5/core/cmd.c 		msg = list_entry(ch->head.next, typeof(*msg), list);
next              206 drivers/net/ethernet/mellanox/mlx5/core/en.h 	struct mlx5_wqe_srq_next_seg  next;
next              868 drivers/net/ethernet/mellanox/mlx5/core/en_main.c 				       &wqe->next.next_wqe_index);
next              551 drivers/net/ethernet/mellanox/mlx5/core/en_rep.c 	struct mlx5e_neigh_hash_entry *next = NULL;
next              555 drivers/net/ethernet/mellanox/mlx5/core/en_rep.c 	for (next = nhe ?
next              563 drivers/net/ethernet/mellanox/mlx5/core/en_rep.c 	     next;
next              564 drivers/net/ethernet/mellanox/mlx5/core/en_rep.c 	     next = list_next_or_null_rcu(&rpriv->neigh_update.neigh_list,
next              565 drivers/net/ethernet/mellanox/mlx5/core/en_rep.c 					  &next->neigh_list,
next              568 drivers/net/ethernet/mellanox/mlx5/core/en_rep.c 		if (mlx5e_rep_neigh_entry_hold(next))
next              576 drivers/net/ethernet/mellanox/mlx5/core/en_rep.c 	return next;
next             1377 drivers/net/ethernet/mellanox/mlx5/core/en_rx.c 	mlx5_wq_ll_pop(wq, cqe->wqe_id, &wqe->next.next_wqe_index);
next             1438 drivers/net/ethernet/mellanox/mlx5/core/en_tc.c 	struct mlx5e_encap_entry *next = NULL;
next             1444 drivers/net/ethernet/mellanox/mlx5/core/en_tc.c 	for (next = e ?
next             1452 drivers/net/ethernet/mellanox/mlx5/core/en_tc.c 	     next;
next             1453 drivers/net/ethernet/mellanox/mlx5/core/en_tc.c 	     next = list_next_or_null_rcu(&nhe->encap_list,
next             1454 drivers/net/ethernet/mellanox/mlx5/core/en_tc.c 					  &next->encap_list,
next             1457 drivers/net/ethernet/mellanox/mlx5/core/en_tc.c 		if (mlx5e_encap_take(next))
next             1465 drivers/net/ethernet/mellanox/mlx5/core/en_tc.c 	if (!next)
next             1466 drivers/net/ethernet/mellanox/mlx5/core/en_tc.c 		return next;
next             1469 drivers/net/ethernet/mellanox/mlx5/core/en_tc.c 	wait_for_completion(&next->res_ready);
next             1471 drivers/net/ethernet/mellanox/mlx5/core/en_tc.c 	if (!(next->flags & MLX5_ENCAP_ENTRY_VALID)) {
next             1472 drivers/net/ethernet/mellanox/mlx5/core/en_tc.c 		e = next;
next             1476 drivers/net/ethernet/mellanox/mlx5/core/en_tc.c 	return next;
next              130 drivers/net/ethernet/mellanox/mlx5/core/fs_counters.c 	struct list_head *next = mlx5_fc_counters_lookup_next(dev, counter->id);
next              132 drivers/net/ethernet/mellanox/mlx5/core/fs_counters.c 	list_add_tail(&counter->list, next);
next              678 drivers/net/ethernet/mellanox/mlx5/core/health.c 	unsigned long next;
next              680 drivers/net/ethernet/mellanox/mlx5/core/health.c 	get_random_bytes(&next, sizeof(next));
next              681 drivers/net/ethernet/mellanox/mlx5/core/health.c 	next %= HZ;
next              682 drivers/net/ethernet/mellanox/mlx5/core/health.c 	next += jiffies + MLX5_HEALTH_POLL_INTERVAL;
next              684 drivers/net/ethernet/mellanox/mlx5/core/health.c 	return next;
next              167 drivers/net/ethernet/mellanox/mlx5/core/pagealloc.c 	fp = list_entry(dev->priv.free_list.next, struct fw_page, list);
next              316 drivers/net/ethernet/mellanox/mlx5/core/steering/dr_icm_pool.c 	struct mlx5dr_icm_chunk *chunk, *next;
next              322 drivers/net/ethernet/mellanox/mlx5/core/steering/dr_icm_pool.c 	list_for_each_entry_safe(chunk, next, &bucket->free_list, chunk_list)
next              328 drivers/net/ethernet/mellanox/mlx5/core/steering/dr_icm_pool.c 	list_for_each_entry_safe(chunk, next, &bucket->used_list, chunk_list)
next              558 drivers/net/ethernet/mellanox/mlx5/core/steering/dr_icm_pool.c 	struct mlx5dr_icm_mr *icm_mr, *next;
next              563 drivers/net/ethernet/mellanox/mlx5/core/steering/dr_icm_pool.c 	list_for_each_entry_safe(icm_mr, next, &pool->icm_mr_list, mr_list)
next              234 drivers/net/ethernet/mellanox/mlx5/core/uar.c 		up = list_entry(head->next, struct mlx5_uars_page, list);
next               28 drivers/net/ethernet/mellanox/mlxfw/mlxfw_mfa2_tlv_multi.c 	void *next;
next               37 drivers/net/ethernet/mellanox/mlxfw/mlxfw_mfa2_tlv_multi.c 	next = (void *) tlv + tlv_len;
next               38 drivers/net/ethernet/mellanox/mlxfw/mlxfw_mfa2_tlv_multi.c 	return mlxfw_mfa2_tlv_get(mfa2_file, next);
next               90 drivers/net/ethernet/mellanox/mlxsw/core_acl_flex_actions.c 	struct mlxsw_afa_set *next; /* Pointer to the next set. */
next              336 drivers/net/ethernet/mellanox/mlxsw/core_acl_flex_actions.c 		block->first_set->next = block->cur_set;
next              357 drivers/net/ethernet/mellanox/mlxsw/core_acl_flex_actions.c 		next_set = set->next;
next              389 drivers/net/ethernet/mellanox/mlxsw/core_acl_flex_actions.c 			prev_set->next = set;
next              417 drivers/net/ethernet/mellanox/mlxsw/core_acl_flex_actions.c 	if (WARN_ON(!block->first_set->next))
next              419 drivers/net/ethernet/mellanox/mlxsw/core_acl_flex_actions.c 	return block->first_set->next->kvdl_index;
next              653 drivers/net/ethernet/mellanox/mlxsw/core_acl_flex_actions.c 		block->cur_set->next = set;
next              276 drivers/net/ethernet/mellanox/mlxsw/spectrum_acl_tcam.c 		if (region->list.next != &group->region_list &&
next              539 drivers/net/ethernet/mellanox/mlxsw/spectrum_acl_tcam.c 		if (pos->next != &vgroup->vregion_list) { /* not last */
next              540 drivers/net/ethernet/mellanox/mlxsw/spectrum_acl_tcam.c 			vregion2 = list_entry(pos->next, typeof(*vregion2),
next              943 drivers/net/ethernet/micrel/ksz884x.c 	u32 next;
next             1008 drivers/net/ethernet/micrel/ksz884x.c 	int next;
next             1628 drivers/net/ethernet/micrel/ksz884x.c 	*desc = &info->ring[info->next];
next             1629 drivers/net/ethernet/micrel/ksz884x.c 	info->next++;
next             1630 drivers/net/ethernet/micrel/ksz884x.c 	info->next &= info->mask;
next             3838 drivers/net/ethernet/micrel/ksz884x.c 		previous->phw->next = cpu_to_le32(phys);
next             3840 drivers/net/ethernet/micrel/ksz884x.c 	previous->phw->next = cpu_to_le32(desc_info->ring_phys);
next             3845 drivers/net/ethernet/micrel/ksz884x.c 	desc_info->last = desc_info->next = 0;
next             3869 drivers/net/ethernet/micrel/ksz884x.c 	info->last = info->next = 0;
next             5032 drivers/net/ethernet/micrel/ksz884x.c 	int next;
next             5041 drivers/net/ethernet/micrel/ksz884x.c 	next = info->next;
next             5044 drivers/net/ethernet/micrel/ksz884x.c 		desc = &info->ring[next];
next             5058 drivers/net/ethernet/micrel/ksz884x.c 		next++;
next             5059 drivers/net/ethernet/micrel/ksz884x.c 		next &= info->mask;
next             5061 drivers/net/ethernet/micrel/ksz884x.c 	info->next = next;
next             5068 drivers/net/ethernet/micrel/ksz884x.c 	int next;
next             5077 drivers/net/ethernet/micrel/ksz884x.c 	next = info->next;
next             5080 drivers/net/ethernet/micrel/ksz884x.c 		desc = &info->ring[next];
next             5103 drivers/net/ethernet/micrel/ksz884x.c 		next++;
next             5104 drivers/net/ethernet/micrel/ksz884x.c 		next &= info->mask;
next             5106 drivers/net/ethernet/micrel/ksz884x.c 	info->next = next;
next             5113 drivers/net/ethernet/micrel/ksz884x.c 	int next;
next             5122 drivers/net/ethernet/micrel/ksz884x.c 	next = info->next;
next             5125 drivers/net/ethernet/micrel/ksz884x.c 		desc = &info->ring[next];
next             5162 drivers/net/ethernet/micrel/ksz884x.c 		next++;
next             5163 drivers/net/ethernet/micrel/ksz884x.c 		next &= info->mask;
next             5165 drivers/net/ethernet/micrel/ksz884x.c 	info->next = next;
next             2906 drivers/net/ethernet/myricom/myri10ge/myri10ge.c 		segs = segs->next;
next             2907 drivers/net/ethernet/myricom/myri10ge/myri10ge.c 		curr->next = NULL;
next             2913 drivers/net/ethernet/myricom/myri10ge/myri10ge.c 				segs = segs->next;
next             2914 drivers/net/ethernet/myricom/myri10ge/myri10ge.c 				curr->next = NULL;
next             4059 drivers/net/ethernet/myricom/myri10ge/myri10ge.c 	.next = NULL,
next              770 drivers/net/ethernet/neterion/s2io.c 			int next = (j + 1) % blk_cnt;
next              772 drivers/net/ethernet/neterion/s2io.c 			tmp_v_addr_next = ring->rx_blocks[next].block_virt_addr;
next              774 drivers/net/ethernet/neterion/s2io.c 			tmp_p_addr_next = ring->rx_blocks[next].block_dma_addr;
next             8489 drivers/net/ethernet/neterion/s2io.c 		lro->last_frag->next = skb;
next              658 drivers/net/ethernet/neterion/vxge/vxge-main.c 	struct list_head *entry, *next;
next              659 drivers/net/ethernet/neterion/vxge/vxge-main.c 	list_for_each_safe(entry, next, &vpath->mac_addr_list) {
next             1068 drivers/net/ethernet/neterion/vxge/vxge-main.c 	struct list_head *entry, *next;
next             1075 drivers/net/ethernet/neterion/vxge/vxge-main.c 	list_for_each_safe(entry, next, &vpath->mac_addr_list) {
next             1131 drivers/net/ethernet/neterion/vxge/vxge-main.c 	struct list_head *entry, *next;
next             1195 drivers/net/ethernet/neterion/vxge/vxge-main.c 			list_for_each_safe(entry, next, list_head) {
next             1237 drivers/net/ethernet/neterion/vxge/vxge-main.c 			list_for_each_safe(entry, next, list_head) {
next             1446 drivers/net/ethernet/neterion/vxge/vxge-main.c 	struct list_head *entry, *next;
next             1451 drivers/net/ethernet/neterion/vxge/vxge-main.c 		list_for_each_safe(entry, next, &vpath->mac_addr_list) {
next             2909 drivers/net/ethernet/neterion/vxge/vxge-main.c 	struct list_head *entry, *next;
next             2913 drivers/net/ethernet/neterion/vxge/vxge-main.c 	list_for_each_safe(entry, next, &vpath->mac_addr_list) {
next             2091 drivers/net/ethernet/neterion/vxge/vxge-traffic.h 		((struct __vxge_hw_vpath_handle *)(vpath)->vpath_handles.next)
next               22 drivers/net/ethernet/netronome/nfp/bpf/jit.c #define nfp_for_each_insn_walk2(nfp_prog, pos, next)			\
next               24 drivers/net/ethernet/netronome/nfp/bpf/jit.c 	     next = list_next_entry(pos, l);			\
next               26 drivers/net/ethernet/netronome/nfp/bpf/jit.c 	     &(nfp_prog)->insns != &next->l;			\
next               28 drivers/net/ethernet/netronome/nfp/bpf/jit.c 	     next = nfp_meta_next(pos))
next               30 drivers/net/ethernet/netronome/nfp/bpf/jit.c #define nfp_for_each_insn_walk3(nfp_prog, pos, next, next2)		\
next               32 drivers/net/ethernet/netronome/nfp/bpf/jit.c 	     next = list_next_entry(pos, l),			\
next               33 drivers/net/ethernet/netronome/nfp/bpf/jit.c 	     next2 = list_next_entry(next, l);			\
next               35 drivers/net/ethernet/netronome/nfp/bpf/jit.c 	     &(nfp_prog)->insns != &next->l &&			\
next               38 drivers/net/ethernet/netronome/nfp/bpf/jit.c 	     next = nfp_meta_next(pos),				\
next               39 drivers/net/ethernet/netronome/nfp/bpf/jit.c 	     next2 = nfp_meta_next(next))
next             3962 drivers/net/ethernet/netronome/nfp/bpf/jit.c 		struct bpf_insn insn, next;
next             3965 drivers/net/ethernet/netronome/nfp/bpf/jit.c 		next = meta2->insn;
next             3973 drivers/net/ethernet/netronome/nfp/bpf/jit.c 		if (next.code != (BPF_ALU64 | BPF_AND | BPF_K))
next             3978 drivers/net/ethernet/netronome/nfp/bpf/jit.c 		if (exp_mask[BPF_SIZE(insn.code)] != next.imm)
next             3981 drivers/net/ethernet/netronome/nfp/bpf/jit.c 		if (next.src_reg || next.dst_reg)
next               97 drivers/net/ethernet/netronome/nfp/nfp_net_main.c 	struct nfp_net *nn, *next;
next               99 drivers/net/ethernet/netronome/nfp/nfp_net_main.c 	list_for_each_entry_safe(nn, next, &pf->vnics, vnic_list)
next              550 drivers/net/ethernet/netronome/nfp/nfp_net_main.c 	struct nfp_net *nn, *next;
next              588 drivers/net/ethernet/netronome/nfp/nfp_net_main.c 	list_for_each_entry_safe(nn, next, &pf->vnics, vnic_list) {
next              764 drivers/net/ethernet/netronome/nfp/nfp_net_main.c 	struct nfp_net *nn, *next;
next              767 drivers/net/ethernet/netronome/nfp/nfp_net_main.c 	list_for_each_entry_safe(nn, next, &pf->vnics, vnic_list) {
next              148 drivers/net/ethernet/ni/nixge.c 	nixge_hw_dma_bd_set_addr((bd), next, (addr))
next              752 drivers/net/ethernet/pasemi/pasemi_mac.c 			goto next;
next              775 drivers/net/ethernet/pasemi/pasemi_mac.c next:
next              319 drivers/net/ethernet/pensando/ionic/ionic_dev.c 			cur->next = cq->info;
next              322 drivers/net/ethernet/pensando/ionic/ionic_dev.c 			cur->next = cur + 1;
next              360 drivers/net/ethernet/pensando/ionic/ionic_dev.c 		cq->tail = cq->tail->next;
next              405 drivers/net/ethernet/pensando/ionic/ionic_dev.c 			cur->next = q->info;
next              407 drivers/net/ethernet/pensando/ionic/ionic_dev.c 			cur->next = cur + 1;
next              448 drivers/net/ethernet/pensando/ionic/ionic_dev.c 	q->head = q->head->next;
next              489 drivers/net/ethernet/pensando/ionic/ionic_dev.c 		q->tail = desc_info->next;
next              141 drivers/net/ethernet/pensando/ionic/ionic_dev.h 	struct ionic_cq_info *next;
next              157 drivers/net/ethernet/pensando/ionic/ionic_dev.h 	struct ionic_desc_info *next;
next              183 drivers/net/ethernet/pensando/ionic/ionic_main.c 		adminq->tail = adminq->tail->next;
next              176 drivers/net/ethernet/pensando/ionic/ionic_txrx.c 	q->tail = desc_info->next;
next              194 drivers/net/ethernet/pensando/ionic/ionic_txrx.c 		rxcq->tail = rxcq->tail->next;
next              290 drivers/net/ethernet/pensando/ionic/ionic_txrx.c 	for (cur = q->tail; cur != q->head; cur = cur->next) {
next              433 drivers/net/ethernet/pensando/ionic/ionic_txrx.c 			q->tail = desc_info->next;
next              443 drivers/net/ethernet/pensando/ionic/ionic_txrx.c 		cq->tail = cq->tail->next;
next              707 drivers/net/ethernet/pensando/ionic/ionic_txrx.c 		rewind = rewind->next;
next              687 drivers/net/ethernet/qlogic/netxen/netxen_nic_hw.c 		cur = list_entry(head->next, nx_mac_list_t, list);
next              721 drivers/net/ethernet/qlogic/netxen/netxen_nic_hw.c 		cur = list_entry(head->next, nx_mac_list_t, list);
next             1217 drivers/net/ethernet/qlogic/netxen/netxen_nic_init.c next:
next             1226 drivers/net/ethernet/qlogic/netxen/netxen_nic_init.c 			goto next;
next             1232 drivers/net/ethernet/qlogic/netxen/netxen_nic_init.c 			goto next;
next             1822 drivers/net/ethernet/qlogic/netxen/netxen_nic_init.c 		buffer = list_entry(head->next, struct netxen_rx_buffer, list);
next             1882 drivers/net/ethernet/qlogic/netxen/netxen_nic_init.c 		buffer = list_entry(head->next, struct netxen_rx_buffer, list);
next              109 drivers/net/ethernet/qlogic/qed/qed_cxt.c 	u64 next;
next              971 drivers/net/ethernet/qlogic/qed/qed_cxt.c 			entries[j].next = cpu_to_be64(val);
next              978 drivers/net/ethernet/qlogic/qed/qed_cxt.c 		entries[j].next = cpu_to_be64(val);
next               81 drivers/net/ethernet/qlogic/qede/qede_rdma.c 		event_node = list_entry(head->next, struct qede_rdma_event_work,
next              299 drivers/net/ethernet/qlogic/qla3xxx.c 	lrg_buf_cb->next = NULL;
next              304 drivers/net/ethernet/qlogic/qla3xxx.c 		qdev->lrg_buf_free_tail->next = lrg_buf_cb;
next              356 drivers/net/ethernet/qlogic/qla3xxx.c 		qdev->lrg_buf_free_head = lrg_buf_cb->next;
next             1836 drivers/net/ethernet/qlogic/qla3xxx.c 		lrg_buf_cb = lrg_buf_cb->next;
next              999 drivers/net/ethernet/qlogic/qla3xxx.h 	struct ql_rcv_buf_cb *next;
next             3901 drivers/net/ethernet/qlogic/qlcnic/qlcnic_83xx_hw.c 		cmd = list_entry(head->next, struct qlcnic_cmd_args, list);
next             4125 drivers/net/ethernet/qlogic/qlcnic/qlcnic_83xx_hw.c 		cmd = list_entry(head->next, struct qlcnic_cmd_args, list);
next              984 drivers/net/ethernet/qlogic/qlcnic/qlcnic_83xx_init.c 	u32 cur, prev, next;
next              988 drivers/net/ethernet/qlogic/qlcnic/qlcnic_83xx_init.c 	next = state;
next              990 drivers/net/ethernet/qlogic/qlcnic/qlcnic_83xx_init.c 	if ((next < QLC_83XX_IDC_DEV_COLD) ||
next              991 drivers/net/ethernet/qlogic/qlcnic/qlcnic_83xx_init.c 	    (next > QLC_83XX_IDC_DEV_QUISCENT)) {
next             1000 drivers/net/ethernet/qlogic/qlcnic/qlcnic_83xx_init.c 		if ((next != QLC_83XX_IDC_DEV_COLD) &&
next             1001 drivers/net/ethernet/qlogic/qlcnic/qlcnic_83xx_init.c 		    (next != QLC_83XX_IDC_DEV_READY)) {
next             1004 drivers/net/ethernet/qlogic/qlcnic/qlcnic_83xx_init.c 				__func__, cur, prev, next);
next             1009 drivers/net/ethernet/qlogic/qlcnic/qlcnic_83xx_init.c 	if (next == QLC_83XX_IDC_DEV_INIT) {
next             1015 drivers/net/ethernet/qlogic/qlcnic/qlcnic_83xx_init.c 				__func__, cur, prev, next);
next              626 drivers/net/ethernet/qlogic/qlcnic/qlcnic_hw.c 		cur = list_entry(head->next, struct qlcnic_mac_vlan_list, list);
next             1279 drivers/net/ethernet/qlogic/qlcnic/qlcnic_init.c next:
next             1289 drivers/net/ethernet/qlogic/qlcnic/qlcnic_init.c 			goto next;
next             1295 drivers/net/ethernet/qlogic/qlcnic/qlcnic_init.c 			goto next;
next              860 drivers/net/ethernet/qlogic/qlcnic/qlcnic_io.c 		buffer = list_entry(head->next, struct qlcnic_rx_buffer, list);
next             1451 drivers/net/ethernet/qlogic/qlcnic/qlcnic_io.c 		buffer = list_entry(head->next, struct qlcnic_rx_buffer, list);
next             1532 drivers/net/ethernet/qlogic/qlcnic/qlcnic_sriov_common.c 		entry = list_entry(head->next, struct qlcnic_async_cmd,
next             2103 drivers/net/ethernet/qlogic/qlcnic/qlcnic_sriov_common.c 		cur = list_entry(head->next, struct qlcnic_mac_vlan_list, list);
next             1222 drivers/net/ethernet/sfc/efx.c 	struct efx_nic *other, *next;
next             1230 drivers/net/ethernet/sfc/efx.c 		list_for_each_entry_safe(other, next, &efx_unassociated_list,
next             1267 drivers/net/ethernet/sfc/efx.c 	struct efx_nic *other, *next;
next             1272 drivers/net/ethernet/sfc/efx.c 	list_for_each_entry_safe(other, next, &efx->secondary_list, node) {
next             1162 drivers/net/ethernet/sfc/falcon/efx.c 	struct ef4_nic *other, *next;
next             1170 drivers/net/ethernet/sfc/falcon/efx.c 		list_for_each_entry_safe(other, next, &ef4_unassociated_list,
next             1207 drivers/net/ethernet/sfc/falcon/efx.c 	struct ef4_nic *other, *next;
next             1212 drivers/net/ethernet/sfc/falcon/efx.c 	list_for_each_entry_safe(other, next, &efx->secondary_list, node) {
next               97 drivers/net/ethernet/sfc/falcon/mtd.c 	struct ef4_mtd_partition *parts, *part, *next;
next              107 drivers/net/ethernet/sfc/falcon/mtd.c 	list_for_each_entry_safe(part, next, &efx->mtd_list, node)
next             1010 drivers/net/ethernet/sfc/mcdi.c 		if (mcdi->async_list.next == &async->list &&
next             1132 drivers/net/ethernet/sfc/mcdi.c 	struct efx_mcdi_async_param *async, *next;
next             1159 drivers/net/ethernet/sfc/mcdi.c 	list_for_each_entry_safe(async, next, &mcdi->async_list, list) {
next              100 drivers/net/ethernet/sfc/mtd.c 	struct efx_mtd_partition *parts, *part, *next;
next              110 drivers/net/ethernet/sfc/mtd.c 	list_for_each_entry_safe(part, next, &efx->mtd_list, node)
next             1156 drivers/net/ethernet/sfc/ptp.c 	struct list_head *next;
next             1164 drivers/net/ethernet/sfc/ptp.c 		list_for_each_safe(cursor, next, &ptp->evt_list) {
next             1185 drivers/net/ethernet/sfc/ptp.c 	struct list_head *next;
next             1201 drivers/net/ethernet/sfc/ptp.c 	list_for_each_safe(cursor, next, &ptp->evt_list) {
next             1353 drivers/net/ethernet/sfc/ptp.c 	struct list_head *next;
next             1369 drivers/net/ethernet/sfc/ptp.c 	list_for_each_safe(cursor, next, &efx->ptp_data->evt_list) {
next              465 drivers/net/ethernet/sfc/tx.c 	struct sk_buff *segments, *next;
next              475 drivers/net/ethernet/sfc/tx.c 		next = skb->next;
next              476 drivers/net/ethernet/sfc/tx.c 		skb->next = NULL;
next              479 drivers/net/ethernet/sfc/tx.c 		skb = next;
next              572 drivers/net/ethernet/sgi/ioc3-eth.c 				goto next;
next              603 drivers/net/ethernet/sgi/ioc3-eth.c next:
next              792 drivers/net/ethernet/silan/sc92031.c 			goto next;
next              813 drivers/net/ethernet/silan/sc92031.c 	next:
next             1245 drivers/net/ethernet/sis/sis190.c 	struct sis190_phy *cur, *next;
next             1247 drivers/net/ethernet/sis/sis190.c 	list_for_each_entry_safe(cur, next, first_phy, list) {
next              148 drivers/net/ethernet/sis/sis900.c 	struct mii_phy * next;
next              631 drivers/net/ethernet/sis/sis900.c 				mii_phy = mii_phy->next;
next              641 drivers/net/ethernet/sis/sis900.c 		mii_phy->next = sis_priv->mii;
next              732 drivers/net/ethernet/sis/sis900.c         for (phy=sis_priv->first_mii; phy; phy=phy->next) {
next             2483 drivers/net/ethernet/sis/sis900.c 		sis_priv->first_mii = phy->next;
next              237 drivers/net/ethernet/smsc/epic100.c 	u32 next;
next              244 drivers/net/ethernet/smsc/epic100.c 	u32 next;
next              906 drivers/net/ethernet/smsc/epic100.c 		ep->rx_ring[i].next = ep->rx_ring_dma +
next              911 drivers/net/ethernet/smsc/epic100.c 	ep->rx_ring[i-1].next = ep->rx_ring_dma;
next              931 drivers/net/ethernet/smsc/epic100.c 		ep->tx_ring[i].next = ep->tx_ring_dma +
next              934 drivers/net/ethernet/smsc/epic100.c 	ep->tx_ring[i-1].next = ep->tx_ring_dma;
next              658 drivers/net/ethernet/socionext/netsec.c 			goto next;
next              666 drivers/net/ethernet/socionext/netsec.c next:
next             1005 drivers/net/ethernet/socionext/netsec.c 				goto next;
next             1031 drivers/net/ethernet/socionext/netsec.c next:
next              640 drivers/net/ethernet/sun/cassini.c 	entry = cp->rx_spare_list.next;
next             2329 drivers/net/ethernet/sun/cassini.c 			goto next;
next             2353 drivers/net/ethernet/sun/cassini.c 	next:
next             3274 drivers/net/ethernet/sun/cassini.c 				goto next;
next             3283 drivers/net/ethernet/sun/cassini.c 						goto next;
next             3294 drivers/net/ethernet/sun/cassini.c 				goto next;
next             3301 drivers/net/ethernet/sun/cassini.c 				goto next;
next             3306 drivers/net/ethernet/sun/cassini.c 				goto next;
next             3325 drivers/net/ethernet/sun/cassini.c 			goto next;
next             3330 drivers/net/ethernet/sun/cassini.c next:
next             3521 drivers/net/ethernet/sun/niu.c 			struct page *next = (struct page *) page->mapping;
next             3531 drivers/net/ethernet/sun/niu.c 			page = next;
next             6445 drivers/net/ethernet/sun/niu.c 					struct page *next =
next             6450 drivers/net/ethernet/sun/niu.c 					page = next;
next              821 drivers/net/ethernet/sun/sunbmac.c 			goto next;
next              879 drivers/net/ethernet/sun/sunbmac.c 	next:
next              812 drivers/net/ethernet/sun/sungem.c 			goto next;
next              870 drivers/net/ethernet/sun/sungem.c 	next:
next             2025 drivers/net/ethernet/sun/sunhme.c 			goto next;
next             2088 drivers/net/ethernet/sun/sunhme.c 	next:
next             2160 drivers/net/ethernet/sun/sunhme.c 				goto next;
next             2178 drivers/net/ethernet/sun/sunhme.c 	next:
next             2578 drivers/net/ethernet/sun/sunhme.c 		qp->next = qfe_sbus_list;
next             2594 drivers/net/ethernet/sun/sunhme.c 	for (qp = qfe_sbus_list; qp != NULL; qp = qp->next) {
next             2623 drivers/net/ethernet/sun/sunhme.c 	for (qp = qfe_sbus_list; qp != NULL; qp = qp->next) {
next             2646 drivers/net/ethernet/sun/sunhme.c 	for (qp = qfe_pci_list; qp != NULL; qp = qp->next) {
next             2660 drivers/net/ethernet/sun/sunhme.c 		qp->next = qfe_pci_list;
next             3256 drivers/net/ethernet/sun/sunhme.c 		struct quattro *next = qfe->next;
next             3260 drivers/net/ethernet/sun/sunhme.c 		qfe_pci_list = next;
next             3354 drivers/net/ethernet/sun/sunhme.c 		struct quattro *next = qfe->next;
next             3358 drivers/net/ethernet/sun/sunhme.c 		qfe_sbus_list = next;
next              489 drivers/net/ethernet/sun/sunhme.h 	struct quattro		*next;
next              484 drivers/net/ethernet/sun/sunqe.c 					goto next;
next              501 drivers/net/ethernet/sun/sunqe.c 	next:
next              983 drivers/net/ethernet/sun/sunqe.c 		struct sunqec *next = root_qec_dev->next_module;
next              991 drivers/net/ethernet/sun/sunqe.c 		root_qec_dev = next;
next             1015 drivers/net/ethernet/sun/sunvnet_common.c 			BUG_ON(port->tx_bufs[txi].skb->next);
next             1017 drivers/net/ethernet/sun/sunvnet_common.c 			port->tx_bufs[txi].skb->next = skb;
next             1034 drivers/net/ethernet/sun/sunvnet_common.c 	struct sk_buff *next;
next             1037 drivers/net/ethernet/sun/sunvnet_common.c 		next = skb->next;
next             1038 drivers/net/ethernet/sun/sunvnet_common.c 		skb->next = NULL;
next             1040 drivers/net/ethernet/sun/sunvnet_common.c 		skb = next;
next             1288 drivers/net/ethernet/sun/sunvnet_common.c 		segs = segs->next;
next             1289 drivers/net/ethernet/sun/sunvnet_common.c 		curr->next = NULL;
next             1570 drivers/net/ethernet/sun/sunvnet_common.c 	for (m = vp->mcast_list; m; m = m->next) {
next             1597 drivers/net/ethernet/sun/sunvnet_common.c 			m->next = vp->mcast_list;
next             1618 drivers/net/ethernet/sun/sunvnet_common.c 	for (m = vp->mcast_list; m; m = m->next) {
next             1644 drivers/net/ethernet/sun/sunvnet_common.c 			pp = &m->next;
next             1657 drivers/net/ethernet/sun/sunvnet_common.c 		*pp = m->next;
next             1792 drivers/net/ethernet/sun/sunvnet_common.c 		port = list_entry(vp->port_list.next, struct vnet_port, list);
next              112 drivers/net/ethernet/sun/sunvnet_common.h 	struct vnet_mcast_entry	*next;
next              181 drivers/net/ethernet/ti/cpmac.c 	struct cpmac_desc *next;
next              245 drivers/net/ethernet/ti/cpmac.c 		dump = dump->next;
next              437 drivers/net/ethernet/ti/cpmac.c 			restart = desc->next;
next              445 drivers/net/ethernet/ti/cpmac.c 		desc = desc->next;
next              693 drivers/net/ethernet/ti/cpmac.c 		desc->hw_next = desc->next->mapping;
next              694 drivers/net/ethernet/ti/cpmac.c 		desc = desc->next;
next              962 drivers/net/ethernet/ti/cpmac.c 		desc->next = &priv->rx_head[(i + 1) % priv->ring_size];
next              963 drivers/net/ethernet/ti/cpmac.c 		desc->next->prev = desc;
next              964 drivers/net/ethernet/ti/cpmac.c 		desc->hw_next = (u32)desc->next->mapping;
next               60 drivers/net/ethernet/ti/cpts.c 	struct list_head *this, *next;
next               64 drivers/net/ethernet/ti/cpts.c 	list_for_each_safe(this, next, &cpts->events) {
next              194 drivers/net/ethernet/ti/cpts.c 	struct list_head *this, *next;
next              201 drivers/net/ethernet/ti/cpts.c 	list_for_each_safe(this, next, &cpts->events) {
next              371 drivers/net/ethernet/ti/cpts.c 	struct list_head *this, *next;
next              382 drivers/net/ethernet/ti/cpts.c 	list_for_each_safe(this, next, &cpts->events) {
next              475 drivers/net/ethernet/ti/netcp_core.c 	struct netcp_hook_list *next;
next              487 drivers/net/ethernet/ti/netcp_core.c 	list_for_each_entry(next, &netcp_priv->txhook_list_head, list) {
next              488 drivers/net/ethernet/ti/netcp_core.c 		if (next->order > order)
next              491 drivers/net/ethernet/ti/netcp_core.c 	__list_add(&entry->list, next->list.prev, &next->list);
next              501 drivers/net/ethernet/ti/netcp_core.c 	struct netcp_hook_list *next, *n;
next              505 drivers/net/ethernet/ti/netcp_core.c 	list_for_each_entry_safe(next, n, &netcp_priv->txhook_list_head, list) {
next              506 drivers/net/ethernet/ti/netcp_core.c 		if ((next->order     == order) &&
next              507 drivers/net/ethernet/ti/netcp_core.c 		    (next->hook_rtn  == hook_rtn) &&
next              508 drivers/net/ethernet/ti/netcp_core.c 		    (next->hook_data == hook_data)) {
next              509 drivers/net/ethernet/ti/netcp_core.c 			list_del(&next->list);
next              511 drivers/net/ethernet/ti/netcp_core.c 			devm_kfree(netcp_priv->dev, next);
next              524 drivers/net/ethernet/ti/netcp_core.c 	struct netcp_hook_list *next;
next              536 drivers/net/ethernet/ti/netcp_core.c 	list_for_each_entry(next, &netcp_priv->rxhook_list_head, list) {
next              537 drivers/net/ethernet/ti/netcp_core.c 		if (next->order > order)
next              540 drivers/net/ethernet/ti/netcp_core.c 	__list_add(&entry->list, next->list.prev, &next->list);
next              550 drivers/net/ethernet/ti/netcp_core.c 	struct netcp_hook_list *next, *n;
next              554 drivers/net/ethernet/ti/netcp_core.c 	list_for_each_entry_safe(next, n, &netcp_priv->rxhook_list_head, list) {
next              555 drivers/net/ethernet/ti/netcp_core.c 		if ((next->order     == order) &&
next              556 drivers/net/ethernet/ti/netcp_core.c 		    (next->hook_rtn  == hook_rtn) &&
next              557 drivers/net/ethernet/ti/netcp_core.c 		    (next->hook_data == hook_data)) {
next              558 drivers/net/ethernet/ti/netcp_core.c 			list_del(&next->list);
next              560 drivers/net/ethernet/ti/netcp_core.c 			devm_kfree(netcp_priv->dev, next);
next              225 drivers/net/ethernet/toshiba/ps3_gelic_net.c 	for (descr = start_descr; start_descr != descr->next; descr++) {
next              227 drivers/net/ethernet/toshiba/ps3_gelic_net.c 		descr->next_descr_addr = cpu_to_be32(descr->next->bus_addr);
next              289 drivers/net/ethernet/toshiba/ps3_gelic_net.c 	for (descr = descr_in; descr && descr->bus_addr; descr = descr->next) {
next              329 drivers/net/ethernet/toshiba/ps3_gelic_net.c 		descr->next = descr + 1;
next              333 drivers/net/ethernet/toshiba/ps3_gelic_net.c 	(descr - 1)->next = start_descr;
next              339 drivers/net/ethernet/toshiba/ps3_gelic_net.c 		descr->next_descr_addr = cpu_to_be32(descr->next->bus_addr);
next              438 drivers/net/ethernet/toshiba/ps3_gelic_net.c 		descr = descr->next;
next              461 drivers/net/ethernet/toshiba/ps3_gelic_net.c 		descr = descr->next;
next              547 drivers/net/ethernet/toshiba/ps3_gelic_net.c 	     tx_chain->tail = tx_chain->tail->next) {
next              681 drivers/net/ethernet/toshiba/ps3_gelic_net.c 	if (card->tx_chain.tail != card->tx_chain.head->next &&
next              799 drivers/net/ethernet/toshiba/ps3_gelic_net.c 	card->tx_chain.head = descr->next;
next             1060 drivers/net/ethernet/toshiba/ps3_gelic_net.c 	chain->head = descr->next;
next              246 drivers/net/ethernet/toshiba/ps3_gelic_net.h 	struct gelic_descr *next;
next             1593 drivers/net/ethernet/toshiba/ps3_gelic_wireless.c 				target = list_entry(wl->network_free_list.next,
next              283 drivers/net/ethernet/toshiba/spider_net.c 		descr = descr->next;
next              329 drivers/net/ethernet/toshiba/spider_net.c 		descr->next = descr + 1;
next              335 drivers/net/ethernet/toshiba/spider_net.c 	(descr-1)->next = chain->ring;
next              364 drivers/net/ethernet/toshiba/spider_net.c 		descr = descr->next;
next              501 drivers/net/ethernet/toshiba/spider_net.c 		chain->head = chain->head->next;
next              523 drivers/net/ethernet/toshiba/spider_net.c 		descr = descr->next;
next              532 drivers/net/ethernet/toshiba/spider_net.c 		chain->head = chain->head->next;
next              663 drivers/net/ethernet/toshiba/spider_net.c 	if (descr->next == chain->tail->prev) {
next              669 drivers/net/ethernet/toshiba/spider_net.c 	chain->head = descr->next;
next              715 drivers/net/ethernet/toshiba/spider_net.c 		descr = descr->next;
next              727 drivers/net/ethernet/toshiba/spider_net.c 		descr = descr->next;
next              810 drivers/net/ethernet/toshiba/spider_net.c 		chain->tail = descr->next;
next              859 drivers/net/ethernet/toshiba/spider_net.c 		descr = descr->next;
next             1034 drivers/net/ethernet/toshiba/spider_net.c 		descr = descr->next;
next             1060 drivers/net/ethernet/toshiba/spider_net.c 		descr = descr->next;
next             1095 drivers/net/ethernet/toshiba/spider_net.c 		descr = descr->next;
next             1116 drivers/net/ethernet/toshiba/spider_net.c 		descr = descr->next;
next             1155 drivers/net/ethernet/toshiba/spider_net.c 	chain->tail = descr->next;
next              388 drivers/net/ethernet/toshiba/spider_net.h 	struct spider_net_descr *next;
next             1260 drivers/net/ethernet/via/via-rhine.c 	dma_addr_t next;
next             1264 drivers/net/ethernet/via/via-rhine.c 	next = rp->rx_ring_dma;
next             1270 drivers/net/ethernet/via/via-rhine.c 		next += sizeof(struct rx_desc);
next             1271 drivers/net/ethernet/via/via-rhine.c 		rp->rx_ring[i].next_desc = cpu_to_le32(next);
next             1318 drivers/net/ethernet/via/via-rhine.c 	dma_addr_t next;
next             1322 drivers/net/ethernet/via/via-rhine.c 	next = rp->tx_ring_dma;
next             1327 drivers/net/ethernet/via/via-rhine.c 		next += sizeof(struct tx_desc);
next             1328 drivers/net/ethernet/via/via-rhine.c 		rp->tx_ring[i].next_desc = cpu_to_le32(next);
next              321 drivers/net/ethernet/xilinx/ll_temac.h 	u32 next;	/* Physical address of next buffer descriptor */
next              353 drivers/net/ethernet/xilinx/ll_temac_main.c 		lp->tx_bd_v[i].next = cpu_to_be32(lp->tx_bd_p
next              358 drivers/net/ethernet/xilinx/ll_temac_main.c 		lp->rx_bd_v[i].next = cpu_to_be32(lp->rx_bd_p
next              364 drivers/net/ethernet/xilinx/xilinx_axienet.h 	u32 next;	/* Physical address of next buffer descriptor */
next              219 drivers/net/ethernet/xilinx/xilinx_axienet_main.c 		lp->tx_bd_v[i].next = lp->tx_bd_p +
next              225 drivers/net/ethernet/xilinx/xilinx_axienet_main.c 		lp->rx_bd_v[i].next = lp->rx_bd_p +
next              195 drivers/net/ethernet/xscale/ixp4xx_eth.c 	u32 next;		/* pointer to next buffer, unused */
next              602 drivers/net/ethernet/xscale/ixp4xx_eth.c 	       phys, desc->next, desc->buf_len, desc->pkt_len,
next              627 drivers/net/ethernet/xscale/ixp4xx_eth.c 	BUG_ON(tab[n_desc].next);
next              120 drivers/net/fddi/skfp/smttimer.c 	struct smt_timer	*next ;
next              141 drivers/net/fddi/skfp/smttimer.c 	next = smc->t.st_queue ;
next              144 drivers/net/fddi/skfp/smttimer.c 	for ( tm = next ; tm ; tm = next) {
next              145 drivers/net/fddi/skfp/smttimer.c 		next = tm->tm_next ;
next               64 drivers/net/geneve.c 	struct list_head   next;	/* geneve's per namespace list */
next             1228 drivers/net/geneve.c 	list_for_each_entry(geneve, &gn->geneve_list, next) {
next             1317 drivers/net/geneve.c 	list_add(&geneve->next, &gn->geneve_list);
next             1660 drivers/net/geneve.c 	list_del(&geneve->next);
next             1832 drivers/net/geneve.c 	struct geneve_dev *geneve, *next;
next             1841 drivers/net/geneve.c 	list_for_each_entry_safe(geneve, next, &gn->geneve_list, next) {
next              418 drivers/net/hamradio/bpqether.c 	.next = bpq_seq_next,
next              577 drivers/net/hamradio/bpqether.c 		bpq = list_entry(bpq_devices.next, struct bpqdev, bpq_list);
next              213 drivers/net/hamradio/dmascc.c 	struct scc_info *next;
next              290 drivers/net/hamradio/dmascc.c 		first = info->next;
next              591 drivers/net/hamradio/dmascc.c 	info->next = first;
next             2083 drivers/net/hamradio/scc.c 	.next   = scc_net_seq_next,
next              138 drivers/net/hamradio/yam.c 	struct yam_mcs *next;
next              385 drivers/net/hamradio/yam.c 		p = p->next;
next              395 drivers/net/hamradio/yam.c 	p->next = yam_data;
next              410 drivers/net/hamradio/yam.c 		p = p->next;
next              825 drivers/net/hamradio/yam.c 	.next = yam_seq_next,
next             1174 drivers/net/hamradio/yam.c 		yam_data = yam_data->next;
next              848 drivers/net/hyperv/hyperv_net.h 	u32 next;	/* next entry for writing */
next             1038 drivers/net/hyperv/netvsc.c 	while (mrc->first != mrc->next) {
next             1070 drivers/net/hyperv/netvsc.c 	if (mrc->next >= mrc->first)
next             1071 drivers/net/hyperv/netvsc.c 		*filled = mrc->next - mrc->first;
next             1073 drivers/net/hyperv/netvsc.c 		*filled = (count - mrc->first) + mrc->next;
next             1101 drivers/net/hyperv/netvsc.c 	rcd = mrc->slots + mrc->next;
next             1105 drivers/net/hyperv/netvsc.c 	if (++mrc->next == nvdev->recv_completion_cnt)
next             1106 drivers/net/hyperv/netvsc.c 		mrc->next = 0;
next             1003 drivers/net/hyperv/rndis_filter.c 		if (nvchan->mrc.first != nvchan->mrc.next)
next              630 drivers/net/ipvlan/ipvlan_main.c 	struct ipvl_addr *addr, *next;
next              633 drivers/net/ipvlan/ipvlan_main.c 	list_for_each_entry_safe(addr, next, &ipvlan->addrs, anode) {
next              694 drivers/net/ipvlan/ipvlan_main.c 	struct ipvl_dev *ipvlan, *next;
next              727 drivers/net/ipvlan/ipvlan_main.c 		list_for_each_entry_safe(ipvlan, next, &port->ipvlans, pnode)
next               67 drivers/net/macsec.c 	     sc = rcu_dereference_bh(sc->next))
next               71 drivers/net/macsec.c 	     sc = rtnl_dereference(sc->next))
next              167 drivers/net/macsec.c 	struct macsec_rx_sc __rcu *next;
next             1381 drivers/net/macsec.c 	     rx_scp = &rx_sc->next, rx_sc = rtnl_dereference(*rx_scp)) {
next             1385 drivers/net/macsec.c 			rcu_assign_pointer(*rx_scp, rx_sc->next);
next             1421 drivers/net/macsec.c 	rcu_assign_pointer(rx_sc->next, secy->rx_sc);
next             2617 drivers/net/macsec.c 			goto next;
next             2620 drivers/net/macsec.c 			goto next;
next             2625 drivers/net/macsec.c next:
next             3107 drivers/net/macsec.c 		rcu_assign_pointer(macsec->secy.rx_sc, rx_sc->next);
next             1668 drivers/net/macvlan.c 	struct macvlan_dev *vlan, *next;
next             1716 drivers/net/macvlan.c 		list_for_each_entry_safe(vlan, next, &port->vlans, list)
next              592 drivers/net/phy/dp83640.c 	struct list_head *this, *next;
next              595 drivers/net/phy/dp83640.c 	list_for_each_safe(this, next, &dp83640->rxts) {
next             1017 drivers/net/phy/dp83640.c 	struct list_head *this, *next;
next             1021 drivers/net/phy/dp83640.c 	list_for_each_safe(this, next, &phyter_clocks) {
next             1193 drivers/net/phy/dp83640.c 	struct list_head *this, *next;
next             1211 drivers/net/phy/dp83640.c 		list_for_each_safe(this, next, &clock->phylist) {
next             1436 drivers/net/phy/dp83640.c 	struct list_head *this, *next;
next             1455 drivers/net/phy/dp83640.c 	list_for_each_safe(this, next, &dp83640->rxts) {
next               31 drivers/net/phy/mdio-mux.c 	struct mdio_mux_child_bus *next;
next              174 drivers/net/phy/mdio-mux.c 			cb->next = pb->children;
next              201 drivers/net/phy/mdio-mux.c 		cb = cb->next;
next             1639 drivers/net/ppp/ppp_generic.c 		list = list->next;
next             1749 drivers/net/ppp/ppp_generic.c 		list = list->next;
next             1759 drivers/net/ppp/ppp_generic.c 		list = list->next;
next             2568 drivers/net/ppp/ppp_generic.c 				p->next = NULL;
next             2569 drivers/net/ppp/ppp_generic.c 				fragpp = &p->next;
next              173 drivers/net/ppp/pppoe.c 		ret = ret->next;
next              190 drivers/net/ppp/pppoe.c 		ret = ret->next;
next              193 drivers/net/ppp/pppoe.c 	po->next = pn->hash_table[hash];
next              211 drivers/net/ppp/pppoe.c 			*src = ret->next;
next              215 drivers/net/ppp/pppoe.c 		src = &ret->next;
next              216 drivers/net/ppp/pppoe.c 		ret = ret->next;
next              288 drivers/net/ppp/pppoe.c 				po = po->next;
next              661 drivers/net/ppp/pppoe.c 		po->next = NULL;
next             1042 drivers/net/ppp/pppoe.c 			po = po->next;
next             1071 drivers/net/ppp/pppoe.c 	if (po->next)
next             1072 drivers/net/ppp/pppoe.c 		po = po->next;
next             1097 drivers/net/ppp/pppoe.c 	.next		= pppoe_seq_next,
next              136 drivers/net/slip/slhc.c 			ts[i].next = &(ts[i - 1]);
next              138 drivers/net/slip/slhc.c 		ts[0].next = &(ts[comp->tslot_limit]);
next              232 drivers/net/slip/slhc.c 	struct cstate *cs = lcs->next;
next              307 drivers/net/slip/slhc.c 		cs = cs->next;
next              334 drivers/net/slip/slhc.c 		lcs->next = cs->next;
next              335 drivers/net/slip/slhc.c 		cs->next = ocs->next;
next              336 drivers/net/slip/slhc.c 		ocs->next = cs;
next              112 drivers/net/tap.c 	struct list_head next;
next              179 drivers/net/tap.c 	list_add_tail(&q->next, &tap->queue_list);
next              236 drivers/net/tap.c 		list_del_init(&q->next);
next              303 drivers/net/tap.c 	list_for_each_entry_safe(q, tmp, &tap->queue_list, next) {
next              304 drivers/net/tap.c 		list_del_init(&q->next);
next              356 drivers/net/tap.c 			struct sk_buff *nskb = segs->next;
next              358 drivers/net/tap.c 			segs->next = NULL;
next              397 drivers/net/tap.c 	list_for_each_entry_rcu(tap_major, &major_list, next) {
next             1303 drivers/net/tap.c 	list_for_each_entry(q, &tap->queue_list, next)
next             1330 drivers/net/tap.c 	list_add_tail_rcu(&tap_major->next, &major_list);
next             1370 drivers/net/tap.c 	list_for_each_entry_safe(tap_major, tmp, &major_list, next) {
next             1373 drivers/net/tap.c 			list_del_rcu(&tap_major->next);
next              175 drivers/net/tun.c 	struct list_head next;
next              638 drivers/net/tun.c 	list_add_tail(&tfile->next, &tun->disabled);
next              647 drivers/net/tun.c 	list_del_init(&tfile->next);
next              761 drivers/net/tun.c 	list_for_each_entry(tfile, &tun->disabled, next) {
next              777 drivers/net/tun.c 	list_for_each_entry_safe(tfile, tmp, &tun->disabled, next) {
next             1211 drivers/net/tun.c 	list_for_each_entry(tfile, &tun->disabled, next) {
next             3450 drivers/net/tun.c 	INIT_LIST_HEAD(&tfile->next);
next             3631 drivers/net/tun.c 	list_for_each_entry(tfile, &tun->disabled, next)
next              219 drivers/net/usb/cdc_eem.c 					goto next;
next              277 drivers/net/usb/cdc_eem.c 				goto next;
next              323 drivers/net/usb/cdc_eem.c next:
next             1695 drivers/net/usb/r8152.c 		cursor = tp->tx_free.next;
next             1725 drivers/net/usb/r8152.c 			segs = segs->next;
next             1726 drivers/net/usb/r8152.c 			nskb->next = NULL;
next             2044 drivers/net/usb/r8152.c 	struct list_head *cursor, *next, rx_queue;
next             2074 drivers/net/usb/r8152.c 	list_for_each_safe(cursor, next, &rx_queue) {
next             2176 drivers/net/usb/r8152.c 			list_add_tail(&agg->list, next);
next              962 drivers/net/virtio_net.c 				curr_skb->next = nskb;
next              177 drivers/net/vxlan.c 	return list_entry_rcu(fdb->remotes.next, struct vxlan_rdst, list);
next             1366 drivers/net/vxlan.c 	list_for_each_entry(vxlan, &vn->vxlan_list, next) {
next             3053 drivers/net/vxlan.c 	INIT_LIST_HEAD(&vxlan->next);
next             3476 drivers/net/vxlan.c 	list_for_each_entry(tmp, &vn->vxlan_list, next) {
next             3640 drivers/net/vxlan.c 	list_add(&vxlan->next, &vn->vxlan_list);
next             4037 drivers/net/vxlan.c 	list_del(&vxlan->next);
next             4229 drivers/net/vxlan.c 	struct vxlan_dev *vxlan, *next;
next             4232 drivers/net/vxlan.c 	list_for_each_entry_safe(vxlan, next, &vn->vxlan_list, next) {
next             4414 drivers/net/vxlan.c 	struct vxlan_dev *vxlan, *next;
next             4422 drivers/net/vxlan.c 	list_for_each_entry_safe(vxlan, next, &vn->vxlan_list, next) {
next              170 drivers/net/wan/hd64572.h 	unsigned long	next;		/* pointer to next block descriptor */
next              214 drivers/net/wan/hdlc.c 		proto = proto->next;
next              321 drivers/net/wan/hdlc.c 	proto->next = first_proto;
next              335 drivers/net/wan/hdlc.c 		p = &((*p)->next);
next              337 drivers/net/wan/hdlc.c 	*p = proto->next;
next              119 drivers/net/wan/hdlc_fr.c 	struct pvc_device *next;	/* Sorted in ascending DLCI order */
next              184 drivers/net/wan/hdlc_fr.c 		pvc = pvc->next;
next              201 drivers/net/wan/hdlc_fr.c 		pvc_p = &(*pvc_p)->next;
next              213 drivers/net/wan/hdlc_fr.c 	pvc->next = *pvc_p;	/* Put it in the chain */
next              255 drivers/net/wan/hdlc_fr.c 			*pvc_p = pvc->next;
next              259 drivers/net/wan/hdlc_fr.c 		pvc_p = &(*pvc_p)->next;
next              551 drivers/net/wan/hdlc_fr.c 			pvc = pvc->next;
next              581 drivers/net/wan/hdlc_fr.c 				pvc = pvc->next;
next              592 drivers/net/wan/hdlc_fr.c 			pvc = pvc->next;
next              756 drivers/net/wan/hdlc_fr.c 				pvc = pvc->next;
next              785 drivers/net/wan/hdlc_fr.c 		pvc = pvc->next;
next              856 drivers/net/wan/hdlc_fr.c 		pvc = pvc->next;
next             1035 drivers/net/wan/hdlc_fr.c 		pvc = pvc->next;
next             1152 drivers/net/wan/hdlc_fr.c 		struct pvc_device *next = pvc->next;
next             1161 drivers/net/wan/hdlc_fr.c 		pvc = next;
next              290 drivers/net/wan/ixp4xx_hss.c 	u32 next;		/* pointer to next buffer, unused */
next              574 drivers/net/wan/ixp4xx_hss.c 	       phys, desc->next, desc->buf_len, desc->pkt_len,
next              594 drivers/net/wan/ixp4xx_hss.c 	BUG_ON(tab[n_desc].next);
next             1259 drivers/net/wimax/i2400m/fw.c 	const void *itr, *next, *top;
next             1264 drivers/net/wimax/i2400m/fw.c 	     headers++, itr = next) {
next             1281 drivers/net/wimax/i2400m/fw.c 			next = itr + size;
next             1283 drivers/net/wimax/i2400m/fw.c 			next = itr + header_len;
next              628 drivers/net/wireless/ath/ar5523/ar5523.c 			data = (struct ar5523_rx_data *) ar->rx_data_free.next;
next              675 drivers/net/wireless/ath/ar5523/ar5523.c 			data = (struct ar5523_rx_data *) ar->rx_data_used.next;
next              695 drivers/net/wireless/ath/ar5523/ar5523.c 		data = (struct ar5523_rx_data *) ar->rx_data_free.next;
next              808 drivers/net/wireless/ath/ar5523/ar5523.c 				ar->tx_queue_pending.next;
next              522 drivers/net/wireless/ath/ath10k/htt_rx.c 			prev_frag_buf->next = NULL;
next              530 drivers/net/wireless/ath/ath10k/htt_rx.c 		prev_frag_buf->next = frag_buf;
next              541 drivers/net/wireless/ath/ath10k/htt_rx.c 	prev_frag_buf->next = NULL;
next              598 drivers/net/wireless/ath/ath10k/htt_rx.c 			prev_frag_buf->next = NULL;
next              606 drivers/net/wireless/ath/ath10k/htt_rx.c 		prev_frag_buf->next = frag_buf;
next              617 drivers/net/wireless/ath/ath10k/htt_rx.c 	prev_frag_buf->next = NULL;
next             1579 drivers/net/wireless/ath/ath10k/mac.c 	const u8 *next;
next             1594 drivers/net/wireless/ath/ath10k/mac.c 	next = ie + len;
next             1596 drivers/net/wireless/ath/ath10k/mac.c 	if (WARN_ON(next > end))
next             1599 drivers/net/wireless/ath/ath10k/mac.c 	memmove(ie, next, end - next);
next             7284 drivers/net/wireless/ath/ath10k/mac.c 		goto next;
next             7319 drivers/net/wireless/ath/ath10k/mac.c next:
next             1582 drivers/net/wireless/ath/ath5k/base.c 				goto next;
next             1595 drivers/net/wireless/ath/ath5k/base.c next:
next              166 drivers/net/wireless/ath/ath5k/debug.c 	.next  = reg_next,
next              824 drivers/net/wireless/ath/ath6kl/htc_pipe.c 	target->pipe.htc_packet_pool = (struct htc_packet *) packet->list.next;
next              828 drivers/net/wireless/ath/ath6kl/htc_pipe.c 	packet->list.next = NULL;
next              841 drivers/net/wireless/ath/ath6kl/htc_pipe.c 		packet->list.next = NULL;
next              844 drivers/net/wireless/ath/ath6kl/htc_pipe.c 		packet->list.next = lh;
next              330 drivers/net/wireless/ath/ath6kl/main.c 				    le32_to_cpu(debug_buf.next));
next              343 drivers/net/wireless/ath/ath6kl/target.h 	__le32 next;
next              160 drivers/net/wireless/ath/ath9k/htc_drv_beacon.c 				goto next;
next              170 drivers/net/wireless/ath/ath9k/htc_drv_beacon.c 			goto next;
next              184 drivers/net/wireless/ath/ath9k/htc_drv_beacon.c 	next:
next              737 drivers/net/wireless/ath/ath9k/recv.c 		tbf = list_entry(bf->list.next, struct ath_rxbuf, list);
next             2593 drivers/net/wireless/ath/ath9k/xmit.c 			bf = list_entry(bf_held->list.next, struct ath_buf,
next              572 drivers/net/wireless/ath/carl9170/tx.c 			goto next;
next              581 drivers/net/wireless/ath/carl9170/tx.c next:
next               59 drivers/net/wireless/ath/wcn36xx/dxe.c 	struct wcn36xx_dxe_ctl *ctl = ch->head_blk_ctl, *next;
next               63 drivers/net/wireless/ath/wcn36xx/dxe.c 		next = ctl->next;
next               65 drivers/net/wireless/ath/wcn36xx/dxe.c 		ctl = next;
next               86 drivers/net/wireless/ath/wcn36xx/dxe.c 			prev_ctl->next = cur_ctl;
next               87 drivers/net/wireless/ath/wcn36xx/dxe.c 			cur_ctl->next = ch->head_blk_ctl;
next               89 drivers/net/wireless/ath/wcn36xx/dxe.c 			prev_ctl->next = cur_ctl;
next              219 drivers/net/wireless/ath/wcn36xx/dxe.c 		cur_ctl = cur_ctl->next;
next              255 drivers/net/wireless/ath/wcn36xx/dxe.c 		cur = cur->next;
next              310 drivers/net/wireless/ath/wcn36xx/dxe.c 		cur_ctl = cur_ctl->next;
next              324 drivers/net/wireless/ath/wcn36xx/dxe.c 		cur = cur->next;
next              389 drivers/net/wireless/ath/wcn36xx/dxe.c 		ctl = ctl->next;
next              579 drivers/net/wireless/ath/wcn36xx/dxe.c 		ctl = ctl->next;
next              692 drivers/net/wireless/ath/wcn36xx/dxe.c 	ctl_skb = ctl_bd->next;
next              753 drivers/net/wireless/ath/wcn36xx/dxe.c 	 ch->head_blk_ctl = ctl_skb->next;
next              420 drivers/net/wireless/ath/wcn36xx/dxe.h 	struct wcn36xx_dxe_ctl	*next;
next             2327 drivers/net/wireless/ath/wil6210/cfg80211.c 		ret = vif->probe_client_pending.next;
next             3379 drivers/net/wireless/ath/wil6210/wmi.c 		ret = wil->pending_wmi_ev.next;
next             1071 drivers/net/wireless/atmel/atmel.c 			goto next;
next             1079 drivers/net/wireless/atmel/atmel.c 			goto next;
next             1137 drivers/net/wireless/atmel/atmel.c 					goto next;
next             1145 drivers/net/wireless/atmel/atmel.c next:
next              453 drivers/net/wireless/broadcom/b43/pio.c 	pack = list_entry(q->packets_list.next,
next              121 drivers/net/wireless/broadcom/b43/sdio.c 		tuple = tuple->next;
next              462 drivers/net/wireless/broadcom/b43legacy/pio.c 	packet = list_entry(queue->txfree.next, struct b43legacy_pio_txpacket,
next             4200 drivers/net/wireless/broadcom/brcm80211/brcmfmac/cfg80211.c 			goto next;
next             4206 drivers/net/wireless/broadcom/brcm80211/brcmfmac/cfg80211.c 			goto next;
next             4213 drivers/net/wireless/broadcom/brcm80211/brcmfmac/cfg80211.c 			goto next;
next             4231 drivers/net/wireless/broadcom/brcm80211/brcmfmac/cfg80211.c next:
next               52 drivers/net/wireless/broadcom/brcm80211/brcmfmac/flowring.c 		search = search->next;
next              400 drivers/net/wireless/broadcom/brcm80211/brcmfmac/flowring.c 		search = search->next;
next              453 drivers/net/wireless/broadcom/brcm80211/brcmfmac/flowring.c 		search = search->next;
next              470 drivers/net/wireless/broadcom/brcm80211/brcmfmac/flowring.c 			prev->next = search->next;
next              472 drivers/net/wireless/broadcom/brcm80211/brcmfmac/flowring.c 			flow->tdls_entry = search->next;
next              491 drivers/net/wireless/broadcom/brcm80211/brcmfmac/flowring.c 	tdls_entry->next = NULL;
next              498 drivers/net/wireless/broadcom/brcm80211/brcmfmac/flowring.c 		while (search->next) {
next              499 drivers/net/wireless/broadcom/brcm80211/brcmfmac/flowring.c 			search = search->next;
next              503 drivers/net/wireless/broadcom/brcm80211/brcmfmac/flowring.c 		search->next = tdls_entry;
next               35 drivers/net/wireless/broadcom/brcm80211/brcmfmac/flowring.h 	struct brcmf_flowring_tdls_entry *next;
next             1257 drivers/net/wireless/broadcom/brcm80211/brcmfmac/sdio.c 	struct sk_buff *cur, *next;
next             1259 drivers/net/wireless/broadcom/brcm80211/brcmfmac/sdio.c 	skb_queue_walk_safe(&bus->glom, cur, next) {
next             1675 drivers/net/wireless/broadcom/brcm80211/brcmfmac/sdio.c 					   pfirst->len, pfirst->next,
next              382 drivers/net/wireless/broadcom/brcm80211/brcmfmac/usb.c 	req = list_entry(q->next, struct brcmf_usbreq, list);
next              383 drivers/net/wireless/broadcom/brcm80211/brcmfmac/usb.c 	list_del_init(q->next);
next              428 drivers/net/wireless/broadcom/brcm80211/brcmfmac/usb.c 		req = list_entry(q->next, struct brcmf_usbreq, list);
next              431 drivers/net/wireless/broadcom/brcm80211/brcmfmac/usb.c 		list_del(q->next);
next              440 drivers/net/wireless/broadcom/brcm80211/brcmfmac/usb.c 	struct brcmf_usbreq *req, *next;
next              442 drivers/net/wireless/broadcom/brcm80211/brcmfmac/usb.c 	list_for_each_entry_safe(req, next, q, list) {
next              153 drivers/net/wireless/broadcom/brcm80211/brcmsmac/aiutils.h 	struct gpioh_item *next;
next              929 drivers/net/wireless/broadcom/brcm80211/brcmsmac/dma.c 	struct sk_buff *p, *next;
next              978 drivers/net/wireless/broadcom/brcm80211/brcmsmac/dma.c 			skb_queue_walk_safe(&dma_frames, p, next) {
next              289 drivers/net/wireless/broadcom/brcm80211/brcmsmac/mac80211_if.c 	struct brcms_timer *t, *next;
next              319 drivers/net/wireless/broadcom/brcm80211/brcmsmac/mac80211_if.c 	for (t = wl->timers; t; t = next) {
next              320 drivers/net/wireless/broadcom/brcm80211/brcmsmac/mac80211_if.c 		next = t->next;
next             1501 drivers/net/wireless/broadcom/brcm80211/brcmsmac/mac80211_if.c 	t->next = wl->timers;
next             1567 drivers/net/wireless/broadcom/brcm80211/brcmsmac/mac80211_if.c 		wl->timers = wl->timers->next;
next             1578 drivers/net/wireless/broadcom/brcm80211/brcmsmac/mac80211_if.c 		if (tmp->next == t) {
next             1579 drivers/net/wireless/broadcom/brcm80211/brcmsmac/mac80211_if.c 			tmp->next = t->next;
next             1586 drivers/net/wireless/broadcom/brcm80211/brcmsmac/mac80211_if.c 		tmp = tmp->next;
next               44 drivers/net/wireless/broadcom/brcm80211/brcmsmac/mac80211_if.h 	struct brcms_timer *next;	/* for freeing on unload */
next             7710 drivers/net/wireless/broadcom/brcm80211/brcmsmac/main.c 	struct sk_buff *next = NULL;
next             7733 drivers/net/wireless/broadcom/brcm80211/brcmsmac/main.c 	skb_queue_walk_safe(&recv_frames, p, next) {
next              564 drivers/net/wireless/broadcom/brcm80211/brcmsmac/phy/phy_cmn.c 	pi->next = pi->sh->phy_head;
next              590 drivers/net/wireless/broadcom/brcm80211/brcmsmac/phy/phy_cmn.c 			pi->sh->phy_head = pi->next;
next              591 drivers/net/wireless/broadcom/brcm80211/brcmsmac/phy/phy_cmn.c 		else if (pi->sh->phy_head->next == pi)
next              592 drivers/net/wireless/broadcom/brcm80211/brcmsmac/phy/phy_cmn.c 			pi->sh->phy_head->next = NULL;
next              542 drivers/net/wireless/broadcom/brcm80211/brcmsmac/phy/phy_int.h 	struct brcms_phy *next;
next               38 drivers/net/wireless/broadcom/brcm80211/brcmutil/utils.c 	WARN_ON(skb->next);
next              111 drivers/net/wireless/broadcom/brcm80211/brcmutil/utils.c 	struct sk_buff *p, *next;
next              114 drivers/net/wireless/broadcom/brcm80211/brcmutil/utils.c 	skb_queue_walk_safe(q, p, next) {
next              145 drivers/net/wireless/broadcom/brcm80211/brcmutil/utils.c 	struct sk_buff *p, *next;
next              148 drivers/net/wireless/broadcom/brcm80211/brcmutil/utils.c 	skb_queue_walk_safe(q, p, next) {
next              305 drivers/net/wireless/broadcom/brcm80211/brcmutil/utils.c 	for (p = p0; p; p = p->next)
next             3024 drivers/net/wireless/cisco/airo.c 			tmp_net = list_entry(ai->network_free_list.next,
next             3026 drivers/net/wireless/cisco/airo.c 			list_del(ai->network_free_list.next);
next             4039 drivers/net/wireless/cisco/airo.c 	u16 next;
next             4043 drivers/net/wireless/cisco/airo.c 	next = IN4500(ai, AUXDATA);
next             4046 drivers/net/wireless/cisco/airo.c 	return next;
next             4056 drivers/net/wireless/cisco/airo.c 	u16 next;
next             4064 drivers/net/wireless/cisco/airo.c 	next = aux_setup(ai, page, offset, &len);
next             4078 drivers/net/wireless/cisco/airo.c 			next = aux_setup(ai, next, 4, &len);
next             5696 drivers/net/wireless/cisco/airo.c 		ai = list_entry(airo_devices.next, struct airo_info, dev_list);
next              760 drivers/net/wireless/intel/ipw2x00/ipw2100.c 	element = priv->msg_free_list.next;
next             2681 drivers/net/wireless/intel/ipw2x00/ipw2100.c 	i = (rxq->next + 1) % rxq->entries;
next             2760 drivers/net/wireless/intel/ipw2x00/ipw2100.c 		rxq->next = (i ? i : rxq->entries) - 1;
next             2763 drivers/net/wireless/intel/ipw2x00/ipw2100.c 			       IPW_MEM_HOST_SHARED_RX_WRITE_INDEX, rxq->next);
next             2819 drivers/net/wireless/intel/ipw2x00/ipw2100.c 	element = priv->fw_pend_list.next;
next             2854 drivers/net/wireless/intel/ipw2x00/ipw2100.c 	if (w != txq->next)
next             3002 drivers/net/wireless/intel/ipw2x00/ipw2100.c 	int next = txq->next;
next             3016 drivers/net/wireless/intel/ipw2x00/ipw2100.c 		element = priv->msg_pend_list.next;
next             3023 drivers/net/wireless/intel/ipw2x00/ipw2100.c 			     &txq->drv[txq->next],
next             3024 drivers/net/wireless/intel/ipw2x00/ipw2100.c 			     (u32) (txq->nic + txq->next *
next             3027 drivers/net/wireless/intel/ipw2x00/ipw2100.c 		packet->index = txq->next;
next             3029 drivers/net/wireless/intel/ipw2x00/ipw2100.c 		tbd = &txq->drv[txq->next];
next             3042 drivers/net/wireless/intel/ipw2x00/ipw2100.c 		txq->next++;
next             3043 drivers/net/wireless/intel/ipw2x00/ipw2100.c 		txq->next %= txq->entries;
next             3051 drivers/net/wireless/intel/ipw2x00/ipw2100.c 	if (txq->next != next) {
next             3057 drivers/net/wireless/intel/ipw2x00/ipw2100.c 			       txq->next);
next             3071 drivers/net/wireless/intel/ipw2x00/ipw2100.c 	int next = txq->next;
next             3083 drivers/net/wireless/intel/ipw2x00/ipw2100.c 		element = priv->tx_pend_list.next;
next             3103 drivers/net/wireless/intel/ipw2x00/ipw2100.c 		tbd = &txq->drv[txq->next];
next             3105 drivers/net/wireless/intel/ipw2x00/ipw2100.c 		packet->index = txq->next;
next             3142 drivers/net/wireless/intel/ipw2x00/ipw2100.c 		txq->next++;
next             3143 drivers/net/wireless/intel/ipw2x00/ipw2100.c 		txq->next %= txq->entries;
next             3154 drivers/net/wireless/intel/ipw2x00/ipw2100.c 			tbd = &txq->drv[txq->next];
next             3181 drivers/net/wireless/intel/ipw2x00/ipw2100.c 				     txq->next, tbd->host_addr,
next             3189 drivers/net/wireless/intel/ipw2x00/ipw2100.c 			txq->next++;
next             3190 drivers/net/wireless/intel/ipw2x00/ipw2100.c 			txq->next %= txq->entries;
next             3200 drivers/net/wireless/intel/ipw2x00/ipw2100.c 	if (txq->next != next) {
next             3205 drivers/net/wireless/intel/ipw2x00/ipw2100.c 			       txq->next);
next             3401 drivers/net/wireless/intel/ipw2x00/ipw2100.c 	element = priv->tx_free_list.next;
next             4398 drivers/net/wireless/intel/ipw2x00/ipw2100.c 	write_register(priv->net_dev, w, q->next);
next             4510 drivers/net/wireless/intel/ipw2x00/ipw2100.c 	priv->tx_queue.next = 0;
next             4632 drivers/net/wireless/intel/ipw2x00/ipw2100.c 	priv->rx_queue.next = priv->rx_queue.entries - 1;
next             5816 drivers/net/wireless/intel/ipw2x00/ipw2100.c 		element = priv->tx_pend_list.next;
next              186 drivers/net/wireless/intel/ipw2x00/ipw2100.h 	u32 next;
next             5149 drivers/net/wireless/intel/ipw2x00/ipw2200.c 		element = rxq->rx_free.next;
next             5187 drivers/net/wireless/intel/ipw2x00/ipw2200.c 		element = rxq->rx_used.next;
next             7598 drivers/net/wireless/intel/ipw2x00/ipw2200.c 		element = priv->ieee->network_free_list.next;
next             1603 drivers/net/wireless/intel/ipw2x00/libipw_rx.c 			target = list_entry(ieee->network_free_list.next,
next             1605 drivers/net/wireless/intel/ipw2x00/libipw_rx.c 			list_del(ieee->network_free_list.next);
next              255 drivers/net/wireless/intel/iwlegacy/3945-mac.c 		element = il->free_frames.next;
next              284 drivers/net/wireless/intel/iwlegacy/3945-mac.c 	element = il->free_frames.next;
next              943 drivers/net/wireless/intel/iwlegacy/3945-mac.c 		element = rxq->rx_free.next;
next             1042 drivers/net/wireless/intel/iwlegacy/3945-mac.c 		element = rxq->rx_used.next;
next              263 drivers/net/wireless/intel/iwlegacy/4965-mac.c 		element = rxq->rx_free.next;
next              364 drivers/net/wireless/intel/iwlegacy/4965-mac.c 		element = rxq->rx_used.next;
next             3694 drivers/net/wireless/intel/iwlegacy/4965-mac.c 		element = il->free_frames.next;
next             3723 drivers/net/wireless/intel/iwlegacy/4965-mac.c 	element = il->free_frames.next;
next             1987 drivers/net/wireless/intel/iwlwifi/fw/dbg.c 			list_entry(list->next, typeof(*entry), list);
next              822 drivers/net/wireless/intel/iwlwifi/mvm/tx.c 	struct sk_buff *tmp, *next;
next              835 drivers/net/wireless/intel/iwlwifi/mvm/tx.c 	next = skb_gso_segment(skb, netdev_flags);
next              837 drivers/net/wireless/intel/iwlwifi/mvm/tx.c 	if (WARN_ON_ONCE(IS_ERR(next)))
next              839 drivers/net/wireless/intel/iwlwifi/mvm/tx.c 	else if (next)
next              842 drivers/net/wireless/intel/iwlwifi/mvm/tx.c 	while (next) {
next              843 drivers/net/wireless/intel/iwlwifi/mvm/tx.c 		tmp = next;
next              844 drivers/net/wireless/intel/iwlwifi/mvm/tx.c 		next = tmp->next;
next              876 drivers/net/wireless/intel/iwlwifi/mvm/tx.c 		tmp->next = NULL;
next              374 drivers/net/wireless/intersil/hostap/hostap_ap.c 	.next	= ap_control_proc_next,
next              404 drivers/net/wireless/intersil/hostap/hostap_ap.c 	for (ptr = mac_restrictions->mac_list.next;
next              405 drivers/net/wireless/intersil/hostap/hostap_ap.c 	     ptr != &mac_restrictions->mac_list; ptr = ptr->next) {
next              455 drivers/net/wireless/intersil/hostap/hostap_ap.c 	for (ptr = mac_restrictions->mac_list.next, n = ptr->next;
next              457 drivers/net/wireless/intersil/hostap/hostap_ap.c 	     ptr = n, n = ptr->next) {
next              504 drivers/net/wireless/intersil/hostap/hostap_ap.c 	for (ptr = ap->sta_list.next, n = ptr->next; ptr != &ap->sta_list;
next              505 drivers/net/wireless/intersil/hostap/hostap_ap.c 	     ptr = n, n = ptr->next) {
next              576 drivers/net/wireless/intersil/hostap/hostap_ap.c 	.next	= prism2_ap_proc_next,
next             1091 drivers/net/wireless/intersil/hostap/hostap_ap.c 		entry = entry->next;
next             1126 drivers/net/wireless/intersil/hostap/hostap_ap.c 			entry->next = ap->add_sta_proc_entries;
next             1975 drivers/net/wireless/intersil/hostap/hostap_ap.c 		entry = entry->next;
next             2335 drivers/net/wireless/intersil/hostap/hostap_ap.c 	for (ptr = ap->sta_list.next; ptr != NULL && ptr != &ap->sta_list;
next             2336 drivers/net/wireless/intersil/hostap/hostap_ap.c 	     ptr = ptr->next) {
next             2387 drivers/net/wireless/intersil/hostap/hostap_ap.c 	for (ptr = ap->sta_list.next; ptr != NULL && ptr != &ap->sta_list;
next             2388 drivers/net/wireless/intersil/hostap/hostap_ap.c 	     ptr = ptr->next) {
next             3263 drivers/net/wireless/intersil/hostap/hostap_ap.c 	entry->next = local->ap->wds_oper_entries;
next              153 drivers/net/wireless/intersil/hostap/hostap_ap.h 	struct add_sta_proc_data *next;
next              161 drivers/net/wireless/intersil/hostap/hostap_ap.h 	struct wds_oper_data *next;
next              219 drivers/net/wireless/intersil/hostap/hostap_download.c 	.next	= prism2_download_aux_dump_proc_next,
next              656 drivers/net/wireless/intersil/hostap/hostap_hw.c 		entry = list_entry(local->cmd_queue.next,
next              702 drivers/net/wireless/intersil/hostap/hostap_hw.c 		entry = list_entry(local->cmd_queue.next,
next             2183 drivers/net/wireless/intersil/hostap/hostap_hw.c 		cb = cb->next;
next             3007 drivers/net/wireless/intersil/hostap/hostap_hw.c 			entry = list_entry(local->set_tim_list.next,
next             3311 drivers/net/wireless/intersil/hostap/hostap_hw.c 		tx_cb = tx_cb->next;
next              250 drivers/net/wireless/intersil/hostap/hostap_main.c 	entry->next = local->tx_callback;
next              267 drivers/net/wireless/intersil/hostap/hostap_main.c 		cb = cb->next;
next              271 drivers/net/wireless/intersil/hostap/hostap_main.c 			local->tx_callback = cb->next;
next              273 drivers/net/wireless/intersil/hostap/hostap_main.c 			prev->next = cb->next;
next              119 drivers/net/wireless/intersil/hostap/hostap_proc.c 	.next	= prism2_wds_proc_next,
next              172 drivers/net/wireless/intersil/hostap/hostap_proc.c 	.next	= prism2_bss_list_proc_next,
next              349 drivers/net/wireless/intersil/hostap/hostap_proc.c 	.next	= prism2_scan_results_proc_next,
next              531 drivers/net/wireless/intersil/hostap/hostap_wlan.h 	struct hostap_tx_callback_info *next;
next               79 drivers/net/wireless/intersil/orinoco/hermes_dld.c 	char next[0];		/* next PDR starts here */
next              169 drivers/net/wireless/intersil/orinoco/hermes_dld.c 		pdr = (struct pdr *) pdr->next;
next              459 drivers/net/wireless/intersil/orinoco/orinoco_usb.c 	    list_entry(upriv->req_pending.next, struct request_context,
next              106 drivers/net/wireless/intersil/p54/main.c 	u8 *next;
next              114 drivers/net/wireless/intersil/p54/main.c 	next = tim + 2 + dtim_len;
next              119 drivers/net/wireless/intersil/p54/main.c 	memmove(tim, next, skb_tail_pointer(skb) - next);
next              433 drivers/net/wireless/intersil/p54/p54spi.c 		entry = list_entry(priv->tx_pending.next,
next             1795 drivers/net/wireless/intersil/prism54/isl_ioctl.c 	struct list_head *ptr, *next;
next             1805 drivers/net/wireless/intersil/prism54/isl_ioctl.c 	for (ptr = acl->mac_list.next, next = ptr->next;
next             1806 drivers/net/wireless/intersil/prism54/isl_ioctl.c 	     ptr != &acl->mac_list; ptr = next, next = ptr->next) {
next               98 drivers/net/wireless/marvell/libertas/firmware.c next:
next              107 drivers/net/wireless/marvell/libertas/firmware.c 		goto next;
next              187 drivers/net/wireless/marvell/libertas/firmware.c 			goto next;
next              192 drivers/net/wireless/marvell/libertas/firmware.c 				goto next;
next              216 drivers/net/wireless/marvell/libertas/firmware.c   next:
next              104 drivers/net/wireless/marvell/libertas/if_sdio.c 	struct if_sdio_packet	*next;
next              409 drivers/net/wireless/marvell/libertas/if_sdio.c 			card->packets = packet->next;
next              937 drivers/net/wireless/marvell/libertas/if_sdio.c 	packet->next = NULL;
next              956 drivers/net/wireless/marvell/libertas/if_sdio.c 		while (cur->next)
next              957 drivers/net/wireless/marvell/libertas/if_sdio.c 			cur = cur->next;
next              958 drivers/net/wireless/marvell/libertas/if_sdio.c 		cur->next = packet;
next             1240 drivers/net/wireless/marvell/libertas/if_sdio.c 		card->packets = card->packets->next;
next             1284 drivers/net/wireless/marvell/libertas/if_sdio.c 		card->packets = card->packets->next;
next               79 drivers/net/wireless/marvell/libertas/if_spi.c 	struct list_head *cursor, *next;
next               82 drivers/net/wireless/marvell/libertas/if_spi.c 	list_for_each_safe(cursor, next, &card->cmd_packet_list) {
next               87 drivers/net/wireless/marvell/libertas/if_spi.c 	list_for_each_safe(cursor, next, &card->data_packet_list) {
next              898 drivers/net/wireless/marvell/libertas/if_spi.c 					cmd_packet_list.next);
next              912 drivers/net/wireless/marvell/libertas/if_spi.c 					data_packet_list.next);
next             5219 drivers/net/wireless/marvell/mwl8k.c 		return list_entry(priv->vif_list.next, struct mwl8k_vif, list);
next              303 drivers/net/wireless/mediatek/mt76/dma.c 	skb->prev = skb->next = NULL;
next              473 drivers/net/wireless/mediatek/mt76/mt76x02_dfs.c 			goto next;
next              498 drivers/net/wireless/mediatek/mt76/mt76x02_dfs.c 			goto next;
next              507 drivers/net/wireless/mediatek/mt76/mt76x02_dfs.c next:
next               49 drivers/net/wireless/mediatek/mt76/mt76x02_usb_core.c 		if (!iter->next) {
next              782 drivers/net/wireless/mediatek/mt76/usb.c 	skb->prev = skb->next = NULL;
next             1936 drivers/net/wireless/realtek/rtlwifi/base.c 	struct rtl_bssid_entry *entry, *next;
next             1938 drivers/net/wireless/realtek/rtlwifi/base.c 	list_for_each_entry_safe(entry, next, &rtlpriv->scan_list.list, list) {
next             1967 drivers/net/wireless/realtek/rtlwifi/base.c 	struct rtl_bssid_entry *entry, *next;
next             1972 drivers/net/wireless/realtek/rtlwifi/base.c 	list_for_each_entry_safe(entry, next, &rtlpriv->scan_list.list, list) {
next              118 drivers/net/wireless/realtek/rtlwifi/pci.h 	u8 next;
next              122 drivers/net/wireless/rsi/rsi_91x_core.c 			skb = skb->next;
next             3916 drivers/net/wireless/ti/wlcore/main.c 	const u8 *next, *end = skb->data + skb->len;
next             3922 drivers/net/wireless/ti/wlcore/main.c 	next = ie + len;
next             3923 drivers/net/wireless/ti/wlcore/main.c 	memmove(ie, next, end - next);
next             3932 drivers/net/wireless/ti/wlcore/main.c 	const u8 *next, *end = skb->data + skb->len;
next             3939 drivers/net/wireless/ti/wlcore/main.c 	next = ie + len;
next             3940 drivers/net/wireless/ti/wlcore/main.c 	memmove(ie, next, end - next);
next              625 drivers/net/wireless/ti/wlcore/tx.c 				goto next;
next              634 drivers/net/wireless/ti/wlcore/tx.c next:
next              296 drivers/net/wireless/wl3501_cs.c 	u16 next, blk_cnt = 0, zero = 0;
next              308 drivers/net/wireless/wl3501_cs.c 		wl3501_get_from_wla(this, this->tx_buffer_head, &next,
next              309 drivers/net/wireless/wl3501_cs.c 				    sizeof(next));
next              313 drivers/net/wireless/wl3501_cs.c 		this->tx_buffer_head = next;
next              316 drivers/net/wireless/wl3501_cs.c 		if (!next && full_len) {
next              339 drivers/net/wireless/wl3501_cs.c 		u16 next;
next              342 drivers/net/wireless/wl3501_cs.c 		wl3501_get_from_wla(this, ptr, &next, sizeof(next));
next              344 drivers/net/wireless/wl3501_cs.c 		ptr = next;
next              468 drivers/net/wireless/wl3501_cs.c 	u16 bf, sig_bf, next, tmplen, pktlen;
next              511 drivers/net/wireless/wl3501_cs.c 			wl3501_get_from_wla(this, bf, &next, sizeof(next));
next              512 drivers/net/wireless/wl3501_cs.c 			bf = next;
next              528 drivers/net/wireless/wl3501_cs.c 			wl3501_get_from_wla(this, bf, &next, sizeof(next));
next              529 drivers/net/wireless/wl3501_cs.c 			bf = next;
next              541 drivers/net/wireless/wl3501_cs.c 			wl3501_get_from_wla(this, bf, &next, sizeof(next));
next              542 drivers/net/wireless/wl3501_cs.c 			bf = next;
next             1179 drivers/net/wireless/wl3501_cs.c 	u16 ptr, next;
next             1208 drivers/net/wireless/wl3501_cs.c 	next = ptr + WL3501_BLKSZ;
next             1209 drivers/net/wireless/wl3501_cs.c 	while ((next - this->tx_buffer_head) < this->tx_buffer_size) {
next             1211 drivers/net/wireless/wl3501_cs.c 		wl3501_set_to_wla(this, ptr, &next, sizeof(next));
next             1212 drivers/net/wireless/wl3501_cs.c 		ptr = next;
next             1213 drivers/net/wireless/wl3501_cs.c 		next = ptr + WL3501_BLKSZ;
next             1216 drivers/net/wireless/wl3501_cs.c 	next = 0;
next             1217 drivers/net/wireless/wl3501_cs.c 	wl3501_set_to_wla(this, ptr, &next, sizeof(next));
next              310 drivers/net/xen-netback/rx.c 			pkt->frag_iter = frag_iter->next;
next              809 drivers/net/xen-netfront.c 			goto next;
next              822 drivers/net/xen-netfront.c 			goto next;
next              832 drivers/net/xen-netfront.c next:
next             1339 drivers/nfc/pn533/pn533.c 	u8 *next, nfcid3[NFC_NFCID3_MAXSIZE];
next             1369 drivers/nfc/pn533/pn533.c 	next = skb_put(skb, 1);  /* Next */
next             1370 drivers/nfc/pn533/pn533.c 	*next = 0;
next             1374 drivers/nfc/pn533/pn533.c 	*next |= 1;
next             1378 drivers/nfc/pn533/pn533.c 	*next |= 2;
next             1381 drivers/nfc/pn533/pn533.c 	*next |= 4; /* We have some Gi */
next             1807 drivers/nfc/pn533/pn533.c 	u8 *next, *arg, nfcid3[NFC_NFCID3_MAXSIZE];
next             1842 drivers/nfc/pn533/pn533.c 	next = skb_put(skb, 1);  /* Next */
next             1843 drivers/nfc/pn533/pn533.c 	*next = 0;
next             1847 drivers/nfc/pn533/pn533.c 	*next |= 1;
next             1855 drivers/nfc/pn533/pn533.c 	*next |= 2;
next             1859 drivers/nfc/pn533/pn533.c 		*next |= 4; /* We have some Gi */
next             1861 drivers/nfc/pn533/pn533.c 		*next = 0;
next              106 drivers/nvdimm/badrange.c 	struct badrange_entry *bre, *next;
next              118 drivers/nvdimm/badrange.c 	list_for_each_entry_safe(bre, next, badrange_list, list) {
next              804 drivers/nvdimm/btt.c 	struct arena_info *arena, *next;
next              806 drivers/nvdimm/btt.c 	list_for_each_entry_safe(arena, next, &btt->arena_list, list) {
next              416 drivers/nvdimm/bus.c 	struct badrange_entry *bre, *next;
next              418 drivers/nvdimm/bus.c 	list_for_each_entry_safe(bre, next, badrange_list, list) {
next              565 drivers/nvdimm/namespace_devs.c 		struct resource *next, struct resource *exist,
next              635 drivers/nvdimm/namespace_devs.c 		struct resource *next = res->sibling, *new_res = NULL;
next              651 drivers/nvdimm/namespace_devs.c 			space_valid(nd_region, ndd, label_id, NULL, next, exist,
next              659 drivers/nvdimm/namespace_devs.c 		if (!loc && next) {
next              661 drivers/nvdimm/namespace_devs.c 			valid.end = min(mapping_end, next->start - 1);
next              662 drivers/nvdimm/namespace_devs.c 			space_valid(nd_region, ndd, label_id, res, next, exist,
next              670 drivers/nvdimm/namespace_devs.c 		if (!loc && !next) {
next              673 drivers/nvdimm/namespace_devs.c 			space_valid(nd_region, ndd, label_id, res, next, exist,
next              694 drivers/nvdimm/namespace_devs.c 			if (strcmp(next->name, label_id->id) == 0) {
next              696 drivers/nvdimm/namespace_devs.c 				rc = adjust_resource(next, next->start
next              697 drivers/nvdimm/namespace_devs.c 						- allocate, resource_size(next)
next              699 drivers/nvdimm/namespace_devs.c 				new_res = next;
next              777 drivers/nvdimm/namespace_devs.c 		struct resource *next = res->sibling;
next              780 drivers/nvdimm/namespace_devs.c 		if (!next || strcmp(res->name, label_id->id) != 0
next              781 drivers/nvdimm/namespace_devs.c 				|| strcmp(next->name, label_id->id) != 0
next              782 drivers/nvdimm/namespace_devs.c 				|| end != next->start)
next              784 drivers/nvdimm/namespace_devs.c 		end += resource_size(next);
next              785 drivers/nvdimm/namespace_devs.c 		nvdimm_free_dpa(ndd, next);
next               99 drivers/nvdimm/nd.h #define for_each_dpa_resource_safe(ndd, res, next) \
next              100 drivers/nvdimm/nd.h 	for (res = (ndd)->dpa.child, next = res ? res->sibling : NULL; \
next              101 drivers/nvdimm/nd.h 			res; res = next, next = next ? next->sibling : NULL)
next              175 drivers/nvdimm/nd.h 	static const unsigned next[] = { 0, 2, 3, 1 };
next              177 drivers/nvdimm/nd.h 	return next[seq & 3];
next             3631 drivers/nvme/host/core.c 	struct nvme_ns *ns, *next;
next             3635 drivers/nvme/host/core.c 	list_for_each_entry_safe(ns, next, &ctrl->namespaces, list) {
next             3641 drivers/nvme/host/core.c 	list_for_each_entry_safe(ns, next, &rm_list, list)
next             3764 drivers/nvme/host/core.c 	struct nvme_ns *ns, *next;
next             3790 drivers/nvme/host/core.c 	list_for_each_entry_safe(ns, next, &ns_list, list)
next              343 drivers/nvme/host/multipath.c 	struct bio *bio, *next;
next              346 drivers/nvme/host/multipath.c 	next = bio_list_get(&head->requeue_list);
next              349 drivers/nvme/host/multipath.c 	while ((bio = next) != NULL) {
next              350 drivers/nvme/host/multipath.c 		next = bio->bi_next;
next             1137 drivers/nvme/host/rdma.c 		.next		    = NULL,
next             1353 drivers/nvme/host/rdma.c 	wr.next       = NULL;
next             1361 drivers/nvme/host/rdma.c 		first->next = &wr;
next             1386 drivers/nvme/host/rdma.c 	wr.next     = NULL;
next             1135 drivers/nvme/target/fc.c 	struct nvmet_fc_tgt_assoc *assoc, *next;
next             1139 drivers/nvme/target/fc.c 	list_for_each_entry_safe(assoc, next,
next             1155 drivers/nvme/target/fc.c 	struct nvmet_fc_tgtport *tgtport, *next;
next             1163 drivers/nvme/target/fc.c 	list_for_each_entry_safe(tgtport, next, &nvmet_fc_target_list,
next              704 drivers/nvme/target/loop.c 	struct nvme_loop_ctrl *ctrl, *next;
next              710 drivers/nvme/target/loop.c 	list_for_each_entry_safe(ctrl, next, &nvme_loop_ctrl_list, list)
next              482 drivers/nvme/target/rdma.c 		rsp = list_entry(queue->rsp_wr_wait_list.next,
next              454 drivers/nvme/target/tcp.c 		node = node->next;
next              261 drivers/of/base.c 	for (pp = np->properties; pp; pp = pp->next) {
next              731 drivers/of/base.c 	struct device_node *next;
next              736 drivers/of/base.c 	next = prev ? prev->sibling : node->child;
next              737 drivers/of/base.c 	for (; next; next = next->sibling)
next              738 drivers/of/base.c 		if (of_node_get(next))
next              741 drivers/of/base.c 	return next;
next              759 drivers/of/base.c 	struct device_node *next;
next              763 drivers/of/base.c 	next = __of_get_next_child(node, prev);
next              765 drivers/of/base.c 	return next;
next              780 drivers/of/base.c 	struct device_node *next;
next              787 drivers/of/base.c 	next = prev ? prev->sibling : node->child;
next              788 drivers/of/base.c 	for (; next; next = next->sibling) {
next              789 drivers/of/base.c 		if (!__of_device_is_available(next))
next              791 drivers/of/base.c 		if (of_node_get(next))
next              796 drivers/of/base.c 	return next;
next              810 drivers/of/base.c 	struct device_node *next = NULL;
next              819 drivers/of/base.c 		next = prev->sibling;
next              821 drivers/of/base.c 		next = node->child;
next              824 drivers/of/base.c 	for (; next; next = next->sibling) {
next              825 drivers/of/base.c 		if (!(of_node_name_eq(next, "cpu") ||
next              826 drivers/of/base.c 		      __of_node_is_type(next, "cpu")))
next              828 drivers/of/base.c 		if (of_node_get(next))
next              833 drivers/of/base.c 	return next;
next             1093 drivers/of/base.c 		for (pp = np->properties; pp; pp = pp->next) {
next             1818 drivers/of/base.c 	struct property **next;
next             1820 drivers/of/base.c 	prop->next = NULL;
next             1821 drivers/of/base.c 	next = &np->properties;
next             1822 drivers/of/base.c 	while (*next) {
next             1823 drivers/of/base.c 		if (strcmp(prop->name, (*next)->name) == 0)
next             1827 drivers/of/base.c 		next = &(*next)->next;
next             1829 drivers/of/base.c 	*next = prop;
next             1861 drivers/of/base.c 	struct property **next;
next             1863 drivers/of/base.c 	for (next = &np->properties; *next; next = &(*next)->next) {
next             1864 drivers/of/base.c 		if (*next == prop)
next             1867 drivers/of/base.c 	if (*next == NULL)
next             1871 drivers/of/base.c 	*next = prop->next;
next             1872 drivers/of/base.c 	prop->next = np->deadprops;
next             1914 drivers/of/base.c 	struct property **next, *oldprop;
next             1916 drivers/of/base.c 	for (next = &np->properties; *next; next = &(*next)->next) {
next             1917 drivers/of/base.c 		if (of_prop_cmp((*next)->name, newprop->name) == 0)
next             1920 drivers/of/base.c 	*oldpropp = oldprop = *next;
next             1924 drivers/of/base.c 		newprop->next = oldprop->next;
next             1925 drivers/of/base.c 		*next = newprop;
next             1926 drivers/of/base.c 		oldprop->next = np->deadprops;
next             1930 drivers/of/base.c 		newprop->next = NULL;
next             1931 drivers/of/base.c 		*next = newprop;
next              310 drivers/of/dynamic.c 	struct property *prop, *next;
next              312 drivers/of/dynamic.c 	for (prop = prop_list; prop != NULL; prop = next) {
next              313 drivers/of/dynamic.c 		next = prop->next;
next              584 drivers/of/dynamic.c 		for (propp = &ce->np->deadprops; *propp; propp = &(*propp)->next) {
next              586 drivers/of/dynamic.c 				*propp = ce->prop->next;
next              587 drivers/of/dynamic.c 				ce->prop->next = NULL;
next              612 drivers/of/dynamic.c 		for (propp = &ce->np->deadprops; *propp; propp = &(*propp)->next) {
next              614 drivers/of/dynamic.c 				*propp = ce->prop->next;
next              615 drivers/of/dynamic.c 				ce->prop->next = NULL;
next              168 drivers/of/fdt.c 		pprev      = &pp->next;
next              196 drivers/of/fdt.c 			pprev      = &pp->next;
next              256 drivers/of/fdt.c 	struct device_node *child, *next;
next              270 drivers/of/fdt.c 		next = child->sibling;
next              274 drivers/of/fdt.c 		child = next;
next              216 drivers/of/of_reserved_mem.c 		struct reserved_mem *this, *next;
next              219 drivers/of/of_reserved_mem.c 		next = &reserved_mem[i + 1];
next              220 drivers/of/of_reserved_mem.c 		if (!(this->base && next->base))
next              222 drivers/of/of_reserved_mem.c 		if (this->base + this->size > next->base) {
next              226 drivers/of/of_reserved_mem.c 			next_end = next->base + next->size;
next              229 drivers/of/of_reserved_mem.c 			       next->name, &next->base, &next_end);
next              354 drivers/of/overlay.c 			new_prop->next = target->np->deadprops;
next              851 drivers/of/overlay.c 	if (ovcs->cset.entries.next)
next              119 drivers/of/pdt.c 	tail->next = of_pdt_build_one_prop(node, NULL, NULL, NULL, 0);
next              120 drivers/of/pdt.c 	tail = tail->next;
next              122 drivers/of/pdt.c 		tail->next = of_pdt_build_one_prop(node, tail->name,
next              124 drivers/of/pdt.c 		tail = tail->next;
next             1126 drivers/of/unittest.c 		save_next = prop->next;
next             1149 drivers/of/unittest.c 	struct device_node *next, *dup, *child;
next             1182 drivers/of/unittest.c 		next = child->sibling;
next             1184 drivers/of/unittest.c 		child = next;
next             1248 drivers/of/unittest.c 		struct device_node *next = np->sibling;
next             1252 drivers/of/unittest.c 		np = next;
next              252 drivers/parisc/ccio-dma.c 	struct ioc *next;		/* Linked list of discovered iocs */
next             1091 drivers/parisc/ccio-dma.c 		ioc = ioc->next;
next             1105 drivers/parisc/ccio-dma.c 		ioc = ioc->next;
next             1130 drivers/parisc/ccio-dma.c 		ioc = ioc->next;
next             1532 drivers/parisc/ccio-dma.c 		ioc_p = &(*ioc_p)->next;
next               28 drivers/parisc/hppb.c 	struct hppb_card *next;
next               33 drivers/parisc/hppb.c 	.next = NULL,
next               52 drivers/parisc/hppb.c 	while(card->next) {
next               53 drivers/parisc/hppb.c 		card = card->next;
next               57 drivers/parisc/hppb.c 		card->next = kzalloc(sizeof(struct hppb_card), GFP_KERNEL);
next               58 drivers/parisc/hppb.c 		if(!card->next) {
next               62 drivers/parisc/hppb.c 		card = card->next;
next             1682 drivers/parisc/sba_iommu.c 	sba_dev->next = sba_list;
next               40 drivers/parport/daisy.c 	struct daisydev *next;
next               65 drivers/parport/daisy.c 		for (p = &topology; *p && (*p)->devnum<devnum; p = &(*p)->next)
next               67 drivers/parport/daisy.c 		newdev->next = *p;
next              184 drivers/parport/daisy.c 			p = &dev->next;
next              187 drivers/parport/daisy.c 		*p = dev->next;
next              222 drivers/parport/daisy.c 		p = p->next;
next              468 drivers/parport/parport_ip32.c 	dma_addr_t			next;
next              510 drivers/parport/parport_ip32.c 			 (unsigned int)parport_ip32_dma.next,
next              514 drivers/parport/parport_ip32.c 		ctxval |= parport_ip32_dma.next &
next              519 drivers/parport/parport_ip32.c 		parport_ip32_dma.next += count;
next              594 drivers/parport/parport_ip32.c 	parport_ip32_dma.next = parport_ip32_dma.buf;
next              603 drivers/parport/parport_ip32.c 		(parport_ip32_dma.next & (MACEPAR_CONTEXT_DATA_BOUND - 1));
next             3348 drivers/parport/parport_pc.c 		priv = list_entry(ports_list.next,
next               52 drivers/parport/procfs.c 	for (dev = port->devices; dev ; dev = dev->next) {
next              501 drivers/parport/share.c 	for (l = all_ports.next, num = 0; l != &all_ports; l = l->next, num++) {
next              800 drivers/parport/share.c 	tmp->next = port->physport->devices;
next              944 drivers/parport/share.c 	par_dev->next = port->physport->devices;
next             1020 drivers/parport/share.c 	if (dev->next)
next             1021 drivers/parport/share.c 		dev->next->prev = dev->prev;
next             1023 drivers/parport/share.c 		dev->prev->next = dev->next;
next             1025 drivers/parport/share.c 		port->devices = dev->next;
next             1374 drivers/parport/share.c 	for (pd = port->devices; !port->cad && pd; pd = pd->next) {
next              383 drivers/pci/bus.c 	struct list_head *next;
next              388 drivers/pci/bus.c 	next = top->devices.next;
next              390 drivers/pci/bus.c 		if (next == &bus->devices) {
next              394 drivers/pci/bus.c 			next = bus->self->bus_list.next;
next              398 drivers/pci/bus.c 		dev = list_entry(next, struct pci_dev, bus_list);
next              401 drivers/pci/bus.c 			next = dev->subordinate->devices.next;
next              404 drivers/pci/bus.c 			next = dev->bus_list.next;
next             2708 drivers/pci/controller/pci-tegra.c 	.next = tegra_pcie_ports_seq_next,
next              817 drivers/pci/controller/pcie-rcar.c 	desc = list_entry(pdev->dev.msi_list.next, struct msi_desc, list);
next              145 drivers/pci/hotplug/acpiphp_glue.c 	struct acpiphp_slot *slot, *next;
next              152 drivers/pci/hotplug/acpiphp_glue.c 	list_for_each_entry_safe(slot, next, &bridge->slots, node) {
next              226 drivers/pci/hotplug/cpqphp.h 	struct pci_func *next;
next              248 drivers/pci/hotplug/cpqphp.h 	struct slot *next;
next              267 drivers/pci/hotplug/cpqphp.h 	struct pci_resource *next;
next              278 drivers/pci/hotplug/cpqphp.h 	struct controller *next;
next              466 drivers/pci/hotplug/cpqphp.h 	node->next = *head;
next              274 drivers/pci/hotplug/cpqphp_core.c 		next_slot = old_slot->next;
next              665 drivers/pci/hotplug/cpqphp_core.c 		slot->next = ctrl->slot;
next             1165 drivers/pci/hotplug/cpqphp_core.c 		ctrl->next = NULL;
next             1167 drivers/pci/hotplug/cpqphp_core.c 		ctrl->next = cpqhp_ctrl_list;
next             1247 drivers/pci/hotplug/cpqphp_core.c 	struct pci_func *next;
next             1278 drivers/pci/hotplug/cpqphp_core.c 			res = res->next;
next             1285 drivers/pci/hotplug/cpqphp_core.c 			res = res->next;
next             1292 drivers/pci/hotplug/cpqphp_core.c 			res = res->next;
next             1299 drivers/pci/hotplug/cpqphp_core.c 			res = res->next;
next             1306 drivers/pci/hotplug/cpqphp_core.c 		ctrl = ctrl->next;
next             1311 drivers/pci/hotplug/cpqphp_core.c 		next = cpqhp_slot_list[loop];
next             1312 drivers/pci/hotplug/cpqphp_core.c 		while (next != NULL) {
next             1313 drivers/pci/hotplug/cpqphp_core.c 			res = next->io_head;
next             1316 drivers/pci/hotplug/cpqphp_core.c 				res = res->next;
next             1320 drivers/pci/hotplug/cpqphp_core.c 			res = next->mem_head;
next             1323 drivers/pci/hotplug/cpqphp_core.c 				res = res->next;
next             1327 drivers/pci/hotplug/cpqphp_core.c 			res = next->p_mem_head;
next             1330 drivers/pci/hotplug/cpqphp_core.c 				res = res->next;
next             1334 drivers/pci/hotplug/cpqphp_core.c 			res = next->bus_head;
next             1337 drivers/pci/hotplug/cpqphp_core.c 				res = res->next;
next             1341 drivers/pci/hotplug/cpqphp_core.c 			TempSlot = next;
next             1342 drivers/pci/hotplug/cpqphp_core.c 			next = next->next;
next              121 drivers/pci/hotplug/cpqphp_ctrl.c 		slot = slot->next;
next              307 drivers/pci/hotplug/cpqphp_ctrl.c 	if (!((*head)->next))
next              314 drivers/pci/hotplug/cpqphp_ctrl.c 		if (((*head)->next) &&
next              315 drivers/pci/hotplug/cpqphp_ctrl.c 		    ((*head)->length > (*head)->next->length)) {
next              318 drivers/pci/hotplug/cpqphp_ctrl.c 			*head = (*head)->next;
next              319 drivers/pci/hotplug/cpqphp_ctrl.c 			current_res->next = (*head)->next;
next              320 drivers/pci/hotplug/cpqphp_ctrl.c 			(*head)->next = current_res;
next              325 drivers/pci/hotplug/cpqphp_ctrl.c 		while (current_res->next && current_res->next->next) {
next              326 drivers/pci/hotplug/cpqphp_ctrl.c 			if (current_res->next->length > current_res->next->next->length) {
next              328 drivers/pci/hotplug/cpqphp_ctrl.c 				next_res = current_res->next;
next              329 drivers/pci/hotplug/cpqphp_ctrl.c 				current_res->next = current_res->next->next;
next              330 drivers/pci/hotplug/cpqphp_ctrl.c 				current_res = current_res->next;
next              331 drivers/pci/hotplug/cpqphp_ctrl.c 				next_res->next = current_res->next;
next              332 drivers/pci/hotplug/cpqphp_ctrl.c 				current_res->next = next_res;
next              334 drivers/pci/hotplug/cpqphp_ctrl.c 				current_res = current_res->next;
next              355 drivers/pci/hotplug/cpqphp_ctrl.c 	if (!((*head)->next))
next              362 drivers/pci/hotplug/cpqphp_ctrl.c 		if (((*head)->next) &&
next              363 drivers/pci/hotplug/cpqphp_ctrl.c 		    ((*head)->length < (*head)->next->length)) {
next              366 drivers/pci/hotplug/cpqphp_ctrl.c 			*head = (*head)->next;
next              367 drivers/pci/hotplug/cpqphp_ctrl.c 			current_res->next = (*head)->next;
next              368 drivers/pci/hotplug/cpqphp_ctrl.c 			(*head)->next = current_res;
next              373 drivers/pci/hotplug/cpqphp_ctrl.c 		while (current_res->next && current_res->next->next) {
next              374 drivers/pci/hotplug/cpqphp_ctrl.c 			if (current_res->next->length < current_res->next->next->length) {
next              376 drivers/pci/hotplug/cpqphp_ctrl.c 				next_res = current_res->next;
next              377 drivers/pci/hotplug/cpqphp_ctrl.c 				current_res->next = current_res->next->next;
next              378 drivers/pci/hotplug/cpqphp_ctrl.c 				current_res = current_res->next;
next              379 drivers/pci/hotplug/cpqphp_ctrl.c 				next_res->next = current_res->next;
next              380 drivers/pci/hotplug/cpqphp_ctrl.c 				current_res->next = next_res;
next              382 drivers/pci/hotplug/cpqphp_ctrl.c 				current_res = current_res->next;
next              446 drivers/pci/hotplug/cpqphp_ctrl.c 		split_node->next = node;
next              454 drivers/pci/hotplug/cpqphp_ctrl.c 		*head = node->next;
next              457 drivers/pci/hotplug/cpqphp_ctrl.c 		while (prevnode->next != node)
next              458 drivers/pci/hotplug/cpqphp_ctrl.c 			prevnode = prevnode->next;
next              460 drivers/pci/hotplug/cpqphp_ctrl.c 		prevnode->next = node->next;
next              462 drivers/pci/hotplug/cpqphp_ctrl.c 	node->next = NULL;
next              487 drivers/pci/hotplug/cpqphp_ctrl.c 	while (node->next) {
next              489 drivers/pci/hotplug/cpqphp_ctrl.c 		node = node->next;
next              542 drivers/pci/hotplug/cpqphp_ctrl.c 	for (node = *head; node; node = node->next) {
next              567 drivers/pci/hotplug/cpqphp_ctrl.c 			split_node->next = node->next;
next              568 drivers/pci/hotplug/cpqphp_ctrl.c 			node->next = split_node;
next              586 drivers/pci/hotplug/cpqphp_ctrl.c 			split_node->next = node->next;
next              587 drivers/pci/hotplug/cpqphp_ctrl.c 			node->next = split_node;
next              598 drivers/pci/hotplug/cpqphp_ctrl.c 			*head = node->next;
next              601 drivers/pci/hotplug/cpqphp_ctrl.c 			while (prevnode->next != node)
next              602 drivers/pci/hotplug/cpqphp_ctrl.c 				prevnode = prevnode->next;
next              604 drivers/pci/hotplug/cpqphp_ctrl.c 			prevnode->next = node->next;
next              606 drivers/pci/hotplug/cpqphp_ctrl.c 		node->next = NULL;
next              636 drivers/pci/hotplug/cpqphp_ctrl.c 	for (max = *head; max; max = max->next) {
next              663 drivers/pci/hotplug/cpqphp_ctrl.c 			split_node->next = max->next;
next              664 drivers/pci/hotplug/cpqphp_ctrl.c 			max->next = split_node;
next              681 drivers/pci/hotplug/cpqphp_ctrl.c 			split_node->next = max->next;
next              682 drivers/pci/hotplug/cpqphp_ctrl.c 			max->next = split_node;
next              692 drivers/pci/hotplug/cpqphp_ctrl.c 			*head = max->next;
next              694 drivers/pci/hotplug/cpqphp_ctrl.c 			while (temp && temp->next != max)
next              695 drivers/pci/hotplug/cpqphp_ctrl.c 				temp = temp->next;
next              698 drivers/pci/hotplug/cpqphp_ctrl.c 				temp->next = max->next;
next              701 drivers/pci/hotplug/cpqphp_ctrl.c 		max->next = NULL;
next              733 drivers/pci/hotplug/cpqphp_ctrl.c 	for (node = *head; node; node = node->next) {
next              760 drivers/pci/hotplug/cpqphp_ctrl.c 			split_node->next = node->next;
next              761 drivers/pci/hotplug/cpqphp_ctrl.c 			node->next = split_node;
next              780 drivers/pci/hotplug/cpqphp_ctrl.c 			split_node->next = node->next;
next              781 drivers/pci/hotplug/cpqphp_ctrl.c 			node->next = split_node;
next              788 drivers/pci/hotplug/cpqphp_ctrl.c 			*head = node->next;
next              791 drivers/pci/hotplug/cpqphp_ctrl.c 			while (prevnode->next != node)
next              792 drivers/pci/hotplug/cpqphp_ctrl.c 				prevnode = prevnode->next;
next              794 drivers/pci/hotplug/cpqphp_ctrl.c 			prevnode->next = node->next;
next              796 drivers/pci/hotplug/cpqphp_ctrl.c 		node->next = NULL;
next              824 drivers/pci/hotplug/cpqphp_ctrl.c 	dbg("*head->next = %p\n", (*head)->next);
next              826 drivers/pci/hotplug/cpqphp_ctrl.c 	if (!(*head)->next)
next              830 drivers/pci/hotplug/cpqphp_ctrl.c 	dbg("*head->next->base = 0x%x\n", (*head)->next->base);
next              835 drivers/pci/hotplug/cpqphp_ctrl.c 		if (((*head)->next) &&
next              836 drivers/pci/hotplug/cpqphp_ctrl.c 		    ((*head)->base > (*head)->next->base)) {
next              838 drivers/pci/hotplug/cpqphp_ctrl.c 			(*head) = (*head)->next;
next              839 drivers/pci/hotplug/cpqphp_ctrl.c 			node1->next = (*head)->next;
next              840 drivers/pci/hotplug/cpqphp_ctrl.c 			(*head)->next = node1;
next              846 drivers/pci/hotplug/cpqphp_ctrl.c 		while (node1->next && node1->next->next) {
next              847 drivers/pci/hotplug/cpqphp_ctrl.c 			if (node1->next->base > node1->next->next->base) {
next              849 drivers/pci/hotplug/cpqphp_ctrl.c 				node2 = node1->next;
next              850 drivers/pci/hotplug/cpqphp_ctrl.c 				node1->next = node1->next->next;
next              851 drivers/pci/hotplug/cpqphp_ctrl.c 				node1 = node1->next;
next              852 drivers/pci/hotplug/cpqphp_ctrl.c 				node2->next = node1->next;
next              853 drivers/pci/hotplug/cpqphp_ctrl.c 				node1->next = node2;
next              855 drivers/pci/hotplug/cpqphp_ctrl.c 				node1 = node1->next;
next              861 drivers/pci/hotplug/cpqphp_ctrl.c 	while (node1 && node1->next) {
next              862 drivers/pci/hotplug/cpqphp_ctrl.c 		if ((node1->base + node1->length) == node1->next->base) {
next              865 drivers/pci/hotplug/cpqphp_ctrl.c 			node1->length += node1->next->length;
next              866 drivers/pci/hotplug/cpqphp_ctrl.c 			node2 = node1->next;
next              867 drivers/pci/hotplug/cpqphp_ctrl.c 			node1->next = node1->next->next;
next              870 drivers/pci/hotplug/cpqphp_ctrl.c 			node1 = node1->next;
next              957 drivers/pci/hotplug/cpqphp_ctrl.c 	struct pci_func *next;
next              963 drivers/pci/hotplug/cpqphp_ctrl.c 	new_slot->next = NULL;
next              969 drivers/pci/hotplug/cpqphp_ctrl.c 		next = cpqhp_slot_list[busnumber];
next              970 drivers/pci/hotplug/cpqphp_ctrl.c 		while (next->next != NULL)
next              971 drivers/pci/hotplug/cpqphp_ctrl.c 			next = next->next;
next              972 drivers/pci/hotplug/cpqphp_ctrl.c 		next->next = new_slot;
next              986 drivers/pci/hotplug/cpqphp_ctrl.c 	struct pci_func *next;
next              991 drivers/pci/hotplug/cpqphp_ctrl.c 	next = cpqhp_slot_list[old_slot->bus];
next              992 drivers/pci/hotplug/cpqphp_ctrl.c 	if (next == NULL)
next              995 drivers/pci/hotplug/cpqphp_ctrl.c 	if (next == old_slot) {
next              996 drivers/pci/hotplug/cpqphp_ctrl.c 		cpqhp_slot_list[old_slot->bus] = old_slot->next;
next             1002 drivers/pci/hotplug/cpqphp_ctrl.c 	while ((next->next != old_slot) && (next->next != NULL))
next             1003 drivers/pci/hotplug/cpqphp_ctrl.c 		next = next->next;
next             1005 drivers/pci/hotplug/cpqphp_ctrl.c 	if (next->next == old_slot) {
next             1006 drivers/pci/hotplug/cpqphp_ctrl.c 		next->next = old_slot->next;
next             1025 drivers/pci/hotplug/cpqphp_ctrl.c 	struct pci_func *next;
next             1031 drivers/pci/hotplug/cpqphp_ctrl.c 		next = cpqhp_slot_list[tempBus];
next             1033 drivers/pci/hotplug/cpqphp_ctrl.c 		while (!slot_remove(next))
next             1034 drivers/pci/hotplug/cpqphp_ctrl.c 			next = cpqhp_slot_list[tempBus];
next             1037 drivers/pci/hotplug/cpqphp_ctrl.c 	next = cpqhp_slot_list[bridge->bus];
next             1039 drivers/pci/hotplug/cpqphp_ctrl.c 	if (next == NULL)
next             1042 drivers/pci/hotplug/cpqphp_ctrl.c 	if (next == bridge) {
next             1043 drivers/pci/hotplug/cpqphp_ctrl.c 		cpqhp_slot_list[bridge->bus] = bridge->next;
next             1047 drivers/pci/hotplug/cpqphp_ctrl.c 	while ((next->next != bridge) && (next->next != NULL))
next             1048 drivers/pci/hotplug/cpqphp_ctrl.c 		next = next->next;
next             1050 drivers/pci/hotplug/cpqphp_ctrl.c 	if (next->next != bridge)
next             1052 drivers/pci/hotplug/cpqphp_ctrl.c 	next->next = bridge->next;
next             1080 drivers/pci/hotplug/cpqphp_ctrl.c 	while (func->next != NULL) {
next             1081 drivers/pci/hotplug/cpqphp_ctrl.c 		func = func->next;
next             1130 drivers/pci/hotplug/cpqphp_ctrl.c 	for (slot = ctrl->slot; slot; slot = slot->next) {
next             1743 drivers/pci/hotplug/cpqphp_ctrl.c 			for (ctrl = cpqhp_ctrl_list; ctrl; ctrl = ctrl->next)
next             2414 drivers/pci/hotplug/cpqphp_ctrl.c 					io_node->length, io_node->next);
next             2417 drivers/pci/hotplug/cpqphp_ctrl.c 					mem_node->length, mem_node->next);
next             2420 drivers/pci/hotplug/cpqphp_ctrl.c 					p_mem_node->length, p_mem_node->next);
next             2468 drivers/pci/hotplug/cpqphp_ctrl.c 		bus_node->next = NULL;
next             2473 drivers/pci/hotplug/cpqphp_ctrl.c 		io_node->next = NULL;
next             2486 drivers/pci/hotplug/cpqphp_ctrl.c 		mem_node->next = NULL;
next             2496 drivers/pci/hotplug/cpqphp_ctrl.c 		p_mem_node->next = NULL;
next             2566 drivers/pci/hotplug/cpqphp_ctrl.c 			hold_bus_node->next = func->bus_head;
next             2608 drivers/pci/hotplug/cpqphp_ctrl.c 					hold_IO_node->next = func->io_head;
next             2625 drivers/pci/hotplug/cpqphp_ctrl.c 				hold_IO_node->next = func->io_head;
next             2630 drivers/pci/hotplug/cpqphp_ctrl.c 			hold_IO_node->next = func->io_head;
next             2658 drivers/pci/hotplug/cpqphp_ctrl.c 					hold_mem_node->next = func->mem_head;
next             2677 drivers/pci/hotplug/cpqphp_ctrl.c 				hold_mem_node->next = func->mem_head;
next             2682 drivers/pci/hotplug/cpqphp_ctrl.c 			hold_mem_node->next = func->mem_head;
next             2711 drivers/pci/hotplug/cpqphp_ctrl.c 					hold_p_mem_node->next = func->p_mem_head;
next             2728 drivers/pci/hotplug/cpqphp_ctrl.c 				hold_p_mem_node->next = func->p_mem_head;
next             2733 drivers/pci/hotplug/cpqphp_ctrl.c 			hold_p_mem_node->next = func->p_mem_head;
next             2785 drivers/pci/hotplug/cpqphp_ctrl.c 					    io_node->base, io_node->length, io_node->next);
next             2790 drivers/pci/hotplug/cpqphp_ctrl.c 					io_node->next = func->io_head;
next             2804 drivers/pci/hotplug/cpqphp_ctrl.c 						p_mem_node->next = func->p_mem_head;
next             2820 drivers/pci/hotplug/cpqphp_ctrl.c 						mem_node->next = func->mem_head;
next               61 drivers/pci/hotplug/cpqphp_nvram.c 	u8 next;
next               72 drivers/pci/hotplug/cpqphp_nvram.c 	u8 next;
next              307 drivers/pci/hotplug/cpqphp_nvram.c 			resNode = resNode->next;
next              332 drivers/pci/hotplug/cpqphp_nvram.c 			resNode = resNode->next;
next              357 drivers/pci/hotplug/cpqphp_nvram.c 			resNode = resNode->next;
next              382 drivers/pci/hotplug/cpqphp_nvram.c 			resNode = resNode->next;
next              388 drivers/pci/hotplug/cpqphp_nvram.c 		ctrl = ctrl->next;
next              455 drivers/pci/hotplug/cpqphp_nvram.c 		p_byte = &(p_EV_header->next);
next              457 drivers/pci/hotplug/cpqphp_nvram.c 		p_ev_ctrl = (struct ev_hrt_ctrl *) &(p_EV_header->next);
next              533 drivers/pci/hotplug/cpqphp_nvram.c 			mem_node->next = ctrl->mem_head;
next              561 drivers/pci/hotplug/cpqphp_nvram.c 			p_mem_node->next = ctrl->p_mem_head;
next              589 drivers/pci/hotplug/cpqphp_nvram.c 			io_node->next = ctrl->io_head;
next              615 drivers/pci/hotplug/cpqphp_nvram.c 			bus_node->next = ctrl->bus_head;
next              558 drivers/pci/hotplug/cpqphp_pci.c 	struct pci_func *next;
next              577 drivers/pci/hotplug/cpqphp_pci.c 			next = cpqhp_slot_list[sub_bus];
next              579 drivers/pci/hotplug/cpqphp_pci.c 			while (next != NULL) {
next              580 drivers/pci/hotplug/cpqphp_pci.c 				rc = cpqhp_save_base_addr_length(ctrl, next);
next              584 drivers/pci/hotplug/cpqphp_pci.c 				next = next->next;
next              738 drivers/pci/hotplug/cpqphp_pci.c 			bus_node->next = func->bus_head;
next              753 drivers/pci/hotplug/cpqphp_pci.c 				io_node->next = func->io_head;
next              769 drivers/pci/hotplug/cpqphp_pci.c 				mem_node->next = func->mem_head;
next              785 drivers/pci/hotplug/cpqphp_pci.c 				p_mem_node->next = func->p_mem_head;
next              818 drivers/pci/hotplug/cpqphp_pci.c 						io_node->next = func->io_head;
next              835 drivers/pci/hotplug/cpqphp_pci.c 						p_mem_node->next = func->p_mem_head;
next              852 drivers/pci/hotplug/cpqphp_pci.c 						mem_node->next = func->mem_head;
next              889 drivers/pci/hotplug/cpqphp_pci.c 						io_node->next = func->io_head;
next              906 drivers/pci/hotplug/cpqphp_pci.c 						p_mem_node->next = func->p_mem_head;
next              923 drivers/pci/hotplug/cpqphp_pci.c 						mem_node->next = func->mem_head;
next              954 drivers/pci/hotplug/cpqphp_pci.c 	struct pci_func *next;
next              981 drivers/pci/hotplug/cpqphp_pci.c 			next = cpqhp_slot_list[sub_bus];
next              983 drivers/pci/hotplug/cpqphp_pci.c 			while (next != NULL) {
next              984 drivers/pci/hotplug/cpqphp_pci.c 				rc = cpqhp_configure_board(ctrl, next);
next              988 drivers/pci/hotplug/cpqphp_pci.c 				next = next->next;
next             1035 drivers/pci/hotplug/cpqphp_pci.c 	struct pci_func *next;
next             1079 drivers/pci/hotplug/cpqphp_pci.c 			next = cpqhp_slot_list[secondary_bus];
next             1081 drivers/pci/hotplug/cpqphp_pci.c 			while (next != NULL) {
next             1082 drivers/pci/hotplug/cpqphp_pci.c 				rc = cpqhp_valid_replace(ctrl, next);
next             1086 drivers/pci/hotplug/cpqphp_pci.c 				next = next->next;
next             1314 drivers/pci/hotplug/cpqphp_pci.c 				io_node->next = ctrl->io_head;
next             1317 drivers/pci/hotplug/cpqphp_pci.c 				io_node->next = func->io_head;
next             1337 drivers/pci/hotplug/cpqphp_pci.c 				mem_node->next = ctrl->mem_head;
next             1340 drivers/pci/hotplug/cpqphp_pci.c 				mem_node->next = func->mem_head;
next             1362 drivers/pci/hotplug/cpqphp_pci.c 				p_mem_node->next = ctrl->p_mem_head;
next             1365 drivers/pci/hotplug/cpqphp_pci.c 				p_mem_node->next = func->p_mem_head;
next             1385 drivers/pci/hotplug/cpqphp_pci.c 				bus_node->next = ctrl->bus_head;
next             1388 drivers/pci/hotplug/cpqphp_pci.c 				bus_node->next = func->bus_head;
next             1431 drivers/pci/hotplug/cpqphp_pci.c 		t_node = node->next;
next             1439 drivers/pci/hotplug/cpqphp_pci.c 		t_node = node->next;
next             1447 drivers/pci/hotplug/cpqphp_pci.c 		t_node = node->next;
next             1455 drivers/pci/hotplug/cpqphp_pci.c 		t_node = node->next;
next             1483 drivers/pci/hotplug/cpqphp_pci.c 		res = res->next;
next             1492 drivers/pci/hotplug/cpqphp_pci.c 		res = res->next;
next             1501 drivers/pci/hotplug/cpqphp_pci.c 		res = res->next;
next             1510 drivers/pci/hotplug/cpqphp_pci.c 		res = res->next;
next             1530 drivers/pci/hotplug/cpqphp_pci.c 		res = res->next;
next             1539 drivers/pci/hotplug/cpqphp_pci.c 		res = res->next;
next             1548 drivers/pci/hotplug/cpqphp_pci.c 		res = res->next;
next             1557 drivers/pci/hotplug/cpqphp_pci.c 		res = res->next;
next               39 drivers/pci/hotplug/cpqphp_sysfs.c 		res = res->next;
next               46 drivers/pci/hotplug/cpqphp_sysfs.c 		res = res->next;
next               53 drivers/pci/hotplug/cpqphp_sysfs.c 		res = res->next;
next               60 drivers/pci/hotplug/cpqphp_sysfs.c 		res = res->next;
next               85 drivers/pci/hotplug/cpqphp_sysfs.c 			res = res->next;
next               92 drivers/pci/hotplug/cpqphp_sysfs.c 			res = res->next;
next               99 drivers/pci/hotplug/cpqphp_sysfs.c 			res = res->next;
next              106 drivers/pci/hotplug/cpqphp_sysfs.c 			res = res->next;
next              108 drivers/pci/hotplug/cpqphp_sysfs.c 		slot = slot->next;
next              214 drivers/pci/hotplug/ibmphp.h 	struct ebda_rsrc_list *next;
next              325 drivers/pci/hotplug/ibmphp.h 	struct range_node *next;
next              356 drivers/pci/hotplug/ibmphp.h 	struct resource_node *next;
next              684 drivers/pci/hotplug/ibmphp.h 	struct pci_func *next;
next              622 drivers/pci/hotplug/ibmphp_core.c 				func_cur = func_cur->next;
next              636 drivers/pci/hotplug/ibmphp_core.c 	struct slot *slot_cur, *next;
next              640 drivers/pci/hotplug/ibmphp_core.c 	list_for_each_entry_safe(slot_cur, next, &ibmphp_slot_head,
next             1036 drivers/pci/hotplug/ibmphp_ebda.c 	struct bus_info *bus_info, *next;
next             1038 drivers/pci/hotplug/ibmphp_ebda.c 	list_for_each_entry_safe(bus_info, next, &bus_info_head,
next             1046 drivers/pci/hotplug/ibmphp_ebda.c 	struct controller *controller = NULL, *next;
next             1049 drivers/pci/hotplug/ibmphp_ebda.c 	list_for_each_entry_safe(controller, next, &ebda_hpc_head,
next             1063 drivers/pci/hotplug/ibmphp_ebda.c 	struct ebda_pci_rsrc *resource, *next;
next             1065 drivers/pci/hotplug/ibmphp_ebda.c 	list_for_each_entry_safe(resource, next, &ibmphp_ebda_pci_rsrc_head,
next              142 drivers/pci/hotplug/ibmphp_pci.c 					cur_func->next = NULL;
next              161 drivers/pci/hotplug/ibmphp_pci.c 					cur_func->next = newfunc;
next              202 drivers/pci/hotplug/ibmphp_pci.c 								for (prev_func = cur_func; prev_func->next; prev_func = prev_func->next) ;
next              203 drivers/pci/hotplug/ibmphp_pci.c 								prev_func->next = newfunc;
next              205 drivers/pci/hotplug/ibmphp_pci.c 								cur_func->next = newfunc;
next              227 drivers/pci/hotplug/ibmphp_pci.c 					for (prev_func = cur_func; prev_func->next; prev_func = prev_func->next);
next              228 drivers/pci/hotplug/ibmphp_pci.c 					prev_func->next = newfunc;
next              274 drivers/pci/hotplug/ibmphp_pci.c 								for (prev_func = cur_func; prev_func->next; prev_func = prev_func->next);
next              275 drivers/pci/hotplug/ibmphp_pci.c 								prev_func->next = newfunc;
next              277 drivers/pci/hotplug/ibmphp_pci.c 								cur_func->next = newfunc;
next             1585 drivers/pci/hotplug/ibmphp_pci.c 			temp_func = cur_func->next;
next              395 drivers/pci/hotplug/ibmphp_res.c 		range_cur = range_cur->next;
next              411 drivers/pci/hotplug/ibmphp_res.c 		range->next = range_cur;
next              416 drivers/pci/hotplug/ibmphp_res.c 		range->next = NULL;
next              417 drivers/pci/hotplug/ibmphp_res.c 		range_prev->next = range;
next              422 drivers/pci/hotplug/ibmphp_res.c 		range_prev->next = range;
next              423 drivers/pci/hotplug/ibmphp_res.c 		range->next = range_cur;
next              430 drivers/pci/hotplug/ibmphp_res.c 		range_cur = range_cur->next;
next              467 drivers/pci/hotplug/ibmphp_res.c 			if (res->next)
next              468 drivers/pci/hotplug/ibmphp_res.c 				res = res->next;
next              481 drivers/pci/hotplug/ibmphp_res.c 				res = res->next;
next              521 drivers/pci/hotplug/ibmphp_res.c 				range = range->next;
next              524 drivers/pci/hotplug/ibmphp_res.c 		if (res->next)
next              525 drivers/pci/hotplug/ibmphp_res.c 			res = res->next;
next              621 drivers/pci/hotplug/ibmphp_res.c 		range_cur = range_cur->next;
next              658 drivers/pci/hotplug/ibmphp_res.c 		res->next = NULL;
next              670 drivers/pci/hotplug/ibmphp_res.c 			if (res_cur->next)
next              671 drivers/pci/hotplug/ibmphp_res.c 				res_cur = res_cur->next;
next              680 drivers/pci/hotplug/ibmphp_res.c 			res->next = NULL;
next              688 drivers/pci/hotplug/ibmphp_res.c 				res_cur = res_cur->next;
next              692 drivers/pci/hotplug/ibmphp_res.c 				res_prev->next = res;
next              693 drivers/pci/hotplug/ibmphp_res.c 				res->next = NULL;
next              711 drivers/pci/hotplug/ibmphp_res.c 					res_prev->next = res;
next              715 drivers/pci/hotplug/ibmphp_res.c 				res->next = res_cur;
next              722 drivers/pci/hotplug/ibmphp_res.c 				res->next = NULL;
next              740 drivers/pci/hotplug/ibmphp_res.c 				res->next = NULL;
next              800 drivers/pci/hotplug/ibmphp_res.c 		if (res_cur->next)
next              801 drivers/pci/hotplug/ibmphp_res.c 			res_cur = res_cur->next;
next              823 drivers/pci/hotplug/ibmphp_res.c 						if (mem_cur->next)
next              824 drivers/pci/hotplug/ibmphp_res.c 							mem_cur = mem_cur->next;
next              835 drivers/pci/hotplug/ibmphp_res.c 						bus_cur->firstPFMemFromMem = res_cur->next;
next              837 drivers/pci/hotplug/ibmphp_res.c 						res_prev->next = res_cur->next;
next              842 drivers/pci/hotplug/ibmphp_res.c 				if (res_cur->next)
next              843 drivers/pci/hotplug/ibmphp_res.c 					res_cur = res_cur->next;
next              858 drivers/pci/hotplug/ibmphp_res.c 		if (res_cur->next) {
next              861 drivers/pci/hotplug/ibmphp_res.c 					bus_cur->firstIO = res_cur->next;
next              864 drivers/pci/hotplug/ibmphp_res.c 					bus_cur->firstMem = res_cur->next;
next              867 drivers/pci/hotplug/ibmphp_res.c 					bus_cur->firstPFMem = res_cur->next;
next              898 drivers/pci/hotplug/ibmphp_res.c 		if (res_cur->next) {
next              900 drivers/pci/hotplug/ibmphp_res.c 				res_prev->next = res_cur->next;
next              902 drivers/pci/hotplug/ibmphp_res.c 				res_prev->nextRange = res_cur->next;
next              904 drivers/pci/hotplug/ibmphp_res.c 			res_prev->next = NULL;
next              907 drivers/pci/hotplug/ibmphp_res.c 			res_prev->next = NULL;
next              938 drivers/pci/hotplug/ibmphp_res.c 		range = range->next;
next             1062 drivers/pci/hotplug/ibmphp_res.c 		if (!res_cur->next) {
next             1181 drivers/pci/hotplug/ibmphp_res.c 		if (res_cur->next)
next             1182 drivers/pci/hotplug/ibmphp_res.c 			res_cur = res_cur->next;
next             1238 drivers/pci/hotplug/ibmphp_res.c 			range = range->next;
next             1304 drivers/pci/hotplug/ibmphp_res.c 				range = range->next;
next             1364 drivers/pci/hotplug/ibmphp_res.c 			if (res_cur->next)
next             1365 drivers/pci/hotplug/ibmphp_res.c 				res_cur = res_cur->next;
next             1377 drivers/pci/hotplug/ibmphp_res.c 			if (res_cur->next)
next             1378 drivers/pci/hotplug/ibmphp_res.c 				res_cur = res_cur->next;
next             1390 drivers/pci/hotplug/ibmphp_res.c 			if (res_cur->next)
next             1391 drivers/pci/hotplug/ibmphp_res.c 				res_cur = res_cur->next;
next             1404 drivers/pci/hotplug/ibmphp_res.c 			res_cur = res_cur->next;
next             1438 drivers/pci/hotplug/ibmphp_res.c 			range_cur = range_cur->next;
next             1452 drivers/pci/hotplug/ibmphp_res.c 			range_cur = range_cur->next;
next             1466 drivers/pci/hotplug/ibmphp_res.c 			range_cur = range_cur->next;
next             1512 drivers/pci/hotplug/ibmphp_res.c 		if (res_cur->next)
next             1513 drivers/pci/hotplug/ibmphp_res.c 			res_cur = res_cur->next;
next             1526 drivers/pci/hotplug/ibmphp_res.c 				res_cur = res_cur->next;
next             1552 drivers/pci/hotplug/ibmphp_res.c 	struct bus_node *bus_cur = NULL, *next;
next             1561 drivers/pci/hotplug/ibmphp_res.c 	list_for_each_entry_safe(bus_cur, next, &gbuses, bus_list) {
next             1568 drivers/pci/hotplug/ibmphp_res.c 				range_cur = range_cur->next;
next             1579 drivers/pci/hotplug/ibmphp_res.c 				range_cur = range_cur->next;
next             1590 drivers/pci/hotplug/ibmphp_res.c 				range_cur = range_cur->next;
next             1600 drivers/pci/hotplug/ibmphp_res.c 				if (res_cur->next)
next             1601 drivers/pci/hotplug/ibmphp_res.c 					res_cur = res_cur->next;
next             1613 drivers/pci/hotplug/ibmphp_res.c 				if (res_cur->next)
next             1614 drivers/pci/hotplug/ibmphp_res.c 					res_cur = res_cur->next;
next             1626 drivers/pci/hotplug/ibmphp_res.c 				if (res_cur->next)
next             1627 drivers/pci/hotplug/ibmphp_res.c 					res_cur = res_cur->next;
next             1640 drivers/pci/hotplug/ibmphp_res.c 				res_cur = res_cur->next;
next             1672 drivers/pci/hotplug/ibmphp_res.c 			for (pfmem_cur = bus_cur->firstPFMem, pfmem_prev = NULL; pfmem_cur; pfmem_prev = pfmem_cur, pfmem_cur = pfmem_cur->next) {
next             1675 drivers/pci/hotplug/ibmphp_res.c 					pfmem_prev->next = pfmem_cur->next;
next             1677 drivers/pci/hotplug/ibmphp_res.c 					bus_cur->firstPFMem = pfmem_cur->next;
next             1680 drivers/pci/hotplug/ibmphp_res.c 					pfmem_cur->next = NULL;
next             1686 drivers/pci/hotplug/ibmphp_res.c 					pfmem_cur->next = bus_cur->firstPFMemFromMem;
next             1719 drivers/pci/hotplug/ibmphp_res.c 		pfmem->next = bus_cur->firstPFMemFromMem;
next             1721 drivers/pci/hotplug/ibmphp_res.c 		pfmem->next = NULL;
next             1778 drivers/pci/hotplug/ibmphp_res.c 				range = range->next;
next             1788 drivers/pci/hotplug/ibmphp_res.c 				range = range->next;
next             1799 drivers/pci/hotplug/ibmphp_res.c 				range = range->next;
next             1812 drivers/pci/hotplug/ibmphp_res.c 				if (res->next)
next             1813 drivers/pci/hotplug/ibmphp_res.c 					res = res->next;
next             1827 drivers/pci/hotplug/ibmphp_res.c 				if (res->next)
next             1828 drivers/pci/hotplug/ibmphp_res.c 					res = res->next;
next             1842 drivers/pci/hotplug/ibmphp_res.c 				if (res->next)
next             1843 drivers/pci/hotplug/ibmphp_res.c 					res = res->next;
next             1858 drivers/pci/hotplug/ibmphp_res.c 				res = res->next;
next             1886 drivers/pci/hotplug/ibmphp_res.c 		range_cur = range_cur->next;
next              197 drivers/pci/hotplug/pnv_php.c 	struct device_node *child, *next;
next              207 drivers/pci/hotplug/pnv_php.c 		next = child->sibling;
next              211 drivers/pci/hotplug/pnv_php.c 		child = next;
next              112 drivers/pci/hotplug/rpadlpar_core.c 	struct slot *slot, *next;
next              114 drivers/pci/hotplug/rpadlpar_core.c 	list_for_each_entry_safe(slot, next, &rpaphp_slot_head,
next              402 drivers/pci/hotplug/rpaphp_core.c 	struct slot *slot, *next;
next              409 drivers/pci/hotplug/rpaphp_core.c 	list_for_each_entry_safe(slot, next, &rpaphp_slot_head,
next              178 drivers/pci/hotplug/s390_pci_hpc.c 	struct slot *slot, *next;
next              180 drivers/pci/hotplug/s390_pci_hpc.c 	list_for_each_entry_safe(slot, next, &s390_hotplug_slot_list,
next              132 drivers/pci/hotplug/shpchp_core.c 	struct slot *slot, *next;
next              134 drivers/pci/hotplug/shpchp_core.c 	list_for_each_entry_safe(slot, next, &ctrl->slot_list, slot_list) {
next               45 drivers/pci/pci-bridge-emul.h 	u8 next;
next             1197 drivers/pci/pci.c 	hlist_for_each_entry(tmp, &pci_dev->saved_cap_space, next) {
next             1511 drivers/pci/pci.c 	hlist_for_each_entry(tmp, &dev->saved_cap_space, next)
next             1522 drivers/pci/pci.c 	hlist_for_each_entry(tmp, &dev->saved_cap_space, next) {
next             3048 drivers/pci/pci.c 	hlist_add_head(&new_cap->next, &pci_dev->saved_cap_space);
next             3125 drivers/pci/pci.c 	hlist_for_each_entry_safe(tmp, n, &dev->saved_cap_space, next)
next              248 drivers/pci/pcie/aspm.c 	child = list_entry(linkbus->devices.next, struct pci_dev, bus_list);
next              397 drivers/pci/proc.c 	.next	= pci_seq_next,
next             5128 drivers/pci/quirks.c 		hlist_add_head(&state->next, &pdev->saved_cap_space);
next              167 drivers/pci/search.c 	n = from ? from->node.next : pci_root_buses.next;
next             2074 drivers/pci/setup-bus.c 	struct pci_dev *next;
next             2082 drivers/pci/setup-bus.c 	next = bridge;
next             2084 drivers/pci/setup-bus.c 		bridge = next;
next             2112 drivers/pci/setup-bus.c 		next = bridge->bus ? bridge->bus->self : NULL;
next             2113 drivers/pci/setup-bus.c 	} while (next);
next              150 drivers/pci/switch/switchtec.c 	stuser = list_entry(stdev->mrpc_queue.next, struct switchtec_user,
next              194 drivers/pci/switch/switchtec.c 	stuser = list_entry(stdev->mrpc_queue.next, struct switchtec_user,
next              561 drivers/pci/xen-pcifront.c 		dev = container_of(bus->devices.next, struct pci_dev,
next               50 drivers/pcmcia/rsrc_nonstatic.c 	struct resource_map	*next;
next              112 drivers/pcmcia/rsrc_nonstatic.c 	for (p = map; ; p = p->next) {
next              117 drivers/pcmcia/rsrc_nonstatic.c 		if ((p->next == map) || (p->next->base > base+num-1))
next              126 drivers/pcmcia/rsrc_nonstatic.c 	q->next = p->next; p->next = q;
next              137 drivers/pcmcia/rsrc_nonstatic.c 		q = p->next;
next              144 drivers/pcmcia/rsrc_nonstatic.c 					p->next = q->next;
next              167 drivers/pcmcia/rsrc_nonstatic.c 				p->next = q->next ; q->next = p;
next              460 drivers/pcmcia/rsrc_nonstatic.c 	ok = inv_probe(m->next, s);
next              489 drivers/pcmcia/rsrc_nonstatic.c 		if (inv_probe(s_data->mem_db.next, s) > 0)
next              491 drivers/pcmcia/rsrc_nonstatic.c 		if (s_data->mem_db_valid.next != &s_data->mem_db_valid)
next              498 drivers/pcmcia/rsrc_nonstatic.c 	for (m = s_data->mem_db.next; m != &s_data->mem_db; m = mm.next) {
next              542 drivers/pcmcia/rsrc_nonstatic.c 	for (m = s_data->mem_db.next; m != &s_data->mem_db; m = mm.next) {
next              577 drivers/pcmcia/rsrc_nonstatic.c 	if (s_data->mem_db_valid.next != &s_data->mem_db_valid)
next              612 drivers/pcmcia/rsrc_nonstatic.c 	for (m = data->map->next; m != data->map; m = m->next) {
next              656 drivers/pcmcia/rsrc_nonstatic.c 	for (m = s_data->io_db.next; m != &s_data->io_db; m = m->next) {
next             1022 drivers/pcmcia/rsrc_nonstatic.c 	data->mem_db.next = &data->mem_db;
next             1023 drivers/pcmcia/rsrc_nonstatic.c 	data->mem_db_valid.next = &data->mem_db_valid;
next             1024 drivers/pcmcia/rsrc_nonstatic.c 	data->io_db.next = &data->io_db;
next             1038 drivers/pcmcia/rsrc_nonstatic.c 	for (p = data->mem_db_valid.next; p != &data->mem_db_valid; p = q) {
next             1039 drivers/pcmcia/rsrc_nonstatic.c 		q = p->next;
next             1042 drivers/pcmcia/rsrc_nonstatic.c 	for (p = data->mem_db.next; p != &data->mem_db; p = q) {
next             1043 drivers/pcmcia/rsrc_nonstatic.c 		q = p->next;
next             1046 drivers/pcmcia/rsrc_nonstatic.c 	for (p = data->io_db.next; p != &data->io_db; p = q) {
next             1047 drivers/pcmcia/rsrc_nonstatic.c 		q = p->next;
next             1076 drivers/pcmcia/rsrc_nonstatic.c 	for (p = data->io_db.next; p != &data->io_db; p = p->next) {
next             1132 drivers/pcmcia/rsrc_nonstatic.c 	for (p = data->mem_db_valid.next; p != &data->mem_db_valid;
next             1133 drivers/pcmcia/rsrc_nonstatic.c 	     p = p->next) {
next             1142 drivers/pcmcia/rsrc_nonstatic.c 	for (p = data->mem_db.next; p != &data->mem_db; p = p->next) {
next              182 drivers/pcmcia/sa1111_generic.c 			s->next = dev_get_drvdata(&dev->dev);
next              243 drivers/pcmcia/sa1111_generic.c 	struct sa1111_pcmcia_socket *next, *s = dev_get_drvdata(&dev->dev);
next              247 drivers/pcmcia/sa1111_generic.c 	for (; s; s = next) {
next              248 drivers/pcmcia/sa1111_generic.c 		next = s->next;
next                8 drivers/pcmcia/sa1111_generic.h 	struct sa1111_pcmcia_socket *next;
next              178 drivers/perf/qcom_l2_pmu.c 	struct list_head next;
next              797 drivers/perf/qcom_l2_pmu.c 	list_for_each_entry(cluster, &l2cache_pmu->clusters, next) {
next              907 drivers/perf/qcom_l2_pmu.c 	INIT_LIST_HEAD(&cluster->next);
next              908 drivers/perf/qcom_l2_pmu.c 	list_add(&cluster->next, &l2cache_pmu->clusters);
next              140 drivers/perf/xgene_pmu.c 	struct list_head next;
next             1256 drivers/perf/xgene_pmu.c 		list_for_each_entry(ctx, &xgene_pmu->mcpmus, next) {
next             1261 drivers/perf/xgene_pmu.c 		list_for_each_entry(ctx, &xgene_pmu->mcbpmus, next) {
next             1266 drivers/perf/xgene_pmu.c 		list_for_each_entry(ctx, &xgene_pmu->l3cpmus, next) {
next             1271 drivers/perf/xgene_pmu.c 		list_for_each_entry(ctx, &xgene_pmu->iobpmus, next) {
next             1586 drivers/perf/xgene_pmu.c 		list_add(&ctx->next, &xgene_pmu->l3cpmus);
next             1589 drivers/perf/xgene_pmu.c 		list_add(&ctx->next, &xgene_pmu->iobpmus);
next             1592 drivers/perf/xgene_pmu.c 		list_add(&ctx->next, &xgene_pmu->iobpmus);
next             1595 drivers/perf/xgene_pmu.c 		list_add(&ctx->next, &xgene_pmu->mcbpmus);
next             1598 drivers/perf/xgene_pmu.c 		list_add(&ctx->next, &xgene_pmu->mcpmus);
next             1708 drivers/perf/xgene_pmu.c 			list_add(&ctx->next, &xgene_pmu->l3cpmus);
next             1711 drivers/perf/xgene_pmu.c 			list_add(&ctx->next, &xgene_pmu->iobpmus);
next             1714 drivers/perf/xgene_pmu.c 			list_add(&ctx->next, &xgene_pmu->iobpmus);
next             1717 drivers/perf/xgene_pmu.c 			list_add(&ctx->next, &xgene_pmu->mcbpmus);
next             1720 drivers/perf/xgene_pmu.c 			list_add(&ctx->next, &xgene_pmu->mcpmus);
next             1821 drivers/perf/xgene_pmu.c 	list_for_each_entry(ctx, &xgene_pmu->mcpmus, next) {
next             1824 drivers/perf/xgene_pmu.c 	list_for_each_entry(ctx, &xgene_pmu->mcbpmus, next) {
next             1827 drivers/perf/xgene_pmu.c 	list_for_each_entry(ctx, &xgene_pmu->l3cpmus, next) {
next             1830 drivers/perf/xgene_pmu.c 	list_for_each_entry(ctx, &xgene_pmu->iobpmus, next) {
next             1958 drivers/perf/xgene_pmu.c 	list_for_each_entry(ctx, pmus, next) {
next              992 drivers/phy/phy-core.c 		struct device_node *parent = of_node_get(children), *next;
next              998 drivers/phy/phy-core.c 			next = of_get_parent(parent);
next             1000 drivers/phy/phy-core.c 			parent = next;
next               44 drivers/phy/socionext/phy-uniphier-usb2.c 	struct uniphier_u2phy_priv *next;
next               89 drivers/phy/socionext/phy-uniphier-usb2.c 		priv = priv->next;
next              110 drivers/phy/socionext/phy-uniphier-usb2.c 	struct uniphier_u2phy_priv *priv = NULL, *next = NULL;
next              169 drivers/phy/socionext/phy-uniphier-usb2.c 		priv->next = next;
next              170 drivers/phy/socionext/phy-uniphier-usb2.c 		next = priv;
next              123 drivers/pinctrl/mediatek/pinctrl-mtk-common-v2.c 	pfd->next = pfd->bitpos + c->x_bits > c->sz_reg ? c->x_addrs : 0;
next              156 drivers/pinctrl/mediatek/pinctrl-mtk-common-v2.c 	mtk_rmw(hw, pf->index, pf->offset + pf->next, BIT(nbits_h) - 1,
next              169 drivers/pinctrl/mediatek/pinctrl-mtk-common-v2.c 	h  = (mtk_r32(hw, pf->index, pf->offset + pf->next))
next              185 drivers/pinctrl/mediatek/pinctrl-mtk-common-v2.c 	if (!pf.next)
next              204 drivers/pinctrl/mediatek/pinctrl-mtk-common-v2.c 	if (!pf.next)
next              101 drivers/pinctrl/mediatek/pinctrl-mtk-common-v2.h 	u8  next;
next               37 drivers/pinctrl/pinctrl-at91.c 	struct at91_gpio_chip	*next;		/* Bank sharing same clock */
next             1702 drivers/pinctrl/pinctrl-at91.c 			if (!at91_gpio->next)
next             1704 drivers/pinctrl/pinctrl-at91.c 			at91_gpio = at91_gpio->next;
next             1779 drivers/pinctrl/pinctrl-at91.c 		if (prev->next) {
next             1780 drivers/pinctrl/pinctrl-at91.c 			prev = prev->next;
next             1782 drivers/pinctrl/pinctrl-at91.c 			prev->next = at91_gpio;
next              341 drivers/platform/mellanox/mlxbf-tmfifo.c 		idx = virtio16_to_cpu(vdev, desc->next);
next              438 drivers/platform/mellanox/mlxbf-tmfifo.c 		idx = virtio16_to_cpu(vdev, desc->next);
next              723 drivers/platform/mellanox/mlxbf-tmfifo.c 			idx = virtio16_to_cpu(vdev, desc->next);
next              605 drivers/platform/x86/dcdbas.c 	.next = NULL,
next              308 drivers/platform/x86/dell_rbu.c 	ptemp_list = (&packet_data_head.list)->next;
next              322 drivers/platform/x86/dell_rbu.c 		ptemp_list = ptemp_list->next;
next              336 drivers/platform/x86/dell_rbu.c 	ptemp_list = (&packet_data_head.list)->next;
next              340 drivers/platform/x86/dell_rbu.c 		pnext_list = ptemp_list->next;
next              469 drivers/platform/x86/intel_menlow.c 	struct intel_menlow_attribute *pos, *next;
next              472 drivers/platform/x86/intel_menlow.c 	list_for_each_entry_safe(pos, next, &intel_menlow_attr_list, node) {
next              813 drivers/platform/x86/wmi.c 	struct wmi_block *next = NULL;
next              815 drivers/platform/x86/wmi.c 	list_for_each_entry_safe(wblock, next, &wmi_block_list, list) {
next             1107 drivers/platform/x86/wmi.c 	struct wmi_block *wblock, *next;
next             1110 drivers/platform/x86/wmi.c 	list_for_each_entry_safe(wblock, next, &wmi_block_list, list) {
next             1147 drivers/platform/x86/wmi.c 	struct wmi_block *wblock, *next;
next             1209 drivers/platform/x86/wmi.c 	list_for_each_entry_safe(wblock, next, &wmi_block_list, list) {
next              124 drivers/pnp/card.c 	dev_id->next = NULL;
next              126 drivers/pnp/card.c 	while (ptr && ptr->next)
next              127 drivers/pnp/card.c 		ptr = ptr->next;
next              129 drivers/pnp/card.c 		ptr->next = dev_id;
next              139 drivers/pnp/card.c 	struct pnp_id *next;
next              143 drivers/pnp/card.c 		next = id->next;
next              145 drivers/pnp/card.c 		id = next;
next              205 drivers/pnp/card.c 		pos = pos->next;
next              342 drivers/pnp/card.c 		pos = card->devices.next;
next              346 drivers/pnp/card.c 		pos = from->card_list.next;
next              352 drivers/pnp/card.c 		pos = pos->next;
next              106 drivers/pnp/core.c 	struct pnp_id *next;
next              110 drivers/pnp/core.c 		next = id->next;
next              112 drivers/pnp/core.c 		id = next;
next              227 drivers/pnp/core.c 	for (id = dev->id; id; id = id->next)
next               39 drivers/pnp/driver.c 		pos = pos->next;
next              302 drivers/pnp/driver.c 	dev_id->next = NULL;
next              304 drivers/pnp/driver.c 	while (ptr && ptr->next)
next              305 drivers/pnp/driver.c 		ptr = ptr->next;
next              307 drivers/pnp/driver.c 		ptr->next = dev_id;
next              449 drivers/pnp/interface.c 		pos = pos->next;
next               34 drivers/pnp/isapnp/compat.c 	list = from ? from->global_list.next : pnp_cards.next;
next               41 drivers/pnp/isapnp/compat.c 		list = list->next;
next               57 drivers/pnp/isapnp/compat.c 		list = pnp_global.next;
next               59 drivers/pnp/isapnp/compat.c 			list = from->global_list.next;
next               67 drivers/pnp/isapnp/compat.c 			list = list->next;
next               72 drivers/pnp/isapnp/compat.c 		list = card->devices.next;
next               74 drivers/pnp/isapnp/compat.c 			list = from->card_list.next;
next               83 drivers/pnp/isapnp/compat.c 			list = list->next;
next              199 drivers/pnp/quirks.c 			new_option = list_entry(new_option->list.next,
next             1204 drivers/pwm/core.c 	.next = pwm_seq_next,
next              958 drivers/rapidio/rio-scan.c 	if (mport->nnode.next || mport->nnode.prev)
next              846 drivers/rapidio/rio.c 	n = from ? from->global_list.next : rio_devices.next;
next              852 drivers/rapidio/rio.c 		n = n->next;
next             1440 drivers/rapidio/rio.c 	n = from ? from->global_list.next : rio_devices.next;
next             1449 drivers/rapidio/rio.c 		n = n->next;
next              364 drivers/remoteproc/qcom_q6v5_mss.c 	struct qcom_scm_vmperm next;
next              373 drivers/remoteproc/qcom_q6v5_mss.c 	next.vmid = remote_owner ? QCOM_SCM_VMID_MSS_MSA : QCOM_SCM_VMID_HLOS;
next              374 drivers/remoteproc/qcom_q6v5_mss.c 	next.perm = remote_owner ? QCOM_SCM_PERM_RW : QCOM_SCM_PERM_RWX;
next              377 drivers/remoteproc/qcom_q6v5_mss.c 				   current_perm, &next, 1);
next              793 drivers/rtc/interface.c 	struct timerqueue_node *next = timerqueue_getnext(&rtc->timerqueue);
next              802 drivers/rtc/interface.c 	while (next) {
next              803 drivers/rtc/interface.c 		if (next->expires >= now)
next              805 drivers/rtc/interface.c 		next = timerqueue_iterate_next(next);
next              810 drivers/rtc/interface.c 	if (!next || ktime_before(timer->node.expires, next->expires)) {
next              853 drivers/rtc/interface.c 	struct timerqueue_node *next = timerqueue_getnext(&rtc->timerqueue);
next              858 drivers/rtc/interface.c 	if (next == &timer->node) {
next              862 drivers/rtc/interface.c 		next = timerqueue_getnext(&rtc->timerqueue);
next              863 drivers/rtc/interface.c 		if (!next) {
next              867 drivers/rtc/interface.c 		alarm.time = rtc_ktime_to_tm(next->expires);
next              890 drivers/rtc/interface.c 	struct timerqueue_node *next;
next              901 drivers/rtc/interface.c 	while ((next = timerqueue_getnext(&rtc->timerqueue))) {
next              902 drivers/rtc/interface.c 		if (next->expires > now)
next              906 drivers/rtc/interface.c 		timer = container_of(next, struct rtc_timer, node);
next              925 drivers/rtc/interface.c 	if (next) {
next              930 drivers/rtc/interface.c 		alarm.time = rtc_ktime_to_tm(next->expires);
next              940 drivers/rtc/interface.c 			timer = container_of(next, struct rtc_timer, node);
next               76 drivers/rtc/rtc-88pm80x.c static void rtc_next_alarm_time(struct rtc_time *next, struct rtc_time *now,
next               82 drivers/rtc/rtc-88pm80x.c 	next->tm_year = now->tm_year;
next               83 drivers/rtc/rtc-88pm80x.c 	next->tm_mon = now->tm_mon;
next               84 drivers/rtc/rtc-88pm80x.c 	next->tm_mday = now->tm_mday;
next               85 drivers/rtc/rtc-88pm80x.c 	next->tm_hour = alrm->tm_hour;
next               86 drivers/rtc/rtc-88pm80x.c 	next->tm_min = alrm->tm_min;
next               87 drivers/rtc/rtc-88pm80x.c 	next->tm_sec = alrm->tm_sec;
next               90 drivers/rtc/rtc-88pm80x.c 	next_time = rtc_tm_to_time64(next);
next               95 drivers/rtc/rtc-88pm80x.c 		rtc_time64_to_tm(next_time, next);
next               83 drivers/rtc/rtc-88pm860x.c static void rtc_next_alarm_time(struct rtc_time *next, struct rtc_time *now,
next               89 drivers/rtc/rtc-88pm860x.c 	next->tm_year = now->tm_year;
next               90 drivers/rtc/rtc-88pm860x.c 	next->tm_mon = now->tm_mon;
next               91 drivers/rtc/rtc-88pm860x.c 	next->tm_mday = now->tm_mday;
next               92 drivers/rtc/rtc-88pm860x.c 	next->tm_hour = alrm->tm_hour;
next               93 drivers/rtc/rtc-88pm860x.c 	next->tm_min = alrm->tm_min;
next               94 drivers/rtc/rtc-88pm860x.c 	next->tm_sec = alrm->tm_sec;
next               97 drivers/rtc/rtc-88pm860x.c 	rtc_tm_to_time(next, &next_time);
next              102 drivers/rtc/rtc-88pm860x.c 		rtc_time_to_tm(next_time, next);
next             1678 drivers/s390/block/dasd.c 	struct dasd_ccw_req *cqr, *next, *fcqr;
next             1830 drivers/s390/block/dasd.c 	next = NULL;
next             1838 drivers/s390/block/dasd.c 		if (cqr->devlist.next != &device->ccw_queue) {
next             1839 drivers/s390/block/dasd.c 			next = list_entry(cqr->devlist.next,
next             1864 drivers/s390/block/dasd.c 			next = cqr;
next             1868 drivers/s390/block/dasd.c 	if (next && (next->status == DASD_CQR_QUEUED) &&
next             1870 drivers/s390/block/dasd.c 		if (device->discipline->start_IO(next) == 0)
next             1871 drivers/s390/block/dasd.c 			expires = next->expires;
next             2018 drivers/s390/block/dasd.c 	cqr = list_entry(device->ccw_queue.next, struct dasd_ccw_req, devlist);
next             2087 drivers/s390/block/dasd.c 	cqr = list_entry(device->ccw_queue.next, struct dasd_ccw_req, devlist);
next             2584 drivers/s390/block/dasd.c 	cqr = list_entry(device->ccw_queue.next, struct dasd_ccw_req, devlist);
next             2621 drivers/s390/block/dasd.c 	list_add(&cqr->devlist, device->ccw_queue.next);
next              374 drivers/s390/block/dasd_alias.c 	if (group->next == device)
next              375 drivers/s390/block/dasd_alias.c 		group->next = NULL;
next              682 drivers/s390/block/dasd_alias.c 	alias_device = group->next;
next              694 drivers/s390/block/dasd_alias.c 		group->next = list_first_entry(&group->aliaslist,
next              697 drivers/s390/block/dasd_alias.c 		group->next = list_first_entry(&alias_device->alias_list,
next              236 drivers/s390/block/dasd_diag.c 	struct dasd_ccw_req *cqr, *next;
next              288 drivers/s390/block/dasd_diag.c 			next = list_entry(device->ccw_queue.next,
next              290 drivers/s390/block/dasd_diag.c 			if (next->status == DASD_CQR_QUEUED) {
next              291 drivers/s390/block/dasd_diag.c 				rc = dasd_start_diag(next);
next              293 drivers/s390/block/dasd_diag.c 					expires = next->expires;
next              650 drivers/s390/block/dasd_eckd.h 	struct dasd_device *next;
next              697 drivers/s390/block/dasd_int.h 	if (left->next != chunk_list) {
next              698 drivers/s390/block/dasd_int.h 		tmp = list_entry(left->next, struct dasd_mchunk, list);
next              712 drivers/s390/block/dasd_int.h 	__list_add(&chunk->list, left, left->next);
next              129 drivers/s390/block/dasd_proc.c 	.next		= dasd_devices_next,
next               73 drivers/s390/char/con3215.c 	struct raw3215_req *next;     /* pointer to next request */
next              116 drivers/s390/char/con3215.c 	raw3215_freelist = req->next;
next              132 drivers/s390/char/con3215.c 	req->next = raw3215_freelist;
next              925 drivers/s390/char/con3215.c 		req->next = raw3215_freelist;
next               91 drivers/s390/char/monwriter.c 	struct mon_buf *entry, *next;
next               93 drivers/s390/char/monwriter.c 	list_for_each_entry_safe(entry, next, &monpriv->list, list)
next              209 drivers/s390/char/monwriter.c 	struct mon_buf *entry, *next;
next              211 drivers/s390/char/monwriter.c 	list_for_each_entry_safe(entry, next, &monpriv->list, list) {
next              370 drivers/s390/char/raw3270.c 		rq = list_entry(rp->req_queue.next,struct raw3270_request,list);
next              652 drivers/s390/char/raw3270.c 		rq = list_entry(rp->req_queue.next,struct raw3270_request,list);
next             1183 drivers/s390/char/raw3270.c 		v = list_entry(rp->view_list.next, struct raw3270_view, list);
next              257 drivers/s390/char/raw3270.h 	if (left->next != free_list) {
next              258 drivers/s390/char/raw3270.h 		tmp = list_entry(left->next, struct string, list);
next              272 drivers/s390/char/raw3270.h 	__list_add(&cs->list, left, left->next);
next              315 drivers/s390/char/sclp.c 		req = list_entry(sclp_req_queue.next, struct sclp_req, list);
next              145 drivers/s390/char/sclp_con.c 	list = sclp_con_outqueue.next;
next              148 drivers/s390/char/sclp_con.c 		list = list->next;
next              190 drivers/s390/char/sclp_con.c 			page = sclp_con_pages.next;
next              123 drivers/s390/char/sclp_tty.c 			buffer = list_entry(sclp_tty_outqueue.next,
next              195 drivers/s390/char/sclp_tty.c 			page = sclp_tty_pages.next;
next              136 drivers/s390/char/sclp_vt220.c 			request = list_entry(sclp_vt220_outqueue.next,
next              379 drivers/s390/char/sclp_vt220.c 	list = sclp_vt220_outqueue.next;
next              382 drivers/s390/char/sclp_vt220.c 		list = list->next;
next              432 drivers/s390/char/sclp_vt220.c 			page = (void *) sclp_vt220_empty.next;
next              146 drivers/s390/char/tape_core.c 		req = list_entry(tdev->req_queue.next, struct tape_request,
next              877 drivers/s390/char/tape_core.c 	request = list_entry(device->req_queue.next, struct tape_request, list);
next             1156 drivers/s390/char/tape_core.c 		req = list_entry(device->req_queue.next,
next               72 drivers/s390/char/tape_proc.c 		request = list_entry(device->req_queue.next,
next              103 drivers/s390/char/tape_proc.c 	.next		= tape_proc_next,
next              465 drivers/s390/char/tty3270.c 		s = list_entry(tp->rcl_lines.next, struct string, list);
next              391 drivers/s390/cio/blacklist.c 	.next  = cio_ignore_proc_seq_next,
next               31 drivers/s390/cio/vfio_ccw_cp.c 	struct list_head	next;
next              335 drivers/s390/cio/vfio_ccw_cp.c 	list_add_tail(&chain->next, &cp->ccwchain_list);
next              342 drivers/s390/cio/vfio_ccw_cp.c 	list_del(&chain->next);
next              411 drivers/s390/cio/vfio_ccw_cp.c 	list_for_each_entry(chain, &cp->ccwchain_list, next) {
next              494 drivers/s390/cio/vfio_ccw_cp.c 	list_for_each_entry(iter, &cp->ccwchain_list, next) {
next              683 drivers/s390/cio/vfio_ccw_cp.c 	list_for_each_entry_safe(chain, temp, &cp->ccwchain_list, next) {
next              738 drivers/s390/cio/vfio_ccw_cp.c 	list_for_each_entry(chain, &cp->ccwchain_list, next) {
next              751 drivers/s390/cio/vfio_ccw_cp.c 	list_for_each_entry_continue(chain, &cp->ccwchain_list, next) {
next              786 drivers/s390/cio/vfio_ccw_cp.c 	chain = list_first_entry(&cp->ccwchain_list, struct ccwchain, next);
next              822 drivers/s390/cio/vfio_ccw_cp.c 	list_for_each_entry(chain, &cp->ccwchain_list, next) {
next              859 drivers/s390/cio/vfio_ccw_cp.c 	list_for_each_entry(chain, &cp->ccwchain_list, next) {
next              242 drivers/s390/crypto/ap_queue.c 	ap_msg = list_entry(aq->requestq.next, struct ap_message, list);
next              723 drivers/s390/crypto/ap_queue.c 	struct ap_message *ap_msg, *next;
next              725 drivers/s390/crypto/ap_queue.c 	list_for_each_entry_safe(ap_msg, next, &aq->pendingq, list) {
next              731 drivers/s390/crypto/ap_queue.c 	list_for_each_entry_safe(ap_msg, next, &aq->requestq, list) {
next              214 drivers/s390/net/ctcm_main.c 			*c = ch->next;
next              236 drivers/s390/net/ctcm_main.c 		c = &((*c)->next);
next              258 drivers/s390/net/ctcm_main.c 		ch = ch->next;
next             1464 drivers/s390/net/ctcm_main.c 		c = &(*c)->next;
next             1484 drivers/s390/net/ctcm_main.c 	ch->next = *c;
next              132 drivers/s390/net/ctcm_main.h 	struct channel *next;
next              441 drivers/s390/net/lcs.c 		ipm = list_entry(card->ipm_list.next,
next              644 drivers/s390/net/lcs.c 	int prev, next;
next              648 drivers/s390/net/lcs.c 	next = (index + 1) & (LCS_NUM_BUFFS - 1);
next              650 drivers/s390/net/lcs.c 	if (channel->ccws[next].flags & CCW_FLAG_SUSPEND) {
next              689 drivers/s390/net/lcs.c 	int index, prev, next;
next              696 drivers/s390/net/lcs.c 	next = (index + 1) & (LCS_NUM_BUFFS - 1);
next              711 drivers/s390/net/lcs.c 	channel->ccws[next].flags &= ~CCW_FLAG_PCI;
next              223 drivers/s390/net/netiucv.c 	u16 next;
next              630 drivers/s390/net/netiucv.c 		if (!header->next)
next              634 drivers/s390/net/netiucv.c 		header->next -= offset;
next              635 drivers/s390/net/netiucv.c 		offset += header->next;
next              636 drivers/s390/net/netiucv.c 		header->next -= NETIUCV_HDRLEN;
next              637 drivers/s390/net/netiucv.c 		if (skb_tailroom(pskb) < header->next) {
next              639 drivers/s390/net/netiucv.c 				header->next, skb_tailroom(pskb));
next              642 drivers/s390/net/netiucv.c 		skb_put(pskb, header->next);
next              664 drivers/s390/net/netiucv.c 		skb_pull(pskb, header->next);
next              747 drivers/s390/net/netiucv.c 		header.next = conn->tx_buff->len + skb->len + NETIUCV_HDRLEN;
next              767 drivers/s390/net/netiucv.c 	header.next = 0;
next             1196 drivers/s390/net/netiucv.c 		header.next = nskb->len + NETIUCV_HDRLEN;
next             1198 drivers/s390/net/netiucv.c 		header.next = 0;
next             2149 drivers/s390/net/netiucv.c 		cp = list_entry(iucv_connection_list.next,
next             2608 drivers/s390/net/qeth_core_main.c 	entry = list_entry(card->qdio.in_buf_pool.entry_list.next,
next             1211 drivers/s390/net/qeth_l3_main.c 	for (im6 = in6_dev->mc_list; im6 != NULL; im6 = im6->next) {
next             1537 drivers/s390/scsi/zfcp_erp.c 	struct list_head *next;
next             1550 drivers/s390/scsi/zfcp_erp.c 		next = adapter->erp_ready_head.next;
next             1553 drivers/s390/scsi/zfcp_erp.c 		if (next != &adapter->erp_ready_head) {
next             1554 drivers/s390/scsi/zfcp_erp.c 			act = list_entry(next, struct zfcp_erp_action, list);
next               65 drivers/s390/scsi/zfcp_fc.c 	unsigned long next = adapter->next_port_scan;
next               69 drivers/s390/scsi/zfcp_fc.c 	if (time_before(now, next)) {
next               70 drivers/s390/scsi/zfcp_fc.c 		delay = next - now;
next               55 drivers/sbus/char/bbc_i2c.h 	struct bbc_i2c_bus		*next;
next              163 drivers/sbus/char/openprom.c 		    !prop->next ||
next              164 drivers/sbus/char/openprom.c 		    (len = strlen(prop->next->name)) + 1 > bufsize)
next              167 drivers/sbus/char/openprom.c 		prop = prop->next;
next              441 drivers/sbus/char/openprom.c 			prop = prop->next;
next              175 drivers/scsi/BusLogic.c 		ccb->next = adapter->free_ccbs;
next              284 drivers/scsi/BusLogic.c 		adapter->free_ccbs = ccb->next;
next              285 drivers/scsi/BusLogic.c 		ccb->next = NULL;
next              296 drivers/scsi/BusLogic.c 	adapter->free_ccbs = ccb->next;
next              297 drivers/scsi/BusLogic.c 	ccb->next = NULL;
next              320 drivers/scsi/BusLogic.c 	ccb->next = adapter->free_ccbs;
next             2600 drivers/scsi/BusLogic.c 	ccb->next = NULL;
next             2605 drivers/scsi/BusLogic.c 		adapter->lastccb->next = ccb;
next             2746 drivers/scsi/BusLogic.c 		adapter->firstccb = ccb->next;
next             3881 drivers/scsi/BusLogic.c 	struct blogic_adapter *ha, *next;
next             3883 drivers/scsi/BusLogic.c 	list_for_each_entry_safe(ha, next, &blogic_host_list, host_list)
next              855 drivers/scsi/BusLogic.h 	struct blogic_ccb *next;
next              812 drivers/scsi/aacraid/aachba.c 	scsicmd->list.next = NULL;
next             4000 drivers/scsi/aacraid/aachba.c 	psg->sg[0].next = 0;
next             4014 drivers/scsi/aacraid/aachba.c 		psg->sg[i].next = 0;
next              476 drivers/scsi/aacraid/aacraid.h 	__le32		next;	/* reserved for F/W use */
next              484 drivers/scsi/aacraid/aacraid.h 	u32		next;	/* reserved for F/W use */
next             1230 drivers/scsi/aacraid/aacraid.h 	struct list_head	next;		// used to link context's into a linked list
next             1292 drivers/scsi/aacraid/aacraid.h 	void			*next;	/* this is used by the allocator */
next              206 drivers/scsi/aacraid/commctrl.c 		entry = dev->fib_list.next;
next              208 drivers/scsi/aacraid/commctrl.c 			context = list_entry(entry, struct aac_fib_context, next);
next              212 drivers/scsi/aacraid/commctrl.c 				entry = dev->fib_list.next;
next              214 drivers/scsi/aacraid/commctrl.c 				entry = entry->next;
next              217 drivers/scsi/aacraid/commctrl.c 		list_add_tail(&fibctx->next, &dev->fib_list);
next              256 drivers/scsi/aacraid/commctrl.c 	entry = dev->fib_list.next;
next              260 drivers/scsi/aacraid/commctrl.c 		fibctx = list_entry(entry, struct aac_fib_context, next);
next              267 drivers/scsi/aacraid/commctrl.c 		entry = entry->next;
next              292 drivers/scsi/aacraid/commctrl.c 		entry = fibctx->fib_list.next;
next              350 drivers/scsi/aacraid/commctrl.c 		entry = fibctx->fib_list.next;
next              363 drivers/scsi/aacraid/commctrl.c 	list_del(&fibctx->next);
next              397 drivers/scsi/aacraid/commctrl.c 	entry = dev->fib_list.next;
next              401 drivers/scsi/aacraid/commctrl.c 		fibctx = list_entry(entry, struct aac_fib_context, next);
next              407 drivers/scsi/aacraid/commctrl.c 		entry = entry->next;
next              176 drivers/scsi/aacraid/commsup.c 		fibptr->next = fibptr+1;	/* Forward chain the fibs */
next              206 drivers/scsi/aacraid/commsup.c 	dev->fibs[dev->scsi_host_ptr->can_queue + AAC_NUM_MGT_FIB - 1].next = NULL;
next              257 drivers/scsi/aacraid/commsup.c 	dev->free_fib = fibptr->next;
next              299 drivers/scsi/aacraid/commsup.c 	fibptr->next = fibptr->dev->free_fib;
next             1748 drivers/scsi/aacraid/commsup.c 	entry = aac->fib_list.next;
next             1760 drivers/scsi/aacraid/commsup.c 		struct aac_fib_context *fibctx = list_entry(entry, struct aac_fib_context, next);
next             1780 drivers/scsi/aacraid/commsup.c 				entry = entry->next;
next             1824 drivers/scsi/aacraid/commsup.c 		entry = entry->next;
next             2029 drivers/scsi/aacraid/commsup.c 	entry = dev->fib_list.next;
next             2031 drivers/scsi/aacraid/commsup.c 		entry = entry->next;
next             2087 drivers/scsi/aacraid/commsup.c 	entry = dev->fib_list.next;
next             2102 drivers/scsi/aacraid/commsup.c 				next);
next             2120 drivers/scsi/aacraid/commsup.c 				entry = entry->next;
next             2131 drivers/scsi/aacraid/commsup.c 			entry = entry->next;
next             2157 drivers/scsi/aacraid/commsup.c 		entry = entry->next;
next             2187 drivers/scsi/aacraid/commsup.c 		entry = dev->queues->queue[HostNormCmdQueue].cmdq.next;
next              101 drivers/scsi/aacraid/src.c 				entry = dev->sync_fib_list.next;
next              502 drivers/scsi/aha152x.c 	struct scsi_cmnd *next;	/* next sc in queue */
next              554 drivers/scsi/aha152x.c #define SCNEXT(SCpnt)		SCDATA(SCpnt)->next
next             1110 drivers/scsi/aha152x.c 		struct scsi_cmnd *next;
next             1113 drivers/scsi/aha152x.c 			next = SCNEXT(ptr);
next             1117 drivers/scsi/aha152x.c 			next = NULL;
next             1127 drivers/scsi/aha152x.c 		ptr = next;
next             2250 drivers/scsi/aha152x.c 		struct scsi_cmnd *next = SCNEXT(ptr);
next             2262 drivers/scsi/aha152x.c 		ptr = next;
next              200 drivers/scsi/aic7xxx/aic79xx_core.c 				     u_int prev, u_int next, u_int tid);
next             2909 drivers/scsi/aic7xxx/aic79xx_core.c 		u_int next;
next             2961 drivers/scsi/aic7xxx/aic79xx_core.c 				next = SCB_LIST_NULL;
next             2964 drivers/scsi/aic7xxx/aic79xx_core.c 				next = ahd_inw_scbram(ahd, SCB_NEXT2);
next             2967 drivers/scsi/aic7xxx/aic79xx_core.c 			ahd_outw(ahd, SCB_NEXT2, next);
next             8382 drivers/scsi/aic7xxx/aic79xx_core.c 	u_int	next;
next             8389 drivers/scsi/aic7xxx/aic79xx_core.c 	next = *list_head;
next             8391 drivers/scsi/aic7xxx/aic79xx_core.c 	for (scbid = next; !SCBID_IS_NULL(scbid); scbid = next) {
next             8407 drivers/scsi/aic7xxx/aic79xx_core.c 		next = ahd_inw_scbram(ahd, SCB_NEXT);
next             8421 drivers/scsi/aic7xxx/aic79xx_core.c 			ahd_rem_wscb(ahd, scbid, prev, next, tid);
next             8424 drivers/scsi/aic7xxx/aic79xx_core.c 				*list_head = next;
next             8482 drivers/scsi/aic7xxx/aic79xx_core.c 	     u_int prev, u_int next, u_int tid)
next             8489 drivers/scsi/aic7xxx/aic79xx_core.c 		ahd_outw(ahd, SCB_NEXT, next);
next             8500 drivers/scsi/aic7xxx/aic79xx_core.c 	if (SCBID_IS_NULL(next)
next             8505 drivers/scsi/aic7xxx/aic79xx_core.c 	return (next);
next              472 drivers/scsi/aic7xxx/aic7xxx.h /*31*/	uint8_t  next;			/*
next              593 drivers/scsi/aic7xxx/aic7xxx_core.c 	q_hscb->next = scb->hscb->tag;
next              612 drivers/scsi/aic7xxx/aic7xxx_core.c 	 || scb->hscb->next == SCB_LIST_NULL)
next              614 drivers/scsi/aic7xxx/aic7xxx_core.c 		      scb->hscb->tag, scb->hscb->next);
next             5804 drivers/scsi/aic7xxx/aic7xxx_core.c 		prev_scb->hscb->next = scb->hscb->tag;
next             5809 drivers/scsi/aic7xxx/aic7xxx_core.c 	scb->hscb->next = ahc->next_queued_scb->hscb->tag;
next             5838 drivers/scsi/aic7xxx/aic7xxx_core.c 	uint8_t next;
next             5949 drivers/scsi/aic7xxx/aic7xxx_core.c 		next = scb->hscb->next;
next             5952 drivers/scsi/aic7xxx/aic7xxx_core.c 		scb->hscb->next = next;
next             5961 drivers/scsi/aic7xxx/aic7xxx_core.c 		scb->hscb->next = ahc->next_queued_scb->hscb->tag;
next             5968 drivers/scsi/aic7xxx/aic7xxx_core.c 	next = ahc_inb(ahc, WAITING_SCBH);  /* Start at head of list. */
next             5971 drivers/scsi/aic7xxx/aic7xxx_core.c 	while (next != SCB_LIST_NULL) {
next             5974 drivers/scsi/aic7xxx/aic7xxx_core.c 		ahc_outb(ahc, SCBPTR, next);
next             5986 drivers/scsi/aic7xxx/aic7xxx_core.c 				scb_index, next);
next             6014 drivers/scsi/aic7xxx/aic7xxx_core.c 				next = ahc_rem_wscb(ahc, next, prev);
next             6017 drivers/scsi/aic7xxx/aic7xxx_core.c 				prev = next;
next             6018 drivers/scsi/aic7xxx/aic7xxx_core.c 				next = ahc_inb(ahc, SCB_NEXT);
next             6023 drivers/scsi/aic7xxx/aic7xxx_core.c 			prev = next;
next             6024 drivers/scsi/aic7xxx/aic7xxx_core.c 			next = ahc_inb(ahc, SCB_NEXT);
next             6141 drivers/scsi/aic7xxx/aic7xxx_core.c 	u_int	next;
next             6147 drivers/scsi/aic7xxx/aic7xxx_core.c 	next = ahc_inb(ahc, DISCONNECTED_SCBH);
next             6157 drivers/scsi/aic7xxx/aic7xxx_core.c 	while (next != SCB_LIST_NULL) {
next             6160 drivers/scsi/aic7xxx/aic7xxx_core.c 		ahc_outb(ahc, SCBPTR, next);
next             6170 drivers/scsi/aic7xxx/aic7xxx_core.c 		if (next == prev) {
next             6173 drivers/scsi/aic7xxx/aic7xxx_core.c 			      next, prev);
next             6180 drivers/scsi/aic7xxx/aic7xxx_core.c 				next =
next             6181 drivers/scsi/aic7xxx/aic7xxx_core.c 				    ahc_rem_scb_from_disc_list(ahc, prev, next);
next             6183 drivers/scsi/aic7xxx/aic7xxx_core.c 				prev = next;
next             6184 drivers/scsi/aic7xxx/aic7xxx_core.c 				next = ahc_inb(ahc, SCB_NEXT);
next             6189 drivers/scsi/aic7xxx/aic7xxx_core.c 			prev = next;
next             6190 drivers/scsi/aic7xxx/aic7xxx_core.c 			next = ahc_inb(ahc, SCB_NEXT);
next             6205 drivers/scsi/aic7xxx/aic7xxx_core.c 	u_int next;
next             6208 drivers/scsi/aic7xxx/aic7xxx_core.c 	next = ahc_inb(ahc, SCB_NEXT);
next             6216 drivers/scsi/aic7xxx/aic7xxx_core.c 		ahc_outb(ahc, SCB_NEXT, next);
next             6218 drivers/scsi/aic7xxx/aic7xxx_core.c 		ahc_outb(ahc, DISCONNECTED_SCBH, next);
next             6220 drivers/scsi/aic7xxx/aic7xxx_core.c 	return (next);
next             6250 drivers/scsi/aic7xxx/aic7xxx_core.c 	u_int curscb, next;
next             6258 drivers/scsi/aic7xxx/aic7xxx_core.c 	next = ahc_inb(ahc, SCB_NEXT);
next             6268 drivers/scsi/aic7xxx/aic7xxx_core.c 		ahc_outb(ahc, WAITING_SCBH, next); 
next             6281 drivers/scsi/aic7xxx/aic7xxx_core.c 		ahc_outb(ahc, SCB_NEXT, next);
next             6288 drivers/scsi/aic7xxx/aic7xxx_core.c 	return next;
next              324 drivers/scsi/aic94xx/aic94xx_hwi.c 	for (i = 0; i < seq->num_escbs; i++, escb = list_entry(escb->list.next,
next             1364 drivers/scsi/aic94xx/aic94xx_hwi.c 		ascb = list_entry(ascb->list.next, struct asd_ascb, list);
next              427 drivers/scsi/aic94xx/aic94xx_sds.c 	u16   next;
next              511 drivers/scsi/aic94xx/aic94xx_sds.c 	__le16 next;
next              535 drivers/scsi/aic94xx/aic94xx_sds.c 	u16   next;
next              544 drivers/scsi/aic94xx/aic94xx_sds.c 	__le16  next;
next              726 drivers/scsi/aic94xx/aic94xx_sds.c 		el = start + le16_to_cpu(el->next);
next             3028 drivers/scsi/arcmsr/arcmsr_hba.c 		ccb = list_entry(head->next, struct CommandControlBlock, list);
next               27 drivers/scsi/arm/msgqueue.c 		msgq->free = mq->next;
next               41 drivers/scsi/arm/msgqueue.c 		mq->next = msgq->free;
next               59 drivers/scsi/arm/msgqueue.c 		msgq->entries[i].next = &msgq->entries[i + 1];
next               61 drivers/scsi/arm/msgqueue.c 	msgq->entries[NR_MESSAGES - 1].next = NULL;
next               85 drivers/scsi/arm/msgqueue.c 	for (mq = msgq->qe; mq; mq = mq->next)
next              102 drivers/scsi/arm/msgqueue.c 	for (mq = msgq->qe; mq && msgno; mq = mq->next, msgno--);
next              131 drivers/scsi/arm/msgqueue.c 		mq->next = NULL;
next              135 drivers/scsi/arm/msgqueue.c 			mqp = &(*mqp)->next;
next              153 drivers/scsi/arm/msgqueue.c 		mqnext = mq->next;
next               20 drivers/scsi/arm/msgqueue.h     struct msgqueue_entry *next;
next              114 drivers/scsi/arm/queue.c 	l = queue->free.next;
next              191 drivers/scsi/arm/queue.c 		SCpnt = __queue_remove(queue, queue->head.next);
next              116 drivers/scsi/bfa/bfa_cs.h #define bfa_q_first(_q) ((void *)(((struct list_head *) (_q))->next))
next              117 drivers/scsi/bfa/bfa_cs.h #define bfa_q_next(_qe) (((struct list_head *) (_qe))->next)
next              983 drivers/scsi/bnx2fc/57xx_hsi_bnx2fc.h 	struct regpair next;
next              152 drivers/scsi/bnx2fc/bnx2fc_fcoe.c 	struct sk_buff *skb, *next;
next              157 drivers/scsi/bnx2fc/bnx2fc_fcoe.c 	skb_queue_walk_safe(list, skb, next) {
next             2764 drivers/scsi/bnx2fc/bnx2fc_fcoe.c 	struct bnx2fc_hba *hba, *next;
next             2781 drivers/scsi/bnx2fc/bnx2fc_fcoe.c 	list_for_each_entry_safe(hba, next, &to_be_deleted, list) {
next             2111 drivers/scsi/bnx2fc/bnx2fc_hwi.c 		hba->t2_hash_tbl[i].next.lo = addr & 0xffffffff;
next             2112 drivers/scsi/bnx2fc/bnx2fc_hwi.c 		hba->t2_hash_tbl[i].next.hi = addr >> 32;
next              438 drivers/scsi/bnx2fc/bnx2fc_io.c 			cmd_mgr->free_list[index].next;
next              493 drivers/scsi/bnx2fc/bnx2fc_io.c 		cmd_mgr->free_list[index].next;
next              618 drivers/scsi/bnx2fc/bnx2fc_tgt.c 	u32 conn_id, next;
next              629 drivers/scsi/bnx2fc/bnx2fc_tgt.c 	next = hba->next_conn_id;
next              639 drivers/scsi/bnx2fc/bnx2fc_tgt.c 		if (conn_id == next) {
next              531 drivers/scsi/bnx2i/bnx2i_init.c 		hba = list_entry(adapter_list.next, struct bnx2i_hba, link);
next               69 drivers/scsi/csiostor/csio_defs.h 	return ((list->next == list) && (list->prev == list));
next               72 drivers/scsi/csiostor/csio_defs.h #define csio_list_next(elem)	(((struct list_head *)(elem))->next)
next              938 drivers/scsi/csiostor/csio_lnode.c 	struct list_head *tmp, *next;
next              941 drivers/scsi/csiostor/csio_lnode.c 	list_for_each_safe(tmp, next, &rnhead->sm.sm_list) {
next              122 drivers/scsi/csiostor/csio_scsi.c 	struct list_head *tmp, *next;
next              133 drivers/scsi/csiostor/csio_scsi.c 	list_for_each_safe(tmp, next, &scm->active_q) {
next             1163 drivers/scsi/csiostor/csio_scsi.c 	struct list_head *tmp, *next;
next             1167 drivers/scsi/csiostor/csio_scsi.c 	list_for_each_safe(tmp, next, q) {
next             1236 drivers/scsi/csiostor/csio_scsi.c 	struct list_head *tmp, *next;
next             1246 drivers/scsi/csiostor/csio_scsi.c 	list_for_each_safe(tmp, next, q) {
next              470 drivers/scsi/cxgbi/libcxgbi.c 	start = idx = pmap->next;
next              477 drivers/scsi/cxgbi/libcxgbi.c 			pmap->next = idx;
next              485 drivers/scsi/cxgbi/libcxgbi.c 				pmap->sport_base + idx, pmap->next);
next              494 drivers/scsi/cxgbi/libcxgbi.c 		pmap->next);
next             1142 drivers/scsi/cxgbi/libcxgbi.c 	struct sk_buff *next;
next             1191 drivers/scsi/cxgbi/libcxgbi.c 		next = skb->next;
next             1192 drivers/scsi/cxgbi/libcxgbi.c 		skb->next = NULL;
next             1198 drivers/scsi/cxgbi/libcxgbi.c 		skb = next;
next              451 drivers/scsi/cxgbi/libcxgbi.h 	unsigned int next;
next              711 drivers/scsi/dc395x.c 	struct DeviceCtlBlk* next = NULL;
next              720 drivers/scsi/dc395x.c 			next = i;
next              726 drivers/scsi/dc395x.c 	if (!next)
next              728 drivers/scsi/dc395x.c         		next = i;
next              732 drivers/scsi/dc395x.c 	return next;
next              800 drivers/scsi/dc395x.c 		start = list_entry(dcb_list_head->next, typeof(*start), list);
next              822 drivers/scsi/dc395x.c 			srb = list_entry(waiting_list_head->next,
next               68 drivers/scsi/dpt/dpti_i2o.h 	struct i2o_device *next;	/* Chain */
next               93 drivers/scsi/dpt/dpti_i2o.h 	struct i2o_controller *next;		/* Controller chain */
next              137 drivers/scsi/dpt_i2o.c 	struct adpt_i2o_post_wait_data *next;
next              194 drivers/scsi/dpt_i2o.c 	adpt_hba *next;
next              212 drivers/scsi/dpt_i2o.c 	for (pHba = hba_chain; pHba; pHba = next) {
next              213 drivers/scsi/dpt_i2o.c 		next = pHba->next;
next              239 drivers/scsi/dpt_i2o.c 	for (pHba = hba_chain; pHba; pHba = pHba->next) {
next              250 drivers/scsi/dpt_i2o.c 	for (pHba = hba_chain; pHba; pHba = next) {
next              251 drivers/scsi/dpt_i2o.c 		next = pHba->next;
next              271 drivers/scsi/dpt_i2o.c 	for (pHba = hba_chain; pHba; pHba = next) {
next              272 drivers/scsi/dpt_i2o.c 		next = pHba->next;
next              551 drivers/scsi/dpt_i2o.c 	for (pHba = hba_chain; pHba; pHba = pHba->next) {
next              842 drivers/scsi/dpt_i2o.c 		pNext = pHba->next;
next              853 drivers/scsi/dpt_i2o.c 		p1 = p1->next;
next              973 drivers/scsi/dpt_i2o.c 		for(p = hba_chain; p->next; p = p->next);
next              974 drivers/scsi/dpt_i2o.c 		p->next = pHba;
next              978 drivers/scsi/dpt_i2o.c 	pHba->next = NULL;
next             1038 drivers/scsi/dpt_i2o.c 	struct i2o_device* next;
next             1050 drivers/scsi/dpt_i2o.c 	for( p1 = hba_chain; p1; p2 = p1,p1=p1->next){
next             1053 drivers/scsi/dpt_i2o.c 				p2->next = p1->next;
next             1055 drivers/scsi/dpt_i2o.c 				hba_chain = p1->next;
next             1090 drivers/scsi/dpt_i2o.c 	for(d = pHba->devices; d ; d = next){
next             1091 drivers/scsi/dpt_i2o.c 		next = d->next;
next             1169 drivers/scsi/dpt_i2o.c 	wait_data->next = adpt_post_wait_queue;
next             1212 drivers/scsi/dpt_i2o.c 	for(p1 = adpt_post_wait_queue; p1; p2 = p1, p1 = p1->next) {
next             1218 drivers/scsi/dpt_i2o.c 				p2->next = p1->next;
next             1220 drivers/scsi/dpt_i2o.c 				adpt_post_wait_queue = p1->next;
next             1283 drivers/scsi/dpt_i2o.c 	for(p1 = adpt_post_wait_queue; p1; p1 = p1->next) {
next             1295 drivers/scsi/dpt_i2o.c 	for(p1 = adpt_post_wait_queue; p1; p1 = p1->next) {
next             1476 drivers/scsi/dpt_i2o.c 		d->next = NULL;
next             1486 drivers/scsi/dpt_i2o.c 	for(d = pHba->devices; d ; d = d->next) {
next             1512 drivers/scsi/dpt_i2o.c 	for(d = pHba->devices; d ; d = d->next) {
next             1581 drivers/scsi/dpt_i2o.c 	d->next=pHba->devices;
next             1607 drivers/scsi/dpt_i2o.c 	for (pHba = hba_chain; pHba; pHba = pHba->next) {
next             1640 drivers/scsi/dpt_i2o.c 	for (pHba = hba_chain; pHba; pHba = pHba->next) {
next             1963 drivers/scsi/dpt_i2o.c 	for (pHba = hba_chain; pHba; pHba = pHba->next) {
next             2508 drivers/scsi/dpt_i2o.c 	for (d = pHba->devices; d; d = d->next) {
next             2559 drivers/scsi/dpt_i2o.c 				d->next = NULL;
next             2632 drivers/scsi/dpt_i2o.c 	for (pI2o_dev = pHba->devices; pI2o_dev; pI2o_dev = pI2o_dev->next) {
next             3101 drivers/scsi/dpt_i2o.c 	for(pHba = hba_chain; pHba; pHba = pHba->next) {
next             3537 drivers/scsi/dpt_i2o.c 	adpt_hba	*pHba, *next;
next             3547 drivers/scsi/dpt_i2o.c 	for (pHba = hba_chain; pHba; pHba = pHba->next) {
next             3555 drivers/scsi/dpt_i2o.c 	for (pHba = hba_chain; pHba; pHba = next) {
next             3556 drivers/scsi/dpt_i2o.c 		next = pHba->next;
next             3564 drivers/scsi/dpt_i2o.c 	adpt_hba	*pHba, *next;
next             3566 drivers/scsi/dpt_i2o.c 	for (pHba = hba_chain; pHba; pHba = next) {
next             3567 drivers/scsi/dpt_i2o.c 		next = pHba->next;
next              203 drivers/scsi/dpti.h 	struct _adpt_hba *next;
next             1156 drivers/scsi/esas2r/esas2r.h 		sgl = a->free_sg_list_head.next;
next             1407 drivers/scsi/esas2r/esas2r.h 	struct list_head *element, *next;
next             1409 drivers/scsi/esas2r/esas2r.h 	list_for_each_safe(element, next, comp_list) {
next              437 drivers/scsi/esas2r/esas2r_init.c 	struct esas2r_mem_desc *memdesc, *next;
next              512 drivers/scsi/esas2r/esas2r_init.c 	list_for_each_entry_safe(memdesc, next, &a->free_sg_list_head,
next              518 drivers/scsi/esas2r/esas2r_init.c 	list_for_each_entry_safe(memdesc, next, &a->vrq_mds_head, next_desc) {
next              342 drivers/scsi/esas2r/esas2r_int.c 		struct list_head *element, *next;
next              346 drivers/scsi/esas2r/esas2r_int.c 		list_for_each_safe(element, next, &a->defer_list) {
next              771 drivers/scsi/esas2r/esas2r_io.c 	struct list_head *next, *element;
next              782 drivers/scsi/esas2r/esas2r_io.c 	list_for_each_safe(element, next, &a->defer_list) {
next              815 drivers/scsi/esas2r/esas2r_io.c 		list_for_each_safe(element, next, &a->active_list) {
next              908 drivers/scsi/esas2r/esas2r_main.c 	struct list_head *element, *next;
next              910 drivers/scsi/esas2r/esas2r_main.c 	list_for_each_safe(element, next, queue) {
next              883 drivers/scsi/esp_scsi.c 		ret = list_entry(head->next, struct esp_cmd_entry, list);
next              272 drivers/scsi/fcoe/fcoe_ctlr.c 	struct fcoe_fcf *next;
next              275 drivers/scsi/fcoe/fcoe_ctlr.c 	list_for_each_entry_safe(fcf, next, &fip->fcfs, list) {
next              820 drivers/scsi/fcoe/fcoe_ctlr.c 	struct fcoe_fcf *next;
next              831 drivers/scsi/fcoe/fcoe_ctlr.c 	list_for_each_entry_safe(fcf, next, &fip->fcfs, list) {
next              867 drivers/scsi/fcoe/fcoe_ctlr.c 	list_for_each_entry_safe(fcf, next, &del_list, list) {
next              873 drivers/scsi/fcoe/fcoe_sysfs.c 	struct fcoe_fcf_device *fcf, *next;
next              876 drivers/scsi/fcoe/fcoe_sysfs.c 	list_for_each_entry_safe(fcf, next,
next              228 drivers/scsi/fnic/fnic_fcs.c 	struct fnic_event *next = NULL;
next              237 drivers/scsi/fnic/fnic_fcs.c 	list_for_each_entry_safe(fevt, next, &fnic->evlist, list) {
next              248 drivers/scsi/fnic/fnic_fcs.c 	struct fnic_event *next = NULL;
next              257 drivers/scsi/fnic/fnic_fcs.c 	list_for_each_entry_safe(fevt, next, &fnic->evlist, list) {
next             1315 drivers/scsi/fnic/fnic_fcs.c 	struct fcoe_vlan *next;
next             1324 drivers/scsi/fnic/fnic_fcs.c 		list_for_each_entry_safe(vlan, next, &fnic->vlans, list) {
next               48 drivers/scsi/fnic/vnic_rq.c 				buf->next = rq->bufs[0];
next               51 drivers/scsi/fnic/vnic_rq.c 				buf->next = rq->bufs[i + 1];
next               53 drivers/scsi/fnic/vnic_rq.c 				buf->next = buf + 1;
next              179 drivers/scsi/fnic/vnic_rq.c 		buf = rq->to_clean = buf->next;
next               83 drivers/scsi/fnic/vnic_rq.h 	struct vnic_rq_buf *next;
next              143 drivers/scsi/fnic/vnic_rq.h 	buf = buf->next;
next              199 drivers/scsi/fnic/vnic_rq.h 		rq->to_clean = buf->next;
next               68 drivers/scsi/fnic/vnic_wq.c 				buf->next = wq->bufs[0];
next               71 drivers/scsi/fnic/vnic_wq.c 				buf->next = wq->bufs[i + 1];
next               73 drivers/scsi/fnic/vnic_wq.c 				buf->next = buf + 1;
next              235 drivers/scsi/fnic/vnic_wq.c 		buf = wq->to_clean = buf->next;
next               70 drivers/scsi/fnic/vnic_wq.h 	struct vnic_wq_buf *next;
next              126 drivers/scsi/fnic/vnic_wq.h 	buf = buf->next;
next              156 drivers/scsi/fnic/vnic_wq.h 		wq->to_clean = buf->next;
next             9638 drivers/scsi/hpsa.c 	struct hpsa_sas_phy *next;
next             9640 drivers/scsi/hpsa.c 	list_for_each_entry_safe(hpsa_sas_phy, next,
next             9665 drivers/scsi/hpsa.c 	struct hpsa_sas_port *next;
next             9670 drivers/scsi/hpsa.c 	list_for_each_entry_safe(hpsa_sas_port, next,
next              704 drivers/scsi/hptiop.c 		hba->req_list = ret->next;
next              712 drivers/scsi/hptiop.c 	req->next = hba->req_list;
next             1226 drivers/scsi/hptiop.c 	hba->u.mvfrey.internal_req.next = NULL;
next             1454 drivers/scsi/hptiop.c 		hba->reqs[i].next = NULL;
next              247 drivers/scsi/hptiop.h 	struct hptiop_request *next;
next             1268 drivers/scsi/ibmvscsi/ibmvfc.c 	evt = list_entry(vhost->free.next, struct ibmvfc_event, queue);
next              447 drivers/scsi/ibmvscsi/ibmvscsi.c 	pool->next = 0;
next              559 drivers/scsi/ibmvscsi/ibmvscsi.c 	int offset = pool->next;
next              564 drivers/scsi/ibmvscsi/ibmvscsi.c 			pool->next = offset;
next               72 drivers/scsi/ibmvscsi/ibmvscsi.h 	int next;
next              666 drivers/scsi/initio.c 		if ((host->first_avail = scb->next) == NULL)
next              668 drivers/scsi/initio.c 		scb->next = NULL;
next              693 drivers/scsi/initio.c 	cmnd->next = NULL;
next              695 drivers/scsi/initio.c 		host->last_avail->next = cmnd;
next              712 drivers/scsi/initio.c 	scbp->next = NULL;
next              714 drivers/scsi/initio.c 		host->last_pending->next = scbp;
next              730 drivers/scsi/initio.c 	if ((scbp->next = host->first_pending) != NULL) {
next              755 drivers/scsi/initio.c 				first = first->next;
next              760 drivers/scsi/initio.c 		first = first->next;
next              777 drivers/scsi/initio.c 				if ((host->first_pending = tmp->next) == NULL)
next              780 drivers/scsi/initio.c 				prev->next = tmp->next;
next              784 drivers/scsi/initio.c 			tmp->next = NULL;
next              788 drivers/scsi/initio.c 		tmp = tmp->next;
next              803 drivers/scsi/initio.c 	scbp->next = NULL;
next              805 drivers/scsi/initio.c 		host->last_busy->next = scbp;
next              820 drivers/scsi/initio.c 		if ((host->first_busy = tmp->next) == NULL)
next              822 drivers/scsi/initio.c 		tmp->next = NULL;
next              847 drivers/scsi/initio.c 				if ((host->first_busy = tmp->next) == NULL)
next              850 drivers/scsi/initio.c 				prev->next = tmp->next;
next              854 drivers/scsi/initio.c 			tmp->next = NULL;
next              862 drivers/scsi/initio.c 		tmp = tmp->next;
next              880 drivers/scsi/initio.c 		tmp = tmp->next;
next              895 drivers/scsi/initio.c 	scbp->next = NULL;
next              897 drivers/scsi/initio.c 		host->last_done->next = scbp;
next              910 drivers/scsi/initio.c 		if ((host->first_done = tmp->next) == NULL)
next              912 drivers/scsi/initio.c 		tmp->next = NULL;
next              947 drivers/scsi/initio.c 				if ((host->first_pending = tmp->next) == NULL)
next              950 drivers/scsi/initio.c 				prev->next = tmp->next;
next              962 drivers/scsi/initio.c 		tmp = tmp->next;
next              977 drivers/scsi/initio.c 					if ((host->first_busy = tmp->next) == NULL)
next              980 drivers/scsi/initio.c 					prev->next = tmp->next;
next              984 drivers/scsi/initio.c 				tmp->next = NULL;
next              996 drivers/scsi/initio.c 		tmp = tmp->next;
next             1250 drivers/scsi/initio.c 	int next;
next             1252 drivers/scsi/initio.c 	next = host->active->next_state;
next             1254 drivers/scsi/initio.c 		switch (next) {
next             1256 drivers/scsi/initio.c 			next = initio_state_1(host);
next             1259 drivers/scsi/initio.c 			next = initio_state_2(host);
next             1262 drivers/scsi/initio.c 			next = initio_state_3(host);
next             1265 drivers/scsi/initio.c 			next = initio_state_4(host);
next             1268 drivers/scsi/initio.c 			next = initio_state_5(host);
next             1271 drivers/scsi/initio.c 			next = initio_state_6(host);
next             1274 drivers/scsi/initio.c 			next = initio_state_7(host);
next             1281 drivers/scsi/initio.c 		if (next <= 0)
next             1282 drivers/scsi/initio.c 			return next;
next             2378 drivers/scsi/initio.c 				if ((host->first_busy = tmp->next) == NULL)
next             2381 drivers/scsi/initio.c 				prev->next = tmp->next;
next             2392 drivers/scsi/initio.c 		tmp = tmp->next;
next             2878 drivers/scsi/initio.c 			prev->next = tmp;
next             2881 drivers/scsi/initio.c 	prev->next = NULL;
next              350 drivers/scsi/initio.h 	struct scsi_ctrl_blk *next;
next              701 drivers/scsi/ipr.c 		ipr_cmd = list_entry(hrrq->hrrq_free_q.next,
next             1461 drivers/scsi/ipr.c 		res = list_entry(ioa_cfg->free_res_q.next,
next             7580 drivers/scsi/ipr.c 	ipr_cmd->u.res = list_entry(ioa_cfg->used_res_q.next,
next             7654 drivers/scsi/ipr.c 		ipr_cmd->u.res = list_entry(ioa_cfg->used_res_q.next,
next             7838 drivers/scsi/ipr.c 			res = list_entry(ioa_cfg->free_res_q.next,
next             8563 drivers/scsi/ipr.c 	hostrcb = list_entry(ioa_cfg->hostrcb_free_q.next,
next              801 drivers/scsi/ips.c 		item = item->next;
next              866 drivers/scsi/ips.c 		item = item->next;
next             1117 drivers/scsi/ips.c 		scratch->next = NULL;
next             2980 drivers/scsi/ips.c 	item->next = NULL;
next             2983 drivers/scsi/ips.c 		queue->tail->next = item;
next             3017 drivers/scsi/ips.c 	queue->head = item->next;
next             3018 drivers/scsi/ips.c 	item->next = NULL;
next             3055 drivers/scsi/ips.c 	while ((p) && (item != p->next))
next             3056 drivers/scsi/ips.c 		p = p->next;
next             3060 drivers/scsi/ips.c 		p->next = item->next;
next             3062 drivers/scsi/ips.c 		if (!item->next)
next             3065 drivers/scsi/ips.c 		item->next = NULL;
next              978 drivers/scsi/ips.h 	struct ips_copp_wait_item *next;
next             2155 drivers/scsi/isci/request.c 						  enum sci_base_request_states next)
next             2163 drivers/scsi/isci/request.c 		sci_change_state(&ireq->sm, next);
next              104 drivers/scsi/isci/task.c 	     task = list_entry(task->list.next, struct sas_task, list))
next               81 drivers/scsi/libfc/fc_disc.c 	struct fc_disc_port *dp, *next;
next              148 drivers/scsi/libfc/fc_disc.c 	list_for_each_entry_safe(dp, next, &disc_ports, peers) {
next             1923 drivers/scsi/libfc/fc_exch.c 	struct fc_exch *next;
next             1927 drivers/scsi/libfc/fc_exch.c 	list_for_each_entry_safe(ep, next, &pool->ex_list, ex_list) {
next             2543 drivers/scsi/libfc/fc_exch.c 	struct fc_exch_mgr_anchor *ema, *next;
next             2546 drivers/scsi/libfc/fc_exch.c 	list_for_each_entry_safe(ema, next, &lport->ema_list, ema_list)
next             1496 drivers/scsi/libiscsi.c 		conn->task = list_entry(conn->mgmtqueue.next,
next             1517 drivers/scsi/libiscsi.c 		conn->task = list_entry(conn->cmdqueue.next, struct iscsi_task,
next             1560 drivers/scsi/libiscsi.c 		task = list_entry(conn->requeue.next, struct iscsi_task,
next              819 drivers/scsi/libsas/sas_ata.c 				list_del_init(sata_q.next);
next               67 drivers/scsi/libsas/sas_discover.c 	phy = container_of(port->phy_list.next, struct asd_sas_phy, port_phy_el);
next              110 drivers/scsi/libsas/sas_event.c 		sas_phy = container_of(port->phy_list.next, struct asd_sas_phy,
next              700 drivers/scsi/libsas/sas_scsi_host.c 		ssp = list_entry(ha->eh_dev_q.next, typeof(*ssp), eh_list_node);
next              613 drivers/scsi/lpfc/lpfc_bsg.c 	prsp = (struct lpfc_dmabuf *)pcmd->list.next;
next              810 drivers/scsi/lpfc/lpfc_bsg.c 		ed = list_entry(evt->events_to_get.next, typeof(*ed), node);
next              817 drivers/scsi/lpfc/lpfc_bsg.c 		ed = list_entry(evt->events_to_see.next, typeof(*ed), node);
next              883 drivers/scsi/lpfc/lpfc_bsg.c 	struct list_head head, *curr, *next;
next              893 drivers/scsi/lpfc/lpfc_bsg.c 	list_for_each_safe(curr, next, &head) {
next             2960 drivers/scsi/lpfc/lpfc_bsg.c 	struct list_head head, *curr, *next;
next             2998 drivers/scsi/lpfc/lpfc_bsg.c 	list_for_each_safe(curr, next, &head) {
next             3042 drivers/scsi/lpfc/lpfc_bsg.c 			dmp = list_entry(next, struct lpfc_dmabuf, list);
next             3056 drivers/scsi/lpfc/lpfc_bsg.c 			dmp = list_entry(next, struct lpfc_dmabuf, list);
next             5622 drivers/scsi/lpfc/lpfc_bsg.c 	struct lpfc_dmabuf *dmabuf, *next;
next             5661 drivers/scsi/lpfc/lpfc_bsg.c 	list_for_each_entry_safe(dmabuf, next,
next             2073 drivers/scsi/lpfc/lpfc_els.c 				   cmdiocb->context2)->list.next,
next             2192 drivers/scsi/lpfc/lpfc_hbadisc.c 		} else if (fcf_pri->list.next == &phba->fcf.fcf_pri_list
next             11092 drivers/scsi/lpfc/lpfc_init.c 	struct lpfc_queue *eq, *next;
next             11107 drivers/scsi/lpfc/lpfc_init.c 	list_for_each_entry_safe(eq, next, &eqlist, _poll_list) {
next             11118 drivers/scsi/lpfc/lpfc_init.c 	struct lpfc_queue *eq, *next;
next             11130 drivers/scsi/lpfc/lpfc_init.c 	list_for_each_entry_safe(eq, next, &phba->poll_list, _poll_list) {
next             12531 drivers/scsi/lpfc/lpfc_init.c 	struct lpfc_dmabuf *dmabuf, *next;
next             12602 drivers/scsi/lpfc/lpfc_init.c 	list_for_each_entry_safe(dmabuf, next, &dma_buffer_list, list) {
next               32 drivers/scsi/lpfc/lpfc_scsi.h 		entry = list_entry((list)->next, type, member);	\
next               39 drivers/scsi/lpfc/lpfc_scsi.h 	list_entry((list)->next, type, member)
next             6215 drivers/scsi/lpfc/lpfc_sli.c 	struct lpfc_dmabuf *dmabuf, *next;
next             6218 drivers/scsi/lpfc/lpfc_sli.c 		list_for_each_entry_safe(dmabuf, next,
next             10964 drivers/scsi/lpfc/lpfc_sli.c 			slp->next, slp->prev, pring->postbufq_cnt);
next             11008 drivers/scsi/lpfc/lpfc_sli.c 			slp->next, slp->prev, pring->postbufq_cnt);
next             14396 drivers/scsi/lpfc/lpfc_sli.c 	struct lpfc_queue *eq, *next;
next             14398 drivers/scsi/lpfc/lpfc_sli.c 	list_for_each_entry_safe(eq, next, &phba->poll_list, _poll_list)
next              431 drivers/scsi/megaraid.c 		scb = list_entry(head->next, scb_t, list);
next              433 drivers/scsi/megaraid.c 		list_del_init(head->next);
next             1068 drivers/scsi/megaraid.c 	struct list_head *pos, *next;
next             1071 drivers/scsi/megaraid.c 	list_for_each_safe(pos, next, &adapter->pending_list) {
next             1934 drivers/scsi/megaraid.c 	struct list_head	*pos, *next;
next             1945 drivers/scsi/megaraid.c 	list_for_each_safe(pos, next, &adapter->pending_list) {
next             1286 drivers/scsi/megaraid/megaraid_mbox.c 	scb = list_entry(head->next, scb_t, list);
next             1886 drivers/scsi/megaraid/megaraid_mbox.c 		scb = list_entry(adapter->pend_list.next, scb_t, list);
next             3584 drivers/scsi/megaraid/megaraid_mbox.c 	scb = list_entry(head->next, scb_t, list);
next              612 drivers/scsi/megaraid/megaraid_mm.c 	kioc = list_entry(head->next, uioc_t, list);
next              314 drivers/scsi/megaraid/megaraid_sas_base.c 		cmd = list_entry((&instance->cmd_pool)->next,
next              358 drivers/scsi/megaraid/megaraid_sas_base.c 	list_add(&cmd->list, (&instance->cmd_pool)->next);
next             2713 drivers/scsi/megaraid/megaraid_sas_base.c 			reset_cmd = list_entry((&clist_local)->next,
next             3699 drivers/scsi/megaraid/megaraid_sas_base.c 		cmd = list_entry((&clist_local)->next,
next              150 drivers/scsi/mesh.c 	struct	mesh_state *next;
next              560 drivers/scsi/mesh.c 	struct scsi_cmnd *cmd, *prev, *next;
next              577 drivers/scsi/mesh.c 		next = (struct scsi_cmnd *) cmd->host_scribble;
next              579 drivers/scsi/mesh.c 			ms->request_q = next;
next              581 drivers/scsi/mesh.c 			prev->host_scribble = (void *) next;
next              582 drivers/scsi/mesh.c 		if (next == NULL)
next             1658 drivers/scsi/mpt3sas/mpt3sas_base.c 	struct adapter_reply_queue *reply_q, *next;
next             1663 drivers/scsi/mpt3sas/mpt3sas_base.c 	list_for_each_entry_safe(reply_q, next, &ioc->reply_queue_list, list) {
next             2816 drivers/scsi/mpt3sas/mpt3sas_base.c 	struct adapter_reply_queue *reply_q, *next;
next             2821 drivers/scsi/mpt3sas/mpt3sas_base.c 	list_for_each_entry_safe(reply_q, next, &ioc->reply_queue_list, list) {
next             3519 drivers/scsi/mpt3sas/mpt3sas_base.c 	request = list_entry(ioc->internal_free_list.next,
next             3572 drivers/scsi/mpt3sas/mpt3sas_base.c 	request = list_entry(ioc->hpr_free_list.next,
next             4096 drivers/scsi/mpt3sas/mpt3sas_scsih.c 		delayed_event_ack = list_entry(ioc->delayed_event_ack_list.next,
next             4106 drivers/scsi/mpt3sas/mpt3sas_scsih.c 		delayed_sc = list_entry(ioc->delayed_sc_list.next,
next             4134 drivers/scsi/mpt3sas/mpt3sas_scsih.c 		delayed_tr = list_entry(ioc->delayed_tr_volume_list.next,
next             4144 drivers/scsi/mpt3sas/mpt3sas_scsih.c 		delayed_tr = list_entry(ioc->delayed_tr_list.next,
next             9627 drivers/scsi/mpt3sas/mpt3sas_scsih.c 	struct _sas_port *mpt3sas_port, *next;
next             9631 drivers/scsi/mpt3sas/mpt3sas_scsih.c 	list_for_each_entry_safe(mpt3sas_port, next,
next             9741 drivers/scsi/mpt3sas/mpt3sas_scsih.c 	struct _raid_device *raid_device, *next;
next             9771 drivers/scsi/mpt3sas/mpt3sas_scsih.c 	list_for_each_entry_safe(raid_device, next, &ioc->raid_device_list,
next              546 drivers/scsi/mpt3sas/mpt3sas_transport.c 	struct _sas_port *mpt3sas_port, *next;
next              552 drivers/scsi/mpt3sas/mpt3sas_transport.c 	list_for_each_entry_safe(mpt3sas_port, next, &sas_node->sas_port_list,
next              607 drivers/scsi/mpt3sas/mpt3sas_transport.c 	struct _sas_phy *mpt3sas_phy, *next;
next              743 drivers/scsi/mpt3sas/mpt3sas_transport.c 	list_for_each_entry_safe(mpt3sas_phy, next, &mpt3sas_port->phy_list,
next              766 drivers/scsi/mpt3sas/mpt3sas_transport.c 	struct _sas_port *mpt3sas_port, *next;
next              778 drivers/scsi/mpt3sas/mpt3sas_transport.c 	list_for_each_entry_safe(mpt3sas_port, next, &sas_node->sas_port_list,
next               70 drivers/scsi/mvsas/mv_sas.c 			phy =  container_of(sha->sas_port[i]->phy_list.next,
next              321 drivers/scsi/mvumi.c 		cmd = list_entry((&mhba->cmd_pool)->next,
next              154 drivers/scsi/ncr53c8xx.c 		struct list_head *elem = head->next;
next              198 drivers/scsi/ncr53c8xx.c 	struct m_link *next;
next              202 drivers/scsi/ncr53c8xx.c 	struct m_vtob *next;
next              218 drivers/scsi/ncr53c8xx.c 	struct m_pool *next;
next              239 drivers/scsi/ncr53c8xx.c 	while (!h[j].next) {
next              241 drivers/scsi/ncr53c8xx.c 			h[j].next = (m_link_s *)mp->getp(mp);
next              242 drivers/scsi/ncr53c8xx.c 			if (h[j].next)
next              243 drivers/scsi/ncr53c8xx.c 				h[j].next->next = NULL;
next              249 drivers/scsi/ncr53c8xx.c 	a = (m_addr_t) h[j].next;
next              251 drivers/scsi/ncr53c8xx.c 		h[j].next = h[j].next->next;
next              255 drivers/scsi/ncr53c8xx.c 			h[j].next = (m_link_s *) (a+s);
next              256 drivers/scsi/ncr53c8xx.c 			h[j].next->next = NULL;
next              296 drivers/scsi/ncr53c8xx.c 		while (q->next && q->next != (m_link_s *) b) {
next              297 drivers/scsi/ncr53c8xx.c 			q = q->next;
next              299 drivers/scsi/ncr53c8xx.c 		if (!q->next) {
next              300 drivers/scsi/ncr53c8xx.c 			((m_link_s *) a)->next = h[i].next;
next              301 drivers/scsi/ncr53c8xx.c 			h[i].next = (m_link_s *) a;
next              304 drivers/scsi/ncr53c8xx.c 		q->next = q->next->next;
next              386 drivers/scsi/ncr53c8xx.c 			vbp->next = mp->vtob[hc];
next              404 drivers/scsi/ncr53c8xx.c 		vbpp = &(*vbpp)->next;
next              407 drivers/scsi/ncr53c8xx.c 		*vbpp = (*vbpp)->next;
next              418 drivers/scsi/ncr53c8xx.c 	for (mp = mp0.next; mp && mp->bush != bush; mp = mp->next);
next              431 drivers/scsi/ncr53c8xx.c 		mp->next = mp0.next;
next              432 drivers/scsi/ncr53c8xx.c 		mp0.next = mp;
next              439 drivers/scsi/ncr53c8xx.c 	struct m_pool **pp = &mp0.next;
next              442 drivers/scsi/ncr53c8xx.c 		pp = &(*pp)->next;
next              444 drivers/scsi/ncr53c8xx.c 		*pp = (*pp)->next;
next              495 drivers/scsi/ncr53c8xx.c 			vp = vp->next;
next              370 drivers/scsi/pmcraid.c 		cmd = list_entry(pinstance->free_cmd_pool.next,
next             1611 drivers/scsi/pmcraid.c 		res = list_entry(pinstance->free_res_q.next,
next             5545 drivers/scsi/pmcraid.c 			res = list_entry(pinstance->free_res_q.next,
next              208 drivers/scsi/qedi/qedi.h 	u16             next;
next              523 drivers/scsi/qedi/qedi_main.c 			    u16 start_id, u16 next)
next              527 drivers/scsi/qedi/qedi_main.c 	id_tbl->next = next;
next              564 drivers/scsi/qedi/qedi_main.c 	id = find_next_zero_bit(id_tbl->table, id_tbl->max, id_tbl->next);
next              567 drivers/scsi/qedi/qedi_main.c 		if (id_tbl->next != 0) {
next              568 drivers/scsi/qedi/qedi_main.c 			id = find_first_zero_bit(id_tbl->table, id_tbl->next);
next              569 drivers/scsi/qedi/qedi_main.c 			if (id >= id_tbl->next)
next              576 drivers/scsi/qedi/qedi_main.c 		id_tbl->next = (id + 1) & (id_tbl->max - 1);
next             1252 drivers/scsi/qla1280.c 		sp = list_entry(done_q->next, struct srb, list);
next             1012 drivers/scsi/qla1280.h 	struct scsi_qla_host *next;
next             3583 drivers/scsi/qla2xxx/qla_os.c 		BUG_ON(base_vha->list.next == &ha->vp_list);
next             1281 drivers/scsi/qla2xxx/qla_sup.c 		goto next;
next             1291 drivers/scsi/qla2xxx/qla_sup.c next:
next             6295 drivers/scsi/qla2xxx/qla_target.c 		    tgt->sess_works_list.next, typeof(*prm),
next              158 drivers/scsi/qlogicfas.c 		priv->next = cards;
next              216 drivers/scsi/qlogicfas.c 	for (priv = cards; priv != NULL; priv = priv->next)
next               87 drivers/scsi/qlogicfas408.h 	struct qlogicfas408_priv *next; /* next private struct */
next              690 drivers/scsi/qlogicpti.c 		while(qlink->next)
next              691 drivers/scsi/qlogicpti.c 			qlink = qlink->next;
next              692 drivers/scsi/qlogicpti.c 		qlink->next = qpti;
next              696 drivers/scsi/qlogicpti.c 	qpti->next = NULL;
next              704 drivers/scsi/qlogicpti.c 		qptichain = qpti->next;
next              707 drivers/scsi/qlogicpti.c 		while(qlink->next != qpti)
next              708 drivers/scsi/qlogicpti.c 			qlink = qlink->next;
next              709 drivers/scsi/qlogicpti.c 		qlink->next = qpti->next;
next              711 drivers/scsi/qlogicpti.c 	qpti->next = NULL;
next             1214 drivers/scsi/qlogicpti.c 			struct scsi_cmnd *next;
next             1216 drivers/scsi/qlogicpti.c 			next = (struct scsi_cmnd *) dq->host_scribble;
next             1218 drivers/scsi/qlogicpti.c 			dq = next;
next              359 drivers/scsi/qlogicpti.h 	struct qlogicpti         *next;
next              505 drivers/scsi/qlogicpti.h         for((qp) = qptichain; (qp); (qp) = (qp)->next)
next               98 drivers/scsi/raid_class.c 	struct raid_component *rc, *next;
next              101 drivers/scsi/raid_class.c 	list_for_each_entry_safe(rc, next, &rd->component_list, node) {
next              568 drivers/scsi/scsi.c 	struct scsi_device *next = NULL;
next              572 drivers/scsi/scsi.c 	while (list->next != &shost->__devices) {
next              573 drivers/scsi/scsi.c 		next = list_entry(list->next, struct scsi_device, siblings);
next              575 drivers/scsi/scsi.c 		if (!scsi_device_get(next))
next              577 drivers/scsi/scsi.c 		next = NULL;
next              578 drivers/scsi/scsi.c 		list = list->next;
next              584 drivers/scsi/scsi.c 	return next;
next             2909 drivers/scsi/scsi_debug.c 	unsigned long next;
next             2915 drivers/scsi/scsi_debug.c 		next = find_next_zero_bit(map_storep, map_size, index);
next             2917 drivers/scsi/scsi_debug.c 		next = find_next_bit(map_storep, map_size, index);
next             2919 drivers/scsi/scsi_debug.c 	end = min_t(sector_t, sdebug_store_sectors,  map_index_to_lba(next));
next              523 drivers/scsi/scsi_devinfo.c 	char *vendor, *model, *strflags, *next;
next              527 drivers/scsi/scsi_devinfo.c 	next = dev_list;
next              528 drivers/scsi/scsi_devinfo.c 	if (next && next[0] == '"') {
next              532 drivers/scsi/scsi_devinfo.c 		next++;
next              542 drivers/scsi/scsi_devinfo.c 	for (vendor = strsep(&next, ":"); vendor && (vendor[0] != '\0')
next              543 drivers/scsi/scsi_devinfo.c 	     && (res == 0); vendor = strsep(&next, ":")) {
next              545 drivers/scsi/scsi_devinfo.c 		model = strsep(&next, ":");
next              547 drivers/scsi/scsi_devinfo.c 			strflags = strsep(&next, next_check);
next              632 drivers/scsi/scsi_devinfo.c 	if (devinfo_table->scsi_dev_info_list.next == dl->bottom &&
next              669 drivers/scsi/scsi_devinfo.c 	dl->bottom = dl->bottom->next;
next              671 drivers/scsi/scsi_devinfo.c 		dl->top = dl->top->next;
next              679 drivers/scsi/scsi_devinfo.c 		dl->bottom = devinfo_table->scsi_dev_info_list.next;
next              692 drivers/scsi/scsi_devinfo.c 	.next	= devinfo_seq_next,
next             1219 drivers/scsi/scsi_error.c 	struct scsi_cmnd *scmd, *next;
next             1227 drivers/scsi/scsi_error.c 	list_for_each_entry_safe(scmd, next, work_q, eh_entry) {
next             1330 drivers/scsi/scsi_error.c 	struct scsi_cmnd *scmd, *next;
next             1335 drivers/scsi/scsi_error.c 		scmd = list_entry(cmd_list->next, struct scsi_cmnd, eh_entry);
next             1355 drivers/scsi/scsi_error.c 		list_for_each_entry_safe(scmd, next, cmd_list, eh_entry)
next             1406 drivers/scsi/scsi_error.c 	struct scsi_cmnd *scmd, *stu_scmd, *next;
next             1436 drivers/scsi/scsi_error.c 				list_for_each_entry_safe(scmd, next,
next             1471 drivers/scsi/scsi_error.c 	struct scsi_cmnd *scmd, *bdr_scmd, *next;
next             1501 drivers/scsi/scsi_error.c 				list_for_each_entry_safe(scmd, next,
next             1538 drivers/scsi/scsi_error.c 		struct scsi_cmnd *next, *scmd;
next             1553 drivers/scsi/scsi_error.c 		scmd = list_entry(tmp_list.next, struct scsi_cmnd, eh_entry);
next             1567 drivers/scsi/scsi_error.c 		list_for_each_entry_safe(scmd, next, &tmp_list, eh_entry) {
next             1594 drivers/scsi/scsi_error.c 	struct scsi_cmnd *scmd, *chan_scmd, *next;
next             1636 drivers/scsi/scsi_error.c 			list_for_each_entry_safe(scmd, next, work_q, eh_entry) {
next             1666 drivers/scsi/scsi_error.c 	struct scsi_cmnd *scmd, *next;
next             1671 drivers/scsi/scsi_error.c 		scmd = list_entry(work_q->next,
next             1683 drivers/scsi/scsi_error.c 			list_for_each_entry_safe(scmd, next, work_q, eh_entry) {
next             1704 drivers/scsi/scsi_error.c 	struct scsi_cmnd *scmd, *next;
next             1707 drivers/scsi/scsi_error.c 	list_for_each_entry_safe(scmd, next, work_q, eh_entry) {
next             2087 drivers/scsi/scsi_error.c 	struct scsi_cmnd *scmd, *next;
next             2089 drivers/scsi/scsi_error.c 	list_for_each_entry_safe(scmd, next, done_q, eh_entry) {
next              468 drivers/scsi/scsi_lib.c 		sdev = list_entry(starved_list.next,
next              377 drivers/scsi/scsi_proc.c 	struct device *next = bus_find_next_device(&scsi_bus_type, start);
next              380 drivers/scsi/scsi_proc.c 	return next;
next              418 drivers/scsi/scsi_proc.c 	.next	= scsi_seq_next,
next              165 drivers/scsi/scsi_scan.c 		struct async_scan_data *next = list_entry(scanning_hosts.next,
next              167 drivers/scsi/scsi_scan.c 		complete(&next->prev_finished);
next             1796 drivers/scsi/scsi_scan.c 		struct async_scan_data *next = list_entry(scanning_hosts.next,
next             1798 drivers/scsi/scsi_scan.c 		complete(&next->prev_finished);
next             1594 drivers/scsi/scsi_transport_fc.c 	pos = list_entry((head)->next, typeof(*pos), member)
next             1039 drivers/scsi/scsi_transport_sas.c 		struct list_head *ent = port->phy_list.next;
next             2332 drivers/scsi/sg.c 	.next  = dev_seq_next,
next             2340 drivers/scsi/sg.c 	.next  = dev_seq_next,
next             2348 drivers/scsi/sg.c 	.next  = dev_seq_next,
next               53 drivers/scsi/sim710.c 	char *pos = str, *next;
next               56 drivers/scsi/sim710.c 	while(pos != NULL && (next = strchr(pos, ':')) != NULL) {
next               57 drivers/scsi/sim710.c 		int val = (int)simple_strtoul(++next, NULL, 0);
next             1747 drivers/scsi/smartpqi/smartpqi_init.c 	struct pqi_scsi_dev *next;
next             1800 drivers/scsi/smartpqi/smartpqi_init.c 	list_for_each_entry_safe(device, next, &ctrl_info->scsi_device_list,
next             1828 drivers/scsi/smartpqi/smartpqi_init.c 	list_for_each_entry_safe(device, next, &delete_list,
next             1857 drivers/scsi/smartpqi/smartpqi_init.c 	list_for_each_entry_safe(device, next, &add_list, add_list_entry) {
next             1957 drivers/scsi/smartpqi/smartpqi_init.c 	struct pqi_scsi_dev *next;
next             2135 drivers/scsi/smartpqi/smartpqi_init.c 	list_for_each_entry_safe(device, next, &new_device_list_head,
next             3882 drivers/scsi/smartpqi/smartpqi_init.c 	struct pqi_io_request *next;
next             3903 drivers/scsi/smartpqi/smartpqi_init.c 	list_for_each_entry_safe(io_request, next,
next             5459 drivers/scsi/smartpqi/smartpqi_init.c 	struct pqi_io_request *next;
next             5470 drivers/scsi/smartpqi/smartpqi_init.c 			list_for_each_entry_safe(io_request, next,
next             5499 drivers/scsi/smartpqi/smartpqi_init.c 	struct pqi_io_request *next;
next             5509 drivers/scsi/smartpqi/smartpqi_init.c 			list_for_each_entry_safe(io_request, next,
next              160 drivers/scsi/smartpqi/smartpqi_sas_transport.c 	struct pqi_sas_phy *next;
next              162 drivers/scsi/smartpqi/smartpqi_sas_transport.c 	list_for_each_entry_safe(pqi_sas_phy, next,
next              187 drivers/scsi/smartpqi/smartpqi_sas_transport.c 	struct pqi_sas_port *next;
next              192 drivers/scsi/smartpqi/smartpqi_sas_transport.c 	list_for_each_entry_safe(pqi_sas_port, next,
next              444 drivers/scsi/snic/snic_debugfs.c 	.next	= snic_trc_seq_next,
next              115 drivers/scsi/snic/snic_io.c 	SNIC_BUG_ON(rqi->list.next == NULL); /* if not added to spl_cmd_list */
next               65 drivers/scsi/snic/vnic_wq.c 				buf->next = wq->bufs[0];
next               68 drivers/scsi/snic/vnic_wq.c 				buf->next = wq->bufs[i + 1];
next               70 drivers/scsi/snic/vnic_wq.c 				buf->next = buf + 1;
next              226 drivers/scsi/snic/vnic_wq.c 		buf = wq->to_clean = buf->next;
next               51 drivers/scsi/snic/vnic_wq.h 	struct vnic_wq_buf *next;
next              113 drivers/scsi/snic/vnic_wq.h 	buf = buf->next;
next              143 drivers/scsi/snic/vnic_wq.h 		wq->to_clean = buf->next;
next             1121 drivers/scsi/sym53c8xx_2/sym_hipd.h 	struct sym_m_link *next;
next             1129 drivers/scsi/sym53c8xx_2/sym_hipd.h 	struct sym_m_vtob *next;
next             1160 drivers/scsi/sym53c8xx_2/sym_hipd.h 	struct sym_m_pool *next;
next               64 drivers/scsi/sym53c8xx_2/sym_malloc.c 	while (!h[j].next) {
next               66 drivers/scsi/sym53c8xx_2/sym_malloc.c 			h[j].next = (m_link_p) M_GET_MEM_CLUSTER();
next               67 drivers/scsi/sym53c8xx_2/sym_malloc.c 			if (h[j].next)
next               68 drivers/scsi/sym53c8xx_2/sym_malloc.c 				h[j].next->next = NULL;
next               74 drivers/scsi/sym53c8xx_2/sym_malloc.c 	a = h[j].next;
next               76 drivers/scsi/sym53c8xx_2/sym_malloc.c 		h[j].next = h[j].next->next;
next               80 drivers/scsi/sym53c8xx_2/sym_malloc.c 			h[j].next = (m_link_p) (a+s);
next               81 drivers/scsi/sym53c8xx_2/sym_malloc.c 			h[j].next->next = NULL;
next              120 drivers/scsi/sym53c8xx_2/sym_malloc.c 			((m_link_p) a)->next = h[i].next;
next              121 drivers/scsi/sym53c8xx_2/sym_malloc.c 			h[i].next = (m_link_p) a;
next              127 drivers/scsi/sym53c8xx_2/sym_malloc.c 		while (q->next && q->next != (m_link_p) b) {
next              128 drivers/scsi/sym53c8xx_2/sym_malloc.c 			q = q->next;
next              130 drivers/scsi/sym53c8xx_2/sym_malloc.c 		if (!q->next) {
next              131 drivers/scsi/sym53c8xx_2/sym_malloc.c 			((m_link_p) a)->next = h[i].next;
next              132 drivers/scsi/sym53c8xx_2/sym_malloc.c 			h[i].next = (m_link_p) a;
next              135 drivers/scsi/sym53c8xx_2/sym_malloc.c 		q->next = q->next->next;
next              222 drivers/scsi/sym53c8xx_2/sym_malloc.c 		vbp->next = mp->vtob[hc];
next              240 drivers/scsi/sym53c8xx_2/sym_malloc.c 		vbpp = &(*vbpp)->next;
next              243 drivers/scsi/sym53c8xx_2/sym_malloc.c 		*vbpp = (*vbpp)->next;
next              255 drivers/scsi/sym53c8xx_2/sym_malloc.c 	for (mp = mp0.next;
next              257 drivers/scsi/sym53c8xx_2/sym_malloc.c 			mp = mp->next);
next              271 drivers/scsi/sym53c8xx_2/sym_malloc.c 		mp->next = mp0.next;
next              272 drivers/scsi/sym53c8xx_2/sym_malloc.c 		mp0.next = mp;
next              282 drivers/scsi/sym53c8xx_2/sym_malloc.c 	m_pool_p *pp = &mp0.next;
next              285 drivers/scsi/sym53c8xx_2/sym_malloc.c 		pp = &(*pp)->next;
next              287 drivers/scsi/sym53c8xx_2/sym_malloc.c 		*pp = (*pp)->next;
next              358 drivers/scsi/sym53c8xx_2/sym_malloc.c 			vp = vp->next;
next              217 drivers/scsi/wd33c93.h     struct Scsi_Host *next;
next              421 drivers/sh/clk/core.c 	if (clk->node.next || clk->node.prev)
next               24 drivers/sh/intc/virq.c 	struct intc_virq_list *next;
next               28 drivers/sh/intc/virq.c 	for (entry = head; entry; entry = entry->next)
next               93 drivers/sh/intc/virq.c 		last = &entry->next;
next               67 drivers/soc/bcm/brcmstb/pm/pm-mips.c 	struct brcmstb_mem_transfer	*next;
next               74 drivers/soc/fsl/dpio/dpio-service.c 	d = list_entry(dpio_list.next, struct dpaa2_io, node);
next              250 drivers/soc/fsl/qbman/bman.c 	struct bman_pool *next;
next              457 drivers/soc/fsl/qbman/qman_test_stash.c 						hp_cpu->iterator->node.next,
next              500 drivers/soc/fsl/qbman/qman_test_stash.c 						hp_cpu->iterator->node.next,
next              266 drivers/spi/spi-bcm63xx.c 		t = list_entry(t->transfer_list.next, struct spi_transfer,
next              317 drivers/spi/spi-bcm63xx.c 		t = list_entry(t->transfer_list.next, struct spi_transfer,
next              214 drivers/spi/spi-mpc52xx-psc.c 		m = container_of(mps->queue.next, struct spi_message, queue);
next              192 drivers/spi/spi-mpc52xx.c 	ms->transfer = container_of(ms->message->transfers.next,
next              287 drivers/spi/spi-mpc52xx.c 	if (ms->transfer->transfer_list.next == &ms->message->transfers) {
next              302 drivers/spi/spi-mpc52xx.c 	ms->transfer = container_of(ms->transfer->transfer_list.next,
next              343 drivers/spi/spi-pic32-sqi.c 	struct ring_desc *rdesc, *next;
next              430 drivers/spi/spi-pic32-sqi.c 	list_for_each_entry_safe_reverse(rdesc, next,
next              781 drivers/spi/spi-pl022.c 	if (trans->transfer_list.next != &msg->transfers) {
next              783 drivers/spi/spi-pl022.c 		    list_entry(trans->transfer_list.next,
next             1591 drivers/spi/spi-pl022.c 	pl022->cur_transfer = list_entry(msg->transfers.next,
next              286 drivers/spi/spi-sh.c 		mesg = list_entry(ss->queue.next, struct spi_message, queue);
next              585 drivers/spi/spi-topcliff-pch.c 		list_for_each_entry_safe(pmsg, tmp, data->queue.next, queue) {
next              670 drivers/spi/spi-topcliff-pch.c 		list_for_each_entry_safe(pmsg, tmp, data->queue.next, queue) {
next             1133 drivers/spi/spi-topcliff-pch.c 		list_for_each_entry_safe(pmsg, tmp, data->queue.next, queue) {
next             1155 drivers/spi/spi-topcliff-pch.c 	data->current_msg = list_entry(data->queue.next, struct spi_message,
next             1178 drivers/spi/spi-topcliff-pch.c 				list_entry(data->current_msg->transfers.next,
next             1185 drivers/spi/spi-topcliff-pch.c 				list_entry(data->cur_trans->transfer_list.next,
next             1242 drivers/spi/spi-topcliff-pch.c 		if ((data->cur_trans->transfer_list.next) ==
next              256 drivers/spi/spi-txx9.c 		if (t->transfer_list.next == &m->transfers)
next              289 drivers/spi/spi-txx9.c 		m = container_of(c->queue.next, struct spi_message, queue);
next             1484 drivers/spi/spi.c 	struct spi_message *next;
next             1489 drivers/spi/spi.c 	next = list_first_entry_or_null(&ctlr->queue, struct spi_message,
next             1493 drivers/spi/spi.c 	return next;
next             2840 drivers/spi/spi.c 		if (rxfer->replaced_after->next == &msg->transfers) {
next             2857 drivers/spi/spi.c 		list_move_tail(rxfer->replaced_after->next,
next              581 drivers/ssb/sdio.c 		tuple = tuple->next;
next              273 drivers/staging/android/ashmem.c 	struct ashmem_range *range, *next;
next              276 drivers/staging/android/ashmem.c 	list_for_each_entry_safe(range, next, &asma->unpinned_list, unpinned)
next              625 drivers/staging/android/ashmem.c 	struct ashmem_range *range, *next;
next              628 drivers/staging/android/ashmem.c 	list_for_each_entry_safe(range, next, &asma->unpinned_list, unpinned) {
next              693 drivers/staging/android/ashmem.c 	struct ashmem_range *range, *next;
next              697 drivers/staging/android/ashmem.c 	list_for_each_entry_safe(range, next, &asma->unpinned_list, unpinned) {
next              439 drivers/staging/comedi/comedidev.h 	struct comedi_driver *next;	/* Next in list of COMEDI drivers. */
next              960 drivers/staging/comedi/drivers.c 	for (driv = comedi_drivers; driv; driv = driv->next) {
next              975 drivers/staging/comedi/drivers.c 		for (driv = comedi_drivers; driv; driv = driv->next) {
next             1129 drivers/staging/comedi/drivers.c 	driver->next = comedi_drivers;
next             1154 drivers/staging/comedi/drivers.c 		comedi_drivers = driver->next;
next             1156 drivers/staging/comedi/drivers.c 		for (prev = comedi_drivers; prev->next; prev = prev->next) {
next             1157 drivers/staging/comedi/drivers.c 			if (prev->next == driver) {
next             1158 drivers/staging/comedi/drivers.c 				prev->next = driver->next;
next             1530 drivers/staging/comedi/drivers/cb_pcidas64.c 		devpriv->ai_dma_desc[i].next =
next             1545 drivers/staging/comedi/drivers/cb_pcidas64.c 			devpriv->ao_dma_desc[i].next =
next             2955 drivers/staging/comedi/drivers/cb_pcidas64.c 	next_bits = le32_to_cpu(devpriv->ao_dma_desc[buffer_index].next);
next             2957 drivers/staging/comedi/drivers/cb_pcidas64.c 	devpriv->ao_dma_desc[buffer_index].next = cpu_to_le32(next_bits);
next             2962 drivers/staging/comedi/drivers/cb_pcidas64.c 	next_bits = le32_to_cpu(devpriv->ao_dma_desc[prev_buffer_index].next);
next             2964 drivers/staging/comedi/drivers/cb_pcidas64.c 	devpriv->ao_dma_desc[prev_buffer_index].next = cpu_to_le32(next_bits);
next              150 drivers/staging/comedi/drivers/gsc_hpdi.c 	unsigned int next;
next              152 drivers/staging/comedi/drivers/gsc_hpdi.c 	next = readl(devpriv->plx9080_mmio + PLX_REG_DMAPADR(channel));
next              157 drivers/staging/comedi/drivers/gsc_hpdi.c 	for (desc = 0; (next < start || next >= start + devpriv->block_size) &&
next              433 drivers/staging/comedi/drivers/gsc_hpdi.c 		devpriv->dma_desc[i].next = cpu_to_le32((phys_addr +
next              447 drivers/staging/comedi/drivers/gsc_hpdi.c 	devpriv->dma_desc[i - 1].next = cpu_to_le32(phys_addr | next_bits);
next              674 drivers/staging/comedi/drivers/mite.c 		desc->next = cpu_to_le32(ring->dma_addr +
next              687 drivers/staging/comedi/drivers/mite.c 	desc->next = cpu_to_le32(ring->dma_addr);
next               25 drivers/staging/comedi/drivers/mite.h 	__le32 next;
next               46 drivers/staging/comedi/drivers/plx9080.h 	__le32 next;
next               50 drivers/staging/comedi/proc.c 	for (driv = comedi_drivers; driv; driv = driv->next) {
next              529 drivers/staging/exfat/exfat.h 	struct buf_cache_t *next;
next               42 drivers/staging/exfat/exfat_cache.c 	bp->next = list->next;
next               44 drivers/staging/exfat/exfat_cache.c 	list->next->prev = bp;
next               45 drivers/staging/exfat/exfat_cache.c 	list->next = bp;
next               51 drivers/staging/exfat/exfat_cache.c 	bp->next = list;
next               52 drivers/staging/exfat/exfat_cache.c 	list->prev->next = bp;
next               58 drivers/staging/exfat/exfat_cache.c 	bp->prev->next = bp->next;
next               59 drivers/staging/exfat/exfat_cache.c 	bp->next->prev = bp->prev;
next               65 drivers/staging/exfat/exfat_cache.c 	bp->prev->next = bp->next;
next               66 drivers/staging/exfat/exfat_cache.c 	bp->next->prev = bp->prev;
next              138 drivers/staging/exfat/exfat_cache.c 	p_fs->FAT_cache_lru_list.next = &p_fs->FAT_cache_lru_list;
next              147 drivers/staging/exfat/exfat_cache.c 		p_fs->FAT_cache_array[i].next = NULL;
next              152 drivers/staging/exfat/exfat_cache.c 	p_fs->buf_cache_lru_list.next = &p_fs->buf_cache_lru_list;
next              161 drivers/staging/exfat/exfat_cache.c 		p_fs->buf_cache_array[i].next = NULL;
next              495 drivers/staging/exfat/exfat_cache.c 	bp = p_fs->FAT_cache_lru_list.next;
next              507 drivers/staging/exfat/exfat_cache.c 		bp = bp->next;
next              520 drivers/staging/exfat/exfat_cache.c 	bp = p_fs->FAT_cache_lru_list.next;
next              526 drivers/staging/exfat/exfat_cache.c 		bp = bp->next;
next              689 drivers/staging/exfat/exfat_cache.c 	bp = p_fs->buf_cache_lru_list.next;
next              701 drivers/staging/exfat/exfat_cache.c 		bp = bp->next;
next              714 drivers/staging/exfat/exfat_cache.c 	bp = p_fs->buf_cache_lru_list.next;
next              720 drivers/staging/exfat/exfat_cache.c 		bp = bp->next;
next              405 drivers/staging/exfat/exfat_core.c 	u32 clu, next;
next              413 drivers/staging/exfat/exfat_core.c 		while ((FAT_read(sb, clu, &next) == 0) &&
next              414 drivers/staging/exfat/exfat_core.c 		       (next != CLUSTER_32(~0))) {
next              417 drivers/staging/exfat/exfat_core.c 			clu = next;
next               98 drivers/staging/fwserial/dma_fifo.c 	struct dma_pending *pending, *next;
next              103 drivers/staging/fwserial/dma_fifo.c 	list_for_each_entry_safe(pending, next, &fifo->pending, link)
next              115 drivers/staging/fwserial/dma_fifo.c 	struct dma_pending *pending, *next;
next              120 drivers/staging/fwserial/dma_fifo.c 	list_for_each_entry_safe(pending, next, &fifo->pending, link)
next              221 drivers/staging/fwserial/dma_fifo.c 	pended->next = fifo->out;
next              244 drivers/staging/fwserial/dma_fifo.c 	struct dma_pending *pending, *next, *tmp;
next              264 drivers/staging/fwserial/dma_fifo.c 	list_for_each_entry_safe(pending, next, &fifo->pending, link) {
next              272 drivers/staging/fwserial/dma_fifo.c 			 addr_check(fifo->in, fifo->done, pending->next),
next              275 drivers/staging/fwserial/dma_fifo.c 			 pending->next))
next              279 drivers/staging/fwserial/dma_fifo.c 		fifo->done = pending->next;
next               61 drivers/staging/fwserial/dma_fifo.h 	unsigned int	 next;
next              220 drivers/staging/gdm724x/gdm_mux.c 		r = list_entry(rx->to_host_list.next, struct mux_rx,
next              152 drivers/staging/gdm724x/gdm_usb.c 	t_sdu = list_entry(tx->free_list.next, struct usb_tx_sdu, list);
next              220 drivers/staging/gdm724x/gdm_usb.c 	r = list_entry(rx->free_list.next, struct usb_rx, free_list);
next              395 drivers/staging/gdm724x/gdm_usb.c 		r = list_entry(rx->to_host_list.next,
next              606 drivers/staging/gdm724x/gdm_usb.c 		t_sdu = list_entry(tx->sdu_list.next, struct usb_tx_sdu, list);
next              659 drivers/staging/gdm724x/gdm_usb.c 		t = list_entry(tx->hci_list.next, struct usb_tx, list);
next               88 drivers/staging/greybus/audio_manager.c 	struct gb_audio_manager_module *module, *next;
next               93 drivers/staging/greybus/audio_manager.c 	list_for_each_entry_safe(module, next, &modules_list, list) {
next               89 drivers/staging/greybus/bootrom.c 				   enum next_request_type next,
next               92 drivers/staging/greybus/bootrom.c 	bootrom->next_request = next;
next             1162 drivers/staging/isdn/gigaset/bas-gigaset.c 	struct isow_urbctx_t *done, *next, *ovfl;
next             1196 drivers/staging/isdn/gigaset/bas-gigaset.c 		next = ubc->isooutfree;
next             1199 drivers/staging/isdn/gigaset/bas-gigaset.c 		if (next) {
next             1200 drivers/staging/isdn/gigaset/bas-gigaset.c 			rc = submit_iso_write_urb(next);
next             1205 drivers/staging/isdn/gigaset/bas-gigaset.c 					ubc->isooutfree = next;
next             1206 drivers/staging/isdn/gigaset/bas-gigaset.c 					next = NULL;
next             1209 drivers/staging/isdn/gigaset/bas-gigaset.c 				if (next) {
next             1261 drivers/staging/isdn/gigaset/bas-gigaset.c 		next = ubc->isooutfree;
next             1264 drivers/staging/isdn/gigaset/bas-gigaset.c 		if (next) {
next             1266 drivers/staging/isdn/gigaset/bas-gigaset.c 			rc = submit_iso_write_urb(next);
next             1751 drivers/staging/isdn/gigaset/bas-gigaset.c 	if (cb->next != NULL) {
next             1752 drivers/staging/isdn/gigaset/bas-gigaset.c 		cs->cmdbuf = cb->next;
next             2016 drivers/staging/isdn/gigaset/bas-gigaset.c 		cs->lastcmdbuf->next = cb;
next              334 drivers/staging/isdn/gigaset/common.c 	unsigned next, tail;
next              342 drivers/staging/isdn/gigaset/common.c 	next = (tail + 1) % MAX_EVENTS;
next              343 drivers/staging/isdn/gigaset/common.c 	if (unlikely(next == cs->ev_head))
next              353 drivers/staging/isdn/gigaset/common.c 		cs->ev_tail = next;
next              374 drivers/staging/isdn/gigaset/common.c 	struct at_state_t *cur, *next;
next              376 drivers/staging/isdn/gigaset/common.c 	list_for_each_entry_safe(cur, next, &cs->temp_at_states, list) {
next              837 drivers/staging/isdn/gigaset/common.c 		cb = cb->next;
next              408 drivers/staging/isdn/gigaset/ev-layer.c 	unsigned next, tail;
next              416 drivers/staging/isdn/gigaset/ev-layer.c 	next = (tail + 1) % MAX_EVENTS;
next              417 drivers/staging/isdn/gigaset/ev-layer.c 	if (unlikely(next == cs->ev_head)) {
next              428 drivers/staging/isdn/gigaset/ev-layer.c 		cs->ev_tail = next;
next              780 drivers/staging/isdn/gigaset/ev-layer.c 	cb->next = NULL;
next             1195 drivers/staging/isdn/gigaset/ev-layer.c 			cb->next = NULL;
next              519 drivers/staging/isdn/gigaset/gigaset.h 	struct cmdbuf_t *next, *prev;
next              322 drivers/staging/isdn/gigaset/interface.c 	cb->next = NULL;
next              133 drivers/staging/isdn/gigaset/ser-gigaset.c 		cs->cmdbuf = cb = cb->next;
next              209 drivers/staging/isdn/gigaset/ser-gigaset.c 		cs->cmdbuf = cb->next;
next              250 drivers/staging/isdn/gigaset/ser-gigaset.c 		cs->lastcmdbuf->next = cb;
next              437 drivers/staging/isdn/gigaset/usb-gigaset.c 			cs->cmdbuf = cb->next;
next              501 drivers/staging/isdn/gigaset/usb-gigaset.c 		cs->lastcmdbuf->next = cb;
next              155 drivers/staging/isdn/hysdn/hysdn_defs.h 	struct HYSDN_CARD *next;	/* pointer to next card */
next               95 drivers/staging/isdn/hysdn/hysdn_init.c 	card->next = NULL;	/*end of chain */
next               97 drivers/staging/isdn/hysdn/hysdn_init.c 		card_last->next = card;		/* pointer to next card */
next              129 drivers/staging/isdn/hysdn/hysdn_init.c 		card_root = card_root->next;
next              135 drivers/staging/isdn/hysdn/hysdn_init.c 			if (tmp->next == card)
next              136 drivers/staging/isdn/hysdn/hysdn_init.c 				tmp->next = card->next;
next              138 drivers/staging/isdn/hysdn/hysdn_init.c 			tmp = tmp->next;
next              381 drivers/staging/isdn/hysdn/hysdn_procconf.c 		card = card->next;	/* next entry */
next              407 drivers/staging/isdn/hysdn/hysdn_procconf.c 		card = card->next;	/* point to next card */
next               33 drivers/staging/isdn/hysdn/hysdn_proclog.c 	struct log_data *next;
next              118 drivers/staging/isdn/hysdn/hysdn_proclog.c 	ib->next = NULL;
next              125 drivers/staging/isdn/hysdn/hysdn_proclog.c 		pd->log_tail->next = ib;	/* follows existing messages */
next              129 drivers/staging/isdn/hysdn/hysdn_proclog.c 	while (pd->log_head->next) {
next              131 drivers/staging/isdn/hysdn/hysdn_proclog.c 		    (pd->log_head->next->usage_cnt <= 0)) {
next              133 drivers/staging/isdn/hysdn/hysdn_proclog.c 			pd->log_head = pd->log_head->next;
next              188 drivers/staging/isdn/hysdn/hysdn_proclog.c 	file->private_data = &inf->next;	/* next structure */
next              218 drivers/staging/isdn/hysdn/hysdn_proclog.c 			filep->private_data = &pd->log_tail->next;
next              265 drivers/staging/isdn/hysdn/hysdn_proclog.c 			inf = inf->next;
next              272 drivers/staging/isdn/hysdn/hysdn_proclog.c 					pd->log_head = pd->log_head->next;
next              498 drivers/staging/kpc2000/kpc2000/cell_probe.c 	struct list_head *next;
next              500 drivers/staging/kpc2000/kpc2000/cell_probe.c 	list_for_each_safe(ptr, next, &pcard->uio_devices_list) {
next               89 drivers/staging/kpc2000/kpc_dma/dma.c 	struct kpc_dma_descriptor *next;
next              121 drivers/staging/kpc2000/kpc_dma/dma.c 		next = dma_pool_alloc(eng->desc_pool, GFP_KERNEL | GFP_DMA, &next_handle);
next              122 drivers/staging/kpc2000/kpc_dma/dma.c 		if (!next)
next              125 drivers/staging/kpc2000/kpc_dma/dma.c 		clear_desc(next);
next              126 drivers/staging/kpc2000/kpc_dma/dma.c 		next->MyDMAAddr = next_handle;
next              129 drivers/staging/kpc2000/kpc_dma/dma.c 		cur->Next = next;
next              130 drivers/staging/kpc2000/kpc_dma/dma.c 		cur = next;
next              219 drivers/staging/kpc2000/kpc_dma/dma.c 		struct kpc_dma_descriptor *next = cur->Next;
next              224 drivers/staging/kpc2000/kpc_dma/dma.c 		cur = next;
next              201 drivers/staging/media/imx/imx-ic-prpencvf.c 	struct imx_media_buffer *done, *next;
next              219 drivers/staging/media/imx/imx-ic-prpencvf.c 	next = imx_media_capture_device_next_buf(vdev);
next              220 drivers/staging/media/imx/imx-ic-prpencvf.c 	if (next) {
next              221 drivers/staging/media/imx/imx-ic-prpencvf.c 		phys = vb2_dma_contig_plane_dma_addr(&next->vbuf.vb2_buf, 0);
next              222 drivers/staging/media/imx/imx-ic-prpencvf.c 		priv->active_vb2_buf[priv->ipu_buf_num] = next;
next              704 drivers/staging/media/imx/imx-media-capture.c 		buf = list_entry(priv->ready_q.next, struct imx_media_buffer,
next              265 drivers/staging/media/imx/imx-media-csi.c 	struct imx_media_buffer *done, *next;
next              283 drivers/staging/media/imx/imx-media-csi.c 	next = imx_media_capture_device_next_buf(vdev);
next              284 drivers/staging/media/imx/imx-media-csi.c 	if (next) {
next              285 drivers/staging/media/imx/imx-media-csi.c 		phys = vb2_dma_contig_plane_dma_addr(&next->vbuf.vb2_buf, 0);
next              286 drivers/staging/media/imx/imx-media-csi.c 		priv->active_vb2_buf[priv->ipu_buf_num] = next;
next              628 drivers/staging/media/imx/imx7-media-csi.c 	struct imx_media_buffer *done, *next;
next              643 drivers/staging/media/imx/imx7-media-csi.c 	next = imx_media_capture_device_next_buf(vdev);
next              644 drivers/staging/media/imx/imx7-media-csi.c 	if (next) {
next              645 drivers/staging/media/imx/imx7-media-csi.c 		phys = vb2_dma_contig_plane_dma_addr(&next->vbuf.vb2_buf, 0);
next              646 drivers/staging/media/imx/imx7-media-csi.c 		csi->active_vb2_buf[csi->buf_num] = next;
next               98 drivers/staging/media/ipu3/ipu3-css-fw.h 	struct imgu_fw_binary_xinfo *next __aligned(8);
next              154 drivers/staging/media/ipu3/ipu3-css-fw.h 	u64 next;
next              181 drivers/staging/most/dim2/dim2.c 		list_del(head->next);
next              189 drivers/staging/most/dim2/dim2.c 	list_move_tail(head->next, &hdm_ch->started_list);
next              288 drivers/staging/most/dim2/dim2.c 		list_del(head->next);
next              421 drivers/staging/most/dim2/dim2.c 		list_del(head->next);
next             2343 drivers/staging/octeon-usb/octeon-hcd.c 	struct cvmx_usb_transaction *transaction, *next;
next             2346 drivers/staging/octeon-usb/octeon-hcd.c 	list_for_each_entry_safe(transaction, next, &pipe->transactions, node) {
next               99 drivers/staging/octeon/ethernet-tx.c 				t->next = to_free_list;
next              109 drivers/staging/octeon/ethernet-tx.c 				to_free_list = to_free_list->next;
next              445 drivers/staging/octeon/ethernet-tx.c 		skb->next = to_free_list;
next              462 drivers/staging/octeon/ethernet-tx.c 		t->next = to_free_list;
next              473 drivers/staging/octeon/ethernet-tx.c 		to_free_list = to_free_list->next;
next             1358 drivers/staging/qlge/qlge.h 	struct tx_ring_desc *next;
next              178 drivers/staging/rtl8188eu/core/rtw_ap.c 	plist = phead->next;
next              183 drivers/staging/rtl8188eu/core/rtw_ap.c 		plist = plist->next;
next              211 drivers/staging/rtl8188eu/core/rtw_ap.c 	plist = phead->next;
next              216 drivers/staging/rtl8188eu/core/rtw_ap.c 		plist = plist->next;
next             1021 drivers/staging/rtl8188eu/core/rtw_ap.c 	plist = phead->next;
next             1025 drivers/staging/rtl8188eu/core/rtw_ap.c 		plist = plist->next;
next             1081 drivers/staging/rtl8188eu/core/rtw_ap.c 	plist = phead->next;
next             1085 drivers/staging/rtl8188eu/core/rtw_ap.c 		plist = plist->next;
next             1330 drivers/staging/rtl8188eu/core/rtw_ap.c 		plist = phead->next;
next             1336 drivers/staging/rtl8188eu/core/rtw_ap.c 			plist = plist->next;
next             1601 drivers/staging/rtl8188eu/core/rtw_ap.c 	plist = phead->next;
next             1607 drivers/staging/rtl8188eu/core/rtw_ap.c 		plist = plist->next;
next             1730 drivers/staging/rtl8188eu/core/rtw_ap.c 	plist = phead->next;
next             1733 drivers/staging/rtl8188eu/core/rtw_ap.c 		plist = plist->next;
next               26 drivers/staging/rtl8188eu/core/rtw_ioctl_set.c 	plist = phead->next;
next              191 drivers/staging/rtl8188eu/core/rtw_mlme.c 	plist = phead->next;
next              197 drivers/staging/rtl8188eu/core/rtw_mlme.c 		plist = plist->next;
next              215 drivers/staging/rtl8188eu/core/rtw_mlme.c 	plist = phead->next;
next              220 drivers/staging/rtl8188eu/core/rtw_mlme.c 		plist = plist->next;
next              325 drivers/staging/rtl8188eu/core/rtw_mlme.c 	for (plist = phead->next; plist != phead; plist = plist->next) {
next              403 drivers/staging/rtl8188eu/core/rtw_mlme.c 	plist = phead->next;
next              413 drivers/staging/rtl8188eu/core/rtw_mlme.c 		plist = plist->next;
next              696 drivers/staging/rtl8188eu/core/rtw_mlme.c 	plist = phead->next;
next              699 drivers/staging/rtl8188eu/core/rtw_mlme.c 		ptemp = plist->next;
next             1468 drivers/staging/rtl8188eu/core/rtw_mlme.c 	pmlmepriv->pscanned = phead->next;
next             1476 drivers/staging/rtl8188eu/core/rtw_mlme.c 		pmlmepriv->pscanned = pmlmepriv->pscanned->next;
next             1769 drivers/staging/rtl8188eu/core/rtw_mlme_ext.c 		plist = phead->next;
next             1778 drivers/staging/rtl8188eu/core/rtw_mlme_ext.c 			plist = plist->next;
next             5415 drivers/staging/rtl8188eu/core/rtw_mlme_ext.c 			xmitframe_plist = xmitframe_phead->next;
next             5420 drivers/staging/rtl8188eu/core/rtw_mlme_ext.c 				xmitframe_plist = xmitframe_plist->next;
next              181 drivers/staging/rtl8188eu/core/rtw_recv.c 	plist = phead->next;
next              186 drivers/staging/rtl8188eu/core/rtw_recv.c 		plist = plist->next;
next              928 drivers/staging/rtl8188eu/core/rtw_recv.c 			xmitframe_plist = xmitframe_phead->next;
next              933 drivers/staging/rtl8188eu/core/rtw_recv.c 				xmitframe_plist = xmitframe_plist->next;
next             1331 drivers/staging/rtl8188eu/core/rtw_recv.c 	plist = phead->next;
next             1348 drivers/staging/rtl8188eu/core/rtw_recv.c 	plist = plist->next;
next             1379 drivers/staging/rtl8188eu/core/rtw_recv.c 		plist = plist->next;
next             1642 drivers/staging/rtl8188eu/core/rtw_recv.c 	plist = phead->next;
next             1649 drivers/staging/rtl8188eu/core/rtw_recv.c 			plist = plist->next;
next             1673 drivers/staging/rtl8188eu/core/rtw_recv.c 	plist = phead->next;
next             1696 drivers/staging/rtl8188eu/core/rtw_recv.c 			plist = plist->next;
next              142 drivers/staging/rtl8188eu/core/rtw_sta_mgt.c 			plist = phead->next;
next              149 drivers/staging/rtl8188eu/core/rtw_sta_mgt.c 				plist = plist->next;
next              315 drivers/staging/rtl8188eu/core/rtw_sta_mgt.c 		plist = phead->next;
next              320 drivers/staging/rtl8188eu/core/rtw_sta_mgt.c 			plist = plist->next;
next              391 drivers/staging/rtl8188eu/core/rtw_sta_mgt.c 		plist = phead->next;
next              396 drivers/staging/rtl8188eu/core/rtw_sta_mgt.c 			plist = plist->next;
next              427 drivers/staging/rtl8188eu/core/rtw_sta_mgt.c 	plist = phead->next;
next              437 drivers/staging/rtl8188eu/core/rtw_sta_mgt.c 		plist = plist->next;
next              487 drivers/staging/rtl8188eu/core/rtw_sta_mgt.c 	plist = phead->next;
next              490 drivers/staging/rtl8188eu/core/rtw_sta_mgt.c 		plist = plist->next;
next             1335 drivers/staging/rtl8188eu/core/rtw_xmit.c 	plist = phead->next;
next             1340 drivers/staging/rtl8188eu/core/rtw_xmit.c 		plist = plist->next;
next             1364 drivers/staging/rtl8188eu/core/rtw_xmit.c 	xmitframe_plist = xmitframe_phead->next;
next             1369 drivers/staging/rtl8188eu/core/rtw_xmit.c 		xmitframe_plist = xmitframe_plist->next;
next             1404 drivers/staging/rtl8188eu/core/rtw_xmit.c 		sta_plist = sta_phead->next;
next             1422 drivers/staging/rtl8188eu/core/rtw_xmit.c 			sta_plist = sta_plist->next;
next             1746 drivers/staging/rtl8188eu/core/rtw_xmit.c 	plist = phead->next;
next             1751 drivers/staging/rtl8188eu/core/rtw_xmit.c 		plist = plist->next;
next             1818 drivers/staging/rtl8188eu/core/rtw_xmit.c 	xmitframe_plist = xmitframe_phead->next;
next             1823 drivers/staging/rtl8188eu/core/rtw_xmit.c 		xmitframe_plist = xmitframe_plist->next;
next             1899 drivers/staging/rtl8188eu/core/rtw_xmit.c 		xmitframe_plist = xmitframe_phead->next;
next             1904 drivers/staging/rtl8188eu/core/rtw_xmit.c 			xmitframe_plist = xmitframe_plist->next;
next             1946 drivers/staging/rtl8188eu/core/rtw_xmit.c 	xmitframe_plist = xmitframe_phead->next;
next             1951 drivers/staging/rtl8188eu/core/rtw_xmit.c 		xmitframe_plist = xmitframe_plist->next;
next              512 drivers/staging/rtl8188eu/hal/rtl8188eu_xmit.c 	xmitframe_plist = xmitframe_phead->next;
next              516 drivers/staging/rtl8188eu/hal/rtl8188eu_xmit.c 		xmitframe_plist = xmitframe_plist->next;
next              982 drivers/staging/rtl8188eu/os_dep/ioctl_linux.c 	pmlmepriv->pscanned = phead->next;
next              987 drivers/staging/rtl8188eu/os_dep/ioctl_linux.c 		pmlmepriv->pscanned = pmlmepriv->pscanned->next;
next             1239 drivers/staging/rtl8188eu/os_dep/ioctl_linux.c 	plist = phead->next;
next             1253 drivers/staging/rtl8188eu/os_dep/ioctl_linux.c 		plist = plist->next;
next             1323 drivers/staging/rtl8188eu/os_dep/ioctl_linux.c 		pmlmepriv->pscanned = phead->next;
next             1328 drivers/staging/rtl8188eu/os_dep/ioctl_linux.c 			pmlmepriv->pscanned = pmlmepriv->pscanned->next;
next              123 drivers/staging/rtl8188eu/os_dep/xmit_linux.c 	plist = phead->next;
next              129 drivers/staging/rtl8188eu/os_dep/xmit_linux.c 		plist = plist->next;
next              258 drivers/staging/rtl8192e/rtl8192e/rtl_core.h 	struct tx_ring *next;
next              354 drivers/staging/rtl8192e/rtl819x_TSProc.c 		(*ppTS) = list_entry(pUnusedList->next,
next              454 drivers/staging/rtl8192e/rtllib_rx.c 	while (pList->next != &pTS->RxPendingPktList) {
next              456 drivers/staging/rtl8192e/rtllib_rx.c 		    list_entry(pList->next, struct rx_reorder_entry,
next              458 drivers/staging/rtl8192e/rtllib_rx.c 			pList = pList->next;
next              460 drivers/staging/rtl8192e/rtllib_rx.c 			((struct rx_reorder_entry *)list_entry(pList->next,
next              466 drivers/staging/rtl8192e/rtllib_rx.c 	pReorderEntry->List.next = pList->next;
next              467 drivers/staging/rtl8192e/rtllib_rx.c 	pReorderEntry->List.next->prev = &pReorderEntry->List;
next              469 drivers/staging/rtl8192e/rtllib_rx.c 	pList->next = &pReorderEntry->List;
next              646 drivers/staging/rtl8192e/rtllib_rx.c 					list_entry(ieee->RxReorder_Unused_List.next,
next             2672 drivers/staging/rtl8192e/rtllib_rx.c 			target = list_entry(ieee->network_free_list.next,
next             2674 drivers/staging/rtl8192e/rtllib_rx.c 			list_del(ieee->network_free_list.next);
next               40 drivers/staging/rtl8192u/ieee80211/ieee80211_crypt.c 	for (ptr = ieee->crypt_deinit_list.next, n = ptr->next;
next               41 drivers/staging/rtl8192u/ieee80211/ieee80211_crypt.c 	     ptr != &ieee->crypt_deinit_list; ptr = n, n = ptr->next) {
next              132 drivers/staging/rtl8192u/ieee80211/ieee80211_crypt.c 	for (ptr = hcrypt->algs.next; ptr != &hcrypt->algs; ptr = ptr->next) {
next              163 drivers/staging/rtl8192u/ieee80211/ieee80211_crypt.c 	for (ptr = hcrypt->algs.next; ptr != &hcrypt->algs; ptr = ptr->next) {
next              224 drivers/staging/rtl8192u/ieee80211/ieee80211_crypt.c 	for (ptr = hcrypt->algs.next, n = ptr->next; ptr != &hcrypt->algs;
next              225 drivers/staging/rtl8192u/ieee80211/ieee80211_crypt.c 	     ptr = n, n = ptr->next) {
next              508 drivers/staging/rtl8192u/ieee80211/ieee80211_rx.c 	while (pList->next != &pTS->rx_pending_pkt_list) {
next              509 drivers/staging/rtl8192u/ieee80211/ieee80211_rx.c 		if (SN_LESS(pReorderEntry->SeqNum, list_entry(pList->next, struct rx_reorder_entry, List)->SeqNum))
next              510 drivers/staging/rtl8192u/ieee80211/ieee80211_rx.c 			pList = pList->next;
next              511 drivers/staging/rtl8192u/ieee80211/ieee80211_rx.c 		else if (SN_EQUAL(pReorderEntry->SeqNum, list_entry(pList->next, struct rx_reorder_entry, List)->SeqNum))
next              516 drivers/staging/rtl8192u/ieee80211/ieee80211_rx.c 	pReorderEntry->List.next = pList->next;
next              517 drivers/staging/rtl8192u/ieee80211/ieee80211_rx.c 	pReorderEntry->List.next->prev = &pReorderEntry->List;
next              519 drivers/staging/rtl8192u/ieee80211/ieee80211_rx.c 	pList->next = &pReorderEntry->List;
next              651 drivers/staging/rtl8192u/ieee80211/ieee80211_rx.c 			pReorderEntry = list_entry(ieee->RxReorder_Unused_List.next, struct rx_reorder_entry, List);
next             2439 drivers/staging/rtl8192u/ieee80211/ieee80211_rx.c 			target = list_entry(ieee->network_free_list.next,
next             2441 drivers/staging/rtl8192u/ieee80211/ieee80211_rx.c 			list_del(ieee->network_free_list.next);
next              372 drivers/staging/rtl8192u/ieee80211/rtl819x_TSProc.c 				(*ppTS) = list_entry(pUnusedList->next, struct ts_common_info, list);
next              459 drivers/staging/rtl8192u/r8192U.h 	struct buffer *next;
next              191 drivers/staging/rtl8712/rtl8712_recv.c 	plist = phead->next;
next              206 drivers/staging/rtl8712/rtl8712_recv.c 	plist = plist->next;
next              232 drivers/staging/rtl8712/rtl8712_recv.c 		plist = plist->next;
next              485 drivers/staging/rtl8712/rtl8712_recv.c 	plist = phead->next;
next              490 drivers/staging/rtl8712/rtl8712_recv.c 			plist = plist->next;
next              513 drivers/staging/rtl8712/rtl8712_recv.c 	plist = phead->next;
next              530 drivers/staging/rtl8712/rtl8712_recv.c 			plist = plist->next;
next              158 drivers/staging/rtl8712/rtl8712_xmit.c 	xmitframe_plist = xmitframe_phead->next;
next              198 drivers/staging/rtl8712/rtl8712_xmit.c 		sta_plist = sta_phead->next;
next              209 drivers/staging/rtl8712/rtl8712_xmit.c 			sta_plist = sta_plist->next;
next             1055 drivers/staging/rtl8712/rtl871x_ioctl_linux.c 	pmlmepriv->pscanned = phead->next;
next             1061 drivers/staging/rtl8712/rtl871x_ioctl_linux.c 		pmlmepriv->pscanned = pmlmepriv->pscanned->next;
next             1207 drivers/staging/rtl8712/rtl871x_ioctl_linux.c 	plist = phead->next;
next             1217 drivers/staging/rtl8712/rtl871x_ioctl_linux.c 		plist = plist->next;
next             1266 drivers/staging/rtl8712/rtl871x_ioctl_linux.c 		pmlmepriv->pscanned = phead->next;
next             1272 drivers/staging/rtl8712/rtl871x_ioctl_linux.c 			pmlmepriv->pscanned = pmlmepriv->pscanned->next;
next             1975 drivers/staging/rtl8712/rtl871x_ioctl_linux.c 	plist = phead->next;
next             2003 drivers/staging/rtl8712/rtl871x_ioctl_linux.c 		plist = plist->next;
next               49 drivers/staging/rtl8712/rtl871x_ioctl_set.c 	plist = phead->next;
next              143 drivers/staging/rtl8712/rtl871x_mlme.c 	plist = phead->next;
next              146 drivers/staging/rtl8712/rtl871x_mlme.c 		plist = plist->next;
next              164 drivers/staging/rtl8712/rtl871x_mlme.c 	plist = phead->next;
next              167 drivers/staging/rtl8712/rtl871x_mlme.c 		plist = plist->next;
next              265 drivers/staging/rtl8712/rtl871x_mlme.c 	plist = phead->next;
next              276 drivers/staging/rtl8712/rtl871x_mlme.c 		plist = plist->next;
next              344 drivers/staging/rtl8712/rtl871x_mlme.c 	plist = phead->next;
next              358 drivers/staging/rtl8712/rtl871x_mlme.c 		plist = plist->next;
next             1080 drivers/staging/rtl8712/rtl871x_mlme.c 	pmlmepriv->pscanned = phead->next;
next             1092 drivers/staging/rtl8712/rtl871x_mlme.c 		pmlmepriv->pscanned = pmlmepriv->pscanned->next;
next              128 drivers/staging/rtl8712/rtl871x_recv.c 	plist = phead->next;
next              131 drivers/staging/rtl8712/rtl871x_recv.c 		plist = plist->next;
next               73 drivers/staging/rtl8712/rtl871x_sta_mgt.c 	plist = phead->next;
next               75 drivers/staging/rtl8712/rtl871x_sta_mgt.c 		plist = plist->next;
next              205 drivers/staging/rtl8712/rtl871x_sta_mgt.c 		plist = phead->next;
next              209 drivers/staging/rtl8712/rtl871x_sta_mgt.c 			plist = plist->next;
next              230 drivers/staging/rtl8712/rtl871x_sta_mgt.c 	plist = phead->next;
next              238 drivers/staging/rtl8712/rtl871x_sta_mgt.c 		plist = plist->next;
next              845 drivers/staging/rtl8712/rtl871x_xmit.c 	plist = phead->next;
next              848 drivers/staging/rtl8712/rtl871x_xmit.c 		plist = plist->next;
next              426 drivers/staging/rtl8723bs/hal/rtl8723bs_xmit.c next:
next              461 drivers/staging/rtl8723bs/hal/rtl8723bs_xmit.c 		goto next;
next              468 drivers/staging/rtl8723bs/hal/rtl8723bs_xmit.c 		goto next;
next               69 drivers/staging/rtl8723bs/include/osdep_service_linux.h 	return list->next;
next              518 drivers/staging/unisys/include/iochannel.h 	struct uiscmdrsp *next;
next               46 drivers/staging/unisys/visorhba/visorhba_main.c 	struct visordisk_info *next;
next              304 drivers/staging/unisys/visornic/visornic_main.c 		     skbinlist = skbinlist->next) {
next             1296 drivers/staging/unisys/visornic/visornic_main.c 			curr->next = NULL;
next             1301 drivers/staging/unisys/visornic/visornic_main.c 				prev->next = curr;
next              164 drivers/staging/uwb/beacon.c 	struct uwb_beca_e *bce, *next;
next              165 drivers/staging/uwb/beacon.c 	list_for_each_entry_safe(bce, next, &rc->uwb_beca.list, node) {
next              179 drivers/staging/uwb/beacon.c 	struct uwb_beca_e *bce, *next;
next              180 drivers/staging/uwb/beacon.c 	list_for_each_entry_safe(bce, next, &rc->uwb_beca.list, node) {
next              274 drivers/staging/uwb/beacon.c 	struct uwb_beca_e *bce, *next;
next              278 drivers/staging/uwb/beacon.c 	list_for_each_entry_safe(bce, next, &rc->uwb_beca.list, node) {
next              290 drivers/staging/uwb/beacon.c 	struct uwb_beca_e *bce, *next;
next              293 drivers/staging/uwb/beacon.c 	list_for_each_entry_safe(bce, next, &rc->uwb_beca.list, node) {
next              131 drivers/staging/uwb/drp-ie.c 	int z, i, num_fields = 0, next = 0;
next              147 drivers/staging/uwb/drp-ie.c 			for (i = 0; i < next; i++) {
next              156 drivers/staging/uwb/drp-ie.c 				zones[next].zone_bm = 1 << z;
next              157 drivers/staging/uwb/drp-ie.c 				zones[next].mas_bm = current_bmp;
next              158 drivers/staging/uwb/drp-ie.c 				next++;
next               28 drivers/staging/uwb/i1480/dfu/mac.c 	struct fw_hdr *next;
next               36 drivers/staging/uwb/i1480/dfu/mac.c 	struct fw_hdr *next;
next               39 drivers/staging/uwb/i1480/dfu/mac.c 		next = hdr->next;
next               41 drivers/staging/uwb/i1480/dfu/mac.c 		hdr = next;
next               61 drivers/staging/uwb/i1480/dfu/mac.c 	hdr->next = NULL;
next              125 drivers/staging/uwb/i1480/dfu/mac.c 		prev_hdr = &hdr->next;
next              208 drivers/staging/uwb/i1480/dfu/mac.c 	for (hdr_itr = hdr; hdr_itr != NULL; hdr_itr = hdr_itr->next) {
next              265 drivers/staging/vc04_services/bcm2835-camera/mmal-msg.h 	u32 next;	/* next header */
next              130 drivers/staging/vc04_services/interface/vchi/vchi_common.h 	void *next;
next              253 drivers/staging/vc04_services/interface/vchiq_arm/vchiq_arm.c 		struct bulk_waiter_node *waiter, *next;
next              255 drivers/staging/vc04_services/interface/vchiq_arm/vchiq_arm.c 		list_for_each_entry_safe(waiter, next,
next             2048 drivers/staging/vc04_services/interface/vchiq_arm/vchiq_arm.c 		struct bulk_waiter_node *waiter, *next;
next             2050 drivers/staging/vc04_services/interface/vchiq_arm/vchiq_arm.c 		list_for_each_entry_safe(waiter, next,
next              191 drivers/staging/vt6655/desc.h 	struct vnt_rx_desc *next __aligned(8);
next              234 drivers/staging/vt6655/desc.h 	struct vnt_tx_desc *next __aligned(8);
next              549 drivers/staging/vt6655/device_main.c 		desc->next = &priv->aRD0Ring[(i + 1) % priv->opts.rx_descs0];
next              595 drivers/staging/vt6655/device_main.c 		desc->next = &priv->aRD1Ring[(i+1) % priv->opts.rx_descs1];
next              662 drivers/staging/vt6655/device_main.c 		desc->next = &(priv->apTD0Rings[(i+1) % priv->opts.tx_descs[0]]);
next              702 drivers/staging/vt6655/device_main.c 		desc->next = &(priv->apTD1Rings[(i + 1) % priv->opts.tx_descs[1]]);
next              756 drivers/staging/vt6655/device_main.c 	     rd = rd->next) {
next              900 drivers/staging/vt6655/device_main.c 	for (desc = priv->apTailTD[idx]; priv->iTDUsed[idx] > 0; desc = desc->next) {
next             1176 drivers/staging/vt6655/device_main.c 	priv->apCurrTD[dma_idx] = head_td->next;
next              282 drivers/staging/wlan-ng/hfa384x_usb.c 	return list_entry(hw->ctlxq.active.next, struct hfa384x_usbctlx, list);
next             2835 drivers/staging/wlan-ng/hfa384x_usb.c 		head = list_entry(hw->ctlxq.pending.next,
next               18 drivers/staging/wusbcore/host/whci/asl.c 			       struct whc_qset **next, struct whc_qset **prev)
next               24 drivers/staging/wusbcore/host/whci/asl.c 	n = qset->list_node.next;
next               26 drivers/staging/wusbcore/host/whci/asl.c 		n = n->next;
next               31 drivers/staging/wusbcore/host/whci/asl.c 	*next = container_of(n, struct whc_qset, list_node);
next               44 drivers/staging/wusbcore/host/whci/asl.c 	struct whc_qset *next, *prev;
next               49 drivers/staging/wusbcore/host/whci/asl.c 	qset_get_next_prev(whc, qset, &next, &prev);
next               50 drivers/staging/wusbcore/host/whci/asl.c 	whc_qset_set_link_ptr(&qset->qh.link, next->qset_dma);
next               57 drivers/staging/wusbcore/host/whci/asl.c 	struct whc_qset *prev, *next;
next               59 drivers/staging/wusbcore/host/whci/asl.c 	qset_get_next_prev(whc, qset, &next, &prev);
next               72 drivers/staging/wusbcore/host/whci/asl.c 	whc_qset_set_link_ptr(&prev->qh.link, next->qset_dma);
next             1740 drivers/staging/wusbcore/wa-xfer.c 	struct wa_xfer *xfer, *next;
next             1754 drivers/staging/wusbcore/wa-xfer.c 	list_for_each_entry_safe(xfer, next, &tmp_list, list_node) {
next             1771 drivers/staging/wusbcore/wa-xfer.c 	struct wa_xfer *xfer, *next;
next             1786 drivers/staging/wusbcore/wa-xfer.c 	list_for_each_entry_safe(xfer, next, &tmp_list, list_node) {
next              120 drivers/target/iscsi/cxgbit/cxgbit.h 	struct np_info *next;
next               90 drivers/target/iscsi/cxgbit/cxgbit_cm.c 		p->next = cdev->np_hash_tab[bucket];
next              105 drivers/target/iscsi/cxgbit/cxgbit_cm.c 	for (p = cdev->np_hash_tab[bucket]; p; p = p->next) {
next              122 drivers/target/iscsi/cxgbit/cxgbit_cm.c 	for (p = *prev; p; prev = &p->next, p = p->next) {
next              125 drivers/target/iscsi/cxgbit/cxgbit_cm.c 			*prev = p->next;
next              638 drivers/target/iscsi/iscsi_target_tmr.c 				goto next;
next              646 drivers/target/iscsi/iscsi_target_tmr.c 				goto next;
next              688 drivers/target/iscsi/iscsi_target_tmr.c 				goto next;
next              700 drivers/target/iscsi/iscsi_target_tmr.c next:
next             1739 drivers/target/target_core_alua.c 	struct se_lun *lun, *next;
next             1770 drivers/target/target_core_alua.c 	list_for_each_entry_safe(lun, next,
next              703 drivers/target/target_core_pr.c 	struct se_lun *lun_tmp, *next, *dest_lun;
next              726 drivers/target/target_core_pr.c 	list_for_each_entry_safe(lun_tmp, next, &dev->dev_sep_list, lun_dev_link) {
next             2488 drivers/target/target_core_pr.c 			list_entry(dev->t10_pr.registration_list.next,
next             1419 drivers/target/target_core_sbc.c 				goto next;
next             1445 drivers/target/target_core_sbc.c next:
next              137 drivers/target/target_core_spc.c 	bool next = true;
next              153 drivers/target/target_core_spc.c 		if (next) {
next              154 drivers/target/target_core_spc.c 			next = false;
next              157 drivers/target/target_core_spc.c 			next = true;
next              259 drivers/target/target_core_tmr.c 	struct se_cmd *cmd, *next;
next              286 drivers/target/target_core_tmr.c 	list_for_each_entry_safe(cmd, next, &dev->state_list, state_list) {
next              316 drivers/target/target_core_tmr.c 		cmd = list_entry(drain_task_list.next, struct se_cmd, state_list);
next             2093 drivers/target/target_core_transport.c 		cmd = list_entry(dev->delayed_cmd_list.next,
next              810 drivers/thermal/thermal_core.c 	struct thermal_instance *pos, *next;
next              814 drivers/thermal/thermal_core.c 	list_for_each_entry_safe(pos, next, &tz->thermal_instances, tz_node) {
next               85 drivers/thunderbolt/cap.c 		offset = header.basic.next;
next              131 drivers/thunderbolt/cap.c 		offset = header.basic.next;
next              171 drivers/thunderbolt/cap.c 			offset = header.extended_long.next;
next              177 drivers/thunderbolt/cap.c 			offset = header.extended_short.next;
next              768 drivers/thunderbolt/switch.c 	struct tb_port *next;
next              782 drivers/thunderbolt/switch.c 			next = prev->remote;
next              784 drivers/thunderbolt/switch.c 			next = tb_port_at(tb_route(end->sw), prev->sw);
next              787 drivers/thunderbolt/switch.c 			next = prev->remote;
next              789 drivers/thunderbolt/switch.c 			next = tb_upstream_port(prev->sw);
next              794 drivers/thunderbolt/switch.c 			if (next->dual_link_port &&
next              795 drivers/thunderbolt/switch.c 			    next->link_nr != prev->link_nr) {
next              796 drivers/thunderbolt/switch.c 				next = next->dual_link_port;
next              801 drivers/thunderbolt/switch.c 	return next;
next               56 drivers/thunderbolt/tb_regs.h 	u8 next;
next               70 drivers/thunderbolt/tb_regs.h 	u8 next;
next               90 drivers/thunderbolt/tb_regs.h 	u16 next;
next               94 drivers/tty/goldfish.c 			unsigned long next =
next               96 drivers/tty/goldfish.c 			unsigned long avail = next - addr;
next               97 drivers/tty/hvc/hvc_console.c 	list_for_each_entry(hp, &hvc_structs, next) {
next              257 drivers/tty/hvc/hvc_console.c 	list_del(&(hp->next));
next              803 drivers/tty/hvc/hvc_console.c 			list_for_each_entry(hp, &hvc_structs, next) {
next              969 drivers/tty/hvc/hvc_console.c 	list_add_tail(&(hp->next), &hvc_structs);
next               49 drivers/tty/hvc/hvc_console.h 	struct list_head next;
next              191 drivers/tty/hvc/hvc_iucv.c 	struct iucv_tty_buffer *ent, *next;
next              193 drivers/tty/hvc/hvc_iucv.c 	list_for_each_entry_safe(ent, next, list, list) {
next              955 drivers/tty/hvc/hvc_iucv.c 	struct iucv_tty_buffer	*ent, *next;
next              959 drivers/tty/hvc/hvc_iucv.c 	list_for_each_entry_safe(ent, next, &priv->tty_outqueue, list)
next              287 drivers/tty/hvc/hvcs.c 	struct list_head next; /* list management */
next              641 drivers/tty/hvc/hvcs.c 		list_for_each_entry(hvcsd, &hvcs_structs, next) {
next              694 drivers/tty/hvc/hvcs.c 	list_del(&(hvcsd->next));
next              805 drivers/tty/hvc/hvcs.c 	list_add_tail(&(hvcsd->next), &hvcs_structs);
next              933 drivers/tty/hvc/hvcs.c 	list_for_each_entry(hvcsd, &hvcs_structs, next) {
next             1071 drivers/tty/hvc/hvcs.c 	list_for_each_entry(hvcsd, &hvcs_structs, next) {
next              240 drivers/tty/n_r3964.c 	pHeader->next = NULL;
next              245 drivers/tty/n_r3964.c 		pInfo->tx_last->next = pHeader;
next              271 drivers/tty/n_r3964.c 	for (pDump = pHeader; pDump; pDump = pDump->next)
next              289 drivers/tty/n_r3964.c 	pInfo->tx_first = pHeader->next;
next              310 drivers/tty/n_r3964.c 	pHeader->next = NULL;
next              315 drivers/tty/n_r3964.c 		pInfo->rx_last->next = pHeader;
next              345 drivers/tty/n_r3964.c 		pInfo->rx_first = pHeader->next;
next              353 drivers/tty/n_r3964.c 		for (pFind = pInfo->rx_first; pFind; pFind = pFind->next) {
next              354 drivers/tty/n_r3964.c 			if (pFind->next == pHeader) {
next              356 drivers/tty/n_r3964.c 				pFind->next = pHeader->next;
next              358 drivers/tty/n_r3964.c 				if (pFind->next == NULL) {
next              535 drivers/tty/n_r3964.c 	pBlock->next = NULL;
next              544 drivers/tty/n_r3964.c 	for (pClient = pInfo->firstClient; pClient; pClient = pClient->next) {
next              747 drivers/tty/n_r3964.c 	for (pClient = pInfo->firstClient; pClient; pClient = pClient->next) {
next              764 drivers/tty/n_r3964.c 		     ppClient = &(*ppClient)->next) {
next              770 drivers/tty/n_r3964.c 				*ppClient = pClient->next;
next              803 drivers/tty/n_r3964.c 			pClient->next = pInfo->firstClient;
next              866 drivers/tty/n_r3964.c 		pMsg->next = NULL;
next              871 drivers/tty/n_r3964.c 			pClient->last_msg->next = pMsg;
next              911 drivers/tty/n_r3964.c 		pClient->first_msg = pMsg->next;
next             1027 drivers/tty/n_r3964.c 		pNext = pClient->next;
next             1047 drivers/tty/n_r3964.c 		pNextHeader = pHeader->next;
next              132 drivers/tty/serial/8250/8250_core.c 		l = l->next;
next              158 drivers/tty/serial/8250/8250_core.c 			i->head = i->head->next;
next               29 drivers/tty/serial/8250/8250_hp300.c 	struct hp300_port *next;	/* next port */
next              273 drivers/tty/serial/8250/8250_hp300.c 		port->next = hp300_ports;
next              309 drivers/tty/serial/8250/8250_hp300.c 		port = port->next;
next               70 drivers/tty/serial/ip22zilog.c 	struct uart_ip22zilog_port	*next;
next              462 drivers/tty/serial/ip22zilog.c 		up = up->next;
next              484 drivers/tty/serial/ip22zilog.c 		up = up->next;
next              712 drivers/tty/serial/ip22zilog.c 	up->next->flags |= IP22ZILOG_FLAG_RESET_DONE;
next             1096 drivers/tty/serial/ip22zilog.c 		up[channel].next = &up[channel - 1];
next             1097 drivers/tty/serial/ip22zilog.c 	up[channel].next = NULL;
next              180 drivers/tty/serial/kgdboc.c 		cons = cons->next;
next             1090 drivers/tty/serial/pch_uart.c 	int next = 1;
next             1095 drivers/tty/serial/pch_uart.c 	while (next) {
next             1137 drivers/tty/serial/pch_uart.c 			next = 0; /* MS ir prioirty is the lowest. So, MS ir
next             1147 drivers/tty/serial/pch_uart.c 			next = 0;
next               78 drivers/tty/serial/sunzilog.c 	struct uart_sunzilog_port	*next;
next              562 drivers/tty/serial/sunzilog.c 		up = up->next;
next              584 drivers/tty/serial/sunzilog.c 		up = up->next;
next             1107 drivers/tty/serial/sunzilog.c 			up->next = up + 1;
next             1109 drivers/tty/serial/sunzilog.c 			up->next = NULL;
next             1597 drivers/tty/serial/sunzilog.c 			up = up->next;
next             1634 drivers/tty/serial/sunzilog.c 			up = up->next;
next              184 drivers/tty/synclink_gt.c 	struct cond_wait *next;
next              202 drivers/tty/synclink_gt.c 	__le32 next;  /* physical address of next descriptor */
next              212 drivers/tty/synclink_gt.c #define set_desc_next(a,b) (a).next   = cpu_to_le32((unsigned int)(b))
next             2316 drivers/tty/synclink_gt.c 	for (w = info->gpio_wait_q, prev = NULL ; w != NULL ; w = w->next) {
next             2321 drivers/tty/synclink_gt.c 				prev->next = w->next;
next             2323 drivers/tty/synclink_gt.c 				info->gpio_wait_q = w->next;
next             3001 drivers/tty/synclink_gt.c 	w->next = *head;
next             3010 drivers/tty/synclink_gt.c 	for (w = *head, prev = NULL ; w != NULL ; prev = w, w = w->next) {
next             3013 drivers/tty/synclink_gt.c 				prev->next = w->next;
next             3015 drivers/tty/synclink_gt.c 				*head = w->next;
next             3025 drivers/tty/synclink_gt.c 		*head = (*head)->next;
next             3364 drivers/tty/synclink_gt.c 			info->rbufs[i].next = cpu_to_le32(pbufs);
next             3366 drivers/tty/synclink_gt.c 			info->rbufs[i].next = cpu_to_le32(pbufs + ((i+1) * sizeof(struct slgt_desc)));
next             3376 drivers/tty/synclink_gt.c 			info->tbufs[i].next = cpu_to_le32(pbufs + info->rbuf_count * sizeof(struct slgt_desc));
next             3378 drivers/tty/synclink_gt.c 			info->tbufs[i].next = cpu_to_le32(pbufs + ((info->rbuf_count + i + 1) * sizeof(struct slgt_desc)));
next              113 drivers/tty/synclinkmp.c 	u16	next;		/* lower l6 bits of next descriptor addr */
next             3447 drivers/tty/synclinkmp.c 		info->rx_buf_list[i].next = info->buffer_list_phys;
next             3449 drivers/tty/synclinkmp.c 			info->rx_buf_list[i].next += (i + 1) * sizeof(SCADESC);
next             3462 drivers/tty/synclinkmp.c 		info->tx_buf_list[i].next = info->buffer_list_phys +
next             3466 drivers/tty/synclinkmp.c 			info->tx_buf_list[i].next += (i + 1) * sizeof(SCADESC);
next              102 drivers/tty/tty_buffer.c 	p->next = NULL;
next              119 drivers/tty/tty_buffer.c 	struct tty_buffer *p, *next;
next              125 drivers/tty/tty_buffer.c 		buf->head = p->next;
next              131 drivers/tty/tty_buffer.c 	llist_for_each_entry_safe(p, next, llist, free)
next              223 drivers/tty/tty_buffer.c 	struct tty_buffer *next;
next              231 drivers/tty/tty_buffer.c 	while ((next = smp_load_acquire(&buf->head->next)) != NULL) {
next              233 drivers/tty/tty_buffer.c 		buf->head = next;
next              285 drivers/tty/tty_buffer.c 			smp_store_release(&b->next, n);
next              509 drivers/tty/tty_buffer.c 		struct tty_buffer *next;
next              520 drivers/tty/tty_buffer.c 		next = smp_load_acquire(&head->next);
next              526 drivers/tty/tty_buffer.c 			if (next == NULL)
next              528 drivers/tty/tty_buffer.c 			buf->head = next;
next              243 drivers/tty/tty_ldisc.c 	.next	= tty_ldiscs_seq_next,
next               76 drivers/tty/tty_ldsem.c 	struct ldsem_waiter *waiter, *next;
next               94 drivers/tty/tty_ldsem.c 	list_for_each_entry_safe(waiter, next, &sem->read_wait, list) {
next              123 drivers/tty/tty_ldsem.c 	waiter = list_entry(sem->write_wait.next, struct ldsem_waiter, list);
next              206 drivers/usb/atm/usbatm.c 	urb = list_entry(channel->list.next, struct urb, urb_list);
next              787 drivers/usb/c67x00/c67x00-sched.c 			urb = list_entry(ep_data->queue.next,
next              894 drivers/usb/c67x00/c67x00-sched.c 	while (td->td_list.next != &c67x00->td_list) {
next              895 drivers/usb/c67x00/c67x00-sched.c 		td = list_entry(td->td_list.next, struct c67x00_td, td_list);
next             1033 drivers/usb/c67x00/c67x00-sched.c 		tmp = list_entry(td->td_list.next, typeof(*td), td_list);
next             1073 drivers/usb/c67x00/c67x00-sched.c 		if (td->td_list.next == &c67x00->td_list)
next               34 drivers/usb/chipidea/otg_fsm.c 	char		*next;
next               38 drivers/usb/chipidea/otg_fsm.c 	next = buf;
next               40 drivers/usb/chipidea/otg_fsm.c 	t = scnprintf(next, size, "%d\n", ci->fsm.a_bus_req);
next               42 drivers/usb/chipidea/otg_fsm.c 	next += t;
next               83 drivers/usb/chipidea/otg_fsm.c 	char		*next;
next               87 drivers/usb/chipidea/otg_fsm.c 	next = buf;
next               89 drivers/usb/chipidea/otg_fsm.c 	t = scnprintf(next, size, "%d\n", ci->fsm.a_bus_drop);
next               91 drivers/usb/chipidea/otg_fsm.c 	next += t;
next              123 drivers/usb/chipidea/otg_fsm.c 	char		*next;
next              127 drivers/usb/chipidea/otg_fsm.c 	next = buf;
next              129 drivers/usb/chipidea/otg_fsm.c 	t = scnprintf(next, size, "%d\n", ci->fsm.b_bus_req);
next              131 drivers/usb/chipidea/otg_fsm.c 	next += t;
next              385 drivers/usb/chipidea/udc.c 		lastnode->ptr->next = cpu_to_le32(node->dma);
next              464 drivers/usb/chipidea/udc.c 	lastnode->ptr->next = cpu_to_le32(TD_TERMINATE);
next              475 drivers/usb/chipidea/udc.c 		u32 next = firstnode->dma & TD_ADDR_MASK;
next              482 drivers/usb/chipidea/udc.c 		prevlastnode->ptr->next = cpu_to_le32(next);
next              496 drivers/usb/chipidea/udc.c 	hwep->qh.ptr->td.next = cpu_to_le32(firstnode->dma);
next              531 drivers/usb/chipidea/udc.c 	hwep->qh.ptr->td.next = cpu_to_le32(node->dma);
next              636 drivers/usb/chipidea/udc.c 		struct ci_hw_req *hwreq = list_entry(hwep->qh.queue.next,
next             1270 drivers/usb/chipidea/udc.c 	hwep->qh.ptr->td.next |= cpu_to_le32(TD_TERMINATE);   /* needed? */
next               22 drivers/usb/chipidea/udc.h 	__le32 next;
next              444 drivers/usb/core/devio.c 		as = list_entry(ps->async_completed.next, struct async,
next              648 drivers/usb/core/devio.c 		as = list_entry(list->next, struct async, asynclist);
next             2760 drivers/usb/core/devio.c 		ps = list_entry(udev->filelist.next, struct usb_dev_state, list);
next             1677 drivers/usb/core/hcd.c 		urb = list_entry(local_list.next, struct urb, urb_list);
next              782 drivers/usb/core/hub.c 		struct list_head	*next;
next              788 drivers/usb/core/hub.c 		next = hub->tt.clear_list.next;
next              789 drivers/usb/core/hub.c 		clear = list_entry(next, struct usb_tt_clear, clear_list);
next              949 drivers/usb/core/urb.c 		victim = list_entry(anchor->urb_list.next, struct urb,
next              259 drivers/usb/dwc2/debugfs.c 		   ep->queue.next, ep->queue.prev);
next             4684 drivers/usb/dwc2/gadget.c 		u32 next = DXEPCTL_NEXTEP((epnum + 1) % 15);
next             4687 drivers/usb/dwc2/gadget.c 			dwc2_writel(hsotg, next, DIEPCTL(epnum));
next             4689 drivers/usb/dwc2/gadget.c 			dwc2_writel(hsotg, next, DOEPCTL(epnum));
next             1626 drivers/usb/dwc2/hcd.c 	if (!qh_list->next)
next             2728 drivers/usb/dwc2/hcd.c 	qh_ptr = hsotg->periodic_sched_ready.next;
next             2745 drivers/usb/dwc2/hcd.c 		qh_ptr = qh_ptr->next;
next             2757 drivers/usb/dwc2/hcd.c 	qh_ptr = hsotg->non_periodic_sched_inactive.next;
next             2779 drivers/usb/dwc2/hcd.c 		qh_ptr = qh_ptr->next;
next             2910 drivers/usb/dwc2/hcd.c 	qh_ptr = hsotg->periodic_sched_assigned.next;
next             2922 drivers/usb/dwc2/hcd.c 			qh_ptr = qh_ptr->next;
next             2928 drivers/usb/dwc2/hcd.c 			qh_ptr = qh_ptr->next;
next             2958 drivers/usb/dwc2/hcd.c 			qh_ptr = qh_ptr->next;
next             3042 drivers/usb/dwc2/hcd.c 		hsotg->non_periodic_qh_ptr = hsotg->non_periodic_qh_ptr->next;
next             3061 drivers/usb/dwc2/hcd.c 			goto next;
next             3065 drivers/usb/dwc2/hcd.c 			goto next;
next             3077 drivers/usb/dwc2/hcd.c next:
next             3079 drivers/usb/dwc2/hcd.c 		hsotg->non_periodic_qh_ptr = hsotg->non_periodic_qh_ptr->next;
next             3083 drivers/usb/dwc2/hcd.c 					hsotg->non_periodic_qh_ptr->next;
next              158 drivers/usb/dwc2/hcd_intr.c 	qh_entry = hsotg->periodic_sched_inactive.next;
next              161 drivers/usb/dwc2/hcd_intr.c 		qh_entry = qh_entry->next;
next             1789 drivers/usb/dwc2/hcd_queue.c 					hsotg->non_periodic_qh_ptr->next;
next             3006 drivers/usb/dwc3/gadget.c 	enum dwc3_link_state	next = evtinfo & DWC3_LINK_STATE_MASK;
next             3030 drivers/usb/dwc3/gadget.c 				(next == DWC3_LINK_STATE_RESUME)) {
next             3054 drivers/usb/dwc3/gadget.c 		if (next == DWC3_LINK_STATE_U0) {
next             3081 drivers/usb/dwc3/gadget.c 	switch (next) {
next             3098 drivers/usb/dwc3/gadget.c 	dwc->link_state = next;
next             3104 drivers/usb/dwc3/gadget.c 	enum dwc3_link_state next = evtinfo & DWC3_LINK_STATE_MASK;
next             3106 drivers/usb/dwc3/gadget.c 	if (dwc->link_state != next && next == DWC3_LINK_STATE_U3)
next             3109 drivers/usb/dwc3/gadget.c 	dwc->link_state = next;
next              454 drivers/usb/gadget/composite.c 	void				*next = buf + USB_DT_CONFIG_SIZE;
next              473 drivers/usb/gadget/composite.c 		status = usb_descriptor_fillbuf(next, len,
next              478 drivers/usb/gadget/composite.c 		next += status;
next              488 drivers/usb/gadget/composite.c 		status = usb_descriptor_fillbuf(next, len,
next              493 drivers/usb/gadget/composite.c 		next += status;
next              496 drivers/usb/gadget/composite.c 	len = next - buf;
next              530 drivers/usb/gadget/composite.c 	while ((pos = pos->next) !=  &cdev->configs) {
next             1221 drivers/usb/gadget/composite.c 	int next = cdev->next_string_id;
next             1224 drivers/usb/gadget/composite.c 		if (unlikely(next >= 254))
next             1226 drivers/usb/gadget/composite.c 		str->id = ++next;
next             1229 drivers/usb/gadget/composite.c 	cdev->next_string_id = next;
next             1369 drivers/usb/gadget/composite.c 	unsigned next = c->next_string_id;
next             1370 drivers/usb/gadget/composite.c 	if (unlikely(n > 254 || (unsigned)next + n > 254))
next             1373 drivers/usb/gadget/composite.c 	return next + 1;
next              439 drivers/usb/gadget/function/f_eem.c 					goto next;
next              494 drivers/usb/gadget/function/f_eem.c 				goto next;
next              515 drivers/usb/gadget/function/f_eem.c next:
next              611 drivers/usb/gadget/function/f_hid.c 	struct f_hidg_req_list *list, *next;
next              618 drivers/usb/gadget/function/f_hid.c 	list_for_each_entry_safe(list, next, &hidg->completed_out_req, list) {
next              738 drivers/usb/gadget/function/f_mass_storage.c 		common->next_buffhd_to_fill = bh->next;
next              843 drivers/usb/gadget/function/f_mass_storage.c 			common->next_buffhd_to_fill = bh->next;
next              857 drivers/usb/gadget/function/f_mass_storage.c 		common->next_buffhd_to_drain = bh->next;
next             1471 drivers/usb/gadget/function/f_mass_storage.c 			common->next_buffhd_to_fill = bh2->next;
next             1483 drivers/usb/gadget/function/f_mass_storage.c 		common->next_buffhd_to_drain = bh->next;
next             1536 drivers/usb/gadget/function/f_mass_storage.c 			common->next_buffhd_to_fill = bh->next;
next             1549 drivers/usb/gadget/function/f_mass_storage.c 			common->next_buffhd_to_fill = bh->next;
next             1646 drivers/usb/gadget/function/f_mass_storage.c 	common->next_buffhd_to_fill = bh->next;
next             2620 drivers/usb/gadget/function/f_mass_storage.c 		bh->next = bh + 1;
next             2627 drivers/usb/gadget/function/f_mass_storage.c 	bh->next = buffhds;
next              382 drivers/usb/gadget/function/f_printer.c 		req = container_of(dev->rx_reqs.next,
next              478 drivers/usb/gadget/function/f_printer.c 			req = container_of(dev->rx_buffers.next,
next              594 drivers/usb/gadget/function/f_printer.c 		req = container_of(dev->tx_reqs.next, struct usb_request,
next              859 drivers/usb/gadget/function/f_printer.c 		req = container_of(dev->rx_buffers.next, struct usb_request,
next              866 drivers/usb/gadget/function/f_printer.c 		req = container_of(dev->rx_buffers.next, struct usb_request,
next              873 drivers/usb/gadget/function/f_printer.c 		req = container_of(dev->tx_reqs_active.next,
next             1107 drivers/usb/gadget/function/f_printer.c 		req = container_of(dev->rx_reqs.next, struct usb_request, list);
next             1114 drivers/usb/gadget/function/f_printer.c 		req = container_of(dev->tx_reqs.next, struct usb_request, list);
next             1378 drivers/usb/gadget/function/f_printer.c 		req = container_of(dev->tx_reqs.next, struct usb_request,
next             1388 drivers/usb/gadget/function/f_printer.c 		req = container_of(dev->rx_reqs.next,
next             1395 drivers/usb/gadget/function/f_printer.c 		req = container_of(dev->rx_buffers.next,
next              146 drivers/usb/gadget/function/storage_common.h 	struct fsg_buffhd		*next;
next              370 drivers/usb/gadget/function/u_ether.c 		struct list_head	*next;
next              372 drivers/usb/gadget/function/u_ether.c 		next = req->list.next;
next              376 drivers/usb/gadget/function/u_ether.c 		if (next == list)
next              379 drivers/usb/gadget/function/u_ether.c 		req = container_of(next, struct usb_request, list);
next              246 drivers/usb/gadget/function/u_serial.c 		req = list_entry(pool->next, struct usb_request, list);
next              319 drivers/usb/gadget/function/u_serial.c 		req = list_entry(pool->next, struct usb_request, list);
next              493 drivers/usb/gadget/function/u_serial.c 		req = list_entry(head->next, struct usb_request, list);
next              472 drivers/usb/gadget/udc/amd5536udc.h 	u32	next;
next              464 drivers/usb/gadget/udc/at91_udc.c 		req = list_entry(ep->queue.next, struct at91_request, queue);
next             1010 drivers/usb/gadget/udc/at91_udc.c 		req = list_entry(ep->queue.next,
next             1297 drivers/usb/gadget/udc/at91_udc.c 		req = list_entry(ep0->queue.next, struct at91_request, queue);
next              471 drivers/usb/gadget/udc/atmel_usba_udc.c 	req = list_entry(ep->queue.next, struct usba_request, queue);
next              501 drivers/usb/gadget/udc/atmel_usba_udc.c 		req = list_entry(ep->queue.next,
next              892 drivers/usb/gadget/udc/atmel_usba_udc.c 		if (ep->queue.next == &req->queue) {
next             1443 drivers/usb/gadget/udc/atmel_usba_udc.c 		req = list_entry(ep->queue.next,
next             1636 drivers/usb/gadget/udc/atmel_usba_udc.c 		req = list_entry(ep->queue.next, struct usba_request, queue);
next             1700 drivers/usb/gadget/udc/atmel_usba_udc.c 		req = list_entry(ep->queue.next, struct usba_request, queue);
next              257 drivers/usb/gadget/udc/atmel_usba_udc.h 	dma_addr_t next;
next             1227 drivers/usb/gadget/udc/bcm63xx_udc.c 			struct bcm63xx_req *next;
next             1229 drivers/usb/gadget/udc/bcm63xx_udc.c 			next = list_first_entry(&bep->queue,
next             1231 drivers/usb/gadget/udc/bcm63xx_udc.c 			iudma_write(udc, bep->iudma, next);
next             2094 drivers/usb/gadget/udc/bcm63xx_udc.c 					struct bcm63xx_req *next;
next             2096 drivers/usb/gadget/udc/bcm63xx_udc.c 					next = list_first_entry(&bep->queue,
next             2098 drivers/usb/gadget/udc/bcm63xx_udc.c 					iudma_write(udc, iudma, next);
next              576 drivers/usb/gadget/udc/bdc/bdc_ep.c 		req = list_entry(ep->queue.next, struct bdc_req,
next              976 drivers/usb/gadget/udc/bdc/bdc_ep.c 	req = list_entry(ep->queue.next, struct bdc_req,
next             1097 drivers/usb/gadget/udc/bdc/bdc_ep.c 		req = list_entry(ep->queue.next, struct bdc_req,
next              334 drivers/usb/gadget/udc/dummy_hcd.c 		req = list_entry(ep->queue.next, struct dummy_request, queue);
next              220 drivers/usb/gadget/udc/fotg210-udc.c 		req = list_entry(ep->queue.next,
next              817 drivers/usb/gadget/udc/fotg210-udc.c 		req = list_entry(ep->queue.next,
next              840 drivers/usb/gadget/udc/fotg210-udc.c 	struct fotg210_request *req = list_entry(ep->queue.next,
next              850 drivers/usb/gadget/udc/fotg210-udc.c 	struct fotg210_request *req = list_entry(ep->queue.next,
next              132 drivers/usb/gadget/udc/fsl_qe_udc.c 		req = list_entry(ep->queue.next, struct qe_req, queue);
next              903 drivers/usb/gadget/udc/fsl_qe_udc.c 		req = list_entry(ep->queue.next, struct qe_req, queue);
next             1171 drivers/usb/gadget/udc/fsl_qe_udc.c 			ep->tx_req = list_entry(ep->queue.next,	struct qe_req,
next              218 drivers/usb/gadget/udc/fsl_udc_core.c 		req = list_entry(ep->queue.next, struct fsl_req, queue);
next              957 drivers/usb/gadget/udc/fsl_udc_core.c 	if (ep->queue.next == &req->queue) {
next              962 drivers/usb/gadget/udc/fsl_udc_core.c 		if (req->queue.next != &ep->queue) {
next              965 drivers/usb/gadget/udc/fsl_udc_core.c 			next_req = list_entry(req->queue.next, struct fsl_req,
next              259 drivers/usb/gadget/udc/fusb300_udc.c 		req = list_entry(ep->queue.next, struct fusb300_request, queue);
next              961 drivers/usb/gadget/udc/fusb300_udc.c 	struct fusb300_request *req = list_entry(ep->queue.next,
next              972 drivers/usb/gadget/udc/fusb300_udc.c 	struct fusb300_request *req = list_entry(ep->queue.next,
next             1033 drivers/usb/gadget/udc/fusb300_udc.c 		req = list_entry(ep->queue.next,
next              489 drivers/usb/gadget/udc/goku_udc.c 				req = list_entry(ep->queue.next,
next              524 drivers/usb/gadget/udc/goku_udc.c 	req = list_entry(ep->queue.next, struct goku_request, queue);
next              604 drivers/usb/gadget/udc/goku_udc.c 	req = list_entry(ep->queue.next, struct goku_request, queue);
next              631 drivers/usb/gadget/udc/goku_udc.c 	req = list_entry(ep->queue.next, struct goku_request, queue);
next              649 drivers/usb/gadget/udc/goku_udc.c 	req = list_entry(ep->queue.next, struct goku_request, queue);
next              801 drivers/usb/gadget/udc/goku_udc.c 		req = list_entry(ep->queue.next, struct goku_request, queue);
next              842 drivers/usb/gadget/udc/goku_udc.c 	if (ep->dma && ep->queue.next == &req->queue && !ep->stopped) {
next              870 drivers/usb/gadget/udc/goku_udc.c 			req = list_entry(ep->queue.next, struct goku_request,
next              165 drivers/usb/gadget/udc/gr_udc.c 		struct gr_dma_desc *next;
next              170 drivers/usb/gadget/udc/gr_udc.c 		next = req->first_desc;
next              172 drivers/usb/gadget/udc/gr_udc.c 			desc = next;
next              173 drivers/usb/gadget/udc/gr_udc.c 			next = desc->next_desc;
next              257 drivers/usb/gadget/udc/gr_udc.c 	struct gr_dma_desc *next;
next              259 drivers/usb/gadget/udc/gr_udc.c 	next = req->first_desc;
next              260 drivers/usb/gadget/udc/gr_udc.c 	if (!next)
next              264 drivers/usb/gadget/udc/gr_udc.c 		desc = next;
next              265 drivers/usb/gadget/udc/gr_udc.c 		next = desc->next_desc;
next              457 drivers/usb/gadget/udc/gr_udc.c 		req->last_desc->next = desc->paddr;
next              115 drivers/usb/gadget/udc/gr_udc.h 	u32 next;
next              995 drivers/usb/gadget/udc/lpc32xx_udc.c 	req = list_entry(ep->queue.next, struct lpc32xx_request, queue);
next             1021 drivers/usb/gadget/udc/lpc32xx_udc.c 	req = list_entry(ep->queue.next, struct lpc32xx_request, queue);
next             1441 drivers/usb/gadget/udc/lpc32xx_udc.c 		req = list_entry(ep->queue.next, struct lpc32xx_request, queue);
next             1462 drivers/usb/gadget/udc/lpc32xx_udc.c 		req = list_entry(ep0->queue.next, struct lpc32xx_request,
next             1499 drivers/usb/gadget/udc/lpc32xx_udc.c 		req = list_entry(ep0->queue.next, struct lpc32xx_request,
next             1973 drivers/usb/gadget/udc/lpc32xx_udc.c 	req = list_entry(ep->queue.next, struct lpc32xx_request, queue);
next             2000 drivers/usb/gadget/udc/lpc32xx_udc.c 	req = list_entry(ep->queue.next, struct lpc32xx_request, queue);
next              733 drivers/usb/gadget/udc/m66592-udc.c 		req = list_entry(ep->queue.next, struct m66592_request, queue);
next              907 drivers/usb/gadget/udc/m66592-udc.c 		req = list_entry(ep->queue.next, struct m66592_request, queue);
next              915 drivers/usb/gadget/udc/m66592-udc.c 				req = list_entry(ep->queue.next,
next              938 drivers/usb/gadget/udc/m66592-udc.c 		req = list_entry(ep->queue.next, struct m66592_request, queue);
next              951 drivers/usb/gadget/udc/m66592-udc.c 					req = list_entry(ep->queue.next,
next             1020 drivers/usb/gadget/udc/m66592-udc.c 		req = list_entry(ep->queue.next,
next             1169 drivers/usb/gadget/udc/m66592-udc.c 		req = list_entry(ep->queue.next, struct m66592_request, queue);
next             1318 drivers/usb/gadget/udc/m66592-udc.c 		req = list_entry(ep->queue.next, struct m66592_request, queue);
next              130 drivers/usb/gadget/udc/mv_u3d_core.c 		curr_trb = list_entry(curr_req->trb_list.next,
next              456 drivers/usb/gadget/udc/mv_u3d_core.c 		req->trb_head = list_entry(req->trb_list.next,
next              483 drivers/usb/gadget/udc/mv_u3d_core.c 		req = list_entry(ep->req_list.next, struct mv_u3d_req, list);
next              875 drivers/usb/gadget/udc/mv_u3d_core.c 	if (ep->queue.next == &req->queue) {
next              880 drivers/usb/gadget/udc/mv_u3d_core.c 		if (req->queue.next != &ep->queue) {
next              884 drivers/usb/gadget/udc/mv_u3d_core.c 			next_req = list_entry(req->queue.next,
next              905 drivers/usb/gadget/udc/mv_u3d_core.c 		curr_req = list_entry(ep->req_list.next,
next             1370 drivers/usb/gadget/udc/mv_u3d_core.c 		req = list_entry(ep->queue.next, struct mv_u3d_req, queue);
next             1645 drivers/usb/gadget/udc/mv_u3d_core.c 			req = list_entry(curr_ep->req_list.next,
next              807 drivers/usb/gadget/udc/mv_udc_core.c 	if (ep->queue.next == &req->queue) {
next              812 drivers/usb/gadget/udc/mv_udc_core.c 		if (req->queue.next != &ep->queue) {
next              815 drivers/usb/gadget/udc/mv_udc_core.c 			next_req = list_entry(req->queue.next,
next             1305 drivers/usb/gadget/udc/mv_udc_core.c 		req = list_entry(ep->queue.next, struct mv_req, queue);
next              466 drivers/usb/gadget/udc/net2272.c 				req = list_entry(ep->queue.next,
next              595 drivers/usb/gadget/udc/net2272.c 				req = list_entry(ep->queue.next,
next              617 drivers/usb/gadget/udc/net2272.c 	req = list_entry(ep->queue.next, struct net2272_request, queue);
next              917 drivers/usb/gadget/udc/net2272.c 		req = list_entry(ep->queue.next,
next              953 drivers/usb/gadget/udc/net2272.c 	if (ep->queue.next == &req->queue) {
next             1169 drivers/usb/gadget/udc/net2272.c 	char *next;
next             1177 drivers/usb/gadget/udc/net2272.c 	next = buf;
next             1182 drivers/usb/gadget/udc/net2272.c 	t = scnprintf(next, size, "%s version %s,"
next             1193 drivers/usb/gadget/udc/net2272.c 	next += t;
next             1197 drivers/usb/gadget/udc/net2272.c 	t = scnprintf(next, size, "\ndmareq %02x: %s %s%s%s%s\n",
next             1204 drivers/usb/gadget/udc/net2272.c 	next += t;
next             1217 drivers/usb/gadget/udc/net2272.c 	t = scnprintf(next, size,
next             1222 drivers/usb/gadget/udc/net2272.c 	next += t;
next             1234 drivers/usb/gadget/udc/net2272.c 		t = scnprintf(next, size,
next             1248 drivers/usb/gadget/udc/net2272.c 		next += t;
next             1250 drivers/usb/gadget/udc/net2272.c 		t = scnprintf(next, size,
next             1261 drivers/usb/gadget/udc/net2272.c 		next += t;
next             1263 drivers/usb/gadget/udc/net2272.c 		t = scnprintf(next, size,
next             1269 drivers/usb/gadget/udc/net2272.c 		next += t;
next             1273 drivers/usb/gadget/udc/net2272.c 		t = scnprintf(next, size,
next             1278 drivers/usb/gadget/udc/net2272.c 		next += t;
next             1514 drivers/usb/gadget/udc/net2272.c 		req = list_entry(ep->queue.next,
next             1547 drivers/usb/gadget/udc/net2272.c 			req = list_entry(ep->queue.next,
next             1591 drivers/usb/gadget/udc/net2272.c 		req = list_entry(ep->queue.next,
next             1734 drivers/usb/gadget/udc/net2272.c 			req = list_entry(ep->queue.next,
next             1144 drivers/usb/gadget/udc/net2280.c 		req = list_entry(ep->queue.next,
next             1206 drivers/usb/gadget/udc/net2280.c 	req = list_entry(ep->queue.next, struct net2280_request, queue);
next             1233 drivers/usb/gadget/udc/net2280.c 		req = list_entry(ep->queue.next,
next             1282 drivers/usb/gadget/udc/net2280.c 	if (ep->queue.next == &req->queue) {
next             1287 drivers/usb/gadget/udc/net2280.c 			if (likely(ep->queue.next == &req->queue)) {
next             1314 drivers/usb/gadget/udc/net2280.c 				start_dma(ep, list_entry(ep->queue.next,
next             1658 drivers/usb/gadget/udc/net2280.c 	char			*next;
next             1666 drivers/usb/gadget/udc/net2280.c 	next = buf;
next             1676 drivers/usb/gadget/udc/net2280.c 	t = scnprintf(next, size, "%s version " DRIVER_VERSION
next             1690 drivers/usb/gadget/udc/net2280.c 	next += t;
next             1705 drivers/usb/gadget/udc/net2280.c 	t = scnprintf(next, size,
next             1711 drivers/usb/gadget/udc/net2280.c 	next += t;
next             1727 drivers/usb/gadget/udc/net2280.c 		t = scnprintf(next, size,
next             1749 drivers/usb/gadget/udc/net2280.c 		next += t;
next             1751 drivers/usb/gadget/udc/net2280.c 		t = scnprintf(next, size,
next             1760 drivers/usb/gadget/udc/net2280.c 		next += t;
next             1765 drivers/usb/gadget/udc/net2280.c 		t = scnprintf(next, size,
next             1774 drivers/usb/gadget/udc/net2280.c 		next += t;
next             1781 drivers/usb/gadget/udc/net2280.c 	t = scnprintf(next, size, "\nirqs:  ");
next             1783 drivers/usb/gadget/udc/net2280.c 	next += t;
next             1790 drivers/usb/gadget/udc/net2280.c 		t = scnprintf(next, size, " %s/%lu", ep->ep.name, ep->irqs);
next             1792 drivers/usb/gadget/udc/net2280.c 		next += t;
next             1795 drivers/usb/gadget/udc/net2280.c 	t = scnprintf(next, size, "\n");
next             1797 drivers/usb/gadget/udc/net2280.c 	next += t;
next             1809 drivers/usb/gadget/udc/net2280.c 	char			*next;
next             1815 drivers/usb/gadget/udc/net2280.c 	next = buf;
next             1831 drivers/usb/gadget/udc/net2280.c 			t = scnprintf(next, size,
next             1840 drivers/usb/gadget/udc/net2280.c 			t = scnprintf(next, size, "ep0 max 64 pio %s\n",
next             1845 drivers/usb/gadget/udc/net2280.c 		next += t;
next             1848 drivers/usb/gadget/udc/net2280.c 			t = scnprintf(next, size, "\t(nothing queued)\n");
next             1852 drivers/usb/gadget/udc/net2280.c 			next += t;
next             1857 drivers/usb/gadget/udc/net2280.c 				t = scnprintf(next, size,
next             1864 drivers/usb/gadget/udc/net2280.c 				t = scnprintf(next, size,
next             1871 drivers/usb/gadget/udc/net2280.c 			next += t;
next             1877 drivers/usb/gadget/udc/net2280.c 				t = scnprintf(next, size, "\t    td %08x "
next             1886 drivers/usb/gadget/udc/net2280.c 				next += t;
next             2520 drivers/usb/gadget/udc/net2280.c 		req = list_entry(ep->queue.next,
next             2620 drivers/usb/gadget/udc/net2280.c 				req = list_entry(ep->queue.next,
next             2732 drivers/usb/gadget/udc/net2280.c 				req = list_entry(ep->queue.next,
next             3135 drivers/usb/gadget/udc/net2280.c 			req = list_entry(ep->queue.next,
next              645 drivers/usb/gadget/udc/omap_udc.c 			req = container_of(ep->queue.next,
next              652 drivers/usb/gadget/udc/omap_udc.c 			req = container_of(ep->queue.next,
next              664 drivers/usb/gadget/udc/omap_udc.c 			req = container_of(ep->queue.next,
next              671 drivers/usb/gadget/udc/omap_udc.c 			req = container_of(ep->queue.next,
next              788 drivers/usb/gadget/udc/omap_udc.c 		req = container_of(ep->queue.next, struct omap_req, queue);
next              813 drivers/usb/gadget/udc/omap_udc.c 		req = container_of(ep->queue.next, struct omap_req, queue);
next             1024 drivers/usb/gadget/udc/omap_udc.c 	if (use_dma && ep->dma_channel && ep->queue.next == &req->queue) {
next             1342 drivers/usb/gadget/udc/omap_udc.c 		req = list_entry(ep->queue.next, struct omap_req, queue);
next             1418 drivers/usb/gadget/udc/omap_udc.c 		req = container_of(ep0->queue.next, struct omap_req, queue);
next             1873 drivers/usb/gadget/udc/omap_udc.c 			req = container_of(ep->queue.next,
next             1914 drivers/usb/gadget/udc/omap_udc.c 				req = container_of(ep->queue.next,
next             1948 drivers/usb/gadget/udc/omap_udc.c 				req = container_of(ep->queue.next,
next             1981 drivers/usb/gadget/udc/omap_udc.c 		req = list_entry(ep->queue.next, struct omap_req, queue);
next              224 drivers/usb/gadget/udc/pch_udc.c 	u32 next;
next             1494 drivers/usb/gadget/udc/pch_udc.c 		req = list_entry(ep->queue.next, struct pch_udc_request, queue);
next             1515 drivers/usb/gadget/udc/pch_udc.c 	dma_addr_t addr = (dma_addr_t)td->next;
next             1516 drivers/usb/gadget/udc/pch_udc.c 	td->next = 0x00;
next             1520 drivers/usb/gadget/udc/pch_udc.c 		addr2 = (dma_addr_t)td->next;
next             1569 drivers/usb/gadget/udc/pch_udc.c 		last->next = dma_addr;
next             1574 drivers/usb/gadget/udc/pch_udc.c 	td->next = req->td_data_phys;
next             1660 drivers/usb/gadget/udc/pch_udc.c 		td_data = phys_to_virt(td_data->next);
next             2095 drivers/usb/gadget/udc/pch_udc.c 	req = list_entry(ep->queue.next, struct pch_udc_request, queue);
next             2109 drivers/usb/gadget/udc/pch_udc.c 		td_data = phys_to_virt(td_data->next);
next             2129 drivers/usb/gadget/udc/pch_udc.c 	req = list_entry(ep->queue.next, struct pch_udc_request, queue);
next             2173 drivers/usb/gadget/udc/pch_udc.c 	req = list_entry(ep->queue.next, struct pch_udc_request, queue);
next             2199 drivers/usb/gadget/udc/pch_udc.c 		addr = (dma_addr_t)td->next;
next             2213 drivers/usb/gadget/udc/pch_udc.c 		req = list_entry(ep->queue.next, struct pch_udc_request, queue);
next             2280 drivers/usb/gadget/udc/pch_udc.c 		req = list_entry(ep->queue.next, struct pch_udc_request,
next             2522 drivers/usb/gadget/udc/pch_udc.c 	ep->td_data->next    = ep->td_data_phys;
next              955 drivers/usb/gadget/udc/pxa25x_udc.c 		req = list_entry(ep->queue.next,
next             1642 drivers/usb/gadget/udc/pxa25x_udc.c 		req = list_entry(ep->queue.next, struct pxa25x_request, queue);
next             1862 drivers/usb/gadget/udc/pxa25x_udc.c 			req = list_entry(ep->queue.next,
next              776 drivers/usb/gadget/udc/pxa27x_udc.c 		req = list_entry(ep->queue.next, struct pxa27x_request, queue);
next             1942 drivers/usb/gadget/udc/pxa27x_udc.c 		req = list_entry(ep->queue.next, struct pxa27x_request, queue);
next             2038 drivers/usb/gadget/udc/pxa27x_udc.c 			req = list_entry(ep->queue.next,
next              894 drivers/usb/gadget/udc/r8a66597-udc.c 	return list_entry(ep->queue.next, struct r8a66597_request, queue);
next              273 drivers/usb/gadget/udc/s3c-hsudc.c 		hsreq = list_entry(hsep->queue.next,
next              444 drivers/usb/gadget/udc/s3c-hsudc.c 		hsreq = list_entry(hsep->queue.next,
next              481 drivers/usb/gadget/udc/s3c-hsudc.c 		hsreq = list_entry(hsep->queue.next,
next              526 drivers/usb/gadget/udc/s3c-hsudc.c 		hsreq = list_entry(hsep->queue.next,
next              718 drivers/usb/gadget/udc/s3c-hsudc.c 			hsreq = list_entry(hsep->queue.next,
next              731 drivers/usb/gadget/udc/s3c-hsudc.c 				hsreq = list_entry(hsep->queue.next,
next              260 drivers/usb/gadget/udc/s3c2410_udc.c 		req = list_entry(ep->queue.next, struct s3c2410_request,
next              733 drivers/usb/gadget/udc/s3c2410_udc.c 		req = list_entry(ep->queue.next, struct s3c2410_request, queue);
next              802 drivers/usb/gadget/udc/s3c2410_udc.c 		req = list_entry(ep->queue.next,
next              572 drivers/usb/gadget/udc/snps_udc_core.c 	dma_addr_t addr = (dma_addr_t)td->next;
next              579 drivers/usb/gadget/udc/snps_udc_core.c 		addr_next = (dma_addr_t)td->next;
next              619 drivers/usb/gadget/udc/snps_udc_core.c 		req->td_data->next = req->td_phys;
next              810 drivers/usb/gadget/udc/snps_udc_core.c 						req->td_data->next);
next              813 drivers/usb/gadget/udc/snps_udc_core.c 			td = (struct udc_data_dma *)phys_to_virt(last->next);
next              833 drivers/usb/gadget/udc/snps_udc_core.c 				req->td_data->next = dma_addr;
next              852 drivers/usb/gadget/udc/snps_udc_core.c 				last->next = dma_addr;
next             1006 drivers/usb/gadget/udc/snps_udc_core.c 		td = phys_to_virt(td->next);
next             1023 drivers/usb/gadget/udc/snps_udc_core.c 		td = phys_to_virt(td->next);
next             1247 drivers/usb/gadget/udc/snps_udc_core.c 		req = list_entry(ep->queue.next,
next             1274 drivers/usb/gadget/udc/snps_udc_core.c 	if (ep->queue.next == &req->queue) {
next             2125 drivers/usb/gadget/udc/snps_udc_core.c 		req = list_entry(ep->queue.next,
next             2143 drivers/usb/gadget/udc/snps_udc_core.c 				req = list_entry(ep->queue.next,
next             2220 drivers/usb/gadget/udc/snps_udc_core.c 				req = list_entry(ep->queue.next,
next             2341 drivers/usb/gadget/udc/snps_udc_core.c 			req = list_entry(ep->queue.next,
next             2381 drivers/usb/gadget/udc/snps_udc_core.c 			req = list_entry(ep->queue.next,
next             2707 drivers/usb/gadget/udc/snps_udc_core.c 				req = list_entry(ep->queue.next,
next              379 drivers/usb/host/ehci-dbg.c 	char			*next = *nextp;
next              398 drivers/usb/host/ehci-dbg.c 	temp = scnprintf(next, size,
next              413 drivers/usb/host/ehci-dbg.c 	next += temp;
next              446 drivers/usb/host/ehci-dbg.c 		temp = scnprintf(next, size,
next              456 drivers/usb/host/ehci-dbg.c 		next += temp;
next              461 drivers/usb/host/ehci-dbg.c 	temp = scnprintf(next, size, "\n");
next              463 drivers/usb/host/ehci-dbg.c 	next += temp;
next              467 drivers/usb/host/ehci-dbg.c 	*nextp = next;
next              476 drivers/usb/host/ehci-dbg.c 	char			*next;
next              481 drivers/usb/host/ehci-dbg.c 	next = buf->output_buf;
next              484 drivers/usb/host/ehci-dbg.c 	*next = 0;
next              493 drivers/usb/host/ehci-dbg.c 		qh_lines(ehci, qh, &next, &size);
next              495 drivers/usb/host/ehci-dbg.c 		temp = scnprintf(next, size, "\nunlink =\n");
next              497 drivers/usb/host/ehci-dbg.c 		next += temp;
next              502 drivers/usb/host/ehci-dbg.c 			qh_lines(ehci, qh, &next, &size);
next              516 drivers/usb/host/ehci-dbg.c 	char			*next;
next              523 drivers/usb/host/ehci-dbg.c 	next = buf->output_buf;
next              526 drivers/usb/host/ehci-dbg.c 	*next = 0;
next              531 drivers/usb/host/ehci-dbg.c 	temp = scnprintf(next, size,
next              534 drivers/usb/host/ehci-dbg.c 	next += temp;
next              537 drivers/usb/host/ehci-dbg.c 		temp = scnprintf(next, size,
next              542 drivers/usb/host/ehci-dbg.c 		next += temp;
next              547 drivers/usb/host/ehci-dbg.c 		temp = scnprintf(next, size,
next              552 drivers/usb/host/ehci-dbg.c 		next += temp;
next              555 drivers/usb/host/ehci-dbg.c 		temp = scnprintf(next, size,
next              560 drivers/usb/host/ehci-dbg.c 		next += temp;
next              562 drivers/usb/host/ehci-dbg.c 		temp = scnprintf(next, size,
next              565 drivers/usb/host/ehci-dbg.c 		next += temp;
next              569 drivers/usb/host/ehci-dbg.c 			temp = scnprintf(next, size,
next              574 drivers/usb/host/ehci-dbg.c 			next += temp;
next              577 drivers/usb/host/ehci-dbg.c 			temp = scnprintf(next, size,
next              585 drivers/usb/host/ehci-dbg.c 			next += temp;
next              590 drivers/usb/host/ehci-dbg.c 	return next - buf->output_buf;
next              628 drivers/usb/host/ehci-dbg.c 	char			*next;
next              639 drivers/usb/host/ehci-dbg.c 	next = buf->output_buf;
next              642 drivers/usb/host/ehci-dbg.c 	temp = scnprintf(next, size, "size = %d\n", ehci->periodic_size);
next              644 drivers/usb/host/ehci-dbg.c 	next += temp;
next              657 drivers/usb/host/ehci-dbg.c 		temp = scnprintf(next, size, "%4d: ", i);
next              659 drivers/usb/host/ehci-dbg.c 		next += temp;
next              667 drivers/usb/host/ehci-dbg.c 				temp = scnprintf(next, size, " qh%d-%04x/%p",
next              675 drivers/usb/host/ehci-dbg.c 				next += temp;
next              681 drivers/usb/host/ehci-dbg.c 						temp = scnprintf(next, size,
next              684 drivers/usb/host/ehci-dbg.c 						next += temp;
next              690 drivers/usb/host/ehci-dbg.c 					temp = output_buf_tds_dir(next, ehci,
next              702 drivers/usb/host/ehci-dbg.c 				temp = scnprintf(next, size,
next              709 drivers/usb/host/ehci-dbg.c 				temp = scnprintf(next, size,
next              715 drivers/usb/host/ehci-dbg.c 				temp = scnprintf(next, size,
next              726 drivers/usb/host/ehci-dbg.c 			next += temp;
next              729 drivers/usb/host/ehci-dbg.c 		temp = scnprintf(next, size, "\n");
next              731 drivers/usb/host/ehci-dbg.c 		next += temp;
next              761 drivers/usb/host/ehci-dbg.c 	char			*next, scratch[80];
next              767 drivers/usb/host/ehci-dbg.c 	next = buf->output_buf;
next              773 drivers/usb/host/ehci-dbg.c 		size = scnprintf(next, size,
next              785 drivers/usb/host/ehci-dbg.c 	temp = scnprintf(next, size,
next              794 drivers/usb/host/ehci-dbg.c 	next += temp;
next              810 drivers/usb/host/ehci-dbg.c 				temp = scnprintf(next, size,
next              815 drivers/usb/host/ehci-dbg.c 				next += temp;
next              819 drivers/usb/host/ehci-dbg.c 				temp = scnprintf(next, size,
next              822 drivers/usb/host/ehci-dbg.c 				next += temp;
next              837 drivers/usb/host/ehci-dbg.c 	temp = scnprintf(next, size, "structural params 0x%08x\n", i);
next              839 drivers/usb/host/ehci-dbg.c 	next += temp;
next              842 drivers/usb/host/ehci-dbg.c 	temp = scnprintf(next, size, "capability params 0x%08x\n", i);
next              844 drivers/usb/host/ehci-dbg.c 	next += temp;
next              849 drivers/usb/host/ehci-dbg.c 	temp = scnprintf(next, size, fmt, temp, scratch);
next              851 drivers/usb/host/ehci-dbg.c 	next += temp;
next              855 drivers/usb/host/ehci-dbg.c 	temp = scnprintf(next, size, fmt, temp, scratch);
next              857 drivers/usb/host/ehci-dbg.c 	next += temp;
next              861 drivers/usb/host/ehci-dbg.c 	temp = scnprintf(next, size, fmt, temp, scratch);
next              863 drivers/usb/host/ehci-dbg.c 	next += temp;
next              865 drivers/usb/host/ehci-dbg.c 	temp = scnprintf(next, size, "uframe %04x\n",
next              868 drivers/usb/host/ehci-dbg.c 	next += temp;
next              874 drivers/usb/host/ehci-dbg.c 		temp = scnprintf(next, size, fmt, temp, scratch);
next              876 drivers/usb/host/ehci-dbg.c 		next += temp;
next              878 drivers/usb/host/ehci-dbg.c 			temp = scnprintf(next, size,
next              883 drivers/usb/host/ehci-dbg.c 			next += temp;
next              888 drivers/usb/host/ehci-dbg.c 		temp = scnprintf(next, size, "async unlink qh %p\n",
next              892 drivers/usb/host/ehci-dbg.c 		next += temp;
next              896 drivers/usb/host/ehci-dbg.c 	temp = scnprintf(next, size,
next              901 drivers/usb/host/ehci-dbg.c 	next += temp;
next              903 drivers/usb/host/ehci-dbg.c 	temp = scnprintf(next, size, "complete %ld unlink %ld\n",
next              906 drivers/usb/host/ehci-dbg.c 	next += temp;
next              117 drivers/usb/host/ehci-q.c 	qtd = list_entry(qh->qtd_list.next, struct ehci_qtd, qtd_list);
next              443 drivers/usb/host/ehci-q.c 					qh->qtd_list.next == &qtd->qtd_list &&
next             1043 drivers/usb/host/ehci-q.c 			qtd = list_entry (qtd_list->next, struct ehci_qtd,
next             1118 drivers/usb/host/ehci-q.c 		qtd = list_entry(qtd_list->next, struct ehci_qtd, qtd_list);
next              643 drivers/usb/host/ehci-sched.c 		ehci->qh_scan_next = list_entry(qh->intr_node.next,
next              690 drivers/usb/host/ehci-sched.c 	else if (ehci->intr_unlink.next == &qh->unlink_node) {
next              711 drivers/usb/host/ehci-sched.c 	else if (ehci->intr_unlink_wait.next == &qh->unlink_node) {
next             1486 drivers/usb/host/ehci-sched.c 	u32			now, base, next, start, period, span, now2;
next             1521 drivers/usb/host/ehci-sched.c 			next = start;
next             1536 drivers/usb/host/ehci-sched.c 			} while (start > next && !done);
next             1565 drivers/usb/host/ehci-sched.c 		next = now + ehci->i_thresh;	/* uframe cache */
next             1567 drivers/usb/host/ehci-sched.c 		next = (now + 2 + 7) & ~0x07;	/* full frame cache */
next             1578 drivers/usb/host/ehci-sched.c 	next = (next - base) & (mod - 1);
next             1604 drivers/usb/host/ehci-sched.c 		if (unlikely(start < next &&
next             1645 drivers/usb/host/ehci-sched.c 	start = next + ((start - next) & (period - 1));
next             1784 drivers/usb/host/ehci-sched.c 			itd = list_entry(iso_sched->td_list.next,
next             2187 drivers/usb/host/ehci-sched.c 		sitd = list_entry(sched->td_list.next,
next               42 drivers/usb/host/fhci-mem.c 		td = list_entry(fhci->empty_tds.next, struct td, node);
next               43 drivers/usb/host/fhci-mem.c 		list_del(fhci->empty_tds.next);
next               66 drivers/usb/host/fhci-mem.c 		ed = list_entry(fhci->empty_eds.next, struct ed, node);
next               67 drivers/usb/host/fhci-mem.c 		list_del(fhci->empty_eds.next);
next               75 drivers/usb/host/fhci-q.c 		td = list_entry(ed->td_list.next, struct td, node);
next               87 drivers/usb/host/fhci-q.c 		td = list_entry(frame->tds_list.next, struct td, frame_lh);
next               88 drivers/usb/host/fhci-q.c 		list_del_init(frame->tds_list.next);
next              100 drivers/usb/host/fhci-q.c 		td = list_entry(frame->tds_list.next, struct td, frame_lh);
next              112 drivers/usb/host/fhci-q.c 		td = list_entry(ed->td_list.next, struct td, node);
next              113 drivers/usb/host/fhci-q.c 		list_del_init(ed->td_list.next);
next              117 drivers/usb/host/fhci-q.c 			ed->td_head = list_entry(ed->td_list.next, struct td,
next              132 drivers/usb/host/fhci-q.c 		td = list_entry(p_list->done_list.next, struct td, node);
next              133 drivers/usb/host/fhci-q.c 		list_del_init(p_list->done_list.next);
next              149 drivers/usb/host/fhci-q.c 		ed->td_head = list_entry(ed->td_list.next, struct td, node);
next              174 drivers/usb/host/fhci-q.c 		ed->td_head = list_entry(ed->td_list.next, struct td, node);
next               63 drivers/usb/host/fhci-sched.c 		if (ed->td_list.next->next != &ed->td_list) {
next               65 drivers/usb/host/fhci-sched.c 			    list_entry(ed->td_list.next->next, struct td,
next              257 drivers/usb/host/fhci-sched.c 	struct list_head *node = list->next;
next              361 drivers/usb/host/fotg210-hcd.c 	char *next = *nextp;
next              380 drivers/usb/host/fotg210-hcd.c 	temp = scnprintf(next, size,
next              391 drivers/usb/host/fotg210-hcd.c 	next += temp;
next              407 drivers/usb/host/fotg210-hcd.c 		temp = snprintf(next, size,
next              430 drivers/usb/host/fotg210-hcd.c 		next += temp;
next              435 drivers/usb/host/fotg210-hcd.c 	temp = snprintf(next, size, "\n");
next              440 drivers/usb/host/fotg210-hcd.c 	next += temp;
next              444 drivers/usb/host/fotg210-hcd.c 	*nextp = next;
next              453 drivers/usb/host/fotg210-hcd.c 	char *next;
next              458 drivers/usb/host/fotg210-hcd.c 	next = buf->output_buf;
next              461 drivers/usb/host/fotg210-hcd.c 	*next = 0;
next              470 drivers/usb/host/fotg210-hcd.c 		qh_lines(fotg210, qh, &next, &size);
next              472 drivers/usb/host/fotg210-hcd.c 		temp = scnprintf(next, size, "\nunlink =\n");
next              474 drivers/usb/host/fotg210-hcd.c 		next += temp;
next              478 drivers/usb/host/fotg210-hcd.c 			qh_lines(fotg210, qh, &next, &size);
next              521 drivers/usb/host/fotg210-hcd.c 	char *next;
next              533 drivers/usb/host/fotg210-hcd.c 	next = buf->output_buf;
next              536 drivers/usb/host/fotg210-hcd.c 	temp = scnprintf(next, size, "size = %d\n", fotg210->periodic_size);
next              538 drivers/usb/host/fotg210-hcd.c 	next += temp;
next              551 drivers/usb/host/fotg210-hcd.c 		temp = scnprintf(next, size, "%4d: ", i);
next              553 drivers/usb/host/fotg210-hcd.c 		next += temp;
next              561 drivers/usb/host/fotg210-hcd.c 				temp = scnprintf(next, size, " qh%d-%04x/%p",
next              569 drivers/usb/host/fotg210-hcd.c 				next += temp;
next              575 drivers/usb/host/fotg210-hcd.c 						temp = scnprintf(next, size,
next              578 drivers/usb/host/fotg210-hcd.c 						next += temp;
next              584 drivers/usb/host/fotg210-hcd.c 					temp = output_buf_tds_dir(next,
next              596 drivers/usb/host/fotg210-hcd.c 				temp = scnprintf(next, size,
next              603 drivers/usb/host/fotg210-hcd.c 				temp = scnprintf(next, size,
next              610 drivers/usb/host/fotg210-hcd.c 			next += temp;
next              613 drivers/usb/host/fotg210-hcd.c 		temp = scnprintf(next, size, "\n");
next              615 drivers/usb/host/fotg210-hcd.c 		next += temp;
next              645 drivers/usb/host/fotg210-hcd.c 	char *next, scratch[80];
next              651 drivers/usb/host/fotg210-hcd.c 	next = buf->output_buf;
next              657 drivers/usb/host/fotg210-hcd.c 		size = scnprintf(next, size,
next              670 drivers/usb/host/fotg210-hcd.c 	temp = scnprintf(next, size,
next              679 drivers/usb/host/fotg210-hcd.c 	next += temp;
next              683 drivers/usb/host/fotg210-hcd.c 	temp = scnprintf(next, size, "structural params 0x%08x\n", i);
next              685 drivers/usb/host/fotg210-hcd.c 	next += temp;
next              688 drivers/usb/host/fotg210-hcd.c 	temp = scnprintf(next, size, "capability params 0x%08x\n", i);
next              690 drivers/usb/host/fotg210-hcd.c 	next += temp;
next              695 drivers/usb/host/fotg210-hcd.c 	temp = scnprintf(next, size, fmt, temp, scratch);
next              697 drivers/usb/host/fotg210-hcd.c 	next += temp;
next              701 drivers/usb/host/fotg210-hcd.c 	temp = scnprintf(next, size, fmt, temp, scratch);
next              703 drivers/usb/host/fotg210-hcd.c 	next += temp;
next              707 drivers/usb/host/fotg210-hcd.c 	temp = scnprintf(next, size, fmt, temp, scratch);
next              709 drivers/usb/host/fotg210-hcd.c 	next += temp;
next              711 drivers/usb/host/fotg210-hcd.c 	temp = scnprintf(next, size, "uframe %04x\n",
next              714 drivers/usb/host/fotg210-hcd.c 	next += temp;
next              717 drivers/usb/host/fotg210-hcd.c 		temp = scnprintf(next, size, "async unlink qh %p\n",
next              720 drivers/usb/host/fotg210-hcd.c 		next += temp;
next              724 drivers/usb/host/fotg210-hcd.c 	temp = scnprintf(next, size,
next              729 drivers/usb/host/fotg210-hcd.c 	next += temp;
next              731 drivers/usb/host/fotg210-hcd.c 	temp = scnprintf(next, size, "complete %ld unlink %ld\n",
next              734 drivers/usb/host/fotg210-hcd.c 	next += temp;
next             2075 drivers/usb/host/fotg210-hcd.c 		qtd = list_entry(qh->qtd_list.next,
next             2966 drivers/usb/host/fotg210-hcd.c 			qtd = list_entry(qtd_list->next, struct fotg210_qtd,
next             3035 drivers/usb/host/fotg210-hcd.c 		qtd = list_entry(qtd_list->next, struct fotg210_qtd, qtd_list);
next             3164 drivers/usb/host/fotg210-hcd.c 	struct fotg210_qh *qh, *next;
next             3169 drivers/usb/host/fotg210-hcd.c 	next = fotg210->async->qh_next.qh;
next             3170 drivers/usb/host/fotg210-hcd.c 	while (next) {
next             3171 drivers/usb/host/fotg210-hcd.c 		qh = next;
next             3172 drivers/usb/host/fotg210-hcd.c 		next = qh->qh_next.qh;
next             3595 drivers/usb/host/fotg210-hcd.c 		fotg210->qh_scan_next = list_entry(qh->intr_node.next,
next             4168 drivers/usb/host/fotg210-hcd.c 	u32 now, next, start, period, span;
next             4198 drivers/usb/host/fotg210-hcd.c 			next = now + fotg210->i_thresh;
next             4200 drivers/usb/host/fotg210-hcd.c 			next = now;
next             4206 drivers/usb/host/fotg210-hcd.c 		excess = (stream->next_uframe - period - next) & (mod - 1);
next             4208 drivers/usb/host/fotg210-hcd.c 			start = next + excess - mod + period *
next             4211 drivers/usb/host/fotg210-hcd.c 			start = next + excess + period;
next             4239 drivers/usb/host/fotg210-hcd.c 		next = start;
next             4247 drivers/usb/host/fotg210-hcd.c 		} while (start > next && !done);
next             4383 drivers/usb/host/fotg210-hcd.c 			itd = list_entry(iso_sched->td_list.next,
next              577 drivers/usb/host/imx21-hcd.c 		td = list_entry(ep_priv->td_list.next, struct td, list);
next              242 drivers/usb/host/isp116x-hcd.c 		urb = container_of(ep->hep->urb_list.next,
next              329 drivers/usb/host/isp116x-hcd.c 			prev = &temp->next;
next              331 drivers/usb/host/isp116x-hcd.c 			*prev = ep->next;
next              361 drivers/usb/host/isp116x-hcd.c 		    container_of(ep->hep->urb_list.next, struct urb, urb_list);
next              510 drivers/usb/host/isp116x-hcd.c 			while (last_ep->next)
next              511 drivers/usb/host/isp116x-hcd.c 				last_ep = (last_ep->active = last_ep->next);
next              518 drivers/usb/host/isp116x-hcd.c 		urb = container_of(ep->hep->urb_list.next,
next              565 drivers/usb/host/isp116x-hcd.c 	if ((&isp116x->async)->next != (&isp116x->async)->prev)
next              566 drivers/usb/host/isp116x-hcd.c 		list_move(&isp116x->async, (&isp116x->async)->next);
next              803 drivers/usb/host/isp116x-hcd.c 				prev = &here->next;
next              807 drivers/usb/host/isp116x-hcd.c 				ep->next = here;
next              856 drivers/usb/host/isp116x-hcd.c 	if (ep->hep->urb_list.next == &urb->urb_list)
next              321 drivers/usb/host/isp116x.h 	struct isp116x_ep *next;
next              781 drivers/usb/host/isp1362-hcd.c 	if (isp1362_hcd->async.next != isp1362_hcd->async.prev) {
next              783 drivers/usb/host/isp1362-hcd.c 		list_move(&isp1362_hcd->async, isp1362_hcd->async.next);
next              914 drivers/usb/host/isp1362-hcd.c 		if (epq->active.next == &ep->active)
next             1422 drivers/usb/host/isp1362-hcd.c 		if (ep->hep->urb_list.next == &urb->urb_list) {
next              578 drivers/usb/host/isp1362.h 	container_of(ep->hep->urb_list.next, struct urb, urb_list);	\
next              128 drivers/usb/host/max3421-hcd.c 	struct max3421_hcd *next;
next              806 drivers/usb/host/max3421-hcd.c 	struct urb *urb, *next;
next              813 drivers/usb/host/max3421-hcd.c 		list_for_each_entry_safe(urb, next, &ep->urb_list, urb_list) {
next             1900 drivers/usb/host/max3421-hcd.c 	max3421_hcd->next = max3421_hcd_list;
next             1956 drivers/usb/host/max3421-hcd.c 	for (prev = &max3421_hcd_list; *prev; prev = &(*prev)->next) {
next             1973 drivers/usb/host/max3421-hcd.c 	*prev = max3421_hcd->next;
next               23 drivers/usb/host/ohci-dbg.c #define ohci_dbg_sw(ohci, next, size, format, arg...) \
next               25 drivers/usb/host/ohci-dbg.c 	if (next != NULL) { \
next               27 drivers/usb/host/ohci-dbg.c 		s_len = scnprintf (*next, *size, format, ## arg ); \
next               28 drivers/usb/host/ohci-dbg.c 		*size -= s_len; *next += s_len; \
next               34 drivers/usb/host/ohci-dbg.c #define ohci_dbg_nosw(ohci, next, size, format, arg...) \
next               37 drivers/usb/host/ohci-dbg.c 		s_len = scnprintf(*next, *size, format, ## arg); \
next               38 drivers/usb/host/ohci-dbg.c 		*size -= s_len; *next += s_len; \
next               46 drivers/usb/host/ohci-dbg.c 	char **next,
next               49 drivers/usb/host/ohci-dbg.c 	ohci_dbg_sw (ohci, next, size, "%s 0x%08x%s%s%s%s%s%s%s%s%s\n",
next               68 drivers/usb/host/ohci-dbg.c 	char **next,
next               72 drivers/usb/host/ohci-dbg.c 		ohci_dbg_sw (ohci, next, size, "%s %08x\n", label, value);
next              101 drivers/usb/host/ohci-dbg.c ohci_dump_status (struct ohci_hcd *controller, char **next, unsigned *size)
next              107 drivers/usb/host/ohci-dbg.c 	ohci_dbg_sw (controller, next, size,
next              114 drivers/usb/host/ohci-dbg.c 	ohci_dbg_sw (controller, next, size,
next              129 drivers/usb/host/ohci-dbg.c 	ohci_dbg_sw (controller, next, size,
next              140 drivers/usb/host/ohci-dbg.c 			next, size);
next              143 drivers/usb/host/ohci-dbg.c 			next, size);
next              148 drivers/usb/host/ohci-dbg.c 			next, size);
next              152 drivers/usb/host/ohci-dbg.c 			next, size);
next              155 drivers/usb/host/ohci-dbg.c 			next, size);
next              159 drivers/usb/host/ohci-dbg.c 			next, size);
next              162 drivers/usb/host/ohci-dbg.c 			next, size);
next              165 drivers/usb/host/ohci-dbg.c 			ohci_readl (controller, &regs->donehead), next, size);
next              168 drivers/usb/host/ohci-dbg.c #define dbg_port_sw(hc,num,value,next,size) \
next              169 drivers/usb/host/ohci-dbg.c 	ohci_dbg_sw (hc, next, size, \
next              194 drivers/usb/host/ohci-dbg.c 	char **next,
next              204 drivers/usb/host/ohci-dbg.c 		ohci_dbg_sw (controller, next, size,
next              215 drivers/usb/host/ohci-dbg.c 		ohci_dbg_sw (controller, next, size,
next              222 drivers/usb/host/ohci-dbg.c 		ohci_dbg_sw (controller, next, size,
next              236 drivers/usb/host/ohci-dbg.c 		dbg_port_sw (controller, i, temp, next, size);
next              492 drivers/usb/host/ohci-dbg.c 	char			*next;
next              501 drivers/usb/host/ohci-dbg.c 	next = buf->page;
next              504 drivers/usb/host/ohci-dbg.c 	temp = scnprintf (next, size, "size = %d\n", NUM_INTS);
next              506 drivers/usb/host/ohci-dbg.c 	next += temp;
next              515 drivers/usb/host/ohci-dbg.c 		temp = scnprintf (next, size, "%2d [%3d]:", i, ohci->load [i]);
next              517 drivers/usb/host/ohci-dbg.c 		next += temp;
next              520 drivers/usb/host/ohci-dbg.c 			temp = scnprintf (next, size, " ed%d/%p",
next              523 drivers/usb/host/ohci-dbg.c 			next += temp;
next              539 drivers/usb/host/ohci-dbg.c 				temp = scnprintf (next, size,
next              555 drivers/usb/host/ohci-dbg.c 				next += temp;
next              570 drivers/usb/host/ohci-dbg.c 		temp = scnprintf (next, size, "\n");
next              572 drivers/usb/host/ohci-dbg.c 		next += temp;
next              588 drivers/usb/host/ohci-dbg.c 	char			*next;
next              594 drivers/usb/host/ohci-dbg.c 	next = buf->page;
next              601 drivers/usb/host/ohci-dbg.c 	ohci_dbg_nosw(ohci, &next, &size,
next              611 drivers/usb/host/ohci-dbg.c 		size -= scnprintf (next, size,
next              616 drivers/usb/host/ohci-dbg.c 	ohci_dump_status(ohci, &next, &size);
next              620 drivers/usb/host/ohci-dbg.c 		ohci_dbg_nosw(ohci, &next, &size,
next              625 drivers/usb/host/ohci-dbg.c 	temp = scnprintf (next, size,
next              630 drivers/usb/host/ohci-dbg.c 	next += temp;
next              633 drivers/usb/host/ohci-dbg.c 	temp = scnprintf (next, size, "fmremaining 0x%08x %sFR=0x%04x\n",
next              637 drivers/usb/host/ohci-dbg.c 	next += temp;
next              640 drivers/usb/host/ohci-dbg.c 	temp = scnprintf (next, size, "periodicstart 0x%04x\n",
next              643 drivers/usb/host/ohci-dbg.c 	next += temp;
next              646 drivers/usb/host/ohci-dbg.c 	temp = scnprintf (next, size, "lsthresh 0x%04x\n",
next              649 drivers/usb/host/ohci-dbg.c 	next += temp;
next              651 drivers/usb/host/ohci-dbg.c 	temp = scnprintf (next, size, "hub poll timer %s\n",
next              654 drivers/usb/host/ohci-dbg.c 	next += temp;
next              657 drivers/usb/host/ohci-dbg.c 	ohci_dump_roothub (ohci, 1, &next, &size);
next              256 drivers/usb/host/ohci-hcd.c 		u16	next = ohci_frame_no(ohci) + 1;
next              261 drivers/usb/host/ohci-hcd.c 		if (unlikely(tick_before(frame, next))) {
next              265 drivers/usb/host/ohci-hcd.c 				frame += (next - frame + ed->interval - 1) &
next              281 drivers/usb/host/ohci-hcd.c 						(u16) (next - frame),
next              287 drivers/usb/host/ohci-hcd.c 							next);
next              838 drivers/usb/host/ohci-q.c 	struct list_head	*tmp = td->td_list.next;
next              853 drivers/usb/host/ohci-q.c 		struct td	*next;
next              855 drivers/usb/host/ohci-q.c 		next = list_entry (tmp, struct td, td_list);
next              856 drivers/usb/host/ohci-q.c 		tmp = next->td_list.next;
next              858 drivers/usb/host/ohci-q.c 		if (next->urb != urb)
next              869 drivers/usb/host/ohci-q.c 		list_del(&next->td_list);
next              871 drivers/usb/host/ohci-q.c 		ed->hwHeadP = next->hwNextTD | toggle;
next             1173 drivers/usb/host/ohci-q.c 		td = list_entry(ed->td_list.next, struct td, td_list);
next             1286 drivers/usb/host/oxu210hp-hcd.c 		qtd = list_entry(qh->qtd_list.next,
next             1980 drivers/usb/host/oxu210hp-hcd.c 			qtd = list_entry(qtd_list->next, struct ehci_qtd,
next             2046 drivers/usb/host/oxu210hp-hcd.c 	qtd = list_entry(qtd_list->next, struct ehci_qtd, qtd_list);
next             2086 drivers/usb/host/oxu210hp-hcd.c 	struct ehci_qh *next;
next             2095 drivers/usb/host/oxu210hp-hcd.c 	next = qh->reclaim;
next             2096 drivers/usb/host/oxu210hp-hcd.c 	oxu->reclaim = next;
next             2116 drivers/usb/host/oxu210hp-hcd.c 	if (next) {
next             2118 drivers/usb/host/oxu210hp-hcd.c 		start_unlink_async(oxu, next);
next              837 drivers/usb/host/r8a66597-hcd.c 	struct r8a66597_td *td, *next;
next              844 drivers/usb/host/r8a66597-hcd.c 	list_for_each_entry_safe(td, next, list, queue) {
next              147 drivers/usb/host/r8a66597.h 	return list_entry(r8a66597->pipe_queue[pipenum].next,
next              312 drivers/usb/host/sl811-hcd.c 		sl811->next_periodic = ep->next;
next              317 drivers/usb/host/sl811-hcd.c 			ep = container_of(sl811->async.next,
next              331 drivers/usb/host/sl811-hcd.c 		if (ep->schedule.next == &sl811->async)
next              334 drivers/usb/host/sl811-hcd.c 			sl811->next_async = container_of(ep->schedule.next,
next              344 drivers/usb/host/sl811-hcd.c 	urb = container_of(ep->hep->urb_list.next, struct urb, urb_list);
next              457 drivers/usb/host/sl811-hcd.c 			prev = &temp->next;
next              459 drivers/usb/host/sl811-hcd.c 			*prev = ep->next;
next              467 drivers/usb/host/sl811-hcd.c 		sl811->next_periodic = ep->next;
next              486 drivers/usb/host/sl811-hcd.c 	urb = container_of(ep->hep->urb_list.next, struct urb, urb_list);
next              700 drivers/usb/host/sl811-hcd.c 						->hep->urb_list.next,
next              710 drivers/usb/host/sl811-hcd.c 						->hep->urb_list.next,
next              933 drivers/usb/host/sl811-hcd.c 				prev = &here->next;
next              937 drivers/usb/host/sl811-hcd.c 				ep->next = here;
next              977 drivers/usb/host/sl811-hcd.c 		if (ep->hep->urb_list.next != &urb->urb_list) {
next             1485 drivers/usb/host/sl811-hcd.c 			ep = ep->next;
next              186 drivers/usb/host/sl811.h 	struct sl811h_ep	*next;
next              347 drivers/usb/host/u132-hcd.c 			struct u132_endp *next_endp = list_entry(head->next,
next              521 drivers/usb/host/u132-hcd.c 		struct list_head *next = endp->urb_more.next;
next              522 drivers/usb/host/u132-hcd.c 		struct u132_urbq *urbq = list_entry(next, struct u132_urbq,
next              524 drivers/usb/host/u132-hcd.c 		list_del(next);
next              560 drivers/usb/host/u132-hcd.c 		struct list_head *next = endp->urb_more.next;
next              561 drivers/usb/host/u132-hcd.c 		struct u132_urbq *urbq = list_entry(next, struct u132_urbq,
next              563 drivers/usb/host/u132-hcd.c 		list_del(next);
next             2472 drivers/usb/host/u132-hcd.c 				struct list_head *next = endp->urb_more.next;
next             2473 drivers/usb/host/u132-hcd.c 				struct u132_urbq *urbq = list_entry(next,
next             2475 drivers/usb/host/u132-hcd.c 				list_del(next);
next              220 drivers/usb/host/uhci-debug.c 		struct urb_priv *urbp = list_entry(qh->queue.next,
next              222 drivers/usb/host/uhci-debug.c 		struct uhci_td *td = list_entry(urbp->td_list.next,
next              435 drivers/usb/host/uhci-debug.c 			tmp = tmp->next;
next              507 drivers/usb/host/uhci-debug.c 		tmp = head->next;
next              511 drivers/usb/host/uhci-debug.c 			tmp = tmp->next;
next              198 drivers/usb/host/uhci-q.c 			ntd = list_entry(td->fl_list.next,
next              334 drivers/usb/host/uhci-q.c 	if (qh->queue.next != &urbp->node) {
next              360 drivers/usb/host/uhci-q.c 	td = list_entry(urbp->td_list.next, struct uhci_td, list);
next              383 drivers/usb/host/uhci-q.c 		urbp = list_entry(qh->queue.next, struct urb_priv, node);
next              398 drivers/usb/host/uhci-q.c 		td = list_entry(urbp->td_list.next, struct uhci_td, list);
next              415 drivers/usb/host/uhci-q.c 	pipe = list_entry(qh->queue.next, struct urb_priv, node)->urb->pipe;
next              487 drivers/usb/host/uhci-q.c 		struct urb_priv *urbp = list_entry(qh->queue.next,
next              489 drivers/usb/host/uhci-q.c 		struct uhci_td *td = list_entry(urbp->td_list.next,
next              506 drivers/usb/host/uhci-q.c 		uhci->next_qh = list_entry(qh->node.next, struct uhci_qh,
next              575 drivers/usb/host/uhci-q.c 		uhci->next_qh = list_entry(qh->node.next, struct uhci_qh,
next              591 drivers/usb/host/uhci-q.c 		uhci->next_qh = list_entry(qh->node.next, struct uhci_qh,
next             1213 drivers/usb/host/uhci-q.c 				if (td->list.next != urbp->td_list.prev)
next             1259 drivers/usb/host/uhci-q.c 	unsigned frame, next;
next             1279 drivers/usb/host/uhci-q.c 		next = uhci->frame_number + 10;
next             1283 drivers/usb/host/uhci-q.c 		frame += (next - frame + qh->period - 1) & -qh->period;
next             1289 drivers/usb/host/uhci-q.c 		next = uhci->frame_number + 1;
next             1305 drivers/usb/host/uhci-q.c 		if (!uhci_frame_before_eq(next, frame)) {
next             1309 drivers/usb/host/uhci-q.c 				frame += (next - frame + qh->period - 1) &
next             1316 drivers/usb/host/uhci-q.c 			else if (!uhci_frame_before_eq(next,
next             1323 drivers/usb/host/uhci-q.c 						next);
next             1462 drivers/usb/host/uhci-q.c 	if (qh->queue.next == &urbp->node && !qh->is_stopped) {
next             1535 drivers/usb/host/uhci-q.c 			urbp->node.next != &qh->queue) {
next             1536 drivers/usb/host/uhci-q.c 		struct urb *nurb = list_entry(urbp->node.next,
next             1582 drivers/usb/host/uhci-q.c 		urbp = list_entry(qh->queue.next, struct urb_priv, node);
next             1641 drivers/usb/host/uhci-q.c 		urbp = list_entry(qh->queue.next, struct urb_priv, node);
next             1643 drivers/usb/host/uhci-q.c 			struct uhci_td *td = list_entry(urbp->td_list.next,
next             1691 drivers/usb/host/uhci-q.c 		urbp = list_entry(qh->queue.next, struct urb_priv, node);
next             1692 drivers/usb/host/uhci-q.c 		td = list_entry(urbp->td_list.next, struct uhci_td, list);
next             1762 drivers/usb/host/uhci-q.c 		uhci->next_qh = list_entry(uhci->skelqh[i]->node.next,
next             1765 drivers/usb/host/uhci-q.c 			uhci->next_qh = list_entry(qh->node.next,
next             1772 drivers/usb/host/uhci-q.c 	list_entry(qh->queue.next, struct urb_priv, node));
next              220 drivers/usb/host/xhci-dbgcap.c 	union xhci_trb		*trb, *next;
next              231 drivers/usb/host/xhci-dbgcap.c 	next = ++(ring->enqueue);
next              232 drivers/usb/host/xhci-dbgcap.c 	if (TRB_TYPE_LINK_LE32(next->link.control)) {
next              233 drivers/usb/host/xhci-dbgcap.c 		next->link.control ^= cpu_to_le32(TRB_CYCLE);
next               41 drivers/usb/host/xhci-dbgtty.c 		req = list_entry(pool->next, struct dbc_request, list_pool);
next               78 drivers/usb/host/xhci-dbgtty.c 		req = list_entry(pool->next, struct dbc_request, list_pool);
next              166 drivers/usb/host/xhci-dbgtty.c 		req = list_entry(head->next, struct dbc_request, list_pool);
next              220 drivers/usb/host/xhci-debugfs.c 		seg = seg->next;
next              100 drivers/usb/host/xhci-ext-caps.h 	u32 next;
next              119 drivers/usb/host/xhci-ext-caps.h 		next = XHCI_EXT_CAPS_NEXT(val);
next              120 drivers/usb/host/xhci-ext-caps.h 		offset += next << 2;
next              121 drivers/usb/host/xhci-ext-caps.h 	} while (next);
next               63 drivers/usb/host/xhci-mem.c 	seg->next = NULL;
next               83 drivers/usb/host/xhci-mem.c 	seg = first->next;
next               85 drivers/usb/host/xhci-mem.c 		struct xhci_segment *next = seg->next;
next               87 drivers/usb/host/xhci-mem.c 		seg = next;
next              100 drivers/usb/host/xhci-mem.c 		struct xhci_segment *next, enum xhci_ring_type type)
next              104 drivers/usb/host/xhci-mem.c 	if (!prev || !next)
next              106 drivers/usb/host/xhci-mem.c 	prev->next = next;
next              109 drivers/usb/host/xhci-mem.c 			cpu_to_le64(next->dma);
next              133 drivers/usb/host/xhci-mem.c 	struct xhci_segment *next;
next              138 drivers/usb/host/xhci-mem.c 	next = ring->enq_seg->next;
next              140 drivers/usb/host/xhci-mem.c 	xhci_link_segments(xhci, last, next, ring->type);
next              238 drivers/usb/host/xhci-mem.c 		seg = seg->next;
next              250 drivers/usb/host/xhci-mem.c 		seg = seg->next;
next              266 drivers/usb/host/xhci-mem.c 		seg = seg->next;
next              332 drivers/usb/host/xhci-mem.c 		struct xhci_segment	*next;
next              334 drivers/usb/host/xhci-mem.c 		next = xhci_segment_alloc(xhci, cycle_state, max_packet, flags);
next              335 drivers/usb/host/xhci-mem.c 		if (!next) {
next              338 drivers/usb/host/xhci-mem.c 				next = prev->next;
next              340 drivers/usb/host/xhci-mem.c 				prev = next;
next              344 drivers/usb/host/xhci-mem.c 		xhci_link_segments(xhci, prev, next, type);
next              346 drivers/usb/host/xhci-mem.c 		prev = next;
next              440 drivers/usb/host/xhci-mem.c 		struct xhci_segment *next;
next              442 drivers/usb/host/xhci-mem.c 			next = first->next;
next              446 drivers/usb/host/xhci-mem.c 			first = next;
next              807 drivers/usb/host/xhci-mem.c 	struct xhci_tt_bw_info *tt_info, *next;
next              820 drivers/usb/host/xhci-mem.c 	list_for_each_entry_safe(tt_info, next, tt_list_head, tt_list) {
next              940 drivers/usb/host/xhci-mem.c 	struct xhci_tt_bw_info *tt_info, *next;
next              954 drivers/usb/host/xhci-mem.c 	list_for_each_entry_safe(tt_info, next, tt_list_head, tt_list) {
next             1821 drivers/usb/host/xhci-mem.c 		seg = seg->next;
next             1869 drivers/usb/host/xhci-mem.c 				list_del_init(ep->next);
next               98 drivers/usb/host/xhci-ring.c 	return last_trb_on_seg(seg, trb) && (seg->next == ring->first_seg);
next              145 drivers/usb/host/xhci-ring.c 		*seg = (*seg)->next;
next              166 drivers/usb/host/xhci-ring.c 		ring->deq_seg = ring->deq_seg->next;
next              177 drivers/usb/host/xhci-ring.c 		ring->deq_seg = ring->deq_seg->next;
next              208 drivers/usb/host/xhci-ring.c 	union xhci_trb *next;
next              214 drivers/usb/host/xhci-ring.c 	next = ++(ring->enqueue);
next              217 drivers/usb/host/xhci-ring.c 	while (trb_is_link(next)) {
next              236 drivers/usb/host/xhci-ring.c 			next->link.control &= cpu_to_le32(~TRB_CHAIN);
next              237 drivers/usb/host/xhci-ring.c 			next->link.control |= cpu_to_le32(chain);
next              241 drivers/usb/host/xhci-ring.c 		next->link.control ^= cpu_to_le32(TRB_CYCLE);
next              244 drivers/usb/host/xhci-ring.c 		if (link_trb_toggles_cycle(next))
next              247 drivers/usb/host/xhci-ring.c 		ring->enq_seg = ring->enq_seg->next;
next              249 drivers/usb/host/xhci-ring.c 		next = ring->enqueue;
next             1019 drivers/usb/host/xhci-ring.c 		ep_ring->deq_seg = ep_ring->deq_seg->next;
next             1031 drivers/usb/host/xhci-ring.c 			ep_ring->deq_seg = ep_ring->deq_seg->next;
next             1820 drivers/usb/host/xhci-ring.c 		cur_seg = cur_seg->next;
next             3004 drivers/usb/host/xhci-ring.c 		ep_ring->enq_seg = ep_ring->enq_seg->next;
next               50 drivers/usb/host/xhci.c 		seg = seg->next;
next              860 drivers/usb/host/xhci.c 		seg = seg->next;
next             2372 drivers/usb/host/xhci.c 			ep_entry = bw_table->interval_bw[i].endpoints.next;
next             1525 drivers/usb/host/xhci.h 	struct xhci_segment	*next;
next               31 drivers/usb/image/microtek.h 	struct mts_desc *next;
next              523 drivers/usb/isp1760/isp1760-hcd.c 	qtd = list_entry(qtd->qtd_list.next, typeof(*qtd), qtd_list);
next              850 drivers/usb/isp1760/isp1760-hcd.c 	if (usb_pipeint(list_entry(qh->qtd_list.next, struct isp1760_qtd,
next             1144 drivers/usb/isp1760/isp1760-hcd.c 				qtd = list_entry(qtd->qtd_list.next,
next              395 drivers/usb/mon/mon_main.c 		p = mon_buses.next;
next              315 drivers/usb/mon/mon_text.c 	p = rp->e_list.next;
next              669 drivers/usb/mon/mon_text.c 		p = rp->e_list.next;
next              227 drivers/usb/mtu3/mtu3_qmu.c 	struct qmu_gpd *next;
next              230 drivers/usb/mtu3/mtu3_qmu.c 		next = enq + 1;
next              232 drivers/usb/mtu3/mtu3_qmu.c 		next = ring->start;
next              235 drivers/usb/mtu3/mtu3_qmu.c 	return next == ring->dequeue;
next               66 drivers/usb/musb/cppi_dma.c 		c->freelist = bd->next;
next               75 drivers/usb/musb/cppi_dma.c 	bd->next = c->freelist;
next              445 drivers/usb/musb/cppi_dma.c 	for (bd = rx->head; bd; bd = bd->next)
next              622 drivers/usb/musb/cppi_dma.c 		if (++i < n_bds && bd->next)
next              623 drivers/usb/musb/cppi_dma.c 			bd->hw_next = bd->next->dma;
next              657 drivers/usb/musb/cppi_dma.c 		bd = bd->next;
next              842 drivers/usb/musb/cppi_dma.c 			tail->next = bd;
next              871 drivers/usb/musb/cppi_dma.c 	tail->next = NULL;
next              885 drivers/usb/musb/cppi_dma.c 	for (d = rx->head; d; d = d->next)
next              891 drivers/usb/musb/cppi_dma.c 		tail->next = bd;
next             1016 drivers/usb/musb/cppi_dma.c 	bd = last ? last->next : rx->head;
next             1023 drivers/usb/musb/cppi_dma.c 			i++, bd = bd->next) {
next             1201 drivers/usb/musb/cppi_dma.c 				i++, bd = bd->next) {
next             1528 drivers/usb/musb/cppi_dma.c 			struct cppi_descriptor	*tmp = queue->next;
next               64 drivers/usb/musb/cppi_dma.h 	struct cppi_descriptor *next;
next             1247 drivers/usb/musb/musb_gadget.c 	if (!musb_ep->busy && &request->list == musb_ep->req_list.next) {
next             1292 drivers/usb/musb/musb_gadget.c 	if (musb_ep->req_list.next != &req->list || musb_ep->busy)
next              107 drivers/usb/musb/musb_gadget.h 	return container_of(queue->next, struct musb_request, list);
next               50 drivers/usb/musb/musb_host.h 	return list_entry(q->next, struct musb_qh, ring);
next              126 drivers/usb/musb/musb_host.h 	return list_entry(queue->next, struct urb, urb_list);
next              967 drivers/usb/phy/phy-fsl-usb.c 	char *next = buf;
next              974 drivers/usb/phy/phy-fsl-usb.c 	t = scnprintf(next, size,
next              978 drivers/usb/phy/phy-fsl-usb.c 	next += t;
next              981 drivers/usb/phy/phy-fsl-usb.c 	t = scnprintf(next, size,
next              995 drivers/usb/phy/phy-fsl-usb.c 	next += t;
next              998 drivers/usb/phy/phy-fsl-usb.c 	t = scnprintf(next, size,
next             1002 drivers/usb/phy/phy-fsl-usb.c 	next += t;
next             1005 drivers/usb/phy/phy-fsl-usb.c 	t = scnprintf(next, size,
next             1037 drivers/usb/phy/phy-fsl-usb.c 	next += t;
next              442 drivers/usb/renesas_usbhs/mod_host.c 	struct usbhsh_ep *uep, *next;
next              444 drivers/usb/renesas_usbhs/mod_host.c 	list_for_each_entry_safe(uep, next, &udev->ep_list_head, ep_list)
next              304 drivers/usb/serial/garmin_gps.c 		result = (struct garmin_packet *)garmin_data_p->pktlist.next;
next              320 drivers/usb/serial/garmin_gps.c 		result = (struct garmin_packet *)garmin_data_p->pktlist.next;
next              291 drivers/usb/serial/mos7720.c 	struct list_head *cursor, *next;
next              322 drivers/usb/serial/mos7720.c 	list_for_each_safe(cursor, next, &mos_parport->deferred_urbs)
next               71 drivers/vfio/mdev/mdev_core.c 	list_for_each_entry(parent, &parent_list, next) {
next              195 drivers/vfio/mdev/mdev_core.c 	list_add(&parent->next, &parent_list);
next              236 drivers/vfio/mdev/mdev_core.c 	list_del(&parent->next);
next              258 drivers/vfio/mdev/mdev_core.c 	list_del(&mdev->next);
next              287 drivers/vfio/mdev/mdev_core.c 	list_for_each_entry(tmp, &mdev_list, next) {
next              303 drivers/vfio/mdev/mdev_core.c 	list_add(&mdev->next, &mdev_list);
next              361 drivers/vfio/mdev/mdev_core.c 	list_for_each_entry(tmp, &mdev_list, next) {
next               20 drivers/vfio/mdev/mdev_private.h 	struct list_head next;
next               32 drivers/vfio/mdev/mdev_private.h 	struct list_head next;
next               45 drivers/vfio/mdev/mdev_private.h 	struct list_head next;
next              171 drivers/vfio/mdev/mdev_sysfs.c 						  next) {
next              172 drivers/vfio/mdev/mdev_sysfs.c 				list_del(&ltype->next);
next              177 drivers/vfio/mdev/mdev_sysfs.c 		list_add(&type->next, &parent->type_list);
next              187 drivers/vfio/mdev/mdev_sysfs.c 	list_for_each_entry_safe(type, tmp, &parent->type_list, next) {
next              188 drivers/vfio/mdev/mdev_sysfs.c 		list_del(&type->next);
next              384 drivers/vfio/pci/vfio_pci.c 				 &vdev->ioeventfds_list, next) {
next              386 drivers/vfio/pci/vfio_pci.c 		list_del(&ioeventfd->next);
next             1451 drivers/vfio/pci/vfio_pci_config.c 		u8 next;
next             1459 drivers/vfio/pci/vfio_pci_config.c 					   pos + PCI_CAP_LIST_NEXT, &next);
next             1475 drivers/vfio/pci/vfio_pci_config.c 			*prev = next;
next             1476 drivers/vfio/pci/vfio_pci_config.c 			pos = next;
next             1497 drivers/vfio/pci/vfio_pci_config.c 		pos = next;
next               33 drivers/vfio/pci/vfio_pci_private.h 	struct list_head	next;
next              333 drivers/vfio/pci/vfio_pci_rdwr.c 	list_for_each_entry(ioeventfd, &vdev->ioeventfds_list, next) {
next              338 drivers/vfio/pci/vfio_pci_rdwr.c 				list_del(&ioeventfd->next);
next              378 drivers/vfio/pci/vfio_pci_rdwr.c 	list_add(&ioeventfd->next, &vdev->ioeventfds_list);
next             1819 drivers/vfio/vfio.c 	for (tmp = buf; tmp->next; tmp = buf + tmp->next)
next             1822 drivers/vfio/vfio.c 	tmp->next = caps->size;
next             1834 drivers/vfio/vfio.c 	for (tmp = buf; tmp->next; tmp = buf + tmp->next - offset)
next             1835 drivers/vfio/vfio.c 		tmp->next += offset;
next               43 drivers/vfio/vfio_iommu_spapr_tce.c 	struct list_head next;
next               52 drivers/vfio/vfio_iommu_spapr_tce.c 	struct list_head next;
next               96 drivers/vfio/vfio_iommu_spapr_tce.c 	list_del(&tcemem->next);
next              117 drivers/vfio/vfio_iommu_spapr_tce.c 	list_for_each_entry(tcemem, &container->prereg_list, next) {
next              148 drivers/vfio/vfio_iommu_spapr_tce.c 		list_for_each_entry(tcemem, &container->prereg_list, next) {
next              167 drivers/vfio/vfio_iommu_spapr_tce.c 	list_add(&tcemem->next, &container->prereg_list);
next              279 drivers/vfio/vfio_iommu_spapr_tce.c 			struct tce_iommu_group, next);
next              351 drivers/vfio/vfio_iommu_spapr_tce.c 				struct tce_iommu_group, next);
next              369 drivers/vfio/vfio_iommu_spapr_tce.c 	list_for_each_entry_safe(tcemem, tmtmp, &container->prereg_list, next)
next              658 drivers/vfio/vfio_iommu_spapr_tce.c 			struct tce_iommu_group, next);
next              683 drivers/vfio/vfio_iommu_spapr_tce.c 	list_for_each_entry(tcegrp, &container->group_list, next) {
next              699 drivers/vfio/vfio_iommu_spapr_tce.c 	list_for_each_entry(tcegrp, &container->group_list, next) {
next              723 drivers/vfio/vfio_iommu_spapr_tce.c 	list_for_each_entry(tcegrp, &container->group_list, next) {
next              761 drivers/vfio/vfio_iommu_spapr_tce.c 			struct tce_iommu_group, next);
next              816 drivers/vfio/vfio_iommu_spapr_tce.c 				struct tce_iommu_group, next);
next             1048 drivers/vfio/vfio_iommu_spapr_tce.c 		list_for_each_entry(tcegrp, &container->group_list, next) {
next             1266 drivers/vfio/vfio_iommu_spapr_tce.c 	list_for_each_entry(tcegrp, &container->group_list, next) {
next             1311 drivers/vfio/vfio_iommu_spapr_tce.c 		list_add(&tcegrp->next, &container->group_list);
next             1334 drivers/vfio/vfio_iommu_spapr_tce.c 	list_for_each_entry(tcegrp, &container->group_list, next) {
next             1347 drivers/vfio/vfio_iommu_spapr_tce.c 	list_del(&tcegrp->next);
next               77 drivers/vfio/vfio_iommu_type1.c 	struct list_head	next;
next               97 drivers/vfio/vfio_iommu_type1.c 	struct list_head	next;
next              666 drivers/vfio/vfio_iommu_type1.c 	struct vfio_regions *entry, *next;
next              670 drivers/vfio/vfio_iommu_type1.c 	list_for_each_entry_safe(entry, next, regions, list) {
next              776 drivers/vfio/vfio_iommu_type1.c 				      struct vfio_domain, next);
next              778 drivers/vfio/vfio_iommu_type1.c 	list_for_each_entry_continue(d, &iommu->domain_list, next) {
next              786 drivers/vfio/vfio_iommu_type1.c 		phys_addr_t phys, next;
next              801 drivers/vfio/vfio_iommu_type1.c 			next = iommu_iova_to_phys(domain->domain, iova + len);
next              802 drivers/vfio/vfio_iommu_type1.c 			if (next != phys + len)
next              851 drivers/vfio/vfio_iommu_type1.c 	list_for_each_entry(domain, &iommu->domain_list, next)
next              991 drivers/vfio/vfio_iommu_type1.c 	list_for_each_entry(d, &iommu->domain_list, next) {
next             1003 drivers/vfio/vfio_iommu_type1.c 	list_for_each_entry_continue_reverse(d, &iommu->domain_list, next)
next             1196 drivers/vfio/vfio_iommu_type1.c 	d = list_first_entry(&iommu->domain_list, struct vfio_domain, next);
next             1300 drivers/vfio/vfio_iommu_type1.c 	list_for_each_entry(g, &domain->group_list, next) {
next             1497 drivers/vfio/vfio_iommu_type1.c 	struct vfio_iova *node, *next;
next             1503 drivers/vfio/vfio_iommu_type1.c 	list_for_each_entry_safe(node, next, iova, list) {
next             1516 drivers/vfio/vfio_iommu_type1.c 	list_for_each_entry_safe(node, next, iova, list) {
next             1559 drivers/vfio/vfio_iommu_type1.c 	struct vfio_iova *n, *next;
next             1570 drivers/vfio/vfio_iommu_type1.c 		list_for_each_entry_safe(n, next, iova, list) {
next             1605 drivers/vfio/vfio_iommu_type1.c 	struct iommu_resv_region *n, *next;
next             1607 drivers/vfio/vfio_iommu_type1.c 	list_for_each_entry_safe(n, next, resv_regions, list) {
next             1615 drivers/vfio/vfio_iommu_type1.c 	struct vfio_iova *n, *next;
next             1617 drivers/vfio/vfio_iommu_type1.c 	list_for_each_entry_safe(n, next, iova, list) {
next             1668 drivers/vfio/vfio_iommu_type1.c 	list_for_each_entry(d, &iommu->domain_list, next) {
next             1712 drivers/vfio/vfio_iommu_type1.c 			list_add(&group->next,
next             1780 drivers/vfio/vfio_iommu_type1.c 	list_add(&group->next, &domain->group_list);
next             1802 drivers/vfio/vfio_iommu_type1.c 	list_for_each_entry(d, &iommu->domain_list, next) {
next             1807 drivers/vfio/vfio_iommu_type1.c 				list_add(&group->next, &d->group_list);
next             1832 drivers/vfio/vfio_iommu_type1.c 	list_add(&domain->next, &iommu->domain_list);
next             1919 drivers/vfio/vfio_iommu_type1.c 	list_for_each_entry(domain, &iommu->domain_list, next) {
next             1954 drivers/vfio/vfio_iommu_type1.c 	list_for_each_entry(d, &iommu->domain_list, next) {
next             1955 drivers/vfio/vfio_iommu_type1.c 		list_for_each_entry(g, &d->group_list, next) {
next             1995 drivers/vfio/vfio_iommu_type1.c 			list_del(&group->next);
next             2018 drivers/vfio/vfio_iommu_type1.c 	list_for_each_entry(domain, &iommu->domain_list, next) {
next             2024 drivers/vfio/vfio_iommu_type1.c 		list_del(&group->next);
next             2041 drivers/vfio/vfio_iommu_type1.c 			list_del(&domain->next);
next             2094 drivers/vfio/vfio_iommu_type1.c 				 &domain->group_list, next) {
next             2097 drivers/vfio/vfio_iommu_type1.c 		list_del(&group->next);
next             2119 drivers/vfio/vfio_iommu_type1.c 				 &iommu->domain_list, next) {
next             2121 drivers/vfio/vfio_iommu_type1.c 		list_del(&domain->next);
next             2136 drivers/vfio/vfio_iommu_type1.c 	list_for_each_entry(domain, &iommu->domain_list, next) {
next             2091 drivers/vhost/vhost.c 	unsigned int next;
next             2098 drivers/vhost/vhost.c 	next = vhost16_to_cpu(vq, READ_ONCE(desc->next));
next             2099 drivers/vhost/vhost.c 	return next;
next              172 drivers/vhost/vringh.c 		*up_next = vringh16_to_cpu(vrh, desc->next);
next              378 drivers/vhost/vringh.c 			i = vringh16_to_cpu(vrh, desc.next);
next              458 drivers/video/fbdev/aty/aty128fb.c 	struct fb_info *next;
next              181 drivers/video/fbdev/core/fb_defio.c 	struct list_head *node, *next;
next              197 drivers/video/fbdev/core/fb_defio.c 	list_for_each_safe(node, next, &fbdefio->pagelist) {
next              733 drivers/video/fbdev/core/fbmem.c 	.next	= fb_seq_next,
next             1039 drivers/video/fbdev/core/fbmem.c 	if (info->modelist.prev && info->modelist.next &&
next             1627 drivers/video/fbdev/core/fbmem.c 	if (!fb_info->modelist.prev || !fb_info->modelist.next)
next             1151 drivers/video/fbdev/core/modedb.c 	if (!head->prev || !head->next || list_empty(head))
next             1960 drivers/video/fbdev/matrox/matroxfb_base.c 	for (minfo = matroxfb_l(matroxfb_list.next);
next             1962 drivers/video/fbdev/matrox/matroxfb_base.c 	     minfo = matroxfb_l(minfo->next_fb.next)) {
next             1980 drivers/video/fbdev/matrox/matroxfb_base.c 	for (minfo = matroxfb_l(matroxfb_list.next);
next             1982 drivers/video/fbdev/matrox/matroxfb_base.c 	     minfo = matroxfb_l(minfo->next_fb.next)) {
next             2001 drivers/video/fbdev/matrox/matroxfb_base.c 	for (drv = matroxfb_driver_l(matroxfb_driver_list.next);
next             2003 drivers/video/fbdev/matrox/matroxfb_base.c 	     drv = matroxfb_driver_l(drv->node.next)) {
next              240 drivers/video/fbdev/omap/hwa742.c 	req = list_entry(hwa742.free_req_list.next,
next              275 drivers/video/fbdev/omap/hwa742.c 		req = list_entry(hwa742.pending_req_list.next,
next              229 drivers/video/fbdev/omap2/omapfb/dss/display.c 			dssdev = list_entry(l->next, struct omap_dss_device,
next               79 drivers/video/fbdev/pxa3xx-gcu.c 	struct pxa3xx_gcu_batch *next;
next              192 drivers/video/fbdev/pxa3xx-gcu.c 	struct pxa3xx_gcu_batch *next;
next              195 drivers/video/fbdev/pxa3xx-gcu.c 		next = running->next;
next              196 drivers/video/fbdev/pxa3xx-gcu.c 		running->next = priv->free;
next              198 drivers/video/fbdev/pxa3xx-gcu.c 		running = next;
next              220 drivers/video/fbdev/pxa3xx-gcu.c 		ready = ready->next;
next              408 drivers/video/fbdev/pxa3xx-gcu.c 	priv->free = buffer->next;
next              416 drivers/video/fbdev/pxa3xx-gcu.c 		buffer->next = priv->free;
next              432 drivers/video/fbdev/pxa3xx-gcu.c 	buffer->next = NULL;
next              437 drivers/video/fbdev/pxa3xx-gcu.c 		priv->ready_last->next = buffer;
next              545 drivers/video/fbdev/pxa3xx-gcu.c 	buffer->next = priv->free;
next              555 drivers/video/fbdev/pxa3xx-gcu.c 	struct pxa3xx_gcu_batch *next, *buffer = priv->free;
next              558 drivers/video/fbdev/pxa3xx-gcu.c 		next = buffer->next;
next              564 drivers/video/fbdev/pxa3xx-gcu.c 		buffer = next;
next              251 drivers/video/fbdev/riva/riva_hw.c     int last, next, cur;
next              292 drivers/video/fbdev/riva/riva_hw.c                 next = VIDEO;
next              294 drivers/video/fbdev/riva/riva_hw.c                 next = MPORT;
next              296 drivers/video/fbdev/riva/riva_hw.c                 next = GRAPHICS;
next              302 drivers/video/fbdev/riva/riva_hw.c                         next = VIDEO;
next              304 drivers/video/fbdev/riva/riva_hw.c                         next = GRAPHICS;
next              306 drivers/video/fbdev/riva/riva_hw.c                         next = MPORT;
next              311 drivers/video/fbdev/riva/riva_hw.c                         next = GRAPHICS;
next              313 drivers/video/fbdev/riva/riva_hw.c                         next = VIDEO;
next              315 drivers/video/fbdev/riva/riva_hw.c                         next = MPORT;
next              320 drivers/video/fbdev/riva/riva_hw.c                         next = MPORT;
next              322 drivers/video/fbdev/riva/riva_hw.c                         next = GRAPHICS;
next              324 drivers/video/fbdev/riva/riva_hw.c                         next = VIDEO;
next              329 drivers/video/fbdev/riva/riva_hw.c         cur = next;
next              650 drivers/video/fbdev/sis/sis.h 	struct sis_video_info *next;
next             5867 drivers/video/fbdev/sis/sis_main.c 		while((countvideo = countvideo->next) != NULL)
next             6494 drivers/video/fbdev/sis/sis_main.c 		ivideo->next = card_list;
next             1828 drivers/video/fbdev/smscufx.c 		node = dev->urbs.list.next; /* have reserved one with sem */
next             1919 drivers/video/fbdev/smscufx.c 	entry = dev->urbs.list.next;
next             1013 drivers/video/fbdev/udlfb.c 		struct dlfb_deferred_free *d = list_entry(dlfb->deferred_free.next, struct dlfb_deferred_free, list);
next             1857 drivers/video/fbdev/udlfb.c 		node = dlfb->urbs.list.next; /* have reserved one with sem */
next             1952 drivers/video/fbdev/udlfb.c 	entry = dlfb->urbs.list.next;
next             1086 drivers/video/fbdev/vermilion/vermilion.c 	list = global_no_mode.next;
next             1124 drivers/video/fbdev/vermilion/vermilion.c 		list = global_no_mode.next;
next             1137 drivers/video/fbdev/vermilion/vermilion.c 	struct vml_info *entry, *next;
next             1146 drivers/video/fbdev/vermilion/vermilion.c 	list_for_each_entry_safe(entry, next, &global_has_mode, head) {
next              901 drivers/video/fbdev/via/hw.c 	bool next = true;
next              904 drivers/video/fbdev/via/hw.c 	while (next) {
next              905 drivers/video/fbdev/via/hw.c 		next = false;
next              913 drivers/video/fbdev/via/hw.c 					next = true;
next              224 drivers/virtio/virtio_balloon.c 	struct page *page, *next;
next              226 drivers/virtio/virtio_balloon.c 	list_for_each_entry_safe(page, next, pages, lru) {
next               77 drivers/virtio/virtio_ring.c 	u16 next;			/* The next desc state in a list. */
next              411 drivers/virtio/virtio_ring.c 		desc[i].next = cpu_to_virtio16(_vq->vdev, i + 1);
next              491 drivers/virtio/virtio_ring.c 			i = virtio16_to_cpu(_vq->vdev, desc[i].next);
next              504 drivers/virtio/virtio_ring.c 			i = virtio16_to_cpu(_vq->vdev, desc[i].next);
next              533 drivers/virtio/virtio_ring.c 					vq->split.vring.desc[head].next);
next              579 drivers/virtio/virtio_ring.c 		i = virtio16_to_cpu(_vq->vdev, desc[i].next);
next              634 drivers/virtio/virtio_ring.c 		i = virtio16_to_cpu(vq->vq.vdev, vq->split.vring.desc[i].next);
next              639 drivers/virtio/virtio_ring.c 	vq->split.vring.desc[i].next = cpu_to_virtio16(vq->vq.vdev,
next             1064 drivers/virtio/virtio_ring.c 	vq->free_head = vq->packed.desc_state[id].next;
next             1172 drivers/virtio/virtio_ring.c 			curr = vq->packed.desc_state[curr].next;
next             1293 drivers/virtio/virtio_ring.c 	vq->packed.desc_state[state->last].next = vq->free_head;
next             1302 drivers/virtio/virtio_ring.c 			curr = vq->packed.desc_state[curr].next;
next             1654 drivers/virtio/virtio_ring.c 		vq->packed.desc_state[i].next = i + 1;
next             2123 drivers/virtio/virtio_ring.c 		vq->split.vring.desc[i].next = cpu_to_virtio16(vdev, i + 1);
next               48 drivers/w1/slaves/w1_ds2413.c next:
next               65 drivers/w1/slaves/w1_ds2413.c 			goto next;
next              182 drivers/xen/balloon.c 	page = list_entry(ballooned_pages.next, struct page, lru);
next              198 drivers/xen/balloon.c 	struct list_head *next = page->lru.next;
next              199 drivers/xen/balloon.c 	if (next == &ballooned_pages)
next              201 drivers/xen/balloon.c 	return list_entry(next, struct page, lru);
next              127 drivers/xen/gntalloc.c 	struct gntalloc_gref *gref, *next;
next              167 drivers/xen/gntalloc.c 	list_for_each_entry_safe(gref, next, &queue_file, next_file) {
next              270 drivers/xen/gntalloc.c 		gref = list_entry(priv->list.next,
next              368 drivers/xen/gntalloc.c 			n = list_entry(gref->next_file.next,
next              477 drivers/xen/gntalloc.c 	struct gntalloc_gref *gref, *next;
next              489 drivers/xen/gntalloc.c 			next = list_entry(gref->next_gref.next,
next              493 drivers/xen/gntalloc.c 			gref = next;
next              553 drivers/xen/gntalloc.c 		gref = list_entry(gref->next_file.next,
next               52 drivers/xen/gntdev-common.h 	struct list_head next;
next               36 drivers/xen/gntdev-dmabuf.c 	struct list_head next;
next               64 drivers/xen/gntdev-dmabuf.c 	struct list_head next;
next              111 drivers/xen/gntdev-dmabuf.c 	list_add(&obj->next, &priv->exp_wait_list);
next              122 drivers/xen/gntdev-dmabuf.c 	list_del(&obj->next);
next              142 drivers/xen/gntdev-dmabuf.c 	list_for_each_entry(obj, &priv->exp_wait_list, next)
next              156 drivers/xen/gntdev-dmabuf.c 	list_for_each_entry(gntdev_dmabuf, &priv->exp_list, next)
next              319 drivers/xen/gntdev-dmabuf.c 	list_del(&gntdev_dmabuf->next);
next              328 drivers/xen/gntdev-dmabuf.c 	list_del(&map->next);
next              431 drivers/xen/gntdev-dmabuf.c 	list_add(&gntdev_dmabuf->next, &args->dmabuf_priv->exp_list);
next              680 drivers/xen/gntdev-dmabuf.c 	list_add(&gntdev_dmabuf->next, &priv->imp_list);
next              708 drivers/xen/gntdev-dmabuf.c 	list_for_each_entry_safe(gntdev_dmabuf, q, &priv->imp_list, next) {
next              712 drivers/xen/gntdev-dmabuf.c 			list_del(&gntdev_dmabuf->next);
next              752 drivers/xen/gntdev-dmabuf.c 	list_for_each_entry_safe(gntdev_dmabuf, q, &priv->imp_list, next)
next               87 drivers/xen/gntdev.c 	list_for_each_entry(map, &priv->maps, next)
next              209 drivers/xen/gntdev.c 	list_for_each_entry(map, &priv->maps, next) {
next              211 drivers/xen/gntdev.c 			list_add_tail(&add->next, &map->next);
next              216 drivers/xen/gntdev.c 	list_add_tail(&add->next, &priv->maps);
next              227 drivers/xen/gntdev.c 	list_for_each_entry(map, &priv->maps, next) {
next              254 drivers/xen/gntdev.c 		list_del(&map->next);
next              530 drivers/xen/gntdev.c 	list_for_each_entry(map, &priv->maps, next) {
next              536 drivers/xen/gntdev.c 	list_for_each_entry(map, &priv->freeable_maps, next) {
next              557 drivers/xen/gntdev.c 	list_for_each_entry(map, &priv->maps, next) {
next              566 drivers/xen/gntdev.c 	list_for_each_entry(map, &priv->freeable_maps, next) {
next              642 drivers/xen/gntdev.c 		map = list_entry(priv->maps.next,
next              643 drivers/xen/gntdev.c 				 struct gntdev_grant_map, next);
next              644 drivers/xen/gntdev.c 		list_del(&map->next);
next              716 drivers/xen/gntdev.c 		list_del(&map->next);
next              718 drivers/xen/gntdev.c 			list_add_tail(&map->next, &priv->freeable_maps);
next              791 drivers/xen/gntdev.c 	list_for_each_entry(map, &priv->maps, next) {
next              198 drivers/xen/grant-table.c 	struct gnttab_free_callback *callback, *next;
next              204 drivers/xen/grant-table.c 		next = callback->next;
next              206 drivers/xen/grant-table.c 			callback->next = NULL;
next              209 drivers/xen/grant-table.c 			callback->next = gnttab_free_callback_list;
next              212 drivers/xen/grant-table.c 		callback = next;
next              625 drivers/xen/grant-table.c 		cb = cb->next;
next              631 drivers/xen/grant-table.c 	callback->next = gnttab_free_callback_list;
next              645 drivers/xen/grant-table.c 	for (pcb = &gnttab_free_callback_list; *pcb; pcb = &(*pcb)->next) {
next              647 drivers/xen/grant-table.c 			*pcb = callback->next;
next              115 drivers/xen/mcelog.c 	num = xen_mcelog.next;
next              131 drivers/xen/mcelog.c 	xen_mcelog.next = 0;
next              146 drivers/xen/mcelog.c 	if (xen_mcelog.next)
next              201 drivers/xen/mcelog.c 	entry = xen_mcelog.next;
next              216 drivers/xen/mcelog.c 	xen_mcelog.next++;
next              171 drivers/xen/privcmd.c 			pos = pos->next;
next              205 drivers/xen/privcmd.c 		pos = pos->next;
next              457 drivers/xen/xen-pciback/pci_stub.c 		psdev = container_of(seized_devices.next,
next               27 drivers/xen/xen-pciback/vpci.c 	return head->next;
next               57 drivers/xen/xenbus/xenbus_client.c 	struct list_head next;
next              598 drivers/xen/xenbus/xenbus_client.c 	list_add(&node->next, &xenbus_valloc_pages);
next              718 drivers/xen/xenbus/xenbus_client.c 	list_add(&node->next, &xenbus_valloc_pages);
next              744 drivers/xen/xenbus/xenbus_client.c 	list_for_each_entry(node, &xenbus_valloc_pages, next) {
next              746 drivers/xen/xenbus/xenbus_client.c 			list_del(&node->next);
next              832 drivers/xen/xenbus/xenbus_client.c 	list_for_each_entry(node, &xenbus_valloc_pages, next) {
next              835 drivers/xen/xenbus/xenbus_client.c 			list_del(&node->next);
next              148 drivers/xen/xenbus/xenbus_dev_frontend.c 	rb = list_entry(u->read_buffers.next, struct read_buffer, list);
next              170 drivers/xen/xenbus/xenbus_dev_frontend.c 			rb = list_entry(u->read_buffers.next,
next              220 drivers/xen/xenbus/xenbus_dev_frontend.c 		rb = list_entry(list->next, struct read_buffer, list);
next              221 drivers/xen/xenbus/xenbus_dev_frontend.c 		list_del(list->next);
next              883 drivers/xen/xenbus/xenbus_xs.c 		ent = watch_events.next;
next              106 drivers/xen/xenfs/xensyms.c 	.next = xensyms_next,
next               94 drivers/zorro/proc.c 	.next  = zorro_seq_next,
next              120 fs/affs/amigaffs.h 	__be32 next;
next              587 fs/affs/file.c 			u32 tmp_next = be32_to_cpu(AFFS_DATA_HEAD(prev_bh)->next);
next              593 fs/affs/file.c 			AFFS_DATA_HEAD(prev_bh)->next = cpu_to_be32(bh->b_blocknr);
next              733 fs/affs/file.c 			AFFS_DATA_HEAD(bh)->next = 0;
next              736 fs/affs/file.c 				u32 tmp_next = be32_to_cpu(AFFS_DATA_HEAD(prev_bh)->next);
next              742 fs/affs/file.c 				AFFS_DATA_HEAD(prev_bh)->next = cpu_to_be32(bh->b_blocknr);
next              767 fs/affs/file.c 			AFFS_DATA_HEAD(bh)->next = 0;
next              770 fs/affs/file.c 				u32 tmp_next = be32_to_cpu(AFFS_DATA_HEAD(prev_bh)->next);
next              776 fs/affs/file.c 				AFFS_DATA_HEAD(prev_bh)->next = cpu_to_be32(bh->b_blocknr);
next              927 fs/affs/file.c 			tmp = be32_to_cpu(AFFS_DATA_HEAD(bh)->next);
next              928 fs/affs/file.c 			AFFS_DATA_HEAD(bh)->next = 0;
next               56 fs/afs/callback.c 	for (pp = &server->cb_volumes.first; *pp; pp = &(*pp)->next) {
next               67 fs/afs/callback.c 	new_vi->srv_link.next = *pp;
next               69 fs/afs/callback.c 		(*pp)->pprev = &new_vi->srv_link.next;
next              605 fs/afs/cell.c  	for (p = &net->proc_cells.first; *p; p = &(*p)->next) {
next              612 fs/afs/cell.c  	cell->proc_link.next = *p;
next              613 fs/afs/cell.c  	rcu_assign_pointer(*p, &cell->proc_link.next);
next              614 fs/afs/cell.c  	if (cell->proc_link.next)
next              615 fs/afs/cell.c  		cell->proc_link.next->pprev = &cell->proc_link.next;
next              356 fs/afs/dir.c   	unsigned offset, next, curr;
next              367 fs/afs/dir.c   	     offset = next
next              369 fs/afs/dir.c   		next = offset + 1;
next              378 fs/afs/dir.c   					next * sizeof(union afs_xdr_dirent);
next              395 fs/afs/dir.c   			if (next >= AFS_DIR_SLOTS_PER_BLOCK) {
next              400 fs/afs/dir.c   				       offset, next, tmp, nlen);
next              403 fs/afs/dir.c   			if (!(block->hdr.bitmap[next / 8] &
next              404 fs/afs/dir.c   			      (1 << (next % 8)))) {
next              408 fs/afs/dir.c   				       offset, next, tmp, nlen);
next              414 fs/afs/dir.c   			       next, tmp, nlen);
next              415 fs/afs/dir.c   			next++;
next              432 fs/afs/dir.c   		ctx->pos = blkoff + next * sizeof(union afs_xdr_dirent);
next              112 fs/afs/flock.c 	struct file_lock *p, *_p, *next = NULL;
next              131 fs/afs/flock.c 		if (next &&
next              132 fs/afs/flock.c 		    (next->fl_type == F_WRLCK || p->fl_type == F_RDLCK))
next              134 fs/afs/flock.c 		next = p;
next              140 fs/afs/flock.c 	if (next) {
next              142 fs/afs/flock.c 		next->fl_u.afs.state = AFS_LOCK_YOUR_TRY;
next              143 fs/afs/flock.c 		trace_afs_flock_op(vnode, next, afs_flock_op_wake);
next              144 fs/afs/flock.c 		wake_up(&next->fl_wait);
next              164 fs/afs/flock.c 		p = list_entry(vnode->pending_locks.next,
next              602 fs/afs/flock.c 		ASSERTCMP(vnode->pending_locks.next, ==, &fl->fl_u.afs.link);
next              786 fs/afs/inode.c 		struct afs_wb_key *wbk = list_entry(vnode->wb_keys.next,
next               77 fs/afs/proc.c  	.next	= afs_proc_cells_next,
next              254 fs/afs/proc.c  	.next	= afs_proc_cell_volumes_next,
next              365 fs/afs/proc.c  	.next	= afs_proc_cell_vlservers_next,
next              418 fs/afs/proc.c  	.next	= afs_proc_servers_next,
next              473 fs/afs/proc.c  	.next	= afs_proc_sysname_next,
next              377 fs/afs/write.c 	p = vnode->wb_keys.next;
next              389 fs/afs/write.c 		p = p->next;
next              436 fs/afs/write.c 		p = wbk->vnode_link.next;
next              700 fs/afs/write.c 	pgoff_t start, end, next;
next              708 fs/afs/write.c 		ret = afs_writepages_region(mapping, wbc, start, end, &next);
next              711 fs/afs/write.c 						    &next);
next              712 fs/afs/write.c 		mapping->writeback_index = next;
next              715 fs/afs/write.c 		ret = afs_writepages_region(mapping, wbc, 0, end, &next);
next              717 fs/afs/write.c 			mapping->writeback_index = next;
next              721 fs/afs/write.c 		ret = afs_writepages_region(mapping, wbc, start, end, &next);
next              841 fs/afs/write.c 		wbk = list_entry(graveyard.next, struct afs_wb_key, vnode_link);
next               86 fs/autofs/autofs_i.h 	struct autofs_wait_queue *next;
next              464 fs/autofs/expire.c 			goto next;
next              472 fs/autofs/expire.c next:
next              152 fs/autofs/root.c 			goto next;
next              157 fs/autofs/root.c 			goto next;
next              159 fs/autofs/root.c 			goto next;
next              162 fs/autofs/root.c 			goto next;
next              164 fs/autofs/root.c 			goto next;
next              172 fs/autofs/root.c next:
next              212 fs/autofs/root.c 			goto next;
next              217 fs/autofs/root.c 			goto next;
next              219 fs/autofs/root.c 			goto next;
next              222 fs/autofs/root.c 			goto next;
next              224 fs/autofs/root.c 			goto next;
next              232 fs/autofs/root.c next:
next              678 fs/autofs/root.c 	if (d_child->next == &parent->d_subdirs &&
next               31 fs/autofs/waitq.c 		nwq = wq->next;
next              228 fs/autofs/waitq.c 	for (wq = sbi->queues; wq; wq = wq->next) {
next              429 fs/autofs/waitq.c 		wq->next = sbi->queues;
next              532 fs/autofs/waitq.c 	for (wql = &sbi->queues; (wq = *wql) != NULL; wql = &wq->next) {
next              542 fs/autofs/waitq.c 	*wql = wq->next;	/* Unlink from chain */
next             1662 fs/binfmt_elf.c 	struct elf_thread_core_info *next;
next             1809 fs/binfmt_elf.c 	for (ct = &dump_task->mm->core_state->dumper; ct; ct = ct->next) {
next             1818 fs/binfmt_elf.c 			t->next = info->thread;
next             1825 fs/binfmt_elf.c 			t->next = info->thread->next;
next             1826 fs/binfmt_elf.c 			info->thread->next = t;
next             1833 fs/binfmt_elf.c 	for (t = info->thread; t != NULL; t = t->next)
next             1892 fs/binfmt_elf.c 		t = t->next;
next             1904 fs/binfmt_elf.c 		threads = t->next;
next             2019 fs/binfmt_elf.c 	for (ct = current->mm->core_state->dumper.next;
next             2020 fs/binfmt_elf.c 					ct; ct = ct->next) {
next             2115 fs/binfmt_elf.c 		struct list_head *tmp = info->thread_list.next;
next             1612 fs/binfmt_elf_fdpic.c 	for (ct = current->mm->core_state->dumper.next;
next             1613 fs/binfmt_elf_fdpic.c 					ct; ct = ct->next) {
next             1795 fs/binfmt_elf_fdpic.c 		struct list_head *tmp = thread_list.next;
next              236 fs/btrfs/async-thread.c 		work = list_entry(list->next, struct btrfs_work,
next               23 fs/btrfs/backref.c 	struct extent_inode_elem *next;
next               56 fs/btrfs/backref.c 	e->next = *eie;
next               69 fs/btrfs/backref.c 		eie_next = eie->next;
next              251 fs/btrfs/backref.c 			while (eie && eie->next)
next              252 fs/btrfs/backref.c 				eie = eie->next;
next              257 fs/btrfs/backref.c 				eie->next = newref->inode_list;
next              473 fs/btrfs/backref.c 				goto next;
next              479 fs/btrfs/backref.c 				while (old->next)
next              480 fs/btrfs/backref.c 					old = old->next;
next              481 fs/btrfs/backref.c 				old->next = eie;
next              485 fs/btrfs/backref.c next:
next             1318 fs/btrfs/backref.c 				while (eie->next)
next             1319 fs/btrfs/backref.c 					eie = eie->next;
next             1320 fs/btrfs/backref.c 				eie->next = ref->inode_list;
next             1872 fs/btrfs/backref.c 	for (eie = inode_list; eie; eie = eie->next) {
next              535 fs/btrfs/block-group.c next:
next              566 fs/btrfs/block-group.c 				goto next;
next              587 fs/btrfs/block-group.c 			goto next;
next             1297 fs/btrfs/block-group.c 			goto next;
next             1306 fs/btrfs/block-group.c 			goto next;
next             1318 fs/btrfs/block-group.c 			goto next;
next             1408 fs/btrfs/block-group.c next:
next             1883 fs/btrfs/block-group.c 			goto next;
next             1899 fs/btrfs/block-group.c next:
next             3102 fs/btrfs/block-group.c 		caching_ctl = list_entry(info->caching_block_groups.next,
next             3163 fs/btrfs/block-group.c 		space_info = list_entry(info->space_info.next,
next              466 fs/btrfs/compression.c 			goto next;
next              476 fs/btrfs/compression.c 			goto next;
next              529 fs/btrfs/compression.c next:
next              875 fs/btrfs/compression.c 		ws = wsman->idle_ws.next;
next              909 fs/btrfs/compression.c 		workspace = idle_ws->next;
next              347 fs/btrfs/ctree.c 	struct rb_node *next;
next              379 fs/btrfs/ctree.c 	for (node = rb_first(tm_root); node; node = next) {
next              380 fs/btrfs/ctree.c 		next = rb_next(node);
next             1224 fs/btrfs/ctree.c 	struct rb_node *next;
next             1279 fs/btrfs/ctree.c 		next = rb_next(&tm->node);
next             1280 fs/btrfs/ctree.c 		if (!next)
next             1282 fs/btrfs/ctree.c 		tm = rb_entry(next, struct tree_mod_elem, node);
next             5283 fs/btrfs/ctree.c next:
next             5316 fs/btrfs/ctree.c 			goto next;
next             5326 fs/btrfs/ctree.c 				goto next;
next             5351 fs/btrfs/ctree.c 	struct extent_buffer *next;
next             5365 fs/btrfs/ctree.c 	next = NULL;
next             5430 fs/btrfs/ctree.c 		if (next) {
next             5431 fs/btrfs/ctree.c 			btrfs_tree_unlock_rw(next, next_rw_lock);
next             5432 fs/btrfs/ctree.c 			free_extent_buffer(next);
next             5435 fs/btrfs/ctree.c 		next = c;
next             5437 fs/btrfs/ctree.c 		ret = read_block_for_search(root, path, &next, level,
next             5448 fs/btrfs/ctree.c 			ret = btrfs_try_tree_read_lock(next);
next             5457 fs/btrfs/ctree.c 				free_extent_buffer(next);
next             5464 fs/btrfs/ctree.c 				btrfs_tree_read_lock(next);
next             5478 fs/btrfs/ctree.c 		path->nodes[level] = next;
next             5485 fs/btrfs/ctree.c 		ret = read_block_for_search(root, path, &next, level,
next             5496 fs/btrfs/ctree.c 			ret = btrfs_try_tree_read_lock(next);
next             5499 fs/btrfs/ctree.c 				btrfs_tree_read_lock(next);
next              215 fs/btrfs/delayed-inode.c 	p = delayed_root->node_list.next;
next              229 fs/btrfs/delayed-inode.c 	struct btrfs_delayed_node *next = NULL;
next              237 fs/btrfs/delayed-inode.c 		p = delayed_root->node_list.next;
next              241 fs/btrfs/delayed-inode.c 		p = node->n_list.next;
next              243 fs/btrfs/delayed-inode.c 	next = list_entry(p, struct btrfs_delayed_node, n_list);
next              244 fs/btrfs/delayed-inode.c 	refcount_inc(&next->refs);
next              248 fs/btrfs/delayed-inode.c 	return next;
next              300 fs/btrfs/delayed-inode.c 	p = delayed_root->prepare_list.next;
next              344 fs/btrfs/delayed-inode.c 				struct btrfs_delayed_item **next)
next              377 fs/btrfs/delayed-inode.c 	if (next) {
next              379 fs/btrfs/delayed-inode.c 			*next = NULL;
next              381 fs/btrfs/delayed-inode.c 			*next = delayed_item;
next              383 fs/btrfs/delayed-inode.c 			*next = rb_entry(node, struct btrfs_delayed_item,
next              386 fs/btrfs/delayed-inode.c 			*next = NULL;
next              538 fs/btrfs/delayed-inode.c 	struct btrfs_delayed_item *next = NULL;
next              542 fs/btrfs/delayed-inode.c 		next = rb_entry(p, struct btrfs_delayed_item, rb_node);
next              544 fs/btrfs/delayed-inode.c 	return next;
next              699 fs/btrfs/delayed-inode.c 	struct btrfs_delayed_item *curr, *next;
next              718 fs/btrfs/delayed-inode.c 	next = item;
next              724 fs/btrfs/delayed-inode.c 	while (total_size + next->data_len + sizeof(struct btrfs_item) <=
next              726 fs/btrfs/delayed-inode.c 		total_data_size += next->data_len;
next              727 fs/btrfs/delayed-inode.c 		total_size += next->data_len + sizeof(struct btrfs_item);
next              728 fs/btrfs/delayed-inode.c 		list_add_tail(&next->tree_list, &head);
next              731 fs/btrfs/delayed-inode.c 		curr = next;
next              732 fs/btrfs/delayed-inode.c 		next = __btrfs_next_delayed_item(curr);
next              733 fs/btrfs/delayed-inode.c 		if (!next)
next              736 fs/btrfs/delayed-inode.c 		if (!btrfs_is_continuous_delayed_item(curr, next))
next              766 fs/btrfs/delayed-inode.c 	list_for_each_entry(next, &head, tree_list) {
next              767 fs/btrfs/delayed-inode.c 		keys[i] = next->key;
next              768 fs/btrfs/delayed-inode.c 		data_size[i] = next->data_len;
next              778 fs/btrfs/delayed-inode.c 	list_for_each_entry_safe(curr, next, &head, tree_list) {
next              879 fs/btrfs/delayed-inode.c 	struct btrfs_delayed_item *curr, *next;
next              895 fs/btrfs/delayed-inode.c 	next = item;
next              902 fs/btrfs/delayed-inode.c 	while (btrfs_comp_cpu_keys(&next->key, &key) == 0) {
next              903 fs/btrfs/delayed-inode.c 		list_add_tail(&next->tree_list, &head);
next              906 fs/btrfs/delayed-inode.c 		curr = next;
next              907 fs/btrfs/delayed-inode.c 		next = __btrfs_next_delayed_item(curr);
next              908 fs/btrfs/delayed-inode.c 		if (!next)
next              911 fs/btrfs/delayed-inode.c 		if (!btrfs_is_continuous_delayed_item(curr, next))
next              927 fs/btrfs/delayed-inode.c 	list_for_each_entry_safe(curr, next, &head, tree_list) {
next             1641 fs/btrfs/delayed-inode.c 	struct btrfs_delayed_item *curr, *next;
next             1643 fs/btrfs/delayed-inode.c 	list_for_each_entry_safe(curr, next, ins_list, readdir_list) {
next             1649 fs/btrfs/delayed-inode.c 	list_for_each_entry_safe(curr, next, del_list, readdir_list) {
next             1687 fs/btrfs/delayed-inode.c 	struct btrfs_delayed_item *curr, *next;
next             1702 fs/btrfs/delayed-inode.c 	list_for_each_entry_safe(curr, next, ins_list, readdir_list) {
next              436 fs/btrfs/delayed-ref.c 	struct btrfs_delayed_ref_node *next;
next              443 fs/btrfs/delayed-ref.c 		next = rb_entry(node, struct btrfs_delayed_ref_node, ref_node);
next              445 fs/btrfs/delayed-ref.c 		if (seq && next->seq >= seq)
next              447 fs/btrfs/delayed-ref.c 		if (comp_refs(ref, next, false))
next              450 fs/btrfs/delayed-ref.c 		if (ref->action == next->action) {
next              451 fs/btrfs/delayed-ref.c 			mod = next->ref_mod;
next              453 fs/btrfs/delayed-ref.c 			if (ref->ref_mod < next->ref_mod) {
next              454 fs/btrfs/delayed-ref.c 				swap(ref, next);
next              457 fs/btrfs/delayed-ref.c 			mod = -next->ref_mod;
next              460 fs/btrfs/delayed-ref.c 		drop_delayed_ref(trans, delayed_refs, head, next);
next             2040 fs/btrfs/disk-io.c 		gang[0] = list_entry(fs_info->dead_roots.next,
next             2727 fs/btrfs/extent-tree.c 	struct btrfs_caching_control *next;
next             2733 fs/btrfs/extent-tree.c 	list_for_each_entry_safe(caching_ctl, next,
next             4804 fs/btrfs/extent-tree.c 	struct extent_buffer *next;
next             4827 fs/btrfs/extent-tree.c 	next = find_extent_buffer(fs_info, bytenr);
next             4828 fs/btrfs/extent-tree.c 	if (!next) {
next             4829 fs/btrfs/extent-tree.c 		next = btrfs_find_create_tree_block(fs_info, bytenr);
next             4830 fs/btrfs/extent-tree.c 		if (IS_ERR(next))
next             4831 fs/btrfs/extent-tree.c 			return PTR_ERR(next);
next             4833 fs/btrfs/extent-tree.c 		btrfs_set_buffer_lockdep_class(root->root_key.objectid, next,
next             4837 fs/btrfs/extent-tree.c 	btrfs_tree_lock(next);
next             4838 fs/btrfs/extent-tree.c 	btrfs_set_lock_blocking_write(next);
next             4879 fs/btrfs/extent-tree.c 	if (!btrfs_buffer_uptodate(next, generation, 0)) {
next             4880 fs/btrfs/extent-tree.c 		btrfs_tree_unlock(next);
next             4881 fs/btrfs/extent-tree.c 		free_extent_buffer(next);
next             4882 fs/btrfs/extent-tree.c 		next = NULL;
next             4886 fs/btrfs/extent-tree.c 	if (!next) {
next             4889 fs/btrfs/extent-tree.c 		next = read_tree_block(fs_info, bytenr, generation, level - 1,
next             4891 fs/btrfs/extent-tree.c 		if (IS_ERR(next)) {
next             4892 fs/btrfs/extent-tree.c 			return PTR_ERR(next);
next             4893 fs/btrfs/extent-tree.c 		} else if (!extent_buffer_uptodate(next)) {
next             4894 fs/btrfs/extent-tree.c 			free_extent_buffer(next);
next             4897 fs/btrfs/extent-tree.c 		btrfs_tree_lock(next);
next             4898 fs/btrfs/extent-tree.c 		btrfs_set_lock_blocking_write(next);
next             4902 fs/btrfs/extent-tree.c 	ASSERT(level == btrfs_header_level(next));
next             4903 fs/btrfs/extent-tree.c 	if (level != btrfs_header_level(next)) {
next             4908 fs/btrfs/extent-tree.c 	path->nodes[level] = next;
next             4958 fs/btrfs/extent-tree.c 			ret = btrfs_qgroup_trace_subtree(trans, next,
next             4988 fs/btrfs/extent-tree.c 	btrfs_tree_unlock(next);
next             4989 fs/btrfs/extent-tree.c 	free_extent_buffer(next);
next               69 fs/btrfs/extent_io.c 		state = list_entry(states.next, struct extent_state, leak_list);
next               79 fs/btrfs/extent_io.c 		eb = list_entry(buffers.next, struct extent_buffer, leak_list);
next              438 fs/btrfs/extent_io.c 	struct rb_node *next= NULL;
next              441 fs/btrfs/extent_io.c 	ret = __etree_search(tree, offset, &next, NULL, p_ret, parent_ret);
next              443 fs/btrfs/extent_io.c 		return next;
next              586 fs/btrfs/extent_io.c 	struct rb_node *next = rb_next(&state->rb_node);
next              587 fs/btrfs/extent_io.c 	if (next)
next              588 fs/btrfs/extent_io.c 		return rb_entry(next, struct extent_state, rb_node);
next              605 fs/btrfs/extent_io.c 	struct extent_state *next;
next              624 fs/btrfs/extent_io.c 		next = next_state(state);
next              634 fs/btrfs/extent_io.c 		next = next_state(state);
next              636 fs/btrfs/extent_io.c 	return next;
next              740 fs/btrfs/extent_io.c 		goto next;
next              772 fs/btrfs/extent_io.c 			goto next;
next              799 fs/btrfs/extent_io.c next:
next             1579 fs/btrfs/extent_io.c 	struct rb_node *node, *prev = NULL, *next;
next             1585 fs/btrfs/extent_io.c 		node = __etree_search(tree, start, &next, &prev, NULL, NULL);
next             1586 fs/btrfs/extent_io.c 		if (!node && !next && !prev) {
next             1594 fs/btrfs/extent_io.c 		} else if (!node && !next) {
next             1604 fs/btrfs/extent_io.c 			node = next;
next             2353 fs/btrfs/extent_io.c 	struct extent_state *state, *next;
next             2366 fs/btrfs/extent_io.c 		next = next_state(state);
next             2372 fs/btrfs/extent_io.c 		state = next;
next              190 fs/btrfs/extent_map.c static int mergable_maps(struct extent_map *prev, struct extent_map *next)
next              203 fs/btrfs/extent_map.c 	    test_bit(EXTENT_FLAG_LOGGING, &next->flags))
next              211 fs/btrfs/extent_map.c 	if (!list_empty(&prev->list) || !list_empty(&next->list))
next              214 fs/btrfs/extent_map.c 	ASSERT(next->block_start != EXTENT_MAP_DELALLOC &&
next              217 fs/btrfs/extent_map.c 	if (extent_map_end(prev) == next->start &&
next              218 fs/btrfs/extent_map.c 	    prev->flags == next->flags &&
next              219 fs/btrfs/extent_map.c 	    prev->bdev == next->bdev &&
next              220 fs/btrfs/extent_map.c 	    ((next->block_start == EXTENT_MAP_HOLE &&
next              222 fs/btrfs/extent_map.c 	     (next->block_start == EXTENT_MAP_INLINE &&
next              224 fs/btrfs/extent_map.c 	     (next->block_start < EXTENT_MAP_LAST_BYTE - 1 &&
next              225 fs/btrfs/extent_map.c 	      next->block_start == extent_map_block_end(prev)))) {
next              420 fs/btrfs/extent_map.c 	struct rb_node *next = NULL;
next              423 fs/btrfs/extent_map.c 	rb_node = __tree_search(&tree->map.rb_root, start, &prev, &next);
next              427 fs/btrfs/extent_map.c 		else if (next)
next              428 fs/btrfs/extent_map.c 			rb_node = next;
next              512 fs/btrfs/extent_map.c 	struct rb_node *next;
next              514 fs/btrfs/extent_map.c 	next = rb_next(&em->rb_node);
next              515 fs/btrfs/extent_map.c 	if (!next)
next              517 fs/btrfs/extent_map.c 	return container_of(next, struct extent_map, rb_node);
next              542 fs/btrfs/extent_map.c 	struct extent_map *next;
next              550 fs/btrfs/extent_map.c 		next = existing;
next              551 fs/btrfs/extent_map.c 		prev = prev_extent_map(next);
next              554 fs/btrfs/extent_map.c 		next = next_extent_map(prev);
next              559 fs/btrfs/extent_map.c 	end = next ? next->start : extent_map_end(em);
next              214 fs/btrfs/file-item.c 			goto next;
next              276 fs/btrfs/file-item.c next:
next              413 fs/btrfs/file-item.c 		sums = list_entry(tmplist.next, struct btrfs_ordered_sum, list);
next              486 fs/btrfs/file.c 			goto next;
next              498 fs/btrfs/file.c next:
next              644 fs/btrfs/file.c 			goto next;
next              722 fs/btrfs/file.c next:
next              647 fs/btrfs/free-space-cache.c 			goto next;
next              649 fs/btrfs/free-space-cache.c 			goto next;
next              660 fs/btrfs/free-space-cache.c next:
next              934 fs/btrfs/free-space-cache.c 		cluster = list_entry(block_group->cluster_list.next,
next             1103 fs/btrfs/free-space-cache.c 	struct btrfs_free_space *entry, *next;
next             1107 fs/btrfs/free-space-cache.c 	list_for_each_entry_safe(entry, next, bitmap_list, list) {
next             1132 fs/btrfs/free-space-cache.c 	struct btrfs_free_space *entry, *next;
next             1134 fs/btrfs/free-space-cache.c 	list_for_each_entry_safe(entry, next, bitmap_list, list)
next             1933 fs/btrfs/free-space-cache.c 		struct rb_node *next = rb_next(&bitmap_info->offset_index);
next             1941 fs/btrfs/free-space-cache.c 		if (!next)
next             1944 fs/btrfs/free-space-cache.c 		bitmap_info = rb_entry(next, struct btrfs_free_space,
next             2076 fs/btrfs/free-space-cache.c 		cluster = list_entry(block_group->cluster_list.next,
next             2614 fs/btrfs/free-space-cache.c 	while ((head = block_group->cluster_list.next) !=
next             3250 fs/btrfs/free-space-cache.c 			goto next;
next             3266 fs/btrfs/free-space-cache.c next:
next             3308 fs/btrfs/free-space-cache.c 			goto next;
next             3317 fs/btrfs/free-space-cache.c 			goto next;
next             3324 fs/btrfs/free-space-cache.c 			goto next;
next             3341 fs/btrfs/free-space-cache.c next:
next              103 fs/btrfs/inode-map.c 			goto next;
next              115 fs/btrfs/inode-map.c next:
next              767 fs/btrfs/inode.c 		async_extent = list_entry(async_chunk->extents.next,
next             1316 fs/btrfs/inode.c 		sums = list_entry(list.next, struct btrfs_ordered_sum, list);
next             3049 fs/btrfs/inode.c 			goto next;
next             3053 fs/btrfs/inode.c 			goto next;
next             3071 fs/btrfs/inode.c next:
next             5255 fs/btrfs/inode.c 				goto next;
next             5282 fs/btrfs/inode.c next:
next             6155 fs/btrfs/inode.c 			goto next;
next             6157 fs/btrfs/inode.c 			goto next;
next             6185 fs/btrfs/inode.c next:
next             7102 fs/btrfs/inode.c 		goto next;
next             7133 fs/btrfs/inode.c next:
next             7153 fs/btrfs/inode.c 			goto next;
next             8133 fs/btrfs/inode.c 			goto next;
next             8143 fs/btrfs/inode.c next:
next             8237 fs/btrfs/inode.c 				goto next;
next             8250 fs/btrfs/inode.c 			goto next;
next             8259 fs/btrfs/inode.c next:
next             10227 fs/btrfs/inode.c 	struct btrfs_delalloc_work *work, *next;
next             10239 fs/btrfs/inode.c 		binode = list_entry(splice.next, struct btrfs_inode,
next             10272 fs/btrfs/inode.c 	list_for_each_entry_safe(work, next, &works, list) {
next             10547 fs/btrfs/inode.c 			goto next;
next             10572 fs/btrfs/inode.c next:
next             10777 fs/btrfs/inode.c 	struct rb_node *node, *next;
next             10782 fs/btrfs/inode.c 		next = rb_next(node);
next             10790 fs/btrfs/inode.c 		node = next;
next             1144 fs/btrfs/ioctl.c 	struct extent_map *next;
next             1151 fs/btrfs/ioctl.c 	next = defrag_lookup_extent(inode, em->start + em->len);
next             1152 fs/btrfs/ioctl.c 	if (!next || next->block_start >= EXTENT_MAP_LAST_BYTE)
next             1154 fs/btrfs/ioctl.c 	else if ((em->block_start + em->block_len == next->block_start) &&
next             1155 fs/btrfs/ioctl.c 		 (em->block_len > SZ_128K && next->block_len > SZ_128K))
next             1158 fs/btrfs/ioctl.c 	free_extent_map(next);
next             1497 fs/btrfs/ioctl.c 			unsigned long next;
next             1502 fs/btrfs/ioctl.c 			next = DIV_ROUND_UP(skip, PAGE_SIZE);
next             1503 fs/btrfs/ioctl.c 			i = max(i + 1, next);
next              447 fs/btrfs/ordered-data.c 			cur = entry->list.next;
next              526 fs/btrfs/ordered-data.c 	struct btrfs_ordered_extent *ordered, *next;
next              564 fs/btrfs/ordered-data.c 	list_for_each_entry_safe(ordered, next, &works, work_list) {
next              357 fs/btrfs/print-tree.c 		struct extent_buffer *next;
next              360 fs/btrfs/print-tree.c 		next = read_tree_block(fs_info, btrfs_node_blockptr(c, i),
next              363 fs/btrfs/print-tree.c 		if (IS_ERR(next)) {
next              365 fs/btrfs/print-tree.c 		} else if (!extent_buffer_uptodate(next)) {
next              366 fs/btrfs/print-tree.c 			free_extent_buffer(next);
next              370 fs/btrfs/print-tree.c 		if (btrfs_is_leaf(next) &&
next              373 fs/btrfs/print-tree.c 		if (btrfs_header_level(next) !=
next              376 fs/btrfs/print-tree.c 		btrfs_print_tree(next, follow);
next              377 fs/btrfs/print-tree.c 		free_extent_buffer(next);
next             3815 fs/btrfs/qgroup.c 		struct btrfs_qgroup_swapped_block *next;
next             3817 fs/btrfs/qgroup.c 		rbtree_postorder_for_each_entry_safe(entry, next, cur_root,
next             4032 fs/btrfs/qgroup.c 	struct btrfs_qgroup_extent_record *next;
next             4036 fs/btrfs/qgroup.c 	rbtree_postorder_for_each_entry_safe(entry, next, root, node) {
next              433 fs/btrfs/raid56.c 		rbio = list_entry(table->stripe_cache.next,
next              797 fs/btrfs/raid56.c 			struct btrfs_raid_bio *next;
next              798 fs/btrfs/raid56.c 			struct list_head *head = rbio->plug_list.next;
next              800 fs/btrfs/raid56.c 			next = list_entry(head, struct btrfs_raid_bio,
next              805 fs/btrfs/raid56.c 			list_add(&next->hash_list, &h->hash_list);
next              806 fs/btrfs/raid56.c 			refcount_inc(&next->refs);
next              810 fs/btrfs/raid56.c 			if (next->operation == BTRFS_RBIO_READ_REBUILD)
next              811 fs/btrfs/raid56.c 				start_async_work(next, read_rebuild_work);
next              812 fs/btrfs/raid56.c 			else if (next->operation == BTRFS_RBIO_REBUILD_MISSING) {
next              813 fs/btrfs/raid56.c 				steal_rbio(rbio, next);
next              814 fs/btrfs/raid56.c 				start_async_work(next, read_rebuild_work);
next              815 fs/btrfs/raid56.c 			} else if (next->operation == BTRFS_RBIO_WRITE) {
next              816 fs/btrfs/raid56.c 				steal_rbio(rbio, next);
next              817 fs/btrfs/raid56.c 				start_async_work(next, rmw_work);
next              818 fs/btrfs/raid56.c 			} else if (next->operation == BTRFS_RBIO_PARITY_SCRUB) {
next              819 fs/btrfs/raid56.c 				steal_rbio(rbio, next);
next              820 fs/btrfs/raid56.c 				start_async_work(next, scrub_parity_work);
next              859 fs/btrfs/raid56.c 	struct bio *next;
next              862 fs/btrfs/raid56.c 		next = cur->bi_next;
next              866 fs/btrfs/raid56.c 		cur = next;
next             1700 fs/btrfs/raid56.c 		cur = list_entry(plug->rbio_list.next,
next              217 fs/btrfs/relocation.c 		node = list_entry(cache->detached.next,
next              223 fs/btrfs/relocation.c 		node = list_entry(cache->leaves.next,
next              348 fs/btrfs/relocation.c 		edge = list_entry(node->upper.next,
next              375 fs/btrfs/relocation.c 		edge = list_entry(edge->list[LOWER].next,
next              429 fs/btrfs/relocation.c 		edge = list_entry(node->upper.next, struct backref_edge,
next              490 fs/btrfs/relocation.c 		node = list_entry(cache->detached.next,
next              496 fs/btrfs/relocation.c 		node = list_entry(cache->changed.next,
next              754 fs/btrfs/relocation.c 		edge = list_entry(cur->upper.next, struct backref_edge,
next              795 fs/btrfs/relocation.c 					goto next;
next              827 fs/btrfs/relocation.c 			goto next;
next              873 fs/btrfs/relocation.c 			goto next;
next              881 fs/btrfs/relocation.c 			goto next;
next             1015 fs/btrfs/relocation.c next:
next             1034 fs/btrfs/relocation.c 		edge = list_entry(list.next, struct backref_edge, list[UPPER]);
next             1058 fs/btrfs/relocation.c 		edge = list_entry(list.next, struct backref_edge, list[UPPER]);
next             1115 fs/btrfs/relocation.c 		upper = list_entry(useless.next, struct backref_node, list);
next             1125 fs/btrfs/relocation.c 			edge = list_entry(upper->lower.next,
next             1149 fs/btrfs/relocation.c 			lower = list_entry(useless.next,
next             1181 fs/btrfs/relocation.c 			lower = list_entry(useless.next,
next             1278 fs/btrfs/relocation.c 		new_edge = list_entry(new_node->lower.next,
next             2225 fs/btrfs/relocation.c 	struct btrfs_root *next;
next             2229 fs/btrfs/relocation.c 	list_for_each_entry_safe(root, next, &rc->dirty_subvol_roots,
next             2457 fs/btrfs/relocation.c 		reloc_root = list_entry(rc->reloc_roots.next,
next             2491 fs/btrfs/relocation.c 		reloc_root = list_entry(list->next, struct btrfs_root,
next             2525 fs/btrfs/relocation.c 		reloc_root = list_entry(reloc_roots.next,
next             2617 fs/btrfs/relocation.c 	struct backref_node *next;
next             2621 fs/btrfs/relocation.c 	next = node;
next             2624 fs/btrfs/relocation.c 		next = walk_up_backref(next, edges, &index);
next             2625 fs/btrfs/relocation.c 		root = next->root;
next             2637 fs/btrfs/relocation.c 		if (next->new_bytenr != root->node->start) {
next             2638 fs/btrfs/relocation.c 			BUG_ON(next->new_bytenr);
next             2639 fs/btrfs/relocation.c 			BUG_ON(!list_empty(&next->list));
next             2640 fs/btrfs/relocation.c 			next->new_bytenr = root->node->start;
next             2641 fs/btrfs/relocation.c 			next->root = root;
next             2642 fs/btrfs/relocation.c 			list_add_tail(&next->list,
next             2644 fs/btrfs/relocation.c 			__mark_block_processed(rc, next);
next             2650 fs/btrfs/relocation.c 		next = walk_down_backref(edges, &index);
next             2651 fs/btrfs/relocation.c 		if (!next || next->level <= node->level)
next             2657 fs/btrfs/relocation.c 	next = node;
next             2660 fs/btrfs/relocation.c 		rc->backref_cache.path[next->level] = next;
next             2663 fs/btrfs/relocation.c 		next = edges[index]->node[UPPER];
next             2677 fs/btrfs/relocation.c 	struct backref_node *next;
next             2683 fs/btrfs/relocation.c 	next = node;
next             2686 fs/btrfs/relocation.c 		next = walk_up_backref(next, edges, &index);
next             2687 fs/btrfs/relocation.c 		root = next->root;
next             2697 fs/btrfs/relocation.c 		if (next != node)
next             2700 fs/btrfs/relocation.c 		next = walk_down_backref(edges, &index);
next             2701 fs/btrfs/relocation.c 		if (!next || next->level <= node->level)
next             2715 fs/btrfs/relocation.c 	struct backref_node *next = node;
next             2723 fs/btrfs/relocation.c 	while (next) {
next             2726 fs/btrfs/relocation.c 			if (next->processed && (reserve || next != node))
next             2731 fs/btrfs/relocation.c 			if (list_empty(&next->upper))
next             2734 fs/btrfs/relocation.c 			edge = list_entry(next->upper.next,
next             2737 fs/btrfs/relocation.c 			next = edge->node[UPPER];
next             2739 fs/btrfs/relocation.c 		next = walk_down_backref(edges, &index);
next             2831 fs/btrfs/relocation.c 					goto next;
next             2836 fs/btrfs/relocation.c 					goto next;
next             2870 fs/btrfs/relocation.c 				goto next;
next             2883 fs/btrfs/relocation.c 				goto next;
next             2887 fs/btrfs/relocation.c 				goto next;
next             2897 fs/btrfs/relocation.c 			goto next;
next             2901 fs/btrfs/relocation.c 			goto next;
next             2913 fs/btrfs/relocation.c 				goto next;
next             2935 fs/btrfs/relocation.c next:
next             2978 fs/btrfs/relocation.c 			node = list_entry(cache->pending[level].next,
next             3020 fs/btrfs/relocation.c 	struct backref_node *next = node;
next             3025 fs/btrfs/relocation.c 	while (next) {
next             3028 fs/btrfs/relocation.c 			if (next->processed)
next             3031 fs/btrfs/relocation.c 			__mark_block_processed(rc, next);
next             3033 fs/btrfs/relocation.c 			if (list_empty(&next->upper))
next             3036 fs/btrfs/relocation.c 			edge = list_entry(next->upper.next,
next             3039 fs/btrfs/relocation.c 			next = edge->node[UPPER];
next             3041 fs/btrfs/relocation.c 		next = walk_down_backref(edges, &index);
next             3144 fs/btrfs/relocation.c 	struct tree_block *next;
next             3155 fs/btrfs/relocation.c 	rbtree_postorder_for_each_entry_safe(block, next, blocks, rb_node) {
next             3161 fs/btrfs/relocation.c 	rbtree_postorder_for_each_entry_safe(block, next, blocks, rb_node) {
next             3170 fs/btrfs/relocation.c 	rbtree_postorder_for_each_entry_safe(block, next, blocks, rb_node) {
next             3769 fs/btrfs/relocation.c 			goto next;
next             3773 fs/btrfs/relocation.c 			goto next;
next             3777 fs/btrfs/relocation.c 			goto next;
next             3782 fs/btrfs/relocation.c 			goto next;
next             3804 fs/btrfs/relocation.c next:
next             3940 fs/btrfs/relocation.c next:
next             3958 fs/btrfs/relocation.c 			goto next;
next             3964 fs/btrfs/relocation.c 			goto next;
next             3971 fs/btrfs/relocation.c 			goto next;
next             4594 fs/btrfs/relocation.c 		reloc_root = list_entry(reloc_roots.next,
next             4681 fs/btrfs/relocation.c 		sums = list_entry(list.next, struct btrfs_ordered_sum, list);
next             2703 fs/btrfs/scrub.c 	struct scrub_page *curr, *next;
next             2714 fs/btrfs/scrub.c 	list_for_each_entry_safe(curr, next, &sparity->spages, list) {
next             2915 fs/btrfs/scrub.c 				goto next;
next             2923 fs/btrfs/scrub.c 				goto next;
next             2948 fs/btrfs/scrub.c 				goto next;
next             3018 fs/btrfs/scrub.c next:
next             3267 fs/btrfs/scrub.c 				goto next;
next             3275 fs/btrfs/scrub.c 				goto next;
next             3299 fs/btrfs/scrub.c 				goto next;
next             3386 fs/btrfs/scrub.c next:
next             2113 fs/btrfs/send.c 		nce = list_entry(sctx->name_cache_list.next,
next             2125 fs/btrfs/send.c 		nce = list_entry(sctx->name_cache_list.next,
next             2819 fs/btrfs/send.c 		cur = list_entry(head->next, struct recorded_ref, list);
next             4102 fs/btrfs/send.c 		cur = list_entry(sctx->deleted_refs.next, struct recorded_ref,
next             5219 fs/btrfs/send.c 			goto next;
next             5319 fs/btrfs/send.c next:
next             5661 fs/btrfs/send.c 			goto next;
next             5679 fs/btrfs/send.c 			goto next;
next             5682 fs/btrfs/send.c 			goto next;
next             5686 fs/btrfs/send.c next:
next             1126 fs/btrfs/transaction.c 	struct list_head *next;
next             1164 fs/btrfs/transaction.c 		next = fs_info->dirty_cowonly_roots.next;
next             1165 fs/btrfs/transaction.c 		list_del_init(next);
next             1166 fs/btrfs/transaction.c 		root = list_entry(next, struct btrfs_root, dirty_list);
next             1675 fs/btrfs/transaction.c 	struct btrfs_pending_snapshot *pending, *next;
next             1679 fs/btrfs/transaction.c 	list_for_each_entry_safe(pending, next, head, list) {
next              807 fs/btrfs/tree-log.c 				sums = list_entry(ordered_sums.next,
next             1106 fs/btrfs/tree-log.c 				goto next;
next             1147 fs/btrfs/tree-log.c next:
next             2170 fs/btrfs/tree-log.c 		goto next;
next             2183 fs/btrfs/tree-log.c next:
next             2689 fs/btrfs/tree-log.c 	struct extent_buffer *next;
next             2719 fs/btrfs/tree-log.c 		next = btrfs_find_create_tree_block(fs_info, bytenr);
next             2720 fs/btrfs/tree-log.c 		if (IS_ERR(next))
next             2721 fs/btrfs/tree-log.c 			return PTR_ERR(next);
next             2724 fs/btrfs/tree-log.c 			ret = wc->process_func(root, next, wc, ptr_gen,
next             2727 fs/btrfs/tree-log.c 				free_extent_buffer(next);
next             2733 fs/btrfs/tree-log.c 				ret = btrfs_read_buffer(next, ptr_gen,
next             2736 fs/btrfs/tree-log.c 					free_extent_buffer(next);
next             2741 fs/btrfs/tree-log.c 					btrfs_tree_lock(next);
next             2742 fs/btrfs/tree-log.c 					btrfs_set_lock_blocking_write(next);
next             2743 fs/btrfs/tree-log.c 					btrfs_clean_tree_block(next);
next             2744 fs/btrfs/tree-log.c 					btrfs_wait_tree_block_writeback(next);
next             2745 fs/btrfs/tree-log.c 					btrfs_tree_unlock(next);
next             2747 fs/btrfs/tree-log.c 					if (test_and_clear_bit(EXTENT_BUFFER_DIRTY, &next->bflags))
next             2748 fs/btrfs/tree-log.c 						clear_extent_buffer_dirty(next);
next             2757 fs/btrfs/tree-log.c 					free_extent_buffer(next);
next             2761 fs/btrfs/tree-log.c 			free_extent_buffer(next);
next             2764 fs/btrfs/tree-log.c 		ret = btrfs_read_buffer(next, ptr_gen, *level - 1, &first_key);
next             2766 fs/btrfs/tree-log.c 			free_extent_buffer(next);
next             2773 fs/btrfs/tree-log.c 		path->nodes[*level-1] = next;
next             2774 fs/btrfs/tree-log.c 		*level = btrfs_header_level(next);
next             2820 fs/btrfs/tree-log.c 				struct extent_buffer *next;
next             2822 fs/btrfs/tree-log.c 				next = path->nodes[*level];
next             2825 fs/btrfs/tree-log.c 					btrfs_tree_lock(next);
next             2826 fs/btrfs/tree-log.c 					btrfs_set_lock_blocking_write(next);
next             2827 fs/btrfs/tree-log.c 					btrfs_clean_tree_block(next);
next             2828 fs/btrfs/tree-log.c 					btrfs_wait_tree_block_writeback(next);
next             2829 fs/btrfs/tree-log.c 					btrfs_tree_unlock(next);
next             2831 fs/btrfs/tree-log.c 					if (test_and_clear_bit(EXTENT_BUFFER_DIRTY, &next->bflags))
next             2832 fs/btrfs/tree-log.c 						clear_extent_buffer_dirty(next);
next             2902 fs/btrfs/tree-log.c 			struct extent_buffer *next;
next             2904 fs/btrfs/tree-log.c 			next = path->nodes[orig_level];
next             2907 fs/btrfs/tree-log.c 				btrfs_tree_lock(next);
next             2908 fs/btrfs/tree-log.c 				btrfs_set_lock_blocking_write(next);
next             2909 fs/btrfs/tree-log.c 				btrfs_clean_tree_block(next);
next             2910 fs/btrfs/tree-log.c 				btrfs_wait_tree_block_writeback(next);
next             2911 fs/btrfs/tree-log.c 				btrfs_tree_unlock(next);
next             2913 fs/btrfs/tree-log.c 				if (test_and_clear_bit(EXTENT_BUFFER_DIRTY, &next->bflags))
next             2914 fs/btrfs/tree-log.c 					clear_extent_buffer_dirty(next);
next             2920 fs/btrfs/tree-log.c 							next->start, next->len);
next             4071 fs/btrfs/tree-log.c 		struct btrfs_ordered_sum *sums = list_entry(ordered_sums.next,
next             4130 fs/btrfs/tree-log.c 		struct btrfs_ordered_sum *sums = list_entry(ordered_sums.next,
next             4436 fs/btrfs/tree-log.c 		em = list_entry(extents.next, struct extent_map, list);
next             6254 fs/btrfs/tree-log.c 				goto next;
next             6293 fs/btrfs/tree-log.c next:
next               64 fs/btrfs/ulist.c 	struct ulist_node *next;
next               66 fs/btrfs/ulist.c 	list_for_each_entry_safe(node, next, &ulist->nodes, list) {
next              267 fs/btrfs/ulist.c 	if (uiter->cur_list && uiter->cur_list->next == &ulist->nodes)
next              270 fs/btrfs/ulist.c 		uiter->cur_list = uiter->cur_list->next;
next              272 fs/btrfs/ulist.c 		uiter->cur_list = ulist->nodes.next;
next              353 fs/btrfs/volumes.c 		device = list_entry(fs_devices->devices.next,
next              366 fs/btrfs/volumes.c 		fs_devices = list_entry(fs_uuids.next,
next             1215 fs/btrfs/volumes.c 	struct btrfs_device *device, *next;
next             1221 fs/btrfs/volumes.c 	list_for_each_entry_safe(device, next, &fs_devices->devices, dev_list) {
next             1668 fs/btrfs/volumes.c 			goto next;
next             1674 fs/btrfs/volumes.c 			goto next;
next             1716 fs/btrfs/volumes.c next:
next             7649 fs/btrfs/volumes.c 	struct btrfs_device *curr, *next;
next             7662 fs/btrfs/volumes.c 	list_for_each_entry_safe(curr, next, &trans->dev_update_list,
next              327 fs/btrfs/xattr.c 				goto next;
next              339 fs/btrfs/xattr.c next:
next              379 fs/btrfs/zlib.c 			goto next;
next              396 fs/btrfs/zlib.c next:
next              107 fs/btrfs/zstd.c 	struct list_head *pos, *next;
next              116 fs/btrfs/zstd.c 	list_for_each_prev_safe(pos, next, &wsm.lru_list) {
next              205 fs/btrfs/zstd.c 			workspace = container_of(wsm.idle_ws[i].next,
next              237 fs/btrfs/zstd.c 			ws = wsm.idle_ws[i].next;
next              686 fs/buffer.c    		bh = BH_ENTRY(list->next);
next              770 fs/buffer.c    			__remove_assoc_queue(BH_ENTRY(list->next));
next              793 fs/buffer.c    			struct buffer_head *bh = BH_ENTRY(list->next);
next             1472 fs/buffer.c    	struct buffer_head *head, *bh, *next;
next             1489 fs/buffer.c    		next = bh->b_this_page;
next             1503 fs/buffer.c    		bh = next;
next             1607 fs/buffer.c    					goto next;
next             1613 fs/buffer.c    next:
next             1776 fs/buffer.c    		struct buffer_head *next = bh->b_this_page;
next             1782 fs/buffer.c    		bh = next;
next             1830 fs/buffer.c    		struct buffer_head *next = bh->b_this_page;
next             1837 fs/buffer.c    		bh = next;
next             3232 fs/buffer.c    		struct buffer_head *next = bh->b_this_page;
next             3236 fs/buffer.c    		bh = next;
next             3285 fs/buffer.c    			struct buffer_head *next = bh->b_this_page;
next             3287 fs/buffer.c    			bh = next;
next              488 fs/cachefiles/namei.c 	struct dentry *dir, *next = NULL;
next              531 fs/cachefiles/namei.c 	next = lookup_one_len(name, dir, nlen);
next              533 fs/cachefiles/namei.c 	if (IS_ERR(next)) {
next              534 fs/cachefiles/namei.c 		trace_cachefiles_lookup(object, next, NULL);
next              538 fs/cachefiles/namei.c 	inode = d_backing_inode(next);
next              539 fs/cachefiles/namei.c 	trace_cachefiles_lookup(object, next, inode);
next              540 fs/cachefiles/namei.c 	_debug("next -> %p %s", next, inode ? "positive" : "negative");
next              548 fs/cachefiles/namei.c 	if (d_is_negative(next))
next              554 fs/cachefiles/namei.c 		if (d_is_negative(next)) {
next              560 fs/cachefiles/namei.c 			ret = security_path_mkdir(&path, next, 0);
next              564 fs/cachefiles/namei.c 			ret = vfs_mkdir(d_inode(dir), next, 0);
next              567 fs/cachefiles/namei.c 				trace_cachefiles_mkdir(object, next, ret);
next              571 fs/cachefiles/namei.c 			if (unlikely(d_unhashed(next))) {
next              572 fs/cachefiles/namei.c 				dput(next);
next              576 fs/cachefiles/namei.c 			ASSERT(d_backing_inode(next));
next              579 fs/cachefiles/namei.c 			       next, d_backing_inode(next), d_backing_inode(next)->i_ino);
next              581 fs/cachefiles/namei.c 		} else if (!d_can_lookup(next)) {
next              583 fs/cachefiles/namei.c 			       d_backing_inode(next)->i_ino);
next              590 fs/cachefiles/namei.c 		if (d_is_negative(next)) {
next              596 fs/cachefiles/namei.c 			ret = security_path_mknod(&path, next, S_IFREG, 0);
next              600 fs/cachefiles/namei.c 			ret = vfs_create(d_inode(dir), next, S_IFREG, true);
next              602 fs/cachefiles/namei.c 			trace_cachefiles_create(object, next, ret);
next              606 fs/cachefiles/namei.c 			ASSERT(d_backing_inode(next));
next              609 fs/cachefiles/namei.c 			       next, d_backing_inode(next), d_backing_inode(next)->i_ino);
next              611 fs/cachefiles/namei.c 		} else if (!d_can_lookup(next) &&
next              612 fs/cachefiles/namei.c 			   !d_is_reg(next)
next              615 fs/cachefiles/namei.c 			       d_backing_inode(next)->i_ino);
next              626 fs/cachefiles/namei.c 		dir = next;
next              627 fs/cachefiles/namei.c 		next = NULL;
next              632 fs/cachefiles/namei.c 	object->dentry = next;
next              637 fs/cachefiles/namei.c 		_debug("validate '%pd'", next);
next              645 fs/cachefiles/namei.c 			ret = cachefiles_bury_object(cache, object, dir, next,
next              648 fs/cachefiles/namei.c 			dput(next);
next              649 fs/cachefiles/namei.c 			next = NULL;
next              682 fs/cachefiles/namei.c 		path.dentry = next;
next              736 fs/cachefiles/namei.c 	_debug("lookup error %ld", PTR_ERR(next));
next              737 fs/cachefiles/namei.c 	ret = PTR_ERR(next);
next              740 fs/cachefiles/namei.c 	next = NULL;
next              743 fs/cachefiles/namei.c 	dput(next);
next               79 fs/cachefiles/proc.c 	.next		= cachefiles_histogram_next,
next              169 fs/cachefiles/rdwr.c 		monitor = list_entry(op->to_do.next,
next              190 fs/cachefiles/rdwr.c 				goto next;
next              208 fs/cachefiles/rdwr.c 	next:
next             1272 fs/ceph/dir.c  			goto next;
next             1297 fs/ceph/dir.c  next:
next             1326 fs/ceph/mds_client.c 	p = session->s_caps.next;
next             1331 fs/ceph/mds_client.c 			p = p->next;
next             1352 fs/ceph/mds_client.c 		p = p->next;
next             1506 fs/ceph/mds_client.c 			cap = list_entry(session->s_caps.next,
next             2632 fs/ceph/mds_client.c 		req = list_entry(tmp_list.next,
next              213 fs/ceph/quota.c 	struct ceph_snap_realm *realm, *next;
next              253 fs/ceph/quota.c 		next = realm->parent;
next              254 fs/ceph/quota.c 		if (has_quota || !next)
next              257 fs/ceph/quota.c 		ceph_get_snap_realm(mdsc, next);
next              259 fs/ceph/quota.c 		realm = next;
next              318 fs/ceph/quota.c 	struct ceph_snap_realm *realm, *next;
next              391 fs/ceph/quota.c 		next = realm->parent;
next              392 fs/ceph/quota.c 		if (exceeded || !next)
next              394 fs/ceph/quota.c 		ceph_get_snap_realm(mdsc, next);
next              396 fs/ceph/quota.c 		realm = next;
next               35 fs/char_dev.c  	struct char_device_struct *next;
next               56 fs/char_dev.c  	for (cd = chrdevs[major_to_index(offset)]; cd; cd = cd->next) {
next               77 fs/char_dev.c  		for (cd = chrdevs[major_to_index(i)]; cd; cd = cd->next)
next              134 fs/char_dev.c  	for (curr = chrdevs[i]; curr; prev = curr, curr = curr->next) {
next              156 fs/char_dev.c  		cd->next = curr;
next              159 fs/char_dev.c  		cd->next = prev->next;
next              160 fs/char_dev.c  		prev->next = cd;
next              178 fs/char_dev.c  	for (cp = &chrdevs[i]; *cp; cp = &(*cp)->next)
next              185 fs/char_dev.c  		*cp = cd->next;
next              204 fs/char_dev.c  	dev_t n, next;
next              206 fs/char_dev.c  	for (n = from; n < to; n = next) {
next              207 fs/char_dev.c  		next = MKDEV(MAJOR(n)+1, 0);
next              208 fs/char_dev.c  		if (next > to)
next              209 fs/char_dev.c  			next = to;
next              211 fs/char_dev.c  			       next - n, name);
next              218 fs/char_dev.c  	for (n = from; n < to; n = next) {
next              219 fs/char_dev.c  		next = MKDEV(MAJOR(n)+1, 0);
next              220 fs/char_dev.c  		kfree(__unregister_chrdev_region(MAJOR(n), MINOR(n), next - n));
next              314 fs/char_dev.c  	dev_t n, next;
next              316 fs/char_dev.c  	for (n = from; n < to; n = next) {
next              317 fs/char_dev.c  		next = MKDEV(MAJOR(n)+1, 0);
next              318 fs/char_dev.c  		if (next > to)
next              319 fs/char_dev.c  			next = to;
next              320 fs/char_dev.c  		kfree(__unregister_chrdev_region(MAJOR(n), MINOR(n), next - n));
next              440 fs/char_dev.c  		inode = container_of(cdev->list.next, struct inode, i_devices);
next             1465 fs/cifs/dfs_cache.c 			goto next;
next             1472 fs/cifs/dfs_cache.c next:
next             1111 fs/cifs/file.c 					(lock->blist.next == &lock->blist));
next             1317 fs/cifs/file.c 	el = locks_to_send.next;
next             2173 fs/cifs/file.c 		    pgoff_t end, pgoff_t *index, pgoff_t *next, bool *done)
next             2203 fs/cifs/file.c 		if (*next && (page->index != *next)) {
next             2231 fs/cifs/file.c 		*next = page->index + 1;
next             2314 fs/cifs/file.c 		pgoff_t next = 0, tofind, saved_index = index;
next             2353 fs/cifs/file.c 					       end, &index, &next, &done);
next             2414 fs/cifs/file.c 		index = next;
next              435 fs/cifs/smb2ops.c 	size_t next = 0;
next              450 fs/cifs/smb2ops.c 		next = le32_to_cpu(p->Next);
next              451 fs/cifs/smb2ops.c 		if (!next) {
next              455 fs/cifs/smb2ops.c 		p = (struct network_interface_info_ioctl_rsp *)((u8 *)p+next);
next              456 fs/cifs/smb2ops.c 		bytes_left -= next;
next              534 fs/cifs/smb2ops.c 		next = le32_to_cpu(p->Next);
next              535 fs/cifs/smb2ops.c 		if (!next)
next              537 fs/cifs/smb2ops.c 		p = (struct network_interface_info_ioctl_rsp *)((u8 *)p+next);
next              538 fs/cifs/smb2ops.c 		bytes_left -= next;
next             3236 fs/cifs/smb2ops.c 	u64 next;
next             3289 fs/cifs/smb2ops.c 		next = le64_to_cpu(out_data[num - 1].file_offset) +
next             3291 fs/cifs/smb2ops.c 		len = len - (next - start);
next             3292 fs/cifs/smb2ops.c 		start = next;
next             1925 fs/cifs/smb2pdu.c 	unsigned int next;
next             1948 fs/cifs/smb2pdu.c 		next = le32_to_cpu(cc->Next);
next             1949 fs/cifs/smb2pdu.c 		if (!next)
next             1951 fs/cifs/smb2pdu.c 		remaining -= next;
next             1952 fs/cifs/smb2pdu.c 		cc = (struct create_context *)((char *)cc + next);
next              748 fs/cifs/smbdirect.c 	send_wr.next = NULL;
next              947 fs/cifs/smbdirect.c 	send_wr.next = NULL;
next             1107 fs/cifs/smbdirect.c 	recv_wr.next = NULL;
next              244 fs/coda/psdev.c 	req = list_entry(vcp->vc_pending.next, struct upc_req,uc_chain);
next             1637 fs/configfs/dir.c 	for (p = q->next; p != &parent_sd->s_children; p = p->next) {
next             1638 fs/configfs/dir.c 		struct configfs_dirent *next;
next             1643 fs/configfs/dir.c 		next = list_entry(p, struct configfs_dirent, s_sibling);
next             1644 fs/configfs/dir.c 		if (!next->s_element)
next             1660 fs/configfs/dir.c 		dentry = next->s_dentry;
next             1669 fs/configfs/dir.c 		name = configfs_get_name(next);
next             1672 fs/configfs/dir.c 		if (!dir_emit(ctx, name, len, ino, dt_type(next)))
next             1709 fs/configfs/dir.c 			p = sd->s_children.next;
next             1711 fs/configfs/dir.c 				struct configfs_dirent *next;
next             1712 fs/configfs/dir.c 				next = list_entry(p, struct configfs_dirent,
next             1714 fs/configfs/dir.c 				if (next->s_element)
next             1716 fs/configfs/dir.c 				p = p->next;
next              446 fs/coredump.c  	core_state->dumper.next = NULL;
next              466 fs/coredump.c  		ptr = core_state->dumper.next;
next              469 fs/coredump.c  			ptr = ptr->next;
next              478 fs/coredump.c  	struct core_thread *curr, *next;
next              488 fs/coredump.c  	next = mm->core_state->dumper.next;
next              489 fs/coredump.c  	while ((curr = next) != NULL) {
next              490 fs/coredump.c  		next = curr->next;
next              514 fs/dcache.c    	struct dentry *next;
next              542 fs/dcache.c    	while (dentry->d_child.next != &parent->d_subdirs) {
next              543 fs/dcache.c    		next = list_entry(dentry->d_child.next, struct dentry, d_child);
next              544 fs/dcache.c    		if (likely(!(next->d_flags & DCACHE_DENTRY_CURSOR)))
next              546 fs/dcache.c    		dentry->d_child.next = next->d_child.next;
next             1271 fs/dcache.c    	struct list_head *next;
next             1293 fs/dcache.c    	next = this_parent->d_subdirs.next;
next             1295 fs/dcache.c    	while (next != &this_parent->d_subdirs) {
next             1296 fs/dcache.c    		struct list_head *tmp = next;
next             1298 fs/dcache.c    		next = tmp->next;
next             1346 fs/dcache.c    			next = child->d_child.next;
next             1347 fs/dcache.c    			if (next == &this_parent->d_subdirs)
next             1349 fs/dcache.c    			child = list_entry(next, struct dentry, d_child);
next             2374 fs/dcache.c    			goto next;
next             2376 fs/dcache.c    			goto next;
next             2379 fs/dcache.c    			goto next;
next             2385 fs/dcache.c    next:
next              313 fs/dcookies.c  	struct list_head next;
next              329 fs/dcookies.c  	list_add(&user->next, &dcookie_users);
next              345 fs/dcookies.c  	list_del(&user->next);
next              504 fs/dlm/debug_fs.c 	struct rb_node *next;
next              518 fs/dlm/debug_fs.c 	next = rb_next(&rp->res_hashnode);
next              520 fs/dlm/debug_fs.c 	if (next) {
next              521 fs/dlm/debug_fs.c 		r = rb_entry(next, struct dlm_rsb, res_hashnode);
next              551 fs/dlm/debug_fs.c 			next = rb_first(tree);
next              552 fs/dlm/debug_fs.c 			r = rb_entry(next, struct dlm_rsb, res_hashnode);
next              576 fs/dlm/debug_fs.c 	.next  = table_seq_next,
next              583 fs/dlm/debug_fs.c 	.next  = table_seq_next,
next              590 fs/dlm/debug_fs.c 	.next  = table_seq_next,
next              597 fs/dlm/debug_fs.c 	.next  = table_seq_next,
next              252 fs/dlm/dir.c   		list = r->res_root_list.next;
next              254 fs/dlm/dir.c   		list = ls->ls_root_list.next;
next              257 fs/dlm/dir.c   	for (offset = 0; list != &ls->ls_root_list; list = list->next) {
next             1650 fs/dlm/lock.c  	struct rb_node *n, *next;
next             1667 fs/dlm/lock.c  	for (n = rb_first(&ls->ls_rsbtbl[b].toss); n; n = next) {
next             1668 fs/dlm/lock.c  		next = rb_next(n);
next             2192 fs/dlm/lock.c  	struct dlm_lkb *first = list_entry(head->next, struct dlm_lkb,
next             2793 fs/dlm/lock.c  			lkb = list_entry(r->res_lookup.next, struct dlm_lkb,
next             6150 fs/dlm/lock.c  	lkb = list_entry(proc->locks.next, struct dlm_lkb, lkb_ownqueue);
next             6223 fs/dlm/lock.c  			lkb = list_entry(proc->locks.next, struct dlm_lkb,
next             1500 fs/dlm/lowcomms.c 		e = list_entry(con->writequeue.next, struct writequeue_entry,
next              308 fs/dlm/member.c 		newlist->next = tmp;
next              309 fs/dlm/member.c 		tmp->prev->next = newlist;
next              367 fs/dlm/member.c 		memb = list_entry(head->next, struct dlm_member, list);
next              392 fs/dlm/plock.c 		op = list_entry(send_list.next, struct plock_op, list);
next              937 fs/dlm/recover.c 	struct rb_node *n, *next;
next              944 fs/dlm/recover.c 		for (n = rb_first(&ls->ls_rsbtbl[i].toss); n; n = next) {
next              945 fs/dlm/recover.c 			next = rb_next(n);
next               77 fs/dlm/requestqueue.c 		e = list_entry(ls->ls_requestqueue.next, struct rq_entry, list);
next              839 fs/dlm/user.c  	lkb = list_entry(proc->asts.next, struct dlm_lkb, lkb_cb_list);
next              278 fs/erofs/utils.c 	p = erofs_sb_list.next;
next              290 fs/erofs/utils.c 			p = p->next;
next              301 fs/erofs/utils.c 		p = p->next;
next              316 fs/erofs/zdata.c 	if (pcl->next == Z_EROFS_PCLUSTER_NIL) {
next              318 fs/erofs/zdata.c 		if (cmpxchg(&pcl->next, Z_EROFS_PCLUSTER_NIL,
next              322 fs/erofs/zdata.c 		*owned_head = &pcl->next;
next              325 fs/erofs/zdata.c 	} else if (pcl->next == Z_EROFS_PCLUSTER_TAIL) {
next              331 fs/erofs/zdata.c 		if (cmpxchg(&pcl->next, Z_EROFS_PCLUSTER_TAIL,
next              355 fs/erofs/zdata.c 	if (clt->owned_head == &pcl->next || pcl == clt->tailpcl) {
next              429 fs/erofs/zdata.c 	pcl->next = clt->owned_head;
next              450 fs/erofs/zdata.c 	clt->owned_head = &pcl->next;
next              947 fs/erofs/zdata.c 	WRITE_ONCE(pcl->next, Z_EROFS_PCLUSTER_NIL);
next              971 fs/erofs/zdata.c 		pcl = container_of(owned, struct z_erofs_pcluster, next);
next              972 fs/erofs/zdata.c 		owned = READ_ONCE(pcl->next);
next             1179 fs/erofs/zdata.c 	WRITE_ONCE(pcl->next, Z_EROFS_PCLUSTER_TAIL_CLOSED);
next             1182 fs/erofs/zdata.c 	WRITE_ONCE(*bypass_qtail, &pcl->next);
next             1184 fs/erofs/zdata.c 	qtail[JQ_BYPASS] = &pcl->next;
next             1241 fs/erofs/zdata.c 		pcl = container_of(owned_head, struct z_erofs_pcluster, next);
next             1246 fs/erofs/zdata.c 		owned_head = cmpxchg(&pcl->next, Z_EROFS_PCLUSTER_TAIL,
next             1292 fs/erofs/zdata.c 			qtail[JQ_SUBMIT] = &pcl->next;
next               60 fs/erofs/zdata.h 	z_erofs_next_pcluster_t next;
next               31 fs/erofs/zpvec.h 	struct page *curr, *next;
next               56 fs/erofs/zpvec.h 	if (ctor->next)
next               57 fs/erofs/zpvec.h 		return ctor->next;
next               74 fs/erofs/zpvec.h 	struct page *next = z_erofs_pagevec_ctor_next_page(ctor, ctor->nr);
next               78 fs/erofs/zpvec.h 	ctor->curr = next;
next               79 fs/erofs/zpvec.h 	ctor->next = NULL;
next               93 fs/erofs/zpvec.h 	ctor->curr = ctor->next = NULL;
next              104 fs/erofs/zpvec.h 	ctor->next = z_erofs_pagevec_ctor_next_page(ctor, i);
next              114 fs/erofs/zpvec.h 	if (!ctor->next && type)
next              126 fs/erofs/zpvec.h 	if (type == (uintptr_t)ctor->next) {
next              127 fs/erofs/zpvec.h 		ctor->next = page;
next              141 fs/erofs/zpvec.h 		DBG_BUGON(!ctor->next);
next              150 fs/erofs/zpvec.h 	if (*type == (uintptr_t)ctor->next)
next              151 fs/erofs/zpvec.h 		ctor->next = tagptr_unfold_ptr(t);
next              152 fs/eventpoll.c 	struct epitem *next;
next              713 fs/eventpoll.c 	     nepi = epi->next, epi->next = EP_UNACTIVE_PTR) {
next              988 fs/eventpoll.c 	struct epitem *epi, *next;
next             1004 fs/eventpoll.c 	list_for_each_entry_safe(epi, next, &file->f_ep_links, fllink) {
next             1145 fs/eventpoll.c 	if (cmpxchg(&new->next, new, head) != new)
next             1163 fs/eventpoll.c 	prev->next = new;
next             1180 fs/eventpoll.c 	if (epi->next != EP_UNACTIVE_PTR)
next             1184 fs/eventpoll.c 	if (cmpxchg(&epi->next, EP_UNACTIVE_PTR, NULL) != EP_UNACTIVE_PTR)
next             1188 fs/eventpoll.c 	epi->next = xchg(&ep->ovflist, epi);
next             1521 fs/eventpoll.c 	epi->next = EP_UNACTIVE_PTR;
next              582 fs/ext2/balloc.c 	ext2_grpblk_t next;
next              584 fs/ext2/balloc.c 	next = ext2_find_next_zero_bit(bh->b_data, maxblocks, start);
next              585 fs/ext2/balloc.c 	if (next >= maxblocks)
next              587 fs/ext2/balloc.c 	return next;
next              605 fs/ext2/balloc.c 	ext2_grpblk_t here, next;
next              632 fs/ext2/balloc.c 	next = (r - ((char *)bh->b_data)) << 3;
next              634 fs/ext2/balloc.c 	if (next < maxblocks && next >= here)
next              635 fs/ext2/balloc.c 		return next;
next              783 fs/ext2/balloc.c 	struct rb_node *next;
next              812 fs/ext2/balloc.c 		next = rb_next(&rsv->rsv_node);
next              813 fs/ext2/balloc.c 		rsv = rb_entry(next,struct ext2_reserve_window_node,rsv_node);
next              819 fs/ext2/balloc.c 		if (!next)
next             1044 fs/ext2/balloc.c 	struct rb_node *next;
next             1050 fs/ext2/balloc.c 	next = rb_next(&my_rsv->rsv_node);
next             1052 fs/ext2/balloc.c 	if (!next)
next             1055 fs/ext2/balloc.c 		next_rsv = rb_entry(next, struct ext2_reserve_window_node, rsv_node);
next              152 fs/ext2/xattr.c 	struct ext2_xattr_entry *next;
next              155 fs/ext2/xattr.c 	next = EXT2_XATTR_NEXT(entry);
next              156 fs/ext2/xattr.c 	if ((char *)next >= end)
next              392 fs/ext4/dir.c  	struct fname	*next;
next              405 fs/ext4/dir.c  	struct fname *fname, *next;
next              407 fs/ext4/dir.c  	rbtree_postorder_for_each_entry_safe(fname, next, root, rb_hash)
next              410 fs/ext4/dir.c  			fname = fname->next;
next              480 fs/ext4/dir.c  			new_fn->next = fname->next;
next              481 fs/ext4/dir.c  			fname->next = new_fn;
next              529 fs/ext4/dir.c  		fname = fname->next;
next             1965 fs/ext4/extents.c 	ext4_lblk_t next;
next             2065 fs/ext4/extents.c 	next = EXT_MAX_BLOCKS;
next             2067 fs/ext4/extents.c 		next = ext4_ext_next_leaf_block(path);
next             2068 fs/ext4/extents.c 	if (next != EXT_MAX_BLOCKS) {
next             2069 fs/ext4/extents.c 		ext_debug("next leaf block - %u\n", next);
next             2071 fs/ext4/extents.c 		npath = ext4_find_extent(inode, next, NULL, 0);
next             2183 fs/ext4/extents.c 	ext4_lblk_t next, next_del, start = 0, end = 0;
next             2210 fs/ext4/extents.c 		next = ext4_ext_next_allocated_block(path);
next             2230 fs/ext4/extents.c 			if (end >= next)
next             2231 fs/ext4/extents.c 				end = next;
next             2290 fs/ext4/extents.c 		if (next == next_del && next == EXT_MAX_BLOCKS) {
next             2293 fs/ext4/extents.c 				     next != EXT_MAX_BLOCKS)) {
next             2297 fs/ext4/extents.c 						 next, next_del);
next             2329 fs/ext4/extents.c 	ext4_lblk_t next, end = block + num - 1;
next             2336 fs/ext4/extents.c 		next = 0;
next             2338 fs/ext4/extents.c 		if (!ext4_es_lookup_extent(inode, block, &next, &es))
next             2347 fs/ext4/extents.c 		if (next == 0)
next             2359 fs/ext4/extents.c 		if (next == 0)
next             2361 fs/ext4/extents.c 		block = next;
next             2401 fs/ext4/extents.c 		ext4_lblk_t next;
next             2404 fs/ext4/extents.c 		next = ext4_ext_next_allocated_block(path);
next             2405 fs/ext4/extents.c 		BUG_ON(next == *lblk);
next             2406 fs/ext4/extents.c 		len = next - *lblk;
next             4236 fs/ext4/extents.c 			ext4_lblk_t next = ext4_ext_next_allocated_block(path);
next             4237 fs/ext4/extents.c 			map->m_len = min(map->m_len, next - map->m_lblk);
next             1696 fs/ext4/extents_status.c 			goto next;
next             1699 fs/ext4/extents_status.c 			goto next;
next             1705 fs/ext4/extents_status.c next:
next               48 fs/ext4/fsync.c 	struct inode *next;
next               59 fs/ext4/fsync.c 		next = igrab(d_inode(dentry->d_parent));
next               61 fs/ext4/fsync.c 		if (!next)
next               64 fs/ext4/fsync.c 		inode = next;
next              712 fs/ext4/ialloc.c next:
next              723 fs/ext4/ialloc.c 			goto next;
next             1102 fs/ext4/inode.c 	struct buffer_head *next;
next             1106 fs/ext4/inode.c 	     block_start = block_end, bh = next) {
next             1107 fs/ext4/inode.c 		next = bh->b_this_page;
next             1496 fs/ext4/mballoc.c 	int next = block;
next             1522 fs/ext4/mballoc.c 	next = next - ex->fe_start;
next             1523 fs/ext4/mballoc.c 	ex->fe_len -= next;
next             1524 fs/ext4/mballoc.c 	ex->fe_start += next;
next             1532 fs/ext4/mballoc.c 		next = (block + 1) * (1 << order);
next             1533 fs/ext4/mballoc.c 		if (mb_test_bit(next, e4b->bd_bitmap))
next             1536 fs/ext4/mballoc.c 		order = mb_find_order_for_block(e4b, next);
next             1538 fs/ext4/mballoc.c 		block = next >> order;
next             2339 fs/ext4/mballoc.c 	.next   = ext4_mb_seq_groups_next,
next             3823 fs/ext4/mballoc.c 	unsigned int next;
next             3839 fs/ext4/mballoc.c 		next = mb_find_next_bit(bitmap_bh->b_data, end, bit);
next             3842 fs/ext4/mballoc.c 			 (unsigned) next - bit, (unsigned) group);
next             3843 fs/ext4/mballoc.c 		free += next - bit;
next             3845 fs/ext4/mballoc.c 		trace_ext4_mballoc_discard(sb, NULL, group, bit, next - bit);
next             3848 fs/ext4/mballoc.c 					       next - bit);
next             3849 fs/ext4/mballoc.c 		mb_free_blocks(pa->pa_inode, e4b, bit, next - bit);
next             3850 fs/ext4/mballoc.c 		bit = next + 1;
next             4035 fs/ext4/mballoc.c 		pa = list_entry(ei->i_prealloc_list.next,
next             5182 fs/ext4/mballoc.c 	ext4_grpblk_t next, count = 0, free_count = 0;
next             5208 fs/ext4/mballoc.c 		next = mb_find_next_bit(bitmap, max + 1, start);
next             5210 fs/ext4/mballoc.c 		if ((next - start) >= minblocks) {
next             5212 fs/ext4/mballoc.c 					       next - start, group, &e4b);
next             5216 fs/ext4/mballoc.c 			count += next - start;
next             5218 fs/ext4/mballoc.c 		free_count += next - start;
next             5219 fs/ext4/mballoc.c 		start = next + 1;
next             5352 fs/ext4/mballoc.c 	ext4_grpblk_t			next;
next             5372 fs/ext4/mballoc.c 		next = mb_find_next_bit(bitmap, end + 1, start);
next             5375 fs/ext4/mballoc.c 		error = formatter(sb, group, start, next - start, priv);
next             5380 fs/ext4/mballoc.c 		start = next + 1;
next             1528 fs/ext4/namei.c 			goto next;
next             1559 fs/ext4/namei.c 	next:
next             1787 fs/ext4/namei.c 	struct ext4_dir_entry_2 *next, *to, *prev, *de = (struct ext4_dir_entry_2 *) base;
next             1792 fs/ext4/namei.c 		next = ext4_next_entry(de, blocksize);
next             1801 fs/ext4/namei.c 		de = next;
next              149 fs/ext4/page-io.c 		   io, inode->i_ino, io->list.next, io->list.prev);
next              179 fs/ext4/page-io.c 		after = cur->next;
next              222 fs/ext4/page-io.c 		io = list_entry(unwritten.next, ext4_io_end_t, list);
next              425 fs/ext4/super.c 		jce = list_entry(txn->t_private_list.next,
next             3173 fs/ext4/super.c 					n = pos->next;
next              190 fs/ext4/xattr.c 		struct ext4_xattr_entry *next = EXT4_XATTR_NEXT(e);
next              191 fs/ext4/xattr.c 		if ((void *)next >= end)
next              195 fs/ext4/xattr.c 		e = next;
next              284 fs/ext4/xattr.c 	struct ext4_xattr_entry *entry, *next;
next              291 fs/ext4/xattr.c 	for (entry = *pentry; !IS_LAST_ENTRY(entry); entry = next) {
next              292 fs/ext4/xattr.c 		next = EXT4_XATTR_NEXT(entry);
next              293 fs/ext4/xattr.c 		if ((void *) next >= end) {
next             1565 fs/ext4/xattr.c 	struct ext4_xattr_entry *last, *next;
next             1600 fs/ext4/xattr.c 	for (; !IS_LAST_ENTRY(last); last = next) {
next             1601 fs/ext4/xattr.c 		next = EXT4_XATTR_NEXT(last);
next             1602 fs/ext4/xattr.c 		if ((void *)next >= s->end) {
next              610 fs/f2fs/data.c next:
next              662 fs/f2fs/data.c 		goto next;
next             1575 fs/f2fs/data.c next:
next             1620 fs/f2fs/data.c 		goto next;
next              565 fs/f2fs/dir.c  next:
next              578 fs/f2fs/dir.c  	goto next;
next              172 fs/f2fs/extent_cache.c 	struct rb_node *cur = rb_first_cached(root), *next;
next              179 fs/f2fs/extent_cache.c 		next = rb_next(cur);
next              180 fs/f2fs/extent_cache.c 		if (!next)
next              184 fs/f2fs/extent_cache.c 		next_re = rb_entry(next, struct rb_entry, rb_node);
next              193 fs/f2fs/extent_cache.c 		cur = next;
next              302 fs/f2fs/extent_cache.c 	struct rb_node *node, *next;
next              308 fs/f2fs/extent_cache.c 		next = rb_next(node);
next              311 fs/f2fs/extent_cache.c 		node = next;
next              633 fs/f2fs/extent_cache.c 	struct extent_tree *et, *next;
next              648 fs/f2fs/extent_cache.c 	list_for_each_entry_safe(et, next, &sbi->zombie_list, list) {
next             1025 fs/f2fs/file.c 		goto next;
next             1054 fs/f2fs/file.c next:
next               87 fs/f2fs/gc.c   			goto next;
next               94 fs/f2fs/gc.c   			goto next;
next              113 fs/f2fs/gc.c   next:
next              392 fs/f2fs/gc.c   			goto next;
next              398 fs/f2fs/gc.c   			goto next;
next              403 fs/f2fs/gc.c   			goto next;
next              405 fs/f2fs/gc.c   			goto next;
next              413 fs/f2fs/gc.c   next:
next             2490 fs/f2fs/node.c 	struct free_nid *i, *next;
next             2500 fs/f2fs/node.c 	list_for_each_entry_safe(i, next, &nm_i->free_nid_list, list) {
next              332 fs/f2fs/recovery.c 			goto next;
next              358 fs/f2fs/recovery.c 					goto next;
next              368 fs/f2fs/recovery.c next:
next              680 fs/f2fs/recovery.c 			goto next;
next              708 fs/f2fs/recovery.c next:
next              253 fs/f2fs/segment.c 				goto next;
next              270 fs/f2fs/segment.c next:
next              293 fs/f2fs/segment.c next:
next              321 fs/f2fs/segment.c 	goto next;
next              608 fs/f2fs/segment.c 		struct flush_cmd *cmd, *next;
next              619 fs/f2fs/segment.c 		llist_for_each_entry_safe(cmd, next,
next              682 fs/f2fs/segment.c 			struct flush_cmd *tmp, *next;
next              686 fs/f2fs/segment.c 			llist_for_each_entry_safe(tmp, next, list, llnode) {
next             1341 fs/f2fs/segment.c 			goto next;
next             1373 fs/f2fs/segment.c  next:
next             1432 fs/f2fs/segment.c 			goto next;
next             1444 fs/f2fs/segment.c next:
next             1492 fs/f2fs/segment.c 			goto next;
next             1516 fs/f2fs/segment.c next:
next             1587 fs/f2fs/segment.c next:
next             1611 fs/f2fs/segment.c 		goto next;
next             1979 fs/f2fs/segment.c next:
next             1989 fs/f2fs/segment.c 			goto next;
next             2755 fs/f2fs/segment.c next:
next             2797 fs/f2fs/segment.c 			goto next;
next             3757 fs/f2fs/segment.c 	struct sit_entry_set *next = ses;
next             3762 fs/f2fs/segment.c 	list_for_each_entry_continue(next, head, set_list)
next             3763 fs/f2fs/segment.c 		if (ses->entry_cnt <= next->entry_cnt)
next             3766 fs/f2fs/segment.c 	list_move_tail(&ses->set_list, &next->set_list);
next              419 fs/f2fs/segment.h 	unsigned int next;
next              425 fs/f2fs/segment.h 	next = find_next_bit(free_i->free_segmap,
next              427 fs/f2fs/segment.h 	if (next >= start_segno + sbi->segs_per_sec) {
next              452 fs/f2fs/segment.h 	unsigned int next;
next              460 fs/f2fs/segment.h 		next = find_next_bit(free_i->free_segmap,
next              462 fs/f2fs/segment.h 		if (next >= start_segno + sbi->segs_per_sec) {
next               47 fs/f2fs/shrinker.c 	p = f2fs_list.next;
next               53 fs/f2fs/shrinker.c 			p = p->next;
next               68 fs/f2fs/shrinker.c 		p = p->next;
next               88 fs/f2fs/shrinker.c 	p = f2fs_list.next;
next               97 fs/f2fs/shrinker.c 			p = p->next;
next              116 fs/f2fs/shrinker.c 		p = p->next;
next              692 fs/f2fs/xattr.c 		struct f2fs_xattr_entry *next = XATTR_NEXT_ENTRY(here);
next              695 fs/f2fs/xattr.c 		memmove(here, next, (char *)last - (char *)next);
next               76 fs/fat/cache.c 	if (MSDOS_I(inode)->cache_lru.next != &cache->cache_list)
next              192 fs/fat/cache.c 		cache = list_entry(i->cache_lru.next,
next               90 fs/fat/dir.c   next:
next              106 fs/fat/dir.c   		goto next;
next              121 fs/fat/fatent.c 	int next;
next              125 fs/fat/fatent.c 		next = (*ent12_p[0] >> 4) | (*ent12_p[1] << 4);
next              127 fs/fat/fatent.c 		next = (*ent12_p[1] << 8) | *ent12_p[0];
next              130 fs/fat/fatent.c 	next &= 0x0fff;
next              131 fs/fat/fatent.c 	if (next >= BAD_FAT12)
next              132 fs/fat/fatent.c 		next = FAT_ENT_EOF;
next              133 fs/fat/fatent.c 	return next;
next              138 fs/fat/fatent.c 	int next = le16_to_cpu(*fatent->u.ent16_p);
next              140 fs/fat/fatent.c 	if (next >= BAD_FAT16)
next              141 fs/fat/fatent.c 		next = FAT_ENT_EOF;
next              142 fs/fat/fatent.c 	return next;
next              147 fs/fat/fatent.c 	int next = le32_to_cpu(*fatent->u.ent32_p) & 0x0fffffff;
next              149 fs/fat/fatent.c 	if (next >= BAD_FAT32)
next              150 fs/fat/fatent.c 		next = FAT_ENT_EOF;
next              151 fs/fat/fatent.c 	return next;
next               52 fs/filesystems.c 	for (p = &file_systems; *p; p = &(*p)->next)
next               81 fs/filesystems.c 	if (fs->next)
next              115 fs/filesystems.c 			*tmp = fs->next;
next              116 fs/filesystems.c 			fs->next = NULL;
next              121 fs/filesystems.c 		tmp = &(*tmp)->next;
next              144 fs/filesystems.c 	for (tmp=file_systems, index=0 ; tmp ; tmp=tmp->next, index++) {
next              161 fs/filesystems.c 	for (tmp = file_systems; tmp; tmp = tmp->next, index--)
next              181 fs/filesystems.c 	for (tmp = file_systems, index = 0 ; tmp ; tmp = tmp->next, index++)
next              222 fs/filesystems.c 		tmp = tmp->next;
next              239 fs/filesystems.c 		tmp = tmp->next;
next             1180 fs/fs-writeback.c 		tail = wb_inode(wb->b_dirty.next);
next             1942 fs/fs-writeback.c 		work = list_entry(wb->work_list.next,
next              160 fs/fscache/cache.c 	cache = list_entry(fscache_cache_list.next,
next              334 fs/fscache/cache.c 			object = list_entry(cache->object_list.next,
next               85 fs/fscache/histogram.c 	.next		= fscache_histogram_next,
next              304 fs/fscache/object-list.c 	.next		= fscache_objlist_next,
next              880 fs/fscache/object.c 		dep = list_entry(object->dependents.next,
next              336 fs/fscache/operation.c 		op = list_entry(object->pending_ops.next,
next              428 fs/fscache/operation.c 		op = list_entry(object->pending_ops.next,
next              578 fs/fscache/operation.c 		op = list_entry(cache->op_gc_list.next,
next             1220 fs/fscache/page.c 	pgoff_t next;
next             1231 fs/fscache/page.c 	next = 0;
next             1233 fs/fscache/page.c 		if (!pagevec_lookup(&pvec, mapping, &next))
next             1244 fs/fscache/page.c 	} while (next);
next              242 fs/fuse/dev.c  		fiq->forget_list_tail->next = forget;
next             1004 fs/fuse/dev.c  	return fiq->forget_list_head.next != NULL;
next             1055 fs/fuse/dev.c  	struct fuse_forget_link *head = fiq->forget_list_head.next;
next             1060 fs/fuse/dev.c  		newhead = &(*newhead)->next;
next             1062 fs/fuse/dev.c  	fiq->forget_list_head.next = *newhead;
next             1064 fs/fuse/dev.c  	if (fiq->forget_list_head.next == NULL)
next             1144 fs/fuse/dev.c  		head = forget->next;
next             1161 fs/fuse/dev.c  	if (fc->minor < 16 || fiq->forget_list_head.next->next == NULL)
next             1227 fs/fuse/dev.c  		req = list_entry(fiq->interrupts.next, struct fuse_req,
next             1240 fs/fuse/dev.c  	req = list_entry(fiq->pending.next, struct fuse_req, list);
next             2047 fs/fuse/dev.c  		req = list_entry(head->next, struct fuse_req, list);
next             2095 fs/fuse/dev.c  		struct fuse_req *req, *next;
next             2110 fs/fuse/dev.c  			list_for_each_entry_safe(req, next, &fpq->io, list) {
next              362 fs/fuse/file.c 	struct fuse_writepage_args *next;
next             1632 fs/fuse/file.c 	struct fuse_writepage_args *aux, *next;
next             1672 fs/fuse/file.c 	for (aux = wpa->next; aux; aux = next) {
next             1673 fs/fuse/file.c 		next = aux->next;
next             1674 fs/fuse/file.c 		aux->next = NULL;
next             1698 fs/fuse/file.c 		wpa = list_entry(fi->queued_writes.next,
next             1715 fs/fuse/file.c 	while (wpa->next) {
next             1718 fs/fuse/file.c 		struct fuse_writepage_args *next = wpa->next;
next             1720 fs/fuse/file.c 		wpa->next = next->next;
next             1721 fs/fuse/file.c 		next->next = NULL;
next             1722 fs/fuse/file.c 		next->ia.ff = fuse_file_get(wpa->ia.ff);
next             1723 fs/fuse/file.c 		list_add(&next->writepages_entry, &fi->writepages);
next             1748 fs/fuse/file.c 		fuse_send_writepage(fc, next, inarg->offset + inarg->size);
next             1763 fs/fuse/file.c 		ff = list_entry(fi->write_files.next, struct fuse_file,
next             1845 fs/fuse/file.c 	wpa->next = NULL;
next             1979 fs/fuse/file.c 	for (tmp = old_wpa->next; tmp; tmp = tmp->next) {
next             1992 fs/fuse/file.c 		new_wpa->next = old_wpa->next;
next             1993 fs/fuse/file.c 		old_wpa->next = new_wpa;
next             2081 fs/fuse/file.c 		wpa->next = NULL;
next               63 fs/fuse/fuse_i.h 	struct fuse_forget_link *next;
next              455 fs/fuse/virtio_fs.c 	struct fuse_req *next;
next              475 fs/fuse/virtio_fs.c 	list_for_each_entry_safe(req, next, &reqs, list) {
next             2227 fs/gfs2/bmap.c 		jext = list_entry(jd->extent_list.next, struct gfs2_journal_extent, list);
next             1009 fs/gfs2/dir.c  	struct gfs2_dirent *dent = NULL, *prev = NULL, *next = NULL, *new;
next             1087 fs/gfs2/dir.c  		next = dent;
next             1088 fs/gfs2/dir.c  		if (dirent_next(dip, obh, &next))
next             1089 fs/gfs2/dir.c  			next = NULL;
next             1119 fs/gfs2/dir.c  		dent = next;
next              998 fs/gfs2/file.c 	loff_t next = (offset + len - 1) >> sdp->sd_sb.sb_bsize_shift;
next             1001 fs/gfs2/file.c 	next = (next + 1) << sdp->sd_sb.sb_bsize_shift;
next             1005 fs/gfs2/file.c 	len = next - offset;
next              284 fs/gfs2/glock.c 	const struct gfs2_holder *gh_head = list_entry(gl->gl_holders.next, const struct gfs2_holder, gh_list);
next              610 fs/gfs2/glock.c 		gh = list_entry(gl->gl_holders.next, struct gfs2_holder, gh_list);
next              831 fs/gfs2/glock.c 	gl->gl_node.next = NULL;
next             1157 fs/gfs2/glock.c 	gh = list_entry(gl->gl_holders.next, struct gfs2_holder, gh_list);
next             1552 fs/gfs2/glock.c 		gl = list_entry(list->next, struct gfs2_glock, gl_lru);
next             1593 fs/gfs2/glock.c 		gl = list_entry(lru_list.next, struct gfs2_glock, gl_lru);
next             2142 fs/gfs2/glock.c 	.next  = gfs2_glock_seq_next,
next             2149 fs/gfs2/glock.c 	.next  = gfs2_glock_seq_next,
next             2156 fs/gfs2/glock.c 	.next  = gfs2_sbstats_seq_next,
next              553 fs/gfs2/log.c  		ip = list_entry(sdp->sd_log_ordered.next, struct gfs2_inode, i_ordered);
next              574 fs/gfs2/log.c  		ip = list_entry(sdp->sd_log_ordered.next, struct gfs2_inode, i_ordered);
next              174 fs/gfs2/lops.c 	struct buffer_head *bh, *next;
next              186 fs/gfs2/lops.c 		next = bh->b_this_page;
next              189 fs/gfs2/lops.c 		bh = next;
next              732 fs/gfs2/lops.c 		bd = list_entry(head->next, struct gfs2_bufdata, bd_list);
next              902 fs/gfs2/lops.c 		bd = list_entry(head->next, struct gfs2_bufdata, bd_list);
next             1081 fs/gfs2/lops.c 		bd = list_entry(head->next, struct gfs2_bufdata, bd_list);
next               76 fs/gfs2/meta_io.c 		struct buffer_head *next = bh->b_this_page;
next               81 fs/gfs2/meta_io.c 		bh = next;
next              199 fs/gfs2/meta_io.c 			struct buffer_head *next = bh->b_this_page;
next              202 fs/gfs2/meta_io.c 			bh = next;
next              118 fs/gfs2/quota.c 		qd = list_entry(list->next, struct gfs2_quota_data, qd_lru);
next             1506 fs/gfs2/quota.c 			ip = list_entry(sdp->sd_trunc_list.next,
next              114 fs/gfs2/recovery.c 		rr = list_entry(head->next, struct gfs2_revoke_replay, rr_list);
next              510 fs/gfs2/rgrp.c 	struct rb_node *n, *next;
next              517 fs/gfs2/rgrp.c 		next = NULL;
next              519 fs/gfs2/rgrp.c 			next = n->rb_left;
next              521 fs/gfs2/rgrp.c 			next = n->rb_right;
next              522 fs/gfs2/rgrp.c 		if (next == NULL) {
next              532 fs/gfs2/rgrp.c 		n = next;
next             1088 fs/gfs2/rgrp.c 	struct gfs2_rgrpd *next = gfs2_rgrpd_get_next(rgd);
next             1095 fs/gfs2/rgrp.c 	if (next == NULL)
next             1097 fs/gfs2/rgrp.c 	else if (next->rd_addr > rgd->rd_addr)
next             1098 fs/gfs2/rgrp.c 		str->rg_skip = cpu_to_be32(next->rd_addr - rgd->rd_addr);
next               65 fs/gfs2/super.c 		jd = list_entry(list.next, struct gfs2_jdesc, jd_list);
next              448 fs/gfs2/super.c 		lfcc = list_entry(list.next, struct lfcc, list);
next              195 fs/hfs/bfind.c 			idx = bnode->next;
next              143 fs/hfs/bnode.c 		be32_to_cpu(desc.next), be32_to_cpu(desc.prev),
next              182 fs/hfs/bnode.c 		tmp->next = node->next;
next              183 fs/hfs/bnode.c 		cnid = cpu_to_be32(tmp->next);
next              184 fs/hfs/bnode.c 		hfs_bnode_write(tmp, &cnid, offsetof(struct hfs_bnode_desc, next), 4);
next              187 fs/hfs/bnode.c 		tree->leaf_head = node->next;
next              189 fs/hfs/bnode.c 	if (node->next) {
next              190 fs/hfs/bnode.c 		tmp = hfs_bnode_find(tree, node->next);
next              201 fs/hfs/bnode.c 	if (!node->prev && !node->next) {
next              340 fs/hfs/bnode.c 	node->next = be32_to_cpu(desc->next);
next              246 fs/hfs/brec.c  		node->this, new_node->this, node->next);
next              247 fs/hfs/brec.c  	new_node->next = node->next;
next              253 fs/hfs/brec.c  	if (node->next)
next              254 fs/hfs/brec.c  		next_node = hfs_bnode_find(tree, node->next);
next              317 fs/hfs/brec.c  	node_desc.next = cpu_to_be32(new_node->next);
next              326 fs/hfs/brec.c  	node->next = new_node->this;
next              328 fs/hfs/brec.c  	node_desc.next = cpu_to_be32(node->next);
next              487 fs/hfs/brec.c  	new_node->next = 0;
next              491 fs/hfs/brec.c  	node_desc.next = cpu_to_be32(new_node->next);
next              202 fs/hfs/btree.c 	prev->next = idx;
next              204 fs/hfs/btree.c 	hfs_bnode_write(prev, &cnid, offsetof(struct hfs_bnode_desc, next), 4);
next              209 fs/hfs/btree.c 	desc.next = 0;
next              301 fs/hfs/btree.c 		nidx = node->next;
next              340 fs/hfs/btree.c 		i = node->next;
next               51 fs/hfs/btree.h 	u32 next;
next              127 fs/hfs/btree.h 	__be32 next;		/* (V) Number of the next node at this level */
next              264 fs/hfsplus/bfind.c 			idx = bnode->next;
next              307 fs/hfsplus/bnode.c 		be32_to_cpu(desc.next), be32_to_cpu(desc.prev),
next              346 fs/hfsplus/bnode.c 		tmp->next = node->next;
next              347 fs/hfsplus/bnode.c 		cnid = cpu_to_be32(tmp->next);
next              349 fs/hfsplus/bnode.c 			offsetof(struct hfs_bnode_desc, next), 4);
next              352 fs/hfsplus/bnode.c 		tree->leaf_head = node->next;
next              354 fs/hfsplus/bnode.c 	if (node->next) {
next              355 fs/hfsplus/bnode.c 		tmp = hfs_bnode_find(tree, node->next);
next              367 fs/hfsplus/bnode.c 	if (!node->prev && !node->next)
next              508 fs/hfsplus/bnode.c 	node->next = be32_to_cpu(desc->next);
next              250 fs/hfsplus/brec.c 		node->this, new_node->this, node->next);
next              251 fs/hfsplus/brec.c 	new_node->next = node->next;
next              257 fs/hfsplus/brec.c 	if (node->next)
next              258 fs/hfsplus/brec.c 		next_node = hfs_bnode_find(tree, node->next);
next              321 fs/hfsplus/brec.c 	node_desc.next = cpu_to_be32(new_node->next);
next              330 fs/hfsplus/brec.c 	node->next = new_node->this;
next              332 fs/hfsplus/brec.c 	node_desc.next = cpu_to_be32(node->next);
next              491 fs/hfsplus/brec.c 	new_node->next = 0;
next              495 fs/hfsplus/brec.c 	node_desc.next = cpu_to_be32(new_node->next);
next              324 fs/hfsplus/btree.c 	prev->next = idx;
next              326 fs/hfsplus/btree.c 	hfs_bnode_write(prev, &cnid, offsetof(struct hfs_bnode_desc, next), 4);
next              331 fs/hfsplus/btree.c 	desc.next = 0;
next              428 fs/hfsplus/btree.c 		nidx = node->next;
next              468 fs/hfsplus/btree.c 		i = node->next;
next              220 fs/hfsplus/dir.c 				goto next;
next              256 fs/hfsplus/dir.c next:
next              108 fs/hfsplus/hfsplus_fs.h 	u32 next;
next              152 fs/hfsplus/hfsplus_raw.h 	__be32 next;
next               98 fs/hfsplus/xattr.c 		desc->next = cpu_to_be32(be32_to_cpu(head->leaf_tail) + 1);
next              277 fs/hostfs/hostfs_kern.c 	unsigned long long next, ino;
next              288 fs/hostfs/hostfs_kern.c 	next = ctx->pos;
next              289 fs/hostfs/hostfs_kern.c 	seek_dir(dir, next);
next              290 fs/hostfs/hostfs_kern.c 	while ((name = read_dir(dir, &next, &ino, &len, &type)) != NULL) {
next              293 fs/hostfs/hostfs_kern.c 		ctx->pos = next;
next              423 fs/hugetlbfs/inode.c 	pgoff_t next, index;
next              430 fs/hugetlbfs/inode.c 	next = start;
next              431 fs/hugetlbfs/inode.c 	while (next < end) {
next              435 fs/hugetlbfs/inode.c 		if (!pagevec_lookup_range(&pvec, mapping, &next, end - 1))
next              626 fs/inode.c     	struct inode *inode, *next;
next              631 fs/inode.c     	list_for_each_entry_safe(inode, next, &sb->s_inodes, i_sb_list) {
next              677 fs/inode.c     	struct inode *inode, *next;
next              682 fs/inode.c     	list_for_each_entry_safe(inode, next, &sb->s_inodes, i_sb_list) {
next              899 fs/inode.c     		int next = atomic_add_return(LAST_INO_BATCH, &shared_last_ino);
next              901 fs/inode.c     		res = next - LAST_INO_BATCH;
next             1757 fs/io_uring.c  	struct io_kiocb *poll_req, *next;
next             1767 fs/io_uring.c  	list_for_each_entry_safe(poll_req, next, &ctx->cancel_list, list) {
next              389 fs/iomap/buffered-io.c 	loff_t last = page_offset(list_entry(pages->next, struct page, lru));
next             1031 fs/jbd2/journal.c 	.next   = jbd2_seq_info_next,
next               70 fs/jbd2/recovery.c 	unsigned int max, nbufs, next;
next               86 fs/jbd2/recovery.c 	for (next = start; next < max; next++) {
next               87 fs/jbd2/recovery.c 		err = jbd2_journal_bmap(journal, next, &blocknr);
next               91 fs/jbd2/recovery.c 				next);
next              169 fs/jbd2/revoke.c 	record = (struct jbd2_revoke_record_s *) hash_list->next;
next              175 fs/jbd2/revoke.c 		record = (struct jbd2_revoke_record_s *) record->hash.next;
next              547 fs/jbd2/revoke.c 				hash_list->next;
next              732 fs/jbd2/revoke.c 			record = (struct jbd2_revoke_record_s*) hash_list->next;
next             2374 fs/jbd2/transaction.c 	struct buffer_head *head, *bh, *next;
next             2395 fs/jbd2/transaction.c 		next = bh->b_this_page;
next             2410 fs/jbd2/transaction.c 		bh = next;
next               40 fs/jffs2/build.c 	if (ic->next)
next               41 fs/jffs2/build.c 		return ic->next;
next               61 fs/jffs2/build.c 	for(fd = ic->scan_dents; fd; fd = fd->next) {
next              151 fs/jffs2/build.c 		dead_fds = fd->next;
next              181 fs/jffs2/build.c 			ic->scan_dents = fd->next;
next              229 fs/jffs2/build.c 				ic->scan_dents = fd->next;
next              250 fs/jffs2/build.c 		struct jffs2_raw_node_ref *next = raw->next_in_ino;
next              253 fs/jffs2/build.c 		raw = next;
next              264 fs/jffs2/build.c 			ic->scan_dents = fd->next;
next              292 fs/jffs2/build.c 				fd->next = *dead_fds;
next               97 fs/jffs2/dir.c 	for (fd_list = dir_f->dents; fd_list && fd_list->nhash <= nhash; fd_list = fd_list->next) {
next              133 fs/jffs2/dir.c 	for (fd = f->dents; fd; fd = fd->next) {
next              593 fs/jffs2/dir.c 	for (fd = f->dents ; fd; fd = fd->next) {
next              779 fs/jffs2/dir.c 			for (fd = victim_f->dents; fd; fd = fd->next) {
next              114 fs/jffs2/erase.c 			jeb = list_entry(c->erase_complete_list.next, struct jffs2_eraseblock, list);
next              127 fs/jffs2/erase.c 			jeb = list_entry(c->erase_pending_list.next, struct jffs2_eraseblock, list);
next              304 fs/jffs2/fs.c  		for (fd=f->dents; fd; fd = fd->next) {
next               97 fs/jffs2/gc.c  	ret = list_entry(nextlist->next, struct jffs2_eraseblock, list);
next              159 fs/jffs2/gc.c  			for (ic = c->inocache_list[bucket]; ic; ic = ic->next) {
next              566 fs/jffs2/gc.c  	for (fd = f->dents; fd; fd=fd->next) {
next              999 fs/jffs2/gc.c  			*fdp = fd->next;
next             1002 fs/jffs2/gc.c  		fdp = &(*fdp)->next;
next               43 fs/jffs2/nodelist.c 				new->next = (*prev)->next;
next               53 fs/jffs2/nodelist.c 		prev = &((*prev)->next);
next               55 fs/jffs2/nodelist.c 	new->next = *prev;
next               73 fs/jffs2/nodelist.c 		struct jffs2_node_frag *next = frag_next(frag);
next               77 fs/jffs2/nodelist.c 		frag = next;
next              395 fs/jffs2/nodelist.c 		struct jffs2_node_frag *next = frag_next(newfrag);
next              397 fs/jffs2/nodelist.c 		if (next) {
next              399 fs/jffs2/nodelist.c 			if (next->node)
next              400 fs/jffs2/nodelist.c 				mark_ref_normal(next->node->raw);
next              427 fs/jffs2/nodelist.c 		ret = ret->next;
next              449 fs/jffs2/nodelist.c 		prev = &(*prev)->next;
next              451 fs/jffs2/nodelist.c 	new->next = *prev;
next              470 fs/jffs2/nodelist.c 		prev = &(*prev)->next;
next              473 fs/jffs2/nodelist.c 		*prev = old->next;
next              490 fs/jffs2/nodelist.c 	struct jffs2_inode_cache *this, *next;
next              495 fs/jffs2/nodelist.c 			next = this->next;
next              498 fs/jffs2/nodelist.c 			this = next;
next              507 fs/jffs2/nodelist.c 	struct jffs2_raw_node_ref *this, *next;
next              513 fs/jffs2/nodelist.c 				next = this[REFS_PER_BLOCK].next_in_ino;
next              515 fs/jffs2/nodelist.c 				next = NULL;
next              518 fs/jffs2/nodelist.c 			this = next;
next              528 fs/jffs2/nodelist.c 	struct rb_node *next;
next              534 fs/jffs2/nodelist.c 	next = fragtree->rb_node;
next              536 fs/jffs2/nodelist.c 	while(next) {
next              537 fs/jffs2/nodelist.c 		frag = rb_entry(next, struct jffs2_node_frag, rb);
next              543 fs/jffs2/nodelist.c 			next = frag->rb.rb_right;
next              545 fs/jffs2/nodelist.c 			next = frag->rb.rb_left;
next              567 fs/jffs2/nodelist.c 	struct jffs2_node_frag *frag, *next;
next              570 fs/jffs2/nodelist.c 	rbtree_postorder_for_each_entry_safe(frag, next, root, rb) {
next              175 fs/jffs2/nodelist.h 	struct jffs2_inode_cache *next;
next              257 fs/jffs2/nodelist.h 	struct jffs2_full_dirent *next;
next              282 fs/jffs2/nodemgmt.c 	struct list_head *next;
next              292 fs/jffs2/nodemgmt.c 			ejeb = list_entry(c->erasable_list.next, struct jffs2_eraseblock, list);
next              334 fs/jffs2/nodemgmt.c 	next = c->free_list.next;
next              335 fs/jffs2/nodemgmt.c 	list_del(next);
next              336 fs/jffs2/nodemgmt.c 	c->nextblock = list_entry(next, struct jffs2_eraseblock, list);
next              177 fs/jffs2/readinode.c 	struct rb_node *next;
next              182 fs/jffs2/readinode.c 	next = tn_root->rb_node;
next              184 fs/jffs2/readinode.c 	while (next) {
next              185 fs/jffs2/readinode.c 		tn = rb_entry(next, struct jffs2_tmp_dnode_info, rb);
next              188 fs/jffs2/readinode.c 			next = tn->rb.rb_right;
next              190 fs/jffs2/readinode.c 			next = tn->rb.rb_left;
next              303 fs/jffs2/readinode.c 				struct jffs2_tmp_dnode_info *next = tn_next(this);
next              311 fs/jffs2/readinode.c 				this = next;
next              546 fs/jffs2/readinode.c 	struct jffs2_tmp_dnode_info *tn, *next;
next              548 fs/jffs2/readinode.c 	rbtree_postorder_for_each_entry_safe(tn, next, list, rb) {
next              558 fs/jffs2/readinode.c 	struct jffs2_full_dirent *next;
next              561 fs/jffs2/readinode.c 		next = fd->next;
next              563 fs/jffs2/readinode.c 		fd = next;
next              678 fs/jffs2/readinode.c 	fd->next = NULL;
next             1420 fs/jffs2/readinode.c 		fds = fd->next;
next              427 fs/jffs2/scan.c 	ref->next = c->xref_temp;
next             1103 fs/jffs2/scan.c 	fd->next = NULL;
next             1132 fs/jffs2/scan.c 	struct list_head *n = head->next;
next             1136 fs/jffs2/scan.c 		n = n->next;
next               68 fs/jffs2/summary.c 		s->sum_list_tail->u.next = (union jffs2_sum_mem *) item;
next              128 fs/jffs2/summary.c 	temp->next = NULL;
next              150 fs/jffs2/summary.c 	temp->next = NULL;
next              171 fs/jffs2/summary.c 	temp->next = NULL;
next              186 fs/jffs2/summary.c 	temp->next = NULL;
next              202 fs/jffs2/summary.c 		s->sum_list_head = s->sum_list_head->u.next;
next              276 fs/jffs2/summary.c 			temp->next = NULL;
next              296 fs/jffs2/summary.c 			temp->next = NULL;
next              326 fs/jffs2/summary.c 			temp->next = NULL;
next              337 fs/jffs2/summary.c 			temp->next = NULL;
next              476 fs/jffs2/summary.c 				fd->next = NULL;
next              537 fs/jffs2/summary.c 				ref->next = c->xref_temp;
next              792 fs/jffs2/summary.c 		c->summary->sum_list_head = temp->u.next;
next               95 fs/jffs2/summary.h 	union jffs2_sum_mem *next;
next              101 fs/jffs2/summary.h 	union jffs2_sum_mem *next;
next              111 fs/jffs2/summary.h 	union jffs2_sum_mem *next;
next              125 fs/jffs2/summary.h 	union jffs2_sum_mem *next;
next              135 fs/jffs2/summary.h 	union jffs2_sum_mem *next;
next               43 fs/jffs2/wbuf.c 	struct jffs2_inodirty *next;
next               64 fs/jffs2/wbuf.c 		this = this->next;
next               77 fs/jffs2/wbuf.c 			struct jffs2_inodirty *next = this->next;
next               79 fs/jffs2/wbuf.c 			this = next;
next              103 fs/jffs2/wbuf.c 	new->next = c->wbuf_inodes;
next              110 fs/jffs2/wbuf.c 	struct list_head *this, *next;
next              116 fs/jffs2/wbuf.c 	list_for_each_safe(this, next, &c->erasable_pending_wbuf_list) {
next              213 fs/jffs2/wbuf.c 		for (fd = f->dents; fd; fd = fd->next) {
next              278 fs/jffs2/wbuf.c 	struct jffs2_raw_node_ref *raw, *next, *first_raw = NULL;
next              298 fs/jffs2/wbuf.c 	for (next = raw = jeb->first_node; next; raw = next) {
next              299 fs/jffs2/wbuf.c 		next = ref_next(raw);
next              302 fs/jffs2/wbuf.c 		    (next && ref_offset(next) <= c->wbuf_ofs)) {
next              609 fs/jffs2/write.c 		for (fd = dir_f->dents; fd; fd = fd->next) {
next              641 fs/jffs2/write.c 				dead_f->dents = fd->next;
next              570 fs/jffs2/xattr.c 	ref->next = ic->xref;
next              586 fs/jffs2/xattr.c 	ref->next = c->xref_dead_list;
next              607 fs/jffs2/xattr.c 		_ref = ref->next;
next              622 fs/jffs2/xattr.c 		_ref = ref->next;
next              648 fs/jffs2/xattr.c 	for (ref=ic->xref, pref=&ic->xref; ref; pref=&ref->next, ref=ref->next) {
next              652 fs/jffs2/xattr.c 				*pref = ref->next;
next              658 fs/jffs2/xattr.c 		for (cmp=ref->next, pcmp=&ref->next; cmp; pcmp=&cmp->next, cmp=cmp->next) {
next              664 fs/jffs2/xattr.c 					*pcmp = cmp->next;
next              673 fs/jffs2/xattr.c 					*pcmp = cmp->next;
next              676 fs/jffs2/xattr.c 					*pref = ref->next;
next              747 fs/jffs2/xattr.c 		_ref = ref->next;
next              752 fs/jffs2/xattr.c 		_ref = ref->next;
next              793 fs/jffs2/xattr.c 		_ref = ref->next;
next              805 fs/jffs2/xattr.c 		for (tmp=xref_tmphash[i]; tmp; tmp=tmp->next) {
next              822 fs/jffs2/xattr.c 			ref->next = xref_tmphash[i];
next              832 fs/jffs2/xattr.c 			_ref = ref->next;
next              834 fs/jffs2/xattr.c 				ref->next = c->xref_dead_list;
next              847 fs/jffs2/xattr.c 				ref->next = c->xref_dead_list;
next              855 fs/jffs2/xattr.c 			ref->next = ic->xref;
next              981 fs/jffs2/xattr.c 	for (ref=ic->xref, pref=&ic->xref; ref; pref=&ref->next, ref=ref->next) {
next              994 fs/jffs2/xattr.c 					*pref = ref->next;
next             1047 fs/jffs2/xattr.c 	for (ref=ic->xref, pref=&ic->xref; ref; pref=&ref->next, ref=ref->next) {
next             1063 fs/jffs2/xattr.c 					*pref = ref->next;
next             1119 fs/jffs2/xattr.c 	for (ref=ic->xref, pref=&ic->xref; ref; pref=&ref->next, ref=ref->next) {
next             1126 fs/jffs2/xattr.c 				*pref = ref->next;
next             1143 fs/jffs2/xattr.c 					*pref = ref->next;
next             1145 fs/jffs2/xattr.c 					ref->next = c->xref_dead_list;
next             1189 fs/jffs2/xattr.c 		*pref = ref->next;
next             1193 fs/jffs2/xattr.c 			ref->next = ic->xref;
next             1340 fs/jffs2/xattr.c 	for (tmp=c->xref_dead_list, ptmp=&c->xref_dead_list; tmp; ptmp=&tmp->next, tmp=tmp->next) {
next             1342 fs/jffs2/xattr.c 			*ptmp = tmp->next;
next               62 fs/jffs2/xattr.h 	struct jffs2_xattr_ref *next;		/* chained from ic->xref_list */
next              483 fs/jfs/jfs_dtree.c static void free_index(tid_t tid, struct inode *ip, u32 index, u32 next)
next              496 fs/jfs/jfs_dtree.c 	dirtab_slot->addr2 = cpu_to_le32(next);
next             1398 fs/jfs/jfs_dtree.c 	nextbn = le64_to_cpu(sp->header.next);
next             1399 fs/jfs/jfs_dtree.c 	rp->header.next = cpu_to_le64(nextbn);
next             1401 fs/jfs/jfs_dtree.c 	sp->header.next = cpu_to_le64(rbn);
next             1446 fs/jfs/jfs_dtree.c 			f->next = fsi;
next             1447 fs/jfs/jfs_dtree.c 		f->next = -1;
next             1563 fs/jfs/jfs_dtree.c 		f->next = fsi;
next             1564 fs/jfs/jfs_dtree.c 	f->next = -1;
next             1775 fs/jfs/jfs_dtree.c 		f->next = last;
next             1788 fs/jfs/jfs_dtree.c 		f->next = fsi;
next             1789 fs/jfs/jfs_dtree.c 	f->next = -1;
next             1798 fs/jfs/jfs_dtree.c 			fsi = f->next;
next             1801 fs/jfs/jfs_dtree.c 		f->next = n;
next             1927 fs/jfs/jfs_dtree.c 	rp->header.next = 0;
next             1960 fs/jfs/jfs_dtree.c 		f->next = fsi;
next             1961 fs/jfs/jfs_dtree.c 	f->next = -1;
next             1972 fs/jfs/jfs_dtree.c 			fsi = f->next;
next             1975 fs/jfs/jfs_dtree.c 		f->next = n;
next             2036 fs/jfs/jfs_dtree.c 	s->next = -1;
next             2049 fs/jfs/jfs_dtree.c 		f->next = fsi;
next             2050 fs/jfs/jfs_dtree.c 	f->next = -1;
next             2114 fs/jfs/jfs_dtree.c 			    || (p->header.next == 0))
next             2118 fs/jfs/jfs_dtree.c 				DT_GETPAGE(ip, le64_to_cpu(p->header.next),
next             2485 fs/jfs/jfs_dtree.c 	if (p->header.next) {
next             2486 fs/jfs/jfs_dtree.c 		nextbn = le64_to_cpu(p->header.next);
next             2523 fs/jfs/jfs_dtree.c 		lp->header.next = cpu_to_le64(nxaddr);
next             2705 fs/jfs/jfs_dtree.c 	if (p->header.next)
next             2706 fs/jfs/jfs_dtree.c 		bn = le64_to_cpu(p->header.next);
next             2744 fs/jfs/jfs_dtree.c 	nextbn = le64_to_cpu(p->header.next);
next             2801 fs/jfs/jfs_dtree.c 		p->header.next = cpu_to_le64(nextbn);
next             2887 fs/jfs/jfs_dtree.c 		f->next = fsi;
next             2888 fs/jfs/jfs_dtree.c 	f->next = -1;
next             3010 fs/jfs/jfs_dtree.c 	int i, next;
next             3243 fs/jfs/jfs_dtree.c 			next = d->next;
next             3244 fs/jfs/jfs_dtree.c 			while (next >= 0) {
next             3245 fs/jfs/jfs_dtree.c 				t = (struct dtslot *) & p->slot[next];
next             3262 fs/jfs/jfs_dtree.c 				next = t->next;
next             3277 fs/jfs/jfs_dtree.c 				bn = le64_to_cpu(p->header.next);
next             3462 fs/jfs/jfs_dtree.c 	bn = le64_to_cpu(p->header.next);
next             3495 fs/jfs/jfs_dtree.c 		bn = le64_to_cpu(p->header.next);
next             3536 fs/jfs/jfs_dtree.c 		bn = le64_to_cpu(p->header.next);
next             3608 fs/jfs/jfs_dtree.c 	si = ih->next;
next             3635 fs/jfs/jfs_dtree.c 		si = t->next;
next             3691 fs/jfs/jfs_dtree.c 		si = lh->next;
next             3704 fs/jfs/jfs_dtree.c 		si = ih->next;
next             3745 fs/jfs/jfs_dtree.c 		si = t->next;
next             3846 fs/jfs/jfs_dtree.c 		si = lh->next;
next             3855 fs/jfs/jfs_dtree.c 		si = ih->next;
next             3880 fs/jfs/jfs_dtree.c 		si = t->next;
next             3916 fs/jfs/jfs_dtree.c 	p->header.freelist = h->next;
next             3929 fs/jfs/jfs_dtree.c 		lh->next = h->next;
next             3944 fs/jfs/jfs_dtree.c 		ih->next = h->next;
next             3964 fs/jfs/jfs_dtree.c 		p->header.freelist = t->next;
next             4004 fs/jfs/jfs_dtree.c 			lh->next = -1;
next             4006 fs/jfs/jfs_dtree.c 			ih->next = -1;
next             4009 fs/jfs/jfs_dtree.c 		t->next = -1;
next             4055 fs/jfs/jfs_dtree.c 	int ssi, next;		/* src slot index */
next             4131 fs/jfs/jfs_dtree.c 			next = slh->next;
next             4135 fs/jfs/jfs_dtree.c 			dlh->next = dsi;
next             4143 fs/jfs/jfs_dtree.c 			next = sih->next;
next             4146 fs/jfs/jfs_dtree.c 			dih->next = dsi;
next             4150 fs/jfs/jfs_dtree.c 		s->next = sfsi;
next             4162 fs/jfs/jfs_dtree.c 		while ((ssi = next) >= 0) {
next             4197 fs/jfs/jfs_dtree.c 			d->next = dsi;
next             4200 fs/jfs/jfs_dtree.c 			next = s->next;
next             4201 fs/jfs/jfs_dtree.c 			s->next = sfsi;
next             4212 fs/jfs/jfs_dtree.c 				dlh->next = -1;
next             4214 fs/jfs/jfs_dtree.c 				dih->next = -1;
next             4217 fs/jfs/jfs_dtree.c 			d->next = -1;
next             4276 fs/jfs/jfs_dtree.c 		si = ((struct ldtentry *) t)->next;
next             4278 fs/jfs/jfs_dtree.c 		si = ((struct idtentry *) t)->next;
next             4279 fs/jfs/jfs_dtree.c 	t->next = si;
next             4311 fs/jfs/jfs_dtree.c 		si = t->next;
next             4321 fs/jfs/jfs_dtree.c 	t->next = p->header.freelist;
next             4372 fs/jfs/jfs_dtree.c 	si = ((struct idtentry *) t)->next;
next             4373 fs/jfs/jfs_dtree.c 	((struct idtentry *) t)->next = -1;
next             4406 fs/jfs/jfs_dtree.c 		si = t->next;
next             4418 fs/jfs/jfs_dtree.c 	t->next = p->header.freelist;
next             4452 fs/jfs/jfs_dtree.c 	si = t->next;
next             4478 fs/jfs/jfs_dtree.c 		si = t->next;
next               35 fs/jfs/jfs_dtree.h 	s8 next;		/* 1: */
next               55 fs/jfs/jfs_dtree.h 	s8 next;		/* 1: */
next               74 fs/jfs/jfs_dtree.h 	s8 next;		/* 1: */
next              182 fs/jfs/jfs_dtree.h 		__le64 next;	/* 8: next sibling */
next              463 fs/jfs/jfs_logmgr.c 	if ((i = linelock->next)) {
next              760 fs/jfs/jfs_logmgr.c 	gcpn = list_entry(log->cqueue.next, struct tblock, cqueue)->pn;
next              959 fs/jfs/jfs_logmgr.c 			lp = list_entry(log->synclist.next,
next             1870 fs/jfs/jfs_logmgr.c 		struct lbuf *next = lbuf->l_freelist;
next             1873 fs/jfs/jfs_logmgr.c 		lbuf = next;
next             2378 fs/jfs/jfs_logmgr.c 	sbi = list_entry(log->sb_list.next, struct jfs_sb_info, log_list);
next              191 fs/jfs/jfs_txnmgr.c 	TxAnchor.freelock = TxLock[lid].next;
next              205 fs/jfs/jfs_txnmgr.c 	TxLock[lid].next = TxAnchor.freelock;
next              276 fs/jfs/jfs_txnmgr.c 		TxBlock[k].next = k + 1;
next              280 fs/jfs/jfs_txnmgr.c 	TxBlock[k].next = 0;
next              304 fs/jfs/jfs_txnmgr.c 		TxLock[k].next = k + 1;
next              305 fs/jfs/jfs_txnmgr.c 	TxLock[k].next = 0;
next              398 fs/jfs/jfs_txnmgr.c 	if ((tblk->next == 0) && !(flag & COMMIT_FORCE)) {
next              406 fs/jfs/jfs_txnmgr.c 	TxAnchor.freetid = tblk->next;
next              418 fs/jfs/jfs_txnmgr.c 	tblk->next = tblk->last = tblk->xflag = tblk->flag = tblk->lsn = 0;
next              523 fs/jfs/jfs_txnmgr.c 	assert(tblk->next == 0);
next              528 fs/jfs/jfs_txnmgr.c 	tblk->next = TxAnchor.freetid;
next              650 fs/jfs/jfs_txnmgr.c 			jfs_ip->atlhead = tlck->next;
next              654 fs/jfs/jfs_txnmgr.c 			     lid_to_tlock(last)->next != lid;
next              655 fs/jfs/jfs_txnmgr.c 			     last = lid_to_tlock(last)->next) {
next              658 fs/jfs/jfs_txnmgr.c 			lid_to_tlock(last)->next = tlck->next;
next              665 fs/jfs/jfs_txnmgr.c 		if (tblk->next)
next              666 fs/jfs/jfs_txnmgr.c 			lid_to_tlock(tblk->last)->next = lid;
next              668 fs/jfs/jfs_txnmgr.c 			tblk->next = lid;
next              669 fs/jfs/jfs_txnmgr.c 		tlck->next = 0;
next              733 fs/jfs/jfs_txnmgr.c 		if (tblk->next)
next              734 fs/jfs/jfs_txnmgr.c 			lid_to_tlock(tblk->last)->next = lid;
next              736 fs/jfs/jfs_txnmgr.c 			tblk->next = lid;
next              737 fs/jfs/jfs_txnmgr.c 		tlck->next = 0;
next              744 fs/jfs/jfs_txnmgr.c 		tlck->next = jfs_ip->atlhead;
next              746 fs/jfs/jfs_txnmgr.c 		if (tlck->next == 0) {
next              758 fs/jfs/jfs_txnmgr.c 	linelock->next = 0;
next              871 fs/jfs/jfs_txnmgr.c 	for (lid = tblk->next; lid; lid = tlck->next) {
next              899 fs/jfs/jfs_txnmgr.c 	lid_t lid, next, llid, k;
next              911 fs/jfs/jfs_txnmgr.c 	for (lid = tblk->next; lid; lid = next) {
next              913 fs/jfs/jfs_txnmgr.c 		next = tlck->next;
next              950 fs/jfs/jfs_txnmgr.c 		llid = ((struct linelock *) & tlck->lock)->next;
next              953 fs/jfs/jfs_txnmgr.c 			k = linelock->next;
next              961 fs/jfs/jfs_txnmgr.c 	tblk->next = tblk->last = 0;
next             1018 fs/jfs/jfs_txnmgr.c 		if (tblk->next)
next             1019 fs/jfs/jfs_txnmgr.c 			lid_to_tlock(tblk->last)->next = lid;
next             1021 fs/jfs/jfs_txnmgr.c 			tblk->next = lid;
next             1022 fs/jfs/jfs_txnmgr.c 		tlck->next = 0;
next             1029 fs/jfs/jfs_txnmgr.c 		tlck->next = jfs_ip->atlhead;
next             1031 fs/jfs/jfs_txnmgr.c 		if (tlck->next == 0) {
next             1043 fs/jfs/jfs_txnmgr.c 	maplock->next = 0;
next             1071 fs/jfs/jfs_txnmgr.c 	linelock->next = 0;
next             1079 fs/jfs/jfs_txnmgr.c 	linelock->next = tlock->next;
next             1080 fs/jfs/jfs_txnmgr.c 	tlock->next = lid;
next             1236 fs/jfs/jfs_txnmgr.c 			lid_to_tlock(jfs_ip->atltail)->next = tblk->next;
next             1237 fs/jfs/jfs_txnmgr.c 			tblk->next = jfs_ip->atlhead;
next             1379 fs/jfs/jfs_txnmgr.c 	for (lid = tblk->next; lid; lid = tlck->next) {
next             1698 fs/jfs/jfs_txnmgr.c 	int next, lwm, hwm;
next             1717 fs/jfs/jfs_txnmgr.c 	next = le16_to_cpu(p->header.nextindex);
next             1751 fs/jfs/jfs_txnmgr.c 		if (lwm == next)
next             1753 fs/jfs/jfs_txnmgr.c 		if (lwm > next) {
next             1759 fs/jfs/jfs_txnmgr.c 		xadlock->count = next - lwm;
next             1955 fs/jfs/jfs_txnmgr.c 		if (twm == next - 1) {
next             1975 fs/jfs/jfs_txnmgr.c 		if (hwm >= next) {
next             1986 fs/jfs/jfs_txnmgr.c 			    cpu_to_le16(hwm - next + 1);
next             1988 fs/jfs/jfs_txnmgr.c 			xtlck->header.offset = next;
next             1989 fs/jfs/jfs_txnmgr.c 			xtlck->header.length = hwm - next + 1;
next             2003 fs/jfs/jfs_txnmgr.c 		if (lwm < next) {
next             2011 fs/jfs/jfs_txnmgr.c 			xadlock->count = next - lwm;
next             2015 fs/jfs/jfs_txnmgr.c 				 tlck->ip, mp, xadlock->count, lwm, next);
next             2023 fs/jfs/jfs_txnmgr.c 		if (twm == next - 1) {
next             2044 fs/jfs/jfs_txnmgr.c 		if (hwm >= next) {
next             2051 fs/jfs/jfs_txnmgr.c 			xadlock->count = hwm - next + 1;
next             2052 fs/jfs/jfs_txnmgr.c 			xadlock->xdlist = &p->xad[next];
next             2055 fs/jfs/jfs_txnmgr.c 				 tlck->ip, mp, xadlock->count, next, hwm);
next             2212 fs/jfs/jfs_txnmgr.c 	lid_t lid, next;
next             2220 fs/jfs/jfs_txnmgr.c 	tlck = lid_to_tlock(tblk->next);
next             2221 fs/jfs/jfs_txnmgr.c 	lid = tlck->next;
next             2222 fs/jfs/jfs_txnmgr.c 	tlck->next = 0;
next             2225 fs/jfs/jfs_txnmgr.c 		next = tlck->next;
next             2226 fs/jfs/jfs_txnmgr.c 		tlck->next = tblk->next;
next             2227 fs/jfs/jfs_txnmgr.c 		tblk->next = lid;
next             2228 fs/jfs/jfs_txnmgr.c 		lid = next;
next             2235 fs/jfs/jfs_txnmgr.c 	for (lid = tblk->next; lid; lid = next) {
next             2237 fs/jfs/jfs_txnmgr.c 		next = tlck->next;
next             2306 fs/jfs/jfs_txnmgr.c 	for (lid = tblk->next; lid; lid = tlck->next) {
next             2591 fs/jfs/jfs_txnmgr.c 	while ((lid = xtlck->next) != 0) {
next             2594 fs/jfs/jfs_txnmgr.c 			xtlck->next = tlck->next;
next             2627 fs/jfs/jfs_txnmgr.c 	lid_t lid, next;
next             2635 fs/jfs/jfs_txnmgr.c 	for (lid = tblk->next; lid; lid = next) {
next             2637 fs/jfs/jfs_txnmgr.c 		next = tlck->next;
next             2663 fs/jfs/jfs_txnmgr.c 	tblk->next = tblk->last = 0;
next             2854 fs/jfs/jfs_txnmgr.c 		jfs_ip = list_entry(TxAnchor.anon_list.next,
next             2925 fs/jfs/jfs_txnmgr.c 			jfs_ip = list_entry(TxAnchor.anon_list.next,
next               35 fs/jfs/jfs_txnmgr.h 	lid_t next;		/* index of first tlock of tid */
next               83 fs/jfs/jfs_txnmgr.h 	lid_t next;		/* 2: index next lockword on tid locklist
next              158 fs/jfs/jfs_txnmgr.h 	lid_t next;		/* 2: next linelock */
next              174 fs/jfs/jfs_txnmgr.h 	lid_t next;		/* 2: */
next              205 fs/jfs/jfs_txnmgr.h 	lid_t next;		/* 2: */
next              233 fs/jfs/jfs_txnmgr.h 	lid_t next;		/* 2: */
next              104 fs/jfs/jfs_xtree.c static int xtSearch(struct inode *ip, s64 xoff, s64 *next, int *cmpp,
next              144 fs/jfs/jfs_xtree.c 	s64 next, size, xoff, xend;
next              163 fs/jfs/jfs_xtree.c 	if ((rc = xtSearch(ip, lstart, &next, &cmp, &btstack, 0))) {
next              182 fs/jfs/jfs_xtree.c 		if (next)
next              183 fs/jfs/jfs_xtree.c 			*plen = min(next - lstart, llen);
next              240 fs/jfs/jfs_xtree.c 	s64 next = 0;
next              309 fs/jfs/jfs_xtree.c 						next = t64;
next              355 fs/jfs/jfs_xtree.c 				*nextp = next;
next              413 fs/jfs/jfs_xtree.c 					next = offsetXAD(&p->xad[index + 1]);
next              414 fs/jfs/jfs_xtree.c 				goto next;
next              430 fs/jfs/jfs_xtree.c 			next = offsetXAD(&p->xad[base]);
next              465 fs/jfs/jfs_xtree.c 				*nextp = next;
next              481 fs/jfs/jfs_xtree.c 	      next:
next              538 fs/jfs/jfs_xtree.c 	s64 next;
next              552 fs/jfs/jfs_xtree.c 	if ((rc = xtSearch(ip, xoff, &next, &cmp, &btstack, XT_INSERT)))
next              560 fs/jfs/jfs_xtree.c 	if ((cmp == 0) || (next && (xlen > next - xoff))) {
next             1023 fs/jfs/jfs_xtree.c 	nextbn = le64_to_cpu(sp->header.next);
next             1024 fs/jfs/jfs_xtree.c 	rp->header.next = cpu_to_le64(nextbn);
next             1026 fs/jfs/jfs_xtree.c 	sp->header.next = cpu_to_le64(rbn);
next             1266 fs/jfs/jfs_xtree.c 	rp->header.next = 0;
next             2002 fs/jfs/jfs_xtree.c 		bn = le64_to_cpu(p->header.next);
next             2176 fs/jfs/jfs_xtree.c 	s64 next;
next             2191 fs/jfs/jfs_xtree.c 	if ((rc = xtSearch(ip, xoff, &next, &cmp, &btstack, XT_INSERT)))
next             2202 fs/jfs/jfs_xtree.c 	if (next)
next             2203 fs/jfs/jfs_xtree.c 		xlen = min(xlen, (int)(next - xoff));
next             2716 fs/jfs/jfs_xtree.c 		if (p->header.next) {
next             2717 fs/jfs/jfs_xtree.c 			nextbn = le64_to_cpu(p->header.next);
next             2747 fs/jfs/jfs_xtree.c 			lp->header.next = cpu_to_le64(nxaddr);
next             2953 fs/jfs/jfs_xtree.c 				goto next;
next             2975 fs/jfs/jfs_xtree.c 	      next:
next             3005 fs/jfs/jfs_xtree.c 	nextbn = le64_to_cpu(p->header.next);
next             3045 fs/jfs/jfs_xtree.c 		p->header.next = le64_to_cpu(nextbn);
next             3244 fs/jfs/jfs_xtree.c 	if (p->header.next) {
next             3253 fs/jfs/jfs_xtree.c 		p->header.next = 0;
next             3539 fs/jfs/jfs_xtree.c 		if (tblk->next == lid)
next             3540 fs/jfs/jfs_xtree.c 			tblk->next = tlck->next;
next             3542 fs/jfs/jfs_xtree.c 			for (prev = lid_to_tlock(tblk->next);
next             3543 fs/jfs/jfs_xtree.c 			     prev->next != lid;
next             3544 fs/jfs/jfs_xtree.c 			     prev = lid_to_tlock(prev->next)) {
next             3545 fs/jfs/jfs_xtree.c 				assert(prev->next);
next             3547 fs/jfs/jfs_xtree.c 			prev->next = tlck->next;
next             3549 fs/jfs/jfs_xtree.c 		lid_to_tlock(tblk->last)->next = lid;
next             3550 fs/jfs/jfs_xtree.c 		tlck->next = 0;
next               73 fs/jfs/jfs_xtree.h 		__le64 next;	/* 8: */
next              117 fs/kernfs/file.c 		void *next = ops->seq_start(sf, ppos);
next              119 fs/kernfs/file.c 		if (next == ERR_PTR(-ENODEV))
next              120 fs/kernfs/file.c 			kernfs_seq_stop_active(sf, next);
next              121 fs/kernfs/file.c 		return next;
next              137 fs/kernfs/file.c 		void *next = ops->seq_next(sf, v, ppos);
next              139 fs/kernfs/file.c 		if (next == ERR_PTR(-ENODEV))
next              140 fs/kernfs/file.c 			kernfs_seq_stop_active(sf, next);
next              141 fs/kernfs/file.c 		return next;
next              172 fs/kernfs/file.c 	.next = kernfs_seq_next,
next              106 fs/libfs.c     	while ((p = p->next) != &dentry->d_subdirs) {
next              189 fs/libfs.c     	struct dentry *next = NULL;
next              202 fs/libfs.c     	while ((next = scan_positives(cursor, p, 1, next)) != NULL) {
next              203 fs/libfs.c     		if (!dir_emit(ctx, next->d_name.name, next->d_name.len,
next              204 fs/libfs.c     			      d_inode(next)->i_ino, dt_type(d_inode(next))))
next              207 fs/libfs.c     		p = &next->d_child;
next              210 fs/libfs.c     	if (next)
next              211 fs/libfs.c     		list_move_tail(&cursor->d_child, &next->d_child);
next              215 fs/libfs.c     	dput(next);
next              234 fs/lockd/clntlock.c 	struct file_lock *fl, *next;
next              260 fs/lockd/clntlock.c 	list_for_each_entry_safe(fl, next, &host->h_reclaim, fl_u.nfs_fl.list) {
next               42 fs/lockd/host.c #define for_each_host_safe(host, next, chain, table) \
next               45 fs/lockd/host.c 		hlist_for_each_entry_safe((host), (next), \
next              657 fs/lockd/host.c 	struct hlist_node *next;
next              671 fs/lockd/host.c 	for_each_host_safe(host, next, chain, nlm_server_hosts) {
next              315 fs/lockd/svclock.c 	struct nlm_block *block, *next;
next              319 fs/lockd/svclock.c 	list_for_each_entry_safe(block, next, &file->f_blocks, b_flist) {
next              995 fs/lockd/svclock.c 		block = list_entry(nlm_blocked.next, struct nlm_block, b_list);
next              256 fs/lockd/svcsubs.c 	struct hlist_node *next;
next              262 fs/lockd/svcsubs.c 		hlist_for_each_entry_safe(file, next, &nlm_files[i], f_list) {
next             2985 fs/locks.c     	.next	= locks_next,
next              139 fs/mbcache.c   		node = entry->e_hash_list.next;
next              149 fs/mbcache.c   		node = node->next;
next              394 fs/mbcache.c   	struct mb_cache_entry *entry, *next;
next              402 fs/mbcache.c   	list_for_each_entry_safe(entry, next, &cache->c_list, e_list) {
next              900 fs/namespace.c 	struct list_head *next = p->mnt_mounts.next;
next              901 fs/namespace.c 	if (next == &p->mnt_mounts) {
next              905 fs/namespace.c 			next = p->mnt_child.next;
next              906 fs/namespace.c 			if (next != &p->mnt_parent->mnt_mounts)
next              911 fs/namespace.c 	return list_entry(next, struct mount, mnt_child);
next             1293 fs/namespace.c 	.next	= m_next,
next             2882 fs/namespace.c 	struct mount *mnt, *next;
next             2897 fs/namespace.c 	list_for_each_entry_safe(mnt, next, mounts, mnt_expire) {
next             2923 fs/namespace.c 	struct list_head *next;
next             2927 fs/namespace.c 	next = this_parent->mnt_mounts.next;
next             2929 fs/namespace.c 	while (next != &this_parent->mnt_mounts) {
next             2930 fs/namespace.c 		struct list_head *tmp = next;
next             2933 fs/namespace.c 		next = tmp->next;
next             2953 fs/namespace.c 		next = this_parent->mnt_child.next;
next             3893 fs/namespace.c 				goto next;
next             3900 fs/namespace.c 	next:	;
next              235 fs/nfs/blocklayout/extent_tree.c 			struct pnfs_block_extent *next = ext_tree_next(be);
next              239 fs/nfs/blocklayout/extent_tree.c 			be = next;
next             1102 fs/nfs/client.c 	.next	= nfs_server_list_next,
next             1114 fs/nfs/client.c 	.next	= nfs_volume_list_next,
next              525 fs/nfs/delegation.c 		server = list_entry_rcu(clp->cl_superblocks.next,
next              532 fs/nfs/delegation.c 			delegation = list_entry_rcu(server->delegations.next,
next             2185 fs/nfs/dir.c   		cache = list_entry(head->next, struct nfs_access_entry, lru);
next             2195 fs/nfs/dir.c   	struct nfs_inode *nfsi, *next;
next             2200 fs/nfs/dir.c   	list_for_each_entry_safe(nfsi, next, &nfs_access_lru_list, access_cache_inode_lru) {
next             2209 fs/nfs/dir.c   		cache = list_entry(nfsi->access_cache_entry_lru.next,
next              401 fs/nfs/direct.c 		struct nfs_page *req = nfs_list_entry(hdr->pages.next);
next              422 fs/nfs/direct.c 		req = nfs_list_entry(head->next);
next              663 fs/nfs/direct.c 		req = nfs_list_entry(failed.next);
next              684 fs/nfs/direct.c 		req = nfs_list_entry(data->pages.next);
next              761 fs/nfs/direct.c 	struct nfs_page *req = nfs_list_entry(hdr->pages.next);
next              792 fs/nfs/direct.c 		req = nfs_list_entry(hdr->pages.next);
next              813 fs/nfs/direct.c 		req = nfs_list_entry(head->next);
next              157 fs/nfs/nfs4client.c 		dss = list_entry(clp->cl_ds_clients.next,
next              171 fs/nfs/nfs4client.c 		cp_state = list_entry(clp->pending_cb_stateids.next,
next               47 fs/nfs/nfs4session.c 		p = &(*p)->next;
next               51 fs/nfs/nfs4session.c 		*p = slot->next;
next              137 fs/nfs/nfs4session.c 		p = &slot->next;
next              284 fs/nfs/nfs4session.c 		p = &(*p)->next;
next               22 fs/nfs/nfs4session.h 	struct nfs4_slot	*next;
next             1094 fs/nfs/nfs4state.c 		struct nfs_seqid *next;
next             1096 fs/nfs/nfs4state.c 		next = list_first_entry(&sequence->list,
next             1098 fs/nfs/nfs4state.c 		rpc_wake_up_queued_task(&sequence->wait, next->task);
next               49 fs/nfs/pagelist.c 	hdr->req = nfs_list_entry(mirror->pg_list.next);
next              279 fs/nfs/pagelist.c 	struct nfs_page *tmp, *next;
next              286 fs/nfs/pagelist.c 		next = tmp->wb_this_page;
next              291 fs/nfs/pagelist.c 		tmp = next;
next              792 fs/nfs/pagelist.c 		req = nfs_list_entry(head->next);
next             1251 fs/nfs/pagelist.c 		struct nfs_page *req = nfs_list_entry(pages.next);
next              419 fs/nfs/pnfs.c  	struct pnfs_layout_segment *lseg, *next;
next              422 fs/nfs/pnfs.c  	list_for_each_entry_safe(lseg, next, &lo->plh_segs, pls_list)
next              677 fs/nfs/pnfs.c  	struct pnfs_layout_segment *lseg, *next;
next              684 fs/nfs/pnfs.c  	list_for_each_entry_safe(lseg, next, &lo->plh_segs, pls_list)
next              703 fs/nfs/pnfs.c  	struct pnfs_layout_segment *lseg, *next;
next              705 fs/nfs/pnfs.c  	list_for_each_entry_safe(lseg, next, &lo->plh_return_segs, pls_list) {
next              774 fs/nfs/pnfs.c  	struct pnfs_layout_hdr *lo, *next;
next              777 fs/nfs/pnfs.c  	list_for_each_entry_safe(lo, next, &server->layouts, plh_layouts) {
next              817 fs/nfs/pnfs.c  		lo = list_entry(layout_list->next, struct pnfs_layout_hdr,
next             1356 fs/nfs/pnfs.c  	struct pnfs_layout_segment *lseg, *next;
next             1402 fs/nfs/pnfs.c  	list_for_each_entry_safe(lseg, next, &lo->plh_segs, pls_list) {
next             2397 fs/nfs/pnfs.c  	struct pnfs_layout_segment *lseg, *next;
next             2404 fs/nfs/pnfs.c  	list_for_each_entry_safe(lseg, next, &lo->plh_segs, pls_list)
next              167 fs/nfs/read.c  		struct nfs_page *req = nfs_list_entry(hdr->pages.next);
next              226 fs/nfs/read.c  		req = nfs_list_entry(head->next);
next             1004 fs/nfs/write.c 		struct nfs_page *req = nfs_list_entry(hdr->pages.next);
next             1020 fs/nfs/write.c 			goto next;
next             1024 fs/nfs/write.c next:
next             1438 fs/nfs/write.c 		req = nfs_list_entry(head->next);
next             1747 fs/nfs/write.c 	struct nfs_page *first = nfs_list_entry(head->next);
next             1785 fs/nfs/write.c 		req = nfs_list_entry(page_list->next);
next             1848 fs/nfs/write.c 		req = nfs_list_entry(data->pages.next);
next             1864 fs/nfs/write.c 			goto next;
next             1875 fs/nfs/write.c 			goto next;
next             1881 fs/nfs/write.c 	next:
next             1235 fs/nfsd/export.c 	.next	= cache_seq_next_rcu,
next              658 fs/nfsd/filecache.c 	struct hlist_node	*next;
next              669 fs/nfsd/filecache.c 		hlist_for_each_entry_safe(nf, next, &nfb->nfb_head, nf_node) {
next             1868 fs/nfsd/nfs4proc.c 	struct nfsd4_op *next = &argp->ops[resp->opcnt];
next             1885 fs/nfsd/nfs4proc.c 	if (next->opnum == OP_ILLEGAL)
next             1887 fs/nfsd/nfs4proc.c 	nextd = OPDESC(next);
next             1392 fs/nfsd/nfs4state.c 		stp = list_entry(open_stp->st_locks.next,
next             1943 fs/nfsd/nfs4state.c 		ses = list_entry(clp->cl_sessions.next, struct nfsd4_session,
next             2013 fs/nfsd/nfs4state.c 		dp = list_entry(clp->cl_delegations.next, struct nfs4_delegation, dl_perclnt);
next             2019 fs/nfsd/nfs4state.c 		dp = list_entry(reaplist.next, struct nfs4_delegation, dl_recall_lru);
next             2024 fs/nfsd/nfs4state.c 		dp = list_entry(clp->cl_revoked.next, struct nfs4_delegation, dl_recall_lru);
next             2029 fs/nfsd/nfs4state.c 		oo = list_entry(clp->cl_openowners.next, struct nfs4_openowner, oo_perclient);
next             2497 fs/nfsd/nfs4state.c 	.next = states_next,
next             5217 fs/nfsd/nfs4state.c 	struct list_head *pos, *next, reaplist;
next             5231 fs/nfsd/nfs4state.c 	list_for_each_safe(pos, next, &nn->client_lru) {
next             5246 fs/nfsd/nfs4state.c 	list_for_each_safe(pos, next, &reaplist) {
next             5254 fs/nfsd/nfs4state.c 	list_for_each_safe(pos, next, &nn->del_recall_lru) {
next             6955 fs/nfsd/nfs4state.c 			crp = list_entry(nn->reclaim_str_hashtbl[i].next,
next             7084 fs/nfsd/nfs4state.c 	struct nfs4_client *clp, *next;
next             7093 fs/nfsd/nfs4state.c 	list_for_each_entry_safe(clp, next, &nn->client_lru, cl_lru) {
next             7102 fs/nfsd/nfs4state.c 	list_for_each_entry_safe(clp, next, &reaplist, cl_lru)
next             7210 fs/nfsd/nfs4state.c 	struct nfs4_ol_stateid *stp, *next;
next             7212 fs/nfsd/nfs4state.c 	list_for_each_entry_safe(stp, next, reaplist, st_locks) {
next             7269 fs/nfsd/nfs4state.c 	struct nfs4_openowner *oop, *next;
next             7277 fs/nfsd/nfs4state.c 	list_for_each_entry_safe(oop, next, &clp->cl_openowners, oo_perclient) {
next             7341 fs/nfsd/nfs4state.c 	struct nfs4_openowner *oop, *next;
next             7343 fs/nfsd/nfs4state.c 	list_for_each_entry_safe(oop, next, reaplist, oo_perclient) {
next             7400 fs/nfsd/nfs4state.c 	struct nfs4_delegation *dp, *next;
next             7408 fs/nfsd/nfs4state.c 	list_for_each_entry_safe(dp, next, &clp->cl_delegations, dl_perclnt) {
next             7470 fs/nfsd/nfs4state.c 	struct nfs4_delegation *dp, *next;
next             7472 fs/nfsd/nfs4state.c 	list_for_each_entry_safe(dp, next, reaplist, dl_recall_lru) {
next             7530 fs/nfsd/nfs4state.c 	struct nfs4_delegation *dp, *next;
next             7532 fs/nfsd/nfs4state.c 	list_for_each_entry_safe(dp, next, reaplist, dl_recall_lru) {
next             7576 fs/nfsd/nfs4state.c 	struct nfs4_client *clp, *next;
next             7585 fs/nfsd/nfs4state.c 	list_for_each_entry_safe(clp, next, &nn->client_lru, cl_lru) {
next             7682 fs/nfsd/nfs4state.c 			clp = list_entry(nn->conf_id_hashtbl[i].next, struct nfs4_client, cl_idhash);
next             7691 fs/nfsd/nfs4state.c 			clp = list_entry(nn->unconf_id_hashtbl[i].next, struct nfs4_client, cl_idhash);
next             7757 fs/nfsd/nfs4state.c 	struct list_head *pos, *next, reaplist;
next             7765 fs/nfsd/nfs4state.c 	list_for_each_safe(pos, next, &nn->del_recall_lru) {
next             7771 fs/nfsd/nfs4state.c 	list_for_each_safe(pos, next, &reaplist) {
next              235 fs/nfsd/nfs4xdr.c 	tb->next = argp->to_free;
next             2065 fs/nfsd/nfs4xdr.c 	char *str, *end, *next;
next             2083 fs/nfsd/nfs4xdr.c 			next = end + 1;
next             2084 fs/nfsd/nfs4xdr.c 			if (*end && (!*next || *next == sep)) {
next             2105 fs/nfsd/nfs4xdr.c 			end = next;
next             4091 fs/nfsd/nfs4xdr.c 	struct nfsd4_test_stateid_id *stateid, *next;
next             4099 fs/nfsd/nfs4xdr.c 	list_for_each_entry_safe(stateid, next, &test_stateid->ts_stateid_list, ts_id_list) {
next             4545 fs/nfsd/nfs4xdr.c 		args->to_free = tb->next;
next              872 fs/nfsd/nfssvc.c 	struct svc_xprt *perm_sock = list_entry(rqstp->rq_server->sv_permsocks.next, typeof(struct svc_xprt), xpt_list);
next              641 fs/nfsd/xdr4.h 	struct svcxdr_tmpbuf *next;
next              483 fs/nilfs2/cpfile.c 	__u64 curr = *cnop, next;
next              525 fs/nilfs2/cpfile.c 		next = le64_to_cpu(cp->cp_snapshot_list.ssl_next);
next              526 fs/nilfs2/cpfile.c 		if (next == 0)
next              529 fs/nilfs2/cpfile.c 		next_blkoff = nilfs_cpfile_get_blkoff(cpfile, next);
next              533 fs/nilfs2/cpfile.c 			ret = nilfs_cpfile_get_checkpoint_block(cpfile, next,
next              541 fs/nilfs2/cpfile.c 		curr = next;
next              742 fs/nilfs2/cpfile.c 	__u64 next, prev;
next              767 fs/nilfs2/cpfile.c 	next = le64_to_cpu(list->ssl_next);
next              774 fs/nilfs2/cpfile.c 	if (next != 0) {
next              775 fs/nilfs2/cpfile.c 		ret = nilfs_cpfile_get_checkpoint_block(cpfile, next, 0,
next              795 fs/nilfs2/cpfile.c 		cpfile, next, next_bh, kaddr);
next              802 fs/nilfs2/cpfile.c 	list->ssl_next = cpu_to_le64(next);
next              288 fs/nilfs2/mdt.c 	__u64 next;
next              302 fs/nilfs2/mdt.c 	ret = nilfs_bmap_seek_key(NILFS_I(inode)->i_bmap, start + 1, &next);
next              304 fs/nilfs2/mdt.c 		if (next <= end) {
next              305 fs/nilfs2/mdt.c 			ret = nilfs_mdt_read_block(inode, next, true, out_bh);
next              307 fs/nilfs2/mdt.c 				*blkoff = next;
next              539 fs/nilfs2/recovery.c 		goto next;
next              552 fs/nilfs2/recovery.c  next:
next               92 fs/nilfs2/segbuf.c 	segbuf->sb_sum.next = nilfs_get_segment_start_blocknr(nilfs, nextnum);
next              149 fs/nilfs2/segbuf.c 	bh_sum = list_entry(segbuf->sb_segsum_buffers.next,
next              158 fs/nilfs2/segbuf.c 	raw_sum->ss_next     = cpu_to_le64(segbuf->sb_sum.next);
next              177 fs/nilfs2/segbuf.c 	bh = list_entry(segbuf->sb_segsum_buffers.next, struct buffer_head,
next              205 fs/nilfs2/segbuf.c 	bh = list_entry(segbuf->sb_segsum_buffers.next, struct buffer_head,
next               41 fs/nilfs2/segbuf.h 	sector_t		next;
next               87 fs/nilfs2/segbuf.h #define NILFS_NEXT_SEGBUF(segbuf)  NILFS_LIST_SEGBUF((segbuf)->sb_list.next)
next               90 fs/nilfs2/segbuf.h #define NILFS_FIRST_SEGBUF(head)   NILFS_LIST_SEGBUF((head)->next)
next               91 fs/nilfs2/segbuf.h #define NILFS_SEGBUF_IS_LAST(segbuf, head)  ((segbuf)->sb_list.next == (head))
next               98 fs/nilfs2/segbuf.h 	(list_entry((head)->next, struct buffer_head, b_assoc_buffers))
next              100 fs/nilfs2/segbuf.h 	(list_entry((bh)->b_assoc_buffers.next, struct buffer_head, \
next              102 fs/nilfs2/segbuf.h #define NILFS_SEGBUF_BH_IS_LAST(bh, head)  ((bh)->b_assoc_buffers.next == head)
next              239 fs/nls/nls_base.c 	if (nls->next)
next              249 fs/nls/nls_base.c 		tmp = &(*tmp)->next;
next              251 fs/nls/nls_base.c 	nls->next = tables;
next              265 fs/nls/nls_base.c 			*tmp = nls->next;
next              269 fs/nls/nls_base.c 		tmp = &(*tmp)->next;
next              279 fs/nls/nls_base.c 	for (nls = tables; nls; nls = nls->next) {
next              255 fs/notify/fsnotify.c 		node = srcu_dereference(mark->obj_list.next,
next              824 fs/notify/mark.c 	struct fsnotify_mark *mark, *next;
next              834 fs/notify/mark.c 	list_for_each_entry_safe(mark, next, &private_destroy_list, g_list) {
next              861 fs/ntfs/aops.c 		struct buffer_head *next = bh->b_this_page;
next              866 fs/ntfs/aops.c 		bh = next;
next             7502 fs/ocfs2/alloc.c 	int ret = 0, count = 0, next;
next             7515 fs/ocfs2/alloc.c 		next = ocfs2_find_next_bit(bitmap, max, start);
next             7517 fs/ocfs2/alloc.c 		if ((next - start) >= minbits) {
next             7519 fs/ocfs2/alloc.c 						start, next - start);
next             7524 fs/ocfs2/alloc.c 			count += next - start;
next             7526 fs/ocfs2/alloc.c 		start = next + 1;
next              437 fs/ocfs2/aops.c 	struct buffer_head *next;
next              441 fs/ocfs2/aops.c 	    	block_start = block_end, bh = next)
next              443 fs/ocfs2/aops.c 		next = bh->b_this_page;
next              766 fs/ocfs2/cluster/heartbeat.c 		event = list_entry(o2hb_node_events.next,
next              161 fs/ocfs2/cluster/netdebug.c 	.next = nst_seq_next,
next              375 fs/ocfs2/cluster/netdebug.c 	.next = sc_seq_next,
next              717 fs/ocfs2/dir.c 			goto next;
next              725 fs/ocfs2/dir.c 			goto next;
next              740 fs/ocfs2/dir.c 	next:
next             3459 fs/ocfs2/dir.c 			goto next;
next             3469 fs/ocfs2/dir.c next:
next              608 fs/ocfs2/dlm/dlmdebug.c 	.next =		lockres_seq_next,
next              435 fs/ocfs2/dlm/dlmdomain.c 				iter = res->hash_node.next;
next             2932 fs/ocfs2/dlm/dlmmaster.c 	struct dlm_lock *lock, *next;
next             2939 fs/ocfs2/dlm/dlmmaster.c 		list_for_each_entry_safe(lock, next, queue, list) {
next              142 fs/ocfs2/dlm/dlmrecovery.c 	struct dlm_work_item *item, *next;
next              155 fs/ocfs2/dlm/dlmrecovery.c 	list_for_each_entry_safe(item, next, &tmp_list, list) {
next              775 fs/ocfs2/dlm/dlmrecovery.c 	struct dlm_reco_node_data *ndata, *next;
next              782 fs/ocfs2/dlm/dlmrecovery.c 	list_for_each_entry_safe(ndata, next, &tmplist, list) {
next             1047 fs/ocfs2/dlm/dlmrecovery.c 	struct dlm_lock_resource *res, *next;
next             1051 fs/ocfs2/dlm/dlmrecovery.c 	list_for_each_entry_safe(res, next, &dlm->reco.resources, recovering) {
next             2058 fs/ocfs2/dlm/dlmrecovery.c 	struct dlm_lock *lock, *next;
next             2077 fs/ocfs2/dlm/dlmrecovery.c 		list_for_each_entry_safe(lock, next, queue, list) {
next             2142 fs/ocfs2/dlm/dlmrecovery.c 	struct dlm_lock_resource *res, *next;
next             2146 fs/ocfs2/dlm/dlmrecovery.c 	list_for_each_entry_safe(res, next, &dlm->reco.resources, recovering) {
next             2264 fs/ocfs2/dlm/dlmrecovery.c 	struct dlm_lock *lock, *next;
next             2278 fs/ocfs2/dlm/dlmrecovery.c 	list_for_each_entry_safe(lock, next, &res->granted, list) {
next             2287 fs/ocfs2/dlm/dlmrecovery.c 	list_for_each_entry_safe(lock, next, &res->converting, list) {
next             2296 fs/ocfs2/dlm/dlmrecovery.c 	list_for_each_entry_safe(lock, next, &res->blocked, list) {
next              293 fs/ocfs2/dlm/dlmthread.c 		lockres = list_entry(dlm->purge_list.next,
next              368 fs/ocfs2/dlm/dlmthread.c 	target = list_entry(res->converting.next, struct dlm_lock, list);
next              439 fs/ocfs2/dlm/dlmthread.c 	target = list_entry(res->blocked.next, struct dlm_lock, list);
next              582 fs/ocfs2/dlm/dlmthread.c 		lock = list_entry(dlm->pending_asts.next,
next              626 fs/ocfs2/dlm/dlmthread.c 		lock = list_entry(dlm->pending_basts.next,
next              712 fs/ocfs2/dlm/dlmthread.c 			res = list_entry(dlm->dirty_list.next,
next             3232 fs/ocfs2/dlmglue.c 	.next =		ocfs2_dlm_seq_next,
next             4354 fs/ocfs2/dlmglue.c 		lockres = list_entry(osb->blocked_lock_list.next,
next             1485 fs/ocfs2/file.c 			goto next;
next             1495 fs/ocfs2/file.c next:
next              850 fs/ocfs2/move_extents.c 			goto next;
next              861 fs/ocfs2/move_extents.c 				goto next;
next              884 fs/ocfs2/move_extents.c next:
next              239 fs/ocfs2/quota_local.c 	struct ocfs2_quota_chunk *pos, *next;
next              241 fs/ocfs2/quota_local.c 	list_for_each_entry_safe(pos, next, head, qc_chunk) {
next              318 fs/ocfs2/quota_local.c 	struct ocfs2_recovery_chunk *next;
next              321 fs/ocfs2/quota_local.c 	list_for_each_entry_safe(rchunk, next, head, rc_list) {
next              471 fs/ocfs2/quota_local.c 	struct ocfs2_recovery_chunk *rchunk, *next;
next              476 fs/ocfs2/quota_local.c 	list_for_each_entry_safe(rchunk, next, &(rec->r_list[type]), rc_list) {
next             1360 fs/ocfs2/refcounttree.c 					   struct ocfs2_refcount_rec *next)
next             1363 fs/ocfs2/refcounttree.c 		ocfs2_get_ref_rec_low_cpos(next))
next              469 fs/ocfs2/reservations.c 	struct rb_node *prev, *next;
next              529 fs/ocfs2/reservations.c 		next = rb_first(root);
next              530 fs/ocfs2/reservations.c 		next_resv = rb_entry(next, struct ocfs2_alloc_reservation,
next              565 fs/ocfs2/reservations.c 		next = rb_next(prev);
next              566 fs/ocfs2/reservations.c 		if (next) {
next              567 fs/ocfs2/reservations.c 			next_resv = rb_entry(next,
next              585 fs/ocfs2/reservations.c 		trace_ocfs2_resv_find_window_next(next ? next_resv->r_start: -1,
next              586 fs/ocfs2/reservations.c 					next ? ocfs2_resv_end(next_resv) : -1);
next              606 fs/ocfs2/reservations.c 		if (!next)
next              609 fs/ocfs2/reservations.c 		prev = next;
next             1200 fs/ocfs2/super.c 	.next           = NULL
next              164 fs/omfs/dir.c  	__be64 *entry, next;
next              184 fs/omfs/dir.c  	next = oi->i_sibling;
next              198 fs/omfs/dir.c  	*entry = next;
next               37 fs/omfs/file.c 	u64 next, last;
next               45 fs/omfs/file.c 	next = inode->i_ino;
next               52 fs/omfs/file.c 	bh = omfs_bread(inode->i_sb, next);
next               61 fs/omfs/file.c 		if (omfs_is_bad(sbi, (struct omfs_header *) bh->b_data, next))
next               69 fs/omfs/file.c 		last = next;
next               70 fs/omfs/file.c 		next = be64_to_cpu(oe->e_next);
next               89 fs/omfs/file.c 		if (next == ~0)
next               92 fs/omfs/file.c 		bh = omfs_bread(inode->i_sb, next);
next              221 fs/omfs/file.c 	sector_t next, offset;
next              239 fs/omfs/file.c 	next = inode->i_ino;
next              243 fs/omfs/file.c 		if (omfs_is_bad(sbi, (struct omfs_header *) bh->b_data, next))
next              247 fs/omfs/file.c 		next = be64_to_cpu(oe->e_next);
next              262 fs/omfs/file.c 		if (next == ~0)
next              266 fs/omfs/file.c 		bh = omfs_bread(inode->i_sb, next);
next              140 fs/openpromfs/inode.c 	.next		= property_next,
next              228 fs/openpromfs/inode.c 		prop = prop->next;
next              311 fs/openpromfs/inode.c 		prop = prop->next;
next              320 fs/openpromfs/inode.c 		prop = prop->next;
next               61 fs/orangefs/devorangefs-req.c 	struct orangefs_kernel_op_s *op, *next;
next               68 fs/orangefs/devorangefs-req.c 				 next,
next               11 fs/orangefs/dir.c 	struct orangefs_dir_part *next;
next              141 fs/orangefs/dir.c 		if (part->next)
next              142 fs/orangefs/dir.c 			part = part->next;
next              148 fs/orangefs/dir.c 	new->next = NULL;
next              154 fs/orangefs/dir.c 		part->next = new;
next              221 fs/orangefs/dir.c 			goto next;
next              224 fs/orangefs/dir.c 			goto next;
next              235 fs/orangefs/dir.c next:
next              251 fs/orangefs/dir.c 	while (part->next && count) {
next              253 fs/orangefs/dir.c 		part = part->next;
next              276 fs/orangefs/dir.c 			part = part->next;
next              293 fs/orangefs/dir.c 			struct orangefs_dir_part *next = part->next;
next              295 fs/orangefs/dir.c 			part = next;
next              390 fs/orangefs/dir.c 		struct orangefs_dir_part *next = part->next;
next              392 fs/orangefs/dir.c 		part = next;
next              111 fs/orangefs/orangefs-debugfs.c 	.next	= help_next,
next              212 fs/orangefs/orangefs-mod.c 		struct orangefs_kernel_op_s *next;
next              216 fs/orangefs/orangefs-mod.c 					 next,
next              866 fs/overlayfs/copy_up.c 		struct dentry *next;
next              872 fs/overlayfs/copy_up.c 		next = dget(dentry);
next              875 fs/overlayfs/copy_up.c 			parent = dget_parent(next);
next              880 fs/overlayfs/copy_up.c 			dput(next);
next              881 fs/overlayfs/copy_up.c 			next = parent;
next              884 fs/overlayfs/copy_up.c 		err = ovl_copy_up_one(parent, next, flags);
next              887 fs/overlayfs/copy_up.c 		dput(next);
next              107 fs/overlayfs/export.c 	struct dentry *next, *parent = NULL;
next              120 fs/overlayfs/export.c 	next = dget(dentry);
next              122 fs/overlayfs/export.c 		parent = dget_parent(next);
next              123 fs/overlayfs/export.c 		if (WARN_ON(parent == next)) {
next              133 fs/overlayfs/export.c 			err = ovl_encode_maybe_copy_up(next);
next              142 fs/overlayfs/export.c 		dput(next);
next              143 fs/overlayfs/export.c 		next = parent;
next              147 fs/overlayfs/export.c 	dput(next);
next              501 fs/overlayfs/export.c 	struct dentry *next, *parent = NULL;
next              508 fs/overlayfs/export.c 	next = dget(real);
next              510 fs/overlayfs/export.c 		parent = dget_parent(next);
next              516 fs/overlayfs/export.c 		ancestor = ovl_lookup_real_inode(sb, next, layer);
next              530 fs/overlayfs/export.c 		if (parent == next) {
next              535 fs/overlayfs/export.c 		dput(next);
next              536 fs/overlayfs/export.c 		next = parent;
next              540 fs/overlayfs/export.c 	dput(next);
next              562 fs/overlayfs/export.c 		struct dentry *next, *this;
next              571 fs/overlayfs/export.c 		next = dget(real);
next              573 fs/overlayfs/export.c 			parent = dget_parent(next);
next              598 fs/overlayfs/export.c 			if (parent == next) {
next              603 fs/overlayfs/export.c 			dput(next);
next              604 fs/overlayfs/export.c 			next = parent;
next              608 fs/overlayfs/export.c 			this = ovl_lookup_real_one(connected, next, layer);
next              634 fs/overlayfs/export.c 		dput(next);
next              294 fs/overlayfs/namei.c 		const char *next = strchrnul(s, '/');
next              295 fs/overlayfs/namei.c 		size_t thislen = next - s;
next              296 fs/overlayfs/namei.c 		bool end = !next[0];
next              303 fs/overlayfs/namei.c 					d->name.len - rem, next, &base);
next              367 fs/overlayfs/readdir.c 	int idx, next;
next              369 fs/overlayfs/readdir.c 	for (idx = 0; idx != -1; idx = next) {
next              370 fs/overlayfs/readdir.c 		next = ovl_path_next(idx, dentry, &realpath);
next              373 fs/overlayfs/readdir.c 		if (next != -1) {
next              777 fs/overlayfs/readdir.c 		od->cursor = p->l_node.next;
next             1122 fs/overlayfs/readdir.c 			goto next;
next             1147 fs/overlayfs/readdir.c next:
next             1506 fs/overlayfs/super.c 	struct dentry *next = dentry, *parent;
next             1512 fs/overlayfs/super.c 	parent = dget_parent(next);
next             1515 fs/overlayfs/super.c 	while (!err && parent != next) {
next             1522 fs/overlayfs/super.c 		next = parent;
next             1523 fs/overlayfs/super.c 		parent = dget_parent(next);
next             1524 fs/overlayfs/super.c 		dput(next);
next              911 fs/overlayfs/util.c 	char *s, *next, *buf = NULL;
next              922 fs/overlayfs/util.c 		for (s = buf; *s++ == '/'; s = next) {
next              923 fs/overlayfs/util.c 			next = strchrnul(s, '/');
next              924 fs/overlayfs/util.c 			if (s == next)
next               19 fs/pnode.c     	return list_entry(p->mnt_share.next, struct mount, mnt_share);
next               24 fs/pnode.c     	return list_entry(p->mnt_slave_list.next, struct mount, mnt_slave);
next               34 fs/pnode.c     	return list_entry(p->mnt_slave.next, struct mount, mnt_slave);
next              160 fs/pnode.c     			struct mount *next = next_peer(m);
next              161 fs/pnode.c     			return (next == origin) ? NULL : next;
next              162 fs/pnode.c     		} else if (m->mnt_slave.next != &master->mnt_slave_list)
next              187 fs/pnode.c     			struct mount *next;
next              190 fs/pnode.c     			next = next_peer(m);
next              192 fs/pnode.c     				if (next == origin)
next              194 fs/pnode.c     			} else if (m->mnt_slave.next != &next->mnt_slave)
next              196 fs/pnode.c     			m = next;
next              201 fs/pnode.c     			if (m->mnt_slave.next != &master->mnt_slave_list)
next              206 fs/pnode.c     			if (master->mnt_slave.next == &m->mnt_slave)
next              751 fs/proc/array.c 	.next	= children_seq_next,
next             2333 fs/proc/base.c 	.next	= timers_next,
next               78 fs/proc/consoles.c 	return con->next;
next               88 fs/proc/consoles.c 	.next	= c_next,
next               49 fs/proc/devices.c 	.next  = devinfo_next,
next              160 fs/proc/generic.c 	const char     		*cp = name, *next;
next              169 fs/proc/generic.c 		next = strchr(cp, '/');
next              170 fs/proc/generic.c 		if (!next)
next              173 fs/proc/generic.c 		len = next - cp;
next              305 fs/proc/generic.c 		struct proc_dir_entry *next;
next              315 fs/proc/generic.c 		next = pde_subdir_next(de);
next              317 fs/proc/generic.c 		de = next;
next              691 fs/proc/generic.c 	struct proc_dir_entry *root = NULL, *de, *next;
next              711 fs/proc/generic.c 		next = pde_subdir_first(de);
next              712 fs/proc/generic.c 		if (next) {
next              713 fs/proc/generic.c 			rb_erase(&next->subdir_node, &de->subdir);
next              714 fs/proc/generic.c 			de = next;
next              720 fs/proc/generic.c 		next = de->parent;
next              722 fs/proc/generic.c 			next->nlink--;
next              729 fs/proc/generic.c 		de = next;
next               32 fs/proc/interrupts.c 	.next  = int_seq_next,
next              107 fs/proc/nommu.c 	.next	= nommu_region_list_next,
next               75 fs/proc/proc_tty.c 	if (&p->tty_drivers == tty_drivers.next) {
next               96 fs/proc/proc_tty.c 		dev_t next = MKDEV(MAJOR(from)+1, 0);
next               97 fs/proc/proc_tty.c 		show_tty_range(m, p, from, next - from);
next               98 fs/proc/proc_tty.c 		from = next;
next              124 fs/proc/proc_tty.c 	.next	= t_next,
next               97 fs/proc/stat.c 	unsigned int i, next = 0;
next              100 fs/proc/stat.c 		show_irq_gap(p, i - next);
next              102 fs/proc/stat.c 		next = i + 1;
next              104 fs/proc/stat.c 	show_irq_gap(p, nr_irqs - next);
next              205 fs/proc/task_mmu.c 	struct vm_area_struct *next;
next              208 fs/proc/task_mmu.c 	next = m_next_vma(priv, v);
next              209 fs/proc/task_mmu.c 	if (!next)
next              211 fs/proc/task_mmu.c 	return next;
next              372 fs/proc/task_mmu.c 	.next	= m_next,
next              909 fs/proc/task_mmu.c 	.next	= m_next,
next             1896 fs/proc/task_mmu.c 	.next   = m_next,
next              253 fs/proc/task_nommu.c 	.next	= m_next,
next              120 fs/pstore/inode.c 	.next	= pstore_ftrace_seq_next,
next             1046 fs/quota/dquot.c 	act_head = tofree_head->next;
next             1049 fs/quota/dquot.c 		act_head = act_head->next;
next              121 fs/reiserfs/bitmap.c 				      int off, int *next)
next              127 fs/reiserfs/bitmap.c 			*next = tmp;
next              130 fs/reiserfs/bitmap.c 			(*next) = off + 1;  /* inc offset to avoid looping. */
next              150 fs/reiserfs/bitmap.c 	int end, next;
next              194 fs/reiserfs/bitmap.c 				next = end;
next              203 fs/reiserfs/bitmap.c 			if (unfm && is_block_in_journal(s, bmap_n, end, &next))
next              269 fs/reiserfs/bitmap.c 			*beg = next;
next              553 fs/reiserfs/bitmap.c 		ei = list_entry(plist->next, struct reiserfs_inode_info,
next              195 fs/reiserfs/inode.c 	struct buffer_head *head, *next, *bh;
next              201 fs/reiserfs/inode.c 			next = bh->b_this_page;
next              205 fs/reiserfs/inode.c 			bh = next;
next             2672 fs/reiserfs/inode.c 		struct buffer_head *next = bh->b_this_page;
next             2678 fs/reiserfs/inode.c 		bh = next;
next             2731 fs/reiserfs/inode.c 		struct buffer_head *next = bh->b_this_page;
next             2738 fs/reiserfs/inode.c 		bh = next;
next             3161 fs/reiserfs/inode.c 	struct buffer_head *head, *bh, *next;
next             3180 fs/reiserfs/inode.c 		next = bh->b_this_page;
next             3195 fs/reiserfs/inode.c 		bh = next;
next              161 fs/reiserfs/journal.c 	struct list_head *entry = journal->j_bitmap_nodes.next;
next              261 fs/reiserfs/journal.c 	struct list_head *next = journal->j_bitmap_nodes.next;
next              264 fs/reiserfs/journal.c 	while (next != &journal->j_bitmap_nodes) {
next              265 fs/reiserfs/journal.c 		bn = list_entry(next, struct reiserfs_bitmap_node, list);
next              266 fs/reiserfs/journal.c 		list_del(next);
next              269 fs/reiserfs/journal.c 		next = journal->j_bitmap_nodes.next;
next              359 fs/reiserfs/journal.c 	head[0].next = head + 1;
next              362 fs/reiserfs/journal.c 		head[i].next = head + (i + 1);	/* if last one, overwrite it after the if */
next              364 fs/reiserfs/journal.c 	head[num_cnodes - 1].next = NULL;
next              385 fs/reiserfs/journal.c 	if (cn->next) {
next              386 fs/reiserfs/journal.c 		cn->next->prev = NULL;
next              388 fs/reiserfs/journal.c 	journal->j_cnode_free_list = cn->next;
next              406 fs/reiserfs/journal.c 	cn->next = journal->j_cnode_free_list;
next              588 fs/reiserfs/journal.c 		jl = JOURNAL_LIST_ENTRY(entry->next);
next              809 fs/reiserfs/journal.c 		jh = JH_ENTRY(list->next);
next              940 fs/reiserfs/journal.c 			entry = entry->next;
next             1222 fs/reiserfs/journal.c 		cn = cn->next;
next             1311 fs/reiserfs/journal.c 	entry = journal->j_journal_list.next;
next             1509 fs/reiserfs/journal.c 		cn = cn->next;
next             1559 fs/reiserfs/journal.c 			cn = cn->next;
next             1649 fs/reiserfs/journal.c 			goto next;
next             1673 fs/reiserfs/journal.c next:
next             1674 fs/reiserfs/journal.c 		cn = cn->next;
next             1712 fs/reiserfs/journal.c 		cn = cn->next;
next             1755 fs/reiserfs/journal.c 		entry = jl->j_list.next;
next             1815 fs/reiserfs/journal.c 		if (tjl->j_list.next == &journal->j_journal_list)
next             1817 fs/reiserfs/journal.c 		tjl = JOURNAL_LIST_ENTRY(tjl->j_list.next);
next             3360 fs/reiserfs/journal.c 	cn->next = NULL;
next             3364 fs/reiserfs/journal.c 		journal->j_last->next = cn;
next             3429 fs/reiserfs/journal.c 		cn->prev->next = cn->next;
next             3431 fs/reiserfs/journal.c 	if (cn->next) {
next             3432 fs/reiserfs/journal.c 		cn->next->prev = cn->prev;
next             3435 fs/reiserfs/journal.c 		journal->j_first = cn->next;
next             3976 fs/reiserfs/journal.c 	struct reiserfs_journal_cnode *cn, *next, *jl_cn;
next             4116 fs/reiserfs/journal.c 	for (i = 0, cn = journal->j_first; cn; cn = cn->next, i++) {
next             4127 fs/reiserfs/journal.c 			jl_cn->next = NULL;
next             4129 fs/reiserfs/journal.c 				last_cn->next = jl_cn;
next             4219 fs/reiserfs/journal.c 		next = cn->next;
next             4221 fs/reiserfs/journal.c 		cn = next;
next             1400 fs/reiserfs/lbalance.c 		int prev, next;
next             1404 fs/reiserfs/lbalance.c 			next =
next             1416 fs/reiserfs/lbalance.c 			if (next && next >= deh_location(&deh[i]))
next              238 fs/reiserfs/reiserfs.h 	struct reiserfs_journal_cnode *next;	/* next in transaction list */
next             2943 fs/reiserfs/reiserfs.h 			 int bit_nr, int searchall, b_blocknr_t *next);
next             1505 fs/reiserfs/stree.c 	struct buffer_head *next;
next             1516 fs/reiserfs/stree.c 				next = bh->b_this_page;
next             1530 fs/reiserfs/stree.c 				bh = next;
next              196 fs/romfs/super.c 		nextfh = be32_to_cpu(ri.next);
next              249 fs/romfs/super.c 			if ((be32_to_cpu(ri.next) & ROMFH_TYPE) == ROMFH_HRD)
next              256 fs/romfs/super.c 		offset = be32_to_cpu(ri.next) & ROMFH_MASK;
next              297 fs/romfs/super.c 		nextfh = be32_to_cpu(ri.next);
next               98 fs/select.c    	struct poll_table_page * next;
next              154 fs/select.c    		p = p->next;
next              176 fs/select.c    		new_table->next = table;
next              827 fs/select.c    	struct poll_list *next;
next              894 fs/select.c    		for (walk = list; walk != NULL; walk = walk->next) {
next              978 fs/select.c    		walk->next = NULL;
next              992 fs/select.c    		walk = walk->next = kmalloc(struct_size(walk, entries, len),
next             1004 fs/select.c    	for (walk = head; walk; walk = walk->next) {
next             1015 fs/select.c    	walk = head->next;
next             1018 fs/select.c    		walk = walk->next;
next              122 fs/seq_file.c  		p = m->op->next(m, p, &m->index);
next              235 fs/seq_file.c  			p = m->op->next(m, p, &m->index);
next              258 fs/seq_file.c  		p = m->op->next(m, p, &m->index);
next              574 fs/seq_file.c  		op->next = single_next;
next              921 fs/seq_file.c  	lh = ((struct list_head *)v)->next;
next              979 fs/seq_file.c  		return node->next;
next             1050 fs/seq_file.c  		return rcu_dereference(node->next);
next             1094 fs/seq_file.c  	if (node->next)
next             1095 fs/seq_file.c  		return node->next;
next              714 fs/ubifs/commit.c 		i = list_entry(list.next, struct idx_node, list);
next             2306 fs/ubifs/debug.c 	for (cur = head->next; cur->next != head; cur = cur->next) {
next             2312 fs/ubifs/debug.c 		sb = container_of(cur->next, struct ubifs_scan_node, list);
next             2373 fs/ubifs/debug.c 	for (cur = head->next; cur->next != head; cur = cur->next) {
next             2379 fs/ubifs/debug.c 		sb = container_of(cur->next, struct ubifs_scan_node, list);
next              546 fs/ubifs/gc.c  	snod = list_entry(sleb->nodes.next, struct ubifs_scan_node, list);
next              978 fs/ubifs/gc.c  		idx_gc = list_entry(c->idx_gc.next, struct ubifs_gced_idx_leb,
next              999 fs/ubifs/gc.c  	idx_gc = list_entry(c->idx_gc.next, struct ubifs_gced_idx_leb, list);
next              522 fs/ubifs/log.c 		bud = list_entry(c->old_buds.next, struct ubifs_bud, list);
next              787 fs/ubifs/lprops.c 	lprops = list_entry(c->empty_list.next, struct ubifs_lprops, list);
next              810 fs/ubifs/lprops.c 	lprops = list_entry(c->freeable_list.next, struct ubifs_lprops, list);
next              834 fs/ubifs/lprops.c 	lprops = list_entry(c->frdi_idx_list.next, struct ubifs_lprops, list);
next              878 fs/ubifs/recovery.c 			snod = list_entry(sleb->nodes.next,
next             1082 fs/ubifs/recovery.c 		ucleb = list_entry(c->unclean_leb_list.next,
next              516 fs/ubifs/replay.c 	struct ubifs_bud *next;
next              550 fs/ubifs/replay.c 	next = list_entry(bud->list.next, struct ubifs_bud, list);
next              551 fs/ubifs/replay.c 	if (!list_is_last(&next->list, &jh->buds_list))
next              554 fs/ubifs/replay.c 	err = ubifs_leb_read(c, next->lnum, (char *)&data, next->start, 4, 1);
next              877 fs/ubifs/replay.c 		b = list_entry(c->replay_buds.next, struct bud_entry, list);
next             1017 fs/ubifs/replay.c 	snod = list_entry(sleb->nodes.next, struct ubifs_scan_node, list);
next              361 fs/ubifs/scan.c 		node = list_entry(head->next, struct ubifs_scan_node, list);
next              156 fs/ubifs/shrinker.c 	p = ubifs_infos.next;
next              168 fs/ubifs/shrinker.c 			p = p->next;
next              178 fs/ubifs/shrinker.c 			p = p->next;
next              191 fs/ubifs/shrinker.c 		p = p->next;
next              878 fs/ubifs/super.c 		orph = list_entry(c->orph_list.next, struct ubifs_orphan, list);
next             1138 fs/ubifs/super.c 		ucleb = list_entry(c->unclean_leb_list.next,
next             1146 fs/ubifs/super.c 		bud = list_entry(c->old_buds.next, struct ubifs_bud, list);
next             1018 fs/udf/inode.c 		int next = laarr[start].extLocation.logicalBlockNum +
next             1024 fs/udf/inode.c 				next, (UDF_DEFAULT_PREALLOC_BLOCKS > length ?
next             1036 fs/udf/inode.c 				laarr[c + 1].extLocation.logicalBlockNum = next;
next               57 fs/udf/unicode.c 		unicode_t next;
next               72 fs/udf/unicode.c 		next = str_i[str_i_idx++] << 8;
next               73 fs/udf/unicode.c 		next |= str_i[str_i_idx++];
next               74 fs/udf/unicode.c 		if ((next & SURROGATE_MASK) != SURROGATE_PAIR ||
next               75 fs/udf/unicode.c 		    !(next & SURROGATE_LOW)) {
next               82 fs/udf/unicode.c 		    (next & SURROGATE_CHAR_MASK);
next              350 fs/unicode/mkutf8data.c 	struct tree *next;
next             1050 fs/unicode/mkutf8data.c 	struct tree *next;
next             1094 fs/unicode/mkutf8data.c 				next = tree->next;
next             1096 fs/unicode/mkutf8data.c 					assert(next);
next             1097 fs/unicode/mkutf8data.c 					n = next->root;
next             1116 fs/unicode/mkutf8data.c 					next = next->next;
next             1650 fs/unicode/mkutf8data.c 	trees[trees_count-2].next = &trees[trees_count-1];
next             1654 fs/unicode/mkutf8data.c 		trees[i].next = &trees[trees_count-2];
next             1656 fs/unicode/mkutf8data.c 		trees[i+1].next = &trees[trees_count-1];
next             1470 fs/userfaultfd.c 			goto next;
next             1482 fs/userfaultfd.c 	next:
next             1632 fs/userfaultfd.c 			goto next;
next             1644 fs/userfaultfd.c 	next:
next             5820 fs/xfs/libxfs/xfs_bmap.c 	struct xfs_bmbt_irec	got, next;
next             5872 fs/xfs/libxfs/xfs_bmap.c 	if (xfs_iext_peek_next_extent(ifp, &icur, &next)) {
next             5873 fs/xfs/libxfs/xfs_bmap.c 		if (new_startoff + got.br_blockcount > next.br_startoff) {
next             5884 fs/xfs/libxfs/xfs_bmap.c 		if (xfs_bmse_can_merge(&got, &next, offset_shift_fsb))
next             1870 fs/xfs/libxfs/xfs_ialloc.c 			goto next;
next             1892 fs/xfs/libxfs/xfs_ialloc.c next:
next              125 fs/xfs/libxfs/xfs_iext_tree.c 	struct xfs_iext_leaf	*next;
next              241 fs/xfs/libxfs/xfs_iext_tree.c 	    cur->leaf->next) {
next              242 fs/xfs/libxfs/xfs_iext_tree.c 		cur->leaf = cur->leaf->next;
next              571 fs/xfs/libxfs/xfs_iext_tree.c 	if (leaf->next)
next              572 fs/xfs/libxfs/xfs_iext_tree.c 		leaf->next->prev = new;
next              573 fs/xfs/libxfs/xfs_iext_tree.c 	new->next = leaf->next;
next              575 fs/xfs/libxfs/xfs_iext_tree.c 	leaf->next = new;
next              700 fs/xfs/libxfs/xfs_iext_tree.c 		struct xfs_iext_node *next = parent->ptrs[*pos + 1];
next              701 fs/xfs/libxfs/xfs_iext_tree.c 		int nr_next = xfs_iext_node_nr_entries(next, 0), i;
next              710 fs/xfs/libxfs/xfs_iext_tree.c 				node->keys[nr_entries + i] = next->keys[i];
next              711 fs/xfs/libxfs/xfs_iext_tree.c 				node->ptrs[nr_entries + i] = next->ptrs[i];
next              715 fs/xfs/libxfs/xfs_iext_tree.c 			return next;
next              818 fs/xfs/libxfs/xfs_iext_tree.c 	if (leaf->next) {
next              819 fs/xfs/libxfs/xfs_iext_tree.c 		int nr_next = xfs_iext_leaf_nr_entries(ifp, leaf->next, 0), i;
next              829 fs/xfs/libxfs/xfs_iext_tree.c 					leaf->next->recs[i];
next              832 fs/xfs/libxfs/xfs_iext_tree.c 			if (cur->leaf == leaf->next) {
next              837 fs/xfs/libxfs/xfs_iext_tree.c 			offset = xfs_iext_leaf_key(leaf->next, 0);
next              838 fs/xfs/libxfs/xfs_iext_tree.c 			leaf = leaf->next;
next              846 fs/xfs/libxfs/xfs_iext_tree.c 		leaf->prev->next = leaf->next;
next              847 fs/xfs/libxfs/xfs_iext_tree.c 	if (leaf->next)
next              848 fs/xfs/libxfs/xfs_iext_tree.c 		leaf->next->prev = leaf->prev;
next              891 fs/xfs/libxfs/xfs_iext_tree.c 		if (ifp->if_height > 1 && leaf->next)
next              892 fs/xfs/libxfs/xfs_iext_tree.c 			cur->leaf = leaf->next;
next              944 fs/xfs/libxfs/xfs_iext_tree.c 	if (ifp->if_height == 1 || !cur->leaf->next)
next              946 fs/xfs/libxfs/xfs_iext_tree.c 	cur->leaf = cur->leaf->next;
next              129 fs/xfs/scrub/bitmap.c 	lp = bitmap->list.next;
next              143 fs/xfs/scrub/bitmap.c 			lp = lp->next;
next              171 fs/xfs/scrub/bitmap.c 			lp = lp->next;
next              175 fs/xfs/scrub/bitmap.c 			lp = lp->next;
next              195 fs/xfs/scrub/bitmap.c 			lp = lp->next;
next               90 fs/xfs/xfs_aops.c 	struct bio		*last = ioend->io_bio, *next;
next               94 fs/xfs/xfs_aops.c 	for (bio = &ioend->io_inline_bio; bio; bio = next) {
next              103 fs/xfs/xfs_aops.c 			next = NULL;
next              105 fs/xfs/xfs_aops.c 			next = bio->bi_private;
next              297 fs/xfs/xfs_aops.c 	struct xfs_ioend	*next)
next              299 fs/xfs/xfs_aops.c 	if (ioend->io_bio->bi_status != next->io_bio->bi_status)
next              301 fs/xfs/xfs_aops.c 	if ((ioend->io_fork == XFS_COW_FORK) ^ (next->io_fork == XFS_COW_FORK))
next              304 fs/xfs/xfs_aops.c 	    (next->io_state == XFS_EXT_UNWRITTEN))
next              306 fs/xfs/xfs_aops.c 	if (ioend->io_offset + ioend->io_size != next->io_offset)
next              320 fs/xfs/xfs_aops.c 	struct xfs_ioend	*next)
next              323 fs/xfs/xfs_aops.c 		ioend->io_append_trans = next->io_append_trans;
next              324 fs/xfs/xfs_aops.c 		next->io_append_trans = NULL;
next              326 fs/xfs/xfs_aops.c 		xfs_setfilesize_ioend(next, -ECANCELED);
next              876 fs/xfs/xfs_aops.c 	struct xfs_ioend	*ioend, *next;
next              951 fs/xfs/xfs_aops.c 	list_for_each_entry_safe(ioend, next, &submit_list, io_list) {
next              494 fs/xfs/xfs_log_cil.c 		struct xfs_log_vec *next = lv->lv_next;
next              496 fs/xfs/xfs_log_cil.c 		lv = next;
next             1001 fs/xfs/xfs_log_cil.c 	struct xfs_log_item	*lip, *next;
next             1036 fs/xfs/xfs_log_cil.c 	list_for_each_entry_safe(lip, next, &tp->t_items, li_trans) {
next             2619 fs/xfs/xfs_log_recover.c 				goto next;
next             2625 fs/xfs/xfs_log_recover.c 				goto next;
next             2633 fs/xfs/xfs_log_recover.c 				goto next;
next             2641 fs/xfs/xfs_log_recover.c  next:
next             4109 fs/xfs/xfs_log_recover.c 	struct xlog_recover_item	*next;
next             4121 fs/xfs/xfs_log_recover.c 	list_for_each_entry_safe(item, next, &trans->r_itemq, ri_list) {
next              232 fs/xfs/xfs_mru_cache.c 	struct xfs_mru_cache_elem *elem, *next;
next              236 fs/xfs/xfs_mru_cache.c 	list_for_each_entry_safe(elem, next, &mru->reap_list, list_node) {
next              249 fs/xfs/xfs_mru_cache.c 	list_for_each_entry_safe(elem, next, &tmp, list_node) {
next              270 fs/xfs/xfs_mru_cache.c 	unsigned long		now, next;
next              277 fs/xfs/xfs_mru_cache.c 	next = _xfs_mru_cache_migrate(mru, jiffies);
next              280 fs/xfs/xfs_mru_cache.c 	mru->queued = next;
next              283 fs/xfs/xfs_mru_cache.c 		if (next <= now)
next              284 fs/xfs/xfs_mru_cache.c 			next = 0;
next              286 fs/xfs/xfs_mru_cache.c 			next -= now;
next              287 fs/xfs/xfs_mru_cache.c 		queue_delayed_work(xfs_mru_reap_wq, &mru->work, next);
next             1424 fs/xfs/xfs_reflink.c 			goto next;
next             1457 fs/xfs/xfs_reflink.c next:
next             1495 fs/xfs/xfs_reflink.c 			goto next;
next             1509 fs/xfs/xfs_reflink.c next:
next              239 fs/xfs/xfs_rtalloc.c 	xfs_rtblock_t	next;		/* next rtblock to try */
next              254 fs/xfs/xfs_rtalloc.c 		error = xfs_rtcheck_range(mp, tp, i, maxlen, 1, &next, &stat);
next              280 fs/xfs/xfs_rtalloc.c 			thislen = next - i;
next              289 fs/xfs/xfs_rtalloc.c 		if (next < end) {
next              290 fs/xfs/xfs_rtalloc.c 			error = xfs_rtfind_forw(mp, tp, next, end, &i);
next              326 fs/xfs/xfs_rtalloc.c 	*nextp = next;
next              353 fs/xfs/xfs_rtalloc.c 	xfs_rtblock_t	next;		/* next block to try (dummy) */
next              359 fs/xfs/xfs_rtalloc.c 	error = xfs_rtcheck_range(mp, tp, bno, maxlen, 1, &next, &isfree);
next              378 fs/xfs/xfs_rtalloc.c 	maxlen = next - bno;
next              772 fs/xfs/xfs_trans.c 	struct xfs_log_item	*lip, *next;
next              776 fs/xfs/xfs_trans.c 	list_for_each_entry_safe(lip, next, &tp->t_items, li_trans) {
next               54 fs/xfs/xfs_trans_ail.c 	next_lip = list_entry(lip->li_ail.next, struct xfs_log_item, li_ail);
next               97 fs/xfs/xfs_trans_ail.c 	if (lip->li_ail.next == &ailp->ail_head)
next               36 include/asm-generic/mmu_context.h 			struct mm_struct *next,
next               21 include/asm-generic/switch_to.h #define switch_to(prev, next, last)					\
next               23 include/asm-generic/switch_to.h 		((last) = __switch_to((prev), (next)));			\
next              210 include/asm-generic/tlb.h 	struct mmu_gather_batch	*next;
next              384 include/drm/drm_bridge.h 	struct drm_bridge *next;
next               92 include/drm/drm_dp_mst_helper.h 	struct list_head next;
next              396 include/drm/drm_dp_mst_helper.h 	struct list_head next;
next              431 include/drm/drm_dp_mst_helper.h 	struct list_head next;
next               61 include/drm/drm_legacy.h 	struct drm_buf *next;	       /**< Kernel-only: used for free list */
next              274 include/drm/drm_mm.h 	return mm->hole_stack.next;
next              371 include/drm/drm_mm.h #define drm_mm_for_each_node_safe(entry, next, mm) \
next              372 include/drm/drm_mm.h 	list_for_each_entry_safe(entry, next, drm_mm_nodes(mm), node_list)
next               35 include/drm/spsc_queue.h 	struct spsc_node *next;
next               69 include/drm/spsc_queue.h 	node->next = NULL;
next               73 include/drm/spsc_queue.h 	tail = (struct spsc_node **)atomic_long_xchg(&queue->tail, (long)&node->next);
next               91 include/drm/spsc_queue.h 	struct spsc_node *next, *node;
next              101 include/drm/spsc_queue.h 	next = READ_ONCE(node->next);
next              102 include/drm/spsc_queue.h 	WRITE_ONCE(queue->head, next);
next              104 include/drm/spsc_queue.h 	if (unlikely(!next)) {
next              108 include/drm/spsc_queue.h 				(long)&node->next, (long) &queue->head) != (long)&node->next) {
next              112 include/drm/spsc_queue.h 			} while (unlikely(!(queue->head = READ_ONCE(node->next))));
next               87 include/keys/rxrpc-type.h 	struct rxrpc_key_token *next;	/* the next token in the list */
next               21 include/linux/adb.h 	struct adb_request *next;
next               70 include/linux/agp_backend.h 	struct agp_memory *next;
next               92 include/linux/agpgart.h 	struct agp_client *next;
next              100 include/linux/agpgart.h 	struct agp_controller *next;
next              115 include/linux/agpgart.h 	struct agp_file_private *next;
next               14 include/linux/atalk.h 	struct atalk_route *next;
next               32 include/linux/atalk.h 	struct atalk_iface	*next;
next              279 include/linux/bpf_verifier.h 	struct bpf_verifier_state_list *next;
next              375 include/linux/compat.h 	compat_uptr_t			next;
next              157 include/linux/console.h 	struct	 console *next;
next              164 include/linux/console.h 	for (con = console_drivers; con != NULL; con = con->next)
next               83 include/linux/debugobjects.h 			  unsigned int expect, unsigned int next);
next               24 include/linux/dm-io.h 	struct page_list *next;
next              508 include/linux/dmaengine.h 	struct dma_async_tx_descriptor *next;
next              555 include/linux/dmaengine.h static inline void txd_chain(struct dma_async_tx_descriptor *txd, struct dma_async_tx_descriptor *next)
next              583 include/linux/dmaengine.h static inline void txd_chain(struct dma_async_tx_descriptor *txd, struct dma_async_tx_descriptor *next)
next              585 include/linux/dmaengine.h 	txd->next = next;
next              586 include/linux/dmaengine.h 	next->parent = txd;
next              594 include/linux/dmaengine.h 	txd->next = NULL;
next              602 include/linux/dmaengine.h 	return txd->next;
next             1783 include/linux/efi.h 	phys_addr_t	next;			// pa of next struct instance
next               23 include/linux/frontswap.h 	struct frontswap_ops *next; /* private pointer to next ops */
next             2233 include/linux/fs.h 	struct file_system_type * next;
next              193 include/linux/ftrace.h 	struct ftrace_ops __rcu		*next;
next               30 include/linux/hdlc.h 	struct hdlc_proto *next; /* next protocol in the list */
next               34 include/linux/huge_mm.h 			pmd_t *pmd, unsigned long addr, unsigned long next);
next               48 include/linux/if_pppox.h 	struct pppox_sock	*next;	  /* for hash table */
next               71 include/linux/if_tap.h 	struct list_head next;
next               15 include/linux/if_tunnel.h 	for (pos = rcu_dereference(start); pos; pos = rcu_dereference(pos->next))
next              147 include/linux/if_vlan.h 	struct vlan_priority_tci_mapping	*next;
next              206 include/linux/if_vlan.h 		mp = mp->next;
next               78 include/linux/igmp.h 		struct ip_mc_list *next;
next              114 include/linux/interrupt.h 	struct irqaction	*next;
next              594 include/linux/interrupt.h 	struct tasklet_struct *next;
next              105 include/linux/jump_label.h 		struct static_key_mod *next;
next               28 include/linux/list.h 	WRITE_ONCE(list->next, list);
next               35 include/linux/list.h 			      struct list_head *next);
next               40 include/linux/list.h 				struct list_head *next)
next               58 include/linux/list.h 			      struct list_head *next)
next               60 include/linux/list.h 	if (!__list_add_valid(new, prev, next))
next               63 include/linux/list.h 	next->prev = new;
next               64 include/linux/list.h 	new->next = next;
next               66 include/linux/list.h 	WRITE_ONCE(prev->next, new);
next               79 include/linux/list.h 	__list_add(new, head, head->next);
next              103 include/linux/list.h static inline void __list_del(struct list_head * prev, struct list_head * next)
next              105 include/linux/list.h 	next->prev = prev;
next              106 include/linux/list.h 	WRITE_ONCE(prev->next, next);
next              119 include/linux/list.h 	__list_del(entry->prev, entry->next);
next              134 include/linux/list.h 	__list_del(entry->prev, entry->next);
next              140 include/linux/list.h 	entry->next = LIST_POISON1;
next              154 include/linux/list.h 	new->next = old->next;
next              155 include/linux/list.h 	new->next->prev = new;
next              157 include/linux/list.h 	new->prev->next = new;
next              230 include/linux/list.h 	first->prev->next = last->next;
next              231 include/linux/list.h 	last->next->prev = first->prev;
next              233 include/linux/list.h 	head->prev->next = first;
next              236 include/linux/list.h 	last->next = head;
next              259 include/linux/list.h 	return list->next == head;
next              268 include/linux/list.h 	return READ_ONCE(head->next) == head;
next              286 include/linux/list.h 	struct list_head *next = head->next;
next              287 include/linux/list.h 	return (next == head) && (next == head->prev);
next              299 include/linux/list.h 		first = head->next;
next              328 include/linux/list.h 	return !list_empty(head) && (head->next == head->prev);
next              334 include/linux/list.h 	struct list_head *new_first = entry->next;
next              335 include/linux/list.h 	list->next = head->next;
next              336 include/linux/list.h 	list->next->prev = list;
next              338 include/linux/list.h 	entry->next = list;
next              339 include/linux/list.h 	head->next = new_first;
next              363 include/linux/list.h 		(head->next != entry && head != entry))
next              389 include/linux/list.h 	if (head->next == entry) {
next              393 include/linux/list.h 	list->next = head->next;
next              394 include/linux/list.h 	list->next->prev = list;
next              396 include/linux/list.h 	list->prev->next = list;
next              397 include/linux/list.h 	head->next = entry;
next              403 include/linux/list.h 				 struct list_head *next)
next              405 include/linux/list.h 	struct list_head *first = list->next;
next              409 include/linux/list.h 	prev->next = first;
next              411 include/linux/list.h 	last->next = next;
next              412 include/linux/list.h 	next->prev = last;
next              424 include/linux/list.h 		__list_splice(list, head, head->next);
next              450 include/linux/list.h 		__list_splice(list, head, head->next);
next              490 include/linux/list.h 	list_entry((ptr)->next, type, member)
next              513 include/linux/list.h 	struct list_head *pos__ = READ_ONCE(head__->next); \
next              523 include/linux/list.h 	list_entry((pos)->member.next, typeof(*(pos)), member)
next              539 include/linux/list.h 	for (pos = (head)->next; pos != (head); pos = pos->next)
next              556 include/linux/list.h 	for (pos = (head)->next, n = pos->next; pos != (head); \
next              557 include/linux/list.h 		pos = n, n = pos->next)
next              743 include/linux/list.h 	h->next = NULL;
next              759 include/linux/list.h 	struct hlist_node *next = n->next;
next              762 include/linux/list.h 	WRITE_ONCE(*pprev, next);
next              763 include/linux/list.h 	if (next)
next              764 include/linux/list.h 		next->pprev = pprev;
next              770 include/linux/list.h 	n->next = LIST_POISON1;
next              785 include/linux/list.h 	n->next = first;
next              787 include/linux/list.h 		first->pprev = &n->next;
next              794 include/linux/list.h 					struct hlist_node *next)
next              796 include/linux/list.h 	n->pprev = next->pprev;
next              797 include/linux/list.h 	n->next = next;
next              798 include/linux/list.h 	next->pprev = &n->next;
next              805 include/linux/list.h 	n->next = prev->next;
next              806 include/linux/list.h 	prev->next = n;
next              807 include/linux/list.h 	n->pprev = &prev->next;
next              809 include/linux/list.h 	if (n->next)
next              810 include/linux/list.h 		n->next->pprev  = &n->next;
next              816 include/linux/list.h 	n->pprev = &n->next;
next              821 include/linux/list.h 	return h->pprev == &h->next;
next              831 include/linux/list.h 	return !n->next && n->pprev == &h->first;
next              850 include/linux/list.h 	for (pos = (head)->first; pos ; pos = pos->next)
next              853 include/linux/list.h 	for (pos = (head)->first; pos && ({ n = pos->next; 1; }); \
next              870 include/linux/list.h 	     pos = hlist_entry_safe((pos)->member.next, typeof(*(pos)), member))
next              878 include/linux/list.h 	for (pos = hlist_entry_safe((pos)->member.next, typeof(*(pos)), member);\
next              880 include/linux/list.h 	     pos = hlist_entry_safe((pos)->member.next, typeof(*(pos)), member))
next              889 include/linux/list.h 	     pos = hlist_entry_safe((pos)->member.next, typeof(*(pos)), member))
next              900 include/linux/list.h 	     pos && ({ n = pos->member.next; 1; });			\
next               39 include/linux/list_bl.h 	struct hlist_bl_node *next, **pprev;
next               46 include/linux/list_bl.h 	h->next = NULL;
next               82 include/linux/list_bl.h 	n->next = first;
next               84 include/linux/list_bl.h 		first->pprev = &n->next;
next               90 include/linux/list_bl.h 				       struct hlist_bl_node *next)
next               92 include/linux/list_bl.h 	struct hlist_bl_node **pprev = next->pprev;
next               95 include/linux/list_bl.h 	n->next = next;
next               96 include/linux/list_bl.h 	next->pprev = &n->next;
next              107 include/linux/list_bl.h 	n->next = prev->next;
next              108 include/linux/list_bl.h 	n->pprev = &prev->next;
next              109 include/linux/list_bl.h 	prev->next = n;
next              111 include/linux/list_bl.h 	if (n->next)
next              112 include/linux/list_bl.h 		n->next->pprev = &n->next;
next              117 include/linux/list_bl.h 	struct hlist_bl_node *next = n->next;
next              125 include/linux/list_bl.h 			((unsigned long)next |
next              127 include/linux/list_bl.h 	if (next)
next              128 include/linux/list_bl.h 		next->pprev = pprev;
next              134 include/linux/list_bl.h 	n->next = LIST_POISON1;
next              173 include/linux/list_bl.h 	     pos = pos->next)
next              185 include/linux/list_bl.h 	     pos && ({ n = pos->next; 1; }) && 				 \
next               26 include/linux/list_nulls.h 	struct hlist_nulls_node *next, **pprev;
next               74 include/linux/list_nulls.h 	n->next = first;
next               78 include/linux/list_nulls.h 		WRITE_ONCE(first->pprev, &n->next);
next               83 include/linux/list_nulls.h 	struct hlist_nulls_node *next = n->next;
next               86 include/linux/list_nulls.h 	WRITE_ONCE(*pprev, next);
next               87 include/linux/list_nulls.h 	if (!is_a_nulls(next))
next               88 include/linux/list_nulls.h 		WRITE_ONCE(next->pprev, pprev);
next              109 include/linux/list_nulls.h 	     pos = pos->next)
next              121 include/linux/list_nulls.h 	     pos = pos->next)
next               59 include/linux/llist.h 	struct llist_node *next;
next              115 include/linux/llist.h 	for ((pos) = (node); pos; (pos) = (pos)->next)
next              134 include/linux/llist.h 	for ((pos) = (node); (pos) && ((n) = (pos)->next, true); (pos) = (n))
next              154 include/linux/llist.h 	     (pos) = llist_entry((pos)->member.next, typeof(*(pos)), member))
next              176 include/linux/llist.h 	        (n = llist_entry(pos->member.next, typeof(*n), member), true); \
next              194 include/linux/llist.h 	return node->next;
next              727 include/linux/mlx5/device.h 	__be64		next;
next              240 include/linux/mlx5/driver.h 	struct mlx5_cmd_mailbox	       *next;
next              320 include/linux/mlx5/driver.h 	struct mlx5_cmd_mailbox *next;
next              107 include/linux/mm_types.h 					struct page *next;
next              360 include/linux/mm_types.h 	struct core_thread *next;
next               25 include/linux/mmc/sdio_func.h 	struct sdio_func_tuple *next;
next               13 include/linux/msg.h 	struct msg_msgseg *next;
next               92 include/linux/n_r3964.h 	struct r3964_client_info *next;
next              110 include/linux/n_r3964.h 	  struct r3964_message *next;
next              124 include/linux/n_r3964.h     struct r3964_block_header *next;
next             2580 include/linux/netdevice.h 	lh = dev->dev_list.next;
next             2597 include/linux/netdevice.h 		net_device_entry(net->dev_base_head.next);
next             4279 include/linux/netdevice.h 	for (iter = (dev)->adj_list.lower.next, \
next             4294 include/linux/netdevice.h 	for (iter = (dev)->adj_list.lower.next, \
next              314 include/linux/netfilter.h 	struct sk_buff *skb, *next;
next              318 include/linux/netfilter.h 	list_for_each_entry_safe(skb, next, head, list) {
next               34 include/linux/nls.h 	struct nls_table *next;
next               56 include/linux/notifier.h 	struct notifier_block __rcu *next;
next               35 include/linux/of.h 	struct property *next;
next              356 include/linux/of.h 	for (pp = dn->properties; pp != NULL; pp = pp->next)
next               10 include/linux/osq_lock.h 	struct optimistic_spin_node *next, *prev;
next               58 include/linux/page_ext.h 	void *next = curr;
next               59 include/linux/page_ext.h 	next += page_ext_size;
next               60 include/linux/page_ext.h 	return next;
next               30 include/linux/pagewalk.h 			 unsigned long next, struct mm_walk *walk);
next               32 include/linux/pagewalk.h 			 unsigned long next, struct mm_walk *walk);
next               34 include/linux/pagewalk.h 			 unsigned long next, struct mm_walk *walk);
next               35 include/linux/pagewalk.h 	int (*pte_hole)(unsigned long addr, unsigned long next,
next               38 include/linux/pagewalk.h 			     unsigned long addr, unsigned long next,
next               40 include/linux/pagewalk.h 	int (*test_walk)(unsigned long addr, unsigned long next,
next              147 include/linux/parport.h 	struct pardevice *next;
next              279 include/linux/pci.h 	struct hlist_node		next;
next               79 include/linux/perf_event.h 		struct perf_raw_frag	*next;
next              897 include/linux/perf_event.h 					struct task_struct *next);
next             1162 include/linux/perf_event.h 					     struct task_struct *next)
next             1167 include/linux/perf_event.h 		__perf_event_task_sched_out(prev, next);
next             1358 include/linux/perf_event.h 			  struct task_struct *next)			{ }
next              282 include/linux/plist.h 	return list_entry(head->node_list.next,
next              371 include/linux/pnp.h 	struct pnp_id *next;
next              296 include/linux/preempt.h 			  struct task_struct *next);
next               11 include/linux/raid/xor.h         struct xor_block_template *next;
next               33 include/linux/rculist.h 	WRITE_ONCE(list->next, list);
next               41 include/linux/rculist.h #define list_next_rcu(list)	(*((struct list_head __rcu **)(&(list)->next)))
next               68 include/linux/rculist.h 		struct list_head *prev, struct list_head *next)
next               70 include/linux/rculist.h 	if (!__list_add_valid(new, prev, next))
next               73 include/linux/rculist.h 	new->next = next;
next               76 include/linux/rculist.h 	next->prev = new;
next               97 include/linux/rculist.h 	__list_add_rcu(new, head, head->next);
next              191 include/linux/rculist.h 	new->next = old->next;
next              194 include/linux/rculist.h 	new->next->prev = new;
next              219 include/linux/rculist.h 					  struct list_head *next,
next              222 include/linux/rculist.h 	struct list_head *first = list->next;
next              250 include/linux/rculist.h 	last->next = next;
next              253 include/linux/rculist.h 	next->prev = last;
next              268 include/linux/rculist.h 		__list_splice_init_rcu(list, head, head->next, sync);
next              334 include/linux/rculist.h 	struct list_head *__next = READ_ONCE(__ptr->next); \
next              354 include/linux/rculist.h 	struct list_head *__next = READ_ONCE(__ptr->next); \
next              372 include/linux/rculist.h 	     pos = list_entry_rcu((head)->next, typeof(*pos), member);	\
next              374 include/linux/rculist.h 		pos = list_entry_rcu(pos->member.next, typeof(*pos), member))
next              406 include/linux/rculist.h 	for (pos = list_entry_lockless((head)->next, typeof(*pos), member); \
next              408 include/linux/rculist.h 	     pos = list_entry_lockless(pos->member.next, typeof(*pos), member))
next              429 include/linux/rculist.h 	for (pos = list_entry_rcu(pos->member.next, typeof(*pos), member); \
next              431 include/linux/rculist.h 	     pos = list_entry_rcu(pos->member.next, typeof(*pos), member))
next              452 include/linux/rculist.h 		pos = list_entry_rcu(pos->member.next, typeof(*(pos)), member))
next              489 include/linux/rculist.h 	struct hlist_node *next = old->next;
next              491 include/linux/rculist.h 	new->next = next;
next              494 include/linux/rculist.h 	if (next)
next              495 include/linux/rculist.h 		new->next->pprev = &new->next;
next              503 include/linux/rculist.h #define hlist_next_rcu(node)	(*((struct hlist_node __rcu **)(&(node)->next)))
next              530 include/linux/rculist.h 	n->next = first;
next              534 include/linux/rculist.h 		first->pprev = &n->next;
next              562 include/linux/rculist.h 	for (i = h->first; i; i = i->next)
next              566 include/linux/rculist.h 		n->next = last->next;
next              567 include/linux/rculist.h 		n->pprev = &last->next;
next              593 include/linux/rculist.h 					struct hlist_node *next)
next              595 include/linux/rculist.h 	n->pprev = next->pprev;
next              596 include/linux/rculist.h 	n->next = next;
next              598 include/linux/rculist.h 	next->pprev = &n->next;
next              622 include/linux/rculist.h 	n->next = prev->next;
next              623 include/linux/rculist.h 	n->pprev = &prev->next;
next              625 include/linux/rculist.h 	if (n->next)
next              626 include/linux/rculist.h 		n->next->pprev = &n->next;
next              107 include/linux/rculist_bl.h 	n->next = first;
next              109 include/linux/rculist_bl.h 		first->pprev = &n->next;
next              127 include/linux/rculist_bl.h 		pos = rcu_dereference_raw(pos->next))
next               45 include/linux/rculist_nulls.h 	(*((struct hlist_nulls_node __rcu __force **)&(node)->next))
next               96 include/linux/rculist_nulls.h 	n->next = first;
next              100 include/linux/rculist_nulls.h 		WRITE_ONCE(first->pprev, &n->next);
next              128 include/linux/rculist_nulls.h 	for (i = h->first; !is_a_nulls(i); i = i->next)
next              132 include/linux/rculist_nulls.h 		n->next = last->next;
next              133 include/linux/rculist_nulls.h 		n->pprev = &last->next;
next               18 include/linux/rhashtable-types.h 	struct rhash_head __rcu		*next;
next               23 include/linux/rhashtable-types.h 	struct rhlist_head __rcu	*next;
next              424 include/linux/rhashtable.h 	     pos = rht_dereference_bucket((pos)->next, tbl, hash))
next              448 include/linux/rhashtable.h 	     pos = rht_dereference_bucket((pos)->next, tbl, hash))
next              475 include/linux/rhashtable.h #define rht_for_each_entry_safe(tpos, pos, next, tbl, hash, member)	      \
next              477 include/linux/rhashtable.h 	     next = !rht_is_a_nulls(pos) ?				      \
next              478 include/linux/rhashtable.h 		       rht_dereference_bucket(pos->next, tbl, hash) : NULL;   \
next              480 include/linux/rhashtable.h 	     pos = next,						      \
next              481 include/linux/rhashtable.h 	     next = !rht_is_a_nulls(pos) ?				      \
next              482 include/linux/rhashtable.h 		       rht_dereference_bucket(pos->next, tbl, hash) : NULL)
next              499 include/linux/rhashtable.h 	     pos = rcu_dereference_raw(pos->next))
next              515 include/linux/rhashtable.h 	     pos = rcu_dereference_raw(pos->next))
next              534 include/linux/rhashtable.h 	     pos = rht_dereference_bucket_rcu(pos->next, tbl, hash))
next              562 include/linux/rhashtable.h 	for (pos = list; pos; pos = rcu_dereference_raw(pos->next))
next              576 include/linux/rhashtable.h 	     pos = rcu_dereference_raw(pos->next))
next              748 include/linux/rhashtable.h 			pprev = &head->next;
next              761 include/linux/rhashtable.h 		RCU_INIT_POINTER(list->next, plist);
next              762 include/linux/rhashtable.h 		head = rht_dereference_bucket(head->next, tbl, hash);
next              763 include/linux/rhashtable.h 		RCU_INIT_POINTER(list->rhead.next, head);
next              786 include/linux/rhashtable.h 	RCU_INIT_POINTER(obj->next, head);
next              791 include/linux/rhashtable.h 		RCU_INIT_POINTER(list->next, NULL);
next             1019 include/linux/rhashtable.h 			pprev = &he->next;
next             1025 include/linux/rhashtable.h 				lpprev = &list->next;
next             1026 include/linux/rhashtable.h 				list = rht_dereference_bucket(list->next,
next             1033 include/linux/rhashtable.h 			list = rht_dereference_bucket(list->next, tbl, hash);
next             1039 include/linux/rhashtable.h 		obj = rht_dereference_bucket(obj->next, tbl, hash);
next             1043 include/linux/rhashtable.h 			list = rht_dereference_bucket(list->next, tbl, hash);
next             1045 include/linux/rhashtable.h 				RCU_INIT_POINTER(list->rhead.next, obj);
next             1172 include/linux/rhashtable.h 			pprev = &he->next;
next             1176 include/linux/rhashtable.h 		rcu_assign_pointer(obj_new->next, obj_old->next);
next              157 include/linux/rmap.h 				  struct vm_area_struct *next)
next              159 include/linux/rmap.h 	VM_BUG_ON_VMA(vma->anon_vma != next->anon_vma, vma);
next              160 include/linux/rmap.h 	unlink_anon_vmas(next);
next              265 include/linux/sbitmap.h 			goto next;
next              282 include/linux/sbitmap.h next:
next              621 include/linux/sched.h 	struct wake_q_node *next;
next              380 include/linux/sched/mm.h 					     struct mm_struct *next,
next              563 include/linux/sched/signal.h 	list_entry_rcu((p)->tasks.next, struct task_struct, tasks)
next              653 include/linux/sched/signal.h 	return list_entry_rcu(p->thread_group.next,
next               35 include/linux/seq_file.h 	void * (*next) (struct seq_file *m, void *v, loff_t *pos);
next              291 include/linux/skbuff.h 	struct sk_buff	*next;
next              689 include/linux/skbuff.h 			struct sk_buff		*next;
next             1480 include/linux/skbuff.h 	skb->next = NULL;
next             1497 include/linux/skbuff.h 	return list->next == (const struct sk_buff *) list;
next             1509 include/linux/skbuff.h 	return READ_ONCE(list->next) == (const struct sk_buff *) list;
next             1523 include/linux/skbuff.h 	return skb->next == (const struct sk_buff *) list;
next             1554 include/linux/skbuff.h 	return skb->next;
next             1748 include/linux/skbuff.h 	struct sk_buff *skb = list_->next;
next             1763 include/linux/skbuff.h 	return list_->next;
next             1778 include/linux/skbuff.h 	struct sk_buff *next = skb->next;
next             1780 include/linux/skbuff.h 	if (next == (struct sk_buff *)list_)
next             1781 include/linux/skbuff.h 		next = NULL;
next             1782 include/linux/skbuff.h 	return next;
next             1831 include/linux/skbuff.h 	list->prev = list->next = (struct sk_buff *)list;
next             1863 include/linux/skbuff.h 				struct sk_buff *prev, struct sk_buff *next,
next             1869 include/linux/skbuff.h 	WRITE_ONCE(newsk->next, next);
next             1871 include/linux/skbuff.h 	WRITE_ONCE(next->prev, newsk);
next             1872 include/linux/skbuff.h 	WRITE_ONCE(prev->next, newsk);
next             1878 include/linux/skbuff.h 				      struct sk_buff *next)
next             1880 include/linux/skbuff.h 	struct sk_buff *first = list->next;
next             1884 include/linux/skbuff.h 	WRITE_ONCE(prev->next, first);
next             1886 include/linux/skbuff.h 	WRITE_ONCE(last->next, next);
next             1887 include/linux/skbuff.h 	WRITE_ONCE(next->prev, last);
next             1899 include/linux/skbuff.h 		__skb_queue_splice(list, (struct sk_buff *) head, head->next);
next             1915 include/linux/skbuff.h 		__skb_queue_splice(list, (struct sk_buff *) head, head->next);
next             1968 include/linux/skbuff.h 	__skb_insert(newsk, prev, prev->next, list);
next             1975 include/linux/skbuff.h 				      struct sk_buff *next,
next             1978 include/linux/skbuff.h 	__skb_insert(newsk, next->prev, next, list);
next             2022 include/linux/skbuff.h 	struct sk_buff *next, *prev;
next             2025 include/linux/skbuff.h 	next	   = skb->next;
next             2027 include/linux/skbuff.h 	skb->next  = skb->prev = NULL;
next             2028 include/linux/skbuff.h 	WRITE_ONCE(next->prev, prev);
next             2029 include/linux/skbuff.h 	WRITE_ONCE(prev->next, next);
next             3405 include/linux/skbuff.h 		for (skb = (queue)->next;					\
next             3407 include/linux/skbuff.h 		     skb = skb->next)
next             3410 include/linux/skbuff.h 		for (skb = (queue)->next, tmp = skb->next;			\
next             3412 include/linux/skbuff.h 		     skb = tmp, tmp = skb->next)
next             3416 include/linux/skbuff.h 		     skb = skb->next)
next             3431 include/linux/skbuff.h 		for (tmp = skb->next;						\
next             3433 include/linux/skbuff.h 		     skb = tmp, tmp = skb->next)
next             3461 include/linux/skbuff.h 	for (iter = skb_shinfo(skb)->frag_list; iter; iter = iter->next)
next               58 include/linux/slub_def.h 	slub_percpu_partial(c) = (p)->next;	\
next              358 include/linux/suspend.h 	struct pbe *next;
next              224 include/linux/swap.h 	unsigned int next; /* Likely next allocation offset */
next               76 include/linux/timer.h 		.entry = { .next = TIMER_ENTRY_STATIC },	\
next               20 include/linux/trace.h 	struct trace_export __rcu	*next;
next               60 include/linux/tty.h 		struct tty_buffer *next;
next              182 include/linux/types.h 	struct list_head *next, *prev;
next              190 include/linux/types.h 	struct hlist_node *next, **pprev;
next              220 include/linux/types.h 	struct callback_head *next;
next               45 include/linux/uprobes.h 	struct uprobe_consumer *next;
next               90 include/linux/uprobes.h 	struct return_instance	*next;		/* keep as stack */
next               20 include/linux/user-return-notifier.h 						struct task_struct *next)
next               24 include/linux/user-return-notifier.h 		set_tsk_thread_flag(next, TIF_USER_RETURN_NOTIFY);
next               40 include/linux/user-return-notifier.h 						struct task_struct *next)
next               42 include/linux/vmalloc.h 	struct vm_struct	*next;
next              528 include/linux/xarray.h 		struct xa_limit, u32 *next, gfp_t);
next              909 include/linux/xarray.h 		struct xa_limit limit, u32 *next, gfp_t gfp)
next              914 include/linux/xarray.h 	err = __xa_alloc_cyclic(xa, id, entry, limit, next, gfp);
next              942 include/linux/xarray.h 		struct xa_limit limit, u32 *next, gfp_t gfp)
next              947 include/linux/xarray.h 	err = __xa_alloc_cyclic(xa, id, entry, limit, next, gfp);
next              975 include/linux/xarray.h 		struct xa_limit limit, u32 *next, gfp_t gfp)
next              980 include/linux/xarray.h 	err = __xa_alloc_cyclic(xa, id, entry, limit, next, gfp);
next               87 include/media/dmxdev.h 	struct list_head next;
next               66 include/media/drv-intf/saa7146_vv.h 			struct saa7146_buf *next);
next              198 include/media/drv-intf/saa7146_vv.h void saa7146_set_capture(struct saa7146_dev *dev, struct saa7146_buf *buf, struct saa7146_buf *next);
next               84 include/media/dvb_demux.h 	struct dvb_demux_filter *next;
next              289 include/media/v4l2-ctrls.h 	struct v4l2_ctrl_ref *next;
next               34 include/net/atmclip.h 	struct clip_vcc	*next;		/* next VCC */
next               50 include/net/atmclip.h 	struct net_device *next;	/* next CLIP interface */
next              189 include/net/ax25.h 	struct ax25_route	*next;
next              231 include/net/ax25.h 	struct ax25_dev		*next;
next              372 include/net/ax25.h 	struct ax25_protocol *next;
next               70 include/net/bond_alb.h 	u32 next;		/* The next Hash table entry index, assigned
next               60 include/net/bonding.h 		netdev_adjacent_get_private(bond_slave_list(bond)->next) : \
next               93 include/net/if_inet6.h 	struct ipv6_mc_socklist __rcu *next;
next              117 include/net/if_inet6.h 	struct ifmcaddr6	*next;
next              122 include/net/ip.h 	struct ip_ra_chain __rcu *next;
next              181 include/net/ip.h 	iter->frag = skb->next;
next               47 include/net/ip6_tunnel.h 	struct ip6_tnl __rcu *next;	/* next tunnel in list */
next               96 include/net/ip_tunnels.h 	struct ip_tunnel_prl_entry __rcu *next;
next              105 include/net/ip_tunnels.h 	struct ip_tunnel __rcu	*next;
next              418 include/net/ip_vs.h 	struct ip_vs_protocol	*next;
next              478 include/net/ip_vs.h 	struct ip_vs_proto_data	*next;
next              172 include/net/ipv6.h 	iter->frag = skb->next;
next              273 include/net/ipv6.h 	struct ip6_ra_chain	*next;
next              313 include/net/ipv6.h 	struct ip6_flowlabel __rcu *next;
next              338 include/net/ipv6.h 	struct ipv6_fl_socklist	__rcu	*next;
next              135 include/net/neighbour.h 	struct neighbour __rcu	*next;
next              172 include/net/neighbour.h 	struct pneigh_entry	*next;
next              296 include/net/neighbour.h 	     n = rcu_dereference_bh(n->next)) {
next              157 include/net/netlabel.h 	struct netlbl_lsm_catmap *next;
next              334 include/net/netlabel.h 		catmap = catmap->next;
next               92 include/net/rose.h 	struct rose_neigh	*next;
next              109 include/net/rose.h 	struct rose_node	*next;
next              118 include/net/rose.h 	struct rose_route	*next;
next               32 include/net/sch_generic.h 	struct qdisc_rate_table *next;
next              236 include/net/sch_generic.h 	struct Qdisc_ops	*next;
next              351 include/net/sch_generic.h 	struct tcf_proto __rcu	*next;
next              959 include/net/sch_generic.h 		skb->next = NULL;
next              960 include/net/sch_generic.h 		last->next = skb;
next              979 include/net/sch_generic.h 	skb->next = qh->head;
next              992 include/net/sch_generic.h 		qh->head = skb->next;
next              996 include/net/sch_generic.h 		skb->next = NULL;
next             1019 include/net/sch_generic.h 	skb->next = *to_free;
next             1027 include/net/sch_generic.h 		skb->prev->next = *to_free;
next             1029 include/net/sch_generic.h 		skb->next = *to_free;
next              388 include/net/sctp/sctp.h 		result = list->next;
next              415 include/net/sctp/sctp.h 	return (head->next != head) && (head->next == head->prev);
next             1398 include/net/sctp/structs.h 	struct sctp_stream_out_ext *next;
next              131 include/net/slhc_vj.h 	struct cstate *next;	/* next in ring (xmit) */
next              594 include/net/sock.h 	return hlist_entry_safe(sk->sk_node.next, struct sock, sk_node);
next              599 include/net/sock.h 	return (!is_a_nulls(sk->sk_nulls_node.next)) ?
next              600 include/net/sock.h 		hlist_nulls_entry(sk->sk_nulls_node.next,
next              910 include/net/sock.h 		sk->sk_backlog.tail->next = skb;
next              913 include/net/sock.h 	skb->next = NULL;
next             1063 include/net/sock.h 	if (offsetof(struct sock, sk_node.next) != 0)
next             1064 include/net/sock.h 		memset(sk, 0, offsetof(struct sock, sk_node.next));
next             1796 include/net/tcp.h 	if (sk->sk_write_queue.next == skb)
next              142 include/net/udp.h 	for (skb = skb_shinfo(skb)->frag_list; skb; skb = skb->next) {
next              236 include/net/vxlan.h 	struct list_head  next;		/* vxlan's per namespace list */
next              991 include/net/xfrm.h 	struct xfrm_if __rcu *next;	/* next interface in list */
next             1402 include/net/xfrm.h 	struct xfrm4_protocol __rcu *next;
next             1412 include/net/xfrm.h 	struct xfrm6_protocol __rcu *next;
next             1421 include/net/xfrm.h 	struct xfrm_tunnel __rcu *next;
next             1429 include/net/xfrm.h 	struct xfrm6_tunnel __rcu *next;
next              490 include/rdma/ib_mad.h 	struct ib_mad_send_buf	*next;
next             1325 include/rdma/ib_verbs.h 	struct ib_send_wr      *next;
next             1396 include/rdma/ib_verbs.h 	struct ib_recv_wr      *next;
next             2096 include/rdma/ib_verbs.h 	struct ib_flow_spec_list	*next;
next              101 include/rdma/rdmavt_mr.h 	u32 next;               /* next unused index (speeds search) */
next              334 include/rdma/rdmavt_qp.h 	struct rvt_qp __rcu *next;           /* link list for QPN hash table */
next              726 include/rdma/rdmavt_qp.h 			qp = rcu_dereference(qp->next))
next             1575 include/sound/emu10k1.h 	struct snd_emu10k1_fx8010_irq *next;
next              183 include/sound/gus.h 	struct snd_gf1_mem_block *next;
next              203 include/sound/gus.h 	struct snd_gf1_dma_block *next;
next              258 include/sound/opl3.h 	struct fm_patch *next;
next              458 include/sound/pcm.h 	struct snd_pcm_substream *next;
next             1386 include/sound/pcm.h 	for (s = info->pcm->streams[info->stream].substream; s; s = s->next)
next               21 include/sound/pcm_oss.h 	struct snd_pcm_oss_setup *next;
next               20 include/sound/soundfont.h 	struct snd_sf_zone *next;	/* Link to next */
next               38 include/sound/soundfont.h 	struct snd_sf_sample *next;
next               45 include/sound/soundfont.h 	struct snd_soundfont *next;	/* Link to next */
next              142 include/trace/events/sched.h 		 struct task_struct *next),
next              144 include/trace/events/sched.h 	TP_ARGS(preempt, prev, next),
next              157 include/trace/events/sched.h 		memcpy(__entry->next_comm, next->comm, TASK_COMM_LEN);
next              162 include/trace/events/sched.h 		__entry->next_pid	= next->pid;
next              163 include/trace/events/sched.h 		__entry->next_prio	= next->prio;
next              116 include/uapi/drm/drm.h 	unsigned char next;
next              151 include/uapi/drm/i810_drm.h 	unsigned char next, prev;	/* indices to form a circular LRU  */
next              180 include/uapi/drm/via_drm.h 	unsigned char next, prev;	/* indices to form a circular LRU  */
next              168 include/uapi/linux/dm-ioctl.h 	__u32 next;
next              193 include/uapi/linux/dm-ioctl.h 	__u32 next;		/* offset to the next record from
next              202 include/uapi/linux/dm-ioctl.h         __u32 next;
next              353 include/uapi/linux/fd.h 	struct floppy_raw_cmd *next; /* used for chaining of raw cmd's 
next               59 include/uapi/linux/futex.h 	struct robust_list __user *next;
next               36 include/uapi/linux/romfs_fs.h 	__be32 next;		/* low 4 bits see ROMFH_ */
next               78 include/uapi/linux/vfio.h 	__u32	next;		/* Offset of next capability */
next             1130 include/uapi/linux/videodev2.h 	struct v4l2_clip	__user *next;
next               98 include/uapi/linux/virtio_ring.h 	__virtio16 next;
next               59 include/xen/grant_table.h 	struct gnttab_free_callback *next;
next              370 include/xen/interface/xen-mca.h 	unsigned next;
next              371 init/do_mounts.c 		char *p, *next;
next              374 init/do_mounts.c 		for (p = page-1; p; p = next) {
next              375 init/do_mounts.c 			next = strchr(++p, '\n');
next               51 init/initramfs.c 	struct hash *next;
next               66 init/initramfs.c 	for (p = head + hash(major, minor, ino); *p; p = &(*p)->next) {
next               85 init/initramfs.c 	q->next = NULL;
next               96 init/initramfs.c 			*p = q->next;
next              207 init/initramfs.c static void __init read_into(char *buf, unsigned size, enum state next)
next              212 init/initramfs.c 		state = next;
next              216 init/initramfs.c 		next_state = next;
next              802 init/main.c    	struct list_head next;
next              829 init/main.c    			list_add(&entry->next, &blacklisted_initcalls);
next              854 init/main.c    	list_for_each_entry(entry, &blacklisted_initcalls, next) {
next              193 ipc/msg.c      	if (mss->list.next)
next              206 ipc/msg.c      			mss->list.next = NULL;
next               38 ipc/msgutil.c  	struct msg_msgseg *next;
next               57 ipc/msgutil.c  	msg->next = NULL;
next               61 ipc/msgutil.c  	pseg = &msg->next;
next               72 ipc/msgutil.c  		seg->next = NULL;
next               73 ipc/msgutil.c  		pseg = &seg->next;
next               99 ipc/msgutil.c  	for (seg = msg->next; seg != NULL; seg = seg->next) {
next              130 ipc/msgutil.c  	for (dst_pseg = dst->next, src_pseg = src->next;
next              132 ipc/msgutil.c  	     dst_pseg = dst_pseg->next, src_pseg = src_pseg->next) {
next              159 ipc/msgutil.c  	for (seg = msg->next; seg != NULL; seg = seg->next) {
next              175 ipc/msgutil.c  	seg = msg->next;
next              178 ipc/msgutil.c  		struct msg_msgseg *tmp = seg->next;
next             2319 ipc/sem.c      		un = list_entry_rcu(ulp->list_proc.next,
next             1635 ipc/shm.c      	struct vm_area_struct *next;
next             1669 ipc/shm.c      		next = vma->vm_next;
next             1695 ipc/shm.c      			vma = next;
next             1698 ipc/shm.c      		vma = next;
next             1708 ipc/shm.c      		next = vma->vm_next;
next             1715 ipc/shm.c      		vma = next;
next              859 ipc/util.c     	.next  = sysvipc_proc_next,
next              546 kernel/audit_tree.c 	struct audit_krule *rule, *next;
next              549 kernel/audit_tree.c 	list_for_each_entry_safe(rule, next, &tree->rules, rlist) {
next              620 kernel/audit_tree.c 	for (p = tree->chunks.next; p != &tree->chunks; p = q) {
next              622 kernel/audit_tree.c 		q = p->next;
next              685 kernel/audit_tree.c 	while (cursor.next != &tree_list) {
next              692 kernel/audit_tree.c 		tree = container_of(cursor.next, struct audit_tree, list);
next              770 kernel/audit_tree.c 			victim = list_entry(prune_list.next,
next              898 kernel/audit_tree.c 	while (cursor.next != &tree_list) {
next              902 kernel/audit_tree.c 		tree = container_of(cursor.next, struct audit_tree, list);
next              987 kernel/audit_tree.c 		victim = list_entry(list->next, struct audit_tree, list);
next             1016 kernel/audit_tree.c 		owner = list_entry(chunk->trees.next,
next              100 kernel/auditsc.c 	struct audit_aux_data	*next;
next              129 kernel/auditsc.c 	struct audit_tree_refs *next;
next              222 kernel/auditsc.c 	p = p->next;
next              241 kernel/auditsc.c 		p->next = ctx->trees;
next              262 kernel/auditsc.c 	for (q = p; q != ctx->trees; q = q->next, n = 31) {
next              280 kernel/auditsc.c 		q = p->next;
next              292 kernel/auditsc.c 	for (p = ctx->first_trees; p != ctx->trees; p = p->next) {
next              863 kernel/auditsc.c 	struct audit_names *n, *next;
next              865 kernel/auditsc.c 	list_for_each_entry_safe(n, next, &context->names_list, list) {
next              883 kernel/auditsc.c 		context->aux = aux->next;
next              887 kernel/auditsc.c 		context->aux_pids = aux->next;
next             1478 kernel/auditsc.c 	for (aux = context->aux; aux; aux = aux->next) {
next             1531 kernel/auditsc.c 	for (aux = context->aux_pids; aux; aux = aux->next) {
next             2414 kernel/auditsc.c 		axp->d.next = ctx->aux_pids;
next             2453 kernel/auditsc.c 	ax->d.next = context->aux;
next              257 kernel/bpf/arraymap.c 	u32 *next = (u32 *)next_key;
next              260 kernel/bpf/arraymap.c 		*next = 0;
next              267 kernel/bpf/arraymap.c 	*next = index + 1;
next              170 kernel/bpf/bpf_lru_list.c 	struct list_head *cur, *last, *next = inactive;
next              177 kernel/bpf/bpf_lru_list.c 	last = l->next_inactive_rotation->next;
next              179 kernel/bpf/bpf_lru_list.c 		last = last->next;
next              189 kernel/bpf/bpf_lru_list.c 		next = cur->prev;
next              194 kernel/bpf/bpf_lru_list.c 		cur = next;
next              198 kernel/bpf/bpf_lru_list.c 	l->next_inactive_rotation = next;
next              579 kernel/bpf/cpumap.c 	u32 *next = next_key;
next              582 kernel/bpf/cpumap.c 		*next = 0;
next              588 kernel/bpf/cpumap.c 	*next = index + 1;
next              240 kernel/bpf/devmap.c 			struct hlist_node *next;
next              244 kernel/bpf/devmap.c 			hlist_for_each_entry_safe(dev, next, head, index_hlist) {
next              277 kernel/bpf/devmap.c 	u32 *next = next_key;
next              280 kernel/bpf/devmap.c 		*next = 0;
next              286 kernel/bpf/devmap.c 	*next = index + 1;
next              308 kernel/bpf/devmap.c 	u32 idx, *next = next_key;
next              326 kernel/bpf/devmap.c 		*next = next_dev->idx;
next              341 kernel/bpf/devmap.c 			*next = next_dev->idx;
next              751 kernel/bpf/devmap.c 		struct hlist_node *next;
next              755 kernel/bpf/devmap.c 		hlist_for_each_entry_safe(dev, next, head, index_hlist) {
next              244 kernel/bpf/hashtab.c 	BUILD_BUG_ON(offsetof(struct htab_elem, fnode.next) !=
next              244 kernel/bpf/inode.c 	.next	= map_seq_next,
next              240 kernel/bpf/local_storage.c 	struct bpf_cgroup_storage_key *next = _next_key;
next              262 kernel/bpf/local_storage.c 	next->attach_type = storage->key.attach_type;
next              263 kernel/bpf/local_storage.c 	next->cgroup_inode_id = storage->key.cgroup_inode_id;
next               32 kernel/bpf/percpu_freelist.c 	node->next = head->first;
next               96 kernel/bpf/percpu_freelist.c 			head->first = node->next;
next               19 kernel/bpf/percpu_freelist.h 	struct pcpu_freelist_node *next;
next              339 kernel/bpf/reuseport_array.c 	u32 *next = (u32 *)next_key;
next              342 kernel/bpf/reuseport_array.c 		*next = 0;
next              349 kernel/bpf/reuseport_array.c 	*next = index + 1;
next              168 kernel/bpf/verifier.c 	struct bpf_verifier_stack_elem *next;
next              794 kernel/bpf/verifier.c 	elem = head->next;
next              816 kernel/bpf/verifier.c 	elem->next = env->head;
next             1164 kernel/bpf/verifier.c 			goto next;
next             1166 kernel/bpf/verifier.c 			goto next;
next             1172 kernel/bpf/verifier.c next:
next             6881 kernel/bpf/verifier.c 			goto next;
next             6884 kernel/bpf/verifier.c 			goto next;
next             6887 kernel/bpf/verifier.c 				goto next;
next             6889 kernel/bpf/verifier.c next:
next             6890 kernel/bpf/verifier.c 		sl = sl->next;
next             7342 kernel/bpf/verifier.c 			goto next;
next             7410 kernel/bpf/verifier.c 			*pprev = sl->next;
next             7425 kernel/bpf/verifier.c 				sl->next = env->free_list;
next             7431 kernel/bpf/verifier.c next:
next             7432 kernel/bpf/verifier.c 		pprev = &sl->next;
next             7477 kernel/bpf/verifier.c 	new_sl->next = *explored_state(env, insn_idx);
next             9246 kernel/bpf/verifier.c 		sln = sl->next;
next             9259 kernel/bpf/verifier.c 			sln = sl->next;
next              150 kernel/bpf/xskmap.c 	u32 *next = next_key;
next              153 kernel/bpf/xskmap.c 		*next = 0;
next              159 kernel/bpf/xskmap.c 	*next = index + 1;
next             1019 kernel/cgroup/cgroup.c 		l1 = l1->next;
next             1020 kernel/cgroup/cgroup.c 		l2 = l2->next;
next             4151 kernel/cgroup/cgroup.c 		unsigned long next = last + CGROUP_FILE_NOTIFY_MIN_INTV;
next             4153 kernel/cgroup/cgroup.c 		if (time_in_range(jiffies, last, next)) {
next             4154 kernel/cgroup/cgroup.c 			timer_reduce(&cfile->notify_timer, next);
next             4183 kernel/cgroup/cgroup.c 	struct cgroup_subsys_state *next;
next             4208 kernel/cgroup/cgroup.c 		next = list_entry_rcu(parent->children.next, struct cgroup_subsys_state, sibling);
next             4210 kernel/cgroup/cgroup.c 		next = list_entry_rcu(pos->sibling.next, struct cgroup_subsys_state, sibling);
next             4212 kernel/cgroup/cgroup.c 		list_for_each_entry_rcu(next, &parent->children, sibling)
next             4213 kernel/cgroup/cgroup.c 			if (next->serial_nr > pos->serial_nr)
next             4221 kernel/cgroup/cgroup.c 	if (&next->sibling != &parent->children)
next             4222 kernel/cgroup/cgroup.c 		return next;
next             4251 kernel/cgroup/cgroup.c 	struct cgroup_subsys_state *next;
next             4260 kernel/cgroup/cgroup.c 	next = css_next_child(NULL, pos);
next             4261 kernel/cgroup/cgroup.c 	if (next)
next             4262 kernel/cgroup/cgroup.c 		return next;
next             4266 kernel/cgroup/cgroup.c 		next = css_next_child(pos, pos->parent);
next             4267 kernel/cgroup/cgroup.c 		if (next)
next             4268 kernel/cgroup/cgroup.c 			return next;
next             4346 kernel/cgroup/cgroup.c 	struct cgroup_subsys_state *next;
next             4359 kernel/cgroup/cgroup.c 	next = css_next_child(pos, pos->parent);
next             4360 kernel/cgroup/cgroup.c 	if (next)
next             4361 kernel/cgroup/cgroup.c 		return css_leftmost_descendant(next);
next             4401 kernel/cgroup/cgroup.c 		l = it->tcset_pos->next;
next             4414 kernel/cgroup/cgroup.c 	l = l->next;
next             4465 kernel/cgroup/cgroup.c 		it->task_pos = cset->tasks.next;
next             4468 kernel/cgroup/cgroup.c 		it->task_pos = cset->mg_tasks.next;
next             4471 kernel/cgroup/cgroup.c 		it->task_pos = cset->dying_tasks.next;
next             4509 kernel/cgroup/cgroup.c 		it->task_pos = it->task_pos->next;
next             4529 kernel/cgroup/cgroup.c 			it->task_pos = it->task_pos->next;
next             4532 kernel/cgroup/cgroup.c 			it->task_pos = it->mg_tasks_head->next;
next             4536 kernel/cgroup/cgroup.c 			it->task_pos = it->dying_tasks_head->next;
next              209 kernel/crash_core.c 					goto next;
next              217 kernel/crash_core.c next:
next              444 kernel/crash_core.c 	VMCOREINFO_OFFSET(list_head, next);
next              950 kernel/debug/debug_core.c 	.next			= NULL,
next              714 kernel/debug/kdb/kdb_io.c 			c = c->next;
next              778 kernel/debug/kdb/kdb_io.c 			c = c->next;
next              693 kernel/debug/kdb/kdb_support.c 	u32 next;	/* offset of next header from start of pool */
next              768 kernel/debug/kdb/kdb_support.c 		if (!h->next)
next              771 kernel/debug/kdb/kdb_support.c 		h = (struct debug_alloc_header *)(debug_alloc_pool + h->next);
next              777 kernel/debug/kdb/kdb_support.c 	if (best->next == 0 && bestprev == NULL && rem < dah_overhead)
next              785 kernel/debug/kdb/kdb_support.c 		h->next = best->next;
next              787 kernel/debug/kdb/kdb_support.c 		h_offset = best->next;
next              792 kernel/debug/kdb/kdb_support.c 		bestprev->next = h_offset;
next              825 kernel/debug/kdb/kdb_support.c 		h->next = dah_first;
next              833 kernel/debug/kdb/kdb_support.c 			if (!prev->next || prev->next > h_offset)
next              836 kernel/debug/kdb/kdb_support.c 				(debug_alloc_pool + prev->next);
next              846 kernel/debug/kdb/kdb_support.c 			h->next = prev->next;
next              847 kernel/debug/kdb/kdb_support.c 			prev->next = h_offset;
next              850 kernel/debug/kdb/kdb_support.c 	if (h_offset + dah_overhead + h->size == h->next) {
next              851 kernel/debug/kdb/kdb_support.c 		struct debug_alloc_header *next;
next              852 kernel/debug/kdb/kdb_support.c 		next = (struct debug_alloc_header *)
next              853 kernel/debug/kdb/kdb_support.c 			(debug_alloc_pool + h->next);
next              854 kernel/debug/kdb/kdb_support.c 		h->size += dah_overhead + next->size;
next              855 kernel/debug/kdb/kdb_support.c 		h->next = next->next;
next              856 kernel/debug/kdb/kdb_support.c 		memset(next, POISON_FREE, dah_overhead - 1);
next              857 kernel/debug/kdb/kdb_support.c 		*((char *)next + dah_overhead - 1) = POISON_END;
next              899 kernel/debug/kdb/kdb_support.c 			  (debug_alloc_pool + h_free->next);
next              900 kernel/debug/kdb/kdb_support.c 	} while (h_free->next);
next              650 kernel/dma/debug.c 	entry = list_entry(free_entries.next, struct dma_debug_entry, list);
next              848 kernel/events/core.c 					 struct task_struct *next)
next              860 kernel/events/core.c 	cgrp2 = perf_cgroup_from_task(next, NULL);
next             1017 kernel/events/core.c 					 struct task_struct *next)
next             1040 kernel/events/core.c perf_cgroup_switch(struct task_struct *task, struct task_struct *next)
next             1647 kernel/events/core.c 	struct perf_event *next;
next             1649 kernel/events/core.c 	next = rb_entry_safe(rb_next(&event->group_node), typeof(*event), group_node);
next             1650 kernel/events/core.c 	if (next && next->cpu == event->cpu)
next             1651 kernel/events/core.c 		return next;
next             3172 kernel/events/core.c 					 struct task_struct *next)
next             3188 kernel/events/core.c 	next_ctx = next->perf_event_ctxp[ctxn];
next             3212 kernel/events/core.c 			WRITE_ONCE(ctx->task, next);
next             3225 kernel/events/core.c 			RCU_INIT_POINTER(next->perf_event_ctxp[ctxn], ctx);
next             3276 kernel/events/core.c 				struct task_struct *next,
next             3282 kernel/events/core.c 	if (prev == next)
next             3319 kernel/events/core.c 				 struct task_struct *next)
next             3324 kernel/events/core.c 		perf_pmu_sched_task(task, next, false);
next             3327 kernel/events/core.c 		perf_event_switch(task, next, false);
next             3330 kernel/events/core.c 		perf_event_context_sched_out(task, ctxn, next);
next             3338 kernel/events/core.c 		perf_cgroup_sched_out(task, next);
next             3741 kernel/events/core.c 			goto next;
next             3763 kernel/events/core.c 	next:
next             6423 kernel/events/core.c 				frag = frag->next;
next             6620 kernel/events/core.c 				frag = frag->next;
next             11573 kernel/events/core.c 	struct perf_event *child_event, *next;
next             11623 kernel/events/core.c 	list_for_each_entry_safe(child_event, next, &child_ctx->event_list, event_entry)
next              769 kernel/events/uprobes.c 	uc->next = uprobe->consumers;
next              785 kernel/events/uprobes.c 	for (con = &uprobe->consumers; *con; con = &(*con)->next) {
next              787 kernel/events/uprobes.c 			*con = uc->next;
next              892 kernel/events/uprobes.c 	for (uc = uprobe->consumers; uc; uc = uc->next) {
next              959 kernel/events/uprobes.c 	struct map_info *next;
next              966 kernel/events/uprobes.c 	struct map_info *next = info->next;
next              968 kernel/events/uprobes.c 	return next;
next              995 kernel/events/uprobes.c 				prev->next = NULL;
next             1006 kernel/events/uprobes.c 		prev = prev->next;
next             1007 kernel/events/uprobes.c 		info->next = curr;
next             1021 kernel/events/uprobes.c 		curr = curr->next;
next             1030 kernel/events/uprobes.c 		info->next = prev;
next             1234 kernel/events/uprobes.c 	for (con = uprobe->consumers; con && con != uc ; con = con->next)
next             1713 kernel/events/uprobes.c 	struct return_instance *next = ri->next;
next             1716 kernel/events/uprobes.c 	return next;
next             1769 kernel/events/uprobes.c 	for (o = o_utask->return_instances; o; o = o->next) {
next             1776 kernel/events/uprobes.c 		n->next = NULL;
next             1779 kernel/events/uprobes.c 		p = &n->next;
next             1925 kernel/events/uprobes.c 	ri->next = utask->return_instances;
next             2086 kernel/events/uprobes.c 	for (uc = uprobe->consumers; uc; uc = uc->next) {
next             2118 kernel/events/uprobes.c 	for (uc = uprobe->consumers; uc; uc = uc->next) {
next             2131 kernel/events/uprobes.c 		ri = ri->next;	/* can't be NULL if chained */
next             2140 kernel/events/uprobes.c 	struct return_instance *ri, *next;
next             2158 kernel/events/uprobes.c 		next = find_next_ret_chain(ri);
next             2159 kernel/events/uprobes.c 		valid = !next || arch_uretprobe_is_alive(next, RP_CHECK_RET, regs);
next             2167 kernel/events/uprobes.c 		} while (ri != next);
next              459 kernel/exit.c  		self.next = xchg(&core_state->dumper.next, &self);
next              212 kernel/fail_function.c 	.next	= fei_seq_next,
next              929 kernel/fork.c  	tsk->wake_q.next = NULL;
next              918 kernel/futex.c 	struct list_head *next, *head = &curr->pi_state_list;
next              932 kernel/futex.c 		next = head->next;
next              933 kernel/futex.c 		pi_state = list_entry(next, struct futex_pi_state, list);
next              962 kernel/futex.c 		if (head->next != next) {
next             1675 kernel/futex.c 	struct futex_q *this, *next;
next             1695 kernel/futex.c 	plist_for_each_entry_safe(this, next, &hb->chain, list) {
next             1777 kernel/futex.c 	struct futex_q *this, *next;
next             1825 kernel/futex.c 	plist_for_each_entry_safe(this, next, &hb1->chain, list) {
next             1839 kernel/futex.c 		plist_for_each_entry_safe(this, next, &hb2->chain, list) {
next             2023 kernel/futex.c 	struct futex_q *this, *next;
next             2199 kernel/futex.c 	plist_for_each_entry_safe(this, next, &hb1->chain, list) {
next             3710 kernel/futex.c 	if (fetch_robust_entry(&entry, &head->list.next, &pi))
next             3730 kernel/futex.c 		rc = fetch_robust_entry(&next_entry, &entry->next, &next_pi);
next             4011 kernel/futex.c 	if (compat_fetch_robust_entry(&uentry, &entry, &head->list.next, &pi))
next             4033 kernel/futex.c 			(compat_uptr_t __user *)&entry->next, &next_pi);
next              135 kernel/gcov/fs.c 	.next	= gcov_seq_next,
next              568 kernel/gcov/fs.c 	char *next;
next              577 kernel/gcov/fs.c 	for (curr = filename; (next = strchr(curr, '/')); curr = next + 1) {
next              578 kernel/gcov/fs.c 		if (curr == next)
next              580 kernel/gcov/fs.c 		*next = 0;
next               75 kernel/gcov/gcc_3_4.c 	struct gcov_info		*next;
next              114 kernel/gcov/gcc_3_4.c 	return info->next;
next              123 kernel/gcov/gcc_3_4.c 	info->next = gcov_info_head;
next              135 kernel/gcov/gcc_3_4.c 		prev->next = info->next;
next              137 kernel/gcov/gcc_3_4.c 		gcov_info_head = info->next;
next               89 kernel/gcov/gcc_4_7.c 	struct gcov_info *next;
next              127 kernel/gcov/gcc_4_7.c 	return info->next;
next              136 kernel/gcov/gcc_4_7.c 	info->next = gcov_info_head;
next              148 kernel/gcov/gcc_4_7.c 		prev->next = info->next;
next              150 kernel/gcov/gcc_4_7.c 		gcov_info_head = info->next;
next              282 kernel/gcov/gcc_4_7.c 	dup->next = NULL;
next              162 kernel/irq/internals.h 	for (act = desc->action; act; act = act->next)
next               56 kernel/irq/ipi.c 		unsigned int next;
next               69 kernel/irq/ipi.c 		next = cpumask_next_zero(offset, dest);
next               70 kernel/irq/ipi.c 		if (next < nr_cpu_ids)
next               71 kernel/irq/ipi.c 			next = cpumask_next(next, dest);
next               72 kernel/irq/ipi.c 		if (next < nr_cpu_ids) {
next              254 kernel/irq/irqdesc.c 	for (action = desc->action; action != NULL; action = action->next) {
next             1437 kernel/irq/manage.c 			old_ptr = &old->next;
next             1709 kernel/irq/manage.c 		action_ptr = &action->next;
next             1713 kernel/irq/manage.c 	*action_ptr = action->next;
next              527 kernel/irq/proc.c 		while ((action = action->next) != NULL)
next              511 kernel/jump_label.c 	struct static_key_mod *next;
next              535 kernel/jump_label.c 	key->next = mod;
next              561 kernel/jump_label.c 	for (mod = static_key_mod(key); mod; mod = mod->next) {
next              650 kernel/jump_label.c 			jlm2->next = NULL;
next              656 kernel/jump_label.c 		jlm->next = static_key_mod(key);
next              689 kernel/jump_label.c 		prev = &key->next;
next              693 kernel/jump_label.c 			prev = &jlm->next;
next              694 kernel/jump_label.c 			jlm = jlm->next;
next              701 kernel/jump_label.c 		if (prev == &key->next)
next              702 kernel/jump_label.c 			static_key_set_mod(key, jlm->next);
next              704 kernel/jump_label.c 			*prev = jlm->next;
next              710 kernel/jump_label.c 		if (jlm->next == NULL) {
next              625 kernel/kallsyms.c 	.next = s_next,
next              342 kernel/kexec_core.c 	struct page *page, *next;
next              344 kernel/kexec_core.c 	list_for_each_entry_safe(page, next, list, lru) {
next              216 kernel/kprobes.c 	struct kprobe_insn_page *kip, *next;
next              221 kernel/kprobes.c 	list_for_each_entry_safe(kip, next, &c->pages, list) {
next             1275 kernel/kprobes.c 	struct hlist_node *next;
next             1277 kernel/kprobes.c 	hlist_for_each_entry_safe(ri, next, &rp->free_instances, hlist) {
next             1287 kernel/kprobes.c 	struct hlist_node *next;
next             1294 kernel/kprobes.c 		hlist_for_each_entry_safe(ri, next, head, hlist) {
next             1786 kernel/kprobes.c 		ap = list_entry(p->list.next, struct kprobe, list);
next             2403 kernel/kprobes.c 	.next  = kprobe_seq_next,
next             2451 kernel/kprobes.c 	.next  = kprobe_blacklist_seq_next,
next              591 kernel/kthread.c 			create = list_entry(kthread_create_list.next,
next              969 kernel/kthread.c 		kthread_insert_work(worker, &fwork.work, work->node.next);
next              972 kernel/kthread.c 				    worker->work_list.next);
next              101 kernel/livepatch/patch.c 			func = list_entry_rcu(func->stack_node.next,
next             2034 kernel/locking/lockdep.c 			 struct held_lock *next,
next             2054 kernel/locking/lockdep.c 	print_lock(next);
next             2061 kernel/locking/lockdep.c 	print_lock_name(hlock_class(next));
next             2080 kernel/locking/lockdep.c 				hlock_class(prev), hlock_class(next));
next             2228 kernel/locking/lockdep.c 			   struct held_lock *next)
next             2261 kernel/locking/lockdep.c 	that.class = hlock_class(next);
next             2298 kernel/locking/lockdep.c 				 prev, next,
next             2320 kernel/locking/lockdep.c 				  struct held_lock *prev, struct held_lock *next)
next             2335 kernel/locking/lockdep.c 	struct lock_class *next = hlock_class(nxt);
next             2345 kernel/locking/lockdep.c 	__print_lock_name(next);
next             2353 kernel/locking/lockdep.c 		   struct held_lock *next)
next             2365 kernel/locking/lockdep.c 	print_lock(next);
next             2370 kernel/locking/lockdep.c 	print_deadlock_scenario(next, prev);
next             2386 kernel/locking/lockdep.c check_deadlock(struct task_struct *curr, struct held_lock *next)
next             2395 kernel/locking/lockdep.c 		if (prev->instance == next->nest_lock)
next             2398 kernel/locking/lockdep.c 		if (hlock_class(prev) != hlock_class(next))
next             2405 kernel/locking/lockdep.c 		if ((next->read == 2) && prev->read)
next             2415 kernel/locking/lockdep.c 		print_deadlock_bug(curr, prev, next);
next             2445 kernel/locking/lockdep.c 	       struct held_lock *next, int distance,
next             2451 kernel/locking/lockdep.c 	if (!hlock_class(prev)->key || !hlock_class(next)->key) {
next             2462 kernel/locking/lockdep.c 		WARN_ONCE(!debug_locks_silent && !hlock_class(next)->key,
next             2464 kernel/locking/lockdep.c 			  hlock_class(next),
next             2465 kernel/locking/lockdep.c 			  hlock_class(next)->name);
next             2479 kernel/locking/lockdep.c 	ret = check_noncircular(next, prev, trace);
next             2483 kernel/locking/lockdep.c 	if (!check_irq_usage(curr, prev, next))
next             2494 kernel/locking/lockdep.c 	if (next->read == 2 || prev->read == 2)
next             2505 kernel/locking/lockdep.c 		if (entry->class == hlock_class(next)) {
next             2516 kernel/locking/lockdep.c 	ret = check_redundant(prev, next);
next             2531 kernel/locking/lockdep.c 	ret = add_lock_to_list(hlock_class(next), hlock_class(prev),
next             2533 kernel/locking/lockdep.c 			       next->acquire_ip, distance, *trace);
next             2538 kernel/locking/lockdep.c 	ret = add_lock_to_list(hlock_class(prev), hlock_class(next),
next             2539 kernel/locking/lockdep.c 			       &hlock_class(next)->locks_before,
next             2540 kernel/locking/lockdep.c 			       next->acquire_ip, distance, *trace);
next             2554 kernel/locking/lockdep.c check_prevs_add(struct task_struct *curr, struct held_lock *next)
next             2584 kernel/locking/lockdep.c 			int ret = check_prev_add(curr, hlock, next, distance,
next             4891 kernel/locking/lockdep.c 	WARN_ON_ONCE(!class->lock_entry.next);
next             4895 kernel/locking/lockdep.c 	WARN_ON_ONCE(!class->lock_entry.next);
next               99 kernel/locking/lockdep_proc.c 	.next	= l_next,
next              157 kernel/locking/lockdep_proc.c 	.next	= lc_next,
next              582 kernel/locking/lockdep_proc.c 	.next	= ls_next,
next               19 kernel/locking/mcs_spinlock.h 	struct mcs_spinlock *next;
next               71 kernel/locking/mcs_spinlock.h 	node->next   = NULL;
next               91 kernel/locking/mcs_spinlock.h 	WRITE_ONCE(prev->next, node);
next              104 kernel/locking/mcs_spinlock.h 	struct mcs_spinlock *next = READ_ONCE(node->next);
next              106 kernel/locking/mcs_spinlock.h 	if (likely(!next)) {
next              113 kernel/locking/mcs_spinlock.h 		while (!(next = READ_ONCE(node->next)))
next              118 kernel/locking/mcs_spinlock.h 	arch_mcs_spin_unlock_contended(&next->locked);
next               76 kernel/locking/mutex-debug.c 		DEBUG_LOCKS_WARN_ON(!lock->wait_list.prev && !lock->wait_list.next);
next             1224 kernel/locking/mutex.c 	struct task_struct *next = NULL;
next             1269 kernel/locking/mutex.c 		next = waiter->task;
next             1272 kernel/locking/mutex.c 		wake_q_add(&wake_q, next);
next             1276 kernel/locking/mutex.c 		__mutex_handoff(lock, next);
next               14 kernel/locking/mutex.h 		__list_del((waiter)->list.prev, (waiter)->list.next)
next               46 kernel/locking/osq_lock.c 	struct optimistic_spin_node *next = NULL;
next               78 kernel/locking/osq_lock.c 		if (node->next) {
next               79 kernel/locking/osq_lock.c 			next = xchg(&node->next, NULL);
next               80 kernel/locking/osq_lock.c 			if (next)
next               87 kernel/locking/osq_lock.c 	return next;
next               93 kernel/locking/osq_lock.c 	struct optimistic_spin_node *prev, *next;
next               98 kernel/locking/osq_lock.c 	node->next = NULL;
next              126 kernel/locking/osq_lock.c 	WRITE_ONCE(prev->next, node);
next              160 kernel/locking/osq_lock.c 		if (prev->next == node &&
next              161 kernel/locking/osq_lock.c 		    cmpxchg(&prev->next, node, NULL) == node)
next              188 kernel/locking/osq_lock.c 	next = osq_wait_next(lock, node, prev);
next              189 kernel/locking/osq_lock.c 	if (!next)
next              200 kernel/locking/osq_lock.c 	WRITE_ONCE(next->prev, prev);
next              201 kernel/locking/osq_lock.c 	WRITE_ONCE(prev->next, next);
next              208 kernel/locking/osq_lock.c 	struct optimistic_spin_node *node, *next;
next              222 kernel/locking/osq_lock.c 	next = xchg(&node->next, NULL);
next              223 kernel/locking/osq_lock.c 	if (next) {
next              224 kernel/locking/osq_lock.c 		WRITE_ONCE(next->locked, 1);
next              228 kernel/locking/osq_lock.c 	next = osq_wait_next(lock, node, NULL);
next              229 kernel/locking/osq_lock.c 	if (next)
next              230 kernel/locking/osq_lock.c 		WRITE_ONCE(next->locked, 1);
next              316 kernel/locking/qspinlock.c 	struct mcs_spinlock *prev, *next, *node;
next              434 kernel/locking/qspinlock.c 	node->next = NULL;
next              460 kernel/locking/qspinlock.c 	next = NULL;
next              470 kernel/locking/qspinlock.c 		WRITE_ONCE(prev->next, node);
next              481 kernel/locking/qspinlock.c 		next = READ_ONCE(node->next);
next              482 kernel/locking/qspinlock.c 		if (next)
next              483 kernel/locking/qspinlock.c 			prefetchw(next);
next              549 kernel/locking/qspinlock.c 	if (!next)
next              550 kernel/locking/qspinlock.c 		next = smp_cond_load_relaxed(&node->next, (VAL));
next              552 kernel/locking/qspinlock.c 	arch_mcs_spin_unlock_contended(&next->locked);
next              553 kernel/locking/qspinlock.c 	pv_kick_node(lock, next);
next             4376 kernel/module.c 	.next	= m_next,
next               29 kernel/notifier.c 		nl = &((*nl)->next);
next               31 kernel/notifier.c 	n->next = *nl;
next               44 kernel/notifier.c 		nl = &((*nl)->next);
next               46 kernel/notifier.c 	n->next = *nl;
next               56 kernel/notifier.c 			rcu_assign_pointer(*nl, n->next);
next               59 kernel/notifier.c 		nl = &((*nl)->next);
next               86 kernel/notifier.c 		next_nb = rcu_dereference_raw(nb->next);
next               78 kernel/padata.c 		padata = list_entry(local_list.next,
next              192 kernel/padata.c 	padata = list_entry(reorder->list.next, struct padata_priv, list);
next              304 kernel/padata.c 		padata = list_entry(local_list.next,
next              272 kernel/pid_namespace.c 	int ret, next;
next              283 kernel/pid_namespace.c 	next = idr_get_cursor(&pid_ns->idr) - 1;
next              285 kernel/pid_namespace.c 	tmp.data = &next;
next              288 kernel/pid_namespace.c 		idr_set_cursor(&pid_ns->idr, next + 1);
next              122 kernel/power/snapshot.c 	struct linked_page *next;
next              180 kernel/power/snapshot.c 		safe_pages_list = safe_pages_list->next;
next              208 kernel/power/snapshot.c 	lp->next = safe_pages_list;
next              239 kernel/power/snapshot.c 		struct linked_page *lp = list->next;
next              287 kernel/power/snapshot.c 		lp->next = ca->chain;
next              550 kernel/power/snapshot.c 	bm->cur.zone = list_entry(bm->zones.next, struct mem_zone_bm_rtree,
next              552 kernel/power/snapshot.c 	bm->cur.node = list_entry(bm->cur.zone->leaves.next,
next              848 kernel/power/snapshot.c 		bm->cur.node = list_entry(bm->cur.node->list.next,
next              858 kernel/power/snapshot.c 		bm->cur.zone = list_entry(bm->cur.zone->list.next,
next              860 kernel/power/snapshot.c 		bm->cur.node = list_entry(bm->cur.zone->leaves.next,
next              937 kernel/power/snapshot.c 		p_list = lp->next;
next             2246 kernel/power/snapshot.c 	struct highmem_pbe *next;
next             2388 kernel/power/snapshot.c 		safe_pages_list = safe_pages_list->next;
next             2391 kernel/power/snapshot.c 	pbe->next = highmem_pblist;
next             2503 kernel/power/snapshot.c 		lp->next = safe_pages_list;
next             2517 kernel/power/snapshot.c 			lp->next = safe_pages_list;
next             2569 kernel/power/snapshot.c 	safe_pages_list = safe_pages_list->next;
next             2570 kernel/power/snapshot.c 	pbe->next = restore_pblist;
next             2729 kernel/power/snapshot.c 		pbe = pbe->next;
next               87 kernel/power/swap.c 	struct swap_map_page_list *next;
next              955 kernel/power/swap.c 		handle->maps = handle->maps->next;
next              985 kernel/power/swap.c 			last->next = tmp;
next             1026 kernel/power/swap.c 		handle->maps = handle->maps->next;
next             1347 kernel/printk/printk.c 		const char *next = memchr(text, '\n', text_size);
next             1350 kernel/printk/printk.c 		if (next) {
next             1351 kernel/printk/printk.c 			text_len = next - text;
next             1352 kernel/printk/printk.c 			next++;
next             1353 kernel/printk/printk.c 			text_size -= next - text;
next             1372 kernel/printk/printk.c 		text = next;
next             2783 kernel/printk/printk.c 		newcon->next = console_drivers;
next             2785 kernel/printk/printk.c 		if (newcon->next)
next             2786 kernel/printk/printk.c 			newcon->next->flags &= ~CON_CONSDEV;
next             2788 kernel/printk/printk.c 		newcon->next = console_drivers->next;
next             2789 kernel/printk/printk.c 		console_drivers->next = newcon;
next             2858 kernel/printk/printk.c 		console_drivers=console->next;
next             2861 kernel/printk/printk.c 		for (a=console_drivers->next, b=console_drivers ;
next             2862 kernel/printk/printk.c 		     a; b=a, a=b->next) {
next             2864 kernel/printk/printk.c 				b->next = a->next;
next               35 kernel/rcu/rcu_segcblist.c 	rclp->tail = &rhp->next;
next               61 kernel/rcu/rcu_segcblist.c 		rhp->next = NULL;
next               63 kernel/rcu/rcu_segcblist.c 		srclp->tail = &rhp->next;
next               84 kernel/rcu/rcu_segcblist.c 	rclp->head = rhp->next;
next              262 kernel/rcu/rcu_segcblist.c 	rhp->next = NULL;
next              264 kernel/rcu/rcu_segcblist.c 	WRITE_ONCE(rsclp->tails[RCU_NEXT_TAIL], &rhp->next);
next              288 kernel/rcu/rcu_segcblist.c 	rhp->next = NULL;
next              294 kernel/rcu/rcu_segcblist.c 		WRITE_ONCE(rsclp->tails[i], &rhp->next);
next              267 kernel/rcu/rcutorture.c 	p = rcu_torture_freelist.next;
next              136 kernel/rcu/srcutiny.c 		lh = lh->next;
next              164 kernel/rcu/srcutiny.c 	rhp->next = NULL;
next              167 kernel/rcu/srcutiny.c 	ssp->srcu_cb_tail = &rhp->next;
next               79 kernel/rcu/tiny.c 	struct rcu_head *next, *list;
next               99 kernel/rcu/tiny.c 		next = list->next;
next              100 kernel/rcu/tiny.c 		prefetch(next);
next              105 kernel/rcu/tiny.c 		list = next;
next              139 kernel/rcu/tiny.c 	head->next = NULL;
next              143 kernel/rcu/tiny.c 	rcu_ctrlblk.curtail = &head->next;
next             2577 kernel/rcu/tree.c 	head->next = NULL;
next              371 kernel/rcu/tree_exp.h 		rnp->exp_tasks = rnp->blkd_tasks.next;
next              404 kernel/rcu/tree_plugin.h 	np = t->rcu_node_entry.next;
next              666 kernel/rcu/tree_plugin.h 		WRITE_ONCE(rnp->gp_tasks, rnp->blkd_tasks.next);
next              527 kernel/rcu/update.c 	rhp->next = NULL;
next              532 kernel/rcu/update.c 	rcu_tasks_cbs_tail = &rhp->next;
next              636 kernel/rcu/update.c 	struct rcu_head *next;
next              777 kernel/rcu/update.c 			next = list->next;
next              781 kernel/rcu/update.c 			list = next;
next              134 kernel/resource.c 	.next	= r_next,
next              616 kernel/resource.c 			goto next;
next              638 kernel/resource.c next:		if (!this || this->end == root->end)
next              783 kernel/resource.c 	struct resource *first, *next;
next              801 kernel/resource.c 	for (next = first; ; next = next->sibling) {
next              803 kernel/resource.c 		if (next->start < new->start || next->end > new->end)
next              804 kernel/resource.c 			return next;
next              805 kernel/resource.c 		if (!next->sibling)
next              807 kernel/resource.c 		if (next->sibling->start > new->end)
next              812 kernel/resource.c 	new->sibling = next->sibling;
next              815 kernel/resource.c 	next->sibling = NULL;
next              816 kernel/resource.c 	for (next = first; next; next = next->sibling)
next              817 kernel/resource.c 		next->parent = new;
next              822 kernel/resource.c 		next = parent->child;
next              823 kernel/resource.c 		while (next->sibling != first)
next              824 kernel/resource.c 			next = next->sibling;
next              825 kernel/resource.c 		next->sibling = new;
next              210 kernel/sched/autogroup.c 	static unsigned long next = INITIAL_JIFFIES;
next              226 kernel/sched/autogroup.c 	if (!capable(CAP_SYS_ADMIN) && time_before(jiffies, next))
next              229 kernel/sched/autogroup.c 	next = HZ / 10 + jiffies;
next              426 kernel/sched/core.c 	if (unlikely(cmpxchg_relaxed(&node->next, NULL, WAKE_Q_TAIL)))
next              433 kernel/sched/core.c 	head->lastp = &node->next;
next              488 kernel/sched/core.c 		node = node->next;
next              489 kernel/sched/core.c 		task->wake_q.next = NULL;
next             3036 kernel/sched/core.c 				   struct task_struct *next)
next             3041 kernel/sched/core.c 		notifier->ops->sched_out(notifier, next);
next             3046 kernel/sched/core.c 				 struct task_struct *next)
next             3049 kernel/sched/core.c 		__fire_sched_out_preempt_notifiers(curr, next);
next             3060 kernel/sched/core.c 				 struct task_struct *next)
next             3066 kernel/sched/core.c static inline void prepare_task(struct task_struct *next)
next             3073 kernel/sched/core.c 	next->on_cpu = 1;
next             3095 kernel/sched/core.c prepare_lock_switch(struct rq *rq, struct task_struct *next, struct rq_flags *rf)
next             3107 kernel/sched/core.c 	rq->lock.owner = next;
next             3127 kernel/sched/core.c # define prepare_arch_switch(next)	do { } while (0)
next             3149 kernel/sched/core.c 		    struct task_struct *next)
next             3152 kernel/sched/core.c 	sched_info_switch(rq, prev, next);
next             3153 kernel/sched/core.c 	perf_event_task_sched_out(prev, next);
next             3155 kernel/sched/core.c 	fire_sched_out_preempt_notifiers(prev, next);
next             3156 kernel/sched/core.c 	prepare_task(next);
next             3157 kernel/sched/core.c 	prepare_arch_switch(next);
next             3265 kernel/sched/core.c 	struct callback_head *head, *next;
next             3274 kernel/sched/core.c 		next = head->next;
next             3275 kernel/sched/core.c 		head->next = NULL;
next             3276 kernel/sched/core.c 		head = next;
next             3330 kernel/sched/core.c 	       struct task_struct *next, struct rq_flags *rf)
next             3332 kernel/sched/core.c 	prepare_task_switch(rq, prev, next);
next             3348 kernel/sched/core.c 	if (!next->mm) {                                // to kernel
next             3349 kernel/sched/core.c 		enter_lazy_tlb(prev->active_mm, next);
next             3351 kernel/sched/core.c 		next->active_mm = prev->active_mm;
next             3357 kernel/sched/core.c 		membarrier_switch_mm(rq, prev->active_mm, next->mm);
next             3366 kernel/sched/core.c 		switch_mm_irqs_off(prev->active_mm, next->mm, next);
next             3377 kernel/sched/core.c 	prepare_lock_switch(rq, next, rf);
next             3380 kernel/sched/core.c 	switch_to(prev, next, prev);
next             3999 kernel/sched/core.c 	struct task_struct *prev, *next;
next             4047 kernel/sched/core.c 	next = pick_next_task(rq, prev, &rf);
next             4051 kernel/sched/core.c 	if (likely(prev != next)) {
next             4057 kernel/sched/core.c 		RCU_INIT_POINTER(rq->curr, next);
next             4074 kernel/sched/core.c 		trace_sched_switch(preempt, prev, next);
next             4077 kernel/sched/core.c 		rq = context_switch(rq, prev, next, &rf);
next             6208 kernel/sched/core.c 	struct task_struct *next;
next             6211 kernel/sched/core.c 		next = class->pick_next_task(rq, NULL, NULL);
next             6212 kernel/sched/core.c 		if (next) {
next             6213 kernel/sched/core.c 			next->sched_class->put_prev_task(rq, next);
next             6214 kernel/sched/core.c 			return next;
next             6233 kernel/sched/core.c 	struct task_struct *next, *stop = rq->stop;
next             6263 kernel/sched/core.c 		next = __pick_migrate_task(rq);
next             6275 kernel/sched/core.c 		raw_spin_lock(&next->pi_lock);
next             6283 kernel/sched/core.c 		if (WARN_ON(task_rq(next) != rq || !task_on_rq_queued(next))) {
next             6284 kernel/sched/core.c 			raw_spin_unlock(&next->pi_lock);
next             6289 kernel/sched/core.c 		dest_cpu = select_fallback_rq(dead_rq->cpu, next);
next             6290 kernel/sched/core.c 		rq = __migrate_task(rq, rf, next, dest_cpu);
next             6297 kernel/sched/core.c 		raw_spin_unlock(&next->pi_lock);
next              362 kernel/sched/deadline.c 	dl_rq->earliest_dl.curr = dl_rq->earliest_dl.next = 0;
next              468 kernel/sched/deadline.c 		dl_rq->earliest_dl.next = p->dl.deadline;
next              487 kernel/sched/deadline.c 			dl_rq->earliest_dl.next = rb_entry(next_node,
next             1355 kernel/sched/deadline.c 		dl_rq->earliest_dl.next = 0;
next             2163 kernel/sched/deadline.c 				   src_rq->dl.earliest_dl.next))
next             2435 kernel/sched/deadline.c 	.next			= &rt_sched_class,
next              800 kernel/sched/debug.c 	.next		= sched_debug_next,
next              612 kernel/sched/fair.c 	struct rb_node *next = rb_next(&se->run_node);
next              614 kernel/sched/fair.c 	if (!next)
next              617 kernel/sched/fair.c 	return rb_entry(next, struct sched_entity, run_node);
next             2497 kernel/sched/fair.c 	work->next = work;
next             2642 kernel/sched/fair.c 	p->numa_work.next		= &p->numa_work;
next             2681 kernel/sched/fair.c 	if ((curr->flags & (PF_EXITING | PF_KTHREAD)) || work->next != work)
next             3171 kernel/sched/fair.c 		      struct cfs_rq *prev, struct cfs_rq *next)
next             3196 kernel/sched/fair.c 			n_last_update_time_copy = next->load_last_update_time_copy;
next             3201 kernel/sched/fair.c 			n_last_update_time = next->avg.last_update_time;
next             3208 kernel/sched/fair.c 	n_last_update_time = next->avg.last_update_time;
next             4036 kernel/sched/fair.c 		if (cfs_rq->next != se)
next             4039 kernel/sched/fair.c 		cfs_rq->next = NULL;
next             4059 kernel/sched/fair.c 	if (cfs_rq->next == se)
next             4246 kernel/sched/fair.c 	if (cfs_rq->next && wakeup_preempt_entity(cfs_rq->next, left) < 1)
next             4247 kernel/sched/fair.c 		se = cfs_rq->next;
next             4634 kernel/sched/fair.c 			goto next;
next             4650 kernel/sched/fair.c next:
next             5021 kernel/sched/fair.c 	if (!cfs_b->throttled_cfs_rq.next)
next             5661 kernel/sched/fair.c 	} while (group = group->next, group != sd->groups);
next             6388 kernel/sched/fair.c 	for (; pd; pd = pd->next) {
next             6679 kernel/sched/fair.c 		cfs_rq_of(se)->next = se;
next             7168 kernel/sched/fair.c 			(&p->se == cfs_rq_of(&p->se)->next ||
next             7408 kernel/sched/fair.c 			goto next;
next             7413 kernel/sched/fair.c 			goto next;
next             7416 kernel/sched/fair.c 			goto next;
next             7442 kernel/sched/fair.c next:
next             7874 kernel/sched/fair.c 			group = group->next;
next             8291 kernel/sched/fair.c 		sg = sg->next;
next             9121 kernel/sched/fair.c 	unsigned long interval, next;
next             9125 kernel/sched/fair.c 	next = sd->last_balance + interval;
next             9127 kernel/sched/fair.c 	if (time_after(*next_balance, next))
next             9128 kernel/sched/fair.c 		*next_balance = next;
next             10466 kernel/sched/fair.c 	.next			= &idle_sched_class,
next              388 kernel/sched/idle.c static void set_next_task_idle(struct rq *rq, struct task_struct *next, bool first)
next              397 kernel/sched/idle.c 	struct task_struct *next = rq->idle;
next              402 kernel/sched/idle.c 	set_next_task_idle(rq, next, true);
next              404 kernel/sched/idle.c 	return next;
next               91 kernel/sched/rt.c 	rt_rq->highest_prio.next = MAX_RT_PRIO;
next              381 kernel/sched/rt.c 	if (p->prio < rq->rt.highest_prio.next)
next              382 kernel/sched/rt.c 		rq->rt.highest_prio.next = p->prio;
next              393 kernel/sched/rt.c 		rq->rt.highest_prio.next = p->prio;
next              395 kernel/sched/rt.c 		rq->rt.highest_prio.next = MAX_RT_PRIO;
next              460 kernel/sched/rt.c 		tg = list_entry_rcu(tg->list.next,
next              664 kernel/sched/rt.c 			goto next;
next              682 kernel/sched/rt.c next:
next             1543 kernel/sched/rt.c 	struct sched_rt_entity *next = NULL;
next             1551 kernel/sched/rt.c 	next = list_entry(queue->next, struct sched_rt_entity, run_list);
next             1553 kernel/sched/rt.c 	return next;
next             1924 kernel/sched/rt.c 	int next;
next             1958 kernel/sched/rt.c 		next = atomic_read_acquire(&rd->rto_loop_next);
next             1960 kernel/sched/rt.c 		if (rd->rto_loop == next)
next             1963 kernel/sched/rt.c 		rd->rto_loop = next;
next             2091 kernel/sched/rt.c 		if (src_rq->rt.highest_prio.next >=
next             2289 kernel/sched/rt.c 		unsigned long next;
next             2296 kernel/sched/rt.c 		next = DIV_ROUND_UP(min(soft, hard), USEC_PER_SEC/HZ);
next             2297 kernel/sched/rt.c 		if (p->rt.timeout > next) {
next             2341 kernel/sched/rt.c 		if (rt_se->run_list.prev != rt_se->run_list.next) {
next             2361 kernel/sched/rt.c 	.next			= &fair_sched_class,
next              482 kernel/sched/sched.h 			     struct cfs_rq *prev, struct cfs_rq *next);
next              485 kernel/sched/sched.h 			     struct cfs_rq *prev, struct cfs_rq *next) { }
next              516 kernel/sched/sched.h 	struct sched_entity	*next;
next              605 kernel/sched/sched.h 		int		next; /* next highest */
next              653 kernel/sched/sched.h 		u64		next;
next              722 kernel/sched/sched.h 	struct perf_domain *next;
next             1321 kernel/sched/sched.h 	if (unlikely(head->next))
next             1325 kernel/sched/sched.h 	head->next = rq->balance_callback;
next             1411 kernel/sched/sched.h 	struct sched_group	*next;			/* Must be a circular list */
next             1709 kernel/sched/sched.h 	const struct sched_class *next;
next             1786 kernel/sched/sched.h static inline void set_next_task(struct rq *rq, struct task_struct *next)
next             1788 kernel/sched/sched.h 	WARN_ON_ONCE(rq->curr != next);
next             1789 kernel/sched/sched.h 	next->sched_class->set_next_task(rq, next, false);
next             1799 kernel/sched/sched.h 	for (class = (_from); class != (_to); class = class->next)
next              118 kernel/sched/stats.c 	.next  = schedstat_next,
next              226 kernel/sched/stats.h __sched_info_switch(struct rq *rq, struct task_struct *prev, struct task_struct *next)
next              236 kernel/sched/stats.h 	if (next != rq->idle)
next              237 kernel/sched/stats.h 		sched_info_arrive(rq, next);
next              241 kernel/sched/stats.h sched_info_switch(struct rq *rq, struct task_struct *prev, struct task_struct *next)
next              244 kernel/sched/stats.h 		__sched_info_switch(rq, prev, next);
next              252 kernel/sched/stats.h # define sched_info_arrive(rq, next)	do { } while (0)
next              253 kernel/sched/stats.h # define sched_info_switch(rq, t, next)	do { } while (0)
next              122 kernel/sched/stop_task.c 	.next			= &dl_sched_class,
next               98 kernel/sched/topology.c 		group = group->next;
next              162 kernel/sched/topology.c 		if (sd->groups != sd->groups->next)
next              185 kernel/sched/topology.c 	if (parent->groups == parent->groups->next) {
next              240 kernel/sched/topology.c 		tmp = pd->next;
next              251 kernel/sched/topology.c 		pd = pd->next;
next              289 kernel/sched/topology.c 		pd = pd->next;
next              385 kernel/sched/topology.c 		tmp->next = pd;
next              567 kernel/sched/topology.c 		tmp = sg->next;
next              971 kernel/sched/topology.c 			last->next = sg;
next              973 kernel/sched/topology.c 		last->next = first;
next             1128 kernel/sched/topology.c 			last->next = sg;
next             1131 kernel/sched/topology.c 	last->next = first;
next             1159 kernel/sched/topology.c 			goto next;
next             1169 kernel/sched/topology.c next:
next             1170 kernel/sched/topology.c 		sg = sg->next;
next             1799 kernel/sched/topology.c 			sg->next = sg;
next               70 kernel/sched/wait.c 	wait_queue_entry_t *curr, *next;
next               86 kernel/sched/wait.c 	list_for_each_entry_safe_from(curr, next, &wq_head->head, entry) {
next              100 kernel/sched/wait.c 				(&next->entry != &wq_head->head)) {
next              102 kernel/sched/wait.c 			list_add_tail(&bookmark->entry, &next->entry);
next              466 kernel/signal.c 		q = list_entry(queue->list.next, struct sigqueue , list);
next              726 kernel/signal.c 			goto next;
next              730 kernel/signal.c next:
next              480 kernel/softirq.c 	t->next = NULL;
next              482 kernel/softirq.c 	head->tail = &(t->next);
next              516 kernel/softirq.c 		list = list->next;
next              531 kernel/softirq.c 		t->next = NULL;
next              533 kernel/softirq.c 		tl_head->tail = &t->next;
next              552 kernel/softirq.c 	t->next = NULL;
next              632 kernel/softirq.c 	for (i = &per_cpu(tasklet_vec, cpu).head; *i; i = &(*i)->next) {
next              634 kernel/softirq.c 			*i = t->next;
next               36 kernel/task_work.c 		work->next = head;
next               73 kernel/task_work.c 			pprev = &work->next;
next               74 kernel/task_work.c 		else if (cmpxchg(pprev, work, work->next) == work)
next               93 kernel/task_work.c 	struct callback_head *work, *head, *next;
next              112 kernel/task_work.c 			next = work->next;
next              114 kernel/task_work.c 			work = next;
next              274 kernel/time/alarmtimer.c 		struct timerqueue_node *next;
next              278 kernel/time/alarmtimer.c 		next = timerqueue_getnext(&base->timerqueue);
next              280 kernel/time/alarmtimer.c 		if (!next)
next              282 kernel/time/alarmtimer.c 		delta = ktime_sub(next->expires, base->gettime());
next              284 kernel/time/alarmtimer.c 			expires = next->expires;
next              348 kernel/time/clockevents.c 		dev = list_entry(clockevents_released.next,
next              512 kernel/time/hrtimer.c 		struct timerqueue_node *next;
next              515 kernel/time/hrtimer.c 		next = timerqueue_getnext(&base->active);
next              516 kernel/time/hrtimer.c 		timer = container_of(next, struct hrtimer, node);
next              519 kernel/time/hrtimer.c 			next = timerqueue_iterate_next(next);
next              520 kernel/time/hrtimer.c 			if (!next)
next              523 kernel/time/hrtimer.c 			timer = container_of(next, struct hrtimer, node);
next              504 kernel/time/ntp.c 	struct timespec64 next;
next              506 kernel/time/ntp.c 	ktime_get_real_ts64(&next);
next              508 kernel/time/ntp.c 		next.tv_sec = 659;
next              516 kernel/time/ntp.c 		next.tv_sec = 0;
next              520 kernel/time/ntp.c 	next.tv_nsec = target_nsec - next.tv_nsec;
next              521 kernel/time/ntp.c 	if (next.tv_nsec <= 0)
next              522 kernel/time/ntp.c 		next.tv_nsec += NSEC_PER_SEC;
next              523 kernel/time/ntp.c 	if (next.tv_nsec >= NSEC_PER_SEC) {
next              524 kernel/time/ntp.c 		next.tv_sec++;
next              525 kernel/time/ntp.c 		next.tv_nsec -= NSEC_PER_SEC;
next              529 kernel/time/ntp.c 			   timespec64_to_jiffies(&next));
next              766 kernel/time/posix-cpu-timers.c 	struct timerqueue_node *next;
next              769 kernel/time/posix-cpu-timers.c 	while ((next = timerqueue_getnext(head))) {
next              773 kernel/time/posix-cpu-timers.c 		ctmr = container_of(next, struct cpu_timer, node);
next             1116 kernel/time/posix-cpu-timers.c 	struct k_itimer *timer, *next;
next             1156 kernel/time/posix-cpu-timers.c 	list_for_each_entry_safe(timer, next, &firing, it.cpu.elist) {
next             1040 kernel/time/posix-timers.c 		tmr = list_entry(sig->posix_timers.next, struct k_itimer, list);
next              334 kernel/time/tick-broadcast.c 		ktime_t next = ktime_add(dev->next_event, tick_period);
next              336 kernel/time/tick-broadcast.c 		clockevents_program_event(dev, next, true);
next              106 kernel/time/tick-common.c 	ktime_t next = dev->next_event;
next              127 kernel/time/tick-common.c 		next = ktime_add(next, tick_period);
next              129 kernel/time/tick-common.c 		if (!clockevents_program_event(dev, next, false))
next              161 kernel/time/tick-common.c 		ktime_t next;
next              165 kernel/time/tick-common.c 			next = tick_next_period;
next              171 kernel/time/tick-common.c 			if (!clockevents_program_event(dev, next, false))
next              173 kernel/time/tick-common.c 			next = ktime_add(next, tick_period);
next             1228 kernel/time/tick-sched.c 	ktime_t next;
next             1242 kernel/time/tick-sched.c 	next = tick_init_jiffy_update();
next             1244 kernel/time/tick-sched.c 	hrtimer_set_expires(&ts->sched_timer, next);
next              613 kernel/time/timer.c 		timer->entry.next == TIMER_ENTRY_STATIC);
next              818 kernel/time/timer.c 	entry->next = LIST_POISON2;
next             1508 kernel/time/timer.c 	unsigned long clk, next, adj;
next             1511 kernel/time/timer.c 	next = base->clk + NEXT_TIMER_MAX_DELTA;
next             1520 kernel/time/timer.c 			if (time_before(tmp, next))
next             1521 kernel/time/timer.c 				next = tmp;
next             1563 kernel/time/timer.c 	return next;
next             1689 kernel/time/timer.c 		unsigned long next = __next_timer_interrupt(base);
next             1695 kernel/time/timer.c 		if (time_after(next, now)) {
next             1703 kernel/time/timer.c 		base->clk = next;
next               77 kernel/time/timer_list.c 	unsigned long next = 0, i;
next               93 kernel/time/timer_list.c 	while (curr && i < next) {
next              105 kernel/time/timer_list.c 		next++;
next              364 kernel/time/timer_list.c 	.next = timer_list_next,
next              525 kernel/trace/bpf_trace.c 				.next	= ctx_size ? &frag : NULL,
next              390 kernel/trace/fgraph.c 			struct task_struct *prev, struct task_struct *next)
next              407 kernel/trace/fgraph.c 	if (!next->ftrace_timestamp)
next              414 kernel/trace/fgraph.c 	timestamp -= next->ftrace_timestamp;
next              416 kernel/trace/fgraph.c 	for (index = next->curr_ret_stack; index >= 0; index--)
next              417 kernel/trace/fgraph.c 		next->ret_stack[index].calltime += timestamp;
next              206 kernel/trace/ftrace.c 	} else if (rcu_dereference_protected(ftrace_ops_list->next,
next              267 kernel/trace/ftrace.c 	rcu_assign_pointer(ops->next, *list);
next              289 kernel/trace/ftrace.c 	    rcu_dereference_protected(ops->next,
next              295 kernel/trace/ftrace.c 	for (p = list; *p != &ftrace_list_end; p = &(*p)->next)
next              302 kernel/trace/ftrace.c 	*p = (*p)->next;
next              400 kernel/trace/ftrace.c 	struct ftrace_profile_page	*next;
next              442 kernel/trace/ftrace.c 		pg = pg->next;
next              581 kernel/trace/ftrace.c 		pg = pg->next;
next              621 kernel/trace/ftrace.c 		pg->next = (void *)get_zeroed_page(GFP_KERNEL);
next              622 kernel/trace/ftrace.c 		if (!pg->next)
next              624 kernel/trace/ftrace.c 		pg = pg->next;
next              634 kernel/trace/ftrace.c 		pg = pg->next;
next              744 kernel/trace/ftrace.c 		if (!stat->pages->next)
next              746 kernel/trace/ftrace.c 		stat->pages = stat->pages->next;
next             1100 kernel/trace/ftrace.c 	struct ftrace_page	*next;
next             1515 kernel/trace/ftrace.c 	for (pg = ftrace_pages_start; pg; pg = pg->next) {		\
next             1558 kernel/trace/ftrace.c 	for (pg = ftrace_pages_start; pg; pg = pg->next) {
next             1613 kernel/trace/ftrace.c 	     ops != &ftrace_list_end; ops = ops->next) {
next             2452 kernel/trace/ftrace.c 		iter->pg = iter->pg->next;
next             2471 kernel/trace/ftrace.c 		iter->pg = iter->pg->next;
next             2476 kernel/trace/ftrace.c 			iter->pg = iter->pg->next;
next             2930 kernel/trace/ftrace.c 	for (pg = new_pgs; pg; pg = pg->next) {
next             3026 kernel/trace/ftrace.c 		pg->next = kzalloc(sizeof(*pg), GFP_KERNEL);
next             3027 kernel/trace/ftrace.c 		if (!pg->next)
next             3030 kernel/trace/ftrace.c 		pg = pg->next;
next             3040 kernel/trace/ftrace.c 		start_pg = pg->next;
next             3075 kernel/trace/ftrace.c 	struct list_head *next;
next             3091 kernel/trace/ftrace.c 		next = func_probes->next;
next             3092 kernel/trace/ftrace.c 		iter->probe = list_entry(next, struct ftrace_func_probe, list);
next             3111 kernel/trace/ftrace.c 		if (iter->probe->list.next == func_probes)
next             3113 kernel/trace/ftrace.c 		next = iter->probe->list.next;
next             3114 kernel/trace/ftrace.c 		iter->probe = list_entry(next, struct ftrace_func_probe, list);
next             3131 kernel/trace/ftrace.c 		hnd = hnd->next;
next             3208 kernel/trace/ftrace.c 	iter->mod_list = iter->mod_list->next;
next             3285 kernel/trace/ftrace.c 		if (iter->pg->next) {
next             3286 kernel/trace/ftrace.c 			iter->pg = iter->pg->next;
next             3481 kernel/trace/ftrace.c 	.next = t_next,
next             4643 kernel/trace/ftrace.c 	char *func, *command, *next = buff;
next             4647 kernel/trace/ftrace.c 	func = strsep(&next, ":");
next             4649 kernel/trace/ftrace.c 	if (!next) {
next             4660 kernel/trace/ftrace.c 	command = strsep(&next, ":");
next             4665 kernel/trace/ftrace.c 			ret = p->func(tr, hash, func, command, next, enable);
next             5214 kernel/trace/ftrace.c 	.next = g_next,
next             5614 kernel/trace/ftrace.c 		if (WARN_ON(ftrace_pages->next)) {
next             5616 kernel/trace/ftrace.c 			while (ftrace_pages->next)
next             5617 kernel/trace/ftrace.c 				ftrace_pages = ftrace_pages->next;
next             5620 kernel/trace/ftrace.c 		ftrace_pages->next = start_pg;
next             5638 kernel/trace/ftrace.c 			if (WARN_ON(!pg->next))
next             5640 kernel/trace/ftrace.c 			pg = pg->next;
next             5648 kernel/trace/ftrace.c 	WARN_ON(pg->next);
next             5692 kernel/trace/ftrace.c #define next_to_ftrace_page(p) container_of(p, struct ftrace_page, next)
next             5701 kernel/trace/ftrace.c 	for (ops = ftrace_ops_list; ops != &ftrace_list_end; ops = ops->next) {
next             5809 kernel/trace/ftrace.c 			*last_pg = pg->next;
next             5811 kernel/trace/ftrace.c 			pg->next = tmp_page;
next             5814 kernel/trace/ftrace.c 			last_pg = &pg->next;
next             5826 kernel/trace/ftrace.c 		tmp_page = pg->next;
next             6146 kernel/trace/ftrace.c 	for (pg = ftrace_pages_start; pg; last_pg = &pg->next, pg = *last_pg) {
next             6166 kernel/trace/ftrace.c 			*last_pg = pg->next;
next             6170 kernel/trace/ftrace.c 			pg = container_of(last_pg, struct ftrace_page, next);
next             6424 kernel/trace/ftrace.c 		    struct task_struct *prev, struct task_struct *next)
next             6432 kernel/trace/ftrace.c 		       trace_ignore_this_task(pid_list, next));
next             6566 kernel/trace/ftrace.c 	.next = fpid_next,
next               24 kernel/trace/ftrace_internal.h 	while (likely(op = rcu_dereference_raw_check((op)->next)) &&	\
next              874 kernel/trace/ring_buffer.c 	val = (unsigned long)list->next;
next              893 kernel/trace/ring_buffer.c 	return rb_list_head(list->next) != &page->list;
next              904 kernel/trace/ring_buffer.c 	ptr = (unsigned long *)&list->next;
next              928 kernel/trace/ring_buffer.c 	unsigned long *ptr = (unsigned long *)&list->next;
next              961 kernel/trace/ring_buffer.c 	ret = cmpxchg((unsigned long *)&list->next,
next             1001 kernel/trace/ring_buffer.c 	struct list_head *p = rb_list_head((*bpage)->list.next);
next             1019 kernel/trace/ring_buffer.c 	if (RB_WARN_ON(cpu_buffer, rb_list_head(list->prev->next) != list))
next             1047 kernel/trace/ring_buffer.c 	unsigned long *ptr = (unsigned long *)&old->list.prev->next;
next             1142 kernel/trace/ring_buffer.c 	if (RB_WARN_ON(cpu_buffer, rb_list_head(list->next) != list->next))
next             1165 kernel/trace/ring_buffer.c 	if (RB_WARN_ON(cpu_buffer, head->next->prev != head))
next             1167 kernel/trace/ring_buffer.c 	if (RB_WARN_ON(cpu_buffer, head->prev->next != head))
next             1175 kernel/trace/ring_buffer.c 			       bpage->list.next->prev != &bpage->list))
next             1178 kernel/trace/ring_buffer.c 			       bpage->list.prev->next != &bpage->list))
next             1275 kernel/trace/ring_buffer.c 	cpu_buffer->pages = pages.next;
next             1524 kernel/trace/ring_buffer.c 		tail_page = rb_list_head(tail_page->next);
next             1528 kernel/trace/ring_buffer.c 	first_page = list_entry(rb_list_head(to_remove->next),
next             1532 kernel/trace/ring_buffer.c 		to_remove = rb_list_head(to_remove)->next;
next             1536 kernel/trace/ring_buffer.c 	next_page = rb_list_head(to_remove)->next;
next             1543 kernel/trace/ring_buffer.c 	tail_page->next = (struct list_head *)((unsigned long)next_page |
next             1639 kernel/trace/ring_buffer.c 		first_page = pages->next;
next             1645 kernel/trace/ring_buffer.c 		last_page->next = head_page_with_bit;
next             1648 kernel/trace/ring_buffer.c 		r = cmpxchg(&prev_page->next, head_page_with_bit, first_page);
next             3730 kernel/trace/ring_buffer.c 	cpu_buffer->reader_page->list.next = rb_list_head(reader->list.next);
next             3779 kernel/trace/ring_buffer.c 	rb_list_head(reader->list.next)->prev = &cpu_buffer->reader_page->list;
next              140 kernel/trace/trace.c 	union trace_eval_map_item	*next;
next             1758 kernel/trace/trace.c 			for (t = trace_types; t; t = t->next) {
next             1760 kernel/trace/trace.c 					*last = t->next;
next             1763 kernel/trace/trace.c 				last = &t->next;
next             1819 kernel/trace/trace.c 	for (t = trace_types; t; t = t->next) {
next             1851 kernel/trace/trace.c 	type->next = trace_types;
next             2293 kernel/trace/trace.c 					  struct task_struct *next, int flags)
next             2305 kernel/trace/trace.c 	done &= !(flags & TRACE_RECORD_CMDLINE) || trace_save_cmdline(next);
next             2307 kernel/trace/trace.c 	done &= !(flags & TRACE_RECORD_TGID) || trace_save_tgid(next);
next             2674 kernel/trace/trace.c 		export = rcu_dereference_raw_check(export->next);
next             2683 kernel/trace/trace.c 	rcu_assign_pointer(export->next, *list);
next             2698 kernel/trace/trace.c 	for (p = list; *p != NULL; p = &(*p)->next)
next             2705 kernel/trace/trace.c 	rcu_assign_pointer(*p, (*p)->next);
next             3301 kernel/trace/trace.c 	struct trace_entry *ent, *next = NULL;
next             3333 kernel/trace/trace.c 		if (ent && (!next || ts < next_ts)) {
next             3334 kernel/trace/trace.c 			next = ent;
next             3353 kernel/trace/trace.c 	return next;
next             4065 kernel/trace/trace.c 	.next		= s_next,
next             4316 kernel/trace/trace.c 		t = t->next;
next             4330 kernel/trace/trace.c 		t = get_tracer_for_array(tr, t->next);
next             4363 kernel/trace/trace.c 	if (t->next)
next             4373 kernel/trace/trace.c 	.next		= t_next,
next             5079 kernel/trace/trace.c 	.next		= saved_tgids_next,
next             5158 kernel/trace/trace.c 	.next		= saved_cmdlines_next,
next             5259 kernel/trace/trace.c 		if (ptr->tail.next) {
next             5260 kernel/trace/trace.c 			ptr = ptr->tail.next;
next             5326 kernel/trace/trace.c 	.next		= eval_map_next,
next             5386 kernel/trace/trace.c 			if (!ptr->tail.next)
next             5388 kernel/trace/trace.c 			ptr = ptr->tail.next;
next             5391 kernel/trace/trace.c 		ptr->tail.next = map_array;
next             5654 kernel/trace/trace.c 	for (t = trace_types; t; t = t->next) {
next             7083 kernel/trace/trace.c 	struct tracing_log_err *err, *next;
next             7086 kernel/trace/trace.c 	list_for_each_entry_safe(err, next, &tr->err_log, list) {
next             7148 kernel/trace/trace.c 	.next   = tracing_err_log_seq_next,
next             8372 kernel/trace/trace.c 	for (t = trace_types; t; t = t->next)
next             8741 kernel/trace/trace.c 		last = &map->tail.next;
next             8742 kernel/trace/trace.c 		map = map->tail.next;
next             8747 kernel/trace/trace.c 	*last = trace_eval_jmp_to_tail(map)->tail.next;
next             8839 kernel/trace/trace.c 	.next           = NULL,
next              498 kernel/trace/trace.h 	struct tracer		*next;
next              130 kernel/trace/trace_dynevent.c 	.next	= dyn_event_seq_next,
next              190 kernel/trace/trace_events.c 	struct ftrace_event_field *field, *next;
next              194 kernel/trace/trace_events.c 	list_for_each_entry_safe(field, next, head, link) {
next              542 kernel/trace/trace_events.c 		    struct task_struct *prev, struct task_struct *next)
next              551 kernel/trace/trace_events.c 		       trace_ignore_this_task(pid_list, next));
next              556 kernel/trace/trace_events.c 		    struct task_struct *prev, struct task_struct *next)
next              564 kernel/trace/trace_events.c 		       trace_ignore_this_task(pid_list, next));
next             1288 kernel/trace/trace_events.c 	.next		= f_next,
next             1661 kernel/trace/trace_events.c 	.next = t_next,
next             1668 kernel/trace/trace_events.c 	.next = s_next,
next             1675 kernel/trace/trace_events.c 	.next = p_next,
next             2878 kernel/trace/trace_events.c 	struct trace_event_file *file, *next;
next             2880 kernel/trace/trace_events.c 	list_for_each_entry_safe(file, next, &tr->events, list)
next              448 kernel/trace/trace_events_filter.c 		const char *next = ptr++;
next              450 kernel/trace/trace_events_filter.c 		if (isspace(*next))
next              453 kernel/trace/trace_events_filter.c 		switch (*next) {
next              462 kernel/trace/trace_events_filter.c 			if (!is_not(next))
next              469 kernel/trace/trace_events_filter.c 			parse_error(pe, FILT_ERR_TOO_MANY_PREDS, next - str);
next              476 kernel/trace/trace_events_filter.c 		len = parse_pred(next, data, ptr - str, pe, &prog[N].pred);
next              481 kernel/trace/trace_events_filter.c 		ptr = next + len;
next              487 kernel/trace/trace_events_filter.c 			next = ptr++;
next              488 kernel/trace/trace_events_filter.c 			if (isspace(*next))
next              491 kernel/trace/trace_events_filter.c 			switch (*next) {
next              498 kernel/trace/trace_events_filter.c 				if (next[1] == next[0]) {
next              505 kernel/trace/trace_events_filter.c 					    next - str);
next              515 kernel/trace/trace_events_filter.c 			if (*next == '&') {			/* #8 */
next              523 kernel/trace/trace_events_filter.c 			if (*next == '|') {			/* #10 */
next              527 kernel/trace/trace_events_filter.c 			if (!*next)				/* #11 */
next             1493 kernel/trace/trace_events_hist.c 	.next	= dyn_event_seq_next,
next              171 kernel/trace/trace_events_trigger.c 	.next = trigger_next,
next              219 kernel/trace/trace_events_trigger.c 	char *command, *next = buff;
next              223 kernel/trace/trace_events_trigger.c 	command = strsep(&next, ": \t");
next              229 kernel/trace/trace_events_trigger.c 			ret = p->func(p, file, buff, command, next);
next              426 kernel/trace/trace_functions_graph.c 	struct ftrace_graph_ret_entry *next;
next              434 kernel/trace/trace_functions_graph.c 		next = &data->ret;
next              456 kernel/trace/trace_functions_graph.c 		next = ring_buffer_event_data(event);
next              469 kernel/trace/trace_functions_graph.c 			if (next->ent.type == TRACE_GRAPH_RET)
next              470 kernel/trace/trace_functions_graph.c 				data->ret = *next;
next              472 kernel/trace/trace_functions_graph.c 				data->ret.ent.type = next->ent.type;
next              476 kernel/trace/trace_functions_graph.c 	if (next->ent.type != TRACE_GRAPH_RET)
next              479 kernel/trace/trace_functions_graph.c 	if (curr->ent.pid != next->ent.pid ||
next              480 kernel/trace/trace_functions_graph.c 			curr->graph_ent.func != next->ret.func)
next              487 kernel/trace/trace_functions_graph.c 	return next;
next              953 kernel/trace/trace_kprobe.c 	.next   = dyn_event_seq_next,
next             1011 kernel/trace/trace_kprobe.c 	.next   = dyn_event_seq_next,
next              154 kernel/trace/trace_printk.c 	if (mod_fmt->list.next == &trace_bprintk_fmt_list)
next              157 kernel/trace/trace_printk.c 	mod_fmt = container_of(mod_fmt->list.next, typeof(*mod_fmt), list);
next              344 kernel/trace/trace_printk.c 	.next = t_next,
next               25 kernel/trace/trace_sched_switch.c 		   struct task_struct *prev, struct task_struct *next)
next               34 kernel/trace/trace_sched_switch.c 	tracing_record_taskinfo_sched_switch(prev, next, flags);
next              377 kernel/trace/trace_sched_wakeup.c 			   struct task_struct *next,
next              393 kernel/trace/trace_sched_wakeup.c 	entry->next_pid			= next->pid;
next              394 kernel/trace/trace_sched_wakeup.c 	entry->next_prio		= next->prio;
next              395 kernel/trace/trace_sched_wakeup.c 	entry->next_state		= task_state_index(next);
next              396 kernel/trace/trace_sched_wakeup.c 	entry->next_cpu	= task_cpu(next);
next              432 kernel/trace/trace_sched_wakeup.c 			  struct task_struct *prev, struct task_struct *next)
next              455 kernel/trace/trace_sched_wakeup.c 	if (next != wakeup_task)
next              470 kernel/trace/trace_sched_wakeup.c 	if (unlikely(!tracer_enabled || next != wakeup_task))
next              477 kernel/trace/trace_sched_wakeup.c 	tracing_sched_switch_trace(wakeup_trace, prev, next, flags, pc);
next              472 kernel/trace/trace_stack.c 	.next		= t_next,
next              229 kernel/trace/trace_stat.c 	.next		= stat_seq_next,
next              757 kernel/trace/trace_uprobe.c 	.next   = dyn_event_seq_next,
next              812 kernel/trace/trace_uprobe.c 	.next   = dyn_event_seq_next,
next               81 kernel/tracepoint.c 		early_probes = tmp->next;
next              102 kernel/tracepoint.c 			tp_probes->rcu.next = early_probes;
next              691 kernel/user_namespace.c 	.next = m_next,
next              698 kernel/user_namespace.c 	.next = m_next,
next              705 kernel/user_namespace.c 	.next = m_next,
next              871 kernel/workqueue.c 	struct worker *next, *worker = kthread_data(task);
next              903 kernel/workqueue.c 		next = first_idle_worker(pool);
next              904 kernel/workqueue.c 		if (next)
next              905 kernel/workqueue.c 			wake_up_process(next->task);
next             1774 kernel/workqueue.c 			 (worker->hentry.next || worker->hentry.pprev)))
next             2678 kernel/workqueue.c 		head = worker->scheduled.next;
next             2682 kernel/workqueue.c 		head = target->entry.next;
next             2858 kernel/workqueue.c 		struct wq_flusher *next, *tmp;
next             2861 kernel/workqueue.c 		list_for_each_entry_safe(next, tmp, &wq->flusher_queue, list) {
next             2862 kernel/workqueue.c 			if (next->flush_color != wq->flush_color)
next             2864 kernel/workqueue.c 			list_del_init(&next->list);
next             2865 kernel/workqueue.c 			complete(&next->done);
next             2902 kernel/workqueue.c 		WARN_ON_ONCE(wq->flush_color != next->flush_color);
next             2904 kernel/workqueue.c 		list_del_init(&next->list);
next             2905 kernel/workqueue.c 		wq->first_flusher = next;
next             4174 kernel/workqueue.c 		WARN(!ret && (wq->pwqs.next != &wq->dfl_pwq->pwqs_node ||
next              482 lib/842/842_compress.c 	u64 last, next, pad, total;
next              517 lib/842/842_compress.c 		next = get_unaligned((u64 *)p->in);
next              528 lib/842/842_compress.c 		if (next == last) {
next              536 lib/842/842_compress.c 			if (next == last) /* reached max repeat bits */
next              540 lib/842/842_compress.c 		if (next == 0)
next              549 lib/842/842_compress.c 		last = next;
next              205 lib/cmdline.c  	char *next;
next              243 lib/cmdline.c  		next = args + i + 1;
next              245 lib/cmdline.c  		next = args + i;
next              248 lib/cmdline.c  	return skip_spaces(next);
next               79 lib/cpumask.c  	int next;
next               82 lib/cpumask.c  	next = cpumask_next(n, mask);
next               84 lib/cpumask.c  	if (wrap && n < start && next >= start) {
next               87 lib/cpumask.c  	} else if (next >= nr_cpumask_bits) {
next               93 lib/cpumask.c  	return next;
next              890 lib/debugobjects.c 			  unsigned int expect, unsigned int next)
next              909 lib/debugobjects.c 				obj->astate = next;
next              734 lib/dynamic_debug.c 	iter->table = list_entry(ddebug_tables.next,
next              757 lib/dynamic_debug.c 		iter->table = list_entry(iter->table->link.next,
next              851 lib/dynamic_debug.c 	.next = ddebug_proc_next,
next              984 lib/dynamic_debug.c 		struct ddebug_table *dt = list_entry(ddebug_tables.next,
next              199 lib/error-inject.c 	.next  = ei_seq_next,
next              395 lib/idr.c      next:
next              433 lib/idr.c      			goto next;
next              107 lib/ioremap.c  	unsigned long next;
next              113 lib/ioremap.c  		next = pmd_addr_end(addr, end);
next              115 lib/ioremap.c  		if (ioremap_try_huge_pmd(pmd, addr, next, phys_addr, prot))
next              118 lib/ioremap.c  		if (ioremap_pte_range(pmd, addr, next, phys_addr, prot))
next              120 lib/ioremap.c  	} while (pmd++, phys_addr += (next - addr), addr = next, addr != end);
next              150 lib/ioremap.c  	unsigned long next;
next              156 lib/ioremap.c  		next = pud_addr_end(addr, end);
next              158 lib/ioremap.c  		if (ioremap_try_huge_pud(pud, addr, next, phys_addr, prot))
next              161 lib/ioremap.c  		if (ioremap_pmd_range(pud, addr, next, phys_addr, prot))
next              163 lib/ioremap.c  	} while (pud++, phys_addr += (next - addr), addr = next, addr != end);
next              193 lib/ioremap.c  	unsigned long next;
next              199 lib/ioremap.c  		next = p4d_addr_end(addr, end);
next              201 lib/ioremap.c  		if (ioremap_try_huge_p4d(p4d, addr, next, phys_addr, prot))
next              204 lib/ioremap.c  		if (ioremap_pud_range(p4d, addr, next, phys_addr, prot))
next              206 lib/ioremap.c  	} while (p4d++, phys_addr += (next - addr), addr = next, addr != end);
next              215 lib/ioremap.c  	unsigned long next;
next              224 lib/ioremap.c  		next = pgd_addr_end(addr, end);
next              225 lib/ioremap.c  		err = ioremap_p4d_range(pgd, addr, next, phys_addr, prot);
next              228 lib/ioremap.c  	} while (pgd++, phys_addr += (next - addr), addr = next, addr != end);
next              329 lib/iov_iter.c 	int next = pipe->curbuf + pipe->nrbufs;
next              334 lib/iov_iter.c 		if (unlikely(idx != ((next - 1) & (pipe->buffers - 1))))
next              341 lib/iov_iter.c 		if (idx != (next & (pipe->buffers - 1)))
next              568 lib/iov_iter.c 	__wsum next = csum_partial_copy_nocheck(from, to, len, 0);
next              569 lib/iov_iter.c 	return csum_block_add(sum, next, off);
next             1408 lib/iov_iter.c 	__wsum sum, next;
next             1417 lib/iov_iter.c 		next = csum_and_copy_from_user(v.iov_base,
next             1421 lib/iov_iter.c 			sum = csum_block_add(sum, next, off);
next             1448 lib/iov_iter.c 	__wsum sum, next;
next             1459 lib/iov_iter.c 		next = csum_and_copy_from_user(v.iov_base,
next             1464 lib/iov_iter.c 		sum = csum_block_add(sum, next, off);
next             1492 lib/iov_iter.c 	__wsum sum, next;
next             1505 lib/iov_iter.c 		next = csum_and_copy_to_user((from += v.iov_len) - v.iov_len,
next             1509 lib/iov_iter.c 			sum = csum_block_add(sum, next, off);
next              105 lib/irq_poll.c 		iop = list_entry(list->next, struct irq_poll, list);
next              379 lib/klist.c    	struct klist_node *next;
next              385 lib/klist.c    		next = to_klist_node(last->n_node.next);
next              389 lib/klist.c    		next = to_klist_node(i->i_klist->k_list.next);
next              392 lib/klist.c    	while (next != to_klist_node(&i->i_klist->k_list)) {
next              393 lib/klist.c    		if (likely(!knode_dead(next))) {
next              394 lib/klist.c    			kref_get(&next->n_ref);
next              395 lib/klist.c    			i->i_cur = next;
next              398 lib/klist.c    		next = to_klist_node(next->n_node.next);
next              102 lib/kobject_uevent.c 	const char *next = buf;
next              104 lib/kobject_uevent.c 	while (next <= buf_end && *next != delim)
next              105 lib/kobject_uevent.c 		if (!isalnum(*next++))
next              108 lib/kobject_uevent.c 	if (next == buf)
next              111 lib/kobject_uevent.c 	return next;
next              118 lib/kobject_uevent.c 	const char *next, *buf_end, *key;
next              141 lib/kobject_uevent.c 	next = buf + UUID_STRING_LEN;
next              144 lib/kobject_uevent.c 	while (next <= buf_end) {
next              145 lib/kobject_uevent.c 		if (*next != ' ')
next              149 lib/kobject_uevent.c 		key = ++next;
next              153 lib/kobject_uevent.c 		buf = next;
next              154 lib/kobject_uevent.c 		next = action_arg_word_end(buf, buf_end, '=');
next              155 lib/kobject_uevent.c 		if (!next || next > buf_end || *next != '=')
next              157 lib/kobject_uevent.c 		key_len = next - buf;
next              160 lib/kobject_uevent.c 		if (++next > buf_end)
next              163 lib/kobject_uevent.c 		buf = next;
next              164 lib/kobject_uevent.c 		next = action_arg_word_end(buf, buf_end, ' ');
next              165 lib/kobject_uevent.c 		if (!next)
next              169 lib/kobject_uevent.c 				   key_len, key, (int) (next - buf), buf))
next               21 lib/list_debug.c 		      struct list_head *next)
next               23 lib/list_debug.c 	if (CHECK_DATA_CORRUPTION(next->prev != prev,
next               25 lib/list_debug.c 			prev, next->prev, next) ||
next               26 lib/list_debug.c 	    CHECK_DATA_CORRUPTION(prev->next != next,
next               28 lib/list_debug.c 			next, prev->next, prev) ||
next               29 lib/list_debug.c 	    CHECK_DATA_CORRUPTION(new == prev || new == next,
next               31 lib/list_debug.c 			new, prev, next))
next               40 lib/list_debug.c 	struct list_head *prev, *next;
next               43 lib/list_debug.c 	next = entry->next;
next               45 lib/list_debug.c 	if (CHECK_DATA_CORRUPTION(next == LIST_POISON1,
next               51 lib/list_debug.c 	    CHECK_DATA_CORRUPTION(prev->next != entry,
next               53 lib/list_debug.c 			entry, prev->next) ||
next               54 lib/list_debug.c 	    CHECK_DATA_CORRUPTION(next->prev != entry,
next               56 lib/list_debug.c 			entry, next->prev))
next               28 lib/list_sort.c 			tail = &a->next;
next               29 lib/list_sort.c 			a = a->next;
next               36 lib/list_sort.c 			tail = &b->next;
next               37 lib/list_sort.c 			b = b->next;
next               64 lib/list_sort.c 			tail->next = a;
next               67 lib/list_sort.c 			a = a->next;
next               71 lib/list_sort.c 			tail->next = b;
next               74 lib/list_sort.c 			b = b->next;
next               83 lib/list_sort.c 	tail->next = b;
next               96 lib/list_sort.c 		b = b->next;
next              100 lib/list_sort.c 	tail->next = head;
next              192 lib/list_sort.c 	struct list_head *list = head->next, *pending = NULL;
next              199 lib/list_sort.c 	head->prev->next = NULL;
next              239 lib/list_sort.c 		list = list->next;
next              240 lib/list_sort.c 		pending->next = NULL;
next              248 lib/list_sort.c 		struct list_head *next = pending->prev;
next              250 lib/list_sort.c 		if (!next)
next              253 lib/list_sort.c 		pending = next;
next               32 lib/llist.c    		new_last->next = first = READ_ONCE(head->first);
next               55 lib/llist.c    	struct llist_node *entry, *old_entry, *next;
next               62 lib/llist.c    		next = READ_ONCE(entry->next);
next               63 lib/llist.c    		entry = cmpxchg(&head->first, old_entry, next);
next               85 lib/llist.c    		head = head->next;
next               86 lib/llist.c    		tmp->next = new_head;
next              327 lib/lru_cache.c 		n = lc->free.next;
next               46 lib/lzo/lzo1x_compress.c next:
next              298 lib/lzo/lzo1x_compress.c 		goto next;
next               44 lib/lzo/lzo1x_decompress_safe.c 	size_t t, next;
next               68 lib/lzo/lzo1x_decompress_safe.c 			next = t;
next              121 lib/lzo/lzo1x_decompress_safe.c 				next = t & 3;
next              132 lib/lzo/lzo1x_decompress_safe.c 				next = t & 3;
next              139 lib/lzo/lzo1x_decompress_safe.c 			next = t & 3;
next              163 lib/lzo/lzo1x_decompress_safe.c 			next = get_unaligned_le16(ip);
next              165 lib/lzo/lzo1x_decompress_safe.c 			m_pos -= next >> 2;
next              166 lib/lzo/lzo1x_decompress_safe.c 			next &= 3;
next              169 lib/lzo/lzo1x_decompress_safe.c 			next = get_unaligned_le16(ip);
next              170 lib/lzo/lzo1x_decompress_safe.c 			if (((next & 0xfffc) == 0xfffc) &&
next              180 lib/lzo/lzo1x_decompress_safe.c 				next &= 3;
next              202 lib/lzo/lzo1x_decompress_safe.c 					next = get_unaligned_le16(ip);
next              205 lib/lzo/lzo1x_decompress_safe.c 				m_pos -= next >> 2;
next              206 lib/lzo/lzo1x_decompress_safe.c 				next &= 3;
next              227 lib/lzo/lzo1x_decompress_safe.c 					state = next;
next              229 lib/lzo/lzo1x_decompress_safe.c 					op += next;
next              230 lib/lzo/lzo1x_decompress_safe.c 					ip += next;
next              253 lib/lzo/lzo1x_decompress_safe.c 		state = next;
next              254 lib/lzo/lzo1x_decompress_safe.c 		t = next;
next              120 lib/mpi/mpi-internal.h 	struct karatsuba_ctx *next;
next              374 lib/mpi/mpih-mul.c 			if (!ctx->next) {
next              375 lib/mpi/mpih-mul.c 				ctx->next = kzalloc(sizeof *ctx, GFP_KERNEL);
next              376 lib/mpi/mpih-mul.c 				if (!ctx->next)
next              382 lib/mpi/mpih-mul.c 						       ctx->next) < 0)
next              401 lib/mpi/mpih-mul.c 	for (ctx = ctx->next; ctx; ctx = ctx2) {
next              402 lib/mpi/mpih-mul.c 		ctx2 = ctx->next;
next               79 lib/oid_registry.c 				goto next;
next               83 lib/oid_registry.c 				goto next;
next               87 lib/oid_registry.c 	next:
next               35 lib/plist.c    	WARN(n->prev != p || p->next != n,
next               39 lib/plist.c    			 t, t->next, t->prev,
next               40 lib/plist.c    			p, p->next, p->prev,
next               41 lib/plist.c    			n, n->next, n->prev);
next               46 lib/plist.c    	struct list_head *prev = top, *next = top->next;
next               48 lib/plist.c    	plist_check_prev_next(top, prev, next);
next               49 lib/plist.c    	while (next != top) {
next               50 lib/plist.c    		prev = next;
next               51 lib/plist.c    		next = prev->next;
next               52 lib/plist.c    		plist_check_prev_next(top, prev, next);
next               94 lib/plist.c    		iter = list_entry(iter->prio_list.next,
next              117 lib/plist.c    		if (node->node_list.next != &head->node_list) {
next              118 lib/plist.c    			struct plist_node *next;
next              120 lib/plist.c    			next = list_entry(node->node_list.next,
next              124 lib/plist.c    			if (list_empty(&next->prio_list))
next              125 lib/plist.c    				list_add(&next->prio_list, &node->prio_list);
next              204 lib/plist.c    		BUG_ON(prio_pos->prio_list.next != &node_pos->prio_list);
next              209 lib/plist.c    	BUG_ON(prio_pos->prio_list.next != &first->prio_list);
next              222 lib/rhashtable.c 	struct rhash_head *head, *next, *entry;
next              234 lib/rhashtable.c 		next = rht_dereference_bucket(entry->next, old_tbl, old_hash);
next              236 lib/rhashtable.c 		if (rht_is_a_nulls(next))
next              239 lib/rhashtable.c 		pprev = &entry->next;
next              251 lib/rhashtable.c 	RCU_INIT_POINTER(entry->next, head);
next              256 lib/rhashtable.c 		rcu_assign_pointer(*pprev, next);
next              259 lib/rhashtable.c 		rht_assign_locked(bkt, next);
next              501 lib/rhashtable.c 			pprev = &head->next;
next              511 lib/rhashtable.c 		RCU_INIT_POINTER(list->next, plist);
next              512 lib/rhashtable.c 		head = rht_dereference_bucket(head->next, tbl, hash);
next              513 lib/rhashtable.c 		RCU_INIT_POINTER(list->rhead.next, head);
next              560 lib/rhashtable.c 	RCU_INIT_POINTER(obj->next, head);
next              565 lib/rhashtable.c 		RCU_INIT_POINTER(list->next, NULL);
next              754 lib/rhashtable.c 			     list = rcu_dereference(list->next)) {
next              800 lib/rhashtable.c 						goto next;
next              802 lib/rhashtable.c 					list = rcu_dereference(list->next);
next              812 lib/rhashtable.c next:
next              860 lib/rhashtable.c 		if (!rhlist || !(list = rcu_dereference(list->next))) {
next              861 lib/rhashtable.c 			p = rcu_dereference(p->next);
next             1103 lib/rhashtable.c 		list = rht_dereference(list->next, ht);
next             1137 lib/rhashtable.c 			struct rhash_head *pos, *next;
next             1141 lib/rhashtable.c 			     next = !rht_is_a_nulls(pos) ?
next             1142 lib/rhashtable.c 					rht_dereference(pos->next, ht) : NULL;
next             1144 lib/rhashtable.c 			     pos = next,
next             1145 lib/rhashtable.c 			     next = !rht_is_a_nulls(pos) ?
next             1146 lib/rhashtable.c 					rht_dereference(pos->next, ht) : NULL)
next              195 lib/scatterlist.c 	struct scatterlist *sgl, *next;
next              213 lib/scatterlist.c 			next = sg_chain_ptr(&sgl[curr_max_ents - 1]);
next              218 lib/scatterlist.c 			next = NULL;
next              226 lib/scatterlist.c 		sgl = next;
next               60 lib/stackdepot.c 	struct stack_record *next;	/* Link in the hashtable */
next              183 lib/stackdepot.c 	for (found = bucket; found; found = found->next) {
next              280 lib/stackdepot.c 			new->next = *bucket;
next              100 lib/test_list_sort.c 	for (cur = head.next; cur->next != &head; cur = cur->next) {
next              104 lib/test_list_sort.c 		if (cur->next->prev != cur) {
next              109 lib/test_list_sort.c 		cmp_result = cmp(NULL, cur, cur->next);
next              116 lib/test_list_sort.c 		el1 = container_of(cur->next, struct debug_el, list);
next              497 lib/test_rhashtable.c 		struct rhash_head *pos, *next;
next              501 lib/test_rhashtable.c 		next = !rht_is_a_nulls(pos) ? rht_dereference(pos->next, ht) : NULL;
next              512 lib/test_rhashtable.c 				list = rht_dereference(list->next, ht);
next              520 lib/test_rhashtable.c 			pos = next,
next              521 lib/test_rhashtable.c 			next = !rht_is_a_nulls(pos) ?
next              522 lib/test_rhashtable.c 				rht_dereference(pos->next, ht) : NULL;
next              209 lib/test_xarray.c 		unsigned long next = base + (1UL << order);
next              216 lib/test_xarray.c 		XA_BUG_ON(xa, xa_store_index(xa, next, GFP_KERNEL));
next              219 lib/test_xarray.c 		for (i = base; i < next; i++) {
next              244 lib/test_xarray.c 		XA_BUG_ON(xa, xa_get_mark(xa, next, XA_MARK_0));
next              245 lib/test_xarray.c 		XA_BUG_ON(xa, xa_get_mark(xa, next, XA_MARK_1));
next              246 lib/test_xarray.c 		XA_BUG_ON(xa, xa_get_mark(xa, next, XA_MARK_2));
next              248 lib/test_xarray.c 		xa_erase_index(xa, next);
next              781 lib/test_xarray.c 	u32 next = 0;
next              787 lib/test_xarray.c 				&next, GFP_KERNEL) != 0);
next              790 lib/test_xarray.c 	next = 0x3ffd;
next              792 lib/test_xarray.c 				&next, GFP_KERNEL) != 0);
next              804 lib/test_xarray.c 					&next, GFP_KERNEL) != (id == 1));
next              812 lib/test_xarray.c 	next = UINT_MAX;
next              814 lib/test_xarray.c 				xa_limit_32b, &next, GFP_KERNEL) != 0);
next              817 lib/test_xarray.c 				xa_limit_32b, &next, GFP_KERNEL) != 1);
next              820 lib/test_xarray.c 				xa_limit_32b, &next, GFP_KERNEL) != 0);
next              913 lib/test_xarray.c 	unsigned long next = 4 << order;
next              917 lib/test_xarray.c 	XA_BUG_ON(xa, xa_store_index(xa, next, GFP_KERNEL) != NULL);
next              918 lib/test_xarray.c 	XA_BUG_ON(xa, xa_store_index(xa, next + 1, GFP_KERNEL) != NULL);
next              927 lib/test_xarray.c 	XA_BUG_ON(xa, (index < multi) || (index >= next));
next              929 lib/test_xarray.c 			xa_mk_value(next));
next              930 lib/test_xarray.c 	XA_BUG_ON(xa, index != next);
next              931 lib/test_xarray.c 	XA_BUG_ON(xa, xa_find_after(xa, &index, next, XA_PRESENT) != NULL);
next              932 lib/test_xarray.c 	XA_BUG_ON(xa, index != next);
next              935 lib/test_xarray.c 	xa_erase_index(xa, next);
next              936 lib/test_xarray.c 	xa_erase_index(xa, next + 1);
next               85 lib/timerqueue.c 	struct rb_node *next;
next               89 lib/timerqueue.c 	next = rb_next(&node->node);
next               90 lib/timerqueue.c 	if (!next)
next               92 lib/timerqueue.c 	return container_of(next, struct timerqueue_node, node);
next               79 lib/ts_bm.c    				     goto next;
next               85 lib/ts_bm.c    next:			bs = bm->bad_shift[text[shift-i]];
next              137 lib/ts_fsm.c   	struct ts_fsm_token *cur = NULL, *next;
next              169 lib/ts_fsm.c   			next = &fsm->tokens[tok_idx + 1];
next              171 lib/ts_fsm.c   			next = NULL;
next              199 lib/ts_fsm.c   			if (next == NULL)
next              205 lib/ts_fsm.c   			while (!match_token(next, data[block_idx])) {
next              222 lib/ts_fsm.c   			while (!match_token(next, data[block_idx])) {
next             3134 lib/vsprintf.c 	char *next;
next             3347 lib/vsprintf.c 				simple_strtol(str, &next, base) :
next             3348 lib/vsprintf.c 				simple_strtoll(str, &next, base);
next             3351 lib/vsprintf.c 				simple_strtoul(str, &next, base) :
next             3352 lib/vsprintf.c 				simple_strtoull(str, &next, base);
next             3354 lib/vsprintf.c 		if (field_width > 0 && next - str > field_width) {
next             3357 lib/vsprintf.c 			while (next - str > field_width) {
next             3362 lib/vsprintf.c 				--next;
next             3403 lib/vsprintf.c 		if (!next)
next             3405 lib/vsprintf.c 		str = next;
next              776 lib/xarray.c   	void *first, *next;
next              794 lib/xarray.c   	next = first;
next              814 lib/xarray.c   		if (xa_is_node(next) && (!node || node->shift))
next              815 lib/xarray.c   			xas_free_nodes(xas, xa_to_node(next));
next              818 lib/xarray.c   		count += !next - !entry;
next              829 lib/xarray.c   		next = xa_entry_locked(xas->xa, node, ++offset);
next              830 lib/xarray.c   		if (!xa_is_sibling(next)) {
next              833 lib/xarray.c   			first = next;
next             1664 lib/xarray.c   		struct xa_limit limit, u32 *next, gfp_t gfp)
next             1669 lib/xarray.c   	limit.min = max(min, *next);
next             1684 lib/xarray.c   		*next = *id + 1;
next             1685 lib/xarray.c   		if (*next == 0)
next             2022 lib/xarray.c   		node->array, node->private_list.prev, node->private_list.next);
next              179 lib/zlib_deflate/deflate.c     char *next;
next              208 lib/zlib_deflate/deflate.c     next = (char *) mem;
next              209 lib/zlib_deflate/deflate.c     next += sizeof(*mem);
next              210 lib/zlib_deflate/deflate.c     mem->window_memory = (Byte *) next;
next              211 lib/zlib_deflate/deflate.c     next += zlib_deflate_window_memsize(windowBits);
next              212 lib/zlib_deflate/deflate.c     mem->prev_memory = (Pos *) next;
next              213 lib/zlib_deflate/deflate.c     next += zlib_deflate_prev_memsize(windowBits);
next              214 lib/zlib_deflate/deflate.c     mem->head_memory = (Pos *) next;
next              215 lib/zlib_deflate/deflate.c     next += zlib_deflate_head_memsize(memLevel);
next              216 lib/zlib_deflate/deflate.c     mem->overlay_memory = next;
next               38 lib/zlib_inflate/inflate.c     state->lencode = state->distcode = state->next = state->codes;
next              170 lib/zlib_inflate/inflate.c         next = strm->next_in; \
next              181 lib/zlib_inflate/inflate.c         strm->next_in = next; \
next              200 lib/zlib_inflate/inflate.c         hold += (unsigned long)(*next++) << bits; \
next              320 lib/zlib_inflate/inflate.c     const unsigned char *next;  /* next input */
next              441 lib/zlib_inflate/inflate.c                 memcpy(put, next, copy);
next              443 lib/zlib_inflate/inflate.c                 next += copy;
next              477 lib/zlib_inflate/inflate.c             state->next = state->codes;
next              478 lib/zlib_inflate/inflate.c             state->lencode = (code const *)(state->next);
next              480 lib/zlib_inflate/inflate.c             ret = zlib_inflate_table(CODES, state->lens, 19, &(state->next),
next              543 lib/zlib_inflate/inflate.c             state->next = state->codes;
next              544 lib/zlib_inflate/inflate.c             state->lencode = (code const *)(state->next);
next              546 lib/zlib_inflate/inflate.c             ret = zlib_inflate_table(LENS, state->lens, state->nlen, &(state->next),
next              553 lib/zlib_inflate/inflate.c             state->distcode = (code const *)(state->next);
next              556 lib/zlib_inflate/inflate.c                             &(state->next), &(state->distbits), state->work);
next              106 lib/zlib_inflate/inflate.h     code *next;             /* next available space in codes[] */
next               40 lib/zlib_inflate/inftrees.c     code *next;             /* next available space in table */
next              190 lib/zlib_inflate/inftrees.c     next = *table;              /* current table to fill in */
next              224 lib/zlib_inflate/inftrees.c             next[(huff >> drop) + fill] = this;
next              252 lib/zlib_inflate/inftrees.c             next += min;            /* here min is 1 << curr */
next              273 lib/zlib_inflate/inftrees.c             (*table)[low].val = (unsigned short)(next - *table);
next              292 lib/zlib_inflate/inftrees.c             next = *table;
next              297 lib/zlib_inflate/inftrees.c         next[huff >> drop] = this;
next              588 mm/backing-dev.c 	    blkcg_cgwb_list->next && memcg_cgwb_list->next) {
next              747 mm/backing-dev.c 	struct bdi_writeback *wb, *next;
next              750 mm/backing-dev.c 	list_for_each_entry_safe(wb, next, memcg_cgwb_list, memcg_node)
next              752 mm/backing-dev.c 	memcg_cgwb_list->next = NULL;	/* prevent new wb's */
next              764 mm/backing-dev.c 	struct bdi_writeback *wb, *next;
next              767 mm/backing-dev.c 	list_for_each_entry_safe(wb, next, &blkcg->cgwb_list, blkcg_node)
next              769 mm/backing-dev.c 	blkcg->cgwb_list.next = NULL;	/* prevent new wb's */
next               55 mm/compaction.c 	struct page *page, *next;
next               58 mm/compaction.c 	list_for_each_entry_safe(page, next, freelist, lru) {
next               72 mm/compaction.c 	struct page *page, *next;
next               75 mm/compaction.c 	list_for_each_entry_safe(page, next, list, lru) {
next             1562 mm/compaction.c 	freepage = list_entry(cc->freepages.next, struct page, lru);
next               69 mm/dmapool.c   	char *next;
next               73 mm/dmapool.c   	next = buf;
next               76 mm/dmapool.c   	temp = scnprintf(next, size, "poolinfo - 0.1\n");
next               78 mm/dmapool.c   	next += temp;
next               93 mm/dmapool.c   		temp = scnprintf(next, size, "%-16s %4u %4zu %4zu %2u\n",
next               98 mm/dmapool.c   		next += temp;
next              211 mm/dmapool.c   		unsigned int next = offset + pool->size;
next              212 mm/dmapool.c   		if (unlikely((next + pool->size) >= next_boundary)) {
next              213 mm/dmapool.c   			next = next_boundary;
next              216 mm/dmapool.c   		*(int *)(page->vaddr + offset) = next;
next              217 mm/dmapool.c   		offset = next;
next              288 mm/dmapool.c   		page = list_entry(pool->page_list.next,
next             2617 mm/filemap.c   			goto next;
next             2624 mm/filemap.c   			goto next;
next             2626 mm/filemap.c   			goto next;
next             2657 mm/filemap.c   		goto next;
next             2662 mm/filemap.c   next:
next               33 mm/frontswap.c 	for ((ops) = frontswap_ops; (ops); (ops) = (ops)->next)
next              140 mm/frontswap.c 		ops->next = frontswap_ops;
next              141 mm/frontswap.c 	} while (cmpxchg(&frontswap_ops, ops->next, ops) != ops->next);
next             2047 mm/gup.c       	unsigned long next;
next             2051 mm/gup.c       		next = hugepte_addr_end(addr, end, sz);
next             2054 mm/gup.c       	} while (ptep++, addr = next, addr != end);
next             2190 mm/gup.c       	unsigned long next;
next             2197 mm/gup.c       		next = pmd_addr_end(addr, end);
next             2211 mm/gup.c       			if (!gup_huge_pmd(pmd, pmdp, addr, next, flags,
next             2221 mm/gup.c       					 PMD_SHIFT, next, flags, pages, nr))
next             2223 mm/gup.c       		} else if (!gup_pte_range(pmd, addr, next, flags, pages, nr))
next             2225 mm/gup.c       	} while (pmdp++, addr = next, addr != end);
next             2233 mm/gup.c       	unsigned long next;
next             2240 mm/gup.c       		next = pud_addr_end(addr, end);
next             2244 mm/gup.c       			if (!gup_huge_pud(pud, pudp, addr, next, flags,
next             2249 mm/gup.c       					 PUD_SHIFT, next, flags, pages, nr))
next             2251 mm/gup.c       		} else if (!gup_pmd_range(pud, addr, next, flags, pages, nr))
next             2253 mm/gup.c       	} while (pudp++, addr = next, addr != end);
next             2261 mm/gup.c       	unsigned long next;
next             2268 mm/gup.c       		next = p4d_addr_end(addr, end);
next             2274 mm/gup.c       					 P4D_SHIFT, next, flags, pages, nr))
next             2276 mm/gup.c       		} else if (!gup_pud_range(p4d, addr, next, flags, pages, nr))
next             2278 mm/gup.c       	} while (p4dp++, addr = next, addr != end);
next             2286 mm/gup.c       	unsigned long next;
next             2293 mm/gup.c       		next = pgd_addr_end(addr, end);
next             2297 mm/gup.c       			if (!gup_huge_pgd(pgd, pgdp, addr, next, flags,
next             2302 mm/gup.c       					 PGDIR_SHIFT, next, flags, pages, nr))
next             2304 mm/gup.c       		} else if (!gup_p4d_range(pgd, addr, next, flags, pages, nr))
next             2306 mm/gup.c       	} while (pgdp++, addr = next, addr != end);
next               26 mm/gup_benchmark.c 	unsigned long i, nr_pages, addr, next;
next               42 mm/gup_benchmark.c 	for (addr = gup->addr; addr < gup->addr + gup->size; addr = next) {
next               46 mm/gup_benchmark.c 		next = addr + nr * PAGE_SIZE;
next               47 mm/gup_benchmark.c 		if (next > gup->addr + gup->size) {
next               48 mm/gup_benchmark.c 			next = gup->addr + gup->size;
next               49 mm/gup_benchmark.c 			nr = (next - addr) / PAGE_SIZE;
next              661 mm/hmm.c       	unsigned long addr = start, next;
next              713 mm/hmm.c       		next = pmd_addr_end(addr, end);
next              714 mm/hmm.c       		ret = hmm_vma_walk_pmd(pmdp, addr, next, walk);
next              717 mm/hmm.c       	} while (pmdp++, addr = next, addr != end);
next             1695 mm/huge_memory.c 		pmd_t *pmd, unsigned long addr, unsigned long next)
next             1734 mm/huge_memory.c 	if (next - addr != HPAGE_PMD_SIZE) {
next             2400 mm/huge_memory.c 		struct vm_area_struct *next = vma->vm_next;
next             2401 mm/huge_memory.c 		unsigned long nstart = next->vm_start;
next             2404 mm/huge_memory.c 		    (nstart & HPAGE_PMD_MASK) >= next->vm_start &&
next             2405 mm/huge_memory.c 		    (nstart & HPAGE_PMD_MASK) + HPAGE_PMD_SIZE <= next->vm_end)
next             2406 mm/huge_memory.c 			split_huge_pmd_address(next, nstart, false, NULL);
next             2907 mm/huge_memory.c 	LIST_HEAD(list), *pos, *next;
next             2918 mm/huge_memory.c 	list_for_each_safe(pos, next, &ds_queue->split_queue) {
next             2933 mm/huge_memory.c 	list_for_each_safe(pos, next, &list) {
next             2936 mm/huge_memory.c 			goto next;
next             2941 mm/huge_memory.c next:
next             2988 mm/huge_memory.c 				goto next;
next             2991 mm/huge_memory.c 				goto next;
next             2998 mm/huge_memory.c next:
next             1344 mm/hugetlb.c   		node = node->next;
next             1573 mm/hugetlb.c   				list_entry(h->hugepage_freelists[node].next,
next             2370 mm/hugetlb.c   		struct page *page, *next;
next             2372 mm/hugetlb.c   		list_for_each_entry_safe(page, next, freel, lru) {
next              117 mm/kasan/init.c 	unsigned long next;
next              120 mm/kasan/init.c 		next = pmd_addr_end(addr, end);
next              140 mm/kasan/init.c 		zero_pte_populate(pmd, addr, next);
next              141 mm/kasan/init.c 	} while (pmd++, addr = next, addr != end);
next              150 mm/kasan/init.c 	unsigned long next;
next              153 mm/kasan/init.c 		next = pud_addr_end(addr, end);
next              177 mm/kasan/init.c 		zero_pmd_populate(pud, addr, next);
next              178 mm/kasan/init.c 	} while (pud++, addr = next, addr != end);
next              187 mm/kasan/init.c 	unsigned long next;
next              190 mm/kasan/init.c 		next = p4d_addr_end(addr, end);
next              218 mm/kasan/init.c 		zero_pud_populate(p4d, addr, next);
next              219 mm/kasan/init.c 	} while (p4d++, addr = next, addr != end);
next              236 mm/kasan/init.c 	unsigned long next;
next              239 mm/kasan/init.c 		next = pgd_addr_end(addr, end);
next              291 mm/kasan/init.c 		zero_p4d_populate(pgd, addr, next);
next              292 mm/kasan/init.c 	} while (pgd++, addr = next, addr != end);
next              360 mm/kasan/init.c 	unsigned long next;
next              362 mm/kasan/init.c 	for (; addr < end; addr = next, pte++) {
next              363 mm/kasan/init.c 		next = (addr + PAGE_SIZE) & PAGE_MASK;
next              364 mm/kasan/init.c 		if (next > end)
next              365 mm/kasan/init.c 			next = end;
next              379 mm/kasan/init.c 	unsigned long next;
next              381 mm/kasan/init.c 	for (; addr < end; addr = next, pmd++) {
next              384 mm/kasan/init.c 		next = pmd_addr_end(addr, end);
next              391 mm/kasan/init.c 			    IS_ALIGNED(next, PMD_SIZE))
next              396 mm/kasan/init.c 		kasan_remove_pte_table(pte, addr, next);
next              404 mm/kasan/init.c 	unsigned long next;
next              406 mm/kasan/init.c 	for (; addr < end; addr = next, pud++) {
next              409 mm/kasan/init.c 		next = pud_addr_end(addr, end);
next              416 mm/kasan/init.c 			    IS_ALIGNED(next, PUD_SIZE))
next              422 mm/kasan/init.c 		kasan_remove_pmd_table(pmd, addr, next);
next              430 mm/kasan/init.c 	unsigned long next;
next              432 mm/kasan/init.c 	for (; addr < end; addr = next, p4d++) {
next              435 mm/kasan/init.c 		next = p4d_addr_end(addr, end);
next              442 mm/kasan/init.c 			    IS_ALIGNED(next, P4D_SIZE))
next              447 mm/kasan/init.c 		kasan_remove_pud_table(pud, addr, next);
next              454 mm/kasan/init.c 	unsigned long addr, end, next;
next              465 mm/kasan/init.c 	for (; addr < end; addr = next) {
next              468 mm/kasan/init.c 		next = pgd_addr_end(addr, end);
next              476 mm/kasan/init.c 			    IS_ALIGNED(next, PGDIR_SIZE))
next              482 mm/kasan/init.c 		kasan_remove_p4d_table(p4d, addr, next);
next              114 mm/kasan/kasan.h 	struct qlist_node *next;
next               67 mm/kasan/quarantine.c 		q->tail->next = qlink;
next               69 mm/kasan/quarantine.c 	qlink->next = NULL;
next               84 mm/kasan/quarantine.c 	to->tail->next = from->head;
next              165 mm/kasan/quarantine.c 		struct qlist_node *next = qlink->next;
next              168 mm/kasan/quarantine.c 		qlink = next;
next              276 mm/kasan/quarantine.c 		struct qlist_node *next = curr->next;
next              284 mm/kasan/quarantine.c 		curr = next;
next             1919 mm/khugepaged.c 		mm_slot = list_entry(khugepaged_scan.mm_head.next,
next             2015 mm/khugepaged.c 		if (mm_slot->mm_node.next != &khugepaged_scan.mm_head) {
next             2017 mm/khugepaged.c 				mm_slot->mm_node.next,
next             1303 mm/kmemleak.c  	void *next;
next             1306 mm/kmemleak.c  		next = min(start + MAX_SCAN_SIZE, end);
next             1307 mm/kmemleak.c  		scan_block(start, next, NULL);
next             1308 mm/kmemleak.c  		start = next;
next             1337 mm/kmemleak.c  		void *next;
next             1340 mm/kmemleak.c  			next = min(start + MAX_SCAN_SIZE, end);
next             1341 mm/kmemleak.c  			scan_block(start, next, object);
next             1343 mm/kmemleak.c  			start = next;
next             1373 mm/kmemleak.c  	object = list_entry(gray_list.next, typeof(*object), gray_list);
next             1381 mm/kmemleak.c  		tmp = list_entry(object->gray_list.next, typeof(*object),
next             1693 mm/kmemleak.c  	.next  = kmemleak_seq_next,
next              639 mm/ksm.c       		if (rmap_item->hlist.next)
next              940 mm/ksm.c       	struct stable_node *stable_node, *next;
next              956 mm/ksm.c       	list_for_each_entry_safe(stable_node, next, &migrate_nodes, list) {
next              972 mm/ksm.c       	ksm_scan.mm_slot = list_entry(ksm_mm_head.mm_list.next,
next              995 mm/ksm.c       		ksm_scan.mm_slot = list_entry(mm_slot->mm_list.next,
next             1422 mm/ksm.c       			BUG_ON(stable_node->hlist.first->next);
next             2020 mm/ksm.c       	if (rmap_item->hlist.next)
next             2256 mm/ksm.c       			struct stable_node *stable_node, *next;
next             2259 mm/ksm.c       			list_for_each_entry_safe(stable_node, next,
next             2273 mm/ksm.c       		slot = list_entry(slot->mm_list.next, struct mm_slot, mm_list);
next             2342 mm/ksm.c       	ksm_scan.mm_slot = list_entry(slot->mm_list.next,
next             2768 mm/ksm.c       	struct stable_node *stable_node, *next;
next             2786 mm/ksm.c       	list_for_each_entry_safe(stable_node, next, &migrate_nodes, list) {
next              320 mm/madvise.c   		unsigned long next = pmd_addr_end(addr, end);
next              343 mm/madvise.c   		if (next - addr != HPAGE_PMD_SIZE) {
next              573 mm/madvise.c   	unsigned long next;
next              575 mm/madvise.c   	next = pmd_addr_end(addr, end);
next              577 mm/madvise.c   		if (madvise_free_huge_pmd(tlb, vma, pmd, addr, next))
next              578 mm/madvise.c   			goto next;
next              691 mm/madvise.c   next:
next              518 mm/memblock.c  		struct memblock_region *next = &type->regions[i + 1];
next              520 mm/memblock.c  		if (this->base + this->size != next->base ||
next              522 mm/memblock.c  		    memblock_get_region_node(next) ||
next              523 mm/memblock.c  		    this->flags != next->flags) {
next              524 mm/memblock.c  			BUG_ON(this->base + this->size > next->base);
next              529 mm/memblock.c  		this->size += next->size;
next              531 mm/memblock.c  		memmove(next, next + 1, (type->cnt - (i + 2)) * sizeof(*next));
next              880 mm/memcontrol.c 	unsigned long val, next;
next              883 mm/memcontrol.c 	next = __this_cpu_read(memcg->vmstats_percpu->targets[target]);
next              885 mm/memcontrol.c 	if ((long)(next - val) < 0) {
next              888 mm/memcontrol.c 			next = val + THRESHOLDS_EVENTS_TARGET;
next              891 mm/memcontrol.c 			next = val + SOFTLIMIT_EVENTS_TARGET;
next              894 mm/memcontrol.c 			next = val + NUMAINFO_EVENTS_TARGET;
next              899 mm/memcontrol.c 		__this_cpu_write(memcg->vmstats_percpu->targets[target], next);
next             6745 mm/memcontrol.c 	struct list_head *next;
next             6753 mm/memcontrol.c 	next = page_list->next;
next             6757 mm/memcontrol.c 		page = list_entry(next, struct page, lru);
next             6758 mm/memcontrol.c 		next = page->lru.next;
next             6761 mm/memcontrol.c 	} while (next != page_list);
next              364 mm/memory-failure.c 	struct to_kill *tk, *next;
next              366 mm/memory-failure.c 	list_for_each_entry_safe (tk, next, to_kill, nd) {
next              209 mm/memory.c    	unsigned long next;
next              215 mm/memory.c    		next = pmd_addr_end(addr, end);
next              219 mm/memory.c    	} while (pmd++, addr = next, addr != end);
next              243 mm/memory.c    	unsigned long next;
next              249 mm/memory.c    		next = pud_addr_end(addr, end);
next              252 mm/memory.c    		free_pmd_range(tlb, pud, addr, next, floor, ceiling);
next              253 mm/memory.c    	} while (pud++, addr = next, addr != end);
next              277 mm/memory.c    	unsigned long next;
next              283 mm/memory.c    		next = p4d_addr_end(addr, end);
next              286 mm/memory.c    		free_pud_range(tlb, p4d, addr, next, floor, ceiling);
next              287 mm/memory.c    	} while (p4d++, addr = next, addr != end);
next              313 mm/memory.c    	unsigned long next;
next              363 mm/memory.c    		next = pgd_addr_end(addr, end);
next              366 mm/memory.c    		free_p4d_range(tlb, pgd, addr, next, floor, ceiling);
next              367 mm/memory.c    	} while (pgd++, addr = next, addr != end);
next              374 mm/memory.c    		struct vm_area_struct *next = vma->vm_next;
next              386 mm/memory.c    				floor, next ? next->vm_start : ceiling);
next              391 mm/memory.c    			while (next && next->vm_start <= vma->vm_end + PMD_SIZE
next              392 mm/memory.c    			       && !is_vm_hugetlb_page(next)) {
next              393 mm/memory.c    				vma = next;
next              394 mm/memory.c    				next = vma->vm_next;
next              399 mm/memory.c    				floor, next ? next->vm_start : ceiling);
next              401 mm/memory.c    		vma = next;
next              852 mm/memory.c    	unsigned long next;
next              859 mm/memory.c    		next = pmd_addr_end(addr, end);
next              863 mm/memory.c    			VM_BUG_ON_VMA(next-addr != HPAGE_PMD_SIZE, vma);
next              875 mm/memory.c    						vma, addr, next))
next              877 mm/memory.c    	} while (dst_pmd++, src_pmd++, addr = next, addr != end);
next              886 mm/memory.c    	unsigned long next;
next              893 mm/memory.c    		next = pud_addr_end(addr, end);
next              897 mm/memory.c    			VM_BUG_ON_VMA(next-addr != HPAGE_PUD_SIZE, vma);
next              909 mm/memory.c    						vma, addr, next))
next              911 mm/memory.c    	} while (dst_pud++, src_pud++, addr = next, addr != end);
next              920 mm/memory.c    	unsigned long next;
next              927 mm/memory.c    		next = p4d_addr_end(addr, end);
next              931 mm/memory.c    						vma, addr, next))
next              933 mm/memory.c    	} while (dst_p4d++, src_p4d++, addr = next, addr != end);
next              941 mm/memory.c    	unsigned long next;
next              989 mm/memory.c    		next = pgd_addr_end(addr, end);
next              993 mm/memory.c    					    vma, addr, next))) {
next              997 mm/memory.c    	} while (dst_pgd++, src_pgd++, addr = next, addr != end);
next             1145 mm/memory.c    	unsigned long next;
next             1149 mm/memory.c    		next = pmd_addr_end(addr, end);
next             1151 mm/memory.c    			if (next - addr != HPAGE_PMD_SIZE)
next             1154 mm/memory.c    				goto next;
next             1165 mm/memory.c    			goto next;
next             1166 mm/memory.c    		next = zap_pte_range(tlb, vma, pmd, addr, next, details);
next             1167 mm/memory.c    next:
next             1169 mm/memory.c    	} while (pmd++, addr = next, addr != end);
next             1180 mm/memory.c    	unsigned long next;
next             1184 mm/memory.c    		next = pud_addr_end(addr, end);
next             1186 mm/memory.c    			if (next - addr != HPAGE_PUD_SIZE) {
next             1190 mm/memory.c    				goto next;
next             1195 mm/memory.c    		next = zap_pmd_range(tlb, vma, pud, addr, next, details);
next             1196 mm/memory.c    next:
next             1198 mm/memory.c    	} while (pud++, addr = next, addr != end);
next             1209 mm/memory.c    	unsigned long next;
next             1213 mm/memory.c    		next = p4d_addr_end(addr, end);
next             1216 mm/memory.c    		next = zap_pud_range(tlb, vma, p4d, addr, next, details);
next             1217 mm/memory.c    	} while (p4d++, addr = next, addr != end);
next             1228 mm/memory.c    	unsigned long next;
next             1234 mm/memory.c    		next = pgd_addr_end(addr, end);
next             1237 mm/memory.c    		next = zap_p4d_range(tlb, vma, pgd, addr, next, details);
next             1238 mm/memory.c    	} while (pgd++, addr = next, addr != end);
next             1822 mm/memory.c    	unsigned long next;
next             1831 mm/memory.c    		next = pmd_addr_end(addr, end);
next             1832 mm/memory.c    		err = remap_pte_range(mm, pmd, addr, next,
next             1836 mm/memory.c    	} while (pmd++, addr = next, addr != end);
next             1845 mm/memory.c    	unsigned long next;
next             1853 mm/memory.c    		next = pud_addr_end(addr, end);
next             1854 mm/memory.c    		err = remap_pmd_range(mm, pud, addr, next,
next             1858 mm/memory.c    	} while (pud++, addr = next, addr != end);
next             1867 mm/memory.c    	unsigned long next;
next             1875 mm/memory.c    		next = p4d_addr_end(addr, end);
next             1876 mm/memory.c    		err = remap_pud_range(mm, p4d, addr, next,
next             1880 mm/memory.c    	} while (p4d++, addr = next, addr != end);
next             1900 mm/memory.c    	unsigned long next;
next             1941 mm/memory.c    		next = pgd_addr_end(addr, end);
next             1942 mm/memory.c    		err = remap_p4d_range(mm, pgd, addr, next,
next             1946 mm/memory.c    	} while (pgd++, addr = next, addr != end);
next             2040 mm/memory.c    	unsigned long next;
next             2049 mm/memory.c    		next = pmd_addr_end(addr, end);
next             2050 mm/memory.c    		err = apply_to_pte_range(mm, pmd, addr, next, fn, data);
next             2053 mm/memory.c    	} while (pmd++, addr = next, addr != end);
next             2062 mm/memory.c    	unsigned long next;
next             2069 mm/memory.c    		next = pud_addr_end(addr, end);
next             2070 mm/memory.c    		err = apply_to_pmd_range(mm, pud, addr, next, fn, data);
next             2073 mm/memory.c    	} while (pud++, addr = next, addr != end);
next             2082 mm/memory.c    	unsigned long next;
next             2089 mm/memory.c    		next = p4d_addr_end(addr, end);
next             2090 mm/memory.c    		err = apply_to_pud_range(mm, p4d, addr, next, fn, data);
next             2093 mm/memory.c    	} while (p4d++, addr = next, addr != end);
next             2105 mm/memory.c    	unsigned long next;
next             2114 mm/memory.c    		next = pgd_addr_end(addr, end);
next             2115 mm/memory.c    		err = apply_to_p4d_range(mm, pgd, addr, next, fn, data);
next             2118 mm/memory.c    	} while (pgd++, addr = next, addr != end);
next              734 mm/mempolicy.c 	struct vm_area_struct *next;
next              750 mm/mempolicy.c 	for (; vma && vma->vm_start < end; prev = vma, vma = next) {
next              751 mm/mempolicy.c 		next = vma->vm_next;
next              765 mm/mempolicy.c 			next = vma->vm_next;
next             1828 mm/mempolicy.c 	unsigned next;
next             1831 mm/mempolicy.c 	next = next_node_in(me->il_prev, policy->v.nodes);
next             1832 mm/mempolicy.c 	if (next < MAX_NUMNODES)
next             1833 mm/mempolicy.c 		me->il_prev = next;
next             1834 mm/mempolicy.c 	return next;
next             2518 mm/mempolicy.c 		struct rb_node *next = rb_next(&n->nd);
next             2541 mm/mempolicy.c 		if (!next)
next             2543 mm/mempolicy.c 		n = rb_entry(next, struct sp_node, nd);
next             2647 mm/mempolicy.c 	struct rb_node *next;
next             2652 mm/mempolicy.c 	next = rb_first(&p->root);
next             2653 mm/mempolicy.c 	while (next) {
next             2654 mm/mempolicy.c 		n = rb_entry(next, struct sp_node, nd);
next             2655 mm/mempolicy.c 		next = rb_next(&n->nd);
next             2257 mm/migrate.c   			goto next;
next             2270 mm/migrate.c   				goto next;
next             2282 mm/migrate.c   				goto next;
next             2292 mm/migrate.c   			goto next;
next             2338 mm/migrate.c   next:
next              500 mm/mlock.c     				goto next;
next              505 mm/mlock.c     next:
next              173 mm/mmap.c      	struct vm_area_struct *next = vma->vm_next;
next              182 mm/mmap.c      	return next;
next              192 mm/mmap.c      	struct vm_area_struct *next;
next              260 mm/mmap.c      	next = find_vma(mm, oldbrk);
next              261 mm/mmap.c      	if (next && newbrk + PAGE_SIZE > vm_start_gap(next))
next              683 mm/mmap.c      	struct vm_area_struct *next;
next              686 mm/mmap.c      	next = vma->vm_next;
next              688 mm/mmap.c      		prev->vm_next = next;
next              692 mm/mmap.c      			prev->vm_next = next;
next              694 mm/mmap.c      			mm->mmap = next;
next              696 mm/mmap.c      	if (next)
next              697 mm/mmap.c      		next->vm_prev = prev;
next              722 mm/mmap.c      	struct vm_area_struct *next = vma->vm_next, *orig_vma = vma;
next              731 mm/mmap.c      	if (next && !insert) {
next              734 mm/mmap.c      		if (end >= next->vm_end) {
next              741 mm/mmap.c      			if (next == expand) {
next              746 mm/mmap.c      				VM_WARN_ON(end != next->vm_end);
next              753 mm/mmap.c      				VM_WARN_ON(file != next->vm_file);
next              754 mm/mmap.c      				swap(vma, next);
next              761 mm/mmap.c      				remove_next = 1 + (end > next->vm_end);
next              763 mm/mmap.c      					   end != next->vm_next->vm_end);
next              765 mm/mmap.c      					   end != next->vm_end);
next              767 mm/mmap.c      				end = next->vm_end;
next              770 mm/mmap.c      			exporter = next;
next              777 mm/mmap.c      			if (remove_next == 2 && !next->anon_vma)
next              778 mm/mmap.c      				exporter = next->vm_next;
next              780 mm/mmap.c      		} else if (end > next->vm_start) {
next              785 mm/mmap.c      			adjust_next = (end - next->vm_start) >> PAGE_SHIFT;
next              786 mm/mmap.c      			exporter = next;
next              797 mm/mmap.c      			importer = next;
next              824 mm/mmap.c      			uprobe_munmap(next, next->vm_start, next->vm_end);
next              840 mm/mmap.c      		anon_vma = next->anon_vma;
next              842 mm/mmap.c      		VM_WARN_ON(adjust_next && next->anon_vma &&
next              843 mm/mmap.c      			   anon_vma != next->anon_vma);
next              847 mm/mmap.c      			anon_vma_interval_tree_pre_update_vma(next);
next              854 mm/mmap.c      			vma_interval_tree_remove(next, root);
next              867 mm/mmap.c      		next->vm_start += adjust_next << PAGE_SHIFT;
next              868 mm/mmap.c      		next->vm_pgoff += adjust_next;
next              873 mm/mmap.c      			vma_interval_tree_insert(next, root);
next              884 mm/mmap.c      			__vma_unlink_prev(mm, next, vma);
next              895 mm/mmap.c      			__vma_unlink_common(mm, next, NULL, false, vma);
next              897 mm/mmap.c      			__remove_shared_vm_struct(next, file, mapping);
next              909 mm/mmap.c      			if (!next)
next              912 mm/mmap.c      				vma_gap_update(next);
next              919 mm/mmap.c      			anon_vma_interval_tree_post_update_vma(next);
next              929 mm/mmap.c      			uprobe_mmap(next);
next              934 mm/mmap.c      			uprobe_munmap(next, next->vm_start, next->vm_end);
next              937 mm/mmap.c      		if (next->anon_vma)
next              938 mm/mmap.c      			anon_vma_merge(vma, next);
next              940 mm/mmap.c      		mpol_put(vma_policy(next));
next              941 mm/mmap.c      		vm_area_free(next);
next              954 mm/mmap.c      			next = vma->vm_next;
next              966 mm/mmap.c      			next = vma;
next              970 mm/mmap.c      			end = next->vm_end;
next              973 mm/mmap.c      		else if (next)
next              974 mm/mmap.c      			vma_gap_update(next);
next             1143 mm/mmap.c      	struct vm_area_struct *area, *next;
next             1154 mm/mmap.c      		next = prev->vm_next;
next             1156 mm/mmap.c      		next = mm->mmap;
next             1157 mm/mmap.c      	area = next;
next             1159 mm/mmap.c      		next = next->vm_next;
next             1177 mm/mmap.c      		if (next && end == next->vm_start &&
next             1178 mm/mmap.c      				mpol_equal(policy, vma_policy(next)) &&
next             1179 mm/mmap.c      				can_vma_merge_before(next, vm_flags,
next             1184 mm/mmap.c      						      next->anon_vma, NULL)) {
next             1187 mm/mmap.c      					 next->vm_end, prev->vm_pgoff, NULL,
next             1201 mm/mmap.c      	if (next && end == next->vm_start &&
next             1202 mm/mmap.c      			mpol_equal(policy, vma_policy(next)) &&
next             1203 mm/mmap.c      			can_vma_merge_before(next, vm_flags,
next             1208 mm/mmap.c      					 addr, prev->vm_pgoff, NULL, next);
next             1210 mm/mmap.c      			err = __vma_adjust(area, addr, next->vm_end,
next             1211 mm/mmap.c      					 next->vm_pgoff - pglen, NULL, next);
next             1217 mm/mmap.c      			area = next;
next             2342 mm/mmap.c      	struct vm_area_struct *next;
next             2362 mm/mmap.c      	next = vma->vm_next;
next             2363 mm/mmap.c      	if (next && next->vm_start < gap_addr &&
next             2364 mm/mmap.c      			(next->vm_flags & (VM_WRITE|VM_READ|VM_EXEC))) {
next             2365 mm/mmap.c      		if (!(next->vm_flags & VM_GROWSUP))
next             2609 mm/mmap.c      	struct vm_area_struct *next = prev ? prev->vm_next : mm->mmap;
next             2617 mm/mmap.c      				 next ? next->vm_start : USER_PGTABLES_CEILING);
next             2928 mm/mmap.c      		struct vm_area_struct *next;
next             2930 mm/mmap.c      		for (next = vma->vm_next; next; next = next->vm_next) {
next             2932 mm/mmap.c      			if (next->vm_start != next->vm_prev->vm_end)
next             2935 mm/mmap.c      			if (next->vm_file != vma->vm_file)
next             2938 mm/mmap.c      			if (next->vm_flags != vma->vm_flags)
next             2941 mm/mmap.c      			if (start + size <= next->vm_end)
next             2945 mm/mmap.c      		if (!next)
next               21 mm/mmu_gather.c 	if (batch->next) {
next               22 mm/mmu_gather.c 		tlb->active = batch->next;
next               34 mm/mmu_gather.c 	batch->next = NULL;
next               38 mm/mmu_gather.c 	tlb->active->next = batch;
next               48 mm/mmu_gather.c 	for (batch = &tlb->local; batch && batch->nr; batch = batch->next) {
next               57 mm/mmu_gather.c 	struct mmu_gather_batch *batch, *next;
next               59 mm/mmu_gather.c 	for (batch = tlb->local.next; batch; batch = next) {
next               60 mm/mmu_gather.c 		next = batch->next;
next               63 mm/mmu_gather.c 	tlb->local.next = NULL;
next              216 mm/mmu_gather.c 	tlb->local.next = NULL;
next              194 mm/mprotect.c  	unsigned long next;
next              205 mm/mprotect.c  		next = pmd_addr_end(addr, end);
next              217 mm/mprotect.c  			goto next;
next              228 mm/mprotect.c  			if (next - addr != HPAGE_PMD_SIZE) {
next              241 mm/mprotect.c  					goto next;
next              246 mm/mprotect.c  		this_pages = change_pte_range(vma, pmd, addr, next, newprot,
next              249 mm/mprotect.c  next:
next              251 mm/mprotect.c  	} while (pmd++, addr = next, addr != end);
next              266 mm/mprotect.c  	unsigned long next;
next              271 mm/mprotect.c  		next = pud_addr_end(addr, end);
next              274 mm/mprotect.c  		pages += change_pmd_range(vma, pud, addr, next, newprot,
next              276 mm/mprotect.c  	} while (pud++, addr = next, addr != end);
next              286 mm/mprotect.c  	unsigned long next;
next              291 mm/mprotect.c  		next = p4d_addr_end(addr, end);
next              294 mm/mprotect.c  		pages += change_pud_range(vma, p4d, addr, next, newprot,
next              296 mm/mprotect.c  	} while (p4d++, addr = next, addr != end);
next              307 mm/mprotect.c  	unsigned long next;
next              316 mm/mprotect.c  		next = pgd_addr_end(addr, end);
next              319 mm/mprotect.c  		pages += change_p4d_range(vma, pgd, addr, next, newprot,
next              321 mm/mprotect.c  	} while (pgd++, addr = next, addr != end);
next              346 mm/mprotect.c  			       unsigned long next, struct mm_walk *walk)
next              353 mm/mprotect.c  				   unsigned long addr, unsigned long next,
next              360 mm/mprotect.c  static int prot_none_test(unsigned long addr, unsigned long next,
next              245 mm/mremap.c    	unsigned long extent, next, old_end;
next              258 mm/mremap.c    		next = (old_addr + PMD_SIZE) & PMD_MASK;
next              260 mm/mremap.c    		extent = next - old_addr;
next              306 mm/mremap.c    		next = (new_addr + PMD_SIZE) & PMD_MASK;
next              307 mm/mremap.c    		if (extent > next - new_addr)
next              308 mm/mremap.c    			extent = next - new_addr;
next              315 mm/oom_kill.c  		goto next;
next              319 mm/oom_kill.c  		goto next;
next              329 mm/oom_kill.c  			goto next;
next              344 mm/oom_kill.c  		goto next;
next              352 mm/oom_kill.c  next:
next             3082 mm/page_alloc.c 	struct page *page, *next;
next             3087 mm/page_alloc.c 	list_for_each_entry_safe(page, next, list, lru) {
next             3095 mm/page_alloc.c 	list_for_each_entry_safe(page, next, list, lru) {
next             6952 mm/page_alloc.c 	phys_addr_t next = 0;
next             6960 mm/page_alloc.c 		if (next < start)
next             6961 mm/page_alloc.c 			pgcnt += zero_pfn_range(PFN_DOWN(next), PFN_UP(start));
next             6962 mm/page_alloc.c 		next = end;
next             6972 mm/page_alloc.c 	pgcnt += zero_pfn_range(PFN_DOWN(next),
next               33 mm/pagewalk.c  	unsigned long next;
next               40 mm/pagewalk.c  		next = pmd_addr_end(addr, end);
next               43 mm/pagewalk.c  				err = ops->pte_hole(addr, next, walk);
next               53 mm/pagewalk.c  			err = ops->pmd_entry(pmd, addr, next, walk);
next               67 mm/pagewalk.c  		err = walk_pte_range(pmd, addr, next, walk);
next               70 mm/pagewalk.c  	} while (pmd++, addr = next, addr != end);
next               79 mm/pagewalk.c  	unsigned long next;
next               86 mm/pagewalk.c  		next = pud_addr_end(addr, end);
next               89 mm/pagewalk.c  				err = ops->pte_hole(addr, next, walk);
next               99 mm/pagewalk.c  				err = ops->pud_entry(pud, addr, next, walk);
next              112 mm/pagewalk.c  			err = walk_pmd_range(pud, addr, next, walk);
next              115 mm/pagewalk.c  	} while (pud++, addr = next, addr != end);
next              124 mm/pagewalk.c  	unsigned long next;
next              130 mm/pagewalk.c  		next = p4d_addr_end(addr, end);
next              133 mm/pagewalk.c  				err = ops->pte_hole(addr, next, walk);
next              139 mm/pagewalk.c  			err = walk_pud_range(p4d, addr, next, walk);
next              142 mm/pagewalk.c  	} while (p4d++, addr = next, addr != end);
next              151 mm/pagewalk.c  	unsigned long next;
next              157 mm/pagewalk.c  		next = pgd_addr_end(addr, end);
next              160 mm/pagewalk.c  				err = ops->pte_hole(addr, next, walk);
next              166 mm/pagewalk.c  			err = walk_p4d_range(pgd, addr, next, walk);
next              169 mm/pagewalk.c  	} while (pgd++, addr = next, addr != end);
next              187 mm/pagewalk.c  	unsigned long next;
next              195 mm/pagewalk.c  		next = hugetlb_entry_end(h, addr, end);
next              199 mm/pagewalk.c  			err = ops->hugetlb_entry(pte, hmask, addr, next, walk);
next              201 mm/pagewalk.c  			err = ops->pte_hole(addr, next, walk);
next              205 mm/pagewalk.c  	} while (addr = next, addr != end);
next              306 mm/pagewalk.c  	unsigned long next;
next              326 mm/pagewalk.c  			next = end;
next              329 mm/pagewalk.c  			next = min(end, vma->vm_start);
next              332 mm/pagewalk.c  			next = min(end, vma->vm_end);
next              335 mm/pagewalk.c  			err = walk_page_test(start, next, &walk);
next              349 mm/pagewalk.c  			err = __walk_page_range(start, next, &walk);
next              352 mm/pagewalk.c  	} while (start = next, start < end);
next             1594 mm/percpu.c    	struct pcpu_chunk *chunk, *next;
next             1656 mm/percpu.c    		list_for_each_entry_safe(chunk, next, &pcpu_slot[slot], list) {
next             1838 mm/percpu.c    	struct pcpu_chunk *chunk, *next;
next             1848 mm/percpu.c    	list_for_each_entry_safe(chunk, next, free_head, list) {
next             1860 mm/percpu.c    	list_for_each_entry_safe(chunk, next, &to_free, list) {
next              379 mm/rmap.c      	struct anon_vma_chain *avc, *next;
next              386 mm/rmap.c      	list_for_each_entry_safe(avc, next, &vma->anon_vma_chain, same_vma) {
next              413 mm/rmap.c      	list_for_each_entry_safe(avc, next, &vma->anon_vma_chain, same_vma) {
next              106 mm/shmem.c     	pgoff_t next;		/* the next page offset to be fallocated */
next              463 mm/shmem.c     	LIST_HEAD(list), *pos, *next;
next              475 mm/shmem.c     	list_for_each_safe(pos, next, &sbinfo->shrinklist) {
next              485 mm/shmem.c     			goto next;
next              493 mm/shmem.c     			goto next;
next              497 mm/shmem.c     next:
next              503 mm/shmem.c     	list_for_each_safe(pos, next, &to_remove) {
next              510 mm/shmem.c     	list_for_each_safe(pos, next, &list) {
next              633 mm/shmem.c     next:
next              637 mm/shmem.c     			goto next;
next             1251 mm/shmem.c     	struct shmem_inode_info *info, *next;
next             1258 mm/shmem.c     	list_for_each_entry_safe(info, next, &shmem_swaplist, swaplist) {
next             1277 mm/shmem.c     		next = list_next_entry(info, swaplist);
next             1343 mm/shmem.c     			    index < shmem_falloc->next)
next             2024 mm/shmem.c     		    vmf->pgoff < shmem_falloc->next) {
next             2754 mm/shmem.c     		shmem_falloc.next = (unmap_end + 1) >> PAGE_SHIFT;
next             2794 mm/shmem.c     	shmem_falloc.next  = start;
next             2828 mm/shmem.c     		shmem_falloc.next++;
next             2754 mm/slab.c      	void *next = *list;
next             2757 mm/slab.c      	while (next) {
next             2758 mm/slab.c      		objp = next - obj_offset(cachep);
next             2759 mm/slab.c      		next = *(void **)next;
next             3780 mm/slab.c      	if (!cachep->list.next) {
next             4009 mm/slab.c      			goto next;
next             4024 mm/slab.c      next:
next             1477 mm/slab_common.c 	if (p == slab_root_caches.next)
next             1544 mm/slab_common.c 	if (p == memcg->kmem_caches.next)
next             1566 mm/slab_common.c 	.next = slab_next,
next              146 mm/slob.c      static void set_slob(slob_t *s, slobidx_t size, slob_t *next)
next              149 mm/slob.c      	slobidx_t offset = next - base;
next              174 mm/slob.c      	slobidx_t next;
next              177 mm/slob.c      		next = -s[0].units;
next              179 mm/slob.c      		next = s[1].units;
next              180 mm/slob.c      	return base+next;
next              261 mm/slob.c      			slob_t *next;
next              264 mm/slob.c      				next = slob_next(cur);
next              265 mm/slob.c      				set_slob(aligned, avail - delta, next);
next              272 mm/slob.c      			next = slob_next(cur);
next              275 mm/slob.c      					set_slob(prev, slob_units(prev), next);
next              277 mm/slob.c      					sp->freelist = next;
next              283 mm/slob.c      				set_slob(cur + units, avail - units, next);
next              385 mm/slob.c      	slob_t *prev, *next, *b = (slob_t *)block;
next              442 mm/slob.c      		next = slob_next(prev);
next              443 mm/slob.c      		while (b > next) {
next              444 mm/slob.c      			prev = next;
next              445 mm/slob.c      			next = slob_next(prev);
next              448 mm/slob.c      		if (!slob_last(prev) && b + units == next) {
next              449 mm/slob.c      			units += slob_units(next);
next              450 mm/slob.c      			set_slob(b, units, slob_next(next));
next              452 mm/slob.c      			set_slob(b, units, next);
next             1432 mm/slub.c      	void *next = *head;
next             1441 mm/slub.c      		object = next;
next             1442 mm/slub.c      		next = get_freepointer(s, object);
next             1575 mm/slub.c      	void *next;
next             1594 mm/slub.c      		next = next_freelist_entry(s, page, &pos, start, page_limit,
next             1596 mm/slub.c      		next = setup_object(s, page, next);
next             1597 mm/slub.c      		set_freepointer(s, cur, next);
next             1598 mm/slub.c      		cur = next;
next             1621 mm/slub.c      	void *start, *p, *next;
next             1674 mm/slub.c      			next = p + s->size;
next             1675 mm/slub.c      			next = setup_object(s, page, next);
next             1676 mm/slub.c      			set_freepointer(s, p, next);
next             1677 mm/slub.c      			p = next;
next             2192 mm/slub.c      		c->partial = page->next;
next             2220 mm/slub.c      			page->next = discard_page;
next             2233 mm/slub.c      		discard_page = discard_page->next;
next             2286 mm/slub.c      		page->next = oldpage;
next             5844 mm/slub.c      	struct saved_alias *next;
next             5867 mm/slub.c      	al->next = alias_list;
next             5898 mm/slub.c      		alias_list = alias_list->next;
next              612 mm/swapfile.c  			cluster->next = cluster_next(&cluster->index) *
next              632 mm/swapfile.c  	tmp = cluster->next;
next              652 mm/swapfile.c  	cluster->next = tmp + 1;
next              995 mm/swapfile.c  	struct swap_info_struct *si, *next;
next             1019 mm/swapfile.c  	plist_for_each_entry_safe(si, next, &swap_avail_heads[node], avail_lists[node]) {
next             1065 mm/swapfile.c  		if (plist_node_empty(&next->avail_lists[node]))
next             1983 mm/swapfile.c  	unsigned long next;
next             1989 mm/swapfile.c  		next = pmd_addr_end(addr, end);
next             1992 mm/swapfile.c  		ret = unuse_pte_range(vma, pmd, addr, next, type,
next             1996 mm/swapfile.c  	} while (pmd++, addr = next, addr != end);
next             2006 mm/swapfile.c  	unsigned long next;
next             2011 mm/swapfile.c  		next = pud_addr_end(addr, end);
next             2014 mm/swapfile.c  		ret = unuse_pmd_range(vma, pud, addr, next, type,
next             2018 mm/swapfile.c  	} while (pud++, addr = next, addr != end);
next             2028 mm/swapfile.c  	unsigned long next;
next             2033 mm/swapfile.c  		next = p4d_addr_end(addr, end);
next             2036 mm/swapfile.c  		ret = unuse_pud_range(vma, p4d, addr, next, type,
next             2040 mm/swapfile.c  	} while (p4d++, addr = next, addr != end);
next             2048 mm/swapfile.c  	unsigned long addr, end, next;
next             2056 mm/swapfile.c  		next = pgd_addr_end(addr, end);
next             2059 mm/swapfile.c  		ret = unuse_p4d_range(vma, pgd, addr, next, type,
next             2063 mm/swapfile.c  	} while (pgd++, addr = next, addr != end);
next             2153 mm/swapfile.c  	       (p = p->next) != &init_mm.mmlist) {
next             2239 mm/swapfile.c  	struct list_head *p, *next;
next             2246 mm/swapfile.c  	list_for_each_safe(p, next, &init_mm.mmlist)
next             2780 mm/swapfile.c  	.next =		swap_next,
next             3649 mm/swapfile.c  	page = list_entry(head->lru.next, struct page, lru);
next             3661 mm/swapfile.c  			page = list_entry(page->lru.next, struct page, lru);
next             3667 mm/swapfile.c  			page = list_entry(page->lru.next, struct page, lru);
next             3693 mm/swapfile.c  			page = list_entry(page->lru.next, struct page, lru);
next             3729 mm/swapfile.c  			struct page *page, *next;
next             3731 mm/swapfile.c  			list_for_each_entry_safe(page, next, &head->lru, lru) {
next             3743 mm/swapfile.c  	struct swap_info_struct *si, *next;
next             3758 mm/swapfile.c  	plist_for_each_entry_safe(si, next, &swap_avail_heads[node],
next              276 mm/util.c      	struct vm_area_struct *next;
next              280 mm/util.c      		next = prev->vm_next;
next              285 mm/util.c      			next = rb_entry(rb_parent,
next              288 mm/util.c      			next = NULL;
next              290 mm/util.c      	vma->vm_next = next;
next              291 mm/util.c      	if (next)
next              292 mm/util.c      		next->vm_prev = vma;
next               78 mm/vmalloc.c   	unsigned long next;
next               82 mm/vmalloc.c   		next = pmd_addr_end(addr, end);
next               87 mm/vmalloc.c   		vunmap_pte_range(pmd, addr, next);
next               88 mm/vmalloc.c   	} while (pmd++, addr = next, addr != end);
next               94 mm/vmalloc.c   	unsigned long next;
next               98 mm/vmalloc.c   		next = pud_addr_end(addr, end);
next              103 mm/vmalloc.c   		vunmap_pmd_range(pud, addr, next);
next              104 mm/vmalloc.c   	} while (pud++, addr = next, addr != end);
next              110 mm/vmalloc.c   	unsigned long next;
next              114 mm/vmalloc.c   		next = p4d_addr_end(addr, end);
next              119 mm/vmalloc.c   		vunmap_pud_range(p4d, addr, next);
next              120 mm/vmalloc.c   	} while (p4d++, addr = next, addr != end);
next              126 mm/vmalloc.c   	unsigned long next;
next              131 mm/vmalloc.c   		next = pgd_addr_end(addr, end);
next              134 mm/vmalloc.c   		vunmap_p4d_range(pgd, addr, next);
next              135 mm/vmalloc.c   	} while (pgd++, addr = next, addr != end);
next              168 mm/vmalloc.c   	unsigned long next;
next              174 mm/vmalloc.c   		next = pmd_addr_end(addr, end);
next              175 mm/vmalloc.c   		if (vmap_pte_range(pmd, addr, next, prot, pages, nr))
next              177 mm/vmalloc.c   	} while (pmd++, addr = next, addr != end);
next              185 mm/vmalloc.c   	unsigned long next;
next              191 mm/vmalloc.c   		next = pud_addr_end(addr, end);
next              192 mm/vmalloc.c   		if (vmap_pmd_range(pud, addr, next, prot, pages, nr))
next              194 mm/vmalloc.c   	} while (pud++, addr = next, addr != end);
next              202 mm/vmalloc.c   	unsigned long next;
next              208 mm/vmalloc.c   		next = p4d_addr_end(addr, end);
next              209 mm/vmalloc.c   		if (vmap_pud_range(p4d, addr, next, prot, pages, nr))
next              211 mm/vmalloc.c   	} while (p4d++, addr = next, addr != end);
next              225 mm/vmalloc.c   	unsigned long next;
next              233 mm/vmalloc.c   		next = pgd_addr_end(addr, end);
next              234 mm/vmalloc.c   		err = vmap_p4d_range(pgd, addr, next, prot, pages, &nr);
next              237 mm/vmalloc.c   	} while (pgd++, addr = next, addr != end);
next              497 mm/vmalloc.c   	return (&parent->rb_right == link ? list->next : list);
next              691 mm/vmalloc.c   	struct list_head *next;
next              705 mm/vmalloc.c   	next = get_va_next_sibling(parent, link);
next              706 mm/vmalloc.c   	if (unlikely(next == NULL))
next              716 mm/vmalloc.c   	if (next != head) {
next              717 mm/vmalloc.c   		sibling = list_entry(next, struct vmap_area, list);
next              740 mm/vmalloc.c   	if (next->prev != head) {
next              741 mm/vmalloc.c   		sibling = list_entry(next->prev, struct vmap_area, list);
next             1824 mm/vmalloc.c   	for (p = &vmlist; (tmp = *p) != NULL; p = &tmp->next) {
next             1831 mm/vmalloc.c   	vm->next = *p;
next             1925 mm/vmalloc.c   	for (tmp = vmlist; tmp; tmp = tmp->next) {
next             3558 mm/vmalloc.c   	.next = s_next,
next             1546 mm/vmscan.c    	struct page *page, *next;
next             1549 mm/vmscan.c    	list_for_each_entry_safe(page, next, page_list, lru) {
next             1533 mm/vmstat.c    	.next	= frag_next,
next             1540 mm/vmstat.c    	.next	= frag_next,
next             1656 mm/vmstat.c    	.next	= frag_next,
next             1742 mm/vmstat.c    	.next	= vmstat_next,
next             2065 mm/vmstat.c    	.next	= frag_next,
next             2117 mm/vmstat.c    	.next	= frag_next,
next             1201 mm/z3fold.c    				goto next;
next             1206 mm/z3fold.c    				goto next;
next             1211 mm/z3fold.c    				goto next;
next             1213 mm/z3fold.c    next:
next              536 mm/zbud.c      				goto next;
next              541 mm/zbud.c      				goto next;
next              543 mm/zbud.c      next:
next              244 mm/zsmalloc.c  		unsigned long next;
next              938 mm/zsmalloc.c  	struct page *page, *next;
next              949 mm/zsmalloc.c  	next = page = get_first_page(zspage);
next              952 mm/zsmalloc.c  		next = get_next_page(page);
next              957 mm/zsmalloc.c  		page = next;
next             1000 mm/zsmalloc.c  			link->next = freeobj++ << OBJ_TAG_BITS;
next             1011 mm/zsmalloc.c  			link->next = freeobj++ << OBJ_TAG_BITS;
next             1017 mm/zsmalloc.c  			link->next = -1UL << OBJ_TAG_BITS;
next             1431 mm/zsmalloc.c  	set_freeobj(zspage, link->next >> OBJ_TAG_BITS);
next             1533 mm/zsmalloc.c  	link->next = get_freeobj(zspage) << OBJ_TAG_BITS;
next              382 net/802/garp.c 	struct rb_node *node, *next;
next              386 net/802/garp.c 	     next = node ? rb_next(node) : NULL, node != NULL;
next              387 net/802/garp.c 	     node = next) {
next              567 net/802/mrp.c  	struct rb_node *node, *next;
next              571 net/802/mrp.c  	     next = node ? rb_next(node) : NULL, node != NULL;
next              572 net/802/mrp.c  	     node = next) {
next              189 net/8021q/vlan_dev.c 		mp = mp->next;
next              198 net/8021q/vlan_dev.c 	np->next = mp;
next              598 net/8021q/vlan_dev.c 			vlan->egress_priority_map[i] = pm->next;
next              256 net/8021q/vlan_netlink.c 			     pm = pm->next) {
next               68 net/8021q/vlanproc.c 	.next = vlan_seq_next,
next              285 net/8021q/vlanproc.c 			mp = mp->next;
next              458 net/9p/trans_fd.c 		req = list_entry(m->unsent_req_list.next, struct p9_req_t,
next              860 net/9p/trans_fd.c 		 m, m->mux_list.prev, m->mux_list.next);
next              401 net/9p/trans_rdma.c 	wr.next = NULL;
next              493 net/9p/trans_rdma.c 	wr.next = NULL;
next               67 net/appletalk/aarp.c 	struct aarp_entry	*next;
next              260 net/appletalk/aarp.c 			*n = (*n)->next;
next              263 net/appletalk/aarp.c 			n = &((*n)->next);
next              279 net/appletalk/aarp.c 			*n = (*n)->next;
next              283 net/appletalk/aarp.c 			n = &((*n)->next);
next              300 net/appletalk/aarp.c 			*n = (*n)->next;
next              303 net/appletalk/aarp.c 			n = &((*n)->next);
next              357 net/appletalk/aarp.c 		*n = (*n)->next;
next              404 net/appletalk/aarp.c 		list = list->next;
next              513 net/appletalk/aarp.c 	entry->next = proxies[hash];
next              641 net/appletalk/aarp.c 	a->next		 = unresolved[hash];
next              693 net/appletalk/aarp.c 			*list = a->next;
next              696 net/appletalk/aarp.c 			a->next = resolved[hash];
next              706 net/appletalk/aarp.c 			list = &((*list)->next);
next              928 net/appletalk/aarp.c 		for (entry = table[ct]; entry; entry = entry->next) {
next              975 net/appletalk/aarp.c 	else if (entry->next)
next              976 net/appletalk/aarp.c 		entry = entry->next;
next             1036 net/appletalk/aarp.c 	.next   = aarp_seq_next,
next               21 net/appletalk/atalk_proc.c 	for (i = atalk_interfaces; pos && i; i = i->next)
next               48 net/appletalk/atalk_proc.c 	i = i->next;
next               82 net/appletalk/atalk_proc.c 	for (r = atalk_routes; pos && r; r = r->next)
next              109 net/appletalk/atalk_proc.c 	r = r->next;
next              191 net/appletalk/atalk_proc.c 	.next   = atalk_seq_interface_next,
next              198 net/appletalk/atalk_proc.c 	.next   = atalk_seq_route_next,
next              205 net/appletalk/atalk_proc.c 	.next   = atalk_seq_socket_next,
next              209 net/appletalk/ddp.c 			*iface = tmp->next;
next              214 net/appletalk/ddp.c 			iface = &tmp->next;
next              234 net/appletalk/ddp.c 	iface->next = atalk_interfaces;
next              351 net/appletalk/ddp.c 	for (iface = atalk_interfaces; iface; iface = iface->next) {
next              399 net/appletalk/ddp.c 	for (iface = atalk_interfaces; iface; iface = iface->next) {
next              434 net/appletalk/ddp.c 	for (r = atalk_routes; r; r = r->next) {
next              515 net/appletalk/ddp.c 	for (rt = atalk_routes; rt; rt = rt->next) {
next              531 net/appletalk/ddp.c 		for (iface = atalk_interfaces; iface; iface = iface->next) {
next              559 net/appletalk/ddp.c 		rt->next = atalk_routes;
next              589 net/appletalk/ddp.c 			*r = tmp->next;
next              594 net/appletalk/ddp.c 		r = &tmp->next;
next              614 net/appletalk/ddp.c 			*r = tmp->next;
next              618 net/appletalk/ddp.c 			r = &tmp->next;
next              288 net/atm/br2684.c 	return list_empty(&brdev->brvccs) ? NULL : list_entry_brvcc(brdev->brvccs.next);	/* 1 vcc/dev right now */
next              376 net/atm/br2684.c 		    brdev->brvccs.next != brdev->brvccs.prev)	/* >1 VCC */
next              379 net/atm/br2684.c 			brvcc = list_entry_brvcc(brdev->brvccs.next);
next              816 net/atm/br2684.c 	.next = br2684_seq_next,
next              852 net/atm/br2684.c 		net_dev = list_entry_brdev(br2684_devs.next);
next              855 net/atm/br2684.c 			brvcc = list_entry_brvcc(brdev->brvccs.next);
next               81 net/atm/clip.c 	clip_vcc->next = entry->vccs;
next               97 net/atm/clip.c 	for (walk = &entry->vccs; *walk; walk = &(*walk)->next)
next              101 net/atm/clip.c 			*walk = clip_vcc->next;	/* atomic */
next              128 net/atm/clip.c 	for (cv = entry->vccs; cv; cv = cv->next) {
next              517 net/atm/clip.c 		for (dev = clip_devs; dev; dev = PRIV(dev)->next)
next              522 net/atm/clip.c 		for (dev = clip_devs; dev; dev = PRIV(dev)->next)
next              539 net/atm/clip.c 	clip_priv->next = clip_devs;
next              801 net/atm/clip.c 	curr = curr->next;
next              862 net/atm/clip.c 	.next	= neigh_seq_next,
next              897 net/atm/clip.c 	struct net_device *dev, *next;
next              911 net/atm/clip.c 		next = PRIV(dev)->next;
next              914 net/atm/clip.c 		dev = next;
next              502 net/atm/common.c 			dev = list_entry(atm_devs.next,
next              844 net/atm/lec.c  	for (; e; e = e->next) {
next              981 net/atm/lec.c  							  next);
next              991 net/atm/lec.c  	.next = lec_seq_next,
next             1288 net/atm/lec.c  	hlist_add_head(&entry->next, tmp);
next             1305 net/atm/lec.c  	hlist_del(&to_remove->next);
next             1318 net/atm/lec.c  					     &priv->lec_arp_tables[i], next) {
next             1362 net/atm/lec.c  				     &priv->lec_arp_tables[i], next) {
next             1393 net/atm/lec.c  	hlist_for_each_entry(rulla, &priv->lec_no_forward, next) {
next             1418 net/atm/lec.c  	hlist_for_each_entry(rulla, &priv->lec_arp_empty_ones, next) {
next             1443 net/atm/lec.c  	hlist_for_each_entry(rulla, &priv->mcast_fwds, next) {
next             1477 net/atm/lec.c  	struct hlist_node *next;
next             1489 net/atm/lec.c  		hlist_for_each_entry_safe(entry, next,
next             1490 net/atm/lec.c  					  &priv->lec_arp_tables[i], next) {
next             1497 net/atm/lec.c  	hlist_for_each_entry_safe(entry, next,
next             1498 net/atm/lec.c  				  &priv->lec_arp_empty_ones, next) {
next             1501 net/atm/lec.c  		hlist_del(&entry->next);
next             1506 net/atm/lec.c  	hlist_for_each_entry_safe(entry, next,
next             1507 net/atm/lec.c  				  &priv->lec_no_forward, next) {
next             1510 net/atm/lec.c  		hlist_del(&entry->next);
next             1515 net/atm/lec.c  	hlist_for_each_entry_safe(entry, next, &priv->mcast_fwds, next) {
next             1518 net/atm/lec.c  		hlist_del(&entry->next);
next             1538 net/atm/lec.c  	hlist_for_each_entry(entry, head, next) {
next             1556 net/atm/lec.c  	INIT_HLIST_NODE(&to_return->next);
next             1602 net/atm/lec.c  	hlist_del(&to_remove->next);
next             1674 net/atm/lec.c  	struct hlist_node *next;
next             1684 net/atm/lec.c  		hlist_for_each_entry_safe(entry, next,
next             1685 net/atm/lec.c  					  &priv->lec_arp_tables[i], next) {
next             1811 net/atm/lec.c  	struct hlist_node *next;
next             1818 net/atm/lec.c  		hlist_for_each_entry_safe(entry, next,
next             1819 net/atm/lec.c  					  &priv->lec_arp_tables[i], next) {
next             1843 net/atm/lec.c  	struct hlist_node *next;
next             1858 net/atm/lec.c  		hlist_for_each_entry_safe(entry, next,
next             1859 net/atm/lec.c  					  &priv->lec_arp_empty_ones, next) {
next             1861 net/atm/lec.c  				hlist_del(&entry->next);
next             1905 net/atm/lec.c  				     &priv->lec_arp_tables[i], next) {
next             1969 net/atm/lec.c  		hlist_add_head(&entry->next, &priv->mcast_fwds);
next             1997 net/atm/lec.c  		hlist_add_head(&entry->next, &priv->lec_no_forward);
next             2015 net/atm/lec.c  				     &priv->lec_arp_tables[i], next) {
next             2078 net/atm/lec.c  	hlist_add_head(&entry->next, &priv->lec_arp_empty_ones);
next             2099 net/atm/lec.c  				     &priv->lec_arp_tables[i], next) {
next             2133 net/atm/lec.c  				     &priv->lec_arp_tables[i], next) {
next             2184 net/atm/lec.c  	struct hlist_node *next;
next             2194 net/atm/lec.c  		hlist_for_each_entry_safe(entry, next,
next             2195 net/atm/lec.c  					  &priv->lec_arp_tables[i], next) {
next             2205 net/atm/lec.c  	hlist_for_each_entry_safe(entry, next,
next             2206 net/atm/lec.c  				  &priv->lec_arp_empty_ones, next) {
next             2210 net/atm/lec.c  			hlist_del(&entry->next);
next             2215 net/atm/lec.c  	hlist_for_each_entry_safe(entry, next,
next             2216 net/atm/lec.c  				  &priv->lec_no_forward, next) {
next             2220 net/atm/lec.c  			hlist_del(&entry->next);
next             2225 net/atm/lec.c  	hlist_for_each_entry_safe(entry, next, &priv->mcast_fwds, next) {
next             2229 net/atm/lec.c  			hlist_del(&entry->next);
next             2243 net/atm/lec.c  	struct hlist_node *next;
next             2249 net/atm/lec.c  	hlist_for_each_entry_safe(entry, next,
next             2250 net/atm/lec.c  				  &priv->lec_arp_empty_ones, next) {
next             2262 net/atm/lec.c  			hlist_del(&entry->next);
next               15 net/atm/lec_arpc.h 	struct hlist_node next;		/* Linked entry list */
next              136 net/atm/mpc.c  		mpc = mpc->next;
next              150 net/atm/mpc.c  		mpc = mpc->next;
next              164 net/atm/mpc.c  		mpc = mpc->next;
next              196 net/atm/mpc.c  	entry->next = qos_head;
next              210 net/atm/mpc.c  		qos = qos->next;
next              226 net/atm/mpc.c  		qos_head = qos_head->next;
next              233 net/atm/mpc.c  		if (curr->next == entry) {
next              234 net/atm/mpc.c  			curr->next = entry->next;
next              238 net/atm/mpc.c  		curr = curr->next;
next              266 net/atm/mpc.c  		qos = qos->next;
next              290 net/atm/mpc.c  	mpc->next = mpcs;
next             1300 net/atm/mpc.c  		entry = entry->next;
next             1403 net/atm/mpc.c  		entry = entry->next;
next             1439 net/atm/mpc.c  		mpc = mpc->next;
next             1503 net/atm/mpc.c  		tmp = mpc->next;
next             1525 net/atm/mpc.c  		nextqos = qos->next;
next               16 net/atm/mpc.h  	struct mpoa_client *next;
next               42 net/atm/mpc.h  	struct atm_mpoa_qos *next;
next               48 net/atm/mpoa_caches.c 		entry = entry->next;
next               69 net/atm/mpoa_caches.c 		entry = entry->next;
next               90 net/atm/mpoa_caches.c 		entry = entry->next;
next              112 net/atm/mpoa_caches.c 	entry->next = client->in_cache;
next              201 net/atm/mpoa_caches.c 		entry->prev->next = entry->next;
next              203 net/atm/mpoa_caches.c 		client->in_cache = entry->next;
next              204 net/atm/mpoa_caches.c 	if (entry->next != NULL)
next              205 net/atm/mpoa_caches.c 		entry->next->prev = entry->prev;
next              237 net/atm/mpoa_caches.c 		next_entry = entry->next;
next              266 net/atm/mpoa_caches.c 				entry = entry->next;	/* Entry in hold down */
next              278 net/atm/mpoa_caches.c 					entry = entry->next;
next              293 net/atm/mpoa_caches.c 		entry = entry->next;
next              319 net/atm/mpoa_caches.c 		entry = entry->next;
next              345 net/atm/mpoa_caches.c 		entry = entry->next;
next              366 net/atm/mpoa_caches.c 		entry = entry->next;
next              388 net/atm/mpoa_caches.c 		entry = entry->next;
next              408 net/atm/mpoa_caches.c 		entry = entry->next;
next              434 net/atm/mpoa_caches.c 		entry->prev->next = entry->next;
next              436 net/atm/mpoa_caches.c 		client->eg_cache = entry->next;
next              437 net/atm/mpoa_caches.c 	if (entry->next != NULL)
next              438 net/atm/mpoa_caches.c 		entry->next->prev = entry->prev;
next              472 net/atm/mpoa_caches.c 	entry->next = client->eg_cache;
next              511 net/atm/mpoa_caches.c 		next_entry = entry->next;
next               18 net/atm/mpoa_caches.h 	struct in_cache_entry *next;
next               55 net/atm/mpoa_caches.h 	struct               eg_cache_entry *next;
next              114 net/atm/mpoa_proc.c 	for (mpc = mpcs; mpc; mpc = mpc->next)
next              124 net/atm/mpoa_proc.c 	return v == SEQ_START_TOKEN ? mpcs : p->next;
next              152 net/atm/mpoa_proc.c 	for (in_entry = mpc->in_cache; in_entry; in_entry = in_entry->next) {
next              171 net/atm/mpoa_proc.c 	for (eg_entry = mpc->eg_cache; eg_entry; eg_entry = eg_entry->next) {
next              199 net/atm/mpoa_proc.c 	.next =		mpc_next,
next              243 net/atm/proc.c 	.next	= atm_dev_seq_next,
next              266 net/atm/proc.c 	.next	= vcc_seq_next,
next              288 net/atm/proc.c 	.next	= vcc_seq_next,
next              311 net/atm/proc.c 	.next	= vcc_seq_next,
next             1923 net/ax25/af_ax25.c 	.next = ax25_info_next,
next               37 net/ax25/ax25_dev.c 	for (ax25_dev = ax25_dev_list; ax25_dev != NULL; ax25_dev = ax25_dev->next)
next               84 net/ax25/ax25_dev.c 	ax25_dev->next = ax25_dev_list;
next              109 net/ax25/ax25_dev.c 	for (s = ax25_dev_list; s != NULL; s = s->next)
next              114 net/ax25/ax25_dev.c 		ax25_dev_list = s->next;
next              122 net/ax25/ax25_dev.c 	while (s != NULL && s->next != NULL) {
next              123 net/ax25/ax25_dev.c 		if (s->next == ax25_dev) {
next              124 net/ax25/ax25_dev.c 			s->next = ax25_dev->next;
next              132 net/ax25/ax25_dev.c 		s = s->next;
next              192 net/ax25/ax25_dev.c 		ax25_dev = ax25_dev->next;
next               35 net/ax25/ax25_iface.c 	struct listen_struct *next;
next               48 net/ax25/ax25_iface.c 	ap->next = protocol_list;
next               65 net/ax25/ax25_iface.c 		protocol_list = protocol->next;
next               69 net/ax25/ax25_iface.c 	while (protocol != NULL && protocol->next != NULL) {
next               70 net/ax25/ax25_iface.c 		if (protocol->next->pid == pid) {
next               71 net/ax25/ax25_iface.c 			protocol->next = protocol->next->next;
next               75 net/ax25/ax25_iface.c 		protocol = protocol->next;
next              115 net/ax25/ax25_iface.c 	listen->next = listen_list;
next              136 net/ax25/ax25_iface.c 		listen_list = listen->next;
next              142 net/ax25/ax25_iface.c 	while (listen != NULL && listen->next != NULL) {
next              143 net/ax25/ax25_iface.c 		if (ax25cmp(&listen->next->callsign, callsign) == 0 && listen->next->dev == dev) {
next              144 net/ax25/ax25_iface.c 			s = listen->next;
next              145 net/ax25/ax25_iface.c 			listen->next = listen->next->next;
next              151 net/ax25/ax25_iface.c 		listen = listen->next;
next              164 net/ax25/ax25_iface.c 	for (protocol = protocol_list; protocol != NULL; protocol = protocol->next)
next              179 net/ax25/ax25_iface.c 	for (listen = listen_list; listen != NULL; listen = listen->next)
next              206 net/ax25/ax25_iface.c 	for (protocol = protocol_list; protocol != NULL; protocol = protocol->next)
next              245 net/ax25/ax25_out.c 	unsigned short start, end, next;
next              287 net/ax25/ax25_out.c 		next = (ax25->vs + 1) % ax25->modulus;
next              288 net/ax25/ax25_out.c 		last = (next == end);
next              308 net/ax25/ax25_out.c 		ax25->vs = next;
next               50 net/ax25/ax25_route.c 		ax25_rt = ax25_rt->next;
next               54 net/ax25/ax25_route.c 				ax25_route_list = s->next;
next               58 net/ax25/ax25_route.c 				for (t = ax25_route_list; t != NULL; t = t->next) {
next               59 net/ax25/ax25_route.c 					if (t->next == s) {
next               60 net/ax25/ax25_route.c 						t->next = s->next;
next              106 net/ax25/ax25_route.c 		ax25_rt = ax25_rt->next;
next              132 net/ax25/ax25_route.c 	ax25_rt->next   = ax25_route_list;
next              158 net/ax25/ax25_route.c 		ax25_rt = ax25_rt->next;
next              162 net/ax25/ax25_route.c 				ax25_route_list = s->next;
next              165 net/ax25/ax25_route.c 				for (t = ax25_route_list; t != NULL; t = t->next) {
next              166 net/ax25/ax25_route.c 					if (t->next == s) {
next              167 net/ax25/ax25_route.c 						t->next = s->next;
next              213 net/ax25/ax25_route.c 		ax25_rt = ax25_rt->next;
next              259 net/ax25/ax25_route.c 	for (ax25_rt = ax25_route_list; ax25_rt != NULL; ax25_rt = ax25_rt->next) {
next              272 net/ax25/ax25_route.c 		((struct ax25_route *) v)->next;
next              325 net/ax25/ax25_route.c 	.next = ax25_rt_seq_next,
next              347 net/ax25/ax25_route.c 	for (ax25_rt = ax25_route_list; ax25_rt != NULL; ax25_rt = ax25_rt->next) {
next              481 net/ax25/ax25_route.c 		ax25_rt = ax25_rt->next;
next              183 net/ax25/ax25_uid.c 	.next = ax25_uid_seq_next,
next             1847 net/batman-adv/bat_iv_ogm.c 				goto next;
next             1850 net/batman-adv/bat_iv_ogm.c 				goto next;
next             1868 net/batman-adv/bat_iv_ogm.c next:
next             2431 net/batman-adv/bat_iv_ogm.c 			goto next;
next             2434 net/batman-adv/bat_iv_ogm.c 			goto next;
next             2479 net/batman-adv/bat_iv_ogm.c next:
next              377 net/batman-adv/bat_v.c 				goto next;
next              396 net/batman-adv/bat_v.c next:
next              756 net/batman-adv/bat_v.c 			goto next;
next              759 net/batman-adv/bat_v.c 			goto next;
next              768 net/batman-adv/bat_v.c next:
next              458 net/batman-adv/multicast.c 	for (pmc6 = in6_dev->mc_list; pmc6; pmc6 = pmc6->next) {
next              588 net/batman-adv/routing.c 			goto next;
next              592 net/batman-adv/routing.c 			goto next;
next              601 net/batman-adv/routing.c 			goto next;
next              605 net/batman-adv/routing.c 			goto next;
next              627 net/batman-adv/routing.c next:
next              673 net/bluetooth/af_bluetooth.c 	.next  = bt_seq_next,
next             3768 net/bluetooth/hci_core.c 			skb = list; list = list->next;
next             2192 net/bluetooth/l2cap_core.c 		frag = &(*frag)->next;
next             2978 net/bluetooth/l2cap_core.c 		frag = &(*frag)->next;
next             5894 net/bluetooth/l2cap_core.c 	new_frag->next = NULL;
next             5896 net/bluetooth/l2cap_core.c 	(*last_frag)->next = new_frag;
next             7367 net/bluetooth/l2cap_core.c 		c = list_entry(chan_list.next, typeof(*c), global_l);
next             7423 net/bluetooth/l2cap_core.c 		struct l2cap_chan *chan, *next;
next             7427 net/bluetooth/l2cap_core.c 			goto next;
next             7441 net/bluetooth/l2cap_core.c next:
next             7442 net/bluetooth/l2cap_core.c 		next = l2cap_global_fixed_chan(pchan, hcon);
next             7444 net/bluetooth/l2cap_core.c 		pchan = next;
next              304 net/bridge/br_forward.c 			p = rcu_dereference(p->next);
next              164 net/bridge/br_mdb.c 		      pp = &p->next) {
next              331 net/bridge/br_mdb.c 	     pp = &p->next) {
next              634 net/bridge/br_mdb.c 	     pp = &p->next) {
next              760 net/bridge/br_mdb.c 	     pp = &p->next) {
next              768 net/bridge/br_mdb.c 		rcu_assign_pointer(*pp, p->next);
next              179 net/bridge/br_multicast.c 	     pp = &p->next) {
next              183 net/bridge/br_multicast.c 		rcu_assign_pointer(*pp, p->next);
next              476 net/bridge/br_multicast.c 			struct net_bridge_port_group __rcu *next,
next              489 net/bridge/br_multicast.c 	rcu_assign_pointer(p->next, next);
next              563 net/bridge/br_multicast.c 	     pp = &p->next) {
next             1280 net/bridge/br_multicast.c 	     pp = &p->next) {
next             1363 net/bridge/br_multicast.c 	     pp = &p->next) {
next             1403 net/bridge/br_multicast.c 		     pp = &p->next) {
next             1410 net/bridge/br_multicast.c 			rcu_assign_pointer(*pp, p->next);
next             1437 net/bridge/br_multicast.c 		     p = mlock_dereference(p->next, br)) {
next             1469 net/bridge/br_multicast.c 	     p = mlock_dereference(p->next, br)) {
next              206 net/bridge/br_private.h 	struct net_bridge_port_group __rcu *next;
next              688 net/bridge/br_private.h 			    struct net_bridge_port_group __rcu *next,
next             1464 net/can/bcm.c  	struct bcm_op *op, *next;
next             1478 net/can/bcm.c  	list_for_each_entry_safe(op, next, &bo->tx_ops, list)
next             1481 net/can/bcm.c  	list_for_each_entry_safe(op, next, &bo->rx_ops, list) {
next              152 net/core/datagram.c 	skb->prev->next = nskb;
next              153 net/core/datagram.c 	skb->next->prev = nskb;
next              155 net/core/datagram.c 	nskb->next = skb->next;
next              364 net/core/datagram.c 		if (skb->next) {
next             2729 net/core/dev.c 	skb->next = __this_cpu_read(softnet_data.completion_queue);
next             3213 net/core/dev.c 		struct sk_buff *next = skb->next;
next             3216 net/core/dev.c 		rc = xmit_one(skb, dev, txq, next != NULL);
next             3218 net/core/dev.c 			skb->next = next;
next             3222 net/core/dev.c 		skb = next;
next             3311 net/core/dev.c 	struct sk_buff *next, *head = NULL, *tail;
next             3313 net/core/dev.c 	for (; skb != NULL; skb = next) {
next             3314 net/core/dev.c 		next = skb->next;
next             3327 net/core/dev.c 			tail->next = skb;
next             4494 net/core/dev.c 			clist = clist->next;
next             4959 net/core/dev.c 	struct sk_buff *skb, *next;
next             4969 net/core/dev.c 		list_for_each_entry_safe(skb, next, head, list) {
next             4991 net/core/dev.c 	struct sk_buff *skb, *next;
next             4994 net/core/dev.c 	list_for_each_entry_safe(skb, next, head, list) {
next             5045 net/core/dev.c 	struct sk_buff *skb, *next;
next             5048 net/core/dev.c 	list_for_each_entry_safe(skb, next, head, list) {
next             5134 net/core/dev.c 	struct sk_buff *skb, *next;
next             5138 net/core/dev.c 	list_for_each_entry_safe(skb, next, head, list) {
next             5149 net/core/dev.c 		list_for_each_entry_safe(skb, next, head, list) {
next             5810 net/core/dev.c 		struct softnet_data *next = remsd->rps_ipi_next;
next             5814 net/core/dev.c 		remsd = next;
next             6570 net/core/dev.c 	upper = list_entry_rcu((*iter)->next, struct netdev_adjacent, list);
next             6587 net/core/dev.c 	upper = list_entry((*iter)->next, struct netdev_adjacent, list);
next             6605 net/core/dev.c 	upper = list_entry_rcu((*iter)->next, struct netdev_adjacent, list);
next             6620 net/core/dev.c 	struct net_device *udev, *next, *now, *dev_stack[MAX_NEST_DEV + 1];
next             6635 net/core/dev.c 		next = NULL;
next             6643 net/core/dev.c 			next = udev;
next             6650 net/core/dev.c 		if (!next) {
next             6653 net/core/dev.c 			next = dev_stack[--cur];
next             6657 net/core/dev.c 		now = next;
next             6669 net/core/dev.c 	struct net_device *udev, *next, *now, *dev_stack[MAX_NEST_DEV + 1];
next             6683 net/core/dev.c 		next = NULL;
next             6689 net/core/dev.c 			next = udev;
next             6696 net/core/dev.c 		if (!next) {
next             6699 net/core/dev.c 			next = dev_stack[--cur];
next             6703 net/core/dev.c 		now = next;
next             6741 net/core/dev.c 	*iter = lower->list.next;
next             6764 net/core/dev.c 	lower = list_entry_rcu((*iter)->next, struct netdev_adjacent, list);
next             6795 net/core/dev.c 	*iter = lower->list.next;
next             6806 net/core/dev.c 	lower = list_entry((*iter)->next, struct netdev_adjacent, list);
next             6822 net/core/dev.c 	lower = list_entry((*iter)->next, struct netdev_adjacent, list);
next             6838 net/core/dev.c 	struct net_device *ldev, *next, *now, *dev_stack[MAX_NEST_DEV + 1];
next             6852 net/core/dev.c 		next = NULL;
next             6858 net/core/dev.c 			next = ldev;
next             6865 net/core/dev.c 		if (!next) {
next             6868 net/core/dev.c 			next = dev_stack[--cur];
next             6872 net/core/dev.c 		now = next;
next             6885 net/core/dev.c 	struct net_device *ldev, *next, *now, *dev_stack[MAX_NEST_DEV + 1];
next             6900 net/core/dev.c 		next = NULL;
next             6908 net/core/dev.c 			next = ldev;
next             6915 net/core/dev.c 		if (!next) {
next             6918 net/core/dev.c 			next = dev_stack[--cur];
next             6922 net/core/dev.c 		now = next;
next             6934 net/core/dev.c 	lower = list_entry_rcu((*iter)->next, struct netdev_adjacent, list);
next             7001 net/core/dev.c 	struct net_device *ldev, *next, *now, *dev_stack[MAX_NEST_DEV + 1];
next             7015 net/core/dev.c 		next = NULL;
next             7021 net/core/dev.c 			next = ldev;
next             7028 net/core/dev.c 		if (!next) {
next             7031 net/core/dev.c 			next = dev_stack[--cur];
next             7035 net/core/dev.c 		now = next;
next             9839 net/core/dev.c 		list_skb = &(*list_skb)->next;
next               75 net/core/fib_rules.c 		pos = ops->rules_list.next;
next               76 net/core/fib_rules.c 		if (pos->next != &ops->rules_list) {
next               77 net/core/fib_rules.c 			rule = list_entry(pos->next, struct fib_rule, list);
next              186 net/core/neighbour.c 		neigh = rcu_dereference_protected(n->next,
next              216 net/core/neighbour.c 		np = &n->next;
next              302 net/core/neighbour.c 				np = &n->next;
next              306 net/core/neighbour.c 				np = &n->next;
next              310 net/core/neighbour.c 				   rcu_dereference_protected(n->next,
next              495 net/core/neighbour.c 		struct neighbour *n, *next;
next              500 net/core/neighbour.c 		     n = next) {
next              505 net/core/neighbour.c 			next = rcu_dereference_protected(n->next,
next              508 net/core/neighbour.c 			rcu_assign_pointer(n->next,
next              557 net/core/neighbour.c 	     n = rcu_dereference_bh(n->next)) {
next              634 net/core/neighbour.c 	     n1 = rcu_dereference_protected(n1->next,
next              650 net/core/neighbour.c 	rcu_assign_pointer(n->next,
next              696 net/core/neighbour.c 		n = n->next;
next              750 net/core/neighbour.c 	n->next = tbl->phash_buckets[hash_val];
next              768 net/core/neighbour.c 	     np = &n->next) {
next              771 net/core/neighbour.c 			*np = n->next;
next              795 net/core/neighbour.c 				*np = n->next;
next              796 net/core/neighbour.c 				n->next = freelist;
next              800 net/core/neighbour.c 			np = &n->next;
next              805 net/core/neighbour.c 		freelist = n->next;
next              806 net/core/neighbour.c 		n->next = NULL;
next              936 net/core/neighbour.c 				*np = n->next;
next              945 net/core/neighbour.c 			np = &n->next;
next             1018 net/core/neighbour.c 	unsigned long now, next;
next             1027 net/core/neighbour.c 	next = now + HZ;
next             1036 net/core/neighbour.c 			next = neigh->confirmed + neigh->parms->reachable_time;
next             1044 net/core/neighbour.c 			next = now + NEIGH_VAR(neigh->parms, DELAY_PROBE_TIME);
next             1061 net/core/neighbour.c 			next = neigh->confirmed + neigh->parms->reachable_time;
next             1068 net/core/neighbour.c 			next = now + NEIGH_VAR(neigh->parms, RETRANS_TIME);
next             1072 net/core/neighbour.c 		next = now + NEIGH_VAR(neigh->parms, RETRANS_TIME);
next             1084 net/core/neighbour.c 		if (time_before(next, jiffies + HZ/2))
next             1085 net/core/neighbour.c 			next = jiffies + HZ/2;
next             1086 net/core/neighbour.c 		if (!mod_timer(&neigh->timer, next))
next             1120 net/core/neighbour.c 			unsigned long next, now = jiffies;
next             1127 net/core/neighbour.c 			next = now + max(NEIGH_VAR(neigh->parms, RETRANS_TIME),
next             1129 net/core/neighbour.c 			neigh_add_timer(neigh, next);
next             2401 net/core/neighbour.c 				goto next;
next             2409 net/core/neighbour.c 		next:
next             2567 net/core/neighbour.c 		     n = rcu_dereference_bh(n->next)) {
next             2569 net/core/neighbour.c 				goto next;
next             2572 net/core/neighbour.c 				goto next;
next             2580 net/core/neighbour.c next:
next             2610 net/core/neighbour.c 		for (n = tbl->phash_buckets[h], idx = 0; n; n = n->next) {
next             2612 net/core/neighbour.c 				goto next;
next             2615 net/core/neighbour.c 				goto next;
next             2623 net/core/neighbour.c 		next:
next             2938 net/core/neighbour.c 		     n = rcu_dereference_bh(n->next))
next             2968 net/core/neighbour.c 					rcu_dereference_protected(n->next,
next             2972 net/core/neighbour.c 				np = &n->next;
next             3041 net/core/neighbour.c 				goto next;
next             3048 net/core/neighbour.c 					goto next;
next             3054 net/core/neighbour.c next:
next             3055 net/core/neighbour.c 			n = rcu_dereference_bh(n->next);
next             3079 net/core/neighbour.c 	n = rcu_dereference_bh(n->next);
next             3084 net/core/neighbour.c 				goto next;
next             3089 net/core/neighbour.c 				goto next;
next             3096 net/core/neighbour.c next:
next             3097 net/core/neighbour.c 			n = rcu_dereference_bh(n->next);
next             3141 net/core/neighbour.c 			pn = pn->next;
next             3159 net/core/neighbour.c 		pn = pn->next;
next             3167 net/core/neighbour.c 			pn = pn->next;
next             3338 net/core/neighbour.c 	.next	= neigh_stat_seq_next,
next              173 net/core/net-procfs.c 	.next  = dev_seq_next,
next              180 net/core/net-procfs.c 	.next  = softnet_seq_next,
next              225 net/core/net-procfs.c 	nxt = pt->list.next;
next              230 net/core/net-procfs.c 		nxt = ptype_base[0].next;
next              237 net/core/net-procfs.c 		nxt = ptype_base[hash].next;
next              270 net/core/net-procfs.c 	.next  = ptype_seq_next,
next              338 net/core/net-procfs.c 	.next  = dev_seq_next,
next             1327 net/core/net_namespace.c 		first_device = first_device->next;
next              256 net/core/netpoll.c 			clist = clist->next;
next               86 net/core/rtnetlink.c 		tail->next = defer_kfree_skb_list;
next              101 net/core/rtnetlink.c 		struct sk_buff *next = head->next;
next              105 net/core/rtnetlink.c 		head = next;
next              704 net/core/skbuff.c 		struct sk_buff *next = segs->next;
next              707 net/core/skbuff.c 		segs = next;
next              985 net/core/skbuff.c 	n->next = n->prev = NULL;
next             1137 net/core/skbuff.c 		u32 bytelen, next;
next             1155 net/core/skbuff.c 		next = (u32)atomic_read(&sk->sk_zckey);
next             1156 net/core/skbuff.c 		if ((u32)(uarg->id + uarg->len) == next) {
next             1161 net/core/skbuff.c 			atomic_set(&sk->sk_zckey, ++next);
next             1360 net/core/skbuff.c 				struct page *next = (struct page *)page_private(head);
next             1362 net/core/skbuff.c 				head = next;
next             1964 net/core/skbuff.c 	     fragp = &frag->next) {
next             1974 net/core/skbuff.c 			nfrag->next = frag->next;
next             1989 net/core/skbuff.c 		if (frag->next)
next             1990 net/core/skbuff.c 			skb_drop_list(&frag->next);
next             2099 net/core/skbuff.c 				list = list->next;
next             2109 net/core/skbuff.c 					insp = list->next;
next             2126 net/core/skbuff.c 			skb_shinfo(skb)->frag_list = list->next;
next             2131 net/core/skbuff.c 			clone->next = list;
next             2524 net/core/skbuff.c 		} else if (skb->next) {
next             2525 net/core/skbuff.c 			skb = skb->next;
next             3522 net/core/skbuff.c 	} else if (st->cur_skb->next) {
next             3523 net/core/skbuff.c 		st->cur_skb = st->cur_skb->next;
next             3719 net/core/skbuff.c 				if (frag_len != iter->len && iter->next)
next             3789 net/core/skbuff.c 			list_skb = list_skb->next;
next             3821 net/core/skbuff.c 			tail->next = nskb;
next             3882 net/core/skbuff.c 				list_skb = list_skb->next;
next             3956 net/core/skbuff.c 		for (iter = segs; iter; iter = iter->next) {
next             4078 net/core/skbuff.c 		NAPI_GRO_CB(p)->last->next = skb;
next             4351 net/core/skbuff.c 		if (skb1->next == NULL && tailbits) {
next             4382 net/core/skbuff.c 			skb2->next = skb1->next;
next             4389 net/core/skbuff.c 		skb_p = &skb1->next;
next             5791 net/core/skbuff.c 			list = list->next;
next             5799 net/core/skbuff.c 				insp = list->next;
next             5815 net/core/skbuff.c 		shinfo->frag_list = list->next;
next             5820 net/core/skbuff.c 		clone->next = list;
next             2430 net/core/sock.c 	struct sk_buff *skb, *next;
next             2438 net/core/sock.c 			next = skb->next;
next             2439 net/core/sock.c 			prefetch(next);
next             2446 net/core/sock.c 			skb = next;
next             3567 net/core/sock.c 	.next   = proto_seq_next,
next              324 net/core/sock_map.c static int sock_map_get_next_key(struct bpf_map *map, void *key, void *next)
next              328 net/core/sock_map.c 	u32 *key_next = next;
next               30 net/dccp/ackvec.c 	struct dccp_ackvec_record *cur, *next;
next               32 net/dccp/ackvec.c 	list_for_each_entry_safe(cur, next, &av->av_records, avr_node)
next              284 net/dccp/ackvec.c 	struct dccp_ackvec_record *avr, *next;
next              338 net/dccp/ackvec.c 	list_for_each_entry_safe_from(avr, next, &av->av_records, avr_node) {
next              364 net/dccp/ackvec.c 	struct dccp_ackvec_parsed *cur, *next;
next              366 net/dccp/ackvec.c 	list_for_each_entry_safe(cur, next, parsed_chunks, node)
next              240 net/dccp/ccids/ccid2.c 	struct ccid2_seq *next;
next              269 net/dccp/ccids/ccid2.c 	next = hc->tx_seqh->ccid2s_next;
next              271 net/dccp/ccids/ccid2.c 	if (next == hc->tx_seqt) {
next              277 net/dccp/ccids/ccid2.c 		next = hc->tx_seqh->ccid2s_next;
next              278 net/dccp/ccids/ccid2.c 		BUG_ON(next == hc->tx_seqt);
next              280 net/dccp/ccids/ccid2.c 	hc->tx_seqh = next;
next              378 net/dccp/ccids/ccid3.c 	tfrc_tx_hist_purge(&acked->next);
next               57 net/dccp/ccids/lib/packet_history.c 	entry->next  = *headp;
next               67 net/dccp/ccids/lib/packet_history.c 		struct tfrc_tx_hist_entry *next = head->next;
next               70 net/dccp/ccids/lib/packet_history.c 		head = next;
next               37 net/dccp/ccids/lib/packet_history.h 	struct tfrc_tx_hist_entry *next;
next               46 net/dccp/ccids/lib/packet_history.h 		head = head->next;
next              541 net/dccp/feat.c 	struct dccp_feat_entry *entry, *next;
next              543 net/dccp/feat.c 	list_for_each_entry_safe(entry, next, fn_list, node)
next              633 net/dccp/feat.c 	struct dccp_feat_entry *pos, *next;
next              638 net/dccp/feat.c 	list_for_each_entry_safe_reverse(pos, next, fn, node) {
next             1500 net/dccp/feat.c 	struct dccp_feat_entry *cur, *next;
next             1544 net/dccp/feat.c 	list_for_each_entry_safe(cur, next, fn_list, node)
next             2305 net/decnet/af_decnet.c 	.next	= dn_socket_seq_next,
next             1381 net/decnet/dn_dev.c 	.next	= dn_dev_seq_next,
next              588 net/decnet/dn_neigh.c 	.next  = neigh_seq_next,
next              357 net/decnet/dn_route.c 	struct dn_route *rt, *next;
next              365 net/decnet/dn_route.c 		for(; rt; rt = next) {
next              366 net/decnet/dn_route.c 			next = rcu_dereference_raw(rt->dn_next);
next             1845 net/decnet/dn_route.c 	.next	= dn_rt_cache_seq_next,
next              127 net/decnet/dn_table.c 	struct dn_fib_node *f, **fp, *next;
next              131 net/decnet/dn_table.c 		for(f = old_ht[i]; f; f = next) {
next              132 net/decnet/dn_table.c 			next = f->fn_next;
next              511 net/decnet/dn_table.c 				goto next;
next              518 net/decnet/dn_table.c next:
next              917 net/decnet/dn_table.c 	struct hlist_node *next;
next              922 net/decnet/dn_table.c 		hlist_for_each_entry_safe(t, next, &dn_fib_table_hash[h],
next             1360 net/ipv4/af_inet.c 			if (skb->next)
next             1387 net/ipv4/af_inet.c 	} while ((skb = skb->next));
next             1414 net/ipv4/arp.c 	.next	= neigh_seq_next,
next              268 net/ipv4/cipso_ipv4.c 				__list_del(entry->list.prev, entry->list.next);
next              699 net/ipv4/devinet.c 	unsigned long now, next, next_sec, next_sched;
next              705 net/ipv4/devinet.c 	next = round_jiffies_up(now + ADDR_CHECK_FREQUENCY);
next              729 net/ipv4/devinet.c 						ifa->ifa_valid_lft * HZ, next))
next              730 net/ipv4/devinet.c 					next = ifa->ifa_tstamp +
next              737 net/ipv4/devinet.c 					       next)) {
next              738 net/ipv4/devinet.c 				next = ifa->ifa_tstamp +
next              783 net/ipv4/devinet.c 	next_sec = round_jiffies_up(next);
next              784 net/ipv4/devinet.c 	next_sched = next;
next              787 net/ipv4/devinet.c 	if (time_before(next_sec, next + ADDRCONF_TIMER_FUZZ))
next             1026 net/ipv4/fib_frontend.c 				goto next;
next             1038 net/ipv4/fib_frontend.c next:
next              175 net/ipv4/fib_semantics.c 			struct fib_nh_exception *next;
next              177 net/ipv4/fib_semantics.c 			next = rcu_dereference_protected(fnhe->fnhe_next, 1);
next              184 net/ipv4/fib_semantics.c 			fnhe = next;
next              481 net/ipv4/fib_trie.c 	tn_info(tn)->rcu.next = NULL;
next              487 net/ipv4/fib_trie.c 	tn_info(n)->rcu.next = tn_info(tn)->rcu.next;
next              488 net/ipv4/fib_trie.c 	tn_info(tn)->rcu.next = &tn_info(n)->rcu;
next              496 net/ipv4/fib_trie.c 		head = head->next;
next             1520 net/ipv4/fib_trie.c 	struct fib_alias *fa = hlist_entry(pprev, typeof(*fa), fa_list.next);
next             2109 net/ipv4/fib_trie.c 			goto next;
next             2114 net/ipv4/fib_trie.c 			goto next;
next             2118 net/ipv4/fib_trie.c 				goto next;
next             2122 net/ipv4/fib_trie.c 				goto next;
next             2126 net/ipv4/fib_trie.c 				goto next;
next             2153 net/ipv4/fib_trie.c next:
next             2653 net/ipv4/fib_trie.c 	.next   = fib_trie_seq_next,
next             2842 net/ipv4/fib_trie.c 	.next   = fib_route_seq_next,
next             1257 net/ipv4/fou.c 	struct fou *fou, *next;
next             1261 net/ipv4/fou.c 	list_for_each_entry_safe(fou, next, &fn->fou_list, list)
next              102 net/ipv4/gre_offload.c 	} while ((skb = skb->next));
next              189 net/ipv4/igmp.c 	struct ip_sf_list *next;
next              192 net/ipv4/igmp.c 		next = psf->sf_next;
next              194 net/ipv4/igmp.c 		psf = next;
next              663 net/ipv4/igmp.c 		pmc_next = pmc->next;
next              683 net/ipv4/igmp.c 				pmc_prev->next = pmc_next;
next             1198 net/ipv4/igmp.c 	pmc->next = in_dev->mc_tomb;
next             1215 net/ipv4/igmp.c 	for (pmc = in_dev->mc_tomb; pmc; pmc = pmc->next) {
next             1222 net/ipv4/igmp.c 			pmc_prev->next = pmc->next;
next             1224 net/ipv4/igmp.c 			in_dev->mc_tomb = pmc->next;
next             1258 net/ipv4/igmp.c 		nextpmc = pmc->next;
next             2865 net/ipv4/igmp.c 	.next	=	igmp_mc_seq_next,
next             2915 net/ipv4/igmp.c 		state->im = state->im->next;
next             2998 net/ipv4/igmp.c 	.next	=	igmp_mcf_seq_next,
next              997 net/ipv4/inet_connection_sock.c 	struct request_sock *next, *req;
next             1030 net/ipv4/inet_connection_sock.c 			next = req->dl_next;
next             1032 net/ipv4/inet_connection_sock.c 			req = next;
next              243 net/ipv4/inet_fragment.c 			struct sk_buff *next = FRAG_CB(skb)->next_frag;
next              247 net/ipv4/inet_fragment.c 			skb = next;
next              468 net/ipv4/inet_fragment.c 		nextp = &clone->next;
next              522 net/ipv4/inet_fragment.c 				nextp = &fp->next;
next              640 net/ipv4/inet_hashtables.c 		if (sk_head(&tb->owners) == sk && !sk->sk_bind_node.next) {
next              106 net/ipv4/inetpeer.c 	struct rb_node **pp, *parent, *next;
next              114 net/ipv4/inetpeer.c 		next = rcu_dereference_raw(*pp);
next              115 net/ipv4/inetpeer.c 		if (!next)
next              117 net/ipv4/inetpeer.c 		parent = next;
next              132 net/ipv4/inetpeer.c 			pp = &next->rb_left;
next              134 net/ipv4/inetpeer.c 			pp = &next->rb_right;
next              156 net/ipv4/ip_input.c 	for (ra = rcu_dereference(net->ipv4.ra_chain); ra; ra = rcu_dereference(ra->next)) {
next              530 net/ipv4/ip_input.c 	struct sk_buff *skb, *next;
next              532 net/ipv4/ip_input.c 	list_for_each_entry_safe(skb, next, head, list) {
next              542 net/ipv4/ip_input.c 	struct sk_buff *skb, *next;
next              546 net/ipv4/ip_input.c 	list_for_each_entry_safe(skb, next, head, list) {
next              589 net/ipv4/ip_input.c 	struct sk_buff *skb, *next;
next              593 net/ipv4/ip_input.c 	list_for_each_entry_safe(skb, next, head, list) {
next              276 net/ipv4/ip_output.c 		struct sk_buff *nskb = segs->next;
next              640 net/ipv4/ip_output.c 	if (frag->next)
next              823 net/ipv4/ip_output.c 			    ((frag->len & 7) && frag->next) ||
next             1491 net/ipv4/ip_output.c 		tail_skb = &(tmp_skb->next);
next              353 net/ipv4/ip_sockglue.c 	     rap = &ra->next) {
next              362 net/ipv4/ip_sockglue.c 			RCU_INIT_POINTER(*rap, ra->next);
next              384 net/ipv4/ip_sockglue.c 	RCU_INIT_POINTER(new_ra->next, ra);
next              191 net/ipv4/ipconfig.c 	struct ic_device *next;
next              254 net/ipv4/ipconfig.c 			last = &d->next;
next              310 net/ipv4/ipconfig.c 	struct ic_device *d, *next;
next              314 net/ipv4/ipconfig.c 	next = ic_first_dev;
next              315 net/ipv4/ipconfig.c 	while ((d = next)) {
next              316 net/ipv4/ipconfig.c 		next = d->next;
next              527 net/ipv4/ipconfig.c 		d = d->next;
next             1030 net/ipv4/ipconfig.c 		d = d->next;
next             1231 net/ipv4/ipconfig.c 		if (!d->next) {
next             1254 net/ipv4/ipconfig.c 		if ((d = d->next))
next             1491 net/ipv4/ipconfig.c 	    ic_first_dev->next) {
next              121 net/ipv4/ipmr.c 		ret = list_entry_rcu(net->ipv4.mr_tables.next,
next              124 net/ipv4/ipmr.c 		ret = list_entry_rcu(mrt->list.next,
next              270 net/ipv4/ipmr.c 	struct mr_table *mrt, *next;
next              273 net/ipv4/ipmr.c 	list_for_each_entry_safe(mrt, next, &net->ipv4.mr_tables, list) {
next              773 net/ipv4/ipmr.c 	struct mr_mfc *c, *next;
next              788 net/ipv4/ipmr.c 	list_for_each_entry_safe(c, next, &mrt->mfc_unres_queue, list) {
next             2964 net/ipv4/ipmr.c 	.next  = mr_vif_seq_next,
next             3024 net/ipv4/ipmr.c 	.next  = mr_mfc_seq_next,
next              185 net/ipv4/ipmr_base.c 	if (c->list.next != it->cache)
next              186 net/ipv4/ipmr_base.c 		return list_entry(c->list.next, struct mr_mfc, list);
next              349 net/ipv4/netfilter/arp_tables.c 						goto next;
next              382 net/ipv4/netfilter/arp_tables.c next:		;
next              413 net/ipv4/netfilter/ip_tables.c 						goto next;
next              446 net/ipv4/netfilter/ip_tables.c next:		;
next              740 net/ipv4/netfilter/ipt_CLUSTERIP.c 	.next	= clusterip_seq_next,
next              145 net/ipv4/nexthop.c 	struct rb_node **pp, *parent = NULL, *next;
next              151 net/ipv4/nexthop.c 		next = rcu_dereference_raw(*pp);
next              152 net/ipv4/nexthop.c 		if (!next)
next              154 net/ipv4/nexthop.c 		parent = next;
next              158 net/ipv4/nexthop.c 			pp = &next->rb_left;
next              160 net/ipv4/nexthop.c 			pp = &next->rb_right;
next             1010 net/ipv4/nexthop.c 	struct rb_node **pp, *parent = NULL, *next;
next             1022 net/ipv4/nexthop.c 		next = rtnl_dereference(*pp);
next             1023 net/ipv4/nexthop.c 		if (!next)
next             1026 net/ipv4/nexthop.c 		parent = next;
next             1030 net/ipv4/nexthop.c 			pp = &next->rb_left;
next             1032 net/ipv4/nexthop.c 			pp = &next->rb_right;
next             1141 net/ipv4/ping.c 	.next		= ping_seq_next,
next             1098 net/ipv4/raw.c 	.next  = raw_seq_next,
next              172 net/ipv4/raw_diag.c 				goto next;
next              174 net/ipv4/raw_diag.c 				goto next;
next              177 net/ipv4/raw_diag.c 				goto next;
next              180 net/ipv4/raw_diag.c 				goto next;
next              183 net/ipv4/raw_diag.c next:
next              230 net/ipv4/route.c 	.next   = rt_cache_seq_next,
next              319 net/ipv4/route.c 	.next   = rt_cpu_seq_next,
next             2886 net/ipv4/route.c 				goto next;
next             2889 net/ipv4/route.c 				goto next;
next             2893 net/ipv4/route.c 				goto next;
next             2899 net/ipv4/route.c 				goto next;
next             2907 net/ipv4/route.c next:
next             1790 net/ipv4/tcp.c 				skb = skb->next;
next             2299 net/ipv4/tcp.c 	int next = (int)new_state[sk->sk_state];
next             2300 net/ipv4/tcp.c 	int ns = next & TCP_STATE_MASK;
next             2304 net/ipv4/tcp.c 	return next & TCP_ACTION_FIN;
next             3079 net/ipv4/tcp_input.c 	struct sk_buff *skb, *next;
next             3091 net/ipv4/tcp_input.c 	for (skb = skb_rb_first(&sk->tcp_rtx_queue); skb; skb = next) {
next             3162 net/ipv4/tcp_input.c 		next = skb_rb_next(skb);
next             3167 net/ipv4/tcp_input.c 		tcp_highest_sack_replace(sk, skb, next);
next             4863 net/ipv4/tcp_input.c 		return !skb_queue_is_last(list, skb) ? skb->next : NULL;
next             4872 net/ipv4/tcp_input.c 	struct sk_buff *next = tcp_skb_next(skb, list);
next             4882 net/ipv4/tcp_input.c 	return next;
next             2535 net/ipv4/tcp_ipv4.c 	.next		= tcp_seq_next,
next               24 net/ipv4/tcp_offload.c 		skb = skb->next;
next              124 net/ipv4/tcp_offload.c 	while (skb->next) {
next              139 net/ipv4/tcp_offload.c 		skb = skb->next;
next             2068 net/ipv4/tcp_output.c 	struct sk_buff *skb, *next;
next             2071 net/ipv4/tcp_output.c 	tcp_for_write_queue_from_safe(skb, next, sk) {
next             2097 net/ipv4/tcp_output.c 	struct sk_buff *skb, *nskb, *next;
next             2180 net/ipv4/tcp_output.c 	tcp_for_write_queue_from_safe(skb, next, sk) {
next               44 net/ipv4/tunnel4.c 	     pprev = &t->next) {
next               51 net/ipv4/tunnel4.c 	handler->next = *pprev;
next               74 net/ipv4/tunnel4.c 	     pprev = &t->next) {
next               76 net/ipv4/tunnel4.c 			*pprev = handler->next;
next               93 net/ipv4/tunnel4.c 	     handler = rcu_dereference(handler->next))	\
next             2103 net/ipv4/udp.c 	struct sk_buff *next, *segs;
next             2112 net/ipv4/udp.c 	for (skb = segs; skb; skb = next) {
next             2113 net/ipv4/udp.c 		next = skb->next;
next             2941 net/ipv4/udp.c 	.next		= udp_seq_next,
next              122 net/ipv4/udp_diag.c 				goto next;
next              124 net/ipv4/udp_diag.c 				goto next;
next              127 net/ipv4/udp_diag.c 				goto next;
next              130 net/ipv4/udp_diag.c 				goto next;
next              133 net/ipv4/udp_diag.c 				goto next;
next              139 net/ipv4/udp_diag.c next:
next              143 net/ipv4/udp_offload.c 	} while ((skb = skb->next));
next              243 net/ipv4/udp_offload.c 		if (!seg->next)
next              255 net/ipv4/udp_offload.c 		seg = seg->next;
next               43 net/ipv4/xfrm4_protocol.c 	     handler = rcu_dereference(handler->next))	\
next              232 net/ipv4/xfrm4_protocol.c 	     pprev = &t->next) {
next              239 net/ipv4/xfrm4_protocol.c 	handler->next = *pprev;
next              273 net/ipv4/xfrm4_protocol.c 	     pprev = &t->next) {
next              275 net/ipv4/xfrm4_protocol.c 			*pprev = handler->next;
next             4350 net/ipv6/addrconf.c 	.next	= if6_seq_next,
next             4413 net/ipv6/addrconf.c 	unsigned long now, next, next_sec, next_sched;
next             4421 net/ipv6/addrconf.c 	next = round_jiffies_up(now + ADDR_CHECK_FREQUENCY);
next             4461 net/ipv6/addrconf.c 				    (time_before(ifp->tstamp + ifp->valid_lft * HZ, next)))
next             4462 net/ipv6/addrconf.c 					next = ifp->tstamp + ifp->valid_lft * HZ;
next             4481 net/ipv6/addrconf.c 					if (time_before(ifp->tstamp + ifp->prefered_lft * HZ, next))
next             4482 net/ipv6/addrconf.c 						next = ifp->tstamp + ifp->prefered_lft * HZ;
next             4499 net/ipv6/addrconf.c 				} else if (time_before(ifp->tstamp + ifp->prefered_lft * HZ - regen_advance * HZ, next))
next             4500 net/ipv6/addrconf.c 					next = ifp->tstamp + ifp->prefered_lft * HZ - regen_advance * HZ;
next             4504 net/ipv6/addrconf.c 				if (time_before(ifp->tstamp + ifp->prefered_lft * HZ, next))
next             4505 net/ipv6/addrconf.c 					next = ifp->tstamp + ifp->prefered_lft * HZ;
next             4511 net/ipv6/addrconf.c 	next_sec = round_jiffies_up(next);
next             4512 net/ipv6/addrconf.c 	next_sched = next;
next             4515 net/ipv6/addrconf.c 	if (time_before(next_sec, next + ADDRCONF_TIMER_FUZZ))
next             4523 net/ipv6/addrconf.c 		 now, next, next_sec, next_sched);
next             5053 net/ipv6/addrconf.c 				goto next;
next             5058 net/ipv6/addrconf.c next:
next             5068 net/ipv6/addrconf.c 		     ifmca = ifmca->next, ip_idx++) {
next              203 net/ipv6/anycast.c 		struct ipv6_ac_socklist *next = pac->acl_next;
next              212 net/ipv6/anycast.c 		pac = next;
next              573 net/ipv6/anycast.c 	.next	=	ac6_seq_next,
next              230 net/ipv6/calipso.c 				__list_del(entry->list.prev, entry->list.next);
next               21 net/ipv6/ila/ila_xlat.c 	struct ila_map __rcu *next;
next              138 net/ipv6/ila/ila_xlat.c 		ila = rcu_access_pointer(ila->next);
next              156 net/ipv6/ila/ila_xlat.c 		ila = rcu_access_pointer(ila->next);
next              169 net/ipv6/ila/ila_xlat.c 	struct ila_map *next;
next              173 net/ipv6/ila/ila_xlat.c 		next = rcu_access_pointer(ila->next);
next              175 net/ipv6/ila/ila_xlat.c 		ila = next;
next              255 net/ipv6/ila/ila_xlat.c 			tila = rcu_dereference_protected(tila->next,
next              261 net/ipv6/ila/ila_xlat.c 			RCU_INIT_POINTER(ila->next, tila);
next              262 net/ipv6/ila/ila_xlat.c 			rcu_assign_pointer(prev->next, ila);
next              265 net/ipv6/ila/ila_xlat.c 			RCU_INIT_POINTER(ila->next, head);
next              301 net/ipv6/ila/ila_xlat.c 			ila = rcu_dereference_protected(ila->next,
next              310 net/ipv6/ila/ila_xlat.c 			rcu_assign_pointer(prev->next, ila->next);
next              315 net/ipv6/ila/ila_xlat.c 			head = rcu_dereference_protected(ila->next,
next              554 net/ipv6/ila/ila_xlat.c 			ila = rcu_access_pointer(ila->next);
next              588 net/ipv6/ila/ila_xlat.c 			ila = rcu_access_pointer(ila->next);
next              640 net/ipv6/ip6_fib.c 				goto next;
next              644 net/ipv6/ip6_fib.c next:
next             1492 net/ipv6/ip6_fib.c 		struct fib6_node *next;
next             1496 net/ipv6/ip6_fib.c 		next = dir ? rcu_dereference(fn->right) :
next             1499 net/ipv6/ip6_fib.c 		if (next) {
next             1500 net/ipv6/ip6_fib.c 			fn = next;
next             1601 net/ipv6/ip6_fib.c 				goto next;
next             1619 net/ipv6/ip6_fib.c next:
next             2547 net/ipv6/ip6_fib.c 	.next	= ipv6_route_seq_next,
next               63 net/ipv6/ip6_flowlabel.c 	     fl = rcu_dereference_bh(fl->next))
next               65 net/ipv6/ip6_flowlabel.c 	for (fl = rcu_dereference_bh(fl->next);			\
next               67 net/ipv6/ip6_flowlabel.c 	     fl = rcu_dereference_bh(fl->next))
next               72 net/ipv6/ip6_flowlabel.c 	     sfl = rcu_dereference_bh(sfl->next))
next              169 net/ipv6/ip6_flowlabel.c 					*flp = fl->next;
next              177 net/ipv6/ip6_flowlabel.c 			flp = &fl->next;
next              202 net/ipv6/ip6_flowlabel.c 				*flp = fl->next;
next              207 net/ipv6/ip6_flowlabel.c 			flp = &fl->next;
next              248 net/ipv6/ip6_flowlabel.c 	fl->next = fl_ht[FL_HASH(fl->label)];
next              292 net/ipv6/ip6_flowlabel.c 		np->ipv6_fl_list = sfl->next;
next              494 net/ipv6/ip6_flowlabel.c 	sfl->next = np->ipv6_fl_list;
next              569 net/ipv6/ip6_flowlabel.c 		     sflp = &sfl->next) {
next              573 net/ipv6/ip6_flowlabel.c 				*sflp = sfl->next;
next              833 net/ipv6/ip6_flowlabel.c 	.next	=	ip6fl_seq_next,
next              301 net/ipv6/ip6_gre.c 	rcu_assign_pointer(t->next, rtnl_dereference(*tp));
next              312 net/ipv6/ip6_gre.c 	     tp = &iter->next) {
next              314 net/ipv6/ip6_gre.c 			rcu_assign_pointer(*tp, t->next);
next              334 net/ipv6/ip6_gre.c 	     tp = &t->next)
next             1551 net/ipv6/ip6_gre.c 				t = rtnl_dereference(t->next);
next               81 net/ipv6/ip6_input.c 	struct sk_buff *skb, *next;
next               83 net/ipv6/ip6_input.c 	list_for_each_entry_safe(skb, next, head, list) {
next               93 net/ipv6/ip6_input.c 	struct sk_buff *skb, *next;
next               97 net/ipv6/ip6_input.c 	list_for_each_entry_safe(skb, next, head, list) {
next              303 net/ipv6/ip6_input.c 	struct sk_buff *skb, *next;
next              307 net/ipv6/ip6_input.c 	list_for_each_entry_safe(skb, next, head, list) {
next              121 net/ipv6/ip6_offload.c 	for (skb = segs; skb; skb = skb->next) {
next              141 net/ipv6/ip6_offload.c 			if (skb->next)
next              302 net/ipv6/ip6_output.c 	for (ra = ip6_ra_chain; ra; ra = ra->next) {
next              657 net/ipv6/ip6_output.c 	if (frag->next)
next              832 net/ipv6/ip6_output.c 			    ((frag->len & 7) && frag->next) ||
next             1740 net/ipv6/ip6_output.c 		tail_skb = &(tmp_skb->next);
next              134 net/ipv6/ip6_tunnel.c 	for (t = rcu_dereference(start); t; t = rcu_dereference(t->next))
next              217 net/ipv6/ip6_tunnel.c 	rcu_assign_pointer(t->next , rtnl_dereference(*tp));
next              237 net/ipv6/ip6_tunnel.c 	     tp = &iter->next) {
next              239 net/ipv6/ip6_tunnel.c 			rcu_assign_pointer(*tp, t->next);
next              352 net/ipv6/ip6_tunnel.c 	     tp = &t->next) {
next              403 net/ipv6/ip6_tunnel.c 	u8 next, nexthdr = ipv6h->nexthdr;
next              426 net/ipv6/ip6_tunnel.c 		next = hdr->nexthdr;
next              453 net/ipv6/ip6_tunnel.c 		nexthdr = next;
next             2209 net/ipv6/ip6_tunnel.c 			t = rtnl_dereference(t->next);
next               75 net/ipv6/ip6_vti.c 	for (t = rcu_dereference(start); t; t = rcu_dereference(t->next))
next              156 net/ipv6/ip6_vti.c 	rcu_assign_pointer(t->next , rtnl_dereference(*tp));
next              168 net/ipv6/ip6_vti.c 	     tp = &iter->next) {
next              170 net/ipv6/ip6_vti.c 			rcu_assign_pointer(*tp, t->next);
next              266 net/ipv6/ip6_vti.c 	     tp = &t->next) {
next             1116 net/ipv6/ip6_vti.c 			t = rtnl_dereference(t->next);
next              108 net/ipv6/ip6mr.c 		ret = list_entry_rcu(net->ipv6.mr6_tables.next,
next              111 net/ipv6/ip6mr.c 		ret = list_entry_rcu(mrt->list.next,
next              257 net/ipv6/ip6mr.c 	struct mr_table *mrt, *next;
next              260 net/ipv6/ip6mr.c 	list_for_each_entry_safe(mrt, next, &net->ipv6.mr6_tables, list) {
next              448 net/ipv6/ip6mr.c 	.next  = mr_vif_seq_next,
next              509 net/ipv6/ip6mr.c 	.next  = mr_mfc_seq_next,
next              796 net/ipv6/ip6mr.c 	struct mr_mfc *c, *next;
next              798 net/ipv6/ip6mr.c 	list_for_each_entry_safe(c, next, &mrt->mfc_unres_queue, list) {
next               71 net/ipv6/ipv6_sockglue.c 	for (rap = &ip6_ra_chain; (ra = *rap) != NULL; rap = &ra->next) {
next               79 net/ipv6/ipv6_sockglue.c 			*rap = ra->next;
next               93 net/ipv6/ipv6_sockglue.c 	new_ra->next = ra;
next              119 net/ipv6/mcast.c 	     pmc = rcu_dereference(pmc->next))
next              162 net/ipv6/mcast.c 	mc_lst->next = NULL;
next              196 net/ipv6/mcast.c 	mc_lst->next = np->ipv6_mc_list;
next              231 net/ipv6/mcast.c 	      lnk = &mc_lst->next) {
next              236 net/ipv6/mcast.c 			*lnk = mc_lst->next;
next              300 net/ipv6/mcast.c 		np->ipv6_mc_list = mc_lst->next;
next              760 net/ipv6/mcast.c 	pmc->next = idev->mc_tomb;
next              773 net/ipv6/mcast.c 	for (pmc = idev->mc_tomb; pmc; pmc = pmc->next) {
next              780 net/ipv6/mcast.c 			pmc_prev->next = pmc->next;
next              782 net/ipv6/mcast.c 			idev->mc_tomb = pmc->next;
next              814 net/ipv6/mcast.c 		nextpmc = pmc->next;
next              822 net/ipv6/mcast.c 	for (pmc = idev->mc_list; pmc; pmc = pmc->next) {
next              904 net/ipv6/mcast.c 	for (mc = idev->mc_list; mc; mc = mc->next) {
next              921 net/ipv6/mcast.c 	mc->next = idev->mc_list;
next              952 net/ipv6/mcast.c 	for (map = &idev->mc_list; (ma = *map) != NULL; map = &ma->next) {
next              955 net/ipv6/mcast.c 				*map = ma->next;
next             1004 net/ipv6/mcast.c 		for (mc = idev->mc_list; mc; mc = mc->next) {
next             1422 net/ipv6/mcast.c 		for (ma = idev->mc_list; ma; ma = ma->next) {
next             1428 net/ipv6/mcast.c 		for (ma = idev->mc_list; ma; ma = ma->next) {
next             1492 net/ipv6/mcast.c 	for (ma = idev->mc_list; ma; ma = ma->next) {
next             1862 net/ipv6/mcast.c 		for (pmc = idev->mc_list; pmc; pmc = pmc->next) {
next             1920 net/ipv6/mcast.c 		pmc_next = pmc->next;
next             1941 net/ipv6/mcast.c 				pmc_prev->next = pmc_next;
next             1952 net/ipv6/mcast.c 	for (pmc = idev->mc_list; pmc; pmc = pmc->next) {
next             2088 net/ipv6/mcast.c 	for (pmc = idev->mc_list; pmc; pmc = pmc->next) {
next             2176 net/ipv6/mcast.c 	for (pmc = idev->mc_list; pmc; pmc = pmc->next) {
next             2340 net/ipv6/mcast.c 	for (pmc = idev->mc_list; pmc; pmc = pmc->next) {
next             2520 net/ipv6/mcast.c 	for (i = idev->mc_list; i; i = i->next)
next             2540 net/ipv6/mcast.c 	for (i = idev->mc_list; i; i = i->next)
next             2571 net/ipv6/mcast.c 	for (i = idev->mc_list; i; i = i->next) {
next             2618 net/ipv6/mcast.c 		idev->mc_list = i->next;
next             2635 net/ipv6/mcast.c 		for (pmc = idev->mc_list; pmc; pmc = pmc->next)
next             2701 net/ipv6/mcast.c 	im = im->next;
next             2774 net/ipv6/mcast.c 	.next	=	igmp6_mc_seq_next,
next             2826 net/ipv6/mcast.c 		state->im = state->im->next;
next             2915 net/ipv6/mcast.c 	.next	=	igmp6_mcf_seq_next,
next              431 net/ipv6/netfilter/ip6_tables.c 						goto next;
next              464 net/ipv6/netfilter/ip6_tables.c next:		;
next              214 net/ipv6/ping.c 	.next		= ping_seq_next,
next             1323 net/ipv6/raw.c 	.next =		raw_seq_next,
next              904 net/ipv6/route.c 		struct fib6_info *next = rcu_dereference(rt0->fib6_next);
next              907 net/ipv6/route.c 		if (!next || next->fib6_metric != rt0->fib6_metric)
next              908 net/ipv6/route.c 			next = leaf;
next              910 net/ipv6/route.c 		if (next != rt0) {
next              913 net/ipv6/route.c 			if (next->fib6_node)
next              914 net/ipv6/route.c 				rcu_assign_pointer(fn->rr_ptr, next);
next             5010 net/ipv6/route.c 	struct list_head next;
next             5021 net/ipv6/route.c 	list_for_each_entry(nh, rt6_nh_list, next) {
next             5032 net/ipv6/route.c 	list_add_tail(&nh->next, rt6_nh_list);
next             5153 net/ipv6/route.c 	list_for_each_entry(nh, &rt6_nh_list, next) {
next             5211 net/ipv6/route.c 	list_for_each_entry(nh, &rt6_nh_list, next) {
next             5218 net/ipv6/route.c 	list_for_each_entry_safe(nh, nh_safe, &rt6_nh_list, next) {
next             5221 net/ipv6/route.c 		list_del(&nh->next);
next              160 net/ipv6/sit.c 	     tp = &iter->next) {
next              162 net/ipv6/sit.c 			rcu_assign_pointer(*tp, t->next);
next              172 net/ipv6/sit.c 	rcu_assign_pointer(t->next, rtnl_dereference(*tp));
next              236 net/ipv6/sit.c 	     tp = &t->next) {
next              280 net/ipv6/sit.c 	     prl = rcu_dereference(prl->next))
next              368 net/ipv6/sit.c 	for (p = rtnl_dereference(t->prl); p; p = rtnl_dereference(p->next)) {
next              390 net/ipv6/sit.c 	p->next = t->prl;
next              405 net/ipv6/sit.c 		n = rcu_dereference_protected(p->next, 1);
next              423 net/ipv6/sit.c 		     p = &x->next) {
next              425 net/ipv6/sit.c 				*p = x->next;
next             1834 net/ipv6/sit.c 				t = rtnl_dereference(t->next);
next             1984 net/ipv6/tcp_ipv6.c 	.next		= tcp_seq_next,
next               38 net/ipv6/tunnel6.c 	     pprev = &t->next) {
next               45 net/ipv6/tunnel6.c 	handler->next = *pprev;
next               68 net/ipv6/tunnel6.c 	     pprev = &t->next) {
next               70 net/ipv6/tunnel6.c 			*pprev = handler->next;
next               87 net/ipv6/tunnel6.c 	     handler = rcu_dereference(handler->next))	\
next              685 net/ipv6/udp.c 	struct sk_buff *next, *segs;
next              693 net/ipv6/udp.c 	for (skb = segs; skb; skb = next) {
next              694 net/ipv6/udp.c 		next = skb->next;
next             1089 net/ipv6/udp.c 		} while ((frags = frags->next));
next             1627 net/ipv6/udp.c 	.next		= udp_seq_next,
next               43 net/ipv6/xfrm6_protocol.c 	     handler = rcu_dereference(handler->next))	\
next              206 net/ipv6/xfrm6_protocol.c 	     pprev = &t->next) {
next              213 net/ipv6/xfrm6_protocol.c 	handler->next = *pprev;
next              247 net/ipv6/xfrm6_protocol.c 	     pprev = &t->next) {
next              249 net/ipv6/xfrm6_protocol.c 			*pprev = handler->next;
next              238 net/kcm/kcmproc.c 	.next	= kcm_seq_next,
next              696 net/kcm/kcmsock.c 		} else if (skb->next) {
next              697 net/kcm/kcmsock.c 			skb = skb->next;
next              809 net/kcm/kcmsock.c 				skb->next = tskb;
next              976 net/kcm/kcmsock.c 					skb->next = tskb;
next             3800 net/key/af_key.c 	.next	= pfkey_seq_next,
next              262 net/l2tp/l2tp_debugfs.c 	.next		= l2tp_dfs_seq_next,
next             1625 net/l2tp/l2tp_ppp.c 	.next		= pppol2tp_seq_next,
next               66 net/lapb/lapb_iface.c 	if (lapb->node.next) {
next               82 net/llc/llc_proc.c 	struct sock* sk, *next;
next               92 net/llc/llc_proc.c 	next = sk_nulls_next(sk);
next               93 net/llc/llc_proc.c 	if (next) {
next               94 net/llc/llc_proc.c 		sk = next;
next              205 net/llc/llc_proc.c 	.next   = llc_seq_next,
next              212 net/llc/llc_proc.c 	.next   = llc_seq_next,
next             1485 net/mac80211/mlme.c 			goto next;
next             1497 net/mac80211/mlme.c  next:
next             1066 net/mac80211/tx.c 			struct sk_buff *next = skb_queue_next(&tx->skbs, skb);
next             1067 net/mac80211/tx.c 			next_len = next->len;
next             3299 net/mac80211/tx.c 		frag_tail = &(*frag_tail)->next;
next             3847 net/mac80211/tx.c 	struct sk_buff *next;
next             3908 net/mac80211/tx.c 	next = skb;
next             3909 net/mac80211/tx.c 	while (next) {
next             3910 net/mac80211/tx.c 		skb = next;
next             3911 net/mac80211/tx.c 		next = skb->next;
next             3914 net/mac80211/tx.c 		skb->next = NULL;
next              915 net/mac802154/llsec.c 		struct mac802154_llsec_device_key *next;
next              917 net/mac802154/llsec.c 		next = kzalloc(sizeof(*devkey), GFP_ATOMIC);
next              918 net/mac802154/llsec.c 		if (!next)
next              921 net/mac802154/llsec.c 		next->devkey.key_id = *in_key;
next              927 net/mac802154/llsec.c 			list_add_rcu(&next->devkey.list, &dev->dev.keys);
next              929 net/mac802154/llsec.c 			kzfree(next);
next             1514 net/mpls/af_mpls.c 				goto next;
next             1530 net/mpls/af_mpls.c next:
next               63 net/mpls/mpls_gso.c 	} while ((skb = skb->next));
next              331 net/netfilter/ipset/ip_set_hash_gen.h 	struct mtype_elem next; /* temporary storage for uadd */
next              956 net/netfilter/ipset/ip_set_hash_gen.h 			mtype_data_next(&h->next, d);
next               67 net/netfilter/ipset/ip_set_hash_ip.c hash_ip4_data_next(struct hash_ip4_elem *next, const struct hash_ip4_elem *e)
next               69 net/netfilter/ipset/ip_set_hash_ip.c 	next->ip = e->ip;
next              147 net/netfilter/ipset/ip_set_hash_ip.c 		ip = ntohl(h->next.ip);
next              200 net/netfilter/ipset/ip_set_hash_ip.c hash_ip6_data_next(struct hash_ip6_elem *next, const struct hash_ip6_elem *e)
next               71 net/netfilter/ipset/ip_set_hash_ipmac.c hash_ipmac4_data_next(struct hash_ipmac4_elem *next,
next               74 net/netfilter/ipset/ip_set_hash_ipmac.c 	next->ip = e->ip;
next              179 net/netfilter/ipset/ip_set_hash_ipmac.c hash_ipmac6_data_next(struct hash_ipmac6_elem *next,
next               68 net/netfilter/ipset/ip_set_hash_ipmark.c hash_ipmark4_data_next(struct hash_ipmark4_elem *next,
next               71 net/netfilter/ipset/ip_set_hash_ipmark.c 	next->ip = d->ip;
next              146 net/netfilter/ipset/ip_set_hash_ipmark.c 		ip = ntohl(h->next.ip);
next              191 net/netfilter/ipset/ip_set_hash_ipmark.c hash_ipmark6_data_next(struct hash_ipmark6_elem *next,
next               75 net/netfilter/ipset/ip_set_hash_ipport.c hash_ipport4_data_next(struct hash_ipport4_elem *next,
next               78 net/netfilter/ipset/ip_set_hash_ipport.c 	next->ip = d->ip;
next               79 net/netfilter/ipset/ip_set_hash_ipport.c 	next->port = d->port;
next              176 net/netfilter/ipset/ip_set_hash_ipport.c 		ip = ntohl(h->next.ip);
next              178 net/netfilter/ipset/ip_set_hash_ipport.c 		p = retried && ip == ntohl(h->next.ip) ? ntohs(h->next.port)
next              230 net/netfilter/ipset/ip_set_hash_ipport.c hash_ipport6_data_next(struct hash_ipport6_elem *next,
next              233 net/netfilter/ipset/ip_set_hash_ipport.c 	next->port = d->port;
next              323 net/netfilter/ipset/ip_set_hash_ipport.c 		port = ntohs(h->next.port);
next               76 net/netfilter/ipset/ip_set_hash_ipportip.c hash_ipportip4_data_next(struct hash_ipportip4_elem *next,
next               79 net/netfilter/ipset/ip_set_hash_ipportip.c 	next->ip = d->ip;
next               80 net/netfilter/ipset/ip_set_hash_ipportip.c 	next->port = d->port;
next              183 net/netfilter/ipset/ip_set_hash_ipportip.c 		ip = ntohl(h->next.ip);
next              185 net/netfilter/ipset/ip_set_hash_ipportip.c 		p = retried && ip == ntohl(h->next.ip) ? ntohs(h->next.port)
next              240 net/netfilter/ipset/ip_set_hash_ipportip.c hash_ipportip6_data_next(struct hash_ipportip6_elem *next,
next              243 net/netfilter/ipset/ip_set_hash_ipportip.c 	next->port = d->port;
next              338 net/netfilter/ipset/ip_set_hash_ipportip.c 		port = ntohs(h->next.port);
next              120 net/netfilter/ipset/ip_set_hash_ipportnet.c hash_ipportnet4_data_next(struct hash_ipportnet4_elem *next,
next              123 net/netfilter/ipset/ip_set_hash_ipportnet.c 	next->ip = d->ip;
next              124 net/netfilter/ipset/ip_set_hash_ipportnet.c 	next->port = d->port;
next              125 net/netfilter/ipset/ip_set_hash_ipportnet.c 	next->ip2 = d->ip2;
next              269 net/netfilter/ipset/ip_set_hash_ipportnet.c 		ip = ntohl(h->next.ip);
next              270 net/netfilter/ipset/ip_set_hash_ipportnet.c 		p = ntohs(h->next.port);
next              271 net/netfilter/ipset/ip_set_hash_ipportnet.c 		ip2 = ntohl(h->next.ip2);
next              369 net/netfilter/ipset/ip_set_hash_ipportnet.c hash_ipportnet6_data_next(struct hash_ipportnet6_elem *next,
next              372 net/netfilter/ipset/ip_set_hash_ipportnet.c 	next->port = d->port;
next              494 net/netfilter/ipset/ip_set_hash_ipportnet.c 		port = ntohs(h->next.port);
next               60 net/netfilter/ipset/ip_set_hash_mac.c hash_mac4_data_next(struct hash_mac4_elem *next,
next              101 net/netfilter/ipset/ip_set_hash_net.c hash_net4_data_next(struct hash_net4_elem *next,
next              104 net/netfilter/ipset/ip_set_hash_net.c 	next->ip = d->ip;
next              191 net/netfilter/ipset/ip_set_hash_net.c 		ip = ntohl(h->next.ip);
next              266 net/netfilter/ipset/ip_set_hash_net.c hash_net6_data_next(struct hash_net6_elem *next,
next              123 net/netfilter/ipset/ip_set_hash_netiface.c hash_netiface4_data_next(struct hash_netiface4_elem *next,
next              126 net/netfilter/ipset/ip_set_hash_netiface.c 	next->ip = d->ip;
next              253 net/netfilter/ipset/ip_set_hash_netiface.c 		ip = ntohl(h->next.ip);
next              346 net/netfilter/ipset/ip_set_hash_netiface.c hash_netiface6_data_next(struct hash_netiface6_elem *next,
next              121 net/netfilter/ipset/ip_set_hash_netnet.c hash_netnet4_data_next(struct hash_netnet4_elem *next,
next              124 net/netfilter/ipset/ip_set_hash_netnet.c 	next->ipcmp = d->ipcmp;
next              248 net/netfilter/ipset/ip_set_hash_netnet.c 		ip = ntohl(h->next.ip[0]);
next              249 net/netfilter/ipset/ip_set_hash_netnet.c 		ip2 = ntohl(h->next.ip[1]);
next              352 net/netfilter/ipset/ip_set_hash_netnet.c hash_netnet6_data_next(struct hash_netnet6_elem *next,
next              116 net/netfilter/ipset/ip_set_hash_netport.c hash_netport4_data_next(struct hash_netport4_elem *next,
next              119 net/netfilter/ipset/ip_set_hash_netport.c 	next->ip = d->ip;
next              120 net/netfilter/ipset/ip_set_hash_netport.c 	next->port = d->port;
next              239 net/netfilter/ipset/ip_set_hash_netport.c 		ip = ntohl(h->next.ip);
next              240 net/netfilter/ipset/ip_set_hash_netport.c 		p = ntohs(h->next.port);
next              329 net/netfilter/ipset/ip_set_hash_netport.c hash_netport6_data_next(struct hash_netport6_elem *next,
next              332 net/netfilter/ipset/ip_set_hash_netport.c 	next->port = d->port;
next              442 net/netfilter/ipset/ip_set_hash_netport.c 		port = ntohs(h->next.port);
next              130 net/netfilter/ipset/ip_set_hash_netportnet.c hash_netportnet4_data_next(struct hash_netportnet4_elem *next,
next              133 net/netfilter/ipset/ip_set_hash_netportnet.c 	next->ipcmp = d->ipcmp;
next              134 net/netfilter/ipset/ip_set_hash_netportnet.c 	next->port = d->port;
next              288 net/netfilter/ipset/ip_set_hash_netportnet.c 		ip = ntohl(h->next.ip[0]);
next              289 net/netfilter/ipset/ip_set_hash_netportnet.c 		p = ntohs(h->next.port);
next              290 net/netfilter/ipset/ip_set_hash_netportnet.c 		ip2 = ntohl(h->next.ip[1]);
next              406 net/netfilter/ipset/ip_set_hash_netportnet.c hash_netportnet6_data_next(struct hash_netportnet6_elem *next,
next              409 net/netfilter/ipset/ip_set_hash_netportnet.c 	next->port = d->port;
next              539 net/netfilter/ipset/ip_set_hash_netportnet.c 		port = ntohs(h->next.port);
next              190 net/netfilter/ipset/ip_set_list_set.c 	struct set_elem *e, *next, *prev = NULL;
next              205 net/netfilter/ipset/ip_set_list_set.c 			next = list_next_entry(e, list);
next              207 net/netfilter/ipset/ip_set_list_set.c 			      next->id == d->refid;
next              237 net/netfilter/ipset/ip_set_list_set.c 	struct set_elem *e, *n, *prev, *next;
next              241 net/netfilter/ipset/ip_set_list_set.c 	n = prev = next = NULL;
next              251 net/netfilter/ipset/ip_set_list_set.c 			next = e;
next              257 net/netfilter/ipset/ip_set_list_set.c 	if ((d->before > 0 && !next) ||
next              280 net/netfilter/ipset/ip_set_list_set.c 		if (!list_is_last(&next->list, &map->members))
next              281 net/netfilter/ipset/ip_set_list_set.c 			n = list_next_entry(next, list);
next              302 net/netfilter/ipset/ip_set_list_set.c 	else if (next)
next              303 net/netfilter/ipset/ip_set_list_set.c 		list_add_tail_rcu(&e->list, &next->list);
next              319 net/netfilter/ipset/ip_set_list_set.c 	struct set_elem *e, *next, *prev = NULL;
next              331 net/netfilter/ipset/ip_set_list_set.c 			next = list_next_entry(e, list);
next              333 net/netfilter/ipset/ip_set_list_set.c 			    next->id != d->refid)
next              557 net/netfilter/ipvs/ip_vs_app.c 	if ((e = inc->a_list.next) != &app->incs_list)
next              561 net/netfilter/ipvs/ip_vs_app.c 	for (e = app->a_list.next; e != &ipvs->app_list; e = e->next) {
next              593 net/netfilter/ipvs/ip_vs_app.c 	.next  = ip_vs_app_seq_next,
next             1143 net/netfilter/ipvs/ip_vs_conn.c 	.next  = ip_vs_conn_seq_next,
next             1209 net/netfilter/ipvs/ip_vs_conn.c 	.next  = ip_vs_conn_seq_next,
next             1236 net/netfilter/ipvs/ip_vs_ctl.c 	struct ip_vs_dest *dest, *next;
next             1240 net/netfilter/ipvs/ip_vs_ctl.c 	list_for_each_entry_safe(dest, next, &ipvs->dest_trash, t_list) {
next             2197 net/netfilter/ipvs/ip_vs_ctl.c 	.next  = ip_vs_info_seq_next,
next              117 net/netfilter/ipvs/ip_vs_dh.c 				p = p->next;
next              123 net/netfilter/ipvs/ip_vs_dh.c 			p = p->next;
next              233 net/netfilter/ipvs/ip_vs_lblc.c 	struct hlist_node *next;
next              239 net/netfilter/ipvs/ip_vs_lblc.c 		hlist_for_each_entry_safe(en, next, &tbl->bucket[i], list) {
next              260 net/netfilter/ipvs/ip_vs_lblc.c 	struct hlist_node *next;
next              268 net/netfilter/ipvs/ip_vs_lblc.c 		hlist_for_each_entry_safe(en, next, &tbl->bucket[j], list) {
next              302 net/netfilter/ipvs/ip_vs_lblc.c 	struct hlist_node *next;
next              324 net/netfilter/ipvs/ip_vs_lblc.c 		hlist_for_each_entry_safe(en, next, &tbl->bucket[j], list) {
next              400 net/netfilter/ipvs/ip_vs_lblcr.c 	struct hlist_node *next;
next              405 net/netfilter/ipvs/ip_vs_lblcr.c 		hlist_for_each_entry_safe(en, next, &tbl->bucket[i], list) {
next              427 net/netfilter/ipvs/ip_vs_lblcr.c 	struct hlist_node *next;
next              433 net/netfilter/ipvs/ip_vs_lblcr.c 		hlist_for_each_entry_safe(en, next, &tbl->bucket[j], list) {
next              466 net/netfilter/ipvs/ip_vs_lblcr.c 	struct hlist_node *next;
next              488 net/netfilter/ipvs/ip_vs_lblcr.c 		hlist_for_each_entry_safe(en, next, &tbl->bucket[j], list) {
next              140 net/netfilter/ipvs/ip_vs_mh.c 	while ((p = p->next) != &svc->destinations) {
next              187 net/netfilter/ipvs/ip_vs_mh.c 			p = p->next;
next              193 net/netfilter/ipvs/ip_vs_mh.c 				p = p->next;
next              223 net/netfilter/ipvs/ip_vs_mh.c 				p = p->next;
next               53 net/netfilter/ipvs/ip_vs_proto.c 	pp->next = ip_vs_proto_table[hash];
next               76 net/netfilter/ipvs/ip_vs_proto.c 	pd->next = ipvs->proto_data_table[hash];
next               84 net/netfilter/ipvs/ip_vs_proto.c 			ipvs->proto_data_table[hash] = pd->next;
next              102 net/netfilter/ipvs/ip_vs_proto.c 	for (; *pp_p; pp_p = &(*pp_p)->next) {
next              104 net/netfilter/ipvs/ip_vs_proto.c 			*pp_p = pp->next;
next              124 net/netfilter/ipvs/ip_vs_proto.c 	for (; *pd_p; pd_p = &(*pd_p)->next) {
next              126 net/netfilter/ipvs/ip_vs_proto.c 			*pd_p = pd->next;
next              145 net/netfilter/ipvs/ip_vs_proto.c 	for (pp = ip_vs_proto_table[hash]; pp; pp = pp->next) {
next              163 net/netfilter/ipvs/ip_vs_proto.c 	for (pd = ipvs->proto_data_table[hash]; pd; pd = pd->next) {
next              181 net/netfilter/ipvs/ip_vs_proto.c 		for (pd = ipvs->proto_data_table[i]; pd; pd = pd->next) {
next               43 net/netfilter/ipvs/ip_vs_rr.c 		svc->sched_data = p->next->prev;
next              184 net/netfilter/ipvs/ip_vs_sh.c 				p = p->next;
next              196 net/netfilter/ipvs/ip_vs_sh.c 				p = p->next;
next              316 net/netfilter/ipvs/ip_vs_sync.c 		sb = list_entry(ms->sync_queue.next, struct ip_vs_sync_buff,
next              469 net/netfilter/nf_conncount.c 		goto next; /* do not bother */
next              488 net/netfilter/nf_conncount.c next:
next              216 net/netfilter/nf_conntrack_expect.c 	struct hlist_node *next;
next              223 net/netfilter/nf_conntrack_expect.c 	hlist_for_each_entry_safe(exp, next, &help->expectations, lnode) {
next              422 net/netfilter/nf_conntrack_expect.c 	struct hlist_node *next;
next              431 net/netfilter/nf_conntrack_expect.c 	hlist_for_each_entry_safe(i, next, &nf_ct_expect_hash[h], hnode) {
next              494 net/netfilter/nf_conntrack_expect.c 	const struct hlist_node *next;
next              500 net/netfilter/nf_conntrack_expect.c 		hlist_for_each_entry_safe(exp, next,
next              520 net/netfilter/nf_conntrack_expect.c 	const struct hlist_node *next;
next              526 net/netfilter/nf_conntrack_expect.c 		hlist_for_each_entry_safe(exp, next,
next              652 net/netfilter/nf_conntrack_expect.c 	.next = exp_seq_next,
next              808 net/netfilter/nf_conntrack_sip.c 	struct hlist_node *next;
next              812 net/netfilter/nf_conntrack_sip.c 	hlist_for_each_entry_safe(exp, next, &help->expectations, lnode) {
next              832 net/netfilter/nf_conntrack_sip.c 	struct hlist_node *next;
next              835 net/netfilter/nf_conntrack_sip.c 	hlist_for_each_entry_safe(exp, next, &help->expectations, lnode) {
next              380 net/netfilter/nf_conntrack_standalone.c 	.next  = ct_seq_next,
next              459 net/netfilter/nf_conntrack_standalone.c 	.next	= ct_cpu_seq_next,
next              394 net/netfilter/nf_log.c 	.next	= seq_next,
next              191 net/netfilter/nf_nat_sip.c 				goto next;
next              196 net/netfilter/nf_nat_sip.c 				goto next;
next              259 net/netfilter/nf_nat_sip.c next:
next              302 net/netfilter/nf_synproxy_core.c 	.next		= synproxy_cpu_seq_next,
next             2350 net/netfilter/nf_tables_api.c 	const struct nft_expr *expr, *next;
next             2381 net/netfilter/nf_tables_api.c 	nft_rule_for_each_expr(expr, next, rule) {
next             2637 net/netfilter/nf_tables_api.c 	struct nft_expr *expr, *next;
next             2645 net/netfilter/nf_tables_api.c 		next = nft_expr_next(expr);
next             2647 net/netfilter/nf_tables_api.c 		expr = next;
next             6590 net/netfilter/nf_tables_api.c 	struct nft_trans *trans, *next;
next             6602 net/netfilter/nf_tables_api.c 	list_for_each_entry_safe(trans, next, &head, list) {
next             6641 net/netfilter/nf_tables_api.c 	struct nft_trans *trans, *next;
next             6643 net/netfilter/nf_tables_api.c 	list_for_each_entry_safe(trans, next, &net->nft.commit_list, list) {
next             6741 net/netfilter/nf_tables_api.c 	struct nft_module_request *req, *next;
next             6744 net/netfilter/nf_tables_api.c 	list_for_each_entry_safe(req, next, &net->nft.module_list, list) {
next             6786 net/netfilter/nf_tables_api.c 	struct nft_trans *trans, *next;
next             6806 net/netfilter/nf_tables_api.c 	list_for_each_entry_safe(trans, next, &net->nft.commit_list, list) {
next             6836 net/netfilter/nf_tables_api.c 	list_for_each_entry_safe(trans, next, &net->nft.commit_list, list) {
next             6972 net/netfilter/nf_tables_api.c 	struct nft_module_request *req, *next;
next             6977 net/netfilter/nf_tables_api.c 	list_for_each_entry_safe(req, next, &module_list, list) {
next             7016 net/netfilter/nf_tables_api.c 	struct nft_trans *trans, *next;
next             7019 net/netfilter/nf_tables_api.c 	list_for_each_entry_safe_reverse(trans, next, &net->nft.commit_list,
next             7130 net/netfilter/nf_tables_api.c 	list_for_each_entry_safe_reverse(trans, next,
next              199 net/netfilter/nf_tables_offload.c 	struct flow_block_cb *block_cb, *next;
next              201 net/netfilter/nf_tables_offload.c 	list_for_each_entry_safe(block_cb, next, &bo->cb_list, list) {
next              274 net/netfilter/nfnetlink.c 	struct nfnl_err *nfnl_err, *next;
next              276 net/netfilter/nfnetlink.c 	list_for_each_entry_safe(nfnl_err, next, err_list, head)
next              282 net/netfilter/nfnetlink.c 	struct nfnl_err *nfnl_err, *next;
next              284 net/netfilter/nfnetlink.c 	list_for_each_entry_safe(nfnl_err, next, err_list, head) {
next             1097 net/netfilter/nfnetlink_log.c 	.next	= seq_next,
next              249 net/netfilter/nfnetlink_queue.c 	struct nf_queue_entry *entry, *next;
next              252 net/netfilter/nfnetlink_queue.c 	list_for_each_entry_safe(entry, next, &queue->queue_list, list) {
next              755 net/netfilter/nfnetlink_queue.c 	if (skb->next == NULL) { /* last packet, no need to copy entry */
next              819 net/netfilter/nfnetlink_queue.c 		struct sk_buff *nskb = segs->next;
next             1431 net/netfilter/nfnetlink_queue.c 	h = h->next;
next             1489 net/netfilter/nfnetlink_queue.c 	.next	= seq_next,
next              618 net/netfilter/nft_set_hash.c 	struct hlist_node *next;
next              622 net/netfilter/nft_set_hash.c 		hlist_for_each_entry_safe(he, next, &priv->table[i], node) {
next             1523 net/netfilter/x_tables.c 	.next	= xt_table_seq_next,
next             1565 net/netfilter/x_tables.c 		trav->curr = trav->curr->next;
next             1575 net/netfilter/x_tables.c 		trav->curr = trav->curr->next;
next             1642 net/netfilter/x_tables.c 	.next	= xt_match_seq_next,
next             1676 net/netfilter/x_tables.c 	.next	= xt_target_seq_next,
next             1214 net/netfilter/xt_hashlimit.c 	.next  = dl_seq_next,
next             1221 net/netfilter/xt_hashlimit.c 	.next  = dl_seq_next,
next             1228 net/netfilter/xt_hashlimit.c 	.next  = dl_seq_next,
next              162 net/netfilter/xt_recent.c 	e = list_entry(t->lru_list.next, struct recent_entry, lru_list);
next              179 net/netfilter/xt_recent.c 		e = list_entry(t->lru_list.next, struct recent_entry, lru_list);
next              224 net/netfilter/xt_recent.c 	struct recent_entry *e, *next;
next              228 net/netfilter/xt_recent.c 		list_for_each_entry_safe(e, next, &t->iphash[i], list)
next              493 net/netfilter/xt_recent.c 	const struct list_head *head = e->list.next;
next              499 net/netfilter/xt_recent.c 		head = t->iphash[st->bucket].next;
next              533 net/netfilter/xt_recent.c 	.next		= recent_seq_next,
next               64 net/netlabel/netlabel_addrlist.h 		i = i->next;
next               83 net/netlabel/netlabel_addrlist.h 	for (iter = __af4list_valid((head)->next, head);		\
next               85 net/netlabel/netlabel_addrlist.h 	     iter = __af4list_valid(iter->list.next, head))
next               88 net/netlabel/netlabel_addrlist.h 	for (iter = __af4list_valid_rcu((head)->next, head);		\
next               90 net/netlabel/netlabel_addrlist.h 	     iter = __af4list_valid_rcu(iter->list.next, head))
next               93 net/netlabel/netlabel_addrlist.h 	for (iter = __af4list_valid((head)->next, head),		\
next               94 net/netlabel/netlabel_addrlist.h 		     tmp = __af4list_valid(iter->list.next, head);	\
next               96 net/netlabel/netlabel_addrlist.h 	     iter = tmp, tmp = __af4list_valid(iter->list.next, head))
next              131 net/netlabel/netlabel_addrlist.h 		i = i->next;
next              150 net/netlabel/netlabel_addrlist.h 	for (iter = __af6list_valid((head)->next, head);		\
next              152 net/netlabel/netlabel_addrlist.h 	     iter = __af6list_valid(iter->list.next, head))
next              155 net/netlabel/netlabel_addrlist.h 	for (iter = __af6list_valid_rcu((head)->next, head);		\
next              157 net/netlabel/netlabel_addrlist.h 	     iter = __af6list_valid_rcu(iter->list.next, head))
next              160 net/netlabel/netlabel_addrlist.h 	for (iter = __af6list_valid((head)->next, head),		\
next              161 net/netlabel/netlabel_addrlist.h 		     tmp = __af6list_valid(iter->list.next, head);	\
next              163 net/netlabel/netlabel_addrlist.h 	     iter = tmp, tmp = __af6list_valid(iter->list.next, head))
next              568 net/netlabel/netlabel_kapi.c 		iter = iter->next;
next              588 net/netlabel/netlabel_kapi.c 		iter->next = *catmap;
next              591 net/netlabel/netlabel_kapi.c 		iter->next = prev->next;
next              592 net/netlabel/netlabel_kapi.c 		prev->next = iter;
next              638 net/netlabel/netlabel_kapi.c 			if (iter->next != NULL) {
next              639 net/netlabel/netlabel_kapi.c 				iter = iter->next;
next              698 net/netlabel/netlabel_kapi.c 			if (iter->next == NULL)
next              701 net/netlabel/netlabel_kapi.c 			iter = iter->next;
next             2647 net/netlink/af_netlink.c 	.next   = netlink_seq_next,
next             1306 net/netrom/af_netrom.c 	.next = nr_info_next,
next              891 net/netrom/nr_route.c 	.next = nr_node_next,
next              945 net/netrom/nr_route.c 	.next = nr_neigh_next,
next             1389 net/nfc/llcp_core.c 			goto next;
next             1402 net/nfc/llcp_core.c next:
next              116 net/nsh/nsh.c  	for (skb = segs; skb; skb = skb->next) {
next              332 net/openvswitch/datapath.c 	} while ((skb = skb->next));
next              337 net/openvswitch/datapath.c 		nskb = skb->next;
next               45 net/openvswitch/flow_netlink.c 	const struct ovs_len_tbl *next;
next              390 net/openvswitch/flow_netlink.c 						.next = ovs_vxlan_ext_key_lens },
next              426 net/openvswitch/flow_netlink.c 				     .next = ovs_tunnel_key_lens, },
next              437 net/openvswitch/flow_netlink.c 				     .next = ovs_nsh_key_attr_lens, },
next             1728 net/openvswitch/flow_netlink.c 			nlattr_set(nla, val, tbl[nla_type(nla)].next ? : tbl);
next             3544 net/packet/af_packet.c 			*mlp = ml->next;
next             3547 net/packet/af_packet.c 			mlp = &ml->next;
next             3575 net/packet/af_packet.c 	for (ml = po->mclist; ml; ml = ml->next) {
next             3593 net/packet/af_packet.c 	i->next = po->mclist;
next             3597 net/packet/af_packet.c 		po->mclist = i->next;
next             3612 net/packet/af_packet.c 	for (mlp = &pkt_sk(sk)->mclist; (ml = *mlp) != NULL; mlp = &ml->next) {
next             3619 net/packet/af_packet.c 				*mlp = ml->next;
next             3644 net/packet/af_packet.c 		po->mclist = ml->next;
next             4619 net/packet/af_packet.c 	.next	= packet_seq_next,
next               48 net/packet/diag.c 	for (ml = po->mclist; ml; ml = ml->next) {
next              205 net/packet/diag.c 			goto next;
next              214 net/packet/diag.c next:
next                8 net/packet/internal.h 	struct packet_mclist	*next;
next              107 net/phonet/pep-gprs.c 		skb->next = skb_shinfo(skb)->frag_list;
next             1218 net/phonet/pep.c 	skb->next = skb_shinfo(skb)->frag_list;
next              612 net/phonet/socket.c 	.next = pn_sock_seq_next,
next              781 net/phonet/socket.c 	.next = pn_res_seq_next,
next              665 net/rds/af_rds.c 	WARN_ON((&rs->rs_item != rs->rs_item.next ||
next              127 net/rds/bind.c 		rs->rs_bound_node.next = NULL;
next              225 net/rds/ib_fmr.c 	struct rds_ib_mr *ibmr, *next;
next              242 net/rds/ib_fmr.c 	list_for_each_entry_safe(ibmr, next, list, unmap_list) {
next              360 net/rds/ib_frmr.c 	struct rds_ib_mr *ibmr, *next;
next              378 net/rds/ib_frmr.c 	list_for_each_entry_safe(ibmr, next, list, unmap_list) {
next              290 net/rds/ib_rdma.c 	struct llist_node *next;
next              295 net/rds/ib_rdma.c 		next = node->next;
next              298 net/rds/ib_rdma.c 		node = next;
next              315 net/rds/ib_rdma.c 	struct llist_node **next = nodes_head;
next              319 net/rds/ib_rdma.c 		*next = cur;
next              320 net/rds/ib_rdma.c 		next = &cur->next;
next              322 net/rds/ib_rdma.c 	*next = NULL;
next              410 net/rds/ib_rdma.c 			clean_nodes = clean_nodes->next;
next               58 net/rds/ib_recv.c 		recv->r_wr.next = NULL;
next              523 net/rds/ib_recv.c 			cache->ready = head->next;
next              543 net/rds/ib_recv.c 	frag = list_entry(ibinc->ii_frags.next, struct rds_page_frag, f_item);
next              548 net/rds/ib_recv.c 			frag = list_entry(frag->f_item.next,
next              797 net/rds/ib_recv.c 	frag = list_entry(ibinc->ii_frags.next, struct rds_page_frag, f_item);
next              831 net/rds/ib_recv.c 			frag = list_entry(frag->f_item.next,
next              630 net/rds/ib_send.c 		send->s_wr.next = NULL;
next              674 net/rds/ib_send.c 			 &send->s_wr, send->s_wr.num_sge, send->s_wr.next);
next              687 net/rds/ib_send.c 			prev->s_wr.next = &send->s_wr;
next              790 net/rds/ib_send.c 	send->s_atomic_wr.wr.next = NULL;
next              915 net/rds/ib_send.c 		send->s_rdma_wr.wr.next = NULL;
next              918 net/rds/ib_send.c 			prev->s_rdma_wr.wr.next = &send->s_rdma_wr.wr;
next              937 net/rds/ib_send.c 			send->s_rdma_wr.wr.next);
next              410 net/rds/recv.c 			*inc = list_entry(rs->rs_recv_queue.next,
next              475 net/rds/recv.c 		notifier = list_entry(rs->rs_notify_queue.next,
next              486 net/rds/recv.c 		notifier = list_entry(copy.next, struct rds_notifier, n_list);
next              613 net/rds/recv.c 		info = list_entry(q->zcookie_head.next,
next              255 net/rds/send.c 				rm = list_entry(cp->cp_send_queue.next,
next              614 net/rds/send.c 		rm = list_entry(messages->next, struct rds_message,
next              792 net/rds/send.c 		rm = list_entry(list.next, struct rds_message, m_sock_item);
next             1423 net/rose/af_rose.c 	.next = rose_info_next,
next               67 net/rose/rose_route.c 		rose_node = rose_node->next;
next               81 net/rose/rose_route.c 		rose_neigh = rose_neigh->next;
next              126 net/rose/rose_route.c 		rose_neigh->next = rose_neigh_list;
next              144 net/rose/rose_route.c 				rose_tmpn = rose_tmpn->next;
next              166 net/rose/rose_route.c 				rose_node->next = NULL;
next              168 net/rose/rose_route.c 				rose_tmpp->next = rose_node;
next              169 net/rose/rose_route.c 				rose_node->next = NULL;
next              173 net/rose/rose_route.c 				rose_node->next = rose_node_list;
next              176 net/rose/rose_route.c 				rose_tmpp->next = rose_node;
next              177 net/rose/rose_route.c 				rose_node->next = rose_tmpn;
next              207 net/rose/rose_route.c 		rose_node_list = rose_node->next;
next              212 net/rose/rose_route.c 	while (s != NULL && s->next != NULL) {
next              213 net/rose/rose_route.c 		if (s->next == rose_node) {
next              214 net/rose/rose_route.c 			s->next = rose_node->next;
next              219 net/rose/rose_route.c 		s = s->next;
next              236 net/rose/rose_route.c 		rose_neigh_list = rose_neigh->next;
next              244 net/rose/rose_route.c 	while (s != NULL && s->next != NULL) {
next              245 net/rose/rose_route.c 		if (s->next == rose_neigh) {
next              246 net/rose/rose_route.c 			s->next = rose_neigh->next;
next              254 net/rose/rose_route.c 		s = s->next;
next              272 net/rose/rose_route.c 		rose_route_list = rose_route->next;
next              277 net/rose/rose_route.c 	while (s != NULL && s->next != NULL) {
next              278 net/rose/rose_route.c 		if (s->next == rose_route) {
next              279 net/rose/rose_route.c 			s->next = rose_route->next;
next              284 net/rose/rose_route.c 		s = s->next;
next              308 net/rose/rose_route.c 		rose_node = rose_node->next;
next              322 net/rose/rose_route.c 		rose_neigh = rose_neigh->next;
next              395 net/rose/rose_route.c 	sn->next = rose_neigh_list;
next              416 net/rose/rose_route.c 		rose_node = rose_node->next;
next              434 net/rose/rose_route.c 	rose_node->next = rose_node_list;
next              460 net/rose/rose_route.c 		rose_node = rose_node->next;
next              488 net/rose/rose_route.c 		rose_neigh = rose_neigh->next;
next              497 net/rose/rose_route.c 			rose_node = rose_node->next;
next              538 net/rose/rose_route.c 		rose_route = rose_route->next;
next              565 net/rose/rose_route.c 		rose_node = rose_node->next;
next              572 net/rose/rose_route.c 		rose_neigh = rose_neigh->next;
next              663 net/rose/rose_route.c 	for (rose_route = rose_route_list; rose_route != NULL; rose_route = rose_route->next)
next              683 net/rose/rose_route.c 	for (node = rose_node_list; node != NULL; node = node->next) {
next              694 net/rose/rose_route.c 		for (node = rose_node_list; node != NULL; node = node->next) {
next              782 net/rose/rose_route.c 			s = rose_route->next;
next              800 net/rose/rose_route.c 		rose_route = rose_route->next;
next              819 net/rose/rose_route.c 		rose_neigh = rose_neigh->next;
next              840 net/rose/rose_route.c 	for (rose_neigh = rose_neigh_list; rose_neigh != NULL; rose_neigh = rose_neigh->next) {
next              890 net/rose/rose_route.c 		rose_neigh = rose_neigh->next;
next             1005 net/rose/rose_route.c 		rose_route = rose_route->next;
next             1039 net/rose/rose_route.c 		rose_route = rose_route->next;
next             1070 net/rose/rose_route.c 	rose_route->next = rose_route_list;
next             1100 net/rose/rose_route.c 	     rose_node = rose_node->next, ++i);
next             1110 net/rose/rose_route.c 		: ((struct rose_node *)v)->next;
next             1150 net/rose/rose_route.c 	.next = rose_node_next,
next             1166 net/rose/rose_route.c 	     rose_neigh = rose_neigh->next, ++i);
next             1176 net/rose/rose_route.c 		: ((struct rose_neigh *)v)->next;
next             1221 net/rose/rose_route.c 	.next = rose_neigh_next,
next             1237 net/rose/rose_route.c 	     rose_route = rose_route->next, ++i);
next             1247 net/rose/rose_route.c 		: ((struct rose_route *)v)->next;
next             1293 net/rose/rose_route.c 	.next = rose_route_next,
next             1310 net/rose/rose_route.c 		rose_neigh = rose_neigh->next;
next             1317 net/rose/rose_route.c 		rose_node = rose_node->next;
next             1324 net/rose/rose_route.c 		rose_route = rose_route->next;
next              501 net/rxrpc/call_accept.c 	call = list_entry(rx->to_be_accepted.next,
next              607 net/rxrpc/call_accept.c 	call = list_entry(rx->to_be_accepted.next,
next              301 net/rxrpc/call_event.c 	unsigned long now, next, t;
next              425 net/rxrpc/call_event.c 	next = call->expect_rx_by;
next              427 net/rxrpc/call_event.c #define set(T) { t = READ_ONCE(T); if (time_before(t, next)) next = t; }
next              438 net/rxrpc/call_event.c 	if (time_after_eq(now, next))
next              441 net/rxrpc/call_event.c 	rxrpc_reduce_call_timer(call, next, now, rxrpc_timer_restart);
next              475 net/rxrpc/call_object.c 	call->recvmsg_link.next = NULL;
next              514 net/rxrpc/call_object.c 		call = list_entry(rx->to_be_accepted.next,
next              522 net/rxrpc/call_object.c 		call = list_entry(rx->sock_calls.next,
next              631 net/rxrpc/call_object.c 			call = list_entry(rxnet->calls.next,
next              552 net/rxrpc/conn_client.c 	struct rxrpc_call *call = list_entry(conn->waiting_calls.next,
next              930 net/rxrpc/conn_client.c 		if (rxnet->idle_client_conns.next == &conn->cache_link &&
next              947 net/rxrpc/conn_client.c 	struct rxrpc_connection *next = NULL;
next              973 net/rxrpc/conn_client.c 			next = list_entry(rxnet->waiting_client_conns.next,
next              975 net/rxrpc/conn_client.c 			rxrpc_get_connection(next);
next              976 net/rxrpc/conn_client.c 			rxrpc_activate_conn(rxnet, next);
next              983 net/rxrpc/conn_client.c 	if (next)
next              984 net/rxrpc/conn_client.c 		rxrpc_activate_channels(next);
next              989 net/rxrpc/conn_client.c 	return next;
next             1035 net/rxrpc/conn_client.c 		conn = list_entry(rxnet->active_client_conns.next,
next             1093 net/rxrpc/conn_client.c next:
next             1099 net/rxrpc/conn_client.c 	conn = list_entry(rxnet->idle_client_conns.next,
next             1136 net/rxrpc/conn_client.c 	goto next;
next             1211 net/rxrpc/conn_client.c 		conn = list_entry(graveyard.next,
next              441 net/rxrpc/conn_object.c 		conn = list_entry(graveyard.next, struct rxrpc_connection,
next              223 net/rxrpc/input.c 		skb->next = list;
next              243 net/rxrpc/input.c 		list = skb->next;
next              157 net/rxrpc/key.c 	     pptoken = &(*pptoken)->next)
next              534 net/rxrpc/key.c 	     pptoken = &(*pptoken)->next)
next              781 net/rxrpc/key.c 		pp = &(*pp)->next;
next              800 net/rxrpc/key.c 	struct rxrpc_key_token *next;
next              802 net/rxrpc/key.c 	for (; token; token = next) {
next              803 net/rxrpc/key.c 		next = token->next;
next             1071 net/rxrpc/key.c 	for (token = key->payload.data[0]; token; token = token->next) {
next             1159 net/rxrpc/key.c 	for (token = key->payload.data[0]; token; token = token->next) {
next              236 net/rxrpc/local_object.c 	for (cursor = rxnet->local_endpoints.next;
next              238 net/rxrpc/local_object.c 	     cursor = cursor->next) {
next              314 net/rxrpc/peer_event.c 		peer = list_entry(collector->next,
next              124 net/rxrpc/proc.c 	.next   = rxrpc_call_seq_next,
next              207 net/rxrpc/proc.c 	.next   = rxrpc_connection_seq_next,
next              333 net/rxrpc/proc.c 	.next   = rxrpc_peer_seq_next,
next              501 net/rxrpc/recvmsg.c 	l = rx->recvmsg_q.next;
next              588 net/sched/cls_api.c 		tp_next = rcu_dereference_protected(tp->next, 1);
next              599 net/sched/cls_api.c 		tp_next = rcu_dereference_protected(tp->next, 1);
next             1005 net/sched/cls_api.c 		for (; tp; tp = tcf_chain_dereference(tp->next, chain))
next             1009 net/sched/cls_api.c 		tp = tcf_chain_dereference(tp->next, chain);
next             1480 net/sched/cls_api.c 	struct flow_block_cb *block_cb, *next;
next             1502 net/sched/cls_api.c 	list_for_each_entry_safe(block_cb, next, &bo->cb_list, list) {
next             1521 net/sched/cls_api.c 	struct flow_block_cb *block_cb, *next;
next             1525 net/sched/cls_api.c 	list_for_each_entry_safe(block_cb, next, &bo->cb_list, list) {
next             1573 net/sched/cls_api.c 	for (; tp; tp = rcu_dereference_bh(tp->next)) {
next             1626 net/sched/cls_api.c 	struct tcf_proto __rcu *next;
next             1645 net/sched/cls_api.c 	RCU_INIT_POINTER(tp->next, tcf_chain_tp_prev(chain, chain_info));
next             1655 net/sched/cls_api.c 	struct tcf_proto *next = tcf_chain_dereference(chain_info->next, chain);
next             1659 net/sched/cls_api.c 		tcf_chain0_head_change(chain, next);
next             1660 net/sched/cls_api.c 	RCU_INIT_POINTER(*chain_info->pprev, next);
next             1714 net/sched/cls_api.c 	struct tcf_proto *next;
next             1721 net/sched/cls_api.c 	     pprev = &tp_iter->next) {
next             1724 net/sched/cls_api.c 			chain_info.next = tp_iter->next;
next             1739 net/sched/cls_api.c 	next = tcf_chain_dereference(chain_info.next, chain);
next             1741 net/sched/cls_api.c 		tcf_chain0_head_change(chain, next);
next             1742 net/sched/cls_api.c 	RCU_INIT_POINTER(*chain_info.pprev, next);
next             1759 net/sched/cls_api.c 	     pprev = &tp->next) {
next             1773 net/sched/cls_api.c 		chain_info->next = tp->next;
next             1776 net/sched/cls_api.c 		chain_info->next = NULL;
next              595 net/sched/cls_flow.c 	struct flow_filter *f, *next;
next              597 net/sched/cls_flow.c 	list_for_each_entry_safe(f, next, &head->filters, list) {
next              567 net/sched/cls_flower.c 	struct cls_fl_filter *f, *next;
next              571 net/sched/cls_flower.c 		list_for_each_entry_safe(f, next, &mask->filters, list) {
next               34 net/sched/cls_fw.c 	struct fw_filter __rcu	*next;
next               62 net/sched/cls_fw.c 		     f = rcu_dereference_bh(f->next)) {
next               98 net/sched/cls_fw.c 	for (; f; f = rtnl_dereference(f->next)) {
next              143 net/sched/cls_fw.c 					 rtnl_dereference(f->next));
next              170 net/sched/cls_fw.c 	     fp = &pfp->next, pfp = rtnl_dereference(*fp)) {
next              172 net/sched/cls_fw.c 			RCU_INIT_POINTER(*fp, rtnl_dereference(f->next));
next              289 net/sched/cls_fw.c 		     fp = &pfp->next, pfp = rtnl_dereference(*fp))
next              293 net/sched/cls_fw.c 		RCU_INIT_POINTER(fnew->next, rtnl_dereference(pfp->next));
next              333 net/sched/cls_fw.c 	RCU_INIT_POINTER(f->next, head->ht[fw_hash(handle)]);
next              361 net/sched/cls_fw.c 		     f = rtnl_dereference(f->next)) {
next               47 net/sched/cls_route.c 	struct route4_filter __rcu	*next;
next              166 net/sched/cls_route.c 		     f = rcu_dereference_bh(f->next))
next              172 net/sched/cls_route.c 		     f = rcu_dereference_bh(f->next))
next              178 net/sched/cls_route.c 		     f = rcu_dereference_bh(f->next))
next              234 net/sched/cls_route.c 		     f = rtnl_dereference(f->next))
next              293 net/sched/cls_route.c 					struct route4_filter *next;
next              295 net/sched/cls_route.c 					next = rtnl_dereference(f->next);
next              296 net/sched/cls_route.c 					RCU_INIT_POINTER(b->ht[h2], next);
next              330 net/sched/cls_route.c 	     fp = &nf->next, nf = rtnl_dereference(*fp)) {
next              333 net/sched/cls_route.c 			RCU_INIT_POINTER(*fp, rtnl_dereference(f->next));
next              440 net/sched/cls_route.c 		     fp = rtnl_dereference(fp->next))
next              521 net/sched/cls_route.c 	     fp = &f1->next)
next              526 net/sched/cls_route.c 	rcu_assign_pointer(f->next, f1);
next              536 net/sched/cls_route.c 			     fp = &pfp->next, pfp = rtnl_dereference(*fp)) {
next              538 net/sched/cls_route.c 					rcu_assign_pointer(*fp, fold->next);
next              579 net/sched/cls_route.c 				     f = rtnl_dereference(f->next)) {
next               74 net/sched/cls_rsvp.h 	struct rsvp_session __rcu	*next;
next               86 net/sched/cls_rsvp.h 	struct rsvp_filter __rcu	*next;
next              171 net/sched/cls_rsvp.h 	     s = rcu_dereference_bh(s->next)) {
next              184 net/sched/cls_rsvp.h 			     f = rcu_dereference_bh(f->next)) {
next              209 net/sched/cls_rsvp.h 			     f = rcu_dereference_bh(f->next)) {
next              230 net/sched/cls_rsvp.h 	     s = rtnl_dereference(s->next)) {
next              232 net/sched/cls_rsvp.h 		     ins = &pins->next, pins = rtnl_dereference(*ins)) {
next              234 net/sched/cls_rsvp.h 				RCU_INIT_POINTER(n->next, pins->next);
next              259 net/sched/cls_rsvp.h 	     s = rtnl_dereference(s->next)) {
next              261 net/sched/cls_rsvp.h 		     f = rtnl_dereference(f->next)) {
next              324 net/sched/cls_rsvp.h 			RCU_INIT_POINTER(data->ht[h1], s->next);
next              330 net/sched/cls_rsvp.h 					rcu_assign_pointer(s->ht[h2], f->next);
next              353 net/sched/cls_rsvp.h 	     fp = &nfp->next, nfp = rtnl_dereference(*fp)) {
next              355 net/sched/cls_rsvp.h 			RCU_INIT_POINTER(*fp, f->next);
next              367 net/sched/cls_rsvp.h 			     sp = &nsp->next, nsp = rtnl_dereference(*sp)) {
next              369 net/sched/cls_rsvp.h 					RCU_INIT_POINTER(*sp, s->next);
next              430 net/sched/cls_rsvp.h 		     s = rtnl_dereference(s->next)) {
next              435 net/sched/cls_rsvp.h 				     f = rtnl_dereference(f->next)) {
next              586 net/sched/cls_rsvp.h 	     sp = &s->next) {
next              610 net/sched/cls_rsvp.h 			     fp = &nfp->next, nfp = rtnl_dereference(*fp)) {
next              616 net/sched/cls_rsvp.h 			RCU_INIT_POINTER(f->next, nfp);
next              639 net/sched/cls_rsvp.h 	     sp = &nsp->next, nsp = rtnl_dereference(*sp)) {
next              643 net/sched/cls_rsvp.h 	RCU_INIT_POINTER(s->next, nsp);
next              669 net/sched/cls_rsvp.h 		     s = rtnl_dereference(s->next)) {
next              674 net/sched/cls_rsvp.h 				     f = rtnl_dereference(f->next)) {
next               42 net/sched/cls_tcindex.c 	struct tcindex_filter __rcu *next;
next               93 net/sched/cls_tcindex.c 		     fp = &f->next, f = rcu_dereference_bh_rtnl(*fp))
next              215 net/sched/cls_tcindex.c 			     walk = &f->next, f = rtnl_dereference(*walk)) {
next              223 net/sched/cls_tcindex.c 		rcu_assign_pointer(*walk, rtnl_dereference(f->next));
next              461 net/sched/cls_tcindex.c 		f->next = NULL;
next              498 net/sched/cls_tcindex.c 		     fp = &nfp->next, nfp = rtnl_dereference(*fp))
next              554 net/sched/cls_tcindex.c 	struct tcindex_filter *f, *next;
next              574 net/sched/cls_tcindex.c 		for (f = rtnl_dereference(p->h[i]); f; f = next) {
next              575 net/sched/cls_tcindex.c 			next = rtnl_dereference(f->next);
next              617 net/sched/cls_tcindex.c 		struct tcindex_filter *f, *next;
next              620 net/sched/cls_tcindex.c 		for (f = rtnl_dereference(p->h[i]); f; f = next) {
next              621 net/sched/cls_tcindex.c 			next = rtnl_dereference(f->next);
next              666 net/sched/cls_tcindex.c 				     fp = &f->next, f = rtnl_dereference(*fp)) {
next               44 net/sched/cls_u32.c 	struct tc_u_knode __rcu	*next;
next               70 net/sched/cls_u32.c 	struct tc_u_hnode __rcu	*next;
next              135 net/sched/cls_u32.c 			n = rcu_dereference_bh(n->next);
next              141 net/sched/cls_u32.c 			n = rcu_dereference_bh(n->next);
next              159 net/sched/cls_u32.c 				n = rcu_dereference_bh(n->next);
next              175 net/sched/cls_u32.c 					n = rcu_dereference_bh(n->next);
next              183 net/sched/cls_u32.c 					n = rcu_dereference_bh(n->next);
next              189 net/sched/cls_u32.c 			n = rcu_dereference_bh(n->next);
next              260 net/sched/cls_u32.c 	     ht = rtnl_dereference(ht->next))
next              278 net/sched/cls_u32.c 	     n = rtnl_dereference(n->next))
next              380 net/sched/cls_u32.c 	RCU_INIT_POINTER(root_ht->next, tp_c->hlist);
next              454 net/sched/cls_u32.c 		     kp = &pkp->next, pkp = rtnl_dereference(*kp)) {
next              456 net/sched/cls_u32.c 				RCU_INIT_POINTER(*kp, key->next);
next              578 net/sched/cls_u32.c 					 rtnl_dereference(n->next));
next              605 net/sched/cls_u32.c 	     hn = &phn->next, phn = rtnl_dereference(*hn)) {
next              610 net/sched/cls_u32.c 			RCU_INIT_POINTER(*hn, ht->next);
next              637 net/sched/cls_u32.c 			RCU_INIT_POINTER(tp_c->hlist, ht->next);
next              783 net/sched/cls_u32.c 	     ins = &pins->next, pins = rtnl_dereference(*ins))
next              788 net/sched/cls_u32.c 	RCU_INIT_POINTER(n->next, pins->next);
next              805 net/sched/cls_u32.c 	RCU_INIT_POINTER(new->next, n->next);
next              972 net/sched/cls_u32.c 		RCU_INIT_POINTER(ht->next, tp_c->hlist);
next             1083 net/sched/cls_u32.c 		     ins = &pins->next, pins = rtnl_dereference(*ins))
next             1087 net/sched/cls_u32.c 		RCU_INIT_POINTER(n->next, pins);
next             1124 net/sched/cls_u32.c 	     ht = rtnl_dereference(ht->next)) {
next             1137 net/sched/cls_u32.c 			     n = rtnl_dereference(n->next)) {
next             1222 net/sched/cls_u32.c 	     ht = rtnl_dereference(ht->next)) {
next             1240 net/sched/cls_u32.c 			     n = rtnl_dereference(n->next)) {
next              134 net/sched/sch_api.c 	for (qp = &qdisc_base; (q = *qp) != NULL; qp = &q->next)
next              159 net/sched/sch_api.c 	qops->next = NULL;
next              178 net/sched/sch_api.c 	for (qp = &qdisc_base; (q = *qp) != NULL; qp = &q->next)
next              182 net/sched/sch_api.c 		*qp = q->next;
next              183 net/sched/sch_api.c 		q->next = NULL;
next              203 net/sched/sch_api.c 	for (q = qdisc_base; q; q = q->next) {
next              353 net/sched/sch_api.c 		for (q = qdisc_base; q; q = q->next) {
next              418 net/sched/sch_api.c 	for (rtab = qdisc_rtab_list; rtab; rtab = rtab->next) {
next              433 net/sched/sch_api.c 		rtab->next = qdisc_rtab_list;
next              451 net/sched/sch_api.c 	     rtabp = &rtab->next) {
next              453 net/sched/sch_api.c 			*rtabp = rtab->next;
next              660 net/sched/sch_api.c 	struct hlist_node *next;
next              679 net/sched/sch_api.c 		hlist_for_each_entry_safe(cl, next, &ohash[i], hnode) {
next              821 net/sched/sch_cake.c 		flow->head = skb->next;
next              835 net/sched/sch_cake.c 		flow->tail->next = skb;
next              837 net/sched/sch_cake.c 	skb->next = NULL;
next             1154 net/sched/sch_cake.c 	     skb_prev = skb_check, skb_check = skb_check->next) {
next             1249 net/sched/sch_cake.c 	if (elig_ack && aggressive && elig_ack->next == skb &&
next             1258 net/sched/sch_cake.c 		elig_ack_prev->next = elig_ack->next;
next             1260 net/sched/sch_cake.c 		flow->head = elig_ack->next;
next             1664 net/sched/sch_cake.c 				u64 next = \
next             1669 net/sched/sch_cake.c 				qdisc_watchdog_schedule_ns(&q->watchdog, next);
next             1687 net/sched/sch_cake.c 			nskb = segs->next;
next             1904 net/sched/sch_cake.c 		u64 next = min(ktime_to_ns(q->time_next_packet),
next             1908 net/sched/sch_cake.c 		qdisc_watchdog_schedule_ns(&q->watchdog, next);
next             2137 net/sched/sch_cake.c 		u64 next = min(ktime_to_ns(q->time_next_packet),
next             2140 net/sched/sch_cake.c 		qdisc_watchdog_schedule_ns(&q->watchdog, next);
next             2146 net/sched/sch_cake.c 				ktime_t next = \
next             2151 net/sched/sch_cake.c 							   ktime_to_ns(next));
next             2966 net/sched/sch_cake.c 				skb = skb->next;
next             1450 net/sched/sch_cbq.c 	struct hlist_node *next;
next             1469 net/sched/sch_cbq.c 		hlist_for_each_entry_safe(cl, next, &q->clhash.hash[h],
next             1791 net/sched/sch_cbq.c 	.next		=	NULL,
next              454 net/sched/sch_drr.c 	struct hlist_node *next;
next              460 net/sched/sch_drr.c 		hlist_for_each_entry_safe(cl, next, &q->clhash.hash[i],
next              494 net/sched/sch_dsmark.c 	.next		=	NULL,
next              124 net/sched/sch_etf.c 	ktime_t next;
next              131 net/sched/sch_etf.c 	next = ktime_sub_ns(skb->tstamp, q->delta);
next              132 net/sched/sch_etf.c 	qdisc_watchdog_schedule_ns(&q->watchdog, ktime_to_ns(next));
next              216 net/sched/sch_etf.c 		skb->next = NULL;
next              240 net/sched/sch_etf.c 	skb->next = NULL;
next              257 net/sched/sch_etf.c 	ktime_t now, next;
next              281 net/sched/sch_etf.c 	next = ktime_sub_ns(skb->tstamp, q->delta);
next              284 net/sched/sch_etf.c 	if (ktime_after(now, next))
next               80 net/sched/sch_fq.c 	struct fq_flow *next;		/* next pointer in RR lists, or &detached */
next              132 net/sched/sch_fq.c 	f->next = &detached;
next              138 net/sched/sch_fq.c 	return f->next == &detached;
next              143 net/sched/sch_fq.c 	return f->next == &throttled;
next              149 net/sched/sch_fq.c 		head->last->next = flow;
next              153 net/sched/sch_fq.c 	flow->next = NULL;
next              182 net/sched/sch_fq.c 	f->next = &throttled;
next              364 net/sched/sch_fq.c 		flow->head = skb->next;
next              399 net/sched/sch_fq.c 			flow->tail->next = skb;
next              401 net/sched/sch_fq.c 		skb->next = NULL;
next              516 net/sched/sch_fq.c 		head->first = f->next;
next              527 net/sched/sch_fq.c 			head->first = f->next;
next              541 net/sched/sch_fq.c 		head->first = f->next;
next              121 net/sched/sch_fq_codel.c 	flow->head = skb->next;
next              133 net/sched/sch_fq_codel.c 		flow->tail->next = skb;
next              135 net/sched/sch_fq_codel.c 	skb->next = NULL;
next              648 net/sched/sch_fq_codel.c 				skb = skb->next;
next              130 net/sched/sch_generic.c 		struct sk_buff *next = skb->next;
next              145 net/sched/sch_generic.c 		skb = next;
next              166 net/sched/sch_generic.c 		skb->next = nskb;
next              192 net/sched/sch_generic.c 		skb->next = nskb;
next              563 net/sched/sch_generic.c 		.next = (struct sk_buff *)&noop_qdisc.gso_skb,
next              569 net/sched/sch_generic.c 		.next = (struct sk_buff *)&noop_qdisc.skb_bad_txq,
next             1495 net/sched/sch_hfsc.c 	struct hlist_node *next;
next             1506 net/sched/sch_hfsc.c 		hlist_for_each_entry_safe(cl, next, &q->clhash.hash[i],
next              184 net/sched/sch_hhf.c 	struct hh_flow_state *flow, *next;
next              190 net/sched/sch_hhf.c 	list_for_each_entry_safe(flow, next, head, flowchain) {
next              333 net/sched/sch_hhf.c 	bucket->head = skb->next;
next              344 net/sched/sch_hhf.c 		bucket->tail->next = skb;
next              346 net/sched/sch_hhf.c 	skb->next = NULL;
next              485 net/sched/sch_hhf.c 		struct hh_flow_state *flow, *next;
next              490 net/sched/sch_hhf.c 		list_for_each_entry_safe(flow, next, head, flowchain) {
next              835 net/sched/sch_htb.c next:
next              845 net/sched/sch_htb.c 			struct htb_class *next;
next              852 net/sched/sch_htb.c 			next = htb_lookup_leaf(hprio, prio);
next              855 net/sched/sch_htb.c 				start = next;
next              856 net/sched/sch_htb.c 			cl = next;
next              857 net/sched/sch_htb.c 			goto next;
next             1221 net/sched/sch_htb.c 	struct hlist_node *next;
next             1241 net/sched/sch_htb.c 		hlist_for_each_entry_safe(cl, next, &q->clhash.hash[i],
next              393 net/sched/sch_multiq.c 	.next		=	NULL,
next              385 net/sched/sch_netem.c 			q->t_tail->next = nskb;
next              499 net/sched/sch_netem.c 			segs = skb->next;
next              522 net/sched/sch_netem.c 		skb->next = segs;
next              603 net/sched/sch_netem.c 			skb2 = segs->next;
next              668 net/sched/sch_netem.c 		q->t_head = skb->next;
next              703 net/sched/sch_netem.c 			skb->next = NULL;
next              416 net/sched/sch_prio.c 	.next		=	NULL,
next              144 net/sched/sch_qfq.c 	struct hlist_node next;	/* Link for the slot list. */
next              751 net/sched/sch_qfq.c 	struct qfq_group *next;
next              754 net/sched/sch_qfq.c 		next = qfq_ffs(q, mask);
next              755 net/sched/sch_qfq.c 		if (qfq_gt(grp->F, next->F))
next              779 net/sched/sch_qfq.c 	struct qfq_group *next;
next              782 net/sched/sch_qfq.c 		next = qfq_ffs(q, mask);
next              783 net/sched/sch_qfq.c 		if (!qfq_gt(next->F, old_F))
next              892 net/sched/sch_qfq.c 	hlist_add_head(&agg->next, &grp->slots[i]);
next              900 net/sched/sch_qfq.c 			   struct qfq_aggregate, next);
next              911 net/sched/sch_qfq.c 	hlist_del(&agg->next);
next             1044 net/sched/sch_qfq.c 			struct qfq_group *next = qfq_ffs(q, mask);
next             1045 net/sched/sch_qfq.c 			if (qfq_gt(roundedF, next->F)) {
next             1046 net/sched/sch_qfq.c 				if (qfq_gt(limit, next->F))
next             1047 net/sched/sch_qfq.c 					agg->S = next->F;
next             1342 net/sched/sch_qfq.c 	hlist_del(&agg->next);
next             1473 net/sched/sch_qfq.c 	struct hlist_node *next;
next             1479 net/sched/sch_qfq.c 		hlist_for_each_entry_safe(cl, next, &q->clhash.hash[i],
next               96 net/sched/sch_sfq.c 	sfq_index	next;
next              104 net/sched/sch_sfq.c 	sfq_index	next; /* next slot in sfq RR chain */
next              210 net/sched/sch_sfq.c 	n = q->dep[qlen].next;
next              212 net/sched/sch_sfq.c 	slot->dep.next = n;
next              215 net/sched/sch_sfq.c 	q->dep[qlen].next = x;		/* sfq_dep_head(q, p)->next = x */
next              221 net/sched/sch_sfq.c 		n = q->slots[x].dep.next;	\
next              223 net/sched/sch_sfq.c 		sfq_dep_head(q, p)->next = n;	\
next              262 net/sched/sch_sfq.c 	skb->prev->next = (struct sk_buff *)slot;
next              263 net/sched/sch_sfq.c 	skb->next = skb->prev = NULL;
next              272 net/sched/sch_sfq.c 	slot->skblist_next = skb->next;
next              273 net/sched/sch_sfq.c 	skb->next->prev = (struct sk_buff *)slot;
next              274 net/sched/sch_sfq.c 	skb->next = skb->prev = NULL;
next              288 net/sched/sch_sfq.c 	skb->next = (struct sk_buff *)slot;
next              289 net/sched/sch_sfq.c 	slot->skblist_prev->next = skb;
next              303 net/sched/sch_sfq.c 		x = q->dep[d].next;
next              318 net/sched/sch_sfq.c 		x = q->tail->next;
next              320 net/sched/sch_sfq.c 		q->tail->next = slot->next;
next              368 net/sched/sch_sfq.c 		x = q->dep[0].next; /* get a free slot */
next              448 net/sched/sch_sfq.c 			slot->next = x;
next              450 net/sched/sch_sfq.c 			slot->next = q->tail->next;
next              451 net/sched/sch_sfq.c 			q->tail->next = x;
next              492 net/sched/sch_sfq.c 	a = q->tail->next;
next              508 net/sched/sch_sfq.c 		next_a = slot->next;
next              513 net/sched/sch_sfq.c 		q->tail->next = next_a;
next              568 net/sched/sch_sfq.c 			x = q->dep[0].next; /* get a free slot */
next              592 net/sched/sch_sfq.c 				slot->next = x;
next              594 net/sched/sch_sfq.c 				slot->next = q->tail->next;
next              595 net/sched/sch_sfq.c 				q->tail->next = x;
next              746 net/sched/sch_sfq.c 		q->dep[i].next = i + SFQ_MAX_FLOWS;
next              695 net/sched/sch_taprio.c 	struct sched_entry *entry, *next;
next              717 net/sched/sch_taprio.c 		next = list_first_entry(&oper->entries, struct sched_entry,
next              719 net/sched/sch_taprio.c 		close_time = next->close_time;
next              724 net/sched/sch_taprio.c 		next = list_first_entry(&oper->entries, struct sched_entry,
next              729 net/sched/sch_taprio.c 		next = list_next_entry(entry, list);
next              732 net/sched/sch_taprio.c 	close_time = ktime_add_ns(entry->close_time, next->interval);
next              743 net/sched/sch_taprio.c 	next->close_time = close_time;
next              744 net/sched/sch_taprio.c 	taprio_set_budget(q, next);
next              747 net/sched/sch_taprio.c 	rcu_assign_pointer(q->current_entry, next);
next              159 net/sched/sch_tbf.c 		nskb = segs->next;
next              533 net/sched/sch_tbf.c 	.next		=	NULL,
next               64 net/sched/sch_teql.c 	struct Qdisc *next;
next               69 net/sched/sch_teql.c #define NEXT_SLAVE(q) (((struct teql_sched_data *)qdisc_priv(q))->next)
next              203 net/sched/sch_teql.c 		q->next = NEXT_SLAVE(m->slaves);
next              206 net/sched/sch_teql.c 		q->next = sch;
next              499 net/sctp/associola.c 	pos = asoc->peer.transport_addr_list.next;
next              281 net/sctp/bind_addr.c 			goto next;
next              290 net/sctp/bind_addr.c next:
next              340 net/sctp/bind_addr.c 				goto next;
next              345 net/sctp/bind_addr.c next:
next              543 net/sctp/bind_addr.c 		addr = list_entry(bp->address_list.next,
next               34 net/sctp/diag.c 	laddr = list_entry(asoc->base.bind_addr.address_list.next,
next              309 net/sctp/diag.c 			goto next;
next              313 net/sctp/diag.c 			goto next;
next              316 net/sctp/diag.c 			goto next;
next              338 net/sctp/diag.c next:
next              356 net/sctp/diag.c 		list_entry(ep->asocs.next, struct sctp_association, asocs);
next              383 net/sctp/diag.c 		goto next;
next              386 net/sctp/diag.c 		goto next;
next              390 net/sctp/diag.c 		goto next;
next              394 net/sctp/diag.c 		goto next;
next              398 net/sctp/diag.c 		goto next;
next              408 net/sctp/diag.c next:
next              128 net/sctp/inqueue.c 			if (chunk->skb->next) {
next              129 net/sctp/inqueue.c 				chunk->skb = chunk->skb->next;
next               91 net/sctp/objcnt.c 	.next  = sctp_objcnt_seq_next,
next               73 net/sctp/offload.c 		for (skb = segs; skb; skb = skb->next) {
next              395 net/sctp/output.c 		SCTP_OUTPUT_CB(head)->last->next = skb;
next              559 net/sctp/output.c 	chunk = list_entry(packet->chunk_list.next, struct sctp_chunk, list);
next              195 net/sctp/proc.c 	.next  = sctp_eps_seq_next,
next              287 net/sctp/proc.c 	.next  = sctp_transport_seq_next,
next              364 net/sctp/proc.c 	.next  = sctp_transport_seq_next,
next             3840 net/sctp/sm_statefuns.c 	entry = asoc->addip_chunk_list.next;
next              245 net/sctp/socket.c 			asoc = list_entry(sctp_sk(sk)->ep->asocs.next,
next              602 net/sctp/socket.c 		p = bp->address_list.next;
next             4920 net/sctp/socket.c 	asoc = list_entry(ep->asocs.next, struct sctp_association, asocs);
next             5204 net/sctp/socket.c 		asoc = list_entry(ep->asocs.next,
next             6308 net/sctp/socket.c 		addr = list_entry(bp->address_list.next,
next             8203 net/sctp/socket.c 					goto next;
next             8205 net/sctp/socket.c 		next:
next               68 net/sctp/stream_sched.c 		ch = list_entry(stream->out_curr->ext->outq.next,
next               71 net/sctp/stream_sched.c 		entry = q->out_chunk_list.next;
next               39 net/sctp/stream_sched_prio.c 	p->next = NULL;
next               84 net/sctp/stream_sched_prio.c 	pos = p->next->prio_list.next;
next               86 net/sctp/stream_sched_prio.c 		pos = pos->next;
next               87 net/sctp/stream_sched_prio.c 	p->next = list_entry(pos, struct sctp_stream_out_ext, prio_list);
next              100 net/sctp/stream_sched_prio.c 		if (prio_head->next == soute)
next              110 net/sctp/stream_sched_prio.c 			prio_head->next = NULL;
next              132 net/sctp/stream_sched_prio.c 	if (prio_head->next) {
next              133 net/sctp/stream_sched_prio.c 		list_add(&soute->prio_list, prio_head->next->prio_list.prev);
next              138 net/sctp/stream_sched_prio.c 	prio_head->next = soute;
next              263 net/sctp/stream_sched_prio.c 		prio = list_entry(stream->prio_list.next,
next              265 net/sctp/stream_sched_prio.c 		soute = prio->next;
next              267 net/sctp/stream_sched_prio.c 	ch = list_entry(soute->outq.next, struct sctp_chunk, stream_list);
next               31 net/sctp/stream_sched_rr.c 	pos = stream->rr_next->rr_list.next;
next               33 net/sctp/stream_sched_rr.c 		pos = pos->next;
next              126 net/sctp/stream_sched_rr.c 	ch = list_entry(soute->outq.next, struct sctp_chunk, stream_list);
next              333 net/sctp/ulpqueue.c 		pos = f_frag->next;
next              336 net/sctp/ulpqueue.c 	for (last = list; list; last = list, list = list->next)
next              343 net/sctp/ulpqueue.c 		last->next = pos;
next              373 net/sctp/ulpqueue.c 		pnext = pos->next;
next              385 net/sctp/ulpqueue.c 		pos->next = pnext;
next              990 net/sctp/ulpqueue.c 		for (last = flist; flist; flist = flist->next) {
next              575 net/smc/smc_core.c 		lower = lower->next;
next              725 net/smc/smc_pnet.c 		lower = lower->next;
next              480 net/smc/smc_wr.c 		lnk->wr_tx_ibs[i].next = NULL;
next              498 net/smc/smc_wr.c 		lnk->wr_rx_ibs[i].next = NULL;
next              502 net/smc/smc_wr.c 	lnk->wr_reg.wr.next = NULL;
next              155 net/strparser/strparser.c 				if (WARN_ON(head->next)) {
next              167 net/strparser/strparser.c 				strp->skb_nextp = &head->next;
next              212 net/strparser/strparser.c 			strp->skb_nextp = &skb->next;
next              381 net/sunrpc/auth.c 		cred = list_entry(head->next, struct rpc_cred, cr_lru);
next              481 net/sunrpc/auth.c 	struct rpc_cred *cred, *next;
next              485 net/sunrpc/auth.c 	list_for_each_entry_safe(cred, next, &cred_unused, cr_lru) {
next              407 net/sunrpc/cache.c 	struct list_head *next;
next              415 net/sunrpc/cache.c 			next = current_detail->others.next;
next              417 net/sunrpc/cache.c 			next = cache_list.next;
next              418 net/sunrpc/cache.c 		if (next == &cache_list) {
next              423 net/sunrpc/cache.c 		current_detail = list_entry(next, struct cache_detail, others);
next              664 net/sunrpc/cache.c 			discard = list_entry(cache_defer_list.next,
next              719 net/sunrpc/cache.c 		dreq = list_entry(pending.next, struct cache_deferred_req, recent);
next              743 net/sunrpc/cache.c 		dreq = list_entry(pending.next, struct cache_deferred_req, recent);
next              812 net/sunrpc/cache.c 	while (rp->q.list.next != &cd->queue &&
next              813 net/sunrpc/cache.c 	       list_entry(rp->q.list.next, struct cache_queue, list)
next              815 net/sunrpc/cache.c 		struct list_head *next = rp->q.list.next;
next              816 net/sunrpc/cache.c 		list_move(&rp->q.list, next);
next              818 net/sunrpc/cache.c 	if (rp->q.list.next == &cd->queue) {
next              824 net/sunrpc/cache.c 	rq = container_of(rp->q.list.next, struct cache_request, q.list);
next              972 net/sunrpc/cache.c 	     cq = list_entry(cq->list.next, struct cache_queue, list))
next              998 net/sunrpc/cache.c 	     cq = list_entry(cq->list.next, struct cache_queue, list))
next             1047 net/sunrpc/cache.c 			     cq = list_entry(cq->list.next, struct cache_queue, list))
next             1094 net/sunrpc/cache.c 		cr = list_entry(dequeued.next, struct cache_request, q.list);
next             1354 net/sunrpc/cache.c 	else if (ch->cache_list.next == NULL) {
next             1425 net/sunrpc/cache.c 	.next	= cache_seq_next_rcu,
next               61 net/sunrpc/debugfs.c 	struct list_head *next = task->tk_task.next;
next               66 net/sunrpc/debugfs.c 	if (next == &clnt->cl_tasks)
next               68 net/sunrpc/debugfs.c 	return list_entry(next, struct rpc_task, tk_task);
next               81 net/sunrpc/debugfs.c 	.next	= tasks_next,
next               72 net/sunrpc/rpc_pipe.c 		msg = list_entry(head->next, struct rpc_pipe_msg, list);
next              298 net/sunrpc/rpc_pipe.c 			msg = list_entry(pipe->pipe.next,
next              666 net/sunrpc/rpc_pipe.c 			goto next;
next              676 net/sunrpc/rpc_pipe.c next:
next              149 net/sunrpc/sched.c 			task->u.tk_wait.list.next = q;
next              176 net/sunrpc/sched.c 		q = t->u.tk_wait.list.next;
next              267 net/sunrpc/stats.c 		struct rpc_clnt *next = clnt;
next              269 net/sunrpc/stats.c 			_add_rpc_iostats(&stats, &next->cl_metrics[op]);
next              270 net/sunrpc/stats.c 			if (next == next->cl_parent)
next              272 net/sunrpc/stats.c 			next = next->cl_parent;
next              273 net/sunrpc/stats.c 		} while (next);
next              696 net/sunrpc/svc.c 		rqstp = list_entry(pool->sp_all_threads.next, struct svc_rqst, rq_all);
next              932 net/sunrpc/svc_xprt.c 	struct list_head *le, *next;
next              943 net/sunrpc/svc_xprt.c 	list_for_each_safe(le, next, &serv->sv_tempsocks) {
next              974 net/sunrpc/svc_xprt.c 	struct list_head *le, *next;
next              978 net/sunrpc/svc_xprt.c 	list_for_each_safe(le, next, &serv->sv_tempsocks) {
next              989 net/sunrpc/svc_xprt.c 		le = to_be_closed.next;
next             1246 net/sunrpc/svc_xprt.c 		dr = list_entry(xprt->xpt_deferred.next,
next             1420 net/sunrpc/svc_xprt.c 	.next	= svc_pool_stats_next,
next             1307 net/sunrpc/xprt.c 			struct rpc_rqst *next = list_first_entry(&req->rq_xmit2,
next             1310 net/sunrpc/xprt.c 			list_add_tail(&next->rq_xmit, &next->rq_xprt->xmit_queue);
next             1504 net/sunrpc/xprt.c 	struct rpc_rqst *next, *req = task->tk_rqstp;
next             1510 net/sunrpc/xprt.c 		next = list_first_entry(&xprt->xmit_queue,
next             1512 net/sunrpc/xprt.c 		xprt_pin_rqst(next);
next             1514 net/sunrpc/xprt.c 		status = xprt_request_transmit(next, task);
next             1515 net/sunrpc/xprt.c 		if (status == -EBADMSG && next != req)
next             1519 net/sunrpc/xprt.c 		xprt_unpin_rqst(next);
next             1593 net/sunrpc/xprt.c 		req = list_entry(xprt->free.next, struct rpc_rqst, rq_list);
next              435 net/sunrpc/xprtrdma/frwr_ops.c 		frwr->fr_regwr.wr.next = post_wr;
next              549 net/sunrpc/xprtrdma/frwr_ops.c 		last->next = NULL;
next              558 net/sunrpc/xprtrdma/frwr_ops.c 		prev = &last->next;
next              591 net/sunrpc/xprtrdma/frwr_ops.c 		bad_wr = bad_wr->next;
next              652 net/sunrpc/xprtrdma/frwr_ops.c 		last->next = NULL;
next              661 net/sunrpc/xprtrdma/frwr_ops.c 		prev = &last->next;
next              686 net/sunrpc/xprtrdma/frwr_ops.c 		bad_wr = bad_wr->next;
next              138 net/sunrpc/xprtrdma/svc_rdma_recvfrom.c 	ctxt->rc_recv_wr.next = NULL;
next              148 net/sunrpc/xprtrdma/svc_rdma_sendto.c 	ctxt->sc_send_wr.next = NULL;
next              404 net/sunrpc/xprtrdma/svc_rdma_transport.c 		newxprt = list_entry(listen_rdma->sc_accept_q.next,
next             1084 net/sunrpc/xprtrdma/verbs.c 	rep->rr_recv_wr.next = NULL;
next             1527 net/sunrpc/xprtrdma/verbs.c 		rep->rr_recv_wr.next = wr;
next             1534 net/sunrpc/xprtrdma/verbs.c 	for (i = wr; i; i = i->next) {
next             1553 net/sunrpc/xprtrdma/verbs.c 			wr = wr->next;
next             1564 net/sunrpc/xprtrdma/verbs.c 		i = i->next;
next               70 net/tipc/msg.c 		skb->next = NULL;
next              137 net/tipc/msg.c 	frag->next = NULL;
next              168 net/tipc/msg.c 			tail->next = frag;
next              684 net/tipc/msg.c 		skb->next = NULL;
next              724 net/tipc/msg.c 		frag->next = NULL;
next               48 net/tipc/name_distr.c 	struct list_head next;
next              239 net/tipc/name_distr.c 	list_for_each_entry_safe(e, tmp, &tn->dist_queue, next) {
next              242 net/tipc/name_distr.c 		list_del(&e->next);
next             2822 net/unix/af_unix.c 	.next   = unix_seq_next,
next              217 net/unix/diag.c 				goto next;
next              219 net/unix/diag.c 				goto next;
next              225 net/unix/diag.c next:
next               95 net/unix/garbage.c 	struct sk_buff *next;
next               98 net/unix/garbage.c 	skb_queue_walk_safe(&x->sk_receive_queue, skb, next) {
next              140 net/unix/garbage.c 		struct sk_buff *next;
next              148 net/unix/garbage.c 		skb_queue_walk_safe(&x->sk_receive_queue, skb, next) {
next              160 net/unix/garbage.c 			u = list_entry(embryos.next, struct unix_sock, link);
next              205 net/unix/garbage.c 	struct unix_sock *next;
next              232 net/unix/garbage.c 	list_for_each_entry_safe(u, next, &gc_inflight_list, link) {
next              262 net/unix/garbage.c 	while (cursor.next != &gc_candidates) {
next              263 net/unix/garbage.c 		u = list_entry(cursor.next, struct unix_sock, link);
next              288 net/unix/garbage.c 		u = list_entry(not_cycle_list.next, struct unix_sock, link);
next              395 net/vmw_vsock/af_vsock.c 	vconnected = list_entry(vlistener->accept_queue.next,
next               91 net/wireless/lib80211.c 	struct lib80211_crypt_data *entry, *next;
next               95 net/wireless/lib80211.c 	list_for_each_entry_safe(entry, next, &info->crypt_deinit_list, list) {
next             2255 net/wireless/scan.c 	const u8 *pos, *end, *next;
next             2269 net/wireless/scan.c 		next = pos + 2 + pos[1];
next             2270 net/wireless/scan.c 		while (next + 2 + next[1] - pos < IW_GENERIC_IE_MAX)
next             2271 net/wireless/scan.c 			next = next + 2 + next[1];
next             2275 net/wireless/scan.c 		iwe.u.data.length = next - pos;
next             2281 net/wireless/scan.c 		pos = next;
next              796 net/wireless/sme.c 	u8 *next;
next              842 net/wireless/sme.c 	next = ((u8 *)ev) + sizeof(*ev);
next              844 net/wireless/sme.c 		ev->cr.bssid = next;
next              846 net/wireless/sme.c 		next += ETH_ALEN;
next              849 net/wireless/sme.c 		ev->cr.req_ie = next;
next              853 net/wireless/sme.c 		next += params->req_ie_len;
next              856 net/wireless/sme.c 		ev->cr.resp_ie = next;
next              860 net/wireless/sme.c 		next += params->resp_ie_len;
next              863 net/wireless/sme.c 		ev->cr.fils.kek = next;
next              867 net/wireless/sme.c 		next += params->fils.kek_len;
next              870 net/wireless/sme.c 		ev->cr.fils.pmk = next;
next              874 net/wireless/sme.c 		next += params->fils.pmk_len;
next              877 net/wireless/sme.c 		ev->cr.fils.pmkid = next;
next              880 net/wireless/sme.c 		next += WLAN_PMKID_LEN;
next              963 net/wireless/sme.c 	u8 *next;
next              985 net/wireless/sme.c 	next = ((u8 *)ev) + sizeof(*ev);
next              987 net/wireless/sme.c 		ev->rm.req_ie = next;
next              990 net/wireless/sme.c 		next += info->req_ie_len;
next              993 net/wireless/sme.c 		ev->rm.resp_ie = next;
next              997 net/wireless/sme.c 		next += info->resp_ie_len;
next             1000 net/wireless/sme.c 		ev->rm.fils.kek = next;
next             1004 net/wireless/sme.c 		next += info->fils.kek_len;
next             1007 net/wireless/sme.c 		ev->rm.fils.pmk = next;
next             1011 net/wireless/sme.c 		next += info->fils.pmk_len;
next             1014 net/wireless/sme.c 		ev->rm.fils.pmkid = next;
next             1017 net/wireless/sme.c 		next += WLAN_PMKID_LEN;
next              124 net/wireless/wext-proc.c 	.next  = wireless_dev_seq_next,
next              283 net/x25/x25_link.c 	if (nb->node.next) {
next              150 net/x25/x25_proc.c 	.next   = x25_seq_route_next,
next              157 net/x25/x25_proc.c 	.next   = x25_seq_socket_next,
next              164 net/x25/x25_proc.c 	.next   = x25_seq_forward_next,
next               71 net/x25/x25_route.c 	if (rt->node.next) {
next              128 net/xfrm/xfrm_device.c 	if (!skb->next) {
next              152 net/xfrm/xfrm_device.c 		struct sk_buff *nskb = skb2->next;
next              164 net/xfrm/xfrm_device.c 			skb2->next = nskb;
next              167 net/xfrm/xfrm_device.c 			skb2->next = nskb;
next               57 net/xfrm/xfrm_interface.c 	for (xi = rcu_dereference(start); xi; xi = rcu_dereference(xi->next))
next              109 net/xfrm/xfrm_interface.c 	rcu_assign_pointer(xi->next , rtnl_dereference(*xip));
next              120 net/xfrm/xfrm_interface.c 	     xip = &iter->next) {
next              122 net/xfrm/xfrm_interface.c 			rcu_assign_pointer(*xip, xi->next);
next              165 net/xfrm/xfrm_interface.c 	     xip = &xi->next)
next              788 net/xfrm/xfrm_interface.c 		     xip = &xi->next)
next              548 net/xfrm/xfrm_output.c 		struct sk_buff *nskb = segs->next;
next              314 net/xfrm/xfrm_policy.c 	time64_t next = TIME64_MAX;
next              330 net/xfrm/xfrm_policy.c 		if (tmo < next)
next              331 net/xfrm/xfrm_policy.c 			next = tmo;
next              338 net/xfrm/xfrm_policy.c 		if (tmo < next)
next              339 net/xfrm/xfrm_policy.c 			next = tmo;
next              348 net/xfrm/xfrm_policy.c 		if (tmo < next)
next              349 net/xfrm/xfrm_policy.c 			next = tmo;
next              358 net/xfrm/xfrm_policy.c 		if (tmo < next)
next              359 net/xfrm/xfrm_policy.c 			next = tmo;
next              364 net/xfrm/xfrm_policy.c 	if (next != TIME64_MAX &&
next              365 net/xfrm/xfrm_policy.c 	    !mod_timer(&xp->timer, jiffies + make_jiffies(next)))
next              526 net/xfrm/xfrm_state.c 	time64_t next = TIME64_MAX;
next              549 net/xfrm/xfrm_state.c 		if (tmo < next)
next              550 net/xfrm/xfrm_state.c 			next = tmo;
next              557 net/xfrm/xfrm_state.c 		if (tmo < next)
next              558 net/xfrm/xfrm_state.c 			next = tmo;
next              568 net/xfrm/xfrm_state.c 		} else if (tmo < next) {
next              569 net/xfrm/xfrm_state.c 			next = tmo;
next              579 net/xfrm/xfrm_state.c 		else if (tmo < next)
next              580 net/xfrm/xfrm_state.c 			next = tmo;
next              587 net/xfrm/xfrm_state.c 	if (next != TIME64_MAX) {
next              588 net/xfrm/xfrm_state.c 		hrtimer_forward_now(&x->mtimer, ktime_set(next, 0));
next             1149 net/xfrm/xfrm_state.c 			xfrm_hash_grow_check(net, x->bydst.next != NULL);
next             1265 net/xfrm/xfrm_state.c 	xfrm_hash_grow_check(net, x->bydst.next != NULL);
next             1376 net/xfrm/xfrm_state.c 		xfrm_hash_grow_check(net, x->bydst.next != NULL);
next               37 samples/bpf/test_lru_dist.c 	struct list_head *next, *prev;
next               42 samples/bpf/test_lru_dist.c 	list->next = list;
next               48 samples/bpf/test_lru_dist.c 	return head->next == head;
next               53 samples/bpf/test_lru_dist.c 			      struct list_head *next)
next               55 samples/bpf/test_lru_dist.c 	next->prev = new;
next               56 samples/bpf/test_lru_dist.c 	new->next = next;
next               58 samples/bpf/test_lru_dist.c 	prev->next = new;
next               63 samples/bpf/test_lru_dist.c 	__list_add(new, head, head->next);
next               66 samples/bpf/test_lru_dist.c static inline void __list_del(struct list_head *prev, struct list_head *next)
next               68 samples/bpf/test_lru_dist.c 	next->prev = prev;
next               69 samples/bpf/test_lru_dist.c 	prev->next = next;
next               74 samples/bpf/test_lru_dist.c 	__list_del(entry->prev, entry->next);
next              328 samples/mic/mpssd/mpssd.c 	unsigned int next;
next              332 samples/mic/mpssd/mpssd.c 	next = le16toh(desc->next);
next              333 samples/mic/mpssd/mpssd.c 	return next;
next             1599 samples/mic/mpssd/mpssd.c 	for (mic = mic_list.next; mic != NULL; mic = mic->next)
next             1713 samples/mic/mpssd/mpssd.c 	for (mic = mic_list.next; mic; mic = mic->next) {
next             1739 samples/mic/mpssd/mpssd.c 			mic->next = calloc(1, sizeof(struct mic_info));
next             1740 samples/mic/mpssd/mpssd.c 			if (mic->next) {
next             1741 samples/mic/mpssd/mpssd.c 				mic = mic->next;
next               82 samples/mic/mpssd/mpssd.h 	struct mic_info *next;
next              157 samples/vfio-mdev/mbochs.c 	struct list_head next;
next              946 samples/vfio-mdev/mbochs.c 	list_add(&dmabuf->next, &mdev_state->dmabufs);
next              968 samples/vfio-mdev/mbochs.c 	list_for_each_entry(dmabuf, &mdev_state->dmabufs, next)
next              982 samples/vfio-mdev/mbochs.c 	list_for_each_entry(dmabuf, &mdev_state->dmabufs, next)
next             1063 samples/vfio-mdev/mbochs.c 		ext->type.header.next = 0;
next             1316 samples/vfio-mdev/mbochs.c 	list_for_each_entry_safe(dmabuf, tmp, &mdev_state->dmabufs, next) {
next             1317 samples/vfio-mdev/mbochs.c 		list_del(&dmabuf->next);
next              139 samples/vfio-mdev/mtty.c 	struct list_head next;
next              755 samples/vfio-mdev/mtty.c 	list_add(&mdev_state->next, &mdev_devices_list);
next              768 samples/vfio-mdev/mtty.c 	list_for_each_entry_safe(mds, tmp_mds, &mdev_devices_list, next) {
next              770 samples/vfio-mdev/mtty.c 			list_del(&mdev_state->next);
next             1350 samples/vfio-mdev/mtty.c 	list_for_each_entry(mds, &mdev_devices_list, next)
next              292 scripts/asn1_compiler.c 	struct action	*next;
next              687 scripts/asn1_compiler.c 	struct element	*next;
next             1098 scripts/asn1_compiler.c 		     ppaction = &(*ppaction)->next
next             1107 scripts/asn1_compiler.c 				action->next = *ppaction;
next             1113 scripts/asn1_compiler.c 		action->next = NULL;
next             1182 scripts/asn1_compiler.c 		child_p = &element->next;
next             1246 scripts/asn1_compiler.c 		for (c = e->children; c; c = c->next)
next             1328 scripts/asn1_compiler.c 	for (action = action_list; action; action = action->next) {
next             1338 scripts/asn1_compiler.c 	for (action = action_list; action; action = action->next)
next             1347 scripts/asn1_compiler.c 	for (action = action_list; action; action = action->next)
next             1406 scripts/asn1_compiler.c 		for (ce = e->children; ce; ce = ce->next)
next             1549 scripts/asn1_compiler.c 			for (ec = e->children; ec; ec = ec->next)
next             1597 scripts/asn1_compiler.c 		for (ec = e->children; ec; ec = ec->next)
next              165 scripts/basic/fixdep.c 	struct item	*next;
next              191 scripts/basic/fixdep.c 	for (aux = hashtab[hash % HASHSZ]; aux; aux = aux->next) {
next              213 scripts/basic/fixdep.c 	aux->next = hashtab[hash % HASHSZ];
next              107 scripts/dtc/checks.c 		while (pos->next) {
next              108 scripts/dtc/checks.c 			pos = pos->next;
next              287 scripts/dtc/checks.c 		for (prop2 = prop->next; prop2; prop2 = prop2->next) {
next              550 scripts/dtc/checks.c 	for (pp = &node->proplist; *pp; pp = &((*pp)->next))
next              566 scripts/dtc/checks.c 		*pp = prop->next;
next               14 scripts/dtc/data.c 		nm = m->next;
next              124 scripts/dtc/data.c 	m = m->next;
next              136 scripts/dtc/data.c 		mp = &((*mp)->next);
next              235 scripts/dtc/data.c 	m->next = NULL;
next               79 scripts/dtc/dtc.h 	struct marker *next;
next               92 scripts/dtc/dtc.h 	for (; (m); (m) = (m)->next)
next              132 scripts/dtc/dtc.h 	struct label *next;
next              144 scripts/dtc/dtc.h 	struct property *next;
next              173 scripts/dtc/dtc.h 	for ((l) = (l0); (l); (l) = (l)->next)
next              180 scripts/dtc/dtc.h 	for ((p) = (n)->proplist; (p); (p) = (p)->next)
next              244 scripts/dtc/dtc.h 	struct reserve_info *next;
next              300 scripts/dtc/flattree.c 	for (re = reservelist; re; re = re->next) {
next              516 scripts/dtc/flattree.c 	for (re = dti->reservelist; re; re = re->next) {
next               27 scripts/dtc/livetree.c 	new->next = *labels;
next               67 scripts/dtc/livetree.c 	assert(first->next == NULL);
next               69 scripts/dtc/livetree.c 	first->next = list;
next               77 scripts/dtc/livetree.c 	struct property *next;
next               80 scripts/dtc/livetree.c 		next = p->next;
next               81 scripts/dtc/livetree.c 		p->next = head;
next               83 scripts/dtc/livetree.c 		p = next;
next              159 scripts/dtc/livetree.c 		new_node->proplist = new_prop->next;
next              160 scripts/dtc/livetree.c 		new_prop->next = NULL;
next              270 scripts/dtc/livetree.c 	prop->next = NULL;
next              274 scripts/dtc/livetree.c 		p = &((*p)->next);
next              288 scripts/dtc/livetree.c 		prop = prop->next;
next              373 scripts/dtc/livetree.c 	assert(first->next == NULL);
next              375 scripts/dtc/livetree.c 	first->next = list;
next              384 scripts/dtc/livetree.c 	new->next = NULL;
next              389 scripts/dtc/livetree.c 	for (last = list; last->next; last = last->next)
next              392 scripts/dtc/livetree.c 	last->next = new;
next              673 scripts/dtc/livetree.c 	     ri = ri->next)
next              683 scripts/dtc/livetree.c 	     ri = ri->next)
next              690 scripts/dtc/livetree.c 		tbl[i]->next = tbl[i+1];
next              691 scripts/dtc/livetree.c 	tbl[n-1]->next = NULL;
next              726 scripts/dtc/livetree.c 		tbl[i]->next = tbl[i+1];
next              727 scripts/dtc/livetree.c 	tbl[n-1]->next = NULL;
next               15 scripts/dtc/srcpos.c 	struct search_path *next;	/* next node in list, NULL for end */
next              140 scripts/dtc/srcpos.c 	for (node = search_path_head; !*fp && node; node = node->next)
next              221 scripts/dtc/srcpos.c 	node->next = NULL;
next              229 scripts/dtc/srcpos.c 	search_path_tail = &node->next;
next              263 scripts/dtc/srcpos.c 	assert(pos->next == NULL);
next              281 scripts/dtc/srcpos.c 	for (p = pos; p->next != NULL; p = p->next);
next              282 scripts/dtc/srcpos.c 	p->next = newtail;
next              351 scripts/dtc/srcpos.c 	if (pos->next != NULL) {
next              352 scripts/dtc/srcpos.c 		rest = srcpos_string_comment(pos->next, first_line, level);
next               63 scripts/dtc/srcpos.h     struct srcpos *next;
next               83 scripts/dtc/srcpos.h 		(Current).next = NULL;						\
next              135 scripts/dtc/treesource.c 		m = m->next;
next              141 scripts/dtc/treesource.c 	struct marker *next = next_type_marker(m->next);
next              143 scripts/dtc/treesource.c 	if (next)
next              144 scripts/dtc/treesource.c 		return next->offset - m->offset;
next              222 scripts/dtc/treesource.c 		dummy_marker.next = prop->val.markers;
next              229 scripts/dtc/treesource.c 		size_t chunk_len = (m->next ? m->next->offset : len) - m->offset;
next              334 scripts/dtc/treesource.c 	for (re = dti->reservelist; re; re = re->next) {
next              326 scripts/gcc-plugins/gcc-common.h 	for (node = cgraph_nodes; node; node = node->next)
next              334 scripts/gcc-plugins/gcc-common.h 	for (node = node->next; node; node = node->next)
next              347 scripts/gcc-plugins/gcc-common.h 	for (alias = node->same_body; alias; alias = alias->next) {
next              368 scripts/gcc-plugins/gcc-common.h 	for (node = cgraph_nodes; node; node = node->next)
next              370 scripts/gcc-plugins/gcc-common.h 	for (node = varpool_nodes; node; node = node->next)
next               96 scripts/gcc-plugins/gcc-generate-gimple-pass.h 		.next			= NULL,
next              164 scripts/gcc-plugins/gcc-generate-ipa-pass.h 		.next			= NULL,
next               96 scripts/gcc-plugins/gcc-generate-rtl-pass.h 		.next			= NULL,
next               96 scripts/gcc-plugins/gcc-generate-simple_ipa-pass.h 		.next			= NULL,
next              193 scripts/gcc-plugins/stackleak_plugin.c 	rtx_insn *insn, *next;
next              212 scripts/gcc-plugins/stackleak_plugin.c 	for (insn = get_insns(); insn; insn = next) {
next              215 scripts/gcc-plugins/stackleak_plugin.c 		next = NEXT_INSN(insn);
next              253 scripts/gcc-plugins/stackleak_plugin.c 		if (GET_CODE(next) == NOTE &&
next              254 scripts/gcc-plugins/stackleak_plugin.c 		    NOTE_KIND(next) == NOTE_INSN_CALL_ARG_LOCATION) {
next              255 scripts/gcc-plugins/stackleak_plugin.c 			insn = next;
next              256 scripts/gcc-plugins/stackleak_plugin.c 			next = NEXT_INSN(insn);
next              179 scripts/genksyms/genksyms.c 		(defn = defn->next) && defn->tag == SYM_NORMAL &&
next              181 scripts/genksyms/genksyms.c 		(defn = defn->next) && defn->tag == SYM_NORMAL &&
next              332 scripts/genksyms/genksyms.c 		struct string_list *next = s->next;
next              334 scripts/genksyms/genksyms.c 		s = next;
next              345 scripts/genksyms/genksyms.c 	newnode->next = NULL;
next              358 scripts/genksyms/genksyms.c 		for (n2 = n; n2->next; n2 = n2->next)
next              360 scripts/genksyms/genksyms.c 		n2->next = start;
next              386 scripts/genksyms/genksyms.c 	for (start = start->next; start != end; start = start->next) {
next              387 scripts/genksyms/genksyms.c 		n->next = copy_node(start);
next              388 scripts/genksyms/genksyms.c 		n = n->next;
next              390 scripts/genksyms/genksyms.c 	n->next = NULL;
next              399 scripts/genksyms/genksyms.c 		a = a->next;
next              400 scripts/genksyms/genksyms.c 		b = b->next;
next              481 scripts/genksyms/genksyms.c 			def->next = defn;
next              513 scripts/genksyms/genksyms.c 	while ((tmp = tmp->next) != NULL)
next              521 scripts/genksyms/genksyms.c 	while ((list = list->next) != NULL)
next              541 scripts/genksyms/genksyms.c 	while ((tmp = tmp->next) != NULL)
next              549 scripts/genksyms/genksyms.c 	while ((list = list->next) != NULL)
next               27 scripts/genksyms/genksyms.h 	struct string_list *next;
next               30 scripts/genksyms/parse.y   *p = node->next;
next               57 scripts/genksyms/parse.y 	r->next = (*keyw)->next; *body = r; (*keyw)->next = NULL;
next              140 scripts/genksyms/parse.y 		    struct string_list *decl = (*$3)->next;
next              141 scripts/genksyms/parse.y 		    (*$3)->next = NULL;
next              410 scripts/genksyms/parse.y 		{ remove_list($2, &(*$1)->next); $$ = $2; }
next              267 scripts/kconfig/conf.c 		for (child = menu->list; child; child = child->next) {
next              322 scripts/kconfig/conf.c 		for (child = menu->list; child; child = child->next) {
next              335 scripts/kconfig/conf.c 		for (child = child->list; child; child = child->next) {
next              405 scripts/kconfig/conf.c 	for (child = menu->list; child; child = child->next)
next              446 scripts/kconfig/conf.c 	for (child = menu->list; child; child = child->next)
next              828 scripts/kconfig/confdata.c 		else if (menu->next != NULL) {
next              829 scripts/kconfig/confdata.c 			menu = menu->next;
next              832 scripts/kconfig/confdata.c 				if (menu->next != NULL) {
next              833 scripts/kconfig/confdata.c 					menu = menu->next;
next              892 scripts/kconfig/confdata.c 				goto next;
next              903 scripts/kconfig/confdata.c 				goto next;
next              912 scripts/kconfig/confdata.c next:
next              917 scripts/kconfig/confdata.c 		if (menu->next)
next              918 scripts/kconfig/confdata.c 			menu = menu->next;
next              926 scripts/kconfig/confdata.c 			if (menu->next) {
next              927 scripts/kconfig/confdata.c 				menu = menu->next;
next              968 scripts/kconfig/confdata.c 	for (file = file_list; file; file = file->next) {
next              969 scripts/kconfig/confdata.c 		if (file->next)
next               21 scripts/kconfig/expr.h 	struct file *next;
next               85 scripts/kconfig/expr.h 	struct symbol *next;
next              134 scripts/kconfig/expr.h #define for_all_symbols(i, sym) for (i = 0; i < SYMBOL_HASHSIZE; i++) for (sym = symbol_hash[i]; sym; sym = sym->next)
next              193 scripts/kconfig/expr.h 	struct property *next;     /* next property - null if last */
next              207 scripts/kconfig/expr.h 	for (st = sym->prop; st; st = st->next) \
next              212 scripts/kconfig/expr.h 	for (st = sym->prop; st; st = st->next) \
next              223 scripts/kconfig/expr.h 	struct menu *next;
next             1096 scripts/kconfig/gconf.c 		for (child = menu->list; child; child = child->next) {
next             1261 scripts/kconfig/gconf.c 	for (child1 = src->list; child1; child1 = child1->next) {
next             1345 scripts/kconfig/gconf.c 	for (child = menu->list; child; child = child->next) {
next             1431 scripts/kconfig/gconf.c 	for (child = menu->list; child; child = child->next) {
next               25 scripts/kconfig/list.h 	struct list_head *next, *prev;
next               50 scripts/kconfig/list.h 	for (pos = list_entry((head)->next, typeof(*pos), member);	\
next               52 scripts/kconfig/list.h 	     pos = list_entry(pos->member.next, typeof(*pos), member))
next               62 scripts/kconfig/list.h 	for (pos = list_entry((head)->next, typeof(*pos), member),	\
next               63 scripts/kconfig/list.h 		n = list_entry(pos->member.next, typeof(*pos), member);	\
next               65 scripts/kconfig/list.h 	     pos = n, n = list_entry(n->member.next, typeof(*n), member))
next               73 scripts/kconfig/list.h 	return head->next == head;
next               84 scripts/kconfig/list.h 			      struct list_head *next)
next               86 scripts/kconfig/list.h 	next->prev = _new;
next               87 scripts/kconfig/list.h 	_new->next = next;
next               89 scripts/kconfig/list.h 	prev->next = _new;
next              112 scripts/kconfig/list.h static inline void __list_del(struct list_head *prev, struct list_head *next)
next              114 scripts/kconfig/list.h 	next->prev = prev;
next              115 scripts/kconfig/list.h 	prev->next = next;
next              128 scripts/kconfig/list.h 	__list_del(entry->prev, entry->next);
next              129 scripts/kconfig/list.h 	entry->next = (struct list_head*)LIST_POISON1;
next               91 scripts/kconfig/lxdialog/dialog.h 	struct subtitle_list *next;
next              163 scripts/kconfig/lxdialog/dialog.h 	struct dialog_list *next;
next              178 scripts/kconfig/lxdialog/dialog.h 	     item_cur && (item_cur != &item_nil); item_cur = item_cur->next)
next              258 scripts/kconfig/lxdialog/util.c 		for (pos = dlg.subtitles; pos != NULL; pos = pos->next) {
next              270 scripts/kconfig/lxdialog/util.c 		for (pos = dlg.subtitles; pos != NULL; pos = pos->next) {
next              583 scripts/kconfig/lxdialog/util.c 	struct dialog_list *p, *next;
next              585 scripts/kconfig/lxdialog/util.c 	for (p = item_head; p; p = next) {
next              586 scripts/kconfig/lxdialog/util.c 		next = p->next;
next              599 scripts/kconfig/lxdialog/util.c 		item_cur->next = p;
next              661 scripts/kconfig/lxdialog/util.c 	for (p = item_head; p; p = p->next)
next              679 scripts/kconfig/lxdialog/util.c 	for (p = item_head; p; p = p->next) {
next              325 scripts/kconfig/mconf.c 		tmp = pos->next;
next              333 scripts/kconfig/mconf.c 				pos->next = xcalloc(1, sizeof(*pos));
next              334 scripts/kconfig/mconf.c 				pos = pos->next;
next              350 scripts/kconfig/mconf.c 		tmp = pos->next;
next              533 scripts/kconfig/mconf.c 		for (child = menu->list; child; child = child->next) {
next              638 scripts/kconfig/mconf.c 	for (child = menu->list; child; child = child->next)
next              808 scripts/kconfig/mconf.c 		for (child = menu->list; child; child = child->next) {
next               59 scripts/kconfig/menu.c 	last_entry_ptr = &menu->next;
next               73 scripts/kconfig/menu.c 	last_entry_ptr = &current_menu->next;
next              232 scripts/kconfig/menu.c 	for (prop = sym->prop; prop; prop = prop->next) {
next              308 scripts/kconfig/menu.c 				for (menu = parent->list; menu; menu = menu->next) {
next              316 scripts/kconfig/menu.c 			for (menu = parent->list; menu; menu = menu->next) {
next              337 scripts/kconfig/menu.c 		for (menu = parent->list; menu; menu = menu->next) {
next              362 scripts/kconfig/menu.c 			for (; prop; prop = prop->next) {
next              417 scripts/kconfig/menu.c 		for (menu = parent->list; menu; menu = menu->next)
next              448 scripts/kconfig/menu.c 		for (menu = parent->next; menu; menu = menu->next) {
next              455 scripts/kconfig/menu.c 				goto next;
next              478 scripts/kconfig/menu.c 		next:
next              485 scripts/kconfig/menu.c 			parent->list = parent->next;
next              486 scripts/kconfig/menu.c 			parent->next = last_menu->next;
next              487 scripts/kconfig/menu.c 			last_menu->next = NULL;
next              492 scripts/kconfig/menu.c 	for (menu = parent->list; menu; menu = menu->next) {
next              499 scripts/kconfig/menu.c 			for (prop = menu->sym->prop; prop; prop = prop->next) {
next              518 scripts/kconfig/menu.c 				for (prop = menu->sym->prop; prop; prop = prop->next) {
next              562 scripts/kconfig/menu.c 			for (last_menu = menu->list; ; last_menu = last_menu->next) {
next              564 scripts/kconfig/menu.c 				if (!last_menu->next)
next              567 scripts/kconfig/menu.c 			last_menu->next = menu->next;
next              568 scripts/kconfig/menu.c 			menu->next = menu->list;
next              617 scripts/kconfig/menu.c 	for (child = menu->list; child; child = child->next) {
next              651 scripts/kconfig/menu.c 	for (child = menu->list; child; child = child->next) {
next              800 scripts/kconfig/nconf.c 		for (child = menu->list; child; child = child->next) {
next              916 scripts/kconfig/nconf.c 	for (child = menu->list; child; child = child->next)
next             1250 scripts/kconfig/nconf.c 		for (i = 0, child = menu->list; child; child = child->next) {
next              621 scripts/kconfig/parser.y 	for (prop = sym->prop; prop; prop = prop->next) {
next              717 scripts/kconfig/parser.y 		else if (menu->next)
next              718 scripts/kconfig/parser.y 			menu = menu->next;
next              722 scripts/kconfig/parser.y 			if (menu->next) {
next              723 scripts/kconfig/parser.y 				menu = menu->next;
next              597 scripts/kconfig/qconf.cc 	for (child = menu->list; child; child = child->next) {
next              662 scripts/kconfig/qconf.cc 	for (child = menu->list; child; child = child->next) {
next             1141 scripts/kconfig/qconf.cc 	for (struct property *prop = sym->prop; prop; prop = prop->next) {
next             1836 scripts/kconfig/qconf.cc 	for (child = menu->list; child; child = child->next) {
next              156 scripts/kconfig/symbol.c 	for (prop = sym->prop; prop; prop = prop->next) {
next              818 scripts/kconfig/symbol.c 		for (symbol = symbol_hash[hash]; symbol; symbol = symbol->next) {
next              837 scripts/kconfig/symbol.c 	symbol->next = symbol_hash[hash];
next              860 scripts/kconfig/symbol.c 	for (symbol = symbol_hash[hash]; symbol; symbol = symbol->next) {
next             1008 scripts/kconfig/symbol.c 	struct dep_stack *prev, *next;
next             1018 scripts/kconfig/symbol.c 		check_top->next = stack;
next             1028 scripts/kconfig/symbol.c 		check_top->next = NULL;
next             1057 scripts/kconfig/symbol.c 	for (; stack; stack = stack->next) {
next             1059 scripts/kconfig/symbol.c 		next_sym = stack->next ? stack->next->sym : last_sym;
next             1066 scripts/kconfig/symbol.c 			for (prop = sym->prop; prop; prop = prop->next) {
next             1185 scripts/kconfig/symbol.c 	for (prop = sym->prop; prop; prop = prop->next) {
next             1290 scripts/kconfig/symbol.c 		for (propp = &sym->prop; *propp; propp = &(*propp)->next)
next               17 scripts/kconfig/util.c 	for (file = file_list; file; file = file->next) {
next               26 scripts/kconfig/util.c 	file->next = file_list;
next              129 scripts/mod/modpost.c 	for (mod = modules; mod; mod = mod->next)
next              153 scripts/mod/modpost.c 	mod->next = modules;
next              165 scripts/mod/modpost.c 	struct symbol *next;
next              200 scripts/mod/modpost.c 				   struct symbol *next)
next              207 scripts/mod/modpost.c 	s->next = next;
next              234 scripts/mod/modpost.c 	for (s = symbolhash[tdb_hash(name) % SYMBOL_HASH_SIZE]; s; s = s->next) {
next              246 scripts/mod/modpost.c 	for (ns_entry = list; ns_entry != NULL; ns_entry = ns_entry->next)
next              261 scripts/mod/modpost.c 		ns_entry->next = *list;
next             2198 scripts/mod/modpost.c 	for (s = mod->unres; s; s = s->next) {
next             2312 scripts/mod/modpost.c 	for (s = mod->unres; s; s = s->next) {
next             2328 scripts/mod/modpost.c 	for (s = mod->unres; s; s = s->next) {
next             2357 scripts/mod/modpost.c 	for (s = mod->unres; s; s = s->next)
next             2363 scripts/mod/modpost.c 	for (s = mod->unres; s; s = s->next) {
next             2523 scripts/mod/modpost.c 			symbol = symbol->next;
next             2536 scripts/mod/modpost.c 	for (mod = modules; mod; mod = mod->next) {
next             2544 scripts/mod/modpost.c 		for (ns = mod->required_namespaces; ns; ns = ns->next)
next             2556 scripts/mod/modpost.c 	struct ext_sym_list *next;
next             2585 scripts/mod/modpost.c 			extsym_iter->next = extsym_start;
next             2627 scripts/mod/modpost.c 		extsym_iter = extsym_start->next;
next             2640 scripts/mod/modpost.c 	for (mod = modules; mod; mod = mod->next) {
next             2679 scripts/mod/modpost.c 		for (s = symbolhash[n]; s; s = s->next) {
next              113 scripts/mod/modpost.h 	struct namespace_list *next;
next              118 scripts/mod/modpost.h 	struct module *next;
next             1980 security/apparmor/apparmorfs.c 	struct aa_ns *parent, *next;
next             1988 security/apparmor/apparmorfs.c 		next = list_first_entry(&ns->sub_ns, typeof(*ns), base.list);
next             1989 security/apparmor/apparmorfs.c 		mutex_lock_nested(&next->lock, next->level);
next             1990 security/apparmor/apparmorfs.c 		return next;
next             1997 security/apparmor/apparmorfs.c 		next = list_next_entry(ns, base.list);
next             1998 security/apparmor/apparmorfs.c 		if (!list_entry_is_head(next, &parent->sub_ns, base.list)) {
next             1999 security/apparmor/apparmorfs.c 			mutex_lock_nested(&next->lock, next->level);
next             2000 security/apparmor/apparmorfs.c 			return next;
next             2082 security/apparmor/apparmorfs.c 	struct aa_profile *next = __next_profile(profile);
next             2083 security/apparmor/apparmorfs.c 	if (next)
next             2084 security/apparmor/apparmorfs.c 		return next;
next             2177 security/apparmor/apparmorfs.c 	.next = p_next,
next              148 security/apparmor/domain.c 		goto next;
next              155 security/apparmor/domain.c next:
next              210 security/apparmor/domain.c 		goto next;
next              216 security/apparmor/domain.c next:
next             1025 security/apparmor/label.c 	struct aa_profile *next;
next             1038 security/apparmor/label.c 	label_for_each_in_merge(i, a, b, next) {
next             1039 security/apparmor/label.c 		AA_BUG(!next);
next             1040 security/apparmor/label.c 		if (profile_is_stale(next)) {
next             1041 security/apparmor/label.c 			new->vec[k] = aa_get_newest_profile(next);
next             1044 security/apparmor/label.c 			if (next->label.proxy != new->vec[k]->label.proxy)
next             1049 security/apparmor/label.c 			new->vec[k++] = aa_get_profile(next);
next             1288 security/apparmor/label.c 		goto next;
next             1295 security/apparmor/label.c next:
next             1348 security/apparmor/label.c 		goto next;
next             1354 security/apparmor/label.c next:
next              370 security/apparmor/match.c #define match_char(state, def, base, next, check, C)	\
next              380 security/apparmor/match.c 	(state) = (next)[pos];				\
next              405 security/apparmor/match.c 	u16 *next = NEXT_TABLE(dfa);
next              417 security/apparmor/match.c 			match_char(state, def, base, next, check,
next              422 security/apparmor/match.c 			match_char(state, def, base, next, check, (u8) *str++);
next              445 security/apparmor/match.c 	u16 *next = NEXT_TABLE(dfa);
next              458 security/apparmor/match.c 			match_char(state, def, base, next, check,
next              463 security/apparmor/match.c 			match_char(state, def, base, next, check, (u8) *str++);
next              484 security/apparmor/match.c 	u16 *next = NEXT_TABLE(dfa);
next              491 security/apparmor/match.c 		match_char(state, def, base, next, check, equiv[(u8) c]);
next              493 security/apparmor/match.c 		match_char(state, def, base, next, check, (u8) c);
next              516 security/apparmor/match.c 	u16 *next = NEXT_TABLE(dfa);
next              532 security/apparmor/match.c 				state = next[pos];
next              543 security/apparmor/match.c 				state = next[pos];
next              577 security/apparmor/match.c 	u16 *next = NEXT_TABLE(dfa);
next              594 security/apparmor/match.c 				state = next[pos];
next              605 security/apparmor/match.c 				state = next[pos];
next              653 security/apparmor/match.c 	u16 *next = NEXT_TABLE(dfa);
next              677 security/apparmor/match.c 				state = next[pos];
next              696 security/apparmor/match.c 				state = next[pos];
next              127 security/apparmor/procattr.c 			char *next = hat + strlen(hat) + 1;
next              131 security/apparmor/procattr.c 			hat = next;
next              103 security/integrity/ima/ima_fs.c 	qe = list_entry_rcu(qe->later.next, struct ima_queue_entry, later);
next              192 security/integrity/ima/ima_fs.c 	.next = ima_measurements_next,
next              258 security/integrity/ima/ima_fs.c 	.next = ima_measurements_next,
next              374 security/integrity/ima/ima_fs.c 		.next = ima_policy_next,
next             1349 security/integrity/ima/ima_policy.c 	entry = list_entry_rcu(entry->list.next, struct ima_rule_entry, list);
next              126 security/keys/gc.c 			list_entry(keys->next, struct key, graveyard_link);
next              417 security/keys/keyring.c 		if (keyring->name_link.next != NULL &&
next               23 security/keys/proc.c 	.next	= proc_keys_next,
next               35 security/keys/proc.c 	.next	= proc_key_users_next,
next               36 security/safesetid/lsm.c 	hash_for_each_possible(policy->rules, rule, next, __kuid_val(src)) {
next               35 security/safesetid/lsm.h 	struct hlist_node next;
next               68 security/safesetid/securityfs.c 	hash_for_each_safe(pol->rules, bucket, tmp, rule, next)
next               81 security/safesetid/securityfs.c 	hash_add(pol->rules, &rule->next, __kuid_val(rule->src_uid));
next               90 security/safesetid/securityfs.c 	hash_for_each(pol->rules, bucket, rule, next) {
next              210 security/security.c 	char *sep, *name, *next;
next              240 security/security.c 	next = sep;
next              242 security/security.c 	while ((name = strsep(&next, ",")) != NULL) {
next               79 security/selinux/avc.c 	struct avc_callback_node *next;
next              808 security/selinux/avc.c 	c->next = avc_callbacks;
next              971 security/selinux/avc.c 	for (c = avc_callbacks; c; c = c->next) {
next             1552 security/selinux/selinuxfs.c 	.next		= sel_avc_stats_seq_next,
next               93 security/selinux/ss/avtab.c 		newnode->next = prev->next;
next               94 security/selinux/ss/avtab.c 		prev->next = newnode;
next               98 security/selinux/ss/avtab.c 		newnode->next = *n;
next              118 security/selinux/ss/avtab.c 	     prev = cur, cur = cur->next) {
next              162 security/selinux/ss/avtab.c 	     prev = cur, cur = cur->next) {
next              192 security/selinux/ss/avtab.c 	     cur = cur->next) {
next              228 security/selinux/ss/avtab.c 	     cur = cur->next) {
next              257 security/selinux/ss/avtab.c 	for (cur = node->next; cur; cur = cur->next) {
next              289 security/selinux/ss/avtab.c 			cur = cur->next;
next              360 security/selinux/ss/avtab.c 				cur = cur->next;
next              648 security/selinux/ss/avtab.c 		     cur = cur->next) {
next               80 security/selinux/ss/avtab.h 	struct avtab_node *next;
next               31 security/selinux/ss/conditional.c 	for (cur = expr; cur; cur = cur->next) {
next               99 security/selinux/ss/conditional.c 		for (cur = node->true_list; cur; cur = cur->next) {
next              106 security/selinux/ss/conditional.c 		for (cur = node->false_list; cur; cur = cur->next) {
next              133 security/selinux/ss/conditional.c 	struct cond_av_list *cur, *next;
next              134 security/selinux/ss/conditional.c 	for (cur = list; cur; cur = next) {
next              135 security/selinux/ss/conditional.c 		next = cur->next;
next              146 security/selinux/ss/conditional.c 		next_expr = cur_expr->next;
next              156 security/selinux/ss/conditional.c 	struct cond_node *next, *cur;
next              161 security/selinux/ss/conditional.c 	for (cur = list; cur; cur = next) {
next              162 security/selinux/ss/conditional.c 		next = cur->next;
next              303 security/selinux/ss/conditional.c 				for (cur = other; cur; cur = cur->next) {
next              339 security/selinux/ss/conditional.c 		data->tail->next = list;
next              433 security/selinux/ss/conditional.c 			last->next = expr;
next              479 security/selinux/ss/conditional.c 			last->next = node;
next              530 security/selinux/ss/conditional.c 	for (cur_list = list; cur_list != NULL; cur_list = cur_list->next)
next              541 security/selinux/ss/conditional.c 	for (cur_list = list; cur_list != NULL; cur_list = cur_list->next) {
next              563 security/selinux/ss/conditional.c 	for (cur_expr = node->expr; cur_expr != NULL; cur_expr = cur_expr->next)
next              571 security/selinux/ss/conditional.c 	for (cur_expr = node->expr; cur_expr != NULL; cur_expr = cur_expr->next) {
next              597 security/selinux/ss/conditional.c 	for (cur = list; cur != NULL; cur = cur->next)
next              604 security/selinux/ss/conditional.c 	for (cur = list; cur != NULL; cur = cur->next) {
next               33 security/selinux/ss/conditional.h 	struct cond_expr *next;
next               43 security/selinux/ss/conditional.h 	struct cond_av_list *next;
next               58 security/selinux/ss/conditional.h 	struct cond_node *next;
next               54 security/selinux/ss/constraint.h 	struct constraint_expr *next;   /* next expression */
next               60 security/selinux/ss/constraint.h 	struct constraint_node *next;	/* next constraint */
next               42 security/selinux/ss/ebitmap.c 		n1 = n1->next;
next               43 security/selinux/ss/ebitmap.c 		n2 = n2->next;
next               67 security/selinux/ss/ebitmap.c 		new->next = NULL;
next               69 security/selinux/ss/ebitmap.c 			prev->next = new;
next               73 security/selinux/ss/ebitmap.c 		n = n->next;
next              123 security/selinux/ss/ebitmap.c 		e_iter = e_iter->next;
next              175 security/selinux/ss/ebitmap.c 				e_prev->next = e_iter;
next              214 security/selinux/ss/ebitmap.c 			n1 = n1->next;
next              232 security/selinux/ss/ebitmap.c 		n1 = n1->next;
next              233 security/selinux/ss/ebitmap.c 		n2 = n2->next;
next              253 security/selinux/ss/ebitmap.c 		n = n->next;
next              279 security/selinux/ss/ebitmap.c 				if (!n->next) {
next              291 security/selinux/ss/ebitmap.c 					prev->next = n->next;
next              293 security/selinux/ss/ebitmap.c 					e->node = n->next;
next              299 security/selinux/ss/ebitmap.c 		n = n->next;
next              317 security/selinux/ss/ebitmap.c 		new->next = prev->next;
next              318 security/selinux/ss/ebitmap.c 		prev->next = new;
next              320 security/selinux/ss/ebitmap.c 		new->next = e->node;
next              337 security/selinux/ss/ebitmap.c 		n = n->next;
next              417 security/selinux/ss/ebitmap.c 				n->next = tmp;
next               35 security/selinux/ss/ebitmap.h 	struct ebitmap_node *next;
next               52 security/selinux/ss/ebitmap.h 	for (*n = e->node; *n; *n = (*n)->next) {
next               75 security/selinux/ss/ebitmap.h 	for (*n = (*n)->next; *n; *n = (*n)->next) {
next               57 security/selinux/ss/hashtab.c 		cur = cur->next;
next               69 security/selinux/ss/hashtab.c 		newnode->next = prev->next;
next               70 security/selinux/ss/hashtab.c 		prev->next = newnode;
next               72 security/selinux/ss/hashtab.c 		newnode->next = h->htable[hvalue];
next               91 security/selinux/ss/hashtab.c 		cur = cur->next;
next              111 security/selinux/ss/hashtab.c 			cur = cur->next;
next              140 security/selinux/ss/hashtab.c 			cur = cur->next;
next              161 security/selinux/ss/hashtab.c 				cur = cur->next;
next               19 security/selinux/ss/hashtab.h 	struct hashtab_node *next;
next              235 security/selinux/ss/policydb.c 				e = e->next;
next              239 security/selinux/ss/policydb.c 			constraint = constraint->next;
next              248 security/selinux/ss/policydb.c 				e = e->next;
next              252 security/selinux/ss/policydb.c 			constraint = constraint->next;
next              791 security/selinux/ss/policydb.c 			c = c->next;
next              804 security/selinux/ss/policydb.c 			c = c->next;
next              808 security/selinux/ss/policydb.c 		g = g->next;
next              815 security/selinux/ss/policydb.c 	for (tr = p->role_tr; tr; tr = tr->next) {
next              822 security/selinux/ss/policydb.c 	for (ra = p->role_allow; ra; ra = ra->next) {
next              862 security/selinux/ss/policydb.c 	for (c = head; c; c = c->next) {
next             1201 security/selinux/ss/policydb.c 			lc->next = c;
next             1218 security/selinux/ss/policydb.c 				le->next = e;
next             1986 security/selinux/ss/policydb.c 		     genfs_p = genfs, genfs = genfs->next) {
next             1996 security/selinux/ss/policydb.c 		newgenfs->next = genfs;
next             1998 security/selinux/ss/policydb.c 			genfs_p->next = newgenfs;
next             2034 security/selinux/ss/policydb.c 			     l = c, c = c->next) {
next             2049 security/selinux/ss/policydb.c 			newc->next = c;
next             2051 security/selinux/ss/policydb.c 				l->next = newc;
next             2091 security/selinux/ss/policydb.c 				l->next = c;
next             2418 security/selinux/ss/policydb.c 			ltr->next = tr;
next             2457 security/selinux/ss/policydb.c 			lra->next = ra;
next             2657 security/selinux/ss/policydb.c 	for (tr = r; tr; tr = tr->next)
next             2663 security/selinux/ss/policydb.c 	for (tr = r; tr; tr = tr->next) {
next             2689 security/selinux/ss/policydb.c 	for (ra = r; ra; ra = ra->next)
next             2695 security/selinux/ss/policydb.c 	for (ra = r; ra; ra = ra->next) {
next             2815 security/selinux/ss/policydb.c 	for (c = node; c; c = c->next) {
next             2817 security/selinux/ss/policydb.c 		for (e = c->expr; e; e = e->next)
next             2824 security/selinux/ss/policydb.c 		for (e = c->expr; e; e = e->next) {
next             2873 security/selinux/ss/policydb.c 	for (c = cladatum->constraints; c; c = c->next)
next             2909 security/selinux/ss/policydb.c 	for (c = cladatum->validatetrans; c; c = c->next)
next             3086 security/selinux/ss/policydb.c 		for (c = p->ocontexts[i]; c; c = c->next)
next             3092 security/selinux/ss/policydb.c 		for (c = p->ocontexts[i]; c; c = c->next) {
next             3214 security/selinux/ss/policydb.c 	for (genfs = p->genfs; genfs; genfs = genfs->next)
next             3220 security/selinux/ss/policydb.c 	for (genfs = p->genfs; genfs; genfs = genfs->next) {
next             3230 security/selinux/ss/policydb.c 		for (c = genfs->head; c; c = c->next)
next             3236 security/selinux/ss/policydb.c 		for (c = genfs->head; c; c = c->next) {
next               88 security/selinux/ss/policydb.h 	struct role_trans *next;
next              105 security/selinux/ss/policydb.h 	struct role_allow *next;
next              202 security/selinux/ss/policydb.h 	struct ocontext *next;
next              208 security/selinux/ss/policydb.h 	struct genfs *next;
next              276 security/selinux/ss/services.c 	for (e = cexpr; e; e = e->next) {
next              689 security/selinux/ss/services.c 		constraint = constraint->next;
next              700 security/selinux/ss/services.c 		for (ra = policydb->role_allow; ra; ra = ra->next) {
next              820 security/selinux/ss/services.c 		constraint = constraint->next;
next             1780 security/selinux/ss/services.c 		     roletr = roletr->next) {
next             2017 security/selinux/ss/services.c 			oc = oc->next;
next             2278 security/selinux/ss/services.c 		c = c->next;
next             2325 security/selinux/ss/services.c 		c = c->next;
next             2372 security/selinux/ss/services.c 		c = c->next;
next             2414 security/selinux/ss/services.c 		c = c->next;
next             2489 security/selinux/ss/services.c 			c = c->next;
next             2503 security/selinux/ss/services.c 			c = c->next;
next             2688 security/selinux/ss/services.c 	for (genfs = policydb->genfs; genfs; genfs = genfs->next) {
next             2698 security/selinux/ss/services.c 	for (c = genfs->head; c; c = c->next) {
next             2767 security/selinux/ss/services.c 		c = c->next;
next             2886 security/selinux/ss/services.c 	for (cur = policydb->cond_list; cur; cur = cur->next) {
next             2943 security/selinux/ss/services.c 	for (cur = policydb->cond_list; cur; cur = cur->next) {
next              725 security/smack/smack_lsm.c 		char *next = strchr(from, ',');
next              729 security/smack/smack_lsm.c 		if (next)
next              730 security/smack/smack_lsm.c 			len = next - from;
next              632 security/smack/smackfs.c 	.next  = load2_seq_next,
next              802 security/smack/smackfs.c 	.next  = cipso_seq_next,
next              985 security/smack/smackfs.c 	.next  = cipso_seq_next,
next             1061 security/smack/smackfs.c 	.next  = net4addr_seq_next,
next             1098 security/smack/smackfs.c 	m = list_entry_rcu(smk_net4addr_list.next,
next             1112 security/smack/smackfs.c 		m_next = list_entry_rcu(m->list.next,
next             1325 security/smack/smackfs.c 	.next  = net6addr_seq_next,
next             1362 security/smack/smackfs.c 	m = list_entry_rcu(smk_net6addr_list.next,
next             1375 security/smack/smackfs.c 		m_next = list_entry_rcu(m->list.next,
next             1885 security/smack/smackfs.c 	.next  = onlycap_seq_next,
next             1913 security/smack/smackfs.c 		first = public->next;
next             1917 security/smack/smackfs.c 		private->prev->next = public;
next             1918 security/smack/smackfs.c 		private->next->prev = public;
next             1919 security/smack/smackfs.c 		rcu_assign_pointer(public->next, private->next);
next             1926 security/smack/smackfs.c 		private->next = first;
next             1929 security/smack/smackfs.c 		last->next = private;
next             2218 security/smack/smackfs.c 	.next  = load_self_seq_next,
next             2349 security/smack/smackfs.c 	.next  = load2_seq_next,
next             2426 security/smack/smackfs.c 	.next  = load_self2_seq_next,
next             2694 security/smack/smackfs.c 	.next  = relabel_self_seq_next,
next              449 security/tomoyo/audit.c 		ptr = list_entry(tomoyo_log.next, typeof(*ptr), list);
next              753 security/tomoyo/common.c  next:
next              833 security/tomoyo/common.c 		goto next;
next             2552 security/tomoyo/common.c 	if (!ns || (head->r.eof && ns->next != &tomoyo_namespace_list)) {
next             2555 security/tomoyo/common.c 		head->r.ns = ns ? ns->next : tomoyo_namespace_list.next;
next             2570 security/tomoyo/common.c 		head->r.ns->next != &tomoyo_namespace_list;
next             1331 security/tomoyo/common.h 		pos =  srcu_dereference((head)->next, &tomoyo_ss);	\
next             1332 security/tomoyo/common.h 	for ( ; pos != (head); pos = srcu_dereference(pos->next, &tomoyo_ss))
next               96 sound/core/control.c 		cread = snd_kctl_event(ctl->events.next);
next             1621 sound/core/control.c 		kev = snd_kctl_event(ctl->events.next);
next             1844 sound/core/control.c 		control = snd_kcontrol(card->controls.next);
next              224 sound/core/device.c 	struct snd_device *dev, *next;
next              228 sound/core/device.c 	list_for_each_entry_safe_reverse(dev, next, &card->devices, list) {
next              237 sound/core/device.c 	list_for_each_entry_safe_reverse(dev, next, &card->devices, list)
next              310 sound/core/info.c 	size_t next;
next              318 sound/core/info.c 	next = pos + count;
next              320 sound/core/info.c 	if (next > 16 * 1024)
next              331 sound/core/info.c 	if (next > buf->len) {
next              332 sound/core/info.c 		char *nbuf = kvzalloc(PAGE_ALIGN(next), GFP_KERNEL);
next              339 sound/core/info.c 		buf->len = PAGE_ALIGN(next);
next              345 sound/core/info.c 	buf->size = next;
next              350 sound/core/info.c 	*offset = next;
next              542 sound/core/oss/pcm_oss.c 	struct snd_pcm_plugin *plugin, *next;
next              546 sound/core/oss/pcm_oss.c 		next = plugin->next;
next              548 sound/core/oss/pcm_oss.c 		plugin = next;
next              557 sound/core/oss/pcm_oss.c 	plugin->next = runtime->oss.plugin_first;
next              572 sound/core/oss/pcm_oss.c 	plugin->next = NULL;
next              575 sound/core/oss/pcm_oss.c 		runtime->oss.plugin_last->next = plugin;
next             2302 sound/core/oss/pcm_oss.c 		     setup = setup->next) {
next             2927 sound/core/oss/pcm_oss.c 		setup = setup->next;
next             2938 sound/core/oss/pcm_oss.c 		setupn = setup->next;
next             2963 sound/core/oss/pcm_oss.c 		for (setup = pstr->oss.setup_list; setup; setup = setup->next) {
next             3007 sound/core/oss/pcm_oss.c 				     setup1->next; setup1 = setup1->next);
next             3008 sound/core/oss/pcm_oss.c 				setup1->next = setup;
next              111 sound/core/oss/pcm_plugin.c 		while (plugin->next) {
next              116 sound/core/oss/pcm_plugin.c 			plugin = plugin->next;
next              225 sound/core/oss/pcm_plugin.c 			plugin_next = plugin->next;
next              255 sound/core/oss/pcm_plugin.c 			plugin_next = plugin->next;
next              618 sound/core/oss/pcm_plugin.c 	struct snd_pcm_plugin *plugin, *next;
next              627 sound/core/oss/pcm_plugin.c 		if ((next = plugin->next) != NULL) {
next              634 sound/core/oss/pcm_plugin.c 			if ((err = next->client_channels(next, frames1, &dst_channels)) < 0) {
next              651 sound/core/oss/pcm_plugin.c 		plugin = next;
next              658 sound/core/oss/pcm_plugin.c 	struct snd_pcm_plugin *plugin, *next;
next              670 sound/core/oss/pcm_plugin.c 		if ((next = plugin->next) != NULL) {
next              681 sound/core/oss/pcm_plugin.c 		plugin = next;
next               60 sound/core/oss/pcm_plugin.h 	struct snd_pcm_plugin *next;
next              137 sound/core/pcm.c 			     substream = substream->next)
next              678 sound/core/pcm.c 			prev->next = substream;
next              688 sound/core/pcm.c 					prev->next = NULL;
next              830 sound/core/pcm.c 		substream_next = substream->next;
next              837 sound/core/pcm.c 		setupn = setup->next;
next              907 sound/core/pcm.c 		     substream = substream->next) {
next              920 sound/core/pcm.c 			     substream = substream->next)
next              933 sound/core/pcm.c 	for (substream = pstr->substream; substream; substream = substream->next) {
next             1075 sound/core/pcm.c 		for (substream = pcm->streams[cidx].substream; substream; substream = substream->next)
next             1097 sound/core/pcm.c 		for (substream = pcm->streams[cidx].substream; substream; substream = substream->next) {
next              477 sound/core/pcm_lib.c 	for (substream = stream->substream; substream != NULL; substream = substream->next)
next               93 sound/core/pcm_memory.c 		for (substream = pcm->streams[stream].substream; substream; substream = substream->next)
next              249 sound/core/pcm_memory.c 		for (substream = pcm->streams[stream].substream; substream; substream = substream->next)
next             1524 sound/core/pcm_native.c 		     substream; substream = substream->next) {
next             1631 sound/core/rawmidi.c 		substream = list_entry(stream->substreams.next, struct snd_rawmidi_substream, list);
next              123 sound/core/seq/seq_fifo.c 		f->tail->next = cell;
next              127 sound/core/seq/seq_fifo.c 	cell->next = NULL;
next              147 sound/core/seq/seq_fifo.c 		f->head = cell->next;
next              153 sound/core/seq/seq_fifo.c 		cell->next = NULL;
next              205 sound/core/seq/seq_fifo.c 		cell->next = f->head;
next              227 sound/core/seq/seq_fifo.c 	struct snd_seq_event_cell *cell, *next, *oldhead;
next              258 sound/core/seq/seq_fifo.c 	for (cell = oldhead; cell; cell = next) {
next              259 sound/core/seq/seq_fifo.c 		next = cell->next;
next               96 sound/core/seq/seq_memory.c 	for (; len > 0 && cell; cell = cell->next) {
next              166 sound/core/seq/seq_memory.c 	cell->next = pool->free;
next              189 sound/core/seq/seq_memory.c 				nextptr = curp->next;
next              190 sound/core/seq/seq_memory.c 				curp->next = pool->free;
next              255 sound/core/seq/seq_memory.c 		pool->free = cell->next;
next              262 sound/core/seq/seq_memory.c 		cell->next = NULL;
next              331 sound/core/seq/seq_memory.c 				tail->next = tmp;
next              336 sound/core/seq/seq_memory.c 				src = src->next;
next              396 sound/core/seq/seq_memory.c 		cellptr->next = pool->free;
next               18 sound/core/seq/seq_memory.h 	struct snd_seq_event_cell *next;	/* next cell */
next              153 sound/core/seq/seq_prioq.c 			f->tail->next = cell;
next              155 sound/core/seq/seq_prioq.c 			cell->next = NULL;
next              180 sound/core/seq/seq_prioq.c 		cur = cur->next;
next              190 sound/core/seq/seq_prioq.c 		prev->next = cell;
next              191 sound/core/seq/seq_prioq.c 	cell->next = cur;
next              228 sound/core/seq/seq_prioq.c 		f->head = cell->next;
next              234 sound/core/seq/seq_prioq.c 		cell->next = NULL;
next              277 sound/core/seq/seq_prioq.c 	register struct snd_seq_event_cell *cell, *next;
next              286 sound/core/seq/seq_prioq.c 		next = cell->next;
next              290 sound/core/seq/seq_prioq.c 				f->head = cell->next;
next              292 sound/core/seq/seq_prioq.c 				prev->next = cell->next;
next              295 sound/core/seq/seq_prioq.c 				f->tail = cell->next;
next              298 sound/core/seq/seq_prioq.c 			cell->next = NULL;
next              302 sound/core/seq/seq_prioq.c 				freeprev->next = cell;
next              316 sound/core/seq/seq_prioq.c 		cell = next;		
next              322 sound/core/seq/seq_prioq.c 		freenext = freefirst->next;
next              387 sound/core/seq/seq_prioq.c 	struct snd_seq_event_cell *cell, *next;
next              397 sound/core/seq/seq_prioq.c 		next = cell->next;
next              403 sound/core/seq/seq_prioq.c 				f->head = cell->next;
next              405 sound/core/seq/seq_prioq.c 				prev->next = cell->next;
next              409 sound/core/seq/seq_prioq.c 				f->tail = cell->next;
next              413 sound/core/seq/seq_prioq.c 			cell->next = NULL;
next              417 sound/core/seq/seq_prioq.c 				freeprev->next = cell;
next              424 sound/core/seq/seq_prioq.c 		cell = next;		
next              430 sound/core/seq/seq_prioq.c 		freenext = freefirst->next;
next              296 sound/core/timer.c 			list_entry(timer->open_list_head.next,
next              782 sound/core/timer.c 		list_del_init(head->next);
next             1528 sound/core/timer.c 			timer = list_entry(snd_timer_list.next,
next              309 sound/drivers/opl3/opl3_synth.c 	for (patch = opl3->patch_table[key]; patch; patch = patch->next) {
next              321 sound/drivers/opl3/opl3_synth.c 	patch->next = opl3->patch_table[key];
next              334 sound/drivers/opl3/opl3_synth.c 		struct fm_patch *patch, *next;
next              335 sound/drivers/opl3/opl3_synth.c 		for (patch = opl3->patch_table[i]; patch; patch = next) {
next              336 sound/drivers/opl3/opl3_synth.c 			next = patch->next;
next              468 sound/drivers/opl4/opl4_synth.c 		return list_entry(opl4->off_voices.next, struct opl4_voice, list);
next              471 sound/drivers/opl4/opl4_synth.c 	return list_entry(opl4->on_voices.next, struct opl4_voice, list);
next             1217 sound/firewire/amdtp-stream.c 	struct amdtp_stream *s, *next;
next             1219 sound/firewire/amdtp-stream.c 	list_for_each_entry_safe(s, next, &d->streams, list) {
next               42 sound/i2c/i2c.c 		device = snd_i2c_device(bus->devices.next);
next               49 sound/i2c/i2c.c 			slave = snd_i2c_slave_bus(bus->buses.next);
next               91 sound/isa/gus/gus_dma.c 			gus->gf1.dma_data_pcm = block->next;
next               99 sound/isa/gus/gus_dma.c 			gus->gf1.dma_data_synth = block->next;
next              166 sound/isa/gus/gus_dma.c 			gus->gf1.dma_data_pcm = block->next;
next              170 sound/isa/gus/gus_dma.c 			gus->gf1.dma_data_synth = block->next;
next              193 sound/isa/gus/gus_dma.c 	block->next = NULL;
next              207 sound/isa/gus/gus_dma.c 			gus->gf1.dma_data_synth_last->next = block;
next              215 sound/isa/gus/gus_dma.c 			gus->gf1.dma_data_pcm_last->next = block;
next               40 sound/isa/gus/gus_mem.c 			nblock->next = pblock;
next               45 sound/isa/gus/gus_mem.c 				nblock->prev->next = nblock;
next               49 sound/isa/gus/gus_mem.c 		pblock = pblock->next;
next               51 sound/isa/gus/gus_mem.c 	nblock->next = NULL;
next               57 sound/isa/gus/gus_mem.c 		alloc->last->next = nblock;
next               71 sound/isa/gus/gus_mem.c 		alloc->first = block->next;
next               72 sound/isa/gus/gus_mem.c 		if (block->next)
next               73 sound/isa/gus/gus_mem.c 			block->next->prev = NULL;
next               75 sound/isa/gus/gus_mem.c 		block->prev->next = block->next;
next               76 sound/isa/gus/gus_mem.c 		if (block->next)
next               77 sound/isa/gus/gus_mem.c 			block->next->prev = block->prev;
next               82 sound/isa/gus/gus_mem.c 			block->prev->next = NULL;
next               84 sound/isa/gus/gus_mem.c 		block->next->prev = block->prev;
next               86 sound/isa/gus/gus_mem.c 			block->prev->next = block->next;
next               98 sound/isa/gus/gus_mem.c 	for (block = alloc->first; block; block = block->next) {
next              114 sound/isa/gus/gus_mem.c 	for (block = alloc->first; block; block = block->next)
next              139 sound/isa/gus/gus_mem.c 	block->prev = block->next = NULL;
next              140 sound/isa/gus/gus_mem.c 	for (pblock = alloc->first, idx = 0; pblock; pblock = pblock->next) {
next              146 sound/isa/gus/gus_mem.c 		if (pblock->next) {
next              147 sound/isa/gus/gus_mem.c 			if (pblock->ptr + pblock->size == pblock->next->ptr)
next              149 sound/isa/gus/gus_mem.c 			if (pblock->next->ptr < boundary)
next              150 sound/isa/gus/gus_mem.c 				ptr2 = pblock->next->ptr;
next              261 sound/isa/gus/gus_mem.c 		nblock = block->next;
next              292 sound/isa/gus/gus_mem.c 	for (block = alloc->first, i = 0; block; block = block->next, i++) {
next              877 sound/isa/gus/gus_pcm.c 	for (substream = pcm->streams[SNDRV_PCM_STREAM_PLAYBACK].substream; substream; substream = substream->next)
next              199 sound/pci/atiixp.c 	__le32 next;	/* address of the next packet descriptor */
next              384 sound/pci/atiixp.c 			desc->next = cpu_to_le32((u32)dma->desc_buf.addr);
next              386 sound/pci/atiixp.c 			desc->next = cpu_to_le32(desc_addr);
next              175 sound/pci/atiixp_modem.c 	__le32 next;	/* address of the next packet descriptor */
next              351 sound/pci/atiixp_modem.c 			desc->next = cpu_to_le32((u32)dma->desc_buf.addr);
next              353 sound/pci/atiixp_modem.c 			desc->next = cpu_to_le32(desc_addr);
next             1390 sound/pci/ca0106/ca0106_main.c 	    substream = substream->next) {
next             1398 sound/pci/ca0106/ca0106_main.c 	      substream = substream->next) {
next              638 sound/pci/ctxfi/ctdaio.c 	hw->daio_mgr_set_imapnxt(mgr->ctrl_blk, entry->next);
next              710 sound/pci/ctxfi/ctdaio.c 	entry->slot = entry->addr = entry->next = entry->user = 0;
next              126 sound/pci/ctxfi/cthardware.h 	int (*srcimp_mgr_set_imapnxt)(void *blk, unsigned int next);
next              172 sound/pci/ctxfi/cthardware.h 	int (*daio_mgr_set_imapnxt)(void *blk, unsigned int next);
next              541 sound/pci/ctxfi/cthw20k1.c static int srcimp_mgr_set_imapnxt(void *blk, unsigned int next)
next              545 sound/pci/ctxfi/cthw20k1.c 	set_field(&ctl->srcimap.srcaim, SRCAIM_NXT, next);
next             1080 sound/pci/ctxfi/cthw20k1.c static int daio_mgr_set_imapnxt(void *blk, unsigned int next)
next             1084 sound/pci/ctxfi/cthw20k1.c 	set_field(&ctl->daoimap.aim, AIM_NXT, next);
next              541 sound/pci/ctxfi/cthw20k2.c static int srcimp_mgr_set_imapnxt(void *blk, unsigned int next)
next              545 sound/pci/ctxfi/cthw20k2.c 	set_field(&ctl->srcimap.srcaim, SRCAIM_NXT, next);
next             1035 sound/pci/ctxfi/cthw20k2.c static int daio_mgr_set_imapnxt(void *blk, unsigned int next)
next             1039 sound/pci/ctxfi/cthw20k2.c 	set_field(&ctl->daoimap.aim, AIM_NXT, next);
next               27 sound/pci/ctxfi/ctimap.c 		entry->next = entry->addr;
next               49 sound/pci/ctxfi/ctimap.c 		pos = head->next;
next               56 sound/pci/ctxfi/ctimap.c 	entry->next = pos_ent->addr;
next               58 sound/pci/ctxfi/ctimap.c 	pre_ent->next = entry->addr;
next               67 sound/pci/ctxfi/ctimap.c 	struct list_head *next, *pre, *head;
next               76 sound/pci/ctxfi/ctimap.c 	next = (entry->list.next == head) ? head->next : entry->list.next;
next               80 sound/pci/ctxfi/ctimap.c 		entry->next = entry->addr = entry->user = entry->slot = 0;
next               87 sound/pci/ctxfi/ctimap.c 	next_ent = list_entry(next, struct imapper, list);
next               89 sound/pci/ctxfi/ctimap.c 	pre_ent->next = next_ent->addr;
next              102 sound/pci/ctxfi/ctimap.c 		pos = head->next;
next               24 sound/pci/ctxfi/ctimap.h 	unsigned short next; /* the next input mapper ram id */
next              784 sound/pci/ctxfi/ctsrc.c 	hw->srcimp_mgr_set_imapnxt(mgr->ctrl_blk, entry->next);
next              848 sound/pci/ctxfi/ctsrc.c 	entry->slot = entry->addr = entry->next = entry->user = 0;
next              223 sound/pci/ctxfi/ctvmem.c 		pos = vm->used.next;
next              229 sound/pci/ctxfi/ctvmem.c 		pos = vm->unused.next;
next              878 sound/pci/echoaudio/echoaudio.c 		for (ss = pcm->streams[stream].substream; ss; ss = ss->next)
next              755 sound/pci/emu10k1/emu10k1_main.c 			goto next;
next              770 sound/pci/emu10k1/emu10k1_main.c 			goto next;
next              789 sound/pci/emu10k1/emu10k1_main.c  next:
next              397 sound/pci/emu10k1/emufx.c 		nirq = irq->next;	/* irq ptr can be removed from list */
next              418 sound/pci/emu10k1/emufx.c 	irq->next = NULL;
next              425 sound/pci/emu10k1/emufx.c 		irq->next = emu->fx8010.irq_handlers;
next              440 sound/pci/emu10k1/emufx.c 		emu->fx8010.irq_handlers = tmp->next;
next              446 sound/pci/emu10k1/emufx.c 		while (tmp && tmp->next != irq)
next              447 sound/pci/emu10k1/emufx.c 			tmp = tmp->next;
next              449 sound/pci/emu10k1/emufx.c 			tmp->next = tmp->next->next;
next             1415 sound/pci/emu10k1/emupcm.c 	for (substream = pcm->streams[SNDRV_PCM_STREAM_PLAYBACK].substream; substream; substream = substream->next)
next             1420 sound/pci/emu10k1/emupcm.c 	for (substream = pcm->streams[SNDRV_PCM_STREAM_CAPTURE].substream; substream; substream = substream->next)
next             1446 sound/pci/emu10k1/emupcm.c 	for (substream = pcm->streams[SNDRV_PCM_STREAM_PLAYBACK].substream; substream; substream = substream->next)
next              137 sound/pci/emu10k1/memory.c 	struct list_head *next;
next              139 sound/pci/emu10k1/memory.c 	page = search_empty_map_area(emu, blk->pages, &next);
next              147 sound/pci/emu10k1/memory.c 	list_add_tail(&blk->mapped_link, next);
next              177 sound/pci/emu10k1/memory.c 	if ((p = blk->mapped_link.next) != &emu->mapped_link_head) {
next              273 sound/pci/emu10k1/memory.c 		p = emu->mapped_order_link_head.next;
next              275 sound/pci/emu10k1/memory.c 			nextp = p->next;
next              463 sound/pci/emu10k1/memory.c 	if ((p = blk->mem.list.next) != &hdr->block) {
next              644 sound/pci/emu10k1/p16v.c 	    substream = substream->next) {
next              657 sound/pci/emu10k1/p16v.c 	      substream = substream->next) {
next             1394 sound/pci/es1968.c 	if (buf->list.next != &chip->buf_list) {
next             1395 sound/pci/es1968.c 		chunk = list_entry(buf->list.next, struct esm_memory, list);
next             1412 sound/pci/es1968.c 	while ((p = chip->buf_list.next) != &chip->buf_list) {
next             1317 sound/pci/hda/hda_controller.c 	struct hda_codec *codec, *next;
next             1322 sound/pci/hda/hda_controller.c 	list_for_each_codec_safe(codec, next, &chip->bus) {
next              130 sound/pci/hda/hda_jack.c 		struct hda_jack_callback *cb, *next;
next              136 sound/pci/hda/hda_jack.c 		for (cb = jack->callback; cb; cb = next) {
next              137 sound/pci/hda/hda_jack.c 			next = cb->next;
next              259 sound/pci/hda/hda_jack.c 		callback->next = jack->callback;
next              536 sound/pci/hda/hda_jack.c 	for (cb = jack->callback; cb; cb = cb->next) {
next              545 sound/pci/hda/hda_jack.c 			for (cb = gated->callback; cb; cb = cb->next) {
next               26 sound/pci/hda/hda_jack.h 	struct hda_jack_callback *next;
next              943 sound/pci/mixart/mixart.c 		for (subs = pcm->streams[stream].substream; subs; subs = subs->next, idx++)
next             2187 sound/pci/trident/trident_main.c 		for (substream = pcm->streams[SNDRV_PCM_STREAM_PLAYBACK].substream; substream; substream = substream->next)
next             2234 sound/pci/trident/trident_main.c 	substream = substream->next;
next             2236 sound/pci/trident/trident_main.c 	substream = substream->next;
next             2239 sound/pci/trident/trident_main.c 		substream = substream->next;
next             1871 sound/pci/ymfpci/ymfpci_main.c 		substream = substream->next;
next              100 sound/soc/au1x/ac97c.c 			goto next;
next              116 sound/soc/au1x/ac97c.c next:
next              139 sound/soc/au1x/ac97c.c 			goto next;
next              148 sound/soc/au1x/ac97c.c next:
next               30 sound/soc/au1x/dma.c 	struct pcm_period *next;
next               56 sound/soc/au1x/dma.c 		pointer_next = pointer->next;
next               92 sound/soc/au1x/dma.c 			pointer->next = kmalloc(sizeof(struct pcm_period),
next               94 sound/soc/au1x/dma.c 			if (!pointer->next) {
next               98 sound/soc/au1x/dma.c 			pointer = pointer->next;
next              101 sound/soc/au1x/dma.c 	pointer->next = stream->buffer;
next              121 sound/soc/au1x/dma.c 		set_dma_addr1(stream->dma, stream->buffer->next->start);
next              127 sound/soc/au1x/dma.c 		set_dma_addr0(stream->dma, stream->buffer->next->start);
next              141 sound/soc/au1x/dma.c 		stream->buffer = stream->buffer->next;
next              143 sound/soc/au1x/dma.c 		set_dma_addr0(stream->dma, stream->buffer->next->start);
next              148 sound/soc/au1x/dma.c 		stream->buffer = stream->buffer->next;
next              150 sound/soc/au1x/dma.c 		set_dma_addr1(stream->dma, stream->buffer->next->start);
next              461 sound/soc/fsl/fsl_dma.c 		dma_private->link[i].next = cpu_to_be64(temp_link);
next              466 sound/soc/fsl/fsl_dma.c 	dma_private->link[i - 1].next = cpu_to_be64(dma_private->ld_buf_phys);
next              103 sound/soc/fsl/fsl_dma.h 	__be64 next;    	/* Address of next list descriptor */
next              121 sound/soc/fsl/fsl_dma.h 	__be64 next;    /* Address of next link descriptor */
next              156 sound/soc/intel/atom/sst/sst_ipc.c 		msg = list_entry(sst_drv_ctx->ipc_dispatch_list.next,
next             1218 sound/soc/sh/rcar/core.c 	     substream = substream->next) {
next             1321 sound/soc/soc-dapm.c 	list_del(widgets.next);
next              113 sound/sound_core.c 	struct sound_unit *next;
next              169 sound/sound_core.c 			list=&((*list)->next);
next              176 sound/sound_core.c 			list=&((*list)->next);
next              189 sound/sound_core.c 			list=&((*list)->next);
next              204 sound/sound_core.c 	s->next=*list;
next              222 sound/sound_core.c 			*list=p->next;
next              225 sound/sound_core.c 		list=&(p->next);
next              531 sound/sound_core.c 		s=s->next;
next              342 sound/sparc/amd7930.c 	struct snd_amd7930	*next;
next             1050 sound/sparc/amd7930.c 	amd->next = amd7930_list;
next             1088 sound/sparc/amd7930.c 		struct snd_amd7930 *next = p->next;
next             1092 sound/sparc/amd7930.c 		p = next;
next              118 sound/sparc/cs4231.c 	struct snd_cs4231	*next;
next              263 sound/synth/emux/soundfont.c 		for (sf = sflist->fonts; sf; sf = sf->next) {
next              278 sound/synth/emux/soundfont.c 	sf->next = sflist->fonts;
next              354 sound/synth/emux/soundfont.c 	zp->next = sf->zones;
next              387 sound/synth/emux/soundfont.c 	sp->next = sf->samples;
next              404 sound/synth/emux/soundfont.c 		sf->samples = sp->next;
next              432 sound/synth/emux/soundfont.c 	for (zp = sf->zones; zp; prevp = zp, zp = zp->next) {
next              443 sound/synth/emux/soundfont.c 				prevp->next = zp->next;
next              444 sound/synth/emux/soundfont.c 				zp->next = sf->zones;
next              480 sound/synth/emux/soundfont.c 	struct snd_sf_zone *prev, *next, *p;
next              484 sound/synth/emux/soundfont.c 	for (p = sf->zones; p; p = next) {
next              485 sound/synth/emux/soundfont.c 		next = p->next;
next              490 sound/synth/emux/soundfont.c 				prev->next = next;
next              492 sound/synth/emux/soundfont.c 				sf->zones = next;
next              550 sound/synth/emux/soundfont.c 		for (zone = sf->zones; zone; zone = zone->next) {
next              680 sound/synth/emux/soundfont.c 	for (p = sf->samples; p; p = p->next) {
next             1144 sound/synth/emux/soundfont.c 	for (sf = sflist->fonts; sf; sf = sf->next) {
next             1145 sound/synth/emux/soundfont.c 		for (cur = sf->zones; cur; cur = cur->next) {
next             1362 sound/synth/emux/soundfont.c 		nextsf = sf->next;
next             1364 sound/synth/emux/soundfont.c 			nextzp = zp->next;
next             1368 sound/synth/emux/soundfont.c 			nextsp = sp->next;
next             1457 sound/synth/emux/soundfont.c 	for (sf = sflist->fonts; sf; sf = sf->next) {
next             1461 sound/synth/emux/soundfont.c 			nextzp = zp->next;
next             1469 sound/synth/emux/soundfont.c 			nextsp = sp->next;
next               49 sound/synth/util_mem.c 	while ((p = hdr->block.next) != &hdr->block) {
next               33 sound/usb/helper.c 	u8 *p, *end, *next;
next               40 sound/usb/helper.c 		next = p + p[0];
next               41 sound/usb/helper.c 		if (next > end)
next               46 sound/usb/helper.c 		p = next;
next              888 tools/bpf/bpf_dbg.c 			    uint32_t pkt_len, int next)
next              896 tools/bpf/bpf_dbg.c 		if (i++ == next)
next                3 tools/firewire/list.h 	struct list *next, *prev;
next                9 tools/firewire/list.h 	list->next = list;
next               16 tools/firewire/list.h 	return list->next == list;
next               23 tools/firewire/list.h 	new_link->next		= link;
next               24 tools/firewire/list.h 	new_link->prev->next	= new_link;
next               25 tools/firewire/list.h 	new_link->next->prev	= new_link;
next               37 tools/firewire/list.h 	list_insert(list->next, new_link);
next               43 tools/firewire/list.h 	link->prev->next = link->next;
next               44 tools/firewire/list.h 	link->next->prev = link->prev;
next               51 tools/firewire/list.h 	list_entry((list)->next, type, member)
next               57 tools/firewire/list.h 	list_entry((elm)->member.next, typeof(*elm), member)
next               29 tools/firmware/ihex2fw.c 	struct ihex_binrec *next; /* not part of the real data structure */
next              259 tools/firmware/ihex2fw.c 		p = &((*p)->next);
next              261 tools/firmware/ihex2fw.c 	record->next = *p;
next              282 tools/firmware/ihex2fw.c 		p = p->next;
next               27 tools/include/linux/list.h 	list->next = list;
next               40 tools/include/linux/list.h 			      struct list_head *next)
next               42 tools/include/linux/list.h 	next->prev = new;
next               43 tools/include/linux/list.h 	new->next = next;
next               45 tools/include/linux/list.h 	prev->next = new;
next               50 tools/include/linux/list.h 			      struct list_head *next);
next               63 tools/include/linux/list.h 	__list_add(new, head, head->next);
next               87 tools/include/linux/list.h static inline void __list_del(struct list_head * prev, struct list_head * next)
next               89 tools/include/linux/list.h 	next->prev = prev;
next               90 tools/include/linux/list.h 	WRITE_ONCE(prev->next, next);
next              102 tools/include/linux/list.h 	__list_del(entry->prev, entry->next);
next              107 tools/include/linux/list.h 	__list_del(entry->prev, entry->next);
next              108 tools/include/linux/list.h 	entry->next = LIST_POISON1;
next              126 tools/include/linux/list.h 	new->next = old->next;
next              127 tools/include/linux/list.h 	new->next->prev = new;
next              129 tools/include/linux/list.h 	new->prev->next = new;
next              180 tools/include/linux/list.h 	return list->next == head;
next              189 tools/include/linux/list.h 	return head->next == head;
next              207 tools/include/linux/list.h 	struct list_head *next = head->next;
next              208 tools/include/linux/list.h 	return (next == head) && (next == head->prev);
next              220 tools/include/linux/list.h 		first = head->next;
next              231 tools/include/linux/list.h 	return !list_empty(head) && (head->next == head->prev);
next              237 tools/include/linux/list.h 	struct list_head *new_first = entry->next;
next              238 tools/include/linux/list.h 	list->next = head->next;
next              239 tools/include/linux/list.h 	list->next->prev = list;
next              241 tools/include/linux/list.h 	entry->next = list;
next              242 tools/include/linux/list.h 	head->next = new_first;
next              266 tools/include/linux/list.h 		(head->next != entry && head != entry))
next              276 tools/include/linux/list.h 				 struct list_head *next)
next              278 tools/include/linux/list.h 	struct list_head *first = list->next;
next              282 tools/include/linux/list.h 	prev->next = first;
next              284 tools/include/linux/list.h 	last->next = next;
next              285 tools/include/linux/list.h 	next->prev = last;
next              297 tools/include/linux/list.h 		__list_splice(list, head, head->next);
next              323 tools/include/linux/list.h 		__list_splice(list, head, head->next);
next              363 tools/include/linux/list.h 	list_entry((ptr)->next, type, member)
next              393 tools/include/linux/list.h 	list_entry((pos)->member.next, typeof(*(pos)), member)
next              409 tools/include/linux/list.h 	for (pos = (head)->next; pos != (head); pos = pos->next)
next              426 tools/include/linux/list.h 	for (pos = (head)->next, n = pos->next; pos != (head); \
next              427 tools/include/linux/list.h 		pos = n, n = pos->next)
next              600 tools/include/linux/list.h 	h->next = NULL;
next              616 tools/include/linux/list.h 	struct hlist_node *next = n->next;
next              619 tools/include/linux/list.h 	WRITE_ONCE(*pprev, next);
next              620 tools/include/linux/list.h 	if (next)
next              621 tools/include/linux/list.h 		next->pprev = pprev;
next              627 tools/include/linux/list.h 	n->next = LIST_POISON1;
next              642 tools/include/linux/list.h 	n->next = first;
next              644 tools/include/linux/list.h 		first->pprev = &n->next;
next              651 tools/include/linux/list.h 					struct hlist_node *next)
next              653 tools/include/linux/list.h 	n->pprev = next->pprev;
next              654 tools/include/linux/list.h 	n->next = next;
next              655 tools/include/linux/list.h 	next->pprev = &n->next;
next              662 tools/include/linux/list.h 	n->next = prev->next;
next              663 tools/include/linux/list.h 	prev->next = n;
next              664 tools/include/linux/list.h 	n->pprev = &prev->next;
next              666 tools/include/linux/list.h 	if (n->next)
next              667 tools/include/linux/list.h 		n->next->pprev  = &n->next;
next              673 tools/include/linux/list.h 	n->pprev = &n->next;
next              678 tools/include/linux/list.h 	return h->pprev == &h->next;
next              697 tools/include/linux/list.h 	for (pos = (head)->first; pos ; pos = pos->next)
next              700 tools/include/linux/list.h 	for (pos = (head)->first; pos && ({ n = pos->next; 1; }); \
next              717 tools/include/linux/list.h 	     pos = hlist_entry_safe((pos)->member.next, typeof(*(pos)), member))
next              725 tools/include/linux/list.h 	for (pos = hlist_entry_safe((pos)->member.next, typeof(*(pos)), member);\
next              727 tools/include/linux/list.h 	     pos = hlist_entry_safe((pos)->member.next, typeof(*(pos)), member))
next              736 tools/include/linux/list.h 	     pos = hlist_entry_safe((pos)->member.next, typeof(*(pos)), member))
next              747 tools/include/linux/list.h 	     pos && ({ n = pos->member.next; 1; });			\
next              760 tools/include/linux/list.h 	begin->prev->next = end->next;
next              761 tools/include/linux/list.h 	end->next->prev = begin->prev;
next              770 tools/include/linux/list.h 	for (; pos != (head); pos = pos->next)
next               70 tools/include/linux/types.h 	struct list_head *next, *prev;
next               78 tools/include/linux/types.h 	struct hlist_node *next, **pprev;
next              116 tools/include/uapi/drm/drm.h 	unsigned char next;
next              144 tools/io_uring/queue.c 	unsigned next = sq->sqe_tail + 1;
next              150 tools/io_uring/queue.c 	if (next - sq->sqe_head > *sq->kring_entries)
next              154 tools/io_uring/queue.c 	sq->sqe_tail = next;
next               21 tools/lib/bpf/hashmap.c 	entry->next = *pprev;
next               28 tools/lib/bpf/hashmap.c 	*pprev = entry->next;
next               29 tools/lib/bpf/hashmap.c 	entry->next = NULL;
next              130 tools/lib/bpf/hashmap.c 	     prev_ptr = &cur->next, cur = cur->next) {
next               32 tools/lib/bpf/hashmap.h 	struct hashmap_entry *next;
next              140 tools/lib/bpf/hashmap.h 		for (cur = map->buckets[bkt]; cur; cur = cur->next)
next              153 tools/lib/bpf/hashmap.h 		     cur && ({tmp = cur->next; true; });		    \
next              167 tools/lib/bpf/hashmap.h 	     cur = cur->next)						    \
next              174 tools/lib/bpf/hashmap.h 	     cur && ({ tmp = cur->next; true; });			    \
next             4229 tools/lib/bpf/libbpf.c 	struct bpf_object *next;
next             4232 tools/lib/bpf/libbpf.c 		next = list_first_entry(&bpf_objects_list,
next             4236 tools/lib/bpf/libbpf.c 		next = list_next_entry(prev, list);
next             4239 tools/lib/bpf/libbpf.c 	if (&next->list == &bpf_objects_list)
next             4242 tools/lib/bpf/libbpf.c 	return next;
next             4320 tools/lib/bpf/libbpf.c bpf_program__prev(struct bpf_program *next, const struct bpf_object *obj)
next             4322 tools/lib/bpf/libbpf.c 	struct bpf_program *prog = next;
next             4759 tools/lib/bpf/libbpf.c bpf_map__prev(const struct bpf_map *next, const struct bpf_object *obj)
next             4761 tools/lib/bpf/libbpf.c 	if (next == NULL) {
next             4767 tools/lib/bpf/libbpf.c 	return __bpf_map__iter(next, obj, -1);
next               75 tools/lib/traceevent/event-parse.c 	struct event_handler		*next;
next               84 tools/lib/traceevent/event-parse.c 	struct func_params	*next;
next               89 tools/lib/traceevent/event-parse.c 	struct tep_function_handler	*next;
next              165 tools/lib/traceevent/event-parse.c 	struct cmdline_list	*next;
next              187 tools/lib/traceevent/event-parse.c 		cmdlist = cmdlist->next;
next              354 tools/lib/traceevent/event-parse.c 	item->next = tep->cmdlist;
next              403 tools/lib/traceevent/event-parse.c 	struct func_list	*next;
next              463 tools/lib/traceevent/event-parse.c 		funclist = funclist->next;
next              622 tools/lib/traceevent/event-parse.c 	item->next = tep->funclist;
next              679 tools/lib/traceevent/event-parse.c 	struct printk_list	*next;
next              716 tools/lib/traceevent/event-parse.c 		printklist = printklist->next;
next              763 tools/lib/traceevent/event-parse.c 	item->next = tep->printklist;
next              857 tools/lib/traceevent/event-parse.c 	struct tep_print_flag_sym *next;
next              860 tools/lib/traceevent/event-parse.c 		next = fsym->next;
next              864 tools/lib/traceevent/event-parse.c 		fsym = next;
next              924 tools/lib/traceevent/event-parse.c 			arg->func.args = farg->next;
next             1698 tools/lib/traceevent/event-parse.c 		fields = &field->next;
next             2579 tools/lib/traceevent/event-parse.c 		list = &field->next;
next             2983 tools/lib/traceevent/event-parse.c 	for (func = tep->func_handlers; func; func = func->next) {
next             2994 tools/lib/traceevent/event-parse.c 	struct tep_function_handler **next;
next             2996 tools/lib/traceevent/event-parse.c 	next = &tep->func_handlers;
next             2997 tools/lib/traceevent/event-parse.c 	while ((func = *next)) {
next             2999 tools/lib/traceevent/event-parse.c 			*next = func->next;
next             3003 tools/lib/traceevent/event-parse.c 		next = &func->next;
next             3050 tools/lib/traceevent/event-parse.c 		next_arg = &(farg->next);
next             3249 tools/lib/traceevent/event-parse.c 			list = &arg->next;
next             3256 tools/lib/traceevent/event-parse.c 			list = &arg->next;
next             3339 tools/lib/traceevent/event-parse.c 	     format; format = format->next) {
next             3361 tools/lib/traceevent/event-parse.c 	     format; format = format->next) {
next             4020 tools/lib/traceevent/event-parse.c 		for (flag = arg->flags.flags; flag; flag = flag->next) {
next             4042 tools/lib/traceevent/event-parse.c 		for (flag = arg->symbol.symbols; flag; flag = flag->next) {
next             4203 tools/lib/traceevent/event-parse.c 		struct save_str *next;
next             4238 tools/lib/traceevent/event-parse.c 			string->next = strings;
next             4258 tools/lib/traceevent/event-parse.c 		farg = farg->next;
next             4259 tools/lib/traceevent/event-parse.c 		param = param->next;
next             4267 tools/lib/traceevent/event-parse.c 		strings = string->next;
next             4279 tools/lib/traceevent/event-parse.c 	struct tep_print_arg *next;
next             4282 tools/lib/traceevent/event-parse.c 		next = args->next;
next             4285 tools/lib/traceevent/event-parse.c 		args = next;
next             4293 tools/lib/traceevent/event-parse.c 	struct tep_print_arg *args, *arg, **next;
next             4329 tools/lib/traceevent/event-parse.c 	arg->next = NULL;
next             4330 tools/lib/traceevent/event-parse.c 	next = &arg->next;
next             4430 tools/lib/traceevent/event-parse.c 				arg->next = NULL;
next             4436 tools/lib/traceevent/event-parse.c 				*next = arg;
next             4437 tools/lib/traceevent/event-parse.c 				next = &arg->next;
next             4454 tools/lib/traceevent/event-parse.c 				arg->next = NULL;
next             4460 tools/lib/traceevent/event-parse.c 				*next = arg;
next             4461 tools/lib/traceevent/event-parse.c 				next = &arg->next;
next             4949 tools/lib/traceevent/event-parse.c 		field = field->next;
next             5038 tools/lib/traceevent/event-parse.c 				arg = arg->next;
next             5057 tools/lib/traceevent/event-parse.c 					arg = arg->next;
next             5066 tools/lib/traceevent/event-parse.c 					arg = arg->next;
next             5074 tools/lib/traceevent/event-parse.c 						arg = arg->next;
next             5105 tools/lib/traceevent/event-parse.c 				arg = arg->next;
next             5194 tools/lib/traceevent/event-parse.c 				arg = arg->next;
next             5370 tools/lib/traceevent/event-parse.c pid_from_cmdlist(struct tep_handle *tep, const char *comm, struct tep_cmdline *next)
next             5372 tools/lib/traceevent/event-parse.c 	struct cmdline_list *cmdlist = (struct cmdline_list *)next;
next             5375 tools/lib/traceevent/event-parse.c 		cmdlist = cmdlist->next;
next             5380 tools/lib/traceevent/event-parse.c 		cmdlist = cmdlist->next;
next             5399 tools/lib/traceevent/event-parse.c 					   struct tep_cmdline *next)
next             5408 tools/lib/traceevent/event-parse.c 		return pid_from_cmdlist(tep, comm, next);
next             5410 tools/lib/traceevent/event-parse.c 	if (next) {
next             5415 tools/lib/traceevent/event-parse.c 		if (next < tep->cmdlines ||
next             5416 tools/lib/traceevent/event-parse.c 		    next >= tep->cmdlines + tep->cmdline_count)
next             5417 tools/lib/traceevent/event-parse.c 			next = NULL;
next             5419 tools/lib/traceevent/event-parse.c 			cmdline  = next++;
next             5422 tools/lib/traceevent/event-parse.c 	if (!next)
next             5872 tools/lib/traceevent/event-parse.c 	for (field = list; field; field = field->next) {
next             5922 tools/lib/traceevent/event-parse.c 	if (field->next) {
next             5924 tools/lib/traceevent/event-parse.c 		print_fields(s, field->next);
next             6013 tools/lib/traceevent/event-parse.c 	if (args->next) {
next             6015 tools/lib/traceevent/event-parse.c 		print_args(args->next);
next             6179 tools/lib/traceevent/event-parse.c 	struct event_handler *handle, **next;
next             6181 tools/lib/traceevent/event-parse.c 	for (next = &tep->handlers; *next;
next             6182 tools/lib/traceevent/event-parse.c 	     next = &(*next)->next) {
next             6183 tools/lib/traceevent/event-parse.c 		handle = *next;
next             6190 tools/lib/traceevent/event-parse.c 	if (!(*next))
next             6199 tools/lib/traceevent/event-parse.c 	*next = handle->next;
next             6291 tools/lib/traceevent/event-parse.c 		for (field = event->format.fields; field; field = field->next) {
next             6306 tools/lib/traceevent/event-parse.c 			list = &arg->next;
next             6625 tools/lib/traceevent/event-parse.c 		func->params = params->next;
next             6705 tools/lib/traceevent/event-parse.c 		param->next = NULL;
next             6708 tools/lib/traceevent/event-parse.c 		next_param = &(param->next);
next             6714 tools/lib/traceevent/event-parse.c 	func_handle->next = tep->func_handlers;
next             6835 tools/lib/traceevent/event-parse.c 	handle->next = tep->handlers;
next             6883 tools/lib/traceevent/event-parse.c 	struct event_handler **next;
next             6899 tools/lib/traceevent/event-parse.c 	for (next = &tep->handlers; *next; next = &(*next)->next) {
next             6900 tools/lib/traceevent/event-parse.c 		handle = *next;
next             6906 tools/lib/traceevent/event-parse.c 	if (!(*next))
next             6909 tools/lib/traceevent/event-parse.c 	*next = handle->next;
next             6953 tools/lib/traceevent/event-parse.c 	struct tep_format_field *next;
next             6956 tools/lib/traceevent/event-parse.c 		next = field->next;
next             6958 tools/lib/traceevent/event-parse.c 		field = next;
next             7012 tools/lib/traceevent/event-parse.c 		cmdnext = cmdlist->next;
next             7027 tools/lib/traceevent/event-parse.c 		funcnext = funclist->next;
next             7036 tools/lib/traceevent/event-parse.c 		tep->func_handlers = func_handler->next;
next             7047 tools/lib/traceevent/event-parse.c 		printknext = printklist->next;
next             7058 tools/lib/traceevent/event-parse.c 		tep->handlers = handle->next;
next               52 tools/lib/traceevent/event-parse.h 	struct tep_record	*next;
next               71 tools/lib/traceevent/event-parse.h 	struct tep_plugin_option	*next;
next              145 tools/lib/traceevent/event-parse.h 	struct tep_format_field	*next;
next              184 tools/lib/traceevent/event-parse.h 	struct tep_print_flag_sym	*next;
next              257 tools/lib/traceevent/event-parse.h 	struct tep_print_arg		*next;
next              536 tools/lib/traceevent/event-parse.h 					   struct tep_cmdline *next);
next               24 tools/lib/traceevent/event-plugin.c 	struct registered_plugin_options	*next;
next               29 tools/lib/traceevent/event-plugin.c 	struct trace_plugin_options	*next;
next               36 tools/lib/traceevent/event-plugin.c 	struct tep_plugin_list	*next;
next              107 tools/lib/traceevent/event-plugin.c 	for (reg = registered_options; reg; reg = reg->next) {
next              175 tools/lib/traceevent/event-plugin.c 	for (op = trace_plugin_options; op; op = op->next) {
next              190 tools/lib/traceevent/event-plugin.c 	for (op = trace_plugin_options; op; op = op->next) {
next              220 tools/lib/traceevent/event-plugin.c 	reg->next = registered_options;
next              240 tools/lib/traceevent/event-plugin.c 	for (last = &registered_options; *last; last = &(*last)->next) {
next              243 tools/lib/traceevent/event-plugin.c 			*last = reg->next;
next              267 tools/lib/traceevent/event-plugin.c 		list = list->next;
next              313 tools/lib/traceevent/event-plugin.c 	list->next = *plugin_list;
next              438 tools/lib/traceevent/event-plugin.c 		plugin_list = list->next;
next               51 tools/lib/traceevent/kbuffer-parse.c 	unsigned int		next;
next              294 tools/lib/traceevent/kbuffer-parse.c 		kbuf->next = kbuf->size;
next              309 tools/lib/traceevent/kbuffer-parse.c 		kbuf->next = kbuf->size;
next              325 tools/lib/traceevent/kbuffer-parse.c 	kbuf->next = kbuf->index + length;
next              335 tools/lib/traceevent/kbuffer-parse.c 		kbuf->curr = kbuf->next;
next              336 tools/lib/traceevent/kbuffer-parse.c 		if (kbuf->next >= kbuf->size)
next              402 tools/lib/traceevent/kbuffer-parse.c 	kbuf->next = kbuf->index + length;
next              453 tools/lib/traceevent/kbuffer-parse.c 		kbuf->curr = kbuf->next;
next              454 tools/lib/traceevent/kbuffer-parse.c 		if (kbuf->next >= kbuf->size)
next              542 tools/lib/traceevent/kbuffer-parse.c 	kbuf->next = 0;
next              671 tools/lib/traceevent/kbuffer-parse.c 	return kbuf->next - kbuf->index;
next              683 tools/lib/traceevent/kbuffer-parse.c 	return kbuf->next - kbuf->curr;
next              753 tools/lib/traceevent/kbuffer-parse.c 	void *ptr = info->next;
next              771 tools/lib/traceevent/kbuffer-parse.c 	info->next = ptr + length;
next               74 tools/lib/traceevent/kbuffer.h 	void			*next;
next               29 tools/lib/traceevent/parse-filter.c 	struct event_list	*next;
next              240 tools/lib/traceevent/parse-filter.c 	list->next = *events;
next              331 tools/lib/traceevent/parse-filter.c 		events = events->next;
next             1332 tools/lib/traceevent/parse-filter.c 	for (event = events; event; event = event->next) {
next               47 tools/objtool/check.c 	struct instruction *next = list_next_entry(insn, list);
next               49 tools/objtool/check.c 	if (!next || &next->list == &file->insn_list || next->sec != insn->sec)
next               52 tools/objtool/check.c 	return next;
next               58 tools/objtool/check.c 	struct instruction *next = list_next_entry(insn, list);
next               64 tools/objtool/check.c 	if (&next->list != &file->insn_list && next->func == func)
next               65 tools/objtool/check.c 		return next;
next              313 tools/perf/builtin-annotate.c 	struct rb_node *nd = rb_first_cached(&hists->entries), *next;
next              365 tools/perf/builtin-annotate.c 				next = rb_next(nd);
next              368 tools/perf/builtin-annotate.c 				next = rb_prev(nd);
next              374 tools/perf/builtin-annotate.c 			if (next != NULL)
next              375 tools/perf/builtin-annotate.c 				nd = next;
next             2106 tools/perf/builtin-c2c.c 	struct rb_node *next = rb_first_cached(&hists->entries);
next             2109 tools/perf/builtin-c2c.c 	while (next) {
next             2112 tools/perf/builtin-c2c.c 		he = rb_entry(next, struct hist_entry, rb_node);
next             2116 tools/perf/builtin-c2c.c 		next = rb_next(&he->rb_node);
next              514 tools/perf/builtin-diff.c 	struct rb_node *next;
next              521 tools/perf/builtin-diff.c 	next = rb_first_cached(root);
next              522 tools/perf/builtin-diff.c 	while (next != NULL) {
next              523 tools/perf/builtin-diff.c 		struct hist_entry *he = rb_entry(next, struct hist_entry, rb_node_in);
next              525 tools/perf/builtin-diff.c 		next = rb_next(&he->rb_node_in);
next              675 tools/perf/builtin-diff.c 	struct rb_node *next = rb_first_cached(root);
next              678 tools/perf/builtin-diff.c 	while (next != NULL) {
next              679 tools/perf/builtin-diff.c 		struct hist_entry *he_pair = rb_entry(next, struct hist_entry,
next              682 tools/perf/builtin-diff.c 		next = rb_next(&he_pair->rb_node_in);
next              707 tools/perf/builtin-diff.c 	struct rb_node *next = rb_first_cached(root);
next              709 tools/perf/builtin-diff.c 	while (next != NULL) {
next              710 tools/perf/builtin-diff.c 		struct hist_entry *he = rb_entry(next, struct hist_entry,
next              714 tools/perf/builtin-diff.c 		next = rb_next(&he->rb_node_in);
next              733 tools/perf/builtin-diff.c 	struct rb_node *next;
next              740 tools/perf/builtin-diff.c 	next = rb_first_cached(root);
next              741 tools/perf/builtin-diff.c 	while (next != NULL) {
next              747 tools/perf/builtin-diff.c 		he   = rb_entry(next, struct hist_entry, rb_node_in);
next              748 tools/perf/builtin-diff.c 		next = rb_next(&he->rb_node_in);
next               29 tools/perf/builtin-help.c 	struct man_viewer_list *next;
next               34 tools/perf/builtin-help.c 	struct man_viewer_info_list *next;
next               63 tools/perf/builtin-help.c 	for (viewer = man_viewer_info_list; viewer; viewer = viewer->next) {
next              195 tools/perf/builtin-help.c 		p = &((*p)->next);
next              215 tools/perf/builtin-help.c 	new->next = man_viewer_info_list;
next              368 tools/perf/builtin-help.c 	for (viewer = man_viewer_list; viewer; viewer = viewer->next)
next              993 tools/perf/builtin-kmem.c 	struct rb_node *next;
next             1001 tools/perf/builtin-kmem.c 	next = rb_first(root);
next             1003 tools/perf/builtin-kmem.c 	while (next && n_lines--) {
next             1004 tools/perf/builtin-kmem.c 		struct alloc_stat *data = rb_entry(next, struct alloc_stat,
next             1034 tools/perf/builtin-kmem.c 		next = rb_next(next);
next             1054 tools/perf/builtin-kmem.c 	struct rb_node *next = rb_first(&page_alloc_sorted);
next             1070 tools/perf/builtin-kmem.c 	while (next && n_lines--) {
next             1077 tools/perf/builtin-kmem.c 		data = rb_entry(next, struct page_stat, node);
next             1090 tools/perf/builtin-kmem.c 		next = rb_next(next);
next             1103 tools/perf/builtin-kmem.c 	struct rb_node *next = rb_first(&page_caller_sorted);
next             1112 tools/perf/builtin-kmem.c 	while (next && n_lines--) {
next             1119 tools/perf/builtin-kmem.c 		data = rb_entry(next, struct page_stat, node);
next             1132 tools/perf/builtin-kmem.c 		next = rb_next(next);
next              838 tools/perf/builtin-sched.c 	struct task_desc *prev, __maybe_unused *next;
next              864 tools/perf/builtin-sched.c 	next = register_pid(sched, next_pid, next_comm);
next             3140 tools/perf/builtin-sched.c 	struct rb_node *next;
next             3154 tools/perf/builtin-sched.c 	next = rb_first_cached(&sched->sorted_atom_root);
next             3156 tools/perf/builtin-sched.c 	while (next) {
next             3159 tools/perf/builtin-sched.c 		work_list = rb_entry(next, struct work_atoms, node);
next             3161 tools/perf/builtin-sched.c 		next = rb_next(next);
next               90 tools/perf/builtin-timechart.c 	struct per_pid *next;
next              107 tools/perf/builtin-timechart.c 	struct per_pidcomm *next;
next              128 tools/perf/builtin-timechart.c 	struct sample_wrapper *next;
next              140 tools/perf/builtin-timechart.c 	struct cpu_sample *next;
next              159 tools/perf/builtin-timechart.c 	struct io_sample *next;
next              174 tools/perf/builtin-timechart.c 	struct power_event *next;
next              183 tools/perf/builtin-timechart.c 	struct wake_event *next;
next              193 tools/perf/builtin-timechart.c 	struct process_filter	*next;
next              206 tools/perf/builtin-timechart.c 		cursor = cursor->next;
next              211 tools/perf/builtin-timechart.c 	cursor->next = tchart->all_data;
next              232 tools/perf/builtin-timechart.c 		c = c->next;
next              238 tools/perf/builtin-timechart.c 	c->next = p->all;
next              281 tools/perf/builtin-timechart.c 		c->next = p->all;
next              290 tools/perf/builtin-timechart.c 	sample->next = c->samples;
next              365 tools/perf/builtin-timechart.c 	pwr->next = tchart->power_events;
next              386 tools/perf/builtin-timechart.c 	pwr->next = tchart->power_events;
next              423 tools/perf/builtin-timechart.c 	we->next = tchart->wake_events;
next              693 tools/perf/builtin-timechart.c 		pwr->next = tchart->power_events;
next              708 tools/perf/builtin-timechart.c 		pwr->next = tchart->power_events;
next              731 tools/perf/builtin-timechart.c 		c->next = p->all;
next              743 tools/perf/builtin-timechart.c 		c->io_samples = prev->next;
next              754 tools/perf/builtin-timechart.c 	sample->next = c->io_samples;
next              792 tools/perf/builtin-timechart.c 	prev = sample->next;
next              829 tools/perf/builtin-timechart.c 		sample->next = prev->next;
next              973 tools/perf/builtin-timechart.c 		tchart->all_data = p->next;
next              974 tools/perf/builtin-timechart.c 		p->next = NULL;
next              978 tools/perf/builtin-timechart.c 			p->next = NULL;
next              988 tools/perf/builtin-timechart.c 					p->next = prev->next;
next              989 tools/perf/builtin-timechart.c 					prev->next = p;
next              993 tools/perf/builtin-timechart.c 					p->next = new_list;
next             1001 tools/perf/builtin-timechart.c 			cursor = cursor->next;
next             1003 tools/perf/builtin-timechart.c 				prev->next = p;
next             1021 tools/perf/builtin-timechart.c 		pwr = pwr->next;
next             1031 tools/perf/builtin-timechart.c 		pwr = pwr->next;
next             1062 tools/perf/builtin-timechart.c 					c = c->next;
next             1074 tools/perf/builtin-timechart.c 					c = c->next;
next             1077 tools/perf/builtin-timechart.c 			p = p->next;
next             1096 tools/perf/builtin-timechart.c 		we = we->next;
next             1123 tools/perf/builtin-timechart.c 				sample = sample->next;
next             1125 tools/perf/builtin-timechart.c 			c = c->next;
next             1127 tools/perf/builtin-timechart.c 		p = p->next;
next             1147 tools/perf/builtin-timechart.c 				c = c->next;
next             1153 tools/perf/builtin-timechart.c 			for (sample = c->io_samples; sample; sample = sample->next) {
next             1240 tools/perf/builtin-timechart.c 			c = c->next;
next             1242 tools/perf/builtin-timechart.c 		p = p->next;
next             1261 tools/perf/builtin-timechart.c 				c = c->next;
next             1283 tools/perf/builtin-timechart.c 				sample = sample->next;
next             1297 tools/perf/builtin-timechart.c 			c = c->next;
next             1299 tools/perf/builtin-timechart.c 		p = p->next;
next             1313 tools/perf/builtin-timechart.c 	filt->next = process_filter;
next             1330 tools/perf/builtin-timechart.c 		filt = filt->next;
next             1368 tools/perf/builtin-timechart.c 			c = c->next;
next             1370 tools/perf/builtin-timechart.c 		p = p->next;
next             1409 tools/perf/builtin-timechart.c 			c = c->next;
next             1411 tools/perf/builtin-timechart.c 		p = p->next;
next             1441 tools/perf/builtin-timechart.c 			c = c->next;
next             1443 tools/perf/builtin-timechart.c 		p = p->next;
next              392 tools/perf/builtin-top.c 	struct rb_node *next;
next              409 tools/perf/builtin-top.c 	next = rb_first_cached(&hists->entries);
next              410 tools/perf/builtin-top.c 	while (next) {
next              411 tools/perf/builtin-top.c 		n = rb_entry(next, struct hist_entry, rb_node);
next              416 tools/perf/builtin-top.c 		next = rb_next(&n->rb_node);
next             1461 tools/perf/builtin-trace.c 	for (field = sc->args; field; field = field->next, ++idx) {
next             1545 tools/perf/builtin-trace.c 		sc->args = sc->args->next;
next             1735 tools/perf/builtin-trace.c 		     field = field->next, ++arg.idx, bit <<= 1) {
next             2935 tools/perf/builtin-trace.c 	for (field = sc->args; field; field = field->next) {
next             2953 tools/perf/builtin-trace.c 		     field && candidate_field; field = field->next, candidate_field = candidate_field->next) {
next             2985 tools/perf/builtin-trace.c 			for (candidate_field = candidate_field->next; candidate_field; candidate_field = candidate_field->next)
next               82 tools/perf/jvmti/libjvmti.c 	for (hdr = compile_info; hdr != NULL; hdr = hdr->next) {
next              113 tools/perf/jvmti/libjvmti.c 	for (hdr = compile_info; hdr != NULL; hdr = hdr->next) {
next               87 tools/perf/lib/evlist.c 	struct perf_evsel *next;
next               90 tools/perf/lib/evlist.c 		next = list_first_entry(&evlist->entries,
next               94 tools/perf/lib/evlist.c 		next = list_next_entry(prev, node);
next               98 tools/perf/lib/evlist.c 	if (&next->node == &evlist->entries)
next              101 tools/perf/lib/evlist.c 	return next;
next               65 tools/perf/lib/include/internal/evlist.h 	return list_entry(evlist->entries.next, struct perf_evsel, node);
next              407 tools/perf/pmu-events/jevents.c 	struct event_struct *es, *next;
next              409 tools/perf/pmu-events/jevents.c 	list_for_each_entry_safe(es, next, &arch_std_events, list) {
next              599 tools/perf/tests/parse-events.c 	term = list_entry(terms->next, struct parse_events_term, list);
next              608 tools/perf/tests/parse-events.c 	term = list_entry(term->list.next, struct parse_events_term, list);
next              617 tools/perf/tests/parse-events.c 	term = list_entry(term->list.next, struct parse_events_term, list);
next              626 tools/perf/tests/parse-events.c 	term = list_entry(term->list.next, struct parse_events_term, list);
next              245 tools/perf/tests/switch-tracking.c 		node = list_entry(events->next, struct event_node, list);
next               84 tools/perf/ui/browser.c 		pos = pos->next;
next              113 tools/perf/ui/browser.c 		pos = ui_browser__list_head_filter_entries(browser, head->next);
next              129 tools/perf/ui/browser.c 			pos = ui_browser__list_head_filter_entries(browser, pos->next);
next              505 tools/perf/ui/browser.c                 browser->top = ui_browser__list_head_filter_entries(browser, head->next);
next              337 tools/perf/ui/browsers/hists.c 				chain->has_children = chain->list.next != &child->val ||
next              340 tools/perf/ui/browsers/hists.c 				chain->has_children = chain->list.next == &child->val &&
next              353 tools/perf/ui/browsers/hists.c 	chain = list_entry(node->val.next, struct callchain_list, list);
next              874 tools/perf/ui/browsers/hists.c 		struct rb_node *next = rb_next(node);
next              900 tools/perf/ui/browsers/hists.c 				goto next;
next              926 tools/perf/ui/browsers/hists.c next:
next              929 tools/perf/ui/browsers/hists.c 		node = next;
next              978 tools/perf/ui/browsers/hists.c 		struct rb_node *next = rb_next(node);
next              986 tools/perf/ui/browsers/hists.c 			goto next;
next             1037 tools/perf/ui/browsers/hists.c next:
next             1040 tools/perf/ui/browsers/hists.c 		node = next;
next             1067 tools/perf/ui/browsers/hists.c 		struct rb_node *next = rb_next(node);
next             1106 tools/perf/ui/browsers/hists.c 		node = next;
next             2800 tools/perf/ui/browsers/hists.c 			goto next;
next             2814 tools/perf/ui/browsers/hists.c next:
next             3322 tools/perf/ui/browsers/hists.c 				if (pos->core.node.next == &evlist->core.entries)
next              486 tools/perf/ui/hist.c 		if (fmt->sort_list.next == NULL)
next              583 tools/perf/ui/hist.c 				goto next;
next              587 tools/perf/ui/hist.c next:
next              602 tools/perf/ui/hist.c 				goto next;
next              606 tools/perf/ui/hist.c next:
next               22 tools/perf/ui/progress.c 	if (p->curr >= p->next) {
next               25 tools/perf/ui/progress.c 		p->next += nr * p->step;
next               34 tools/perf/ui/progress.c 	p->next = p->step = total / 16 ?: 1;
next               11 tools/perf/ui/progress.h 	u64 curr, next, step, total;
next              112 tools/perf/ui/stdio/hist.c 	struct rb_node *node, *next;
next              141 tools/perf/ui/stdio/hist.c 		next = rb_next(node);
next              142 tools/perf/ui/stdio/hist.c 		if (!next && (callchain_param.mode != CHAIN_GRAPH_REL || !remaining))
next              168 tools/perf/ui/stdio/hist.c 		node = next;
next               12 tools/perf/ui/tui/progress.c 	p->next = p->step = p->total / (SLtt_Screen_Cols - 2) ?: 1;
next             2039 tools/perf/util/annotate.c 	struct annotation_line *al, *next;
next             2049 tools/perf/util/annotate.c 		next = annotation_line__next(al, &notes->src->source);
next             2050 tools/perf/util/annotate.c 		end  = next ? next->offset : len;
next             2191 tools/perf/util/annotate.c 		struct rb_node *next;
next             2194 tools/perf/util/annotate.c 		next = rb_next(node);
next             2198 tools/perf/util/annotate.c 		node = next;
next             2352 tools/perf/util/annotate.c 				queue = list_entry(queue->node.next, typeof(*queue), node);
next              412 tools/perf/util/auxtrace.c 			buffer = list_entry(queues->queue_array[i].head.next,
next              804 tools/perf/util/auxtrace.c 		return list_entry(buffer->list.next, struct auxtrace_buffer,
next              809 tools/perf/util/auxtrace.c 		return list_entry(queue->head.next, struct auxtrace_buffer,
next               85 tools/perf/util/block-range.c 	struct block_range *next, *entry = NULL;
next              117 tools/perf/util/block-range.c 		next = rb_entry(n, struct block_range, node);
next              119 tools/perf/util/block-range.c 		if (next->start <= end) { /* add head: [start...][n->start...] */
next              126 tools/perf/util/block-range.c 				.end		= next->start - 1,
next              131 tools/perf/util/block-range.c 			rb_link_left_of_node(&head->node, &next->node);
next              242 tools/perf/util/block-range.c 		next = block_range__next(entry);
next              243 tools/perf/util/block-range.c 		if (!next)
next              249 tools/perf/util/block-range.c 		if (end < next->start) { /* add tail: [...e->end][...end] */
next              274 tools/perf/util/block-range.c 		if (entry->end + 1 != next->start) {
next              281 tools/perf/util/block-range.c 				.end		= next->start - 1,
next              286 tools/perf/util/block-range.c 			rb_link_left_of_node(&hole->node, &next->node);
next              291 tools/perf/util/block-range.c 		entry = next;
next              514 tools/perf/util/bpf-loader.c 			ref1 = ref1->next;
next              515 tools/perf/util/bpf-loader.c 			ref2 = ref2->next;
next              289 tools/perf/util/bpf-prologue.c 			ref = ref->next;
next              427 tools/perf/util/bpf-prologue.c 			ref = ref->next;
next               58 tools/perf/util/call-path.c 	if (cpr->next < cpr->sz) {
next               69 tools/perf/util/call-path.c 	n = cpr->next++ & CALL_PATH_BLOCK_MASK;
next               57 tools/perf/util/call-path.h 	size_t next;
next              196 tools/perf/util/callchain.c 			goto next;
next              206 tools/perf/util/callchain.c 			goto next;
next              229 tools/perf/util/callchain.c next:
next              806 tools/perf/util/callchain.c 	new->val.next = &to_split->list;
next              809 tools/perf/util/callchain.c 	old_tail->next = &new->val;
next             1078 tools/perf/util/callchain.c 	cursor->last = &node->next;
next             1599 tools/perf/util/callchain.c 	for (node = cursor->first; node != NULL; node = node->next)
next              152 tools/perf/util/callchain.h 	struct callchain_cursor_node	*next;
next              222 tools/perf/util/callchain.h 	cursor->curr = cursor->curr->next;
next              405 tools/perf/util/data-convert-bt.c 	for (field = fields; field; field = field->next) {
next             1018 tools/perf/util/data-convert-bt.c 	for (field = fields; field; field = field->next) {
next              757 tools/perf/util/dso.c 	struct rb_node *next = rb_first(root);
next              760 tools/perf/util/dso.c 	while (next) {
next              763 tools/perf/util/dso.c 		cache = rb_entry(next, struct dso_cache, rb_node);
next              764 tools/perf/util/dso.c 		next = rb_next(&cache->rb_node);
next              130 tools/perf/util/env.c 	struct rb_node *next;
next              135 tools/perf/util/env.c 	next = rb_first(root);
next              137 tools/perf/util/env.c 	while (next) {
next              140 tools/perf/util/env.c 		node = rb_entry(next, struct bpf_prog_info_node, rb_node);
next              141 tools/perf/util/env.c 		next = rb_next(&node->rb_node);
next              149 tools/perf/util/env.c 	next = rb_first(root);
next              151 tools/perf/util/env.c 	while (next) {
next              154 tools/perf/util/env.c 		node = rb_entry(next, struct btf_node, rb_node);
next              155 tools/perf/util/env.c 		next = rb_next(&node->rb_node);
next              193 tools/perf/util/evlist.c 	leader = list_entry(list->next, struct evsel, core.node);
next              309 tools/perf/util/evsel.h 	return list_entry(evsel->core.node.next, struct evsel, core.node);
next              370 tools/perf/util/evsel.h for ((_evsel) = list_entry((_leader)->core.node.next, struct evsel, core.node); \
next              372 tools/perf/util/evsel.h      (_evsel) = list_entry((_evsel)->core.node.next, struct evsel, core.node))
next              378 tools/perf/util/evsel.h      (_evsel) = list_entry((_evsel)->core.node.next, struct evsel, core.node))
next               93 tools/perf/util/evsel_fprintf.c 		field = field->next;
next               96 tools/perf/util/evsel_fprintf.c 			field = field->next;
next              135 tools/perf/util/evsel_fprintf.c 				goto next;
next              185 tools/perf/util/evsel_fprintf.c next:
next              920 tools/perf/util/header.c 	struct rb_node *next;
next              931 tools/perf/util/header.c 	next = rb_first(root);
next              932 tools/perf/util/header.c 	while (next) {
next              936 tools/perf/util/header.c 		node = rb_entry(next, struct bpf_prog_info_node, rb_node);
next              937 tools/perf/util/header.c 		next = rb_next(&node->rb_node);
next              969 tools/perf/util/header.c 	struct rb_node *next;
next              981 tools/perf/util/header.c 	next = rb_first(root);
next              982 tools/perf/util/header.c 	while (next) {
next              985 tools/perf/util/header.c 		node = rb_entry(next, struct btf_node, rb_node);
next              986 tools/perf/util/header.c 		next = rb_next(&node->rb_node);
next             1536 tools/perf/util/header.c 	struct rb_node *next;
next             1541 tools/perf/util/header.c 	next = rb_first(root);
next             1543 tools/perf/util/header.c 	while (next) {
next             1546 tools/perf/util/header.c 		node = rb_entry(next, struct bpf_prog_info_node, rb_node);
next             1547 tools/perf/util/header.c 		next = rb_next(&node->rb_node);
next             1560 tools/perf/util/header.c 	struct rb_node *next;
next             1565 tools/perf/util/header.c 	next = rb_first(root);
next             1567 tools/perf/util/header.c 	while (next) {
next             1570 tools/perf/util/header.c 		node = rb_entry(next, struct btf_node, rb_node);
next             1571 tools/perf/util/header.c 		next = rb_next(&node->rb_node);
next              226 tools/perf/util/hist.c 	struct rb_node *next = rb_first_cached(&hists->entries);
next              232 tools/perf/util/hist.c 	while (next && row++ < max_rows) {
next              233 tools/perf/util/hist.c 		n = rb_entry(next, struct hist_entry, rb_node);
next              236 tools/perf/util/hist.c 		next = rb_next(&n->rb_node);
next              362 tools/perf/util/hist.c 	struct rb_node *next = rb_first_cached(&hists->entries);
next              365 tools/perf/util/hist.c 	while (next) {
next              366 tools/perf/util/hist.c 		n = rb_entry(next, struct hist_entry, rb_node);
next              367 tools/perf/util/hist.c 		next = rb_next(&n->rb_node);
next              378 tools/perf/util/hist.c 	struct rb_node *next = rb_first_cached(&hists->entries);
next              381 tools/perf/util/hist.c 	while (next) {
next              382 tools/perf/util/hist.c 		n = rb_entry(next, struct hist_entry, rb_node);
next              383 tools/perf/util/hist.c 		next = rb_next(&n->rb_node);
next              391 tools/perf/util/hist.c 	struct rb_node *next = rb_first_cached(&hists->entries);
next              395 tools/perf/util/hist.c 	while (next) {
next              396 tools/perf/util/hist.c 		n = rb_entry(next, struct hist_entry, rb_node);
next              400 tools/perf/util/hist.c 		next = rb_next(&n->rb_node);
next             1590 tools/perf/util/hist.c 	struct rb_node *next;
next             1601 tools/perf/util/hist.c 	next = rb_first_cached(root);
next             1603 tools/perf/util/hist.c 	while (next) {
next             1606 tools/perf/util/hist.c 		n = rb_entry(next, struct hist_entry, rb_node_in);
next             1607 tools/perf/util/hist.c 		next = rb_next(&n->rb_node_in);
next             1836 tools/perf/util/hist.c 	struct rb_node *next;
next             1865 tools/perf/util/hist.c 	next = rb_first_cached(root);
next             1868 tools/perf/util/hist.c 	while (next) {
next             1869 tools/perf/util/hist.c 		n = rb_entry(next, struct hist_entry, rb_node_in);
next             1870 tools/perf/util/hist.c 		next = rb_next(&n->rb_node_in);
next             2012 tools/perf/util/hist.c 				goto next;
next             2019 tools/perf/util/hist.c next:
next             1886 tools/perf/util/intel-pt-decoder/intel-pt-decoder.c next:
next             1975 tools/perf/util/intel-pt-decoder/intel-pt-decoder.c 				goto next;
next             2042 tools/perf/util/intel-pt-decoder/intel-pt-decoder.c 			goto next;
next             2065 tools/perf/util/intel-pt-decoder/intel-pt-decoder.c 			goto next;
next             2103 tools/perf/util/intel-pt-decoder/intel-pt-decoder.c 			goto next;
next             2143 tools/perf/util/intel-pt-decoder/intel-pt-decoder.c 			goto next;
next             2490 tools/perf/util/intel-pt-decoder/intel-pt-decoder.c 	unsigned char *next;
next             2501 tools/perf/util/intel-pt-decoder/intel-pt-decoder.c 		next = memmem(decoder->buf, decoder->len, INTEL_PT_PSB_STR,
next             2503 tools/perf/util/intel-pt-decoder/intel-pt-decoder.c 		if (!next) {
next             2518 tools/perf/util/intel-pt-decoder/intel-pt-decoder.c 		decoder->pkt_step = next - decoder->buf;
next             2667 tools/perf/util/intel-pt-decoder/intel-pt-decoder.c 	unsigned char *next;
next             2669 tools/perf/util/intel-pt-decoder/intel-pt-decoder.c 	next = memmem(*buf, *len, INTEL_PT_PSB_STR, INTEL_PT_PSB_LEN);
next             2670 tools/perf/util/intel-pt-decoder/intel-pt-decoder.c 	if (next) {
next             2671 tools/perf/util/intel-pt-decoder/intel-pt-decoder.c 		*len -= next - *buf;
next             2672 tools/perf/util/intel-pt-decoder/intel-pt-decoder.c 		*buf = next;
next             2692 tools/perf/util/intel-pt-decoder/intel-pt-decoder.c 	unsigned char *next;
next             2697 tools/perf/util/intel-pt-decoder/intel-pt-decoder.c 	next = memmem(*buf + 1, *len - 1, INTEL_PT_PSB_STR, INTEL_PT_PSB_LEN);
next             2698 tools/perf/util/intel-pt-decoder/intel-pt-decoder.c 	if (next) {
next             2699 tools/perf/util/intel-pt-decoder/intel-pt-decoder.c 		*len -= next - *buf;
next             2700 tools/perf/util/intel-pt-decoder/intel-pt-decoder.c 		*buf = next;
next             1178 tools/perf/util/machine.c 	struct rb_node *next = rb_first_cached(&machines->guests);
next             1182 tools/perf/util/machine.c 	while (next) {
next             1183 tools/perf/util/machine.c 		struct machine *pos = rb_entry(next, struct machine, rb_node);
next             1185 tools/perf/util/machine.c 		next = rb_next(&pos->rb_node);
next              598 tools/perf/util/map.c 	struct rb_node *next = rb_first(root);
next              600 tools/perf/util/map.c 	while (next) {
next              601 tools/perf/util/map.c 		struct map *pos = rb_entry(next, struct map, rb_node);
next              603 tools/perf/util/map.c 		next = rb_next(&pos->rb_node);
next              612 tools/perf/util/map.c 	struct rb_node *next = rb_first(root);
next              614 tools/perf/util/map.c 	while (next) {
next              615 tools/perf/util/map.c 		struct map *pos = rb_entry(next, struct map, rb_node_name);
next              617 tools/perf/util/map.c 		next = rb_next(&pos->rb_node_name);
next              776 tools/perf/util/map.c 	struct rb_node *next, *first;
next              787 tools/perf/util/map.c 	next = root->rb_node;
next              789 tools/perf/util/map.c 	while (next) {
next              790 tools/perf/util/map.c 		struct map *pos = rb_entry(next, struct map, rb_node);
next              793 tools/perf/util/map.c 			first = next;
next              796 tools/perf/util/map.c 			next = next->rb_left;
next              798 tools/perf/util/map.c 			next = next->rb_right;
next              801 tools/perf/util/map.c 	next = first;
next              802 tools/perf/util/map.c 	while (next) {
next              803 tools/perf/util/map.c 		struct map *pos = rb_entry(next, struct map, rb_node);
next              804 tools/perf/util/map.c 		next = rb_next(&pos->rb_node);
next             1012 tools/perf/util/map.c 	struct rb_node *next = rb_next(&map->rb_node);
next             1014 tools/perf/util/map.c 	if (next)
next             1015 tools/perf/util/map.c 		return rb_entry(next, struct map, rb_node);
next              306 tools/perf/util/metricgroup.c 	struct rb_node *node, *next;
next              387 tools/perf/util/metricgroup.c 	for (node = rb_first_cached(&groups.entries); node; node = next) {
next              394 tools/perf/util/metricgroup.c 		next = rb_next(node);
next               42 tools/perf/util/ordered-events.c 			p = last->list.next;
next              141 tools/perf/util/ordered-events.c 		new = list_entry(cache->next, struct ordered_event, list);
next              296 tools/perf/util/ordered-events.c 		first = list_entry(head->next, struct ordered_event, list);
next              207 tools/perf/util/parse-events.c 			goto next;
next              248 tools/perf/util/parse-events.c next:
next             1575 tools/perf/util/parse-events.c 	leader = list_entry(list->next, struct evsel, core.node);
next             2225 tools/perf/util/parse-events.c 			goto next;
next             2248 tools/perf/util/parse-events.c next:
next             2309 tools/perf/util/parse-events.c 			goto next;
next             2321 tools/perf/util/parse-events.c next:
next               24 tools/perf/util/parse-events.h 	struct tracepoint_path *next;
next               39 tools/perf/util/parse-events.y 	if (!list_is_last(list->next, list))
next             1642 tools/perf/util/probe-event.c 			fieldp = &(*fieldp)->next;
next             1884 tools/perf/util/probe-event.c 		field = field->next;
next             1974 tools/perf/util/probe-event.c 	if (ref->next) {
next             1975 tools/perf/util/probe-event.c 		depth = __synthesize_probe_trace_arg_ref(ref->next, buf,
next             2000 tools/perf/util/probe-event.c 			ref = ref->next;
next             2214 tools/perf/util/probe-event.c 	struct perf_probe_arg_field *field, *next;
next             2228 tools/perf/util/probe-event.c 			next = field->next;
next             2231 tools/perf/util/probe-event.c 			field = next;
next             2275 tools/perf/util/probe-event.c 		field = field->next;
next             2276 tools/perf/util/probe-event.c 		ppfield = &((*ppfield)->next);
next             2313 tools/perf/util/probe-event.c 	struct probe_trace_arg_ref *ref, *next;
next             2327 tools/perf/util/probe-event.c 			next = ref->next;
next             2329 tools/perf/util/probe-event.c 			ref = next;
next             2542 tools/perf/util/probe-event.c 				goto next;
next             2546 tools/perf/util/probe-event.c 				goto next;
next             2551 tools/perf/util/probe-event.c next:
next               38 tools/perf/util/probe-event.h 	struct probe_trace_arg_ref	*next;	/* Next reference */
next               74 tools/perf/util/probe-event.h 	struct perf_probe_arg_field	*next;	/* Next field */
next              341 tools/perf/util/probe-finder.c 				ref_ptr = &(*ref_ptr)->next;
next              435 tools/perf/util/probe-finder.c 				(*ref_ptr)->next = ref;
next              441 tools/perf/util/probe-finder.c 		goto next;
next              466 tools/perf/util/probe-finder.c 			(*ref_ptr)->next = ref;
next              519 tools/perf/util/probe-finder.c next:
next              521 tools/perf/util/probe-finder.c 	if (field->next)
next              523 tools/perf/util/probe-finder.c 				field->next, &ref, die_mem, user_access);
next              112 tools/perf/util/rblist.c 	struct rb_node *pos, *next = rb_first_cached(&rblist->entries);
next              114 tools/perf/util/rblist.c 	while (next) {
next              115 tools/perf/util/rblist.c 		pos = next;
next              116 tools/perf/util/rblist.c 		next = rb_next(pos);
next              109 tools/perf/util/scripting-engines/trace-event-perl.c 	if (field->next)
next              110 tools/perf/util/scripting-engines/trace-event-perl.c 		define_symbolic_values(field->next, ev_name, field_name);
next              167 tools/perf/util/scripting-engines/trace-event-perl.c 	if (field->next)
next              168 tools/perf/util/scripting-engines/trace-event-perl.c 		define_flag_values(field->next, ev_name, field_name);
next              256 tools/perf/util/scripting-engines/trace-event-perl.c 	if (args->next)
next              257 tools/perf/util/scripting-engines/trace-event-perl.c 		define_event_symbols(event, ev_name, args->next);
next              392 tools/perf/util/scripting-engines/trace-event-perl.c 	for (field = event->format.fields; field; field = field->next) {
next              627 tools/perf/util/scripting-engines/trace-event-perl.c 		for (f = event->format.fields; f; f = f->next) {
next              646 tools/perf/util/scripting-engines/trace-event-perl.c 		for (f = event->format.fields; f; f = f->next) {
next              670 tools/perf/util/scripting-engines/trace-event-perl.c 		for (f = event->format.fields; f; f = f->next) {
next              238 tools/perf/util/scripting-engines/trace-event-python.c 	if (field->next)
next              239 tools/perf/util/scripting-engines/trace-event-python.c 		define_values(field_type, field->next, ev_name, field_name);
next              335 tools/perf/util/scripting-engines/trace-event-python.c 	if (args->next)
next              336 tools/perf/util/scripting-engines/trace-event-python.c 		define_event_symbols(event, ev_name, args->next);
next              869 tools/perf/util/scripting-engines/trace-event-python.c 	for (field = event->format.fields; field; field = field->next) {
next             1757 tools/perf/util/scripting-engines/trace-event-python.c 		for (f = event->format.fields; f; f = f->next) {
next             1782 tools/perf/util/scripting-engines/trace-event-python.c 		for (f = event->format.fields; f; f = f->next) {
next             1807 tools/perf/util/scripting-engines/trace-event-python.c 		for (f = event->format.fields; f; f = f->next) {
next               87 tools/perf/util/session.c 		session->decomp_last->next = decomp;
next              275 tools/perf/util/session.c 	struct decomp *next, *decomp;
next              277 tools/perf/util/session.c 	next = session->decomp;
next              279 tools/perf/util/session.c 		decomp = next;
next              282 tools/perf/util/session.c 		next = decomp->next;
next               47 tools/perf/util/session.h 	struct decomp *next;
next             2385 tools/perf/util/sort.c 		field = field->next;
next              171 tools/perf/util/sort.h 		return list_entry(he->pairs.node.next, struct hist_entry, pairs.node);
next              115 tools/perf/util/srccode.c 	       srcfile_list.next != &srcfile_list) {
next              666 tools/perf/util/srcline.c 	struct rb_node *next = rb_first_cached(tree);
next              668 tools/perf/util/srcline.c 	while (next) {
next              669 tools/perf/util/srcline.c 		pos = rb_entry(next, struct srcline_node, rb_node);
next              670 tools/perf/util/srcline.c 		next = rb_next(&pos->rb_node);
next              750 tools/perf/util/srcline.c 	struct rb_node *next = rb_first_cached(tree);
next              752 tools/perf/util/srcline.c 	while (next) {
next              753 tools/perf/util/srcline.c 		pos = rb_entry(next, struct inline_node, rb_node);
next              754 tools/perf/util/srcline.c 		next = rb_next(&pos->rb_node);
next              170 tools/perf/util/stat-shadow.c 	struct rb_node *pos, *next;
next              173 tools/perf/util/stat-shadow.c 	next = rb_first_cached(&rblist->entries);
next              174 tools/perf/util/stat-shadow.c 	while (next) {
next              175 tools/perf/util/stat-shadow.c 		pos = next;
next              176 tools/perf/util/stat-shadow.c 		next = rb_next(pos);
next             2172 tools/perf/util/symbol-elf.c 	size_t shstrndx, next;
next             2201 tools/perf/util/symbol-elf.c 	for (offset = 0; (next = gelf_getnote(data, offset, &nhdr, &name_off,
next             2202 tools/perf/util/symbol-elf.c 					      &desc_off)) > 0; offset = next) {
next              188 tools/perf/util/symbol.c 	struct symbol *curr, *next;
next              199 tools/perf/util/symbol.c 		next = rb_entry(nd, struct symbol, rb_node);
next              204 tools/perf/util/symbol.c 		if (curr->start != next->start)
next              207 tools/perf/util/symbol.c 		if (choose_best_symbol(curr, next) == SYMBOL_A) {
next              208 tools/perf/util/symbol.c 			rb_erase_cached(&next->rb_node, symbols);
next              209 tools/perf/util/symbol.c 			symbol__delete(next);
next              245 tools/perf/util/symbol.c 	struct map *next, *curr;
next              253 tools/perf/util/symbol.c 	for (next = map__next(curr); next; next = map__next(curr)) {
next              255 tools/perf/util/symbol.c 			curr->end = next->start;
next              256 tools/perf/util/symbol.c 		curr = next;
next              307 tools/perf/util/symbol.c 	struct rb_node *next = rb_first_cached(symbols);
next              309 tools/perf/util/symbol.c 	while (next) {
next              310 tools/perf/util/symbol.c 		pos = rb_entry(next, struct symbol, rb_node);
next              311 tools/perf/util/symbol.c 		next = rb_next(&pos->rb_node);
next              712 tools/perf/util/symbol.c 	struct rb_node *next = rb_first_cached(root);
next              719 tools/perf/util/symbol.c 	while (next) {
next              722 tools/perf/util/symbol.c 		pos = rb_entry(next, struct symbol, rb_node);
next              723 tools/perf/util/symbol.c 		next = rb_next(&pos->rb_node);
next              766 tools/perf/util/symbol.c 	struct rb_node *next = rb_first_cached(root);
next              777 tools/perf/util/symbol.c 	while (next) {
next              780 tools/perf/util/symbol.c 		pos = rb_entry(next, struct symbol, rb_node);
next              781 tools/perf/util/symbol.c 		next = rb_next(&pos->rb_node);
next              946 tools/perf/util/symbol.c 	struct rb_node *next = rb_first(modules);
next              948 tools/perf/util/symbol.c 	while (next) {
next              949 tools/perf/util/symbol.c 		mi = rb_entry(next, struct module_info, rb_node);
next              950 tools/perf/util/symbol.c 		next = rb_next(&mi->rb_node);
next             1074 tools/perf/util/symbol.c 		struct map *next = map_groups__next(old_map);
next             1078 tools/perf/util/symbol.c 			old_map = next;
next             1089 tools/perf/util/symbol.c 		old_map = next;
next             1249 tools/perf/util/symbol.c 		old_map = list_entry(merged.next, struct map, node);
next             1316 tools/perf/util/symbol.c 		struct map *next = map_groups__next(old_map);
next             1325 tools/perf/util/symbol.c 		old_map = next;
next             1340 tools/perf/util/symbol.c 		replacement_map = list_entry(md.maps.next, struct map, node);
next             1344 tools/perf/util/symbol.c 		new_map = list_entry(md.maps.next, struct map, node);
next             1403 tools/perf/util/symbol.c 		map = list_entry(md.maps.next, struct map, node);
next              149 tools/perf/util/trace-event-info.c 		tps = tps->next;
next              247 tools/perf/util/trace-event-info.c 		tps = tps->next;
next              396 tools/perf/util/trace-event-info.c 		tps = tps->next;
next              416 tools/perf/util/trace-event-info.c 			ppath->next = tracepoint_name_to_path(pos->name);
next              417 tools/perf/util/trace-event-info.c 			if (ppath->next)
next              418 tools/perf/util/trace-event-info.c 				goto next;
next              427 tools/perf/util/trace-event-info.c 		ppath->next = tracepoint_id_to_path(pos->core.attr.config);
next              428 tools/perf/util/trace-event-info.c 		if (!ppath->next) {
next              434 tools/perf/util/trace-event-info.c next:
next              435 tools/perf/util/trace-event-info.c 		ppath = ppath->next;
next              438 tools/perf/util/trace-event-info.c 	return nr_tracepoints > 0 ? path.next : NULL;
next              129 tools/perf/util/trace-event-parse.c 	char *next = NULL;
next              133 tools/perf/util/trace-event-parse.c 	line = strtok_r(file, "\n", &next);
next              143 tools/perf/util/trace-event-parse.c 		line = strtok_r(NULL, "\n", &next);
next              154 tools/perf/util/trace-event-parse.c 	char *next = NULL;
next              157 tools/perf/util/trace-event-parse.c 	line = strtok_r(file, "\n", &next);
next              161 tools/perf/util/trace-event-parse.c 		line = strtok_r(NULL, "\n", &next);
next               20 tools/power/acpi/os_specific/service_layers/oslinuxtbl.c 	struct osl_table_info *next;
next              280 tools/power/acpi/os_specific/service_layers/oslinuxtbl.c 	struct osl_table_info *next;
next              294 tools/power/acpi/os_specific/service_layers/oslinuxtbl.c 		next = gbl_table_list_head;
next              296 tools/power/acpi/os_specific/service_layers/oslinuxtbl.c 			if (ACPI_COMPARE_NAMESEG(next->signature, signature)) {
next              297 tools/power/acpi/os_specific/service_layers/oslinuxtbl.c 				if (next->instance == instance) {
next              300 tools/power/acpi/os_specific/service_layers/oslinuxtbl.c 				if (next->instance >= next_instance) {
next              301 tools/power/acpi/os_specific/service_layers/oslinuxtbl.c 					next_instance = next->instance + 1;
next              305 tools/power/acpi/os_specific/service_layers/oslinuxtbl.c 			if (!next->next) {
next              308 tools/power/acpi/os_specific/service_layers/oslinuxtbl.c 			next = next->next;
next              310 tools/power/acpi/os_specific/service_layers/oslinuxtbl.c 		next->next = new_info;
next              373 tools/power/acpi/os_specific/service_layers/oslinuxtbl.c 		info = info->next;
next              281 tools/power/cpupower/lib/cpufreq.c 				current->next = malloc(sizeof(*current));
next              282 tools/power/cpupower/lib/cpufreq.c 				if (!current->next)
next              284 tools/power/cpupower/lib/cpufreq.c 				current = current->next;
next              292 tools/power/cpupower/lib/cpufreq.c 			current->next = NULL;
next              308 tools/power/cpupower/lib/cpufreq.c 		current = first->next;
next              319 tools/power/cpupower/lib/cpufreq.c 	struct cpufreq_available_governors *tmp, *next;
next              326 tools/power/cpupower/lib/cpufreq.c 		next = tmp->next;
next              330 tools/power/cpupower/lib/cpufreq.c 		tmp = next;
next              358 tools/power/cpupower/lib/cpufreq.c 				current->next = malloc(sizeof(*current));
next              359 tools/power/cpupower/lib/cpufreq.c 				if (!current->next)
next              361 tools/power/cpupower/lib/cpufreq.c 				current = current->next;
next              369 tools/power/cpupower/lib/cpufreq.c 			current->next = NULL;
next              384 tools/power/cpupower/lib/cpufreq.c 		current = first->next;
next              414 tools/power/cpupower/lib/cpufreq.c 				current->next = malloc(sizeof(*current));
next              415 tools/power/cpupower/lib/cpufreq.c 				if (!current->next)
next              417 tools/power/cpupower/lib/cpufreq.c 				current = current->next;
next              425 tools/power/cpupower/lib/cpufreq.c 			current->next = NULL;
next              440 tools/power/cpupower/lib/cpufreq.c 		current = first->next;
next              449 tools/power/cpupower/lib/cpufreq.c 	struct cpufreq_available_frequencies *tmp, *next;
next              456 tools/power/cpupower/lib/cpufreq.c 		next = tmp->next;
next              458 tools/power/cpupower/lib/cpufreq.c 		tmp = next;
next              489 tools/power/cpupower/lib/cpufreq.c 				current->next = malloc(sizeof(*current));
next              490 tools/power/cpupower/lib/cpufreq.c 				if (!current->next)
next              492 tools/power/cpupower/lib/cpufreq.c 				current = current->next;
next              500 tools/power/cpupower/lib/cpufreq.c 			current->next = NULL;
next              516 tools/power/cpupower/lib/cpufreq.c 		current = first->next;
next              530 tools/power/cpupower/lib/cpufreq.c 	struct cpufreq_affected_cpus *tmp, *next;
next              537 tools/power/cpupower/lib/cpufreq.c 		next = tmp->next;
next              539 tools/power/cpupower/lib/cpufreq.c 		tmp = next;
next              722 tools/power/cpupower/lib/cpufreq.c 				current->next = malloc(sizeof(*current));
next              723 tools/power/cpupower/lib/cpufreq.c 				if (!current->next)
next              725 tools/power/cpupower/lib/cpufreq.c 				current = current->next;
next              733 tools/power/cpupower/lib/cpufreq.c 			current->next = NULL;
next              751 tools/power/cpupower/lib/cpufreq.c 		current = first->next;
next              760 tools/power/cpupower/lib/cpufreq.c 	struct cpufreq_stats *tmp, *next;
next              767 tools/power/cpupower/lib/cpufreq.c 		next = tmp->next;
next              769 tools/power/cpupower/lib/cpufreq.c 		tmp = next;
next               19 tools/power/cpupower/lib/cpufreq.h 	struct cpufreq_available_governors *next;
next               25 tools/power/cpupower/lib/cpufreq.h 	struct cpufreq_available_frequencies *next;
next               32 tools/power/cpupower/lib/cpufreq.h 	struct cpufreq_affected_cpus *next;
next               39 tools/power/cpupower/lib/cpufreq.h 	struct cpufreq_stats *next;
next              257 tools/power/cpupower/utils/cpufreq-info.c 		while (freqs->next) {
next              260 tools/power/cpupower/utils/cpufreq-info.c 			freqs = freqs->next;
next              379 tools/power/cpupower/utils/cpufreq-info.c 	while (governors->next) {
next              381 tools/power/cpupower/utils/cpufreq-info.c 		governors = governors->next;
next              401 tools/power/cpupower/utils/cpufreq-info.c 	while (cpus->next) {
next              403 tools/power/cpupower/utils/cpufreq-info.c 		cpus = cpus->next;
next              422 tools/power/cpupower/utils/cpufreq-info.c 	while (cpus->next) {
next              424 tools/power/cpupower/utils/cpufreq-info.c 		cpus = cpus->next;
next              446 tools/power/cpupower/utils/cpufreq-info.c 		stats = stats->next;
next              489 tools/power/cpupower/utils/cpufreq-info.c 		while (freqs->next) {
next              492 tools/power/cpupower/utils/cpufreq-info.c 			freqs = freqs->next;
next              304 tools/power/cpupower/utils/cpufreq-set.c 			while (cpus->next) {
next              306 tools/power/cpupower/utils/cpufreq-set.c 				cpus = cpus->next;
next               47 tools/power/cpupower/utils/helpers/pci.c 	for (device = (*pacc)->devices; device; device = device->next) {
next             1343 tools/power/x86/intel-speed-select/isst-config.c 	char *next;
next             1345 tools/power/x86/intel-speed-select/isst-config.c 	next = optarg;
next             1347 tools/power/x86/intel-speed-select/isst-config.c 	while (next && *next) {
next             1348 tools/power/x86/intel-speed-select/isst-config.c 		if (*next == '-') /* no negative cpu numbers */
next             1351 tools/power/x86/intel-speed-select/isst-config.c 		start = strtoul(next, &next, 10);
next             1356 tools/power/x86/intel-speed-select/isst-config.c 		if (*next == '\0')
next             1359 tools/power/x86/intel-speed-select/isst-config.c 		if (*next == ',') {
next             1360 tools/power/x86/intel-speed-select/isst-config.c 			next += 1;
next             1364 tools/power/x86/intel-speed-select/isst-config.c 		if (*next == '-') {
next             1365 tools/power/x86/intel-speed-select/isst-config.c 			next += 1; /* start range */
next             1366 tools/power/x86/intel-speed-select/isst-config.c 		} else if (*next == '.') {
next             1367 tools/power/x86/intel-speed-select/isst-config.c 			next += 1;
next             1368 tools/power/x86/intel-speed-select/isst-config.c 			if (*next == '.')
next             1369 tools/power/x86/intel-speed-select/isst-config.c 				next += 1; /* start range */
next             1374 tools/power/x86/intel-speed-select/isst-config.c 		end = strtoul(next, &next, 10);
next             1383 tools/power/x86/intel-speed-select/isst-config.c 		if (*next == ',')
next             1384 tools/power/x86/intel-speed-select/isst-config.c 			next += 1;
next             1385 tools/power/x86/intel-speed-select/isst-config.c 		else if (*next != '\0')
next              255 tools/power/x86/turbostat/turbostat.c 	struct msr_counter *next;
next              666 tools/power/x86/turbostat/turbostat.c 	for (mp = sys.tp; mp; mp = mp->next) {
next              704 tools/power/x86/turbostat/turbostat.c 	for (mp = sys.cp; mp; mp = mp->next) {
next              782 tools/power/x86/turbostat/turbostat.c 	for (mp = sys.pp; mp; mp = mp->next) {
next              820 tools/power/x86/turbostat/turbostat.c 		for (i = 0, mp = sys.tp; mp; i++, mp = mp->next) {
next              834 tools/power/x86/turbostat/turbostat.c 		for (i = 0, mp = sys.cp; mp; i++, mp = mp->next) {
next              871 tools/power/x86/turbostat/turbostat.c 		for (i = 0, mp = sys.pp; mp; i++, mp = mp->next) {
next             1007 tools/power/x86/turbostat/turbostat.c 	for (i = 0, mp = sys.tp; mp; i++, mp = mp->next) {
next             1049 tools/power/x86/turbostat/turbostat.c 	for (i = 0, mp = sys.cp; mp; i++, mp = mp->next) {
next             1152 tools/power/x86/turbostat/turbostat.c 	for (i = 0, mp = sys.pp; mp; i++, mp = mp->next) {
next             1264 tools/power/x86/turbostat/turbostat.c 	for (i = 0, mp = sys.pp; mp; i++, mp = mp->next) {
next             1288 tools/power/x86/turbostat/turbostat.c 	for (i = 0, mp = sys.cp; mp; i++, mp = mp->next) {
next             1383 tools/power/x86/turbostat/turbostat.c 	for (i = 0, mp = sys.tp; mp; i++, mp = mp->next) {
next             1472 tools/power/x86/turbostat/turbostat.c 	for (i = 0, mp = sys.tp; mp; i++, mp = mp->next)
next             1475 tools/power/x86/turbostat/turbostat.c 	for (i = 0, mp = sys.cp; mp; i++, mp = mp->next)
next             1478 tools/power/x86/turbostat/turbostat.c 	for (i = 0, mp = sys.pp; mp; i++, mp = mp->next)
next             1508 tools/power/x86/turbostat/turbostat.c 	for (i = 0, mp = sys.tp; mp; i++, mp = mp->next) {
next             1527 tools/power/x86/turbostat/turbostat.c 	for (i = 0, mp = sys.cp; mp; i++, mp = mp->next) {
next             1573 tools/power/x86/turbostat/turbostat.c 	for (i = 0, mp = sys.pp; mp; i++, mp = mp->next) {
next             1631 tools/power/x86/turbostat/turbostat.c 	for (i = 0, mp = sys.tp; mp; i++, mp = mp->next) {
next             1641 tools/power/x86/turbostat/turbostat.c 	for (i = 0, mp = sys.cp; mp; i++, mp = mp->next) {
next             1650 tools/power/x86/turbostat/turbostat.c 	for (i = 0, mp = sys.pp; mp; i++, mp = mp->next) {
next             1867 tools/power/x86/turbostat/turbostat.c 	for (i = 0, mp = sys.tp; mp; i++, mp = mp->next) {
next             1909 tools/power/x86/turbostat/turbostat.c 	for (i = 0, mp = sys.cp; mp; i++, mp = mp->next) {
next             2015 tools/power/x86/turbostat/turbostat.c 	for (i = 0, mp = sys.pp; mp; i++, mp = mp->next) {
next             5356 tools/power/x86/turbostat/turbostat.c 		msrp->next = sys.tp;
next             5367 tools/power/x86/turbostat/turbostat.c 		msrp->next = sys.cp;
next             5378 tools/power/x86/turbostat/turbostat.c 		msrp->next = sys.pp;
next             5406 tools/power/x86/turbostat/turbostat.c 			goto next;
next             5409 tools/power/x86/turbostat/turbostat.c 			goto next;
next             5413 tools/power/x86/turbostat/turbostat.c 			goto next;
next             5418 tools/power/x86/turbostat/turbostat.c 				goto next;
next             5423 tools/power/x86/turbostat/turbostat.c 			goto next;
next             5427 tools/power/x86/turbostat/turbostat.c 			goto next;
next             5431 tools/power/x86/turbostat/turbostat.c 			goto next;
next             5435 tools/power/x86/turbostat/turbostat.c 			goto next;
next             5439 tools/power/x86/turbostat/turbostat.c 			goto next;
next             5443 tools/power/x86/turbostat/turbostat.c 			goto next;
next             5447 tools/power/x86/turbostat/turbostat.c 			goto next;
next             5451 tools/power/x86/turbostat/turbostat.c 			goto next;
next             5455 tools/power/x86/turbostat/turbostat.c 			goto next;
next             5464 tools/power/x86/turbostat/turbostat.c 			goto next;
next             5467 tools/power/x86/turbostat/turbostat.c next:
next             5581 tools/power/x86/turbostat/turbostat.c 	char *next;
next             5605 tools/power/x86/turbostat/turbostat.c 	next = optarg;
next             5607 tools/power/x86/turbostat/turbostat.c 	while (next && *next) {
next             5609 tools/power/x86/turbostat/turbostat.c 		if (*next == '-')	/* no negative cpu numbers */
next             5612 tools/power/x86/turbostat/turbostat.c 		start = strtoul(next, &next, 10);
next             5618 tools/power/x86/turbostat/turbostat.c 		if (*next == '\0')
next             5621 tools/power/x86/turbostat/turbostat.c 		if (*next == ',') {
next             5622 tools/power/x86/turbostat/turbostat.c 			next += 1;
next             5626 tools/power/x86/turbostat/turbostat.c 		if (*next == '-') {
next             5627 tools/power/x86/turbostat/turbostat.c 			next += 1;	/* start range */
next             5628 tools/power/x86/turbostat/turbostat.c 		} else if (*next == '.') {
next             5629 tools/power/x86/turbostat/turbostat.c 			next += 1;
next             5630 tools/power/x86/turbostat/turbostat.c 			if (*next == '.')
next             5631 tools/power/x86/turbostat/turbostat.c 				next += 1;	/* start range */
next             5636 tools/power/x86/turbostat/turbostat.c 		end = strtoul(next, &next, 10);
next             5646 tools/power/x86/turbostat/turbostat.c 		if (*next == ',')
next             5647 tools/power/x86/turbostat/turbostat.c 			next += 1;
next             5648 tools/power/x86/turbostat/turbostat.c 		else if (*next != '\0')
next               32 tools/testing/selftests/bpf/progs/btf_dump_test_case_ordering.c 	struct list_head *next;
next               37 tools/testing/selftests/bpf/progs/btf_dump_test_case_ordering.c 	struct hlist_node *next;
next               46 tools/testing/selftests/bpf/progs/btf_dump_test_case_ordering.c 	struct callback_head *next;
next               32 tools/testing/selftests/bpf/test_lpm_map.c 	struct tlpm_node *next;
next               62 tools/testing/selftests/bpf/test_lpm_map.c 	node->next = list;
next               76 tools/testing/selftests/bpf/test_lpm_map.c 		list = list->next;
next               94 tools/testing/selftests/bpf/test_lpm_map.c 	for ( ; list; list = list->next) {
next              121 tools/testing/selftests/bpf/test_lpm_map.c 		node = best->next;
next              126 tools/testing/selftests/bpf/test_lpm_map.c 	for (node = list; node; node = node->next) {
next              127 tools/testing/selftests/bpf/test_lpm_map.c 		if (node->next == best) {
next              128 tools/testing/selftests/bpf/test_lpm_map.c 			node->next = best->next;
next              187 tools/testing/selftests/bpf/test_lpm_map.c 	for (t1 = l1; t1; t1 = t1->next)
next              283 tools/testing/selftests/bpf/test_lpm_map.c 	for (i = 0, t = list; t; i++, t = t->next)
next              135 tools/testing/selftests/bpf/test_lru_map.c 	int next = *next_to_try;
next              138 tools/testing/selftests/bpf/test_lru_map.c 	while (next < nr_cpus) {
next              140 tools/testing/selftests/bpf/test_lru_map.c 		CPU_SET(next++, &cpuset);
next              147 tools/testing/selftests/bpf/test_lru_map.c 	*next_to_try = next;
next              366 tools/testing/selftests/bpf/test_progs.c 	char *next;
next              370 tools/testing/selftests/bpf/test_progs.c 		num = strtol(s, &next, 10);
next              379 tools/testing/selftests/bpf/test_progs.c 		if (!parsing_end && *next == '-') {
next              380 tools/testing/selftests/bpf/test_progs.c 			s = next + 1;
next              383 tools/testing/selftests/bpf/test_progs.c 		} else if (*next == ',') {
next              385 tools/testing/selftests/bpf/test_progs.c 			s = next + 1;
next              387 tools/testing/selftests/bpf/test_progs.c 		} else if (*next == '\0') {
next              389 tools/testing/selftests/bpf/test_progs.c 			s = next;
next              644 tools/testing/selftests/kselftest_harness.h 	struct __test_metadata *prev, *next;
next              671 tools/testing/selftests/kselftest_harness.h 		t->next = NULL;
next              676 tools/testing/selftests/kselftest_harness.h 		t->next = NULL;
next              678 tools/testing/selftests/kselftest_harness.h 		t->prev->next = t;
next              681 tools/testing/selftests/kselftest_harness.h 		t->next = __test_list;
next              682 tools/testing/selftests/kselftest_harness.h 		t->next->prev = t;
next              769 tools/testing/selftests/kselftest_harness.h 	for (t = __test_list; t; t = t->next) {
next              264 tools/testing/selftests/kvm/lib/kvm_util.c 		region = region->next) {
next              325 tools/testing/selftests/kvm/lib/kvm_util.c 		region = region->next) {
next              385 tools/testing/selftests/kvm/lib/kvm_util.c 	for (vcpup = vm->vcpu_head; vcpup; vcpup = vcpup->next) {
next              418 tools/testing/selftests/kvm/lib/kvm_util.c 	if (vcpu->next)
next              419 tools/testing/selftests/kvm/lib/kvm_util.c 		vcpu->next->prev = vcpu->prev;
next              421 tools/testing/selftests/kvm/lib/kvm_util.c 		vcpu->prev->next = vcpu->next;
next              423 tools/testing/selftests/kvm/lib/kvm_util.c 		vm->vcpu_head = vcpu->next;
next              464 tools/testing/selftests/kvm/lib/kvm_util.c 		vmp->userspace_mem_region_head = region->next;
next              614 tools/testing/selftests/kvm/lib/kvm_util.c 		region = region->next) {
next              689 tools/testing/selftests/kvm/lib/kvm_util.c 	region->next = vm->userspace_mem_region_head;
next              714 tools/testing/selftests/kvm/lib/kvm_util.c 		region = region->next) {
next              836 tools/testing/selftests/kvm/lib/kvm_util.c 	vcpu->next = vm->vcpu_head;
next             1032 tools/testing/selftests/kvm/lib/kvm_util.c 	     region = region->next) {
next             1065 tools/testing/selftests/kvm/lib/kvm_util.c 	     region = region->next) {
next             1458 tools/testing/selftests/kvm/lib/kvm_util.c 		region = region->next) {
next             1477 tools/testing/selftests/kvm/lib/kvm_util.c 	for (vcpu = vm->vcpu_head; vcpu; vcpu = vcpu->next)
next               27 tools/testing/selftests/kvm/lib/kvm_util_internal.h 	struct userspace_mem_region *next, *prev;
next               38 tools/testing/selftests/kvm/lib/kvm_util_internal.h 	struct vcpu *next, *prev;
next              605 tools/testing/selftests/kvm/lib/sparsebit.c 		struct node *prev, *next, *tmp;
next              746 tools/testing/selftests/kvm/lib/sparsebit.c 		next = node_next(s, nodep);
next              747 tools/testing/selftests/kvm/lib/sparsebit.c 		if (next) {
next              749 tools/testing/selftests/kvm/lib/sparsebit.c 			if (next->mask == 0 && next->num_after == 0) {
next              750 tools/testing/selftests/kvm/lib/sparsebit.c 				node_rm(s, next);
next              759 tools/testing/selftests/kvm/lib/sparsebit.c 			if (next->idx == nodep->idx + MASK_BITS + nodep->num_after &&
next              760 tools/testing/selftests/kvm/lib/sparsebit.c 			    next->mask == ~(mask_t) 0) {
next              762 tools/testing/selftests/kvm/lib/sparsebit.c 				next->mask = 0;
next              763 tools/testing/selftests/kvm/lib/sparsebit.c 				nodep->num_after += next->num_after;
next              764 tools/testing/selftests/kvm/lib/sparsebit.c 				next->num_after = 0;
next              766 tools/testing/selftests/kvm/lib/sparsebit.c 				node_rm(s, next);
next              767 tools/testing/selftests/kvm/lib/sparsebit.c 				next = NULL;
next             1374 tools/testing/selftests/kvm/lib/sparsebit.c 	struct node *nodep, *next;
next             1421 tools/testing/selftests/kvm/lib/sparsebit.c 		for (next = node_next(s, nodep);
next             1422 tools/testing/selftests/kvm/lib/sparsebit.c 			next && (next->idx < middle_end);
next             1423 tools/testing/selftests/kvm/lib/sparsebit.c 			next = node_next(s, nodep)) {
next             1424 tools/testing/selftests/kvm/lib/sparsebit.c 			assert(next->idx + MASK_BITS + next->num_after - 1 <= middle_end);
next             1425 tools/testing/selftests/kvm/lib/sparsebit.c 			node_rm(s, next);
next             1426 tools/testing/selftests/kvm/lib/sparsebit.c 			next = NULL;
next             1456 tools/testing/selftests/kvm/lib/sparsebit.c 	struct node *nodep, *next;
next             1484 tools/testing/selftests/kvm/lib/sparsebit.c 		for (next = node_next(s, nodep);
next             1485 tools/testing/selftests/kvm/lib/sparsebit.c 			next && (next->idx < middle_end);
next             1486 tools/testing/selftests/kvm/lib/sparsebit.c 			next = node_next(s, nodep)) {
next             1487 tools/testing/selftests/kvm/lib/sparsebit.c 			assert(next->idx + MASK_BITS + next->num_after - 1 <= middle_end);
next             1488 tools/testing/selftests/kvm/lib/sparsebit.c 			node_rm(s, next);
next             1489 tools/testing/selftests/kvm/lib/sparsebit.c 			next = NULL;
next             1918 tools/testing/selftests/kvm/lib/sparsebit.c 	sparsebit_idx_t next;
next             1982 tools/testing/selftests/kvm/lib/sparsebit.c 		next = sparsebit_next_set(s, first);
next             1983 tools/testing/selftests/kvm/lib/sparsebit.c 		assert(next == 0 || next > first);
next             1984 tools/testing/selftests/kvm/lib/sparsebit.c 		assert(next == 0 || get_value(next));
next             1987 tools/testing/selftests/kvm/lib/sparsebit.c 		next = sparsebit_next_clear(s, first);
next             1988 tools/testing/selftests/kvm/lib/sparsebit.c 		assert(next == 0 || next > first);
next             1989 tools/testing/selftests/kvm/lib/sparsebit.c 		assert(next == 0 || !get_value(next));
next             1992 tools/testing/selftests/kvm/lib/sparsebit.c 		next = sparsebit_next_clear(s, first);
next             1994 tools/testing/selftests/kvm/lib/sparsebit.c 			assert(next == 0 || next > last);
next             1996 tools/testing/selftests/kvm/lib/sparsebit.c 				next = sparsebit_next_set(s, first - 1);
next             1998 tools/testing/selftests/kvm/lib/sparsebit.c 				next = sparsebit_first_set(s);
next             2001 tools/testing/selftests/kvm/lib/sparsebit.c 			assert(next == first);
next             2003 tools/testing/selftests/kvm/lib/sparsebit.c 			assert(sparsebit_is_clear(s, first) || next <= last);
next             2007 tools/testing/selftests/kvm/lib/sparsebit.c 		next = sparsebit_next_set(s, first);
next             2009 tools/testing/selftests/kvm/lib/sparsebit.c 			assert(next == 0 || next > last);
next             2011 tools/testing/selftests/kvm/lib/sparsebit.c 				next = sparsebit_next_clear(s, first - 1);
next             2013 tools/testing/selftests/kvm/lib/sparsebit.c 				next = sparsebit_first_clear(s);
next             2016 tools/testing/selftests/kvm/lib/sparsebit.c 			assert(next == first);
next             2018 tools/testing/selftests/kvm/lib/sparsebit.c 			assert(sparsebit_is_set(s, first) || next <= last);
next              419 tools/testing/selftests/net/psock_tpacket.c 		void *next = get_next_frame(ring, frame_num);
next              421 tools/testing/selftests/net/psock_tpacket.c 		while (__tx_kernel_ready(next, ring->version) &&
next              423 tools/testing/selftests/net/psock_tpacket.c 			ppd.raw = next;
next              446 tools/testing/selftests/net/psock_tpacket.c 				struct tpacket3_hdr *tx = next;
next              463 tools/testing/selftests/net/psock_tpacket.c 			__tx_user_ready(next, ring->version);
next              315 tools/testing/selftests/networking/timestamping/timestamping.c 	struct timeval next;
next              455 tools/testing/selftests/networking/timestamping/timestamping.c 	gettimeofday(&next, 0);
next              456 tools/testing/selftests/networking/timestamping/timestamping.c 	next.tv_sec = (next.tv_sec + 1) / 5 * 5;
next              457 tools/testing/selftests/networking/timestamping/timestamping.c 	next.tv_usec = 0;
next              466 tools/testing/selftests/networking/timestamping/timestamping.c 		delta_us = (long)(next.tv_sec - now.tv_sec) * 1000000 +
next              467 tools/testing/selftests/networking/timestamping/timestamping.c 			(long)(next.tv_usec - now.tv_usec);
next              503 tools/testing/selftests/networking/timestamping/timestamping.c 			next.tv_sec += 5;
next              122 tools/testing/selftests/proc/fd-001-lookup.c next:
next              137 tools/testing/selftests/proc/fd-001-lookup.c 				goto next;
next              109 tools/testing/selftests/rcutorture/formal/srcu-cbmc/include/linux/types.h 	struct list_head *next, *prev;
next              117 tools/testing/selftests/rcutorture/formal/srcu-cbmc/include/linux/types.h 	struct hlist_node *next, **pprev;
next              140 tools/testing/selftests/rcutorture/formal/srcu-cbmc/include/linux/types.h 	struct callback_head *next;
next               36 tools/testing/selftests/rseq/basic_percpu_ops_test.c 	struct percpu_list_node *next;
next              150 tools/testing/selftests/rseq/basic_percpu_ops_test.c 		node->next = (struct percpu_list_node *)expect;
next              177 tools/testing/selftests/rseq/basic_percpu_ops_test.c 		offset = offsetof(struct percpu_list_node, next);
next              203 tools/testing/selftests/rseq/basic_percpu_ops_test.c 	list->c[cpu].head = node->next;
next              260 tools/testing/selftests/rseq/basic_percpu_ops_test.c 			node->next = list.c[i].head;
next              281 tools/testing/selftests/rseq/param_test.c 	struct percpu_list_node *next;
next              528 tools/testing/selftests/rseq/param_test.c 		node->next = (struct percpu_list_node *)expect;
next              558 tools/testing/selftests/rseq/param_test.c 		offset = offsetof(struct percpu_list_node, next);
next              586 tools/testing/selftests/rseq/param_test.c 	list->c[cpu].head = node->next;
next              642 tools/testing/selftests/rseq/param_test.c 			node->next = list.c[i].head;
next              152 tools/testing/selftests/timers/nanosleep.c 				goto next;
next              161 tools/testing/selftests/timers/nanosleep.c next:
next               25 tools/testing/selftests/vm/compaction_test.c 	struct map_list *next;
next              207 tools/testing/selftests/vm/compaction_test.c 		entry->next = list;
next              219 tools/testing/selftests/vm/compaction_test.c 	for (entry = list; entry != NULL; entry = entry->next) {
next              221 tools/testing/selftests/vm/compaction_test.c 		if (!entry->next)
next              223 tools/testing/selftests/vm/compaction_test.c 		entry = entry->next;
next               46 tools/testing/selftests/vm/mlock2.h 			goto next;
next               51 tools/testing/selftests/vm/mlock2.h next:
next               61 tools/testing/vsock/control.c 				goto next;
next               75 tools/testing/vsock/control.c 			goto next;
next               77 tools/testing/vsock/control.c 			goto next;
next               93 tools/testing/vsock/control.c next:
next              332 tools/testing/vsock/vsock_diag_test.c 	struct vsock_stat *next;
next              334 tools/testing/vsock/vsock_diag_test.c 	list_for_each_entry_safe(st, next, sockets, list)
next              116 tools/usb/testusb.c 	struct testdev		*next;
next              274 tools/usb/testusb.c 	entry->next = testdevs;
next              483 tools/usb/testusb.c 	if (testdevs && testdevs->next == 0 && !device)
next              485 tools/usb/testusb.c 	for (entry = testdevs; entry; entry = entry->next) {
next              517 tools/usb/testusb.c 	for (entry = testdevs; entry; entry = entry->next) {
next               20 tools/usb/usbip/libsrc/list.h 	struct list_head *next, *prev;
next               30 tools/usb/usbip/libsrc/list.h 	list->next = list;
next               42 tools/usb/usbip/libsrc/list.h 			      struct list_head *next)
next               44 tools/usb/usbip/libsrc/list.h 	next->prev = new;
next               45 tools/usb/usbip/libsrc/list.h 	new->next = next;
next               47 tools/usb/usbip/libsrc/list.h 	prev->next = new;
next               60 tools/usb/usbip/libsrc/list.h 	__list_add(new, head, head->next);
next               70 tools/usb/usbip/libsrc/list.h static inline void __list_del(struct list_head * prev, struct list_head * next)
next               72 tools/usb/usbip/libsrc/list.h 	next->prev = prev;
next               73 tools/usb/usbip/libsrc/list.h 	prev->next = next;
next               88 tools/usb/usbip/libsrc/list.h 	__list_del(entry->prev, entry->next);
next               93 tools/usb/usbip/libsrc/list.h 	__list_del(entry->prev, entry->next);
next               94 tools/usb/usbip/libsrc/list.h 	entry->next = LIST_POISON1;
next              112 tools/usb/usbip/libsrc/list.h 	for (pos = (head)->next; pos != (head); pos = pos->next)
next              121 tools/usb/usbip/libsrc/list.h 	for (pos = (head)->next, n = pos->next; pos != (head); \
next              122 tools/usb/usbip/libsrc/list.h 		pos = n, n = pos->next)
next               26 tools/usb/usbip/libsrc/names.c 	struct vendor *next;
next               32 tools/usb/usbip/libsrc/names.c 	struct product *next;
next               38 tools/usb/usbip/libsrc/names.c 	struct class *next;
next               44 tools/usb/usbip/libsrc/names.c 	struct subclass *next;
next               50 tools/usb/usbip/libsrc/names.c 	struct protocol *next;
next               56 tools/usb/usbip/libsrc/names.c 	struct genericstrtable *next;
next               88 tools/usb/usbip/libsrc/names.c 	for (; v; v = v->next)
next               99 tools/usb/usbip/libsrc/names.c 	for (; p; p = p->next)
next              110 tools/usb/usbip/libsrc/names.c 	for (; c; c = c->next)
next              121 tools/usb/usbip/libsrc/names.c 	for (; s; s = s->next)
next              134 tools/usb/usbip/libsrc/names.c 	for (; p; p = p->next)
next              143 tools/usb/usbip/libsrc/names.c 	struct pool *next;
next              163 tools/usb/usbip/libsrc/names.c 	p->next = pool_head;
next              183 tools/usb/usbip/libsrc/names.c 		pool = pool->next;
next              194 tools/usb/usbip/libsrc/names.c 	for (; v; v = v->next)
next              202 tools/usb/usbip/libsrc/names.c 	v->next = vendors[h];
next              214 tools/usb/usbip/libsrc/names.c 	for (; p; p = p->next)
next              223 tools/usb/usbip/libsrc/names.c 	p->next = products[h];
next              234 tools/usb/usbip/libsrc/names.c 	for (; c; c = c->next)
next              242 tools/usb/usbip/libsrc/names.c 	c->next = classes[h];
next              253 tools/usb/usbip/libsrc/names.c 	for (; s; s = s->next)
next              262 tools/usb/usbip/libsrc/names.c 	s->next = subclasses[h];
next              275 tools/usb/usbip/libsrc/names.c 	for (; p; p = p->next)
next              286 tools/usb/usbip/libsrc/names.c 	p->next = protocols[h];
next               20 tools/virtio/ringtest/ring.c 			      unsigned short next,
next               23 tools/virtio/ringtest/ring.c 	return (unsigned short)(next - event - 1) < (unsigned short)(next - prev);
next               86 tools/virtio/ringtest/virtio_ring_0_9.c 		ring.desc[i].next = i + 1;
next              127 tools/virtio/ringtest/virtio_ring_0_9.c 	guest.free_head = desc[head].next;
next              189 tools/virtio/ringtest/virtio_ring_0_9.c 	ring.desc[index].next = guest.free_head;
next              698 tools/virtio/vringh_test.c 		d[0].next = 1;
next              712 tools/virtio/vringh_test.c 		d[3].next = 1;
next              716 tools/virtio/vringh_test.c 		d[4].next = 2;
next              271 virt/kvm/arm/mmu.c 	phys_addr_t next, start_addr = addr;
next              276 virt/kvm/arm/mmu.c 		next = stage2_pmd_addr_end(kvm, addr, end);
next              288 virt/kvm/arm/mmu.c 				unmap_stage2_ptes(kvm, pmd, addr, next);
next              291 virt/kvm/arm/mmu.c 	} while (pmd++, addr = next, addr != end);
next              300 virt/kvm/arm/mmu.c 	phys_addr_t next, start_addr = addr;
next              305 virt/kvm/arm/mmu.c 		next = stage2_pud_addr_end(kvm, addr, end);
next              315 virt/kvm/arm/mmu.c 				unmap_stage2_pmds(kvm, pud, addr, next);
next              318 virt/kvm/arm/mmu.c 	} while (pud++, addr = next, addr != end);
next              339 virt/kvm/arm/mmu.c 	phys_addr_t next;
next              353 virt/kvm/arm/mmu.c 		next = stage2_pgd_addr_end(kvm, addr, end);
next              355 virt/kvm/arm/mmu.c 			unmap_stage2_puds(kvm, pgd, addr, next);
next              360 virt/kvm/arm/mmu.c 		if (next != end)
next              362 virt/kvm/arm/mmu.c 	} while (pgd++, addr = next, addr != end);
next              381 virt/kvm/arm/mmu.c 	phys_addr_t next;
next              385 virt/kvm/arm/mmu.c 		next = stage2_pmd_addr_end(kvm, addr, end);
next              390 virt/kvm/arm/mmu.c 				stage2_flush_ptes(kvm, pmd, addr, next);
next              392 virt/kvm/arm/mmu.c 	} while (pmd++, addr = next, addr != end);
next              399 virt/kvm/arm/mmu.c 	phys_addr_t next;
next              403 virt/kvm/arm/mmu.c 		next = stage2_pud_addr_end(kvm, addr, end);
next              408 virt/kvm/arm/mmu.c 				stage2_flush_pmds(kvm, pud, addr, next);
next              410 virt/kvm/arm/mmu.c 	} while (pud++, addr = next, addr != end);
next              418 virt/kvm/arm/mmu.c 	phys_addr_t next;
next              423 virt/kvm/arm/mmu.c 		next = stage2_pgd_addr_end(kvm, addr, end);
next              425 virt/kvm/arm/mmu.c 			stage2_flush_puds(kvm, pgd, addr, next);
next              426 virt/kvm/arm/mmu.c 	} while (pgd++, addr = next, addr != end);
next              497 virt/kvm/arm/mmu.c 	phys_addr_t next;
next              502 virt/kvm/arm/mmu.c 		next = pmd_addr_end(addr, end);
next              505 virt/kvm/arm/mmu.c 			unmap_hyp_ptes(pmd, addr, next);
next              506 virt/kvm/arm/mmu.c 	} while (pmd++, addr = next, addr != end);
next              514 virt/kvm/arm/mmu.c 	phys_addr_t next;
next              519 virt/kvm/arm/mmu.c 		next = pud_addr_end(addr, end);
next              522 virt/kvm/arm/mmu.c 			unmap_hyp_pmds(pud, addr, next);
next              523 virt/kvm/arm/mmu.c 	} while (pud++, addr = next, addr != end);
next              539 virt/kvm/arm/mmu.c 	phys_addr_t next;
next              547 virt/kvm/arm/mmu.c 		next = pgd_addr_end(addr, end);
next              549 virt/kvm/arm/mmu.c 			unmap_hyp_puds(pgd, addr, next);
next              550 virt/kvm/arm/mmu.c 	} while (pgd++, addr = next, addr != end);
next              632 virt/kvm/arm/mmu.c 	unsigned long addr, next;
next              650 virt/kvm/arm/mmu.c 		next = pmd_addr_end(addr, end);
next              652 virt/kvm/arm/mmu.c 		create_hyp_pte_mappings(pmd, addr, next, pfn, prot);
next              653 virt/kvm/arm/mmu.c 		pfn += (next - addr) >> PAGE_SHIFT;
next              654 virt/kvm/arm/mmu.c 	} while (addr = next, addr != end);
next              665 virt/kvm/arm/mmu.c 	unsigned long addr, next;
next              682 virt/kvm/arm/mmu.c 		next = pud_addr_end(addr, end);
next              683 virt/kvm/arm/mmu.c 		ret = create_hyp_pmd_mappings(pud, addr, next, pfn, prot);
next              686 virt/kvm/arm/mmu.c 		pfn += (next - addr) >> PAGE_SHIFT;
next              687 virt/kvm/arm/mmu.c 	} while (addr = next, addr != end);
next              698 virt/kvm/arm/mmu.c 	unsigned long addr, next;
next              718 virt/kvm/arm/mmu.c 		next = pgd_addr_end(addr, end);
next              719 virt/kvm/arm/mmu.c 		err = create_hyp_pud_mappings(pgd, addr, next, pfn, prot);
next              722 virt/kvm/arm/mmu.c 		pfn += (next - addr) >> PAGE_SHIFT;
next              723 virt/kvm/arm/mmu.c 	} while (addr = next, addr != end);
next             1453 virt/kvm/arm/mmu.c 	phys_addr_t next;
next             1458 virt/kvm/arm/mmu.c 		next = stage2_pmd_addr_end(kvm, addr, end);
next             1464 virt/kvm/arm/mmu.c 				stage2_wp_ptes(pmd, addr, next);
next             1467 virt/kvm/arm/mmu.c 	} while (pmd++, addr = next, addr != end);
next             1480 virt/kvm/arm/mmu.c 	phys_addr_t next;
next             1484 virt/kvm/arm/mmu.c 		next = stage2_pud_addr_end(kvm, addr, end);
next             1490 virt/kvm/arm/mmu.c 				stage2_wp_pmds(kvm, pud, addr, next);
next             1493 virt/kvm/arm/mmu.c 	} while (pud++, addr = next, addr != end);
next             1505 virt/kvm/arm/mmu.c 	phys_addr_t next;
next             1521 virt/kvm/arm/mmu.c 		next = stage2_pgd_addr_end(kvm, addr, end);
next             1523 virt/kvm/arm/mmu.c 			stage2_wp_puds(kvm, pgd, addr, next);
next             1524 virt/kvm/arm/mmu.c 	} while (pgd++, addr = next, addr != end);
next              253 virt/kvm/arm/vgic/vgic-debug.c 	.next  = vgic_debug_next,
next              332 virt/kvm/arm/vgic/vgic-init.c 	struct vgic_redist_region *rdreg, *next;
next              342 virt/kvm/arm/vgic/vgic-init.c 		list_for_each_entry_safe(rdreg, next, &dist->rd_regions, list) {
next             2027 virt/kvm/arm/vgic/vgic-its.c 	struct its_device *next;
next             2032 virt/kvm/arm/vgic/vgic-its.c 	next = list_next_entry(dev, dev_list);
next             2033 virt/kvm/arm/vgic/vgic-its.c 	next_offset = next->device_id - dev->device_id;
next             2040 virt/kvm/arm/vgic/vgic-its.c 	struct its_ite *next;
next             2045 virt/kvm/arm/vgic/vgic-its.c 	next = list_next_entry(ite, ite_list);
next             2046 virt/kvm/arm/vgic/vgic-its.c 	next_offset = next->event_id - ite->event_id;
next             4385 virt/kvm/kvm_main.c 			  struct task_struct *next)