prev               69 arch/alpha/include/asm/futex.h 	u32 prev;
prev               87 arch/alpha/include/asm/futex.h 	:	"+r"(ret), "=&r"(prev), "=&r"(cmp)
prev               91 arch/alpha/include/asm/futex.h 	*uval = prev;
prev               25 arch/alpha/include/asm/gct.h 	gct6_handle prev;
prev              131 arch/alpha/include/asm/xchg.h 	unsigned long prev, tmp, cmp, addr64;
prev              148 arch/alpha/include/asm/xchg.h 	: "=&r" (prev), "=&r" (new), "=&r" (tmp), "=&r" (cmp), "=&r" (addr64)
prev              151 arch/alpha/include/asm/xchg.h 	return prev;
prev              157 arch/alpha/include/asm/xchg.h 	unsigned long prev, tmp, cmp, addr64;
prev              174 arch/alpha/include/asm/xchg.h 	: "=&r" (prev), "=&r" (new), "=&r" (tmp), "=&r" (cmp), "=&r" (addr64)
prev              177 arch/alpha/include/asm/xchg.h 	return prev;
prev              183 arch/alpha/include/asm/xchg.h 	unsigned long prev, cmp;
prev              196 arch/alpha/include/asm/xchg.h 	: "=&r"(prev), "=&r"(cmp), "=m"(*m)
prev              199 arch/alpha/include/asm/xchg.h 	return prev;
prev              205 arch/alpha/include/asm/xchg.h 	unsigned long prev, cmp;
prev              218 arch/alpha/include/asm/xchg.h 	: "=&r"(prev), "=&r"(cmp), "=m"(*m)
prev              221 arch/alpha/include/asm/xchg.h 	return prev;
prev               95 arch/alpha/kernel/core_marvel.c marvel_next_io7(struct io7 *prev)
prev               97 arch/alpha/kernel/core_marvel.c 	return (prev ? prev->next : io7_head);
prev               60 arch/alpha/kernel/proto.h struct io7 *marvel_next_io7(struct io7 *prev);
prev              447 arch/arc/include/asm/atomic.h 	s64 prev;
prev              458 arch/arc/include/asm/atomic.h 	: "=&r"(prev)
prev              464 arch/arc/include/asm/atomic.h 	return prev;
prev              469 arch/arc/include/asm/atomic.h 	s64 prev;
prev              478 arch/arc/include/asm/atomic.h 	: "=&r"(prev)
prev              484 arch/arc/include/asm/atomic.h 	return prev;
prev               19 arch/arc/include/asm/cmpxchg.h 	unsigned long prev;
prev               33 arch/arc/include/asm/cmpxchg.h 	: "=&r"(prev)	/* Early clobber, to prevent reg reuse */
prev               41 arch/arc/include/asm/cmpxchg.h 	return prev;
prev               50 arch/arc/include/asm/cmpxchg.h 	int prev;
prev               57 arch/arc/include/asm/cmpxchg.h 	prev = *p;
prev               58 arch/arc/include/asm/cmpxchg.h 	if (prev == expected)
prev               61 arch/arc/include/asm/cmpxchg.h 	return prev;
prev              129 arch/arc/include/asm/mmu_context.h static inline void switch_mm(struct mm_struct *prev, struct mm_struct *next,
prev              162 arch/arc/include/asm/mmu_context.h #define activate_mm(prev, next)		switch_mm(prev, next, NULL)
prev               36 arch/arc/include/asm/switch_to.h #define switch_to(prev, next, last)	\
prev               38 arch/arc/include/asm/switch_to.h 	ARC_EZNPS_DP_PREV(prev, next);	\
prev               39 arch/arc/include/asm/switch_to.h 	ARC_FPU_PREV(prev, next);	\
prev               40 arch/arc/include/asm/switch_to.h 	last = __switch_to(prev, next);\
prev               27 arch/arc/kernel/ctx_sw.c 	unsigned int prev = (unsigned int)prev_task;
prev              117 arch/arc/kernel/ctx_sw.c 		: "n"(KSP_WORD_OFF), "r"(next), "r"(prev)
prev               30 arch/arc/kernel/fpu.c void fpu_save_restore(struct task_struct *prev, struct task_struct *next)
prev               32 arch/arc/kernel/fpu.c 	unsigned int *saveto = &prev->thread.fpu.aux_dpfp[0].l;
prev              413 arch/arc/kernel/unwind.c 	struct unwind_table *table = info->table, *prev;
prev              415 arch/arc/kernel/unwind.c 	for (prev = &root_table; prev->link && prev->link != table;
prev              416 arch/arc/kernel/unwind.c 	     prev = prev->link)
prev              419 arch/arc/kernel/unwind.c 	if (prev->link) {
prev              425 arch/arc/kernel/unwind.c 			prev->link = table->link;
prev              426 arch/arc/kernel/unwind.c 			if (!prev->link)
prev              427 arch/arc/kernel/unwind.c 				last_table = prev;
prev               10 arch/arc/plat-eznps/ctop.c void dp_save_restore(struct task_struct *prev, struct task_struct *next)
prev               12 arch/arc/plat-eznps/ctop.c 	struct eznps_dp *prev_task_dp = &prev->thread.dp;
prev              105 arch/arm/include/asm/mmu_context.h #define activate_mm(prev,next)		switch_mm(prev, next, NULL)
prev              128 arch/arm/include/asm/mmu_context.h switch_mm(struct mm_struct *prev, struct mm_struct *next,
prev              144 arch/arm/include/asm/mmu_context.h 	if (!cpumask_test_and_set_cpu(cpu, mm_cpumask(next)) || prev != next) {
prev              147 arch/arm/include/asm/mmu_context.h 			cpumask_clear_cpu(cpu, mm_cpumask(prev));
prev               26 arch/arm/include/asm/switch_to.h #define switch_to(prev,next,last)					\
prev               29 arch/arm/include/asm/switch_to.h 	last = __switch_to(prev,task_thread_info(prev), task_thread_info(next));	\
prev              120 arch/arm/kernel/module-plts.c 	const Elf32_Rel *prev;
prev              130 arch/arm/kernel/module-plts.c 	prev = rel + num - 1;
prev              131 arch/arm/kernel/module-plts.c 	return cmp_rel(rel + num, prev) == 0 &&
prev              132 arch/arm/kernel/module-plts.c 	       is_zero_addend_relocation(base, prev);
prev              800 arch/arm/mach-omap1/clock.c 	if (clk->node.next || clk->node.prev)
prev               54 arch/arm/mach-omap2/pm-debug.c void pm_dbg_update_time(struct powerdomain *pwrdm, int prev)
prev               64 arch/arm/mach-omap2/pm-debug.c 	pwrdm->state_timer[prev] += t - pwrdm->timer;
prev               59 arch/arm/mach-omap2/pm.h extern void pm_dbg_update_time(struct powerdomain *pwrdm, int prev);
prev               61 arch/arm/mach-omap2/pm.h #define pm_dbg_update_time(pwrdm, prev) do {} while (0);
prev              164 arch/arm/mach-omap2/powerdomain.c 	int prev, next, state, trace_state = 0;
prev              173 arch/arm/mach-omap2/powerdomain.c 		prev = pwrdm->state;
prev              176 arch/arm/mach-omap2/powerdomain.c 		prev = pwrdm_read_prev_pwrst(pwrdm);
prev              177 arch/arm/mach-omap2/powerdomain.c 		if (pwrdm->state != prev)
prev              178 arch/arm/mach-omap2/powerdomain.c 			pwrdm->state_counter[prev]++;
prev              179 arch/arm/mach-omap2/powerdomain.c 		if (prev == PWRDM_POWER_RET)
prev              186 arch/arm/mach-omap2/powerdomain.c 		if (next != prev) {
prev              189 arch/arm/mach-omap2/powerdomain.c 				       ((prev & OMAP_POWERSTATE_MASK) << 0));
prev              199 arch/arm/mach-omap2/powerdomain.c 	if (state != prev)
prev              202 arch/arm/mach-omap2/powerdomain.c 	pm_dbg_update_time(pwrdm, prev);
prev               82 arch/arm/mach-omap2/sdrc2xxx.c 	u32 prev = curr_perf_level;
prev               86 arch/arm/mach-omap2/sdrc2xxx.c 		return prev;
prev               93 arch/arm/mach-omap2/sdrc2xxx.c 		return prev;
prev              110 arch/arm/mach-omap2/sdrc2xxx.c 	return prev;
prev              230 arch/arm64/include/asm/mmu_context.h switch_mm(struct mm_struct *prev, struct mm_struct *next,
prev              233 arch/arm64/include/asm/mmu_context.h 	if (prev != next)
prev              246 arch/arm64/include/asm/mmu_context.h #define activate_mm(prev,next)	switch_mm(prev, next, current)
prev              251 arch/arm64/include/asm/processor.h extern struct task_struct *cpu_switch_to(struct task_struct *prev,
prev              139 arch/arm64/kernel/armv8_deprecated.c 				       enum insn_emulation_mode prev)
prev              143 arch/arm64/kernel/armv8_deprecated.c 	switch (prev) {
prev              504 arch/arm64/kernel/process.c __notrace_funcgraph struct task_struct *__switch_to(struct task_struct *prev,
prev              527 arch/arm64/kernel/process.c 	last = cpu_switch_to(prev, next);
prev               17 arch/c6x/include/asm/switch_to.h asmlinkage void *__switch_to(struct thread_struct *prev,
prev               21 arch/c6x/include/asm/switch_to.h #define switch_to(prev, next, last)				\
prev               24 arch/c6x/include/asm/switch_to.h 		(last) = __switch_to(&(prev)->thread,		\
prev               25 arch/c6x/include/asm/switch_to.h 				     &(next)->thread, (prev));	\
prev               27 arch/csky/include/asm/mmu_context.h #define activate_mm(prev,next)		switch_mm(prev, next, current)
prev               36 arch/csky/include/asm/mmu_context.h switch_mm(struct mm_struct *prev, struct mm_struct *next,
prev               41 arch/csky/include/asm/mmu_context.h 	if (prev != next)
prev               10 arch/csky/include/asm/switch_to.h static inline void __switch_to_fpu(struct task_struct *prev,
prev               13 arch/csky/include/asm/switch_to.h 	save_to_user_fp(&prev->thread.user_fp);
prev               17 arch/csky/include/asm/switch_to.h static inline void __switch_to_fpu(struct task_struct *prev,
prev               28 arch/csky/include/asm/switch_to.h #define switch_to(prev, next, last)					\
prev               30 arch/csky/include/asm/switch_to.h 		struct task_struct *__prev = (prev);			\
prev               33 arch/csky/include/asm/switch_to.h 		((last) = __switch_to((prev), (next)));			\
prev               35 arch/h8300/include/asm/switch_to.h #define switch_to(prev, next, last) \
prev               45 arch/h8300/include/asm/switch_to.h 		: "r" (&(prev->thread)),		    \
prev               47 arch/h8300/include/asm/switch_to.h 		  "g" (prev)				    \
prev               77 arch/hexagon/include/asm/futex.h 	int prev;
prev               99 arch/hexagon/include/asm/futex.h 	: "+r" (ret), "=&r" (prev), "+m" (*uaddr)
prev              103 arch/hexagon/include/asm/futex.h 	*uval = prev;
prev               55 arch/hexagon/include/asm/mmu_context.h static inline void switch_mm(struct mm_struct *prev, struct mm_struct *next,
prev               64 arch/hexagon/include/asm/mmu_context.h 	if (next->context.generation < prev->context.generation) {
prev               68 arch/hexagon/include/asm/mmu_context.h 		next->context.generation = prev->context.generation;
prev               77 arch/hexagon/include/asm/mmu_context.h static inline void activate_mm(struct mm_struct *prev, struct mm_struct *next)
prev               82 arch/hexagon/include/asm/mmu_context.h 	switch_mm(prev, next, current_thread_info()->task);
prev               94 arch/ia64/include/asm/futex.h 		unsigned long prev;
prev              101 arch/ia64/include/asm/futex.h 			: "+r" (r8), "=&r" (prev)
prev              105 arch/ia64/include/asm/futex.h 		*uval = prev;
prev              187 arch/ia64/include/asm/mmu_context.h activate_mm (struct mm_struct *prev, struct mm_struct *next)
prev               45 arch/ia64/include/asm/switch_to.h #define __switch_to(prev,next,last) do {							 \
prev               46 arch/ia64/include/asm/switch_to.h 	if (IA64_HAS_EXTRA_STATE(prev))								 \
prev               47 arch/ia64/include/asm/switch_to.h 		ia64_save_extra(prev);								 \
prev               61 arch/ia64/include/asm/switch_to.h # define switch_to(prev,next,last) do {						\
prev               62 arch/ia64/include/asm/switch_to.h 	if (ia64_psr(task_pt_regs(prev))->mfh && ia64_is_local_fpu_owner(prev)) {				\
prev               63 arch/ia64/include/asm/switch_to.h 		ia64_psr(task_pt_regs(prev))->mfh = 0;			\
prev               64 arch/ia64/include/asm/switch_to.h 		(prev)->thread.flags |= IA64_THREAD_FPH_VALID;			\
prev               65 arch/ia64/include/asm/switch_to.h 		__ia64_save_fpu((prev)->thread.fph);				\
prev               67 arch/ia64/include/asm/switch_to.h 	__switch_to(prev, next, last);						\
prev               76 arch/ia64/include/asm/switch_to.h # define switch_to(prev,next,last)	__switch_to(prev, next, last)
prev             1051 arch/ia64/kernel/efi.c 	struct kern_memdesc *k, *prev = NULL;
prev             1152 arch/ia64/kernel/efi.c 		if (prev && kmd_end(prev) == md->phys_addr) {
prev             1153 arch/ia64/kernel/efi.c 			prev->num_pages += (ae - as) >> EFI_PAGE_SHIFT;
prev             1161 arch/ia64/kernel/efi.c 		prev = k++;
prev              108 arch/ia64/kernel/time.c void arch_vtime_task_switch(struct task_struct *prev)
prev              110 arch/ia64/kernel/time.c 	struct thread_info *pi = task_thread_info(prev);
prev              646 arch/ia64/kernel/unwind.c 	struct unw_reg_state *copy, *prev = NULL, *first = NULL;
prev              656 arch/ia64/kernel/unwind.c 			prev->next = copy;
prev              660 arch/ia64/kernel/unwind.c 		prev = copy;
prev             1269 arch/ia64/kernel/unwind.c 	struct unw_script *script, *prev, *tmp;
prev             1300 arch/ia64/kernel/unwind.c 		prev = NULL;
prev             1303 arch/ia64/kernel/unwind.c 				if (prev)
prev             1304 arch/ia64/kernel/unwind.c 					prev->coll_chain = tmp->coll_chain;
prev             1309 arch/ia64/kernel/unwind.c 				prev = tmp;
prev             1534 arch/ia64/kernel/unwind.c 	struct unw_table *table, *prev;
prev             1563 arch/ia64/kernel/unwind.c 	prev = NULL;
prev             1573 arch/ia64/kernel/unwind.c 			if (prev && prev != unw.tables) {
prev             1575 arch/ia64/kernel/unwind.c 				prev->next = table->next;
prev             1582 arch/ia64/kernel/unwind.c 		prev = table;
prev             2130 arch/ia64/kernel/unwind.c 	struct unw_table *table, *prev;
prev             2152 arch/ia64/kernel/unwind.c 		for (prev = (struct unw_table *) &unw.tables; prev; prev = prev->next)
prev             2153 arch/ia64/kernel/unwind.c 			if (prev->next == table)
prev             2155 arch/ia64/kernel/unwind.c 		if (!prev) {
prev             2161 arch/ia64/kernel/unwind.c 		prev->next = table->next;
prev              173 arch/m68k/include/asm/atomic.h 	int prev;
prev              176 arch/m68k/include/asm/atomic.h 	prev = atomic_read(v);
prev              177 arch/m68k/include/asm/atomic.h 	if (prev == old)
prev              180 arch/m68k/include/asm/atomic.h 	return prev;
prev              186 arch/m68k/include/asm/atomic.h 	int prev;
prev              189 arch/m68k/include/asm/atomic.h 	prev = atomic_read(v);
prev              192 arch/m68k/include/asm/atomic.h 	return prev;
prev               75 arch/m68k/include/asm/mmu_context.h static inline void switch_mm(struct mm_struct *prev, struct mm_struct *next,
prev              197 arch/m68k/include/asm/mmu_context.h static inline void switch_mm(struct mm_struct *prev, struct mm_struct *next,
prev              283 arch/m68k/include/asm/mmu_context.h static inline void switch_mm(struct mm_struct *prev, struct mm_struct *next, struct task_struct *tsk)
prev              285 arch/m68k/include/asm/mmu_context.h 	if (prev != next) {
prev              316 arch/m68k/include/asm/mmu_context.h static inline void switch_mm(struct mm_struct *prev, struct mm_struct *next, struct task_struct *tsk)
prev               31 arch/m68k/include/asm/switch_to.h #define switch_to(prev,next,last) do { \
prev               32 arch/m68k/include/asm/switch_to.h   register void *_prev __asm__ ("a0") = (prev); \
prev              110 arch/m68k/sun3/sun3dvma.c 	struct hole *prev = NULL;
prev              117 arch/m68k/sun3/sun3dvma.c 		if(!prev) {
prev              118 arch/m68k/sun3/sun3dvma.c 			prev = hole;
prev              122 arch/m68k/sun3/sun3dvma.c 		if(hole->end == prev->start) {
prev              123 arch/m68k/sun3/sun3dvma.c 			hole->size += prev->size;
prev              124 arch/m68k/sun3/sun3dvma.c 			hole->end = prev->end;
prev              125 arch/m68k/sun3/sun3dvma.c 			list_move(&(prev->list), &hole_cache);
prev               72 arch/microblaze/include/asm/futex.h 	u32 prev;
prev               91 arch/microblaze/include/asm/futex.h 		: "+r" (ret), "=&r" (prev), "=&r"(cmp)	\
prev               94 arch/microblaze/include/asm/futex.h 	*uval = prev;
prev              120 arch/microblaze/include/asm/mmu_context_mm.h static inline void switch_mm(struct mm_struct *prev, struct mm_struct *next,
prev               15 arch/microblaze/include/asm/switch_to.h extern struct task_struct *_switch_to(struct thread_info *prev,
prev               18 arch/microblaze/include/asm/switch_to.h #define switch_to(prev, next, last)					\
prev               20 arch/microblaze/include/asm/switch_to.h 		(last) = _switch_to(task_thread_info(prev),		\
prev              158 arch/mips/include/asm/mmu_context.h static inline void switch_mm(struct mm_struct *prev, struct mm_struct *next,
prev              172 arch/mips/include/asm/mmu_context.h 	cpumask_clear_cpu(cpu, mm_cpumask(prev));
prev              188 arch/mips/include/asm/mmu_context.h #define activate_mm(prev, next)	switch_mm(prev, next, current)
prev               32 arch/mips/include/asm/switch_to.h extern asmlinkage struct task_struct *resume(struct task_struct *prev,
prev               52 arch/mips/include/asm/switch_to.h #define __mips_mt_fpaff_switch_to(prev)					\
prev               54 arch/mips/include/asm/switch_to.h 	struct thread_info *__prev_ti = task_thread_info(prev);		\
prev               58 arch/mips/include/asm/switch_to.h 	    (!(KSTK_STATUS(prev) & ST0_CU1))) {				\
prev               60 arch/mips/include/asm/switch_to.h 		prev->cpus_mask = prev->thread.user_cpus_allowed;	\
prev               66 arch/mips/include/asm/switch_to.h #define __mips_mt_fpaff_switch_to(prev) do { (void) (prev); } while (0)
prev              109 arch/mips/include/asm/switch_to.h #define switch_to(prev, next, last)					\
prev              111 arch/mips/include/asm/switch_to.h 	__mips_mt_fpaff_switch_to(prev);				\
prev              112 arch/mips/include/asm/switch_to.h 	lose_fpu_inatomic(1, prev);					\
prev              116 arch/mips/include/asm/switch_to.h 		__save_dsp(prev);					\
prev              121 arch/mips/include/asm/switch_to.h 		if ((KSTK_STATUS(prev) & ST0_CU2)) {			\
prev              123 arch/mips/include/asm/switch_to.h 				KSTK_STATUS(prev) &= ~ST0_CU2;		\
prev              124 arch/mips/include/asm/switch_to.h 			cop2_save(prev);				\
prev              137 arch/mips/include/asm/switch_to.h 	(last) = resume(prev, next, task_thread_info(next));		\
prev               47 arch/mips/kernel/csrc-r4k.c 	unsigned int prev, curr, i;
prev               55 arch/mips/kernel/csrc-r4k.c 	for (i = 0, prev = rdhwr_count(); i < 100; i++) {
prev               58 arch/mips/kernel/csrc-r4k.c 		if (curr != prev)
prev               61 arch/mips/kernel/csrc-r4k.c 		prev = curr;
prev               54 arch/nds32/include/asm/mmu_context.h static inline void switch_mm(struct mm_struct *prev, struct mm_struct *next,
prev               59 arch/nds32/include/asm/mmu_context.h 	if (!cpumask_test_and_set_cpu(cpu, mm_cpumask(next)) || prev != next) {
prev               66 arch/nds32/include/asm/mmu_context.h #define activate_mm(prev,next)	switch_mm(prev, next, NULL)
prev              205 arch/nds32/kernel/process.c struct task_struct *_switch_fpu(struct task_struct *prev, struct task_struct *next)
prev              208 arch/nds32/kernel/process.c 	unlazy_fpu(prev);
prev              212 arch/nds32/kernel/process.c 	return prev;
prev               54 arch/nios2/include/asm/mmu_context.h void switch_mm(struct mm_struct *prev, struct mm_struct *next,
prev               66 arch/nios2/include/asm/mmu_context.h void activate_mm(struct mm_struct *prev, struct mm_struct *next);
prev               17 arch/nios2/include/asm/switch_to.h #define switch_to(prev, next, last)			\
prev               26 arch/nios2/include/asm/switch_to.h 		: "r" (prev), "r" (next)		\
prev               80 arch/nios2/mm/mmu_context.c void switch_mm(struct mm_struct *prev, struct mm_struct *next,
prev              106 arch/nios2/mm/mmu_context.c void activate_mm(struct mm_struct *prev, struct mm_struct *next)
prev               73 arch/openrisc/include/asm/futex.h 	u32 prev;
prev               94 arch/openrisc/include/asm/futex.h 		: "+r" (ret), "=&r" (prev), "+m" (*uaddr) \
prev               99 arch/openrisc/include/asm/futex.h 	*uval = prev;
prev               22 arch/openrisc/include/asm/mmu_context.h extern void switch_mm(struct mm_struct *prev, struct mm_struct *next,
prev               27 arch/openrisc/include/asm/mmu_context.h #define activate_mm(prev, next) switch_mm((prev), (next), NULL)
prev              138 arch/openrisc/mm/tlb.c void switch_mm(struct mm_struct *prev, struct mm_struct *next,
prev              153 arch/openrisc/mm/tlb.c 	if (prev != next)
prev              154 arch/openrisc/mm/tlb.c 		local_flush_tlb_mm(prev);
prev               53 arch/parisc/include/asm/mmu_context.h static inline void switch_mm_irqs_off(struct mm_struct *prev,
prev               56 arch/parisc/include/asm/mmu_context.h 	if (prev != next) {
prev               62 arch/parisc/include/asm/mmu_context.h static inline void switch_mm(struct mm_struct *prev,
prev               67 arch/parisc/include/asm/mmu_context.h 	if (prev == next)
prev               71 arch/parisc/include/asm/mmu_context.h 	switch_mm_irqs_off(prev, next, tsk);
prev               78 arch/parisc/include/asm/mmu_context.h static inline void activate_mm(struct mm_struct *prev, struct mm_struct *next)
prev               93 arch/parisc/include/asm/mmu_context.h 	switch_mm(prev,next,current);
prev                9 arch/parisc/include/asm/switch_to.h #define switch_to(prev, next, last) do {			\
prev               10 arch/parisc/include/asm/switch_to.h 	(last) = _switch_to(prev, next);			\
prev               86 arch/parisc/kernel/sys_parisc.c 	struct vm_area_struct *vma, *prev;
prev              113 arch/parisc/kernel/sys_parisc.c 		vma = find_vma_prev(mm, addr, &prev);
prev              116 arch/parisc/kernel/sys_parisc.c 		    (!prev || addr >= vm_end_gap(prev)))
prev              140 arch/parisc/kernel/sys_parisc.c 	struct vm_area_struct *vma, *prev;
prev              170 arch/parisc/kernel/sys_parisc.c 		vma = find_vma_prev(mm, addr, &prev);
prev              173 arch/parisc/kernel/sys_parisc.c 		    (!prev || addr >= vm_end_gap(prev)))
prev               62 arch/parisc/lib/bitops.c 	u64 prev;
prev               65 arch/parisc/lib/bitops.c 	if ((prev = *ptr) == old)
prev               68 arch/parisc/lib/bitops.c 	return prev;
prev               74 arch/parisc/lib/bitops.c 	unsigned int prev;
prev               77 arch/parisc/lib/bitops.c 	if ((prev = *ptr) == old)
prev               80 arch/parisc/lib/bitops.c 	return (unsigned long)prev;
prev              130 arch/parisc/mm/fault.c 					prev = tree;
prev              131 arch/parisc/mm/fault.c 					if (prev->vm_next == NULL)
prev              133 arch/parisc/mm/fault.c 					if (prev->vm_next->vm_start > addr)
prev              117 arch/powerpc/boot/libfdt-wrapper.c static void *fdt_wrapper_find_node_by_prop_value(const void *prev,
prev              122 arch/powerpc/boot/libfdt-wrapper.c 	int offset = fdt_node_offset_by_prop_value(fdt, devp_offset_find(prev),
prev              127 arch/powerpc/boot/libfdt-wrapper.c static void *fdt_wrapper_find_node_by_compatible(const void *prev,
prev              130 arch/powerpc/boot/libfdt-wrapper.c 	int offset = fdt_node_offset_by_compatible(fdt, devp_offset_find(prev),
prev               48 arch/powerpc/boot/ops.h 	void *(*find_node_by_prop_value)(const void *prev,
prev               51 arch/powerpc/boot/ops.h 	void *(*find_node_by_compatible)(const void *prev,
prev              147 arch/powerpc/boot/ops.h static inline void *find_node_by_prop_value(const void *prev,
prev              152 arch/powerpc/boot/ops.h 		return dt_ops.find_node_by_prop_value(prev, propname,
prev              158 arch/powerpc/boot/ops.h static inline void *find_node_by_prop_value_str(const void *prev,
prev              162 arch/powerpc/boot/ops.h 	return find_node_by_prop_value(prev, propname, propval,
prev              166 arch/powerpc/boot/ops.h static inline void *find_node_by_devtype(const void *prev,
prev              169 arch/powerpc/boot/ops.h 	return find_node_by_prop_value_str(prev, "device_type", type);
prev              185 arch/powerpc/boot/ops.h static inline void *find_node_by_compatible(const void *prev,
prev              189 arch/powerpc/boot/ops.h 		return dt_ops.find_node_by_compatible(prev, compat);
prev               20 arch/powerpc/include/asm/cmpxchg.h 	unsigned int prev, prev_mask, tmp, bitoff, off;		\
prev               35 arch/powerpc/include/asm/cmpxchg.h 	: "=&r" (prev), "=&r" (tmp), "+m" (*(u32*)p)		\
prev               39 arch/powerpc/include/asm/cmpxchg.h 	return prev >> bitoff;					\
prev               46 arch/powerpc/include/asm/cmpxchg.h 	unsigned int prev, prev_mask, tmp, bitoff, off;		\
prev               69 arch/powerpc/include/asm/cmpxchg.h 	: "=&r" (prev), "=&r" (tmp), "+m" (*(u32*)p)		\
prev               73 arch/powerpc/include/asm/cmpxchg.h 	return prev >> bitoff;					\
prev               91 arch/powerpc/include/asm/cmpxchg.h 	unsigned long prev;
prev               98 arch/powerpc/include/asm/cmpxchg.h 	: "=&r" (prev), "+m" (*(volatile unsigned int *)p)
prev              102 arch/powerpc/include/asm/cmpxchg.h 	return prev;
prev              108 arch/powerpc/include/asm/cmpxchg.h 	unsigned long prev;
prev              115 arch/powerpc/include/asm/cmpxchg.h 	: "=&r" (prev), "+m" (*p)
prev              119 arch/powerpc/include/asm/cmpxchg.h 	return prev;
prev              126 arch/powerpc/include/asm/cmpxchg.h 	unsigned long prev;
prev              133 arch/powerpc/include/asm/cmpxchg.h 	: "=&r" (prev), "+m" (*(volatile unsigned long *)p)
prev              137 arch/powerpc/include/asm/cmpxchg.h 	return prev;
prev              143 arch/powerpc/include/asm/cmpxchg.h 	unsigned long prev;
prev              150 arch/powerpc/include/asm/cmpxchg.h 	: "=&r" (prev), "+m" (*p)
prev              154 arch/powerpc/include/asm/cmpxchg.h 	return prev;
prev              225 arch/powerpc/include/asm/cmpxchg.h 	unsigned int prev;
prev              238 arch/powerpc/include/asm/cmpxchg.h 	: "=&r" (prev), "+m" (*p)
prev              242 arch/powerpc/include/asm/cmpxchg.h 	return prev;
prev              249 arch/powerpc/include/asm/cmpxchg.h 	unsigned int prev;
prev              260 arch/powerpc/include/asm/cmpxchg.h 	: "=&r" (prev), "+m" (*p)
prev              264 arch/powerpc/include/asm/cmpxchg.h 	return prev;
prev              270 arch/powerpc/include/asm/cmpxchg.h 	unsigned long prev;
prev              280 arch/powerpc/include/asm/cmpxchg.h 	: "=&r" (prev), "+m" (*p)
prev              284 arch/powerpc/include/asm/cmpxchg.h 	return prev;
prev              298 arch/powerpc/include/asm/cmpxchg.h 	unsigned long prev;
prev              310 arch/powerpc/include/asm/cmpxchg.h 	: "=&r" (prev), "+m" (*p)
prev              314 arch/powerpc/include/asm/cmpxchg.h 	return prev;
prev              321 arch/powerpc/include/asm/cmpxchg.h 	unsigned long prev;
prev              333 arch/powerpc/include/asm/cmpxchg.h 	: "=&r" (prev), "+m" (*p)
prev              337 arch/powerpc/include/asm/cmpxchg.h 	return prev;
prev              344 arch/powerpc/include/asm/cmpxchg.h 	unsigned long prev;
prev              354 arch/powerpc/include/asm/cmpxchg.h 	: "=&r" (prev), "+m" (*p)
prev              358 arch/powerpc/include/asm/cmpxchg.h 	return prev;
prev              364 arch/powerpc/include/asm/cmpxchg.h 	unsigned long prev;
prev              373 arch/powerpc/include/asm/cmpxchg.h 	: "=&r" (prev), "+m" (*p)
prev              377 arch/powerpc/include/asm/cmpxchg.h 	return prev;
prev              383 arch/powerpc/include/asm/cmpxchg.h 	unsigned long prev;
prev              394 arch/powerpc/include/asm/cmpxchg.h 	: "=&r" (prev), "+m" (*p)
prev              398 arch/powerpc/include/asm/cmpxchg.h 	return prev;
prev               54 arch/powerpc/include/asm/cputime.h static inline void arch_vtime_task_switch(struct task_struct *prev)
prev               57 arch/powerpc/include/asm/cputime.h 	struct cpu_accounting_data *acct0 = get_accounting(prev);
prev               74 arch/powerpc/include/asm/futex.h 	u32 prev;
prev               96 arch/powerpc/include/asm/futex.h         : "+r" (ret), "=&r" (prev), "+m" (*uaddr)
prev              100 arch/powerpc/include/asm/futex.h 	*uval = prev;
prev                4 arch/powerpc/include/asm/membarrier.h static inline void membarrier_arch_switch_mm(struct mm_struct *prev,
prev               17 arch/powerpc/include/asm/membarrier.h 		      MEMBARRIER_STATE_GLOBAL_EXPEDITED)) || !prev))
prev               61 arch/powerpc/include/asm/mmu_context.h extern void radix__switch_mmu_context(struct mm_struct *prev,
prev               63 arch/powerpc/include/asm/mmu_context.h static inline void switch_mmu_context(struct mm_struct *prev,
prev               68 arch/powerpc/include/asm/mmu_context.h 		return radix__switch_mmu_context(prev, next);
prev              104 arch/powerpc/include/asm/mmu_context.h extern void switch_mmu_context(struct mm_struct *prev, struct mm_struct *next,
prev              196 arch/powerpc/include/asm/mmu_context.h extern void switch_mm_irqs_off(struct mm_struct *prev, struct mm_struct *next,
prev              199 arch/powerpc/include/asm/mmu_context.h static inline void switch_mm(struct mm_struct *prev, struct mm_struct *next,
prev              205 arch/powerpc/include/asm/mmu_context.h 	switch_mm_irqs_off(prev, next, tsk);
prev              217 arch/powerpc/include/asm/mmu_context.h static inline void activate_mm(struct mm_struct *prev, struct mm_struct *next)
prev              219 arch/powerpc/include/asm/mmu_context.h 	switch_mm(prev, next, current);
prev               84 arch/powerpc/include/asm/pgtable-be-types.h 	__be64 prev;
prev               87 arch/powerpc/include/asm/pgtable-be-types.h 	prev = (__force __be64)__cmpxchg_u64(p, (__force unsigned long)pte_raw(old),
prev               90 arch/powerpc/include/asm/pgtable-be-types.h 	return pte_raw(old) == prev;
prev               96 arch/powerpc/include/asm/pgtable-be-types.h 	__be64 prev;
prev               98 arch/powerpc/include/asm/pgtable-be-types.h 	prev = (__force __be64)__cmpxchg_u64(p, (__force unsigned long)pmd_raw(old),
prev              101 arch/powerpc/include/asm/pgtable-be-types.h 	return pmd_raw(old) == prev;
prev               16 arch/powerpc/include/asm/switch_to.h #define switch_to(prev, next, last)	((last) = __switch_to((prev), (next)))
prev               18 arch/powerpc/include/asm/switch_to.h extern struct task_struct *_switch(struct thread_struct *prev,
prev              788 arch/powerpc/kernel/nvram_64.c 	struct nvram_partition *part, *prev, *tmp;
prev              807 arch/powerpc/kernel/nvram_64.c 	prev = NULL;
prev              810 arch/powerpc/kernel/nvram_64.c 			prev = NULL;
prev              813 arch/powerpc/kernel/nvram_64.c 		if (prev) {
prev              814 arch/powerpc/kernel/nvram_64.c 			prev->header.length += part->header.length;
prev              815 arch/powerpc/kernel/nvram_64.c 			prev->header.checksum = nvram_checksum(&prev->header);
prev              816 arch/powerpc/kernel/nvram_64.c 			rc = nvram_write_header(prev);
prev              824 arch/powerpc/kernel/nvram_64.c 			prev = part;
prev              985 arch/powerpc/kernel/process.c static inline void __switch_to_tm(struct task_struct *prev,
prev              989 arch/powerpc/kernel/process.c 		if (tm_enabled(prev) || tm_enabled(new))
prev              992 arch/powerpc/kernel/process.c 		if (tm_enabled(prev)) {
prev              993 arch/powerpc/kernel/process.c 			prev->thread.load_tm++;
prev              994 arch/powerpc/kernel/process.c 			tm_reclaim_task(prev);
prev              995 arch/powerpc/kernel/process.c 			if (!MSR_TM_ACTIVE(prev->thread.regs->msr) && prev->thread.load_tm == 0)
prev              996 arch/powerpc/kernel/process.c 				prev->thread.regs->msr &= ~MSR_TM;
prev             1048 arch/powerpc/kernel/process.c #define __switch_to_tm(prev, new)
prev             1122 arch/powerpc/kernel/process.c struct task_struct *__switch_to(struct task_struct *prev,
prev             1163 arch/powerpc/kernel/process.c 	save_sprs(&prev->thread);
prev             1166 arch/powerpc/kernel/process.c 	giveup_all(prev);
prev             1168 arch/powerpc/kernel/process.c 	__switch_to_tm(prev, new);
prev              164 arch/powerpc/kvm/book3s_hv_rm_mmu.c 	struct revmap_entry *next, *prev;
prev              180 arch/powerpc/kvm/book3s_hv_rm_mmu.c 	prev = real_vmalloc_addr(&kvm->arch.hpt.rev[rev->back]);
prev              182 arch/powerpc/kvm/book3s_hv_rm_mmu.c 	prev->forw = rev->forw;
prev               39 arch/powerpc/lib/rheap.c 	pp = (unsigned long *)&l->prev;
prev              489 arch/powerpc/lib/rheap.c 			list_add(&spblk->list, blk->list.prev);
prev              304 arch/powerpc/mm/book3s64/mmu_context.c void radix__switch_mmu_context(struct mm_struct *prev, struct mm_struct *next)
prev              201 arch/powerpc/mm/mem.c 	struct memblock_region *reg, *prev = NULL;
prev              204 arch/powerpc/mm/mem.c 		if (prev &&
prev              205 arch/powerpc/mm/mem.c 		    memblock_region_memory_end_pfn(prev) < memblock_region_memory_base_pfn(reg))
prev              206 arch/powerpc/mm/mem.c 			register_nosave_region(memblock_region_memory_end_pfn(prev),
prev              208 arch/powerpc/mm/mem.c 		prev = reg;
prev               34 arch/powerpc/mm/mmu_context.c void switch_mm_irqs_off(struct mm_struct *prev, struct mm_struct *next,
prev               74 arch/powerpc/mm/mmu_context.c 	if (prev == next)
prev               87 arch/powerpc/mm/mmu_context.c 		membarrier_arch_switch_mm(prev, next, tsk);
prev               93 arch/powerpc/mm/mmu_context.c 	switch_mmu_context(prev, next, tsk);
prev              261 arch/powerpc/mm/nohash/mmu_context.c void switch_mmu_context(struct mm_struct *prev, struct mm_struct *next,
prev              279 arch/powerpc/mm/nohash/mmu_context.c 	if (prev) {
prev              280 arch/powerpc/mm/nohash/mmu_context.c 		pr_hardcont(" (old=0x%p a=%d)", prev, prev->context.active);
prev              281 arch/powerpc/mm/nohash/mmu_context.c 		WARN_ON(prev->context.active < 1);
prev              282 arch/powerpc/mm/nohash/mmu_context.c 		prev->context.active--;
prev              330 arch/powerpc/mm/slice.c 	unsigned long addr, found, prev;
prev              363 arch/powerpc/mm/slice.c 		else if (slice_scan_available(addr - 1, available, 0, &prev)) {
prev              364 arch/powerpc/mm/slice.c 			addr = prev;
prev              126 arch/powerpc/perf/8xx-pmu.c 	s64 prev, val = 0, delta = 0;
prev              132 arch/powerpc/perf/8xx-pmu.c 		prev = local64_read(&event->hw.prev_count);
prev              136 arch/powerpc/perf/8xx-pmu.c 			delta = 16 * (val - prev);
prev              140 arch/powerpc/perf/8xx-pmu.c 			delta = prev - val;
prev              146 arch/powerpc/perf/8xx-pmu.c 			delta = (s64)((s32)val - (s32)prev);
prev              150 arch/powerpc/perf/8xx-pmu.c 			delta = (s64)((s32)val - (s32)prev);
prev              153 arch/powerpc/perf/8xx-pmu.c 	} while (local64_cmpxchg(&event->hw.prev_count, prev, val) != prev);
prev             1034 arch/powerpc/perf/core-book3s.c static u64 check_and_compute_delta(u64 prev, u64 val)
prev             1036 arch/powerpc/perf/core-book3s.c 	u64 delta = (val - prev) & 0xfffffffful;
prev             1047 arch/powerpc/perf/core-book3s.c 	if (prev > val && (prev - val) < 256)
prev             1055 arch/powerpc/perf/core-book3s.c 	s64 val, delta, prev;
prev             1075 arch/powerpc/perf/core-book3s.c 		prev = local64_read(&event->hw.prev_count);
prev             1078 arch/powerpc/perf/core-book3s.c 		delta = check_and_compute_delta(prev, val);
prev             1081 arch/powerpc/perf/core-book3s.c 	} while (local64_cmpxchg(&event->hw.prev_count, prev, val) != prev);
prev             1095 arch/powerpc/perf/core-book3s.c 		prev = local64_read(&event->hw.period_left);
prev             1096 arch/powerpc/perf/core-book3s.c 		val = prev - delta;
prev             1099 arch/powerpc/perf/core-book3s.c 	} while (local64_cmpxchg(&event->hw.period_left, prev, val) != prev);
prev             1117 arch/powerpc/perf/core-book3s.c 	u64 val, prev, delta;
prev             1125 arch/powerpc/perf/core-book3s.c 		prev = local64_read(&event->hw.prev_count);
prev             1127 arch/powerpc/perf/core-book3s.c 		delta = check_and_compute_delta(prev, val);
prev             1137 arch/powerpc/perf/core-book3s.c 	u64 val, prev;
prev             1144 arch/powerpc/perf/core-book3s.c 		prev = local64_read(&event->hw.prev_count);
prev             1145 arch/powerpc/perf/core-book3s.c 		if (check_and_compute_delta(prev, val))
prev             2042 arch/powerpc/perf/core-book3s.c 	s64 prev, delta, left;
prev             2051 arch/powerpc/perf/core-book3s.c 	prev = local64_read(&event->hw.prev_count);
prev             2052 arch/powerpc/perf/core-book3s.c 	delta = check_and_compute_delta(prev, val);
prev              177 arch/powerpc/perf/core-fsl-emb.c 	s64 val, delta, prev;
prev              188 arch/powerpc/perf/core-fsl-emb.c 		prev = local64_read(&event->hw.prev_count);
prev              191 arch/powerpc/perf/core-fsl-emb.c 	} while (local64_cmpxchg(&event->hw.prev_count, prev, val) != prev);
prev              194 arch/powerpc/perf/core-fsl-emb.c 	delta = (val - prev) & 0xfffffffful;
prev              606 arch/powerpc/perf/core-fsl-emb.c 	s64 prev, delta, left;
prev              615 arch/powerpc/perf/core-fsl-emb.c 	prev = local64_read(&event->hw.prev_count);
prev              616 arch/powerpc/perf/core-fsl-emb.c 	delta = (val - prev) & 0xfffffffful;
prev             1359 arch/powerpc/perf/hv-24x7.c 	s64 prev;
prev             1361 arch/powerpc/perf/hv-24x7.c 	prev = local64_xchg(&event->hw.prev_count, now);
prev             1362 arch/powerpc/perf/hv-24x7.c 	local64_add(now - prev, &event->count);
prev              192 arch/powerpc/perf/hv-gpci.c 	s64 prev;
prev              194 arch/powerpc/perf/hv-gpci.c 	prev = local64_xchg(&event->hw.prev_count, now);
prev              195 arch/powerpc/perf/hv-gpci.c 	local64_add(now - prev, &event->count);
prev               72 arch/powerpc/platforms/85xx/smp.c 		u64 prev;
prev               78 arch/powerpc/platforms/85xx/smp.c 			prev = timebase;
prev               81 arch/powerpc/platforms/85xx/smp.c 		} while (prev != timebase);
prev              283 arch/powerpc/platforms/cell/spufs/sched.c 	list_for_each_entry(ctx, gang->aff_ref_ctx->aff_list.prev, aff_list) {
prev              374 arch/powerpc/platforms/cell/spufs/sched.c 		list_for_each_entry(spu, ref->aff_list.prev, aff_list) {
prev              350 arch/powerpc/platforms/cell/spufs/spufs.h extern int spu_save(struct spu_state *prev, struct spu *spu);
prev              352 arch/powerpc/platforms/cell/spufs/spufs.h extern int spu_switch(struct spu_state *prev, struct spu_state *new,
prev             1777 arch/powerpc/platforms/cell/spufs/switch.c static int quiece_spu(struct spu_state *prev, struct spu *spu)
prev             1789 arch/powerpc/platforms/cell/spufs/switch.c 	if (check_spu_isolate(prev, spu)) {	/* Step 2. */
prev             1792 arch/powerpc/platforms/cell/spufs/switch.c 	disable_interrupts(prev, spu);	        /* Step 3. */
prev             1793 arch/powerpc/platforms/cell/spufs/switch.c 	set_watchdog_timer(prev, spu);	        /* Step 4. */
prev             1794 arch/powerpc/platforms/cell/spufs/switch.c 	inhibit_user_access(prev, spu);	        /* Step 5. */
prev             1795 arch/powerpc/platforms/cell/spufs/switch.c 	if (check_spu_isolate(prev, spu)) {	/* Step 6. */
prev             1798 arch/powerpc/platforms/cell/spufs/switch.c 	set_switch_pending(prev, spu);	        /* Step 7. */
prev             1799 arch/powerpc/platforms/cell/spufs/switch.c 	save_mfc_cntl(prev, spu);		/* Step 8. */
prev             1800 arch/powerpc/platforms/cell/spufs/switch.c 	save_spu_runcntl(prev, spu);	        /* Step 9. */
prev             1801 arch/powerpc/platforms/cell/spufs/switch.c 	save_mfc_sr1(prev, spu);	        /* Step 10. */
prev             1802 arch/powerpc/platforms/cell/spufs/switch.c 	save_spu_status(prev, spu);	        /* Step 11. */
prev             1803 arch/powerpc/platforms/cell/spufs/switch.c 	save_mfc_stopped_status(prev, spu);     /* Step 12. */
prev             1804 arch/powerpc/platforms/cell/spufs/switch.c 	halt_mfc_decr(prev, spu);	        /* Step 13. */
prev             1805 arch/powerpc/platforms/cell/spufs/switch.c 	save_timebase(prev, spu);		/* Step 14. */
prev             1806 arch/powerpc/platforms/cell/spufs/switch.c 	remove_other_spu_access(prev, spu);	/* Step 15. */
prev             1807 arch/powerpc/platforms/cell/spufs/switch.c 	do_mfc_mssync(prev, spu);	        /* Step 16. */
prev             1808 arch/powerpc/platforms/cell/spufs/switch.c 	issue_mfc_tlbie(prev, spu);	        /* Step 17. */
prev             1809 arch/powerpc/platforms/cell/spufs/switch.c 	handle_pending_interrupts(prev, spu);	/* Step 18. */
prev             1814 arch/powerpc/platforms/cell/spufs/switch.c static void save_csa(struct spu_state *prev, struct spu *spu)
prev             1821 arch/powerpc/platforms/cell/spufs/switch.c 	save_mfc_queues(prev, spu);	/* Step 19. */
prev             1822 arch/powerpc/platforms/cell/spufs/switch.c 	save_ppu_querymask(prev, spu);	/* Step 20. */
prev             1823 arch/powerpc/platforms/cell/spufs/switch.c 	save_ppu_querytype(prev, spu);	/* Step 21. */
prev             1824 arch/powerpc/platforms/cell/spufs/switch.c 	save_ppu_tagstatus(prev, spu);  /* NEW.     */
prev             1825 arch/powerpc/platforms/cell/spufs/switch.c 	save_mfc_csr_tsq(prev, spu);	/* Step 22. */
prev             1826 arch/powerpc/platforms/cell/spufs/switch.c 	save_mfc_csr_cmd(prev, spu);	/* Step 23. */
prev             1827 arch/powerpc/platforms/cell/spufs/switch.c 	save_mfc_csr_ato(prev, spu);	/* Step 24. */
prev             1828 arch/powerpc/platforms/cell/spufs/switch.c 	save_mfc_tclass_id(prev, spu);	/* Step 25. */
prev             1829 arch/powerpc/platforms/cell/spufs/switch.c 	set_mfc_tclass_id(prev, spu);	/* Step 26. */
prev             1830 arch/powerpc/platforms/cell/spufs/switch.c 	save_mfc_cmd(prev, spu);	/* Step 26a - moved from 44. */
prev             1831 arch/powerpc/platforms/cell/spufs/switch.c 	purge_mfc_queue(prev, spu);	/* Step 27. */
prev             1832 arch/powerpc/platforms/cell/spufs/switch.c 	wait_purge_complete(prev, spu);	/* Step 28. */
prev             1833 arch/powerpc/platforms/cell/spufs/switch.c 	setup_mfc_sr1(prev, spu);	/* Step 30. */
prev             1834 arch/powerpc/platforms/cell/spufs/switch.c 	save_spu_npc(prev, spu);	/* Step 31. */
prev             1835 arch/powerpc/platforms/cell/spufs/switch.c 	save_spu_privcntl(prev, spu);	/* Step 32. */
prev             1836 arch/powerpc/platforms/cell/spufs/switch.c 	reset_spu_privcntl(prev, spu);	/* Step 33. */
prev             1837 arch/powerpc/platforms/cell/spufs/switch.c 	save_spu_lslr(prev, spu);	/* Step 34. */
prev             1838 arch/powerpc/platforms/cell/spufs/switch.c 	reset_spu_lslr(prev, spu);	/* Step 35. */
prev             1839 arch/powerpc/platforms/cell/spufs/switch.c 	save_spu_cfg(prev, spu);	/* Step 36. */
prev             1840 arch/powerpc/platforms/cell/spufs/switch.c 	save_pm_trace(prev, spu);	/* Step 37. */
prev             1841 arch/powerpc/platforms/cell/spufs/switch.c 	save_mfc_rag(prev, spu);	/* Step 38. */
prev             1842 arch/powerpc/platforms/cell/spufs/switch.c 	save_ppu_mb_stat(prev, spu);	/* Step 39. */
prev             1843 arch/powerpc/platforms/cell/spufs/switch.c 	save_ppu_mb(prev, spu);	        /* Step 40. */
prev             1844 arch/powerpc/platforms/cell/spufs/switch.c 	save_ppuint_mb(prev, spu);	/* Step 41. */
prev             1845 arch/powerpc/platforms/cell/spufs/switch.c 	save_ch_part1(prev, spu);	/* Step 42. */
prev             1846 arch/powerpc/platforms/cell/spufs/switch.c 	save_spu_mb(prev, spu);	        /* Step 43. */
prev             1847 arch/powerpc/platforms/cell/spufs/switch.c 	reset_ch(prev, spu);	        /* Step 45. */
prev             1850 arch/powerpc/platforms/cell/spufs/switch.c static void save_lscsa(struct spu_state *prev, struct spu *spu)
prev             1858 arch/powerpc/platforms/cell/spufs/switch.c 	resume_mfc_queue(prev, spu);	/* Step 46. */
prev             1860 arch/powerpc/platforms/cell/spufs/switch.c 	setup_mfc_slbs(prev, spu, spu_save_code, sizeof(spu_save_code));
prev             1861 arch/powerpc/platforms/cell/spufs/switch.c 	set_switch_active(prev, spu);	/* Step 48. */
prev             1862 arch/powerpc/platforms/cell/spufs/switch.c 	enable_interrupts(prev, spu);	/* Step 49. */
prev             1863 arch/powerpc/platforms/cell/spufs/switch.c 	save_ls_16kb(prev, spu);	/* Step 50. */
prev             1864 arch/powerpc/platforms/cell/spufs/switch.c 	set_spu_npc(prev, spu);	        /* Step 51. */
prev             1865 arch/powerpc/platforms/cell/spufs/switch.c 	set_signot1(prev, spu);		/* Step 52. */
prev             1866 arch/powerpc/platforms/cell/spufs/switch.c 	set_signot2(prev, spu);		/* Step 53. */
prev             1867 arch/powerpc/platforms/cell/spufs/switch.c 	send_save_code(prev, spu);	/* Step 54. */
prev             1868 arch/powerpc/platforms/cell/spufs/switch.c 	set_ppu_querymask(prev, spu);	/* Step 55. */
prev             1869 arch/powerpc/platforms/cell/spufs/switch.c 	wait_tag_complete(prev, spu);	/* Step 56. */
prev             1870 arch/powerpc/platforms/cell/spufs/switch.c 	wait_spu_stopped(prev, spu);	/* Step 57. */
prev             1918 arch/powerpc/platforms/cell/spufs/switch.c static void harvest(struct spu_state *prev, struct spu *spu)
prev             1926 arch/powerpc/platforms/cell/spufs/switch.c 	disable_interrupts(prev, spu);	        /* Step 2.  */
prev             1927 arch/powerpc/platforms/cell/spufs/switch.c 	inhibit_user_access(prev, spu);	        /* Step 3.  */
prev             1928 arch/powerpc/platforms/cell/spufs/switch.c 	terminate_spu_app(prev, spu);	        /* Step 4.  */
prev             1929 arch/powerpc/platforms/cell/spufs/switch.c 	set_switch_pending(prev, spu);	        /* Step 5.  */
prev             1931 arch/powerpc/platforms/cell/spufs/switch.c 	remove_other_spu_access(prev, spu);	/* Step 6.  */
prev             1932 arch/powerpc/platforms/cell/spufs/switch.c 	suspend_mfc_and_halt_decr(prev, spu);	/* Step 7.  */
prev             1933 arch/powerpc/platforms/cell/spufs/switch.c 	wait_suspend_mfc_complete(prev, spu);	/* Step 8.  */
prev             1934 arch/powerpc/platforms/cell/spufs/switch.c 	if (!suspend_spe(prev, spu))	        /* Step 9.  */
prev             1935 arch/powerpc/platforms/cell/spufs/switch.c 		clear_spu_status(prev, spu);	/* Step 10. */
prev             1936 arch/powerpc/platforms/cell/spufs/switch.c 	do_mfc_mssync(prev, spu);	        /* Step 11. */
prev             1937 arch/powerpc/platforms/cell/spufs/switch.c 	issue_mfc_tlbie(prev, spu);	        /* Step 12. */
prev             1938 arch/powerpc/platforms/cell/spufs/switch.c 	handle_pending_interrupts(prev, spu);	/* Step 13. */
prev             1939 arch/powerpc/platforms/cell/spufs/switch.c 	purge_mfc_queue(prev, spu);	        /* Step 14. */
prev             1940 arch/powerpc/platforms/cell/spufs/switch.c 	wait_purge_complete(prev, spu);	        /* Step 15. */
prev             1941 arch/powerpc/platforms/cell/spufs/switch.c 	reset_spu_privcntl(prev, spu);	        /* Step 16. */
prev             1942 arch/powerpc/platforms/cell/spufs/switch.c 	reset_spu_lslr(prev, spu);              /* Step 17. */
prev             1943 arch/powerpc/platforms/cell/spufs/switch.c 	setup_mfc_sr1(prev, spu);	        /* Step 18. */
prev             1945 arch/powerpc/platforms/cell/spufs/switch.c 	reset_ch_part1(prev, spu);	        /* Step 20. */
prev             1946 arch/powerpc/platforms/cell/spufs/switch.c 	reset_ch_part2(prev, spu);	        /* Step 21. */
prev             1947 arch/powerpc/platforms/cell/spufs/switch.c 	enable_interrupts(prev, spu);	        /* Step 22. */
prev             1948 arch/powerpc/platforms/cell/spufs/switch.c 	set_switch_active(prev, spu);	        /* Step 23. */
prev             1949 arch/powerpc/platforms/cell/spufs/switch.c 	set_mfc_tclass_id(prev, spu);	        /* Step 24. */
prev             1950 arch/powerpc/platforms/cell/spufs/switch.c 	resume_mfc_queue(prev, spu);	        /* Step 25. */
prev             2024 arch/powerpc/platforms/cell/spufs/switch.c static int __do_spu_save(struct spu_state *prev, struct spu *spu)
prev             2040 arch/powerpc/platforms/cell/spufs/switch.c 	rc = quiece_spu(prev, spu);	        /* Steps 2-16. */
prev             2045 arch/powerpc/platforms/cell/spufs/switch.c 		harvest(prev, spu);
prev             2051 arch/powerpc/platforms/cell/spufs/switch.c 	save_csa(prev, spu);	                /* Steps 17-43. */
prev             2052 arch/powerpc/platforms/cell/spufs/switch.c 	save_lscsa(prev, spu);	                /* Steps 44-53. */
prev             2053 arch/powerpc/platforms/cell/spufs/switch.c 	return check_save_status(prev, spu);	/* Step 54.     */
prev             2094 arch/powerpc/platforms/cell/spufs/switch.c int spu_save(struct spu_state *prev, struct spu *spu)
prev             2099 arch/powerpc/platforms/cell/spufs/switch.c 	rc = __do_spu_save(prev, spu);	/* Steps 2-53. */
prev              958 arch/powerpc/platforms/powermac/low_i2c.c 	struct device_node *prev = NULL;
prev              964 arch/powerpc/platforms/powermac/low_i2c.c 				if (prev && bus->flags & pmac_i2c_multibus) {
prev              966 arch/powerpc/platforms/powermac/low_i2c.c 					reg = of_get_property(prev, "reg",
prev              974 arch/powerpc/platforms/powermac/low_i2c.c 				of_node_put(prev);
prev              978 arch/powerpc/platforms/powermac/low_i2c.c 		of_node_put(prev);
prev              979 arch/powerpc/platforms/powermac/low_i2c.c 		prev = p;
prev              329 arch/powerpc/sysdev/fsl_msi.c 	if (msi->list.prev != NULL)
prev              417 arch/powerpc/xmon/xmon.c 	int last_speaker = 0, prev;
prev              440 arch/powerpc/xmon/xmon.c 			prev = cmpxchg(&xmon_speaker, last_speaker, me);
prev              441 arch/powerpc/xmon/xmon.c 			if (prev == last_speaker)
prev              204 arch/riscv/include/asm/atomic.h        int prev, rc;
prev              214 arch/riscv/include/asm/atomic.h 		: [p]"=&r" (prev), [rc]"=&r" (rc), [c]"+A" (v->counter)
prev              217 arch/riscv/include/asm/atomic.h 	return prev;
prev              224 arch/riscv/include/asm/atomic.h        s64 prev;
prev              235 arch/riscv/include/asm/atomic.h 		: [p]"=&r" (prev), [rc]"=&r" (rc), [c]"+A" (v->counter)
prev              238 arch/riscv/include/asm/atomic.h 	return prev;
prev              317 arch/riscv/include/asm/atomic.h        int prev, rc;
prev              327 arch/riscv/include/asm/atomic.h 		: [p]"=&r" (prev), [rc]"=&r" (rc), [c]"+A" (v->counter)
prev              330 arch/riscv/include/asm/atomic.h 	return prev - offset;
prev              338 arch/riscv/include/asm/atomic.h        s64 prev;
prev              349 arch/riscv/include/asm/atomic.h 		: [p]"=&r" (prev), [rc]"=&r" (rc), [c]"+A" (v->counter)
prev              352 arch/riscv/include/asm/atomic.h 	return prev - offset;
prev               32 arch/riscv/include/asm/mmu_context.h void switch_mm(struct mm_struct *prev, struct mm_struct *next,
prev               35 arch/riscv/include/asm/mmu_context.h static inline void activate_mm(struct mm_struct *prev,
prev               38 arch/riscv/include/asm/mmu_context.h 	switch_mm(prev, next, NULL);
prev               47 arch/riscv/include/asm/switch_to.h static inline void __switch_to_aux(struct task_struct *prev,
prev               52 arch/riscv/include/asm/switch_to.h 	regs = task_pt_regs(prev);
prev               54 arch/riscv/include/asm/switch_to.h 		fstate_save(prev, regs);
prev               69 arch/riscv/include/asm/switch_to.h #define switch_to(prev, next, last)			\
prev               71 arch/riscv/include/asm/switch_to.h 	struct task_struct *__prev = (prev);		\
prev               23 arch/riscv/kernel/cacheinfo.c 	struct device_node *prev = NULL;
prev               35 arch/riscv/kernel/cacheinfo.c 	prev = np;
prev               37 arch/riscv/kernel/cacheinfo.c 		of_node_put(prev);
prev               38 arch/riscv/kernel/cacheinfo.c 		prev = np;
prev               66 arch/riscv/kernel/cacheinfo.c 	struct device_node *prev = NULL;
prev               76 arch/riscv/kernel/cacheinfo.c 	prev = np;
prev               78 arch/riscv/kernel/cacheinfo.c 		of_node_put(prev);
prev               79 arch/riscv/kernel/cacheinfo.c 		prev = np;
prev               43 arch/riscv/mm/context.c void switch_mm(struct mm_struct *prev, struct mm_struct *next,
prev               48 arch/riscv/mm/context.c 	if (unlikely(prev == next))
prev               58 arch/riscv/mm/context.c 	cpumask_clear_cpu(cpu, mm_cpumask(prev));
prev              102 arch/s390/crypto/prng.c 	u8 *prev;
prev              428 arch/s390/crypto/prng.c 		prng_data->prev = prng_data->buf + prng_chunk_size;
prev              431 arch/s390/crypto/prng.c 			   prng_data->prev, prng_chunk_size, NULL, 0);
prev              493 arch/s390/crypto/prng.c 		if (!memcmp(prng_data->prev, buf, nbytes)) {
prev              497 arch/s390/crypto/prng.c 		memcpy(prng_data->prev, buf, nbytes);
prev               35 arch/s390/include/asm/debug.h 	struct debug_info *prev;
prev               88 arch/s390/include/asm/mmu_context.h static inline void switch_mm(struct mm_struct *prev, struct mm_struct *next,
prev              104 arch/s390/include/asm/mmu_context.h 	if (prev != next)
prev              105 arch/s390/include/asm/mmu_context.h 		cpumask_clear_cpu(cpu, &prev->context.cpu_attach_mask);
prev              128 arch/s390/include/asm/mmu_context.h static inline void activate_mm(struct mm_struct *prev,
prev              131 arch/s390/include/asm/mmu_context.h 	switch_mm(prev, next, current);
prev               99 arch/s390/include/asm/stacktrace.h 	unsigned long prev;						\
prev              107 arch/s390/include/asm/stacktrace.h 		: [_prev] "=&a" (prev), CALL_FMT_##nr			\
prev               33 arch/s390/include/asm/switch_to.h #define switch_to(prev, next, last) do {				\
prev               39 arch/s390/include/asm/switch_to.h 	save_access_regs(&prev->thread.acrs[0]);			\
prev               40 arch/s390/include/asm/switch_to.h 	save_ri_cb(prev->thread.ri_cb);					\
prev               41 arch/s390/include/asm/switch_to.h 	save_gs_cb(prev->thread.gs_cb);					\
prev               44 arch/s390/include/asm/switch_to.h 	restore_ri_cb(next->thread.ri_cb, prev->thread.ri_cb);		\
prev               46 arch/s390/include/asm/switch_to.h 	prev = __switch_to(prev, next);					\
prev              338 arch/s390/kernel/debug.c 		rc->prev = NULL;
prev              342 arch/s390/kernel/debug.c 		rc->prev = debug_area_last;
prev              420 arch/s390/kernel/debug.c 			debug_area_last = db_info->prev;
prev              421 arch/s390/kernel/debug.c 		if (db_info->prev)
prev              422 arch/s390/kernel/debug.c 			db_info->prev->next = db_info->next;
prev              424 arch/s390/kernel/debug.c 			db_info->next->prev = db_info->prev;
prev              316 arch/s390/kernel/perf_cpum_cf.c 	u64 prev, new;
prev              320 arch/s390/kernel/perf_cpum_cf.c 		prev = local64_read(&event->hw.prev_count);
prev              332 arch/s390/kernel/perf_cpum_cf.c 	} while (local64_cmpxchg(&event->hw.prev_count, prev, new) != prev);
prev              339 arch/s390/kernel/perf_cpum_cf.c 	u64 prev, new, delta;
prev              343 arch/s390/kernel/perf_cpum_cf.c 		prev = local64_read(&event->hw.prev_count);
prev              347 arch/s390/kernel/perf_cpum_cf.c 	} while (local64_cmpxchg(&event->hw.prev_count, prev, new) != prev);
prev              349 arch/s390/kernel/perf_cpum_cf.c 	delta = (prev <= new) ? new - prev
prev              350 arch/s390/kernel/perf_cpum_cf.c 			      : (-1ULL - prev) + new + 1;	 /* overflow */
prev              188 arch/s390/kernel/vtime.c void vtime_task_switch(struct task_struct *prev)
prev              190 arch/s390/kernel/vtime.c 	do_account_vtime(prev);
prev              191 arch/s390/kernel/vtime.c 	prev->thread.user_timer = S390_lowcore.user_timer;
prev              192 arch/s390/kernel/vtime.c 	prev->thread.guest_timer = S390_lowcore.guest_timer;
prev              193 arch/s390/kernel/vtime.c 	prev->thread.system_timer = S390_lowcore.system_timer;
prev              194 arch/s390/kernel/vtime.c 	prev->thread.hardirq_timer = S390_lowcore.hardirq_timer;
prev              195 arch/s390/kernel/vtime.c 	prev->thread.softirq_timer = S390_lowcore.softirq_timer;
prev               40 arch/s390/lib/spinlock.c 	struct spin_wait *next, *prev;
prev              110 arch/s390/lib/spinlock.c 	if (node == NULL || node->prev == NULL)
prev              112 arch/s390/lib/spinlock.c 	while (node->prev)
prev              113 arch/s390/lib/spinlock.c 		node = node->prev;
prev              126 arch/s390/lib/spinlock.c 	node->prev = node->next = NULL;
prev              156 arch/s390/lib/spinlock.c 		node->prev = arch_spin_decode_tail(tail_id);
prev              157 arch/s390/lib/spinlock.c 		WRITE_ONCE(node->prev->next, node);
prev              168 arch/s390/lib/spinlock.c 		while (READ_ONCE(node->prev) != NULL) {
prev              204 arch/s390/lib/spinlock.c 		next->prev = NULL;
prev               44 arch/s390/numa/toptree.h 	 (ptree->parent->children.prev == &ptree->sibling))
prev              273 arch/sh/include/asm/dwarf.h 	struct dwarf_frame *prev, *next;
prev               11 arch/sh/include/asm/futex-irq.h 	u32 prev = 0;
prev               15 arch/sh/include/asm/futex-irq.h 	ret = get_user(prev, uaddr);
prev               16 arch/sh/include/asm/futex-irq.h 	if (!ret && oldval == prev)
prev               21 arch/sh/include/asm/futex-irq.h 	*uval = prev;
prev               34 arch/sh/include/asm/futex.h 	u32 oldval, newval, prev;
prev               67 arch/sh/include/asm/futex.h 		ret = futex_atomic_cmpxchg_inatomic(&prev, uaddr, oldval, newval);
prev               68 arch/sh/include/asm/futex.h 	} while (!ret && prev != oldval);
prev              121 arch/sh/include/asm/mmu_context.h static inline void switch_mm(struct mm_struct *prev,
prev              127 arch/sh/include/asm/mmu_context.h 	if (likely(prev != next)) {
prev              136 arch/sh/include/asm/mmu_context.h #define activate_mm(prev, next)		switch_mm((prev),(next),NULL)
prev               64 arch/sh/include/asm/switch_to_32.h struct task_struct *__switch_to(struct task_struct *prev,
prev               70 arch/sh/include/asm/switch_to_32.h #define switch_to(prev, next, last)				\
prev               80 arch/sh/include/asm/switch_to_32.h 	if (is_dsp_enabled(prev))				\
prev               81 arch/sh/include/asm/switch_to_32.h 		__save_dsp(prev);				\
prev               85 arch/sh/include/asm/switch_to_32.h 	__ts1 = (u32 *)&prev->thread.sp;			\
prev               86 arch/sh/include/asm/switch_to_32.h 	__ts2 = (u32 *)&prev->thread.pc;			\
prev               87 arch/sh/include/asm/switch_to_32.h 	__ts4 = (u32 *)prev;					\
prev               16 arch/sh/include/asm/switch_to_64.h struct task_struct *sh64_switch_to(struct task_struct *prev,
prev               21 arch/sh/include/asm/switch_to_64.h #define switch_to(prev,next,last)				\
prev               27 arch/sh/include/asm/switch_to_64.h 	last = sh64_switch_to(prev, &prev->thread, next,	\
prev              575 arch/sh/kernel/dwarf.c 				       struct dwarf_frame *prev)
prev              598 arch/sh/kernel/dwarf.c 	if (!pc || !prev)
prev              631 arch/sh/kernel/dwarf.c 	frame->prev = prev;
prev              670 arch/sh/kernel/dwarf.c 		if (prev) {
prev              671 arch/sh/kernel/dwarf.c 			reg = dwarf_frame_reg(prev, frame->cfa_register);
prev              675 arch/sh/kernel/dwarf.c 			addr = prev->cfa + reg->addr;
prev              725 arch/sh/kernel/dwarf.c 	if (prev && prev->pc == (unsigned long)ret_from_irq)
prev              173 arch/sh/kernel/process_32.c __switch_to(struct task_struct *prev, struct task_struct *next)
prev              181 arch/sh/kernel/process_32.c 	unlazy_fpu(prev, task_pt_regs(prev));
prev              205 arch/sh/kernel/process_32.c 	return prev;
prev               35 arch/sh/kernel/unwinder.c 		.prev = &unwinder_list,
prev               53 arch/sh/kernel/unwinder.c 	.prev = &stack_reader.list,
prev              143 arch/sparc/include/asm/mmu_context_64.h static inline void arch_start_context_switch(struct task_struct *prev)
prev              158 arch/sparc/include/asm/mmu_context_64.h 			set_tsk_thread_flag(prev, TIF_MCDPER);
prev              160 arch/sparc/include/asm/mmu_context_64.h 			clear_tsk_thread_flag(prev, TIF_MCDPER);
prev               58 arch/sparc/include/asm/switch_to_32.h #define switch_to(prev, next, last) do {						\
prev               59 arch/sparc/include/asm/switch_to_32.h 	SWITCH_ENTER(prev);								\
prev               21 arch/sparc/include/asm/switch_to_64.h #define switch_to(prev, next, last)					\
prev              234 arch/sparc/kernel/leon_smp.c 	int *prev;
prev              238 arch/sparc/kernel/leon_smp.c 	prev = &first;
prev              241 arch/sparc/kernel/leon_smp.c 			*prev = i;
prev              242 arch/sparc/kernel/leon_smp.c 			prev = &cpu_data(i).next;
prev              245 arch/sparc/kernel/leon_smp.c 	*prev = first;
prev              132 arch/sparc/kernel/prom_common.c static int __init prom_common_nextprop(phandle node, char *prev, char *buf)
prev              137 arch/sparc/kernel/prom_common.c 	name = prom_nextprop(node, prev, buf);
prev              166 arch/sparc/kernel/sun4d_smp.c 	int *prev;
prev              170 arch/sparc/kernel/sun4d_smp.c 	prev = &first;
prev              172 arch/sparc/kernel/sun4d_smp.c 		*prev = i;
prev              173 arch/sparc/kernel/sun4d_smp.c 		prev = &cpu_data(i).next;
prev              175 arch/sparc/kernel/sun4d_smp.c 	*prev = first;
prev              124 arch/sparc/kernel/sun4m_smp.c 	int *prev;
prev              128 arch/sparc/kernel/sun4m_smp.c 	prev = &first;
prev              130 arch/sparc/kernel/sun4m_smp.c 		*prev = i;
prev              131 arch/sparc/kernel/sun4m_smp.c 		prev = &cpu_data(i).next;
prev              133 arch/sparc/kernel/sun4m_smp.c 	*prev = first;
prev              165 arch/sparc/lib/atomic32.c 	u32 prev;
prev              168 arch/sparc/lib/atomic32.c 	if ((prev = *ptr) == old)
prev              172 arch/sparc/lib/atomic32.c 	return (unsigned long)prev;
prev              179 arch/sparc/lib/atomic32.c 	u64 prev;
prev              182 arch/sparc/lib/atomic32.c 	if ((prev = *ptr) == old)
prev              186 arch/sparc/lib/atomic32.c 	return prev;
prev              193 arch/sparc/lib/atomic32.c 	u32 prev;
prev              196 arch/sparc/lib/atomic32.c 	prev = *ptr;
prev              200 arch/sparc/lib/atomic32.c 	return (unsigned long)prev;
prev              407 arch/sparc/mm/srmmu.c 	struct ctx_list *prev;
prev              421 arch/sparc/mm/srmmu.c 	entry->next->prev = entry->prev;
prev              422 arch/sparc/mm/srmmu.c 	entry->prev->next = entry->next;
prev              428 arch/sparc/mm/srmmu.c 	(entry->prev = head->prev)->next = entry;
prev              429 arch/sparc/mm/srmmu.c 	head->prev = entry;
prev              487 arch/sparc/mm/srmmu.c 	ctx_free.next = ctx_free.prev = &ctx_free;
prev              488 arch/sparc/mm/srmmu.c 	ctx_used.next = ctx_used.prev = &ctx_used;
prev               60 arch/um/include/asm/mmu_context.h static inline void switch_mm(struct mm_struct *prev, struct mm_struct *next, 
prev               65 arch/um/include/asm/mmu_context.h 	if(prev != next){
prev               66 arch/um/include/asm/mmu_context.h 		cpumask_clear_cpu(cpu, mm_cpumask(prev));
prev               46 arch/unicore32/include/asm/mmu_context.h switch_mm(struct mm_struct *prev, struct mm_struct *next,
prev               51 arch/unicore32/include/asm/mmu_context.h 	if (!cpumask_test_and_set_cpu(cpu, mm_cpumask(next)) || prev != next)
prev               56 arch/unicore32/include/asm/mmu_context.h #define activate_mm(prev, next)	switch_mm(prev, next, NULL)
prev               21 arch/unicore32/include/asm/switch_to.h #define switch_to(prev, next, last)					\
prev               23 arch/unicore32/include/asm/switch_to.h 		last = __switch_to(prev, task_thread_info(prev),	\
prev              516 arch/x86/boot/compressed/eboot.c 	struct boot_e820_entry *prev = NULL;
prev              573 arch/x86/boot/compressed/eboot.c 		if (prev && prev->type == e820_type &&
prev              574 arch/x86/boot/compressed/eboot.c 		    (prev->addr + prev->size) == d->phys_addr) {
prev              575 arch/x86/boot/compressed/eboot.c 			prev->size += d->num_pages << 12;
prev              593 arch/x86/boot/compressed/eboot.c 		prev = entry++;
prev              303 arch/x86/events/amd/iommu.c 	u64 count, prev, delta;
prev              314 arch/x86/events/amd/iommu.c 	prev = local64_read(&hwc->prev_count);
prev              315 arch/x86/events/amd/iommu.c 	if (local64_cmpxchg(&hwc->prev_count, prev, count) != prev)
prev              319 arch/x86/events/amd/iommu.c 	delta = (count << COUNTER_SHIFT) - (prev << COUNTER_SHIFT);
prev               85 arch/x86/events/amd/uncore.c 	u64 prev, new;
prev               93 arch/x86/events/amd/uncore.c 	prev = local64_read(&hwc->prev_count);
prev               96 arch/x86/events/amd/uncore.c 	delta = (new << COUNTER_SHIFT) - (prev << COUNTER_SHIFT);
prev              229 arch/x86/events/msr.c 	u64 prev, now;
prev              234 arch/x86/events/msr.c 	prev = local64_read(&event->hw.prev_count);
prev              237 arch/x86/events/msr.c 	if (local64_cmpxchg(&event->hw.prev_count, prev, now) != prev)
prev              240 arch/x86/events/msr.c 	delta = now - prev;
prev               28 arch/x86/include/asm/cmpxchg_32.h 	u64 prev = *ptr;
prev               33 arch/x86/include/asm/cmpxchg_32.h 		     : "=m" (*ptr), "+A" (prev)
prev               49 arch/x86/include/asm/cmpxchg_32.h 	u64 prev;
prev               51 arch/x86/include/asm/cmpxchg_32.h 		     : "=A" (prev),
prev               57 arch/x86/include/asm/cmpxchg_32.h 	return prev;
prev               62 arch/x86/include/asm/cmpxchg_32.h 	u64 prev;
prev               64 arch/x86/include/asm/cmpxchg_32.h 		     : "=A" (prev),
prev               70 arch/x86/include/asm/cmpxchg_32.h 	return prev;
prev               21 arch/x86/include/asm/mmu_context.h static inline void paravirt_activate_mm(struct mm_struct *prev,
prev              152 arch/x86/include/asm/mmu_context.h static inline void switch_ldt(struct mm_struct *prev, struct mm_struct *next)
prev              172 arch/x86/include/asm/mmu_context.h 	if (unlikely((unsigned long)prev->context.ldt |
prev              210 arch/x86/include/asm/mmu_context.h extern void switch_mm(struct mm_struct *prev, struct mm_struct *next,
prev              213 arch/x86/include/asm/mmu_context.h extern void switch_mm_irqs_off(struct mm_struct *prev, struct mm_struct *next,
prev              217 arch/x86/include/asm/mmu_context.h #define activate_mm(prev, next)			\
prev              219 arch/x86/include/asm/mmu_context.h 	paravirt_activate_mm((prev), (next));	\
prev              220 arch/x86/include/asm/mmu_context.h 	switch_mm((prev), (next), NULL);	\
prev              302 arch/x86/include/asm/paravirt.h static inline void paravirt_activate_mm(struct mm_struct *prev,
prev              305 arch/x86/include/asm/paravirt.h 	PVOP_VCALL2(mmu.activate_mm, prev, next);
prev              604 arch/x86/include/asm/paravirt.h static inline void arch_start_context_switch(struct task_struct *prev)
prev              606 arch/x86/include/asm/paravirt.h 	PVOP_VCALL1(cpu.start_context_switch, prev);
prev              178 arch/x86/include/asm/paravirt_types.h 	void (*start_context_switch)(struct task_struct *prev);
prev              225 arch/x86/include/asm/paravirt_types.h 	void (*activate_mm)(struct mm_struct *prev,
prev              656 arch/x86/include/asm/paravirt_types.h void paravirt_start_context_switch(struct task_struct *prev);
prev              116 arch/x86/include/asm/pgtable.h #define arch_end_context_switch(prev)	do {} while(0)
prev                9 arch/x86/include/asm/switch_to.h struct task_struct *__switch_to_asm(struct task_struct *prev,
prev               12 arch/x86/include/asm/switch_to.h __visible struct task_struct *__switch_to(struct task_struct *prev,
prev               68 arch/x86/include/asm/switch_to.h #define switch_to(prev, next, last)					\
prev               72 arch/x86/include/asm/switch_to.h 	((last) = __switch_to_asm((prev), (next)));			\
prev              792 arch/x86/kernel/alternative.c 	temp_mm_state_t prev;
prev              850 arch/x86/kernel/alternative.c 	prev = use_temporary_mm(poking_mm);
prev              871 arch/x86/kernel/alternative.c 	unuse_temporary_mm(prev);
prev              630 arch/x86/kernel/cpu/microcode/amd.c 		__list_del(p->plist.prev, p->plist.next);
prev              764 arch/x86/kernel/cpu/microcode/intel.c 	static struct cpu_signature prev;
prev              781 arch/x86/kernel/cpu/microcode/intel.c 	if (csig->sig != prev.sig || csig->pf != prev.pf || csig->rev != prev.rev) {
prev              784 arch/x86/kernel/cpu/microcode/intel.c 		prev = *csig;
prev               83 arch/x86/kernel/cpu/mtrr/generic.c static int check_type_overlap(u8 *prev, u8 *curr)
prev               85 arch/x86/kernel/cpu/mtrr/generic.c 	if (*prev == MTRR_TYPE_UNCACHABLE || *curr == MTRR_TYPE_UNCACHABLE) {
prev               86 arch/x86/kernel/cpu/mtrr/generic.c 		*prev = MTRR_TYPE_UNCACHABLE;
prev               91 arch/x86/kernel/cpu/mtrr/generic.c 	if ((*prev == MTRR_TYPE_WRBACK && *curr == MTRR_TYPE_WRTHROUGH) ||
prev               92 arch/x86/kernel/cpu/mtrr/generic.c 	    (*prev == MTRR_TYPE_WRTHROUGH && *curr == MTRR_TYPE_WRBACK)) {
prev               93 arch/x86/kernel/cpu/mtrr/generic.c 		*prev = MTRR_TYPE_WRTHROUGH;
prev               97 arch/x86/kernel/cpu/mtrr/generic.c 	if (*prev != *curr) {
prev               98 arch/x86/kernel/cpu/mtrr/generic.c 		*prev = MTRR_TYPE_UNCACHABLE;
prev              142 arch/x86/kernel/jump_label.c 		int prev = tp_vec_nr - 1;
prev              143 arch/x86/kernel/jump_label.c 		struct text_poke_loc *prev_tp = &tp_vec[prev];
prev              251 arch/x86/kernel/paravirt.c void paravirt_start_context_switch(struct task_struct *prev)
prev              257 arch/x86/kernel/paravirt.c 		set_ti_thread_flag(task_thread_info(prev), TIF_LAZY_MMU_UPDATES);
prev              272 arch/x86/kernel/process.c static inline void switch_to_bitmap(struct thread_struct *prev,
prev              284 arch/x86/kernel/process.c 		       max(prev->io_bitmap_max, next->io_bitmap_max));
prev              294 arch/x86/kernel/process.c 		memset(tss->io_bitmap, 0xff, prev->io_bitmap_max);
prev              492 arch/x86/kernel/process.c 	struct thread_struct *prev, *next;
prev              495 arch/x86/kernel/process.c 	prev = &prev_p->thread;
prev              500 arch/x86/kernel/process.c 	switch_to_bitmap(prev, next, tifp, tifn);
prev               13 arch/x86/kernel/process.h static inline void switch_to_extra(struct task_struct *prev,
prev               17 arch/x86/kernel/process.h 	unsigned long prev_tif = task_thread_info(prev)->flags;
prev               38 arch/x86/kernel/process.h 		__switch_to_xtra(prev, next);
prev              230 arch/x86/kernel/process_32.c 	struct thread_struct *prev = &prev_p->thread,
prev              232 arch/x86/kernel/process_32.c 	struct fpu *prev_fpu = &prev->fpu;
prev              251 arch/x86/kernel/process_32.c 	lazy_save_gs(prev->gs);
prev              264 arch/x86/kernel/process_32.c 	if (get_kernel_rpl() && unlikely(prev->iopl != next->iopl))
prev              290 arch/x86/kernel/process_32.c 	if (prev->gs | next->gs)
prev              280 arch/x86/kernel/process_64.c static __always_inline void x86_fsgsbase_load(struct thread_struct *prev,
prev              283 arch/x86/kernel/process_64.c 	load_seg_legacy(prev->fsindex, prev->fsbase,
prev              285 arch/x86/kernel/process_64.c 	load_seg_legacy(prev->gsindex, prev->gsbase,
prev              506 arch/x86/kernel/process_64.c 	struct thread_struct *prev = &prev_p->thread;
prev              508 arch/x86/kernel/process_64.c 	struct fpu *prev_fpu = &prev->fpu;
prev              552 arch/x86/kernel/process_64.c 	savesegment(es, prev->es);
prev              553 arch/x86/kernel/process_64.c 	if (unlikely(next->es | prev->es))
prev              556 arch/x86/kernel/process_64.c 	savesegment(ds, prev->ds);
prev              557 arch/x86/kernel/process_64.c 	if (unlikely(next->ds | prev->ds))
prev              560 arch/x86/kernel/process_64.c 	x86_fsgsbase_load(prev, next);
prev              582 arch/x86/kernel/process_64.c 		     prev->iopl != next->iopl))
prev              228 arch/x86/kernel/tsc_sync.c 	cycles_t start, now, prev, end, cur_max_warp = 0;
prev              245 arch/x86/kernel/tsc_sync.c 		prev = last_tsc;
prev              266 arch/x86/kernel/tsc_sync.c 		if (unlikely(prev > now)) {
prev              268 arch/x86/kernel/tsc_sync.c 			max_warp = max(max_warp, prev - now);
prev              275 arch/x86/kvm/vmx/nested.c 				     struct loaded_vmcs *prev)
prev              282 arch/x86/kvm/vmx/nested.c 	src = &prev->host_state;
prev              296 arch/x86/kvm/vmx/nested.c 	struct loaded_vmcs *prev;
prev              303 arch/x86/kvm/vmx/nested.c 	prev = vmx->loaded_vmcs;
prev              305 arch/x86/kvm/vmx/nested.c 	vmx_vcpu_load_vmcs(vcpu, cpu, prev);
prev              306 arch/x86/kvm/vmx/nested.c 	vmx_sync_vmcs_host_state(vmx, prev);
prev             1294 arch/x86/kvm/vmx/vmx.c 	struct vmcs *prev;
prev             1313 arch/x86/kvm/vmx/vmx.c 	prev = per_cpu(current_vmcs, cpu);
prev             1314 arch/x86/kvm/vmx/vmx.c 	if (prev != vmx->loaded_vmcs->vmcs) {
prev             1323 arch/x86/kvm/vmx/vmx.c 		if (!buddy || WARN_ON_ONCE(buddy->vmcs != prev))
prev              778 arch/x86/mm/mpx.c 	struct vm_area_struct *prev;
prev              794 arch/x86/mm/mpx.c 	next = find_vma_prev(mm, start, &prev);
prev              805 arch/x86/mm/mpx.c 	while (prev && (prev->vm_flags & VM_MPX))
prev              806 arch/x86/mm/mpx.c 		prev = prev->vm_prev;
prev              814 arch/x86/mm/mpx.c 	next = find_vma_prev(mm, start, &prev);
prev              815 arch/x86/mm/mpx.c 	if ((!prev || prev->vm_end <= bta_start_vaddr) &&
prev              154 arch/x86/mm/tlb.c void switch_mm(struct mm_struct *prev, struct mm_struct *next,
prev              160 arch/x86/mm/tlb.c 	switch_mm_irqs_off(prev, next, tsk);
prev              275 arch/x86/mm/tlb.c void switch_mm_irqs_off(struct mm_struct *prev, struct mm_struct *next,
prev               90 arch/x86/platform/olpc/olpc_dt.c static int __init olpc_dt_nextprop(phandle node, char *prev, char *buf)
prev               92 arch/x86/platform/olpc/olpc_dt.c 	const void *args[] = { (void *)node, prev, buf };
prev              977 arch/x86/xen/mmu_pv.c static void xen_activate_mm(struct mm_struct *prev, struct mm_struct *next)
prev              125 arch/xtensa/include/asm/mmu_context.h static inline void switch_mm(struct mm_struct *prev, struct mm_struct *next,
prev              135 arch/xtensa/include/asm/mmu_context.h 	if (migrated || prev != next)
prev              139 arch/xtensa/include/asm/mmu_context.h #define activate_mm(prev, next)	switch_mm((prev), (next), NULL)
prev               23 arch/xtensa/include/asm/nommu_context.h static inline void activate_mm(struct mm_struct *prev, struct mm_struct *next)
prev               27 arch/xtensa/include/asm/nommu_context.h static inline void switch_mm(struct mm_struct *prev, struct mm_struct *next,
prev               17 arch/xtensa/include/asm/switch_to.h #define switch_to(prev,next,last)		\
prev               19 arch/xtensa/include/asm/switch_to.h 	(last) = _switch_to(prev, next);	\
prev              896 block/bfq-iosched.c 	struct request *next, *prev = NULL;
prev              904 block/bfq-iosched.c 		prev = rb_entry_rq(rbprev);
prev              914 block/bfq-iosched.c 	return bfq_choose_req(bfqd, next, prev, blk_rq_pos(last));
prev             1868 block/bfq-iosched.c 	struct request *next_rq, *prev;
prev             2055 block/bfq-iosched.c 	prev = bfqq->next_rq;
prev             2063 block/bfq-iosched.c 	if (unlikely(!bfqd->nonrot_with_queueing && prev != bfqq->next_rq))
prev             2080 block/bfq-iosched.c 		if (prev != bfqq->next_rq)
prev             2164 block/bfq-iosched.c 	if (rq->queuelist.prev != &rq->queuelist)
prev             2273 block/bfq-iosched.c 		struct request *prev, *next_rq;
prev             2285 block/bfq-iosched.c 		prev = bfqq->next_rq;
prev             2294 block/bfq-iosched.c 		if (prev != bfqq->next_rq) {
prev               33 block/blk-integrity.c 	int prev = 0;
prev               37 block/blk-integrity.c 		if (prev) {
prev               50 block/blk-integrity.c 		prev = 1;
prev               75 block/blk-integrity.c 	int prev = 0;
prev               79 block/blk-integrity.c 		if (prev) {
prev               99 block/blk-integrity.c 		prev = 1;
prev               16 block/blk-merge.c 		struct request *prev_rq, struct bio *prev, struct bio *next)
prev               20 block/blk-merge.c 	if (!bio_has_data(prev) || !queue_virt_boundary(q))
prev               31 block/blk-merge.c 		bio_get_first_bvec(prev, &pb);
prev               44 block/blk-merge.c 	bio_get_last_bvec(prev, &pb);
prev              826 block/blk-merge.c 	struct request *prev = elv_former_request(q, rq);
prev              828 block/blk-merge.c 	if (prev)
prev              829 block/blk-merge.c 		return attempt_merge(q, prev, rq);
prev             2016 block/blk-mq.c 			last = list_entry_rq(plug->mq_list.prev);
prev              493 crypto/af_alg.c 	sgl = list_entry(ctx->tsgl_list.prev, struct af_alg_tsgl, list);
prev              870 crypto/af_alg.c 			sgl = list_entry(ctx->tsgl_list.prev,
prev              905 crypto/af_alg.c 		sgl = list_entry(ctx->tsgl_list.prev, struct af_alg_tsgl,
prev              991 crypto/af_alg.c 	sgl = list_entry(ctx->tsgl_list.prev, struct af_alg_tsgl, list);
prev              109 crypto/cmac.c  	u8 *prev = PTR_ALIGN((void *)ctx->ctx, alignmask + 1) + bs;
prev              112 crypto/cmac.c  	memset(prev, 0, bs);
prev              127 crypto/cmac.c  	u8 *prev = odds + bs;
prev              141 crypto/cmac.c  	crypto_xor(prev, odds, bs);
prev              142 crypto/cmac.c  	crypto_cipher_encrypt_one(tfm, prev, prev);
prev              149 crypto/cmac.c  		crypto_xor(prev, p, bs);
prev              150 crypto/cmac.c  		crypto_cipher_encrypt_one(tfm, prev, prev);
prev              175 crypto/cmac.c  	u8 *prev = odds + bs;
prev              192 crypto/cmac.c  	crypto_xor(prev, odds, bs);
prev              193 crypto/cmac.c  	crypto_xor(prev, consts + offset, bs);
prev              195 crypto/cmac.c  	crypto_cipher_encrypt_one(tfm, out, prev);
prev              259 crypto/drbg.c  		memcpy(drbg->prev, entropy, entropylen);
prev              264 crypto/drbg.c  	ret = memcmp(drbg->prev, entropy, entropylen);
prev              267 crypto/drbg.c  	memcpy(drbg->prev, entropy, entropylen);
prev             1221 crypto/drbg.c  		kzfree(drbg->prev);
prev             1222 crypto/drbg.c  		drbg->prev = NULL;
prev             1295 crypto/drbg.c  		drbg->prev = kzalloc(drbg_sec_strength(drbg->core->flags),
prev             1297 crypto/drbg.c  		if (!drbg->prev) {
prev               76 crypto/xcbc.c  	u8 *prev = PTR_ALIGN(&ctx->ctx[0], alignmask + 1) + bs;
prev               79 crypto/xcbc.c  	memset(prev, 0, bs);
prev               94 crypto/xcbc.c  	u8 *prev = odds + bs;
prev              108 crypto/xcbc.c  	crypto_xor(prev, odds, bs);
prev              109 crypto/xcbc.c  	crypto_cipher_encrypt_one(tfm, prev, prev);
prev              116 crypto/xcbc.c  		crypto_xor(prev, p, bs);
prev              117 crypto/xcbc.c  		crypto_cipher_encrypt_one(tfm, prev, prev);
prev              141 crypto/xcbc.c  	u8 *prev = odds + bs;
prev              158 crypto/xcbc.c  	crypto_xor(prev, odds, bs);
prev              159 crypto/xcbc.c  	crypto_xor(prev, consts + offset, bs);
prev              161 crypto/xcbc.c  	crypto_cipher_encrypt_one(tfm, out, prev);
prev              128 drivers/acpi/acpica/acobject.h 	union acpi_operand_object *prev;	/* Link for list of acquired mutexes */
prev               46 drivers/acpi/acpica/exmutex.c 		(obj_desc->mutex.next)->mutex.prev = obj_desc->mutex.prev;
prev               49 drivers/acpi/acpica/exmutex.c 	if (obj_desc->mutex.prev) {
prev               50 drivers/acpi/acpica/exmutex.c 		(obj_desc->mutex.prev)->mutex.next = obj_desc->mutex.next;
prev               58 drivers/acpi/acpica/exmutex.c 		(obj_desc->mutex.prev)->mutex.original_sync_level =
prev               88 drivers/acpi/acpica/exmutex.c 	obj_desc->mutex.prev = NULL;
prev               94 drivers/acpi/acpica/exmutex.c 		list_head->mutex.prev = obj_desc;
prev              498 drivers/acpi/acpica/exmutex.c 		obj_desc->mutex.prev = NULL;
prev              708 drivers/acpi/acpica/psargs.c 	union acpi_parse_object *prev = NULL;
prev              757 drivers/acpi/acpica/psargs.c 				if (prev) {
prev              758 drivers/acpi/acpica/psargs.c 					prev->common.next = field;
prev              762 drivers/acpi/acpica/psargs.c 				prev = field;
prev              103 drivers/acpi/acpica/psparse.c 	union acpi_parse_object *prev;
prev              130 drivers/acpi/acpica/psparse.c 		prev = op->common.parent->common.value.arg;
prev              131 drivers/acpi/acpica/psparse.c 		if (!prev) {
prev              224 drivers/acpi/acpica/psparse.c 		if (prev == op) {
prev              245 drivers/acpi/acpica/psparse.c 			while (prev) {
prev              249 drivers/acpi/acpica/psparse.c 				next = prev->common.next;
prev              258 drivers/acpi/acpica/psparse.c 						prev->common.next =
prev              264 drivers/acpi/acpica/psparse.c 						prev->common.next =
prev              269 drivers/acpi/acpica/psparse.c 				prev = next;
prev               99 drivers/acpi/acpica/utaddress.c 	struct acpi_address_range *prev;
prev              110 drivers/acpi/acpica/utaddress.c 	range_info = prev = acpi_gbl_address_range_list[space_id];
prev              113 drivers/acpi/acpica/utaddress.c 			if (range_info == prev) {	/* Found at list head */
prev              117 drivers/acpi/acpica/utaddress.c 				prev->next = range_info->next;
prev              133 drivers/acpi/acpica/utaddress.c 		prev = range_info;
prev              679 drivers/acpi/nfit/core.c 		struct nfit_table_prev *prev,
prev              688 drivers/acpi/nfit/core.c 	list_for_each_entry(nfit_spa, &prev->spas, list) {
prev              709 drivers/acpi/nfit/core.c 		struct nfit_table_prev *prev,
prev              718 drivers/acpi/nfit/core.c 	list_for_each_entry(nfit_memdev, &prev->memdevs, list)
prev              780 drivers/acpi/nfit/core.c 		struct nfit_table_prev *prev,
prev              789 drivers/acpi/nfit/core.c 	list_for_each_entry(nfit_dcr, &prev->dcrs, list)
prev              808 drivers/acpi/nfit/core.c 		struct nfit_table_prev *prev,
prev              816 drivers/acpi/nfit/core.c 	list_for_each_entry(nfit_bdw, &prev->bdws, list)
prev              842 drivers/acpi/nfit/core.c 		struct nfit_table_prev *prev,
prev              851 drivers/acpi/nfit/core.c 	list_for_each_entry(nfit_idt, &prev->idts, list) {
prev              881 drivers/acpi/nfit/core.c 		struct nfit_table_prev *prev,
prev              890 drivers/acpi/nfit/core.c 	list_for_each_entry(nfit_flush, &prev->flushes, list) {
prev              926 drivers/acpi/nfit/core.c 		struct nfit_table_prev *prev, void *table, const void *end)
prev              944 drivers/acpi/nfit/core.c 		if (!add_spa(acpi_desc, prev, table))
prev              948 drivers/acpi/nfit/core.c 		if (!add_memdev(acpi_desc, prev, table))
prev              952 drivers/acpi/nfit/core.c 		if (!add_dcr(acpi_desc, prev, table))
prev              956 drivers/acpi/nfit/core.c 		if (!add_bdw(acpi_desc, prev, table))
prev              960 drivers/acpi/nfit/core.c 		if (!add_idt(acpi_desc, prev, table))
prev              964 drivers/acpi/nfit/core.c 		if (!add_flush(acpi_desc, prev, table))
prev             3325 drivers/acpi/nfit/core.c 		struct nfit_table_prev *prev)
prev             3329 drivers/acpi/nfit/core.c 	if (!list_empty(&prev->spas) ||
prev             3330 drivers/acpi/nfit/core.c 			!list_empty(&prev->memdevs) ||
prev             3331 drivers/acpi/nfit/core.c 			!list_empty(&prev->dcrs) ||
prev             3332 drivers/acpi/nfit/core.c 			!list_empty(&prev->bdws) ||
prev             3333 drivers/acpi/nfit/core.c 			!list_empty(&prev->idts) ||
prev             3334 drivers/acpi/nfit/core.c 			!list_empty(&prev->flushes)) {
prev             3376 drivers/acpi/nfit/core.c 	struct nfit_table_prev prev;
prev             3405 drivers/acpi/nfit/core.c 	INIT_LIST_HEAD(&prev.spas);
prev             3406 drivers/acpi/nfit/core.c 	INIT_LIST_HEAD(&prev.memdevs);
prev             3407 drivers/acpi/nfit/core.c 	INIT_LIST_HEAD(&prev.dcrs);
prev             3408 drivers/acpi/nfit/core.c 	INIT_LIST_HEAD(&prev.bdws);
prev             3409 drivers/acpi/nfit/core.c 	INIT_LIST_HEAD(&prev.idts);
prev             3410 drivers/acpi/nfit/core.c 	INIT_LIST_HEAD(&prev.flushes);
prev             3412 drivers/acpi/nfit/core.c 	list_cut_position(&prev.spas, &acpi_desc->spas,
prev             3413 drivers/acpi/nfit/core.c 				acpi_desc->spas.prev);
prev             3414 drivers/acpi/nfit/core.c 	list_cut_position(&prev.memdevs, &acpi_desc->memdevs,
prev             3415 drivers/acpi/nfit/core.c 				acpi_desc->memdevs.prev);
prev             3416 drivers/acpi/nfit/core.c 	list_cut_position(&prev.dcrs, &acpi_desc->dcrs,
prev             3417 drivers/acpi/nfit/core.c 				acpi_desc->dcrs.prev);
prev             3418 drivers/acpi/nfit/core.c 	list_cut_position(&prev.bdws, &acpi_desc->bdws,
prev             3419 drivers/acpi/nfit/core.c 				acpi_desc->bdws.prev);
prev             3420 drivers/acpi/nfit/core.c 	list_cut_position(&prev.idts, &acpi_desc->idts,
prev             3421 drivers/acpi/nfit/core.c 				acpi_desc->idts.prev);
prev             3422 drivers/acpi/nfit/core.c 	list_cut_position(&prev.flushes, &acpi_desc->flushes,
prev             3423 drivers/acpi/nfit/core.c 				acpi_desc->flushes.prev);
prev             3427 drivers/acpi/nfit/core.c 		data = add_table(acpi_desc, &prev, data, end);
prev             3435 drivers/acpi/nfit/core.c 	rc = acpi_nfit_check_deletions(acpi_desc, &prev);
prev             1170 drivers/acpi/processor_idle.c 	struct acpi_lpi_states_array info[2], *tmp, *prev, *curr;
prev             1179 drivers/acpi/processor_idle.c 	prev = &info[0];
prev             1182 drivers/acpi/processor_idle.c 	ret = acpi_processor_evaluate_lpi(handle, prev);
prev             1185 drivers/acpi/processor_idle.c 	flatten_lpi_states(pr, prev, NULL);
prev             1204 drivers/acpi/processor_idle.c 		flatten_lpi_states(pr, curr, prev);
prev             1206 drivers/acpi/processor_idle.c 		tmp = prev, prev = curr, curr = tmp;
prev             1142 drivers/acpi/property.c 	const struct fwnode_handle *fwnode, struct fwnode_handle *prev)
prev             1147 drivers/acpi/property.c 	if (!prev) {
prev             1161 drivers/acpi/property.c 		port = fwnode_get_parent(prev);
prev             1167 drivers/acpi/property.c 	endpoint = fwnode_get_next_child_node(port, prev);
prev               57 drivers/android/binder_alloc.c 	return list_entry(buffer->entry.prev, struct binder_buffer, entry);
prev              549 drivers/android/binder_alloc.c 	struct binder_buffer *prev, *next = NULL;
prev              552 drivers/android/binder_alloc.c 	prev = binder_buffer_prev(buffer);
prev              553 drivers/android/binder_alloc.c 	BUG_ON(!prev->free);
prev              554 drivers/android/binder_alloc.c 	if (prev_buffer_end_page(prev) == buffer_start_page(buffer)) {
prev              559 drivers/android/binder_alloc.c 				   prev->user_data);
prev              585 drivers/android/binder_alloc.c 				   prev->user_data,
prev              639 drivers/android/binder_alloc.c 		struct binder_buffer *prev = binder_buffer_prev(buffer);
prev              641 drivers/android/binder_alloc.c 		if (prev->free) {
prev              643 drivers/android/binder_alloc.c 			rb_erase(&prev->rb_node, &alloc->free_buffers);
prev              644 drivers/android/binder_alloc.c 			buffer = prev;
prev              253 drivers/android/binder_alloc_selftest.c 	size_t end, prev;
prev              259 drivers/android/binder_alloc_selftest.c 	prev = index == 0 ? 0 : end_offset[index - 1];
prev              260 drivers/android/binder_alloc_selftest.c 	end = prev;
prev              135 drivers/ata/pata_cs5530.c 	struct ata_device *prev = ap->private_data;
prev              138 drivers/ata/pata_cs5530.c 	if (ata_dma_enabled(adev) && adev != prev && prev != NULL) {
prev              140 drivers/ata/pata_cs5530.c 		if ((ata_using_udma(adev) && !ata_using_udma(prev)) ||
prev              141 drivers/ata/pata_cs5530.c 		    (ata_using_udma(prev) && !ata_using_udma(adev)))
prev              156 drivers/ata/pata_sc1200.c 	struct ata_device *prev = ap->private_data;
prev              159 drivers/ata/pata_sc1200.c 	if (ata_dma_enabled(adev) && adev != prev && prev != NULL) {
prev              161 drivers/ata/pata_sc1200.c 		if ((ata_using_udma(adev) && !ata_using_udma(prev)) ||
prev              162 drivers/ata/pata_sc1200.c 		    (ata_using_udma(prev) && !ata_using_udma(adev)))
prev              323 drivers/atm/idt77105.c 	struct idt77105_priv *walk, *prev;
prev              331 drivers/atm/idt77105.c 	for (prev = NULL, walk = idt77105_all ;
prev              333 drivers/atm/idt77105.c              prev = walk, walk = walk->next) {
prev              335 drivers/atm/idt77105.c                 if (prev != NULL)
prev              336 drivers/atm/idt77105.c                     prev->next = walk->next;
prev              332 drivers/atm/suni.c printk(KERN_DEBUG "[u] p=0x%lx,n=0x%lx\n",(unsigned long) poll_timer.list.prev,
prev             3192 drivers/base/core.c 		dev = list_entry(devices_kset->list.prev, struct device,
prev             1225 drivers/base/dd.c 		dev_prv = list_entry(drv->p->klist_devices.k_list.prev,
prev              380 drivers/base/power/domain.c 	unsigned int prev;
prev              397 drivers/base/power/domain.c 	prev = gpd_data->performance_state;
prev              403 drivers/base/power/domain.c 		gpd_data->performance_state = prev;
prev             1168 drivers/base/power/main.c 		struct device *dev = to_device(dpm_prepared_list.prev);
prev             1400 drivers/base/power/main.c 		struct device *dev = to_device(dpm_late_early_list.prev);
prev             1600 drivers/base/power/main.c 		struct device *dev = to_device(dpm_suspended_list.prev);
prev             1872 drivers/base/power/main.c 		struct device *dev = to_device(dpm_prepared_list.prev);
prev              200 drivers/base/power/trace.c 	entry = dpm_list.prev;
prev              208 drivers/base/power/trace.c 		entry = entry->prev;
prev              227 drivers/base/power/trace.c 	entry = dpm_list.prev;
prev              241 drivers/base/power/trace.c 		entry = entry->prev;
prev              900 drivers/base/property.c 			       struct fwnode_handle *prev)
prev              902 drivers/base/property.c 	return fwnode_call_ptr_op(fwnode, graph_get_next_endpoint, prev);
prev              939 drivers/block/drbd/drbd_nl.c 	} prev;
prev              963 drivers/block/drbd/drbd_nl.c 	prev.last_agreed_sect = md->la_size_sect;
prev              964 drivers/block/drbd/drbd_nl.c 	prev.md_offset = md->md_offset;
prev              965 drivers/block/drbd/drbd_nl.c 	prev.al_offset = md->al_offset;
prev              966 drivers/block/drbd/drbd_nl.c 	prev.bm_offset = md->bm_offset;
prev              967 drivers/block/drbd/drbd_nl.c 	prev.md_size_sect = md->md_size_sect;
prev              968 drivers/block/drbd/drbd_nl.c 	prev.al_stripes = md->al_stripes;
prev              969 drivers/block/drbd/drbd_nl.c 	prev.al_stripe_size_4k = md->al_stripe_size_4k;
prev              985 drivers/block/drbd/drbd_nl.c 	if (size < prev.last_agreed_sect) {
prev             1023 drivers/block/drbd/drbd_nl.c 	la_size_changed = (prev.last_agreed_sect != md->la_size_sect);
prev             1025 drivers/block/drbd/drbd_nl.c 	md_moved = prev.md_offset    != md->md_offset
prev             1026 drivers/block/drbd/drbd_nl.c 		|| prev.md_size_sect != md->md_size_sect;
prev             1066 drivers/block/drbd/drbd_nl.c 	if (size > prev.last_agreed_sect)
prev             1067 drivers/block/drbd/drbd_nl.c 		rv = prev.last_agreed_sect ? DS_GREW : DS_GREW_FROM_ZERO;
prev             1068 drivers/block/drbd/drbd_nl.c 	if (size < prev.last_agreed_sect)
prev             1074 drivers/block/drbd/drbd_nl.c 		md->la_size_sect = prev.last_agreed_sect;
prev             1075 drivers/block/drbd/drbd_nl.c 		md->md_offset = prev.md_offset;
prev             1076 drivers/block/drbd/drbd_nl.c 		md->al_offset = prev.al_offset;
prev             1077 drivers/block/drbd/drbd_nl.c 		md->bm_offset = prev.bm_offset;
prev             1078 drivers/block/drbd/drbd_nl.c 		md->md_size_sect = prev.md_size_sect;
prev             1079 drivers/block/drbd/drbd_nl.c 		md->al_stripes = prev.al_stripes;
prev             1080 drivers/block/drbd/drbd_nl.c 		md->al_stripe_size_4k = prev.al_stripe_size_4k;
prev             1081 drivers/block/drbd/drbd_nl.c 		md->al_size_4k = (u64)prev.al_stripes * prev.al_stripe_size_4k;
prev               68 drivers/char/agp/frontend.c 	struct agp_memory *prev;
prev               76 drivers/char/agp/frontend.c 		prev = temp->prev;
prev               78 drivers/char/agp/frontend.c 		if (prev != NULL) {
prev               79 drivers/char/agp/frontend.c 			prev->next = next;
prev               81 drivers/char/agp/frontend.c 				next->prev = prev;
prev               86 drivers/char/agp/frontend.c 				next->prev = NULL;
prev              201 drivers/char/agp/frontend.c 	struct agp_memory *prev;
prev              203 drivers/char/agp/frontend.c 	prev = agp_fe.current_controller->pool;
prev              205 drivers/char/agp/frontend.c 	if (prev != NULL) {
prev              206 drivers/char/agp/frontend.c 		prev->prev = temp;
prev              207 drivers/char/agp/frontend.c 		temp->next = prev;
prev              232 drivers/char/agp/frontend.c 	struct agp_file_private *prev;
prev              234 drivers/char/agp/frontend.c 	prev = agp_fe.file_priv_list;
prev              236 drivers/char/agp/frontend.c 	if (prev != NULL)
prev              237 drivers/char/agp/frontend.c 		prev->prev = priv;
prev              238 drivers/char/agp/frontend.c 	priv->next = prev;
prev              245 drivers/char/agp/frontend.c 	struct agp_file_private *prev;
prev              248 drivers/char/agp/frontend.c 	prev = priv->prev;
prev              250 drivers/char/agp/frontend.c 	if (prev != NULL) {
prev              251 drivers/char/agp/frontend.c 		prev->next = next;
prev              254 drivers/char/agp/frontend.c 			next->prev = prev;
prev              258 drivers/char/agp/frontend.c 			next->prev = NULL;
prev              328 drivers/char/agp/frontend.c 		prev_controller->prev = controller;
prev              377 drivers/char/agp/frontend.c 	prev_controller = controller->prev;
prev              383 drivers/char/agp/frontend.c 			next_controller->prev = prev_controller;
prev              387 drivers/char/agp/frontend.c 			next_controller->prev = NULL;
prev              507 drivers/char/agp/frontend.c 		prev_client->prev = client;
prev              541 drivers/char/agp/frontend.c 	prev_client = client->prev;
prev              547 drivers/char/agp/frontend.c 			next_client->prev = prev_client;
prev              551 drivers/char/agp/frontend.c 			next_client->prev = NULL;
prev              402 drivers/char/virtio_console.c 			  pending_free_dma_bufs.prev);
prev              165 drivers/clk/clk-max9485.c 	const struct max9485_rate *curr, *prev = NULL;
prev              183 drivers/clk/clk-max9485.c 			if (!prev)
prev              190 drivers/clk/clk-max9485.c 			mid = prev->out + ((curr->out - prev->out) / 2);
prev              192 drivers/clk/clk-max9485.c 			return (mid > rate) ? prev->out : curr->out;
prev              195 drivers/clk/clk-max9485.c 		prev = curr;
prev              199 drivers/clk/clk-max9485.c 	return prev->out;
prev              508 drivers/clk/keystone/sci-clk.c 	struct sci_clk *sci_clk, *prev;
prev              607 drivers/clk/keystone/sci-clk.c 	prev = NULL;
prev              610 drivers/clk/keystone/sci-clk.c 		if (prev && prev->dev_id == sci_clk->dev_id &&
prev              611 drivers/clk/keystone/sci-clk.c 		    prev->clk_id == sci_clk->clk_id)
prev              615 drivers/clk/keystone/sci-clk.c 		prev = sci_clk;
prev              296 drivers/cpufreq/freq_table.c 	struct cpufreq_frequency_table *prev = NULL;
prev              302 drivers/cpufreq/freq_table.c 		if (!prev) {
prev              303 drivers/cpufreq/freq_table.c 			prev = pos;
prev              307 drivers/cpufreq/freq_table.c 		if (pos->frequency == prev->frequency) {
prev              314 drivers/cpufreq/freq_table.c 		if (pos->frequency > prev->frequency) {
prev              334 drivers/cpufreq/freq_table.c 		prev = pos;
prev              127 drivers/cpufreq/qcom-cpufreq-hw.c 			struct cpufreq_frequency_table *prev = &table[i - 1];
prev              133 drivers/cpufreq/qcom-cpufreq-hw.c 			if (prev->frequency == CPUFREQ_ENTRY_INVALID) {
prev              134 drivers/cpufreq/qcom-cpufreq-hw.c 				prev->frequency = prev_freq;
prev              135 drivers/cpufreq/qcom-cpufreq-hw.c 				prev->flags = CPUFREQ_BOOST_FREQ;
prev              993 drivers/crypto/chelsio/chcr_algo.c 	u32 c, prev;
prev              997 drivers/crypto/chelsio/chcr_algo.c 		prev = be32_to_cpu(*--b);
prev              998 drivers/crypto/chelsio/chcr_algo.c 		c = prev + add;
prev             1000 drivers/crypto/chelsio/chcr_algo.c 		if (prev < c)
prev              227 drivers/crypto/chelsio/chtls/chtls.h 		struct sk_buff *prev;
prev              541 drivers/crypto/chelsio/chtls/chtls_cm.c 	struct listen_info *p, **prev;
prev              546 drivers/crypto/chelsio/chtls/chtls_cm.c 	prev = &cdev->listen_hash_tab[key];
prev              549 drivers/crypto/chelsio/chtls/chtls_cm.c 	for (p = *prev; p; prev = &p->next, p = p->next)
prev              552 drivers/crypto/chelsio/chtls/chtls_cm.c 			*prev = p->next;
prev               23 drivers/dma-buf/dma-fence-chain.c 	struct dma_fence *prev;
prev               26 drivers/dma-buf/dma-fence-chain.c 	prev = dma_fence_get_rcu_safe(&chain->prev);
prev               28 drivers/dma-buf/dma-fence-chain.c 	return prev;
prev               42 drivers/dma-buf/dma-fence-chain.c 	struct dma_fence *prev, *replacement, *tmp;
prev               50 drivers/dma-buf/dma-fence-chain.c 	while ((prev = dma_fence_chain_get_prev(chain))) {
prev               52 drivers/dma-buf/dma-fence-chain.c 		prev_chain = to_dma_fence_chain(prev);
prev               59 drivers/dma-buf/dma-fence-chain.c 			if (!dma_fence_is_signaled(prev))
prev               65 drivers/dma-buf/dma-fence-chain.c 		tmp = cmpxchg((void **)&chain->prev, (void *)prev, (void *)replacement);
prev               66 drivers/dma-buf/dma-fence-chain.c 		if (tmp == prev)
prev               70 drivers/dma-buf/dma-fence-chain.c 		dma_fence_put(prev);
prev               74 drivers/dma-buf/dma-fence-chain.c 	return prev;
prev              181 drivers/dma-buf/dma-fence-chain.c 	struct dma_fence *prev;
prev              186 drivers/dma-buf/dma-fence-chain.c 	while ((prev = rcu_dereference_protected(chain->prev, true))) {
prev              189 drivers/dma-buf/dma-fence-chain.c 		if (kref_read(&prev->refcount) > 1)
prev              192 drivers/dma-buf/dma-fence-chain.c 		prev_chain = to_dma_fence_chain(prev);
prev              199 drivers/dma-buf/dma-fence-chain.c 		chain->prev = prev_chain->prev;
prev              200 drivers/dma-buf/dma-fence-chain.c 		RCU_INIT_POINTER(prev_chain->prev, NULL);
prev              201 drivers/dma-buf/dma-fence-chain.c 		dma_fence_put(prev);
prev              203 drivers/dma-buf/dma-fence-chain.c 	dma_fence_put(prev);
prev              229 drivers/dma-buf/dma-fence-chain.c 			  struct dma_fence *prev,
prev              233 drivers/dma-buf/dma-fence-chain.c 	struct dma_fence_chain *prev_chain = to_dma_fence_chain(prev);
prev              237 drivers/dma-buf/dma-fence-chain.c 	rcu_assign_pointer(chain->prev, prev);
prev              243 drivers/dma-buf/dma-fence-chain.c 	if (prev_chain && __dma_fence_is_later(seqno, prev->seqno, prev->ops)) {
prev              244 drivers/dma-buf/dma-fence-chain.c 		context = prev->context;
prev              245 drivers/dma-buf/dma-fence-chain.c 		chain->prev_seqno = prev->seqno;
prev              250 drivers/dma-buf/dma-fence-chain.c 			seqno = max(prev->seqno, seqno);
prev              196 drivers/dma/at_hdmac.c static void atc_desc_chain(struct at_desc **first, struct at_desc **prev,
prev              203 drivers/dma/at_hdmac.c 		(*prev)->lli.dscr = desc->txd.phys;
prev              208 drivers/dma/at_hdmac.c 	*prev = desc;
prev              555 drivers/dma/at_hdmac.c 	list_splice_init(&atchan->queue, atchan->active_list.prev);
prev              807 drivers/dma/at_hdmac.c 	struct at_desc		*prev = NULL;
prev              853 drivers/dma/at_hdmac.c 		atc_desc_chain(&first, &prev, desc);
prev              985 drivers/dma/at_hdmac.c 	struct at_desc		*desc = NULL, *first = NULL, *prev = NULL;
prev             1026 drivers/dma/at_hdmac.c 		atc_desc_chain(&first, &prev, desc);
prev             1072 drivers/dma/at_hdmac.c 	struct at_desc		*prev = NULL;
prev             1133 drivers/dma/at_hdmac.c 			atc_desc_chain(&first, &prev, desc);
prev             1174 drivers/dma/at_hdmac.c 			atc_desc_chain(&first, &prev, desc);
prev             1183 drivers/dma/at_hdmac.c 	set_desc_eol(prev);
prev             1292 drivers/dma/at_hdmac.c 	struct at_desc		*prev = NULL;
prev             1338 drivers/dma/at_hdmac.c 		atc_desc_chain(&first, &prev, desc);
prev             1342 drivers/dma/at_hdmac.c 	prev->lli.dscr = first->txd.phys;
prev              485 drivers/dma/at_xdmac.c 				struct at_xdmac_desc *prev,
prev              488 drivers/dma/at_xdmac.c 	if (!prev || !desc)
prev              491 drivers/dma/at_xdmac.c 	prev->lld.mbr_nda = desc->tx_dma_desc.phys;
prev              492 drivers/dma/at_xdmac.c 	prev->lld.mbr_ubc |= AT_XDMAC_MBR_UBC_NDE;
prev              495 drivers/dma/at_xdmac.c 		__func__, prev, &prev->lld.mbr_nda);
prev              638 drivers/dma/at_xdmac.c 	struct at_xdmac_desc		*first = NULL, *prev = NULL;
prev              709 drivers/dma/at_xdmac.c 		if (prev)
prev              710 drivers/dma/at_xdmac.c 			at_xdmac_queue_desc(chan, prev, desc);
prev              712 drivers/dma/at_xdmac.c 		prev = desc;
prev              740 drivers/dma/at_xdmac.c 	struct at_xdmac_desc	*first = NULL, *prev = NULL;
prev              797 drivers/dma/at_xdmac.c 		if (prev)
prev              798 drivers/dma/at_xdmac.c 			at_xdmac_queue_desc(chan, prev, desc);
prev              800 drivers/dma/at_xdmac.c 		prev = desc;
prev              809 drivers/dma/at_xdmac.c 	at_xdmac_queue_desc(chan, prev, first);
prev              849 drivers/dma/at_xdmac.c 				struct at_xdmac_desc *prev,
prev              885 drivers/dma/at_xdmac.c 	if (prev)
prev              887 drivers/dma/at_xdmac.c 			"Adding items at the end of desc 0x%p\n", prev);
prev              932 drivers/dma/at_xdmac.c 	if (prev)
prev              933 drivers/dma/at_xdmac.c 		at_xdmac_queue_desc(chan, prev, desc);
prev              944 drivers/dma/at_xdmac.c 	struct at_xdmac_desc	*prev = NULL, *first = NULL;
prev              998 drivers/dma/at_xdmac.c 							       prev,
prev             1021 drivers/dma/at_xdmac.c 			prev = desc;
prev             1037 drivers/dma/at_xdmac.c 	struct at_xdmac_desc	*first = NULL, *prev = NULL;
prev             1118 drivers/dma/at_xdmac.c 		if (prev)
prev             1119 drivers/dma/at_xdmac.c 			at_xdmac_queue_desc(chan, prev, desc);
prev             1121 drivers/dma/at_xdmac.c 		prev = desc;
prev              428 drivers/dma/dw-axi-dmac/dw-axi-dmac-platform.c 	struct axi_dma_desc *first = NULL, *desc = NULL, *prev = NULL;
prev              495 drivers/dma/dw-axi-dmac/dw-axi-dmac-platform.c 			write_desc_llp(prev, desc->vd.tx.phys | lms);
prev              497 drivers/dma/dw-axi-dmac/dw-axi-dmac-platform.c 		prev = desc;
prev              336 drivers/dma/dw/core.c 					desc->residue -= to_dw_desc(active->prev)->len;
prev              445 drivers/dma/dw/core.c 	list_move(dwc->queue.next, dwc->active_list.prev);
prev              552 drivers/dma/dw/core.c 	struct dw_desc		*prev;
prev              581 drivers/dma/dw/core.c 	prev = first = NULL;
prev              599 drivers/dma/dw/core.c 			lli_write(prev, llp, desc->txd.phys | lms);
prev              602 drivers/dma/dw/core.c 		prev = desc;
prev              607 drivers/dma/dw/core.c 		lli_set(prev, ctllo, DWC_CTLL_INT_EN);
prev              609 drivers/dma/dw/core.c 	prev->lli.llp = 0;
prev              610 drivers/dma/dw/core.c 	lli_clear(prev, ctllo, DWC_CTLL_LLP_D_EN | DWC_CTLL_LLP_S_EN);
prev              629 drivers/dma/dw/core.c 	struct dw_desc		*prev;
prev              649 drivers/dma/dw/core.c 	prev = first = NULL;
prev              689 drivers/dma/dw/core.c 				lli_write(prev, llp, desc->txd.phys | lms);
prev              692 drivers/dma/dw/core.c 			prev = desc;
prev              738 drivers/dma/dw/core.c 				lli_write(prev, llp, desc->txd.phys | lms);
prev              741 drivers/dma/dw/core.c 			prev = desc;
prev              757 drivers/dma/dw/core.c 		lli_set(prev, ctllo, DWC_CTLL_INT_EN);
prev              759 drivers/dma/dw/core.c 	prev->lli.llp = 0;
prev              760 drivers/dma/dw/core.c 	lli_clear(prev, ctllo, DWC_CTLL_LLP_D_EN | DWC_CTLL_LLP_S_EN);
prev              384 drivers/dma/fsldma.c 	struct fsl_desc_sw *tail = to_fsl_desc(chan->ld_pending.prev);
prev              764 drivers/dma/fsldma.c 	struct fsl_desc_sw *first = NULL, *prev = NULL, *new;
prev              793 drivers/dma/fsldma.c 			set_desc_next(chan, &prev->hw, new->async_tx.phys);
prev              798 drivers/dma/fsldma.c 		prev = new;
prev              292 drivers/dma/img-mdc-dma.c 	struct mdc_hw_list_desc *curr, *prev = NULL;
prev              311 drivers/dma/img-mdc-dma.c 		if (prev) {
prev              312 drivers/dma/img-mdc-dma.c 			prev->node_addr = curr_phys;
prev              313 drivers/dma/img-mdc-dma.c 			prev->next_desc = curr;
prev              324 drivers/dma/img-mdc-dma.c 		prev = curr;
prev              374 drivers/dma/img-mdc-dma.c 	struct mdc_hw_list_desc *curr, *prev = NULL;
prev              406 drivers/dma/img-mdc-dma.c 			if (!prev) {
prev              410 drivers/dma/img-mdc-dma.c 				prev->node_addr = curr_phys;
prev              411 drivers/dma/img-mdc-dma.c 				prev->next_desc = curr;
prev              429 drivers/dma/img-mdc-dma.c 			prev = curr;
prev              437 drivers/dma/img-mdc-dma.c 	prev->node_addr = mdesc->list_phys;
prev              456 drivers/dma/img-mdc-dma.c 	struct mdc_hw_list_desc *curr, *prev = NULL;
prev              486 drivers/dma/img-mdc-dma.c 			if (!prev) {
prev              490 drivers/dma/img-mdc-dma.c 				prev->node_addr = curr_phys;
prev              491 drivers/dma/img-mdc-dma.c 				prev->next_desc = curr;
prev              507 drivers/dma/img-mdc-dma.c 			prev = curr;
prev              376 drivers/dma/iop-adma.c 	old_chain_tail = list_entry(iop_chan->chain.prev,
prev              418 drivers/dma/mediatek/mtk-hsdma.c 	u16 reserved, prev, tlen, num_sgs;
prev              477 drivers/dma/mediatek/mtk-hsdma.c 		prev = MTK_HSDMA_LAST_DESP_IDX(ring->cur_tptr, MTK_DMA_SIZE);
prev              478 drivers/dma/mediatek/mtk-hsdma.c 		ring->cb[prev].flag = MTK_HSDMA_VDESC_FINISHED;
prev              450 drivers/dma/mmp_pdma.c 	struct mmp_pdma_desc_sw *first = NULL, *prev = NULL, *new;
prev              487 drivers/dma/mmp_pdma.c 			prev->desc.ddadr = new->async_tx.phys;
prev              492 drivers/dma/mmp_pdma.c 		prev = new;
prev              531 drivers/dma/mmp_pdma.c 	struct mmp_pdma_desc_sw *first = NULL, *prev = NULL, *new = NULL;
prev              572 drivers/dma/mmp_pdma.c 				prev->desc.ddadr = new->async_tx.phys;
prev              576 drivers/dma/mmp_pdma.c 			prev = new;
prev              612 drivers/dma/mmp_pdma.c 	struct mmp_pdma_desc_sw *first = NULL, *prev = NULL, *new;
prev              660 drivers/dma/mmp_pdma.c 			prev->desc.ddadr = new->async_tx.phys;
prev              665 drivers/dma/mmp_pdma.c 		prev = new;
prev              258 drivers/dma/mpc512x_dma.c 	struct mpc_dma_desc *prev = NULL;
prev              285 drivers/dma/mpc512x_dma.c 		if (!prev) {
prev              286 drivers/dma/mpc512x_dma.c 			prev = mdesc;
prev              290 drivers/dma/mpc512x_dma.c 		prev->tcd->dlast_sga = mdesc->tcd_paddr;
prev              291 drivers/dma/mpc512x_dma.c 		prev->tcd->e_sg = 1;
prev              294 drivers/dma/mpc512x_dma.c 		prev = mdesc;
prev              297 drivers/dma/mpc512x_dma.c 	prev->tcd->int_maj = 1;
prev              302 drivers/dma/mpc512x_dma.c 	if (first != prev)
prev              402 drivers/dma/mv_xor.c 		old_chain_tail = list_entry(mv_chan->chain.prev,
prev              779 drivers/dma/nbpfaxi.c 	struct nbpf_link_desc *ldesc, *prev = NULL;
prev              816 drivers/dma/nbpfaxi.c 			if (prev)
prev              817 drivers/dma/nbpfaxi.c 				prev->hwdesc->next = (u32)ldesc->hwdesc_dma_addr;
prev              819 drivers/dma/nbpfaxi.c 			prev = ldesc;
prev              826 drivers/dma/nbpfaxi.c 	prev->hwdesc->next = 0;
prev              344 drivers/dma/owl-dma.c 					   struct owl_dma_lli *prev,
prev              351 drivers/dma/owl-dma.c 	if (prev) {
prev              352 drivers/dma/owl-dma.c 		prev->hw.next_lli = next->phys;
prev              353 drivers/dma/owl-dma.c 		prev->hw.ctrla |= llc_hw_ctrla(OWL_DMA_MODE_LME, 0);
prev              833 drivers/dma/owl-dma.c 	struct owl_dma_lli *lli, *prev = NULL;
prev              864 drivers/dma/owl-dma.c 		prev = owl_dma_add_lli(txd, prev, lli, false);
prev              885 drivers/dma/owl-dma.c 	struct owl_dma_lli *lli, *prev = NULL;
prev              928 drivers/dma/owl-dma.c 		prev = owl_dma_add_lli(txd, prev, lli, false);
prev              950 drivers/dma/owl-dma.c 	struct owl_dma_lli *lli, *prev = NULL, *first = NULL;
prev              987 drivers/dma/owl-dma.c 		prev = owl_dma_add_lli(txd, prev, lli, false);
prev              991 drivers/dma/owl-dma.c 	owl_dma_add_lli(txd, prev, first, true);
prev              385 drivers/dma/pch_dma.c 	list_splice_init(&pd_chan->queue, pd_chan->active_list.prev);
prev              574 drivers/dma/pch_dma.c 	struct pch_dma_desc *prev = NULL;
prev              629 drivers/dma/pch_dma.c 			prev->regs.next |= desc->txd.phys;
prev              633 drivers/dma/pch_dma.c 		prev = desc;
prev             1921 drivers/dma/ppc4xx/adma.c 		old_chain_tail = list_entry(chan->chain.prev,
prev              574 drivers/dma/pxa_dma.c 		vd_last_issued = list_entry(vc->desc_issued.prev,
prev              802 drivers/dma/pxa_dma.c 		vd_chained = list_entry(vc->desc_submitted.prev,
prev               86 drivers/dma/sh/shdma-base.c 	list_for_each_entry_safe(chunk, c, desc->node.prev, node) {
prev              365 drivers/dma/sun6i-dma.c static void *sun6i_dma_lli_add(struct sun6i_dma_lli *prev,
prev              370 drivers/dma/sun6i-dma.c 	if ((!prev && !txd) || !next)
prev              373 drivers/dma/sun6i-dma.c 	if (!prev) {
prev              377 drivers/dma/sun6i-dma.c 		prev->p_lli_next = next_phy;
prev              378 drivers/dma/sun6i-dma.c 		prev->v_lli_next = next;
prev              690 drivers/dma/sun6i-dma.c 	struct sun6i_dma_lli *v_lli, *prev = NULL;
prev              745 drivers/dma/sun6i-dma.c 		prev = sun6i_dma_lli_add(prev, v_lli, p_lli, txd);
prev              749 drivers/dma/sun6i-dma.c 	for (prev = txd->v_lli; prev; prev = prev->v_lli_next)
prev              750 drivers/dma/sun6i-dma.c 		sun6i_dma_dump_lli(vchan, prev);
prev              755 drivers/dma/sun6i-dma.c 	for (prev = txd->v_lli; prev; prev = prev->v_lli_next)
prev              756 drivers/dma/sun6i-dma.c 		dma_pool_free(sdev->pool, prev, virt_to_phys(prev));
prev              772 drivers/dma/sun6i-dma.c 	struct sun6i_dma_lli *v_lli, *prev = NULL;
prev              813 drivers/dma/sun6i-dma.c 		prev = sun6i_dma_lli_add(prev, v_lli, p_lli, txd);
prev              816 drivers/dma/sun6i-dma.c 	prev->p_lli_next = txd->p_lli;		/* cyclic list */
prev              823 drivers/dma/sun6i-dma.c 	for (prev = txd->v_lli; prev; prev = prev->v_lli_next)
prev              824 drivers/dma/sun6i-dma.c 		dma_pool_free(sdev->pool, prev, virt_to_phys(prev));
prev              171 drivers/dma/txx9dmac.c 	return list_entry(dc->active_list.prev,
prev              183 drivers/dma/txx9dmac.c 		desc = list_entry(desc->tx_list.prev, typeof(*desc), desc_node);
prev              428 drivers/dma/txx9dmac.c 	struct txx9dmac_desc *prev = NULL;
prev              433 drivers/dma/txx9dmac.c 		if (prev) {
prev              434 drivers/dma/txx9dmac.c 			desc_write_CHAR(dc, prev, desc->txd.phys);
prev              436 drivers/dma/txx9dmac.c 				prev->txd.phys, ddev->descsize,
prev              439 drivers/dma/txx9dmac.c 		prev = txx9dmac_last_child(desc);
prev              716 drivers/dma/txx9dmac.c 	struct txx9dmac_desc *prev;
prev              728 drivers/dma/txx9dmac.c 	prev = first = NULL;
prev              778 drivers/dma/txx9dmac.c 			desc_write_CHAR(dc, prev, desc->txd.phys);
prev              780 drivers/dma/txx9dmac.c 					prev->txd.phys, ddev->descsize,
prev              784 drivers/dma/txx9dmac.c 		prev = desc;
prev              789 drivers/dma/txx9dmac.c 		txx9dmac_desc_set_INTENT(ddev, prev);
prev              791 drivers/dma/txx9dmac.c 	desc_write_CHAR(dc, prev, 0);
prev              793 drivers/dma/txx9dmac.c 			prev->txd.phys, ddev->descsize,
prev              810 drivers/dma/txx9dmac.c 	struct txx9dmac_desc *prev;
prev              825 drivers/dma/txx9dmac.c 	prev = first = NULL;
prev              872 drivers/dma/txx9dmac.c 			desc_write_CHAR(dc, prev, desc->txd.phys);
prev              874 drivers/dma/txx9dmac.c 					prev->txd.phys,
prev              879 drivers/dma/txx9dmac.c 		prev = desc;
prev              884 drivers/dma/txx9dmac.c 		txx9dmac_desc_set_INTENT(ddev, prev);
prev              886 drivers/dma/txx9dmac.c 	desc_write_CHAR(dc, prev, 0);
prev              888 drivers/dma/txx9dmac.c 			prev->txd.phys, ddev->descsize,
prev              940 drivers/dma/txx9dmac.c 				   struct txx9dmac_desc *prev)
prev              946 drivers/dma/txx9dmac.c 	prev = txx9dmac_last_child(prev);
prev              949 drivers/dma/txx9dmac.c 	desc_write_CHAR(dc, prev, desc->txd.phys);
prev              951 drivers/dma/txx9dmac.c 				   prev->txd.phys, ddev->descsize,
prev              954 drivers/dma/txx9dmac.c 	    channel_read_CHAR(dc) == prev->txd.phys)
prev              973 drivers/dma/txx9dmac.c 			struct txx9dmac_desc *prev = txx9dmac_last_active(dc);
prev              975 drivers/dma/txx9dmac.c 			if (!(prev->txd.flags & DMA_PREP_INTERRUPT) ||
prev              977 drivers/dma/txx9dmac.c 				txx9dmac_chain_dynamic(dc, prev);
prev             1881 drivers/dma/xilinx/xilinx_dma.c 	struct xilinx_axidma_tx_segment *segment, *head_segment, *prev = NULL;
prev             1929 drivers/dma/xilinx/xilinx_dma.c 			if (prev)
prev             1930 drivers/dma/xilinx/xilinx_dma.c 				prev->hw.next_desc = segment->phys;
prev             1932 drivers/dma/xilinx/xilinx_dma.c 			prev = segment;
prev              305 drivers/dma/xilinx/zynqmp_dma.c 				   struct zynqmp_dma_desc_ll *prev)
prev              319 drivers/dma/xilinx/zynqmp_dma.c 	if (prev) {
prev              322 drivers/dma/xilinx/zynqmp_dma.c 		ddesc = prev + 1;
prev              323 drivers/dma/xilinx/zynqmp_dma.c 		prev->nxtdscraddr = addr;
prev              806 drivers/dma/xilinx/zynqmp_dma.c 	void *desc = NULL, *prev = NULL;
prev              831 drivers/dma/xilinx/zynqmp_dma.c 					     dma_dst, copy, prev);
prev              832 drivers/dma/xilinx/zynqmp_dma.c 		prev = desc;
prev             1305 drivers/edac/i7core_edac.c static int i7core_get_onedevice(struct pci_dev **prev,
prev             1318 drivers/edac/i7core_edac.c 			      dev_descr->dev_id, *prev);
prev             1326 drivers/edac/i7core_edac.c 		pci_dev_get(*prev);	/* pci_get_device will put it */
prev             1328 drivers/edac/i7core_edac.c 				      PCI_DEVICE_ID_INTEL_I7_NONCORE_ALT, *prev);
prev             1333 drivers/edac/i7core_edac.c 		pci_dev_get(*prev);	/* pci_get_device will put it */
prev             1336 drivers/edac/i7core_edac.c 				      *prev);
prev             1340 drivers/edac/i7core_edac.c 		if (*prev) {
prev             1341 drivers/edac/i7core_edac.c 			*prev = pdev;
prev             1418 drivers/edac/i7core_edac.c 	*prev = pdev;
prev              734 drivers/edac/sb_edac.c 					   struct sbridge_dev *prev)
prev              747 drivers/edac/sb_edac.c 	sbridge_dev = list_entry(prev ? prev->list.next
prev             2329 drivers/edac/sb_edac.c static int sbridge_get_onedevice(struct pci_dev **prev,
prev             2347 drivers/edac/sb_edac.c 			      dev_descr->dev_id, *prev);
prev             2350 drivers/edac/sb_edac.c 		if (*prev) {
prev             2351 drivers/edac/sb_edac.c 			*prev = pdev;
prev             2431 drivers/edac/sb_edac.c 	*prev = pdev;
prev               77 drivers/edac/skx_base.c 	struct pci_dev *pdev, *prev;
prev               82 drivers/edac/skx_base.c 	prev = NULL;
prev               84 drivers/edac/skx_base.c 		pdev = pci_get_device(PCI_VENDOR_ID_INTEL, m->did, prev);
prev              139 drivers/edac/skx_base.c 		prev = pdev;
prev              186 drivers/edac/skx_common.c 	struct pci_dev *pdev, *prev;
prev              191 drivers/edac/skx_common.c 	prev = NULL;
prev              193 drivers/edac/skx_common.c 		pdev = pci_get_device(PCI_VENDOR_ID_INTEL, did, prev);
prev              223 drivers/edac/skx_common.c 		prev = pdev;
prev              215 drivers/firewire/core-topology.c 			h = h->prev;
prev              279 drivers/firewire/core-topology.c 		__list_del(h->prev, &stack);
prev              325 drivers/firewire/net.c 			fi2 = list_entry(fi->fi_link.prev,
prev              345 drivers/firewire/net.c 			list = fi->fi_link.prev;
prev              148 drivers/firewire/ohci.c 	struct descriptor *prev;
prev             1161 drivers/firewire/ohci.c 	ctx->prev = ctx->buffer_tail->buffer;
prev             1235 drivers/firewire/ohci.c 	d_branch = find_branch_descriptor(ctx->prev, ctx->prev_z);
prev             1248 drivers/firewire/ohci.c 	    d_branch != ctx->prev &&
prev             1249 drivers/firewire/ohci.c 	    (ctx->prev->control & cpu_to_le16(DESCRIPTOR_CMD)) ==
prev             1251 drivers/firewire/ohci.c 		ctx->prev->branch_address = cpu_to_le32(d_bus | z);
prev             1254 drivers/firewire/ohci.c 	ctx->prev = d;
prev              341 drivers/firmware/efi/libstub/arm-stub.c 	efi_memory_desc_t *in, *prev = NULL, *out = runtime_map;
prev              357 drivers/firmware/efi/libstub/arm-stub.c 	for (l = 0; l < map_size; l += desc_size, prev = in) {
prev              378 drivers/firmware/efi/libstub/arm-stub.c 		     !regions_are_adjacent(prev, in)) ||
prev              379 drivers/firmware/efi/libstub/arm-stub.c 		    !regions_have_compatible_memory_type_attrs(prev, in)) {
prev             1087 drivers/firmware/efi/vars.c 			struct efivar_entry **prev)
prev             1092 drivers/firmware/efi/vars.c 	if (!prev || !*prev) {
prev             1099 drivers/firmware/efi/vars.c 		if (prev)
prev             1100 drivers/firmware/efi/vars.c 			*prev = entry;
prev             1106 drivers/firmware/efi/vars.c 	list_for_each_entry_safe_continue((*prev), n, head, list) {
prev             1107 drivers/firmware/efi/vars.c 		err = func(*prev, data);
prev              257 drivers/gpio/gpiolib.c 	struct gpio_device *prev, *next;
prev              272 drivers/gpio/gpiolib.c 	prev = list_entry(gpio_devices.prev, struct gpio_device, list);
prev              273 drivers/gpio/gpiolib.c 	if (prev->base + prev->ngpio <= gdev->base) {
prev              279 drivers/gpio/gpiolib.c 	list_for_each_entry_safe(prev, next, &gpio_devices, list) {
prev              285 drivers/gpio/gpiolib.c 		if (prev->base + prev->ngpio <= gdev->base
prev              287 drivers/gpio/gpiolib.c 			list_add(&gdev->list, &prev->list);
prev               97 drivers/gpu/drm/amd/amdgpu/amdgpu_pmu.c 	u64 count, prev;
prev              100 drivers/gpu/drm/amd/amdgpu/amdgpu_pmu.c 		prev = local64_read(&hwc->prev_count);
prev              111 drivers/gpu/drm/amd/amdgpu/amdgpu_pmu.c 	} while (local64_cmpxchg(&hwc->prev_count, prev, count) != prev);
prev              113 drivers/gpu/drm/amd/amdgpu/amdgpu_pmu.c 	local64_add(count - prev, &event->count);
prev              106 drivers/gpu/drm/amd/amdgpu/amdgpu_sa.c 		sa_manager->hole = sa_bo->olist.prev;
prev              265 drivers/gpu/drm/amd/amdgpu/amdgpu_sa.c 		sa_manager->hole = best_bo->olist.prev;
prev               42 drivers/gpu/drm/amd/powerplay/inc/power_state.h 	struct pp_power_state *prev;
prev               62 drivers/gpu/drm/drm_lock.c 	unsigned int old, new, prev;
prev               75 drivers/gpu/drm/drm_lock.c 		prev = cmpxchg(lock, old, new);
prev               76 drivers/gpu/drm/drm_lock.c 	} while (prev != old);
prev              111 drivers/gpu/drm/drm_lock.c 	unsigned int old, new, prev;
prev              118 drivers/gpu/drm/drm_lock.c 		prev = cmpxchg(lock, old, new);
prev              119 drivers/gpu/drm/drm_lock.c 	} while (prev != old);
prev              126 drivers/gpu/drm/drm_lock.c 	unsigned int old, new, prev;
prev              141 drivers/gpu/drm/drm_lock.c 		prev = cmpxchg(lock, old, new);
prev              142 drivers/gpu/drm/drm_lock.c 	} while (prev != old);
prev              312 drivers/gpu/drm/drm_lock.c 	unsigned int old, prev;
prev              320 drivers/gpu/drm/drm_lock.c 				prev = cmpxchg(lock, old, DRM_KERNEL_CONTEXT);
prev              321 drivers/gpu/drm/drm_lock.c 			} while (prev != old);
prev              229 drivers/gpu/drm/drm_syncobj.c 	struct dma_fence *prev;
prev              235 drivers/gpu/drm/drm_syncobj.c 	prev = drm_syncobj_fence_get(syncobj);
prev              237 drivers/gpu/drm/drm_syncobj.c 	if (prev && prev->seqno >= point)
prev              239 drivers/gpu/drm/drm_syncobj.c 	dma_fence_chain_init(chain, prev, fence, point);
prev              247 drivers/gpu/drm/drm_syncobj.c 	dma_fence_chain_for_each(fence, prev);
prev              248 drivers/gpu/drm/drm_syncobj.c 	dma_fence_put(prev);
prev              365 drivers/gpu/drm/exynos/exynos_drm_g2d.c 	lnode = list_entry(file_priv->inuse_cmdlist.prev,
prev              254 drivers/gpu/drm/i915/gt/intel_engine.h 				       u32 next, u32 prev)
prev              257 drivers/gpu/drm/i915/gt/intel_engine.h 	typecheck(typeof(ring->size), prev);
prev              258 drivers/gpu/drm/i915/gt/intel_engine.h 	return (next - prev) << ring->wrap;
prev             1170 drivers/gpu/drm/i915/gt/intel_engine_cs.c 	const void *prev = NULL;
prev             1177 drivers/gpu/drm/i915/gt/intel_engine_cs.c 		if (prev && !memcmp(prev, buf + pos, rowsize)) {
prev             1191 drivers/gpu/drm/i915/gt/intel_engine_cs.c 		prev = buf + pos;
prev              198 drivers/gpu/drm/i915/gt/intel_engine_user.c 	struct rb_node **p, *prev;
prev              203 drivers/gpu/drm/i915/gt/intel_engine_user.c 	prev = NULL;
prev              224 drivers/gpu/drm/i915/gt/intel_engine_user.c 		rb_link_node(&engine->uabi_node, prev, p);
prev              234 drivers/gpu/drm/i915/gt/intel_engine_user.c 		prev = &engine->uabi_node;
prev              235 drivers/gpu/drm/i915/gt/intel_engine_user.c 		p = &prev->rb_right;
prev              387 drivers/gpu/drm/i915/gt/intel_lrc.c assert_priority_queue(const struct i915_request *prev,
prev              397 drivers/gpu/drm/i915/gt/intel_lrc.c 	if (i915_request_is_active(prev))
prev              400 drivers/gpu/drm/i915/gt/intel_lrc.c 	return rq_prio(prev) >= rq_prio(next);
prev              650 drivers/gpu/drm/i915/gt/intel_lrc.c 	u32 tail, prev;
prev              670 drivers/gpu/drm/i915/gt/intel_lrc.c 	prev = ce->lrc_reg_state[CTX_RING_TAIL + 1];
prev              671 drivers/gpu/drm/i915/gt/intel_lrc.c 	if (unlikely(intel_ring_direction(rq->ring, tail, prev) <= 0))
prev              803 drivers/gpu/drm/i915/gt/intel_lrc.c static bool can_merge_ctx(const struct intel_context *prev,
prev              806 drivers/gpu/drm/i915/gt/intel_lrc.c 	if (prev != next)
prev              809 drivers/gpu/drm/i915/gt/intel_lrc.c 	if (ctx_single_port_submission(prev))
prev              815 drivers/gpu/drm/i915/gt/intel_lrc.c static bool can_merge_rq(const struct i915_request *prev,
prev              818 drivers/gpu/drm/i915/gt/intel_lrc.c 	GEM_BUG_ON(prev == next);
prev              819 drivers/gpu/drm/i915/gt/intel_lrc.c 	GEM_BUG_ON(!assert_priority_queue(prev, next));
prev              832 drivers/gpu/drm/i915/gt/intel_lrc.c 	if (!can_merge_ctx(prev->hw_context, next->hw_context))
prev               15 drivers/gpu/drm/i915/gt/selftest_engine_cs.c 		u8 prev = U8_MAX;
prev               21 drivers/gpu/drm/i915/gt/selftest_engine_cs.c 			if (gen >= prev) {
prev               26 drivers/gpu/drm/i915/gt/selftest_engine_cs.c 				       prev, gen);
prev               42 drivers/gpu/drm/i915/gt/selftest_engine_cs.c 			prev = gen;
prev               49 drivers/gpu/drm/i915/gt/selftest_engine_cs.c 			 prev);
prev             1412 drivers/gpu/drm/i915/gt/selftest_hangcheck.c 		struct i915_request *prev;
prev             1419 drivers/gpu/drm/i915/gt/selftest_hangcheck.c 		prev = hang_create_request(&h, engine);
prev             1420 drivers/gpu/drm/i915/gt/selftest_hangcheck.c 		if (IS_ERR(prev)) {
prev             1421 drivers/gpu/drm/i915/gt/selftest_hangcheck.c 			err = PTR_ERR(prev);
prev             1425 drivers/gpu/drm/i915/gt/selftest_hangcheck.c 		i915_request_get(prev);
prev             1426 drivers/gpu/drm/i915/gt/selftest_hangcheck.c 		i915_request_add(prev);
prev             1457 drivers/gpu/drm/i915/gt/selftest_hangcheck.c 				i915_request_put(prev);
prev             1464 drivers/gpu/drm/i915/gt/selftest_hangcheck.c 			if (!wait_until_running(&h, prev)) {
prev             1469 drivers/gpu/drm/i915/gt/selftest_hangcheck.c 				       prev->fence.seqno, hws_seqno(&h, prev));
prev             1474 drivers/gpu/drm/i915/gt/selftest_hangcheck.c 				i915_request_put(prev);
prev             1484 drivers/gpu/drm/i915/gt/selftest_hangcheck.c 			if (prev->fence.error != -EIO) {
prev             1486 drivers/gpu/drm/i915/gt/selftest_hangcheck.c 				       prev->fence.error);
prev             1488 drivers/gpu/drm/i915/gt/selftest_hangcheck.c 				i915_request_put(prev);
prev             1497 drivers/gpu/drm/i915/gt/selftest_hangcheck.c 				i915_request_put(prev);
prev             1505 drivers/gpu/drm/i915/gt/selftest_hangcheck.c 				i915_request_put(prev);
prev             1510 drivers/gpu/drm/i915/gt/selftest_hangcheck.c 			i915_request_put(prev);
prev             1511 drivers/gpu/drm/i915/gt/selftest_hangcheck.c 			prev = rq;
prev             1519 drivers/gpu/drm/i915/gt/selftest_hangcheck.c 		i915_request_put(prev);
prev               58 drivers/gpu/drm/i915/i915_active.c 	return (struct intel_engine_cs *)READ_ONCE(node->base.link.prev);
prev              502 drivers/gpu/drm/i915/i915_active.c 	struct rb_node *prev, *p;
prev              522 drivers/gpu/drm/i915/i915_active.c 	prev = NULL;
prev              531 drivers/gpu/drm/i915/i915_active.c 		prev = p;
prev              544 drivers/gpu/drm/i915/i915_active.c 	for (p = prev; p; p = rb_next(p)) {
prev              634 drivers/gpu/drm/i915/i915_active.c 			node->base.link.prev = (void *)engine;
prev             1132 drivers/gpu/drm/i915/i915_irq.c 	const struct intel_rps_ei *prev = &rps->ei;
prev             1141 drivers/gpu/drm/i915/i915_irq.c 	if (prev->ktime) {
prev             1145 drivers/gpu/drm/i915/i915_irq.c 		time = ktime_us_delta(now.ktime, prev->ktime);
prev             1154 drivers/gpu/drm/i915/i915_irq.c 		render = now.render_c0 - prev->render_c0;
prev             1155 drivers/gpu/drm/i915/i915_irq.c 		media = now.media_c0 - prev->media_c0;
prev              577 drivers/gpu/drm/i915/i915_pmu.c 	u64 prev, new;
prev              580 drivers/gpu/drm/i915/i915_pmu.c 	prev = local64_read(&hwc->prev_count);
prev              583 drivers/gpu/drm/i915/i915_pmu.c 	if (local64_cmpxchg(&hwc->prev_count, prev, new) != prev)
prev              586 drivers/gpu/drm/i915/i915_pmu.c 	local64_add(new - prev, &event->count);
prev             1129 drivers/gpu/drm/i915/i915_request.c 	struct i915_request *prev;
prev             1151 drivers/gpu/drm/i915/i915_request.c 	prev = rcu_dereference_protected(timeline->last_request.request,
prev             1153 drivers/gpu/drm/i915/i915_request.c 	if (prev && !i915_request_completed(prev)) {
prev             1154 drivers/gpu/drm/i915/i915_request.c 		if (is_power_of_2(prev->engine->mask | rq->engine->mask))
prev             1156 drivers/gpu/drm/i915/i915_request.c 						     &prev->submit,
prev             1160 drivers/gpu/drm/i915/i915_request.c 							&prev->fence,
prev             1164 drivers/gpu/drm/i915/i915_request.c 							 &prev->sched,
prev             1179 drivers/gpu/drm/i915/i915_request.c 	return prev;
prev             1242 drivers/gpu/drm/i915/i915_request.c 	struct i915_request *prev;
prev             1249 drivers/gpu/drm/i915/i915_request.c 	prev = __i915_request_commit(rq);
prev             1296 drivers/gpu/drm/i915/i915_request.c 	if (prev && i915_request_completed(prev) && prev->timeline == tl)
prev             1297 drivers/gpu/drm/i915/i915_request.c 		i915_request_retire_upto(prev);
prev              305 drivers/gpu/drm/i915/i915_scheduler.c 		if (stack.dfs_link.next == stack.dfs_link.prev)
prev              244 drivers/gpu/drm/i915/i915_utils.h 	first->prev = head;
prev              119 drivers/gpu/drm/i915/selftests/i915_buddy.c 	struct i915_buddy_block *prev;
prev              124 drivers/gpu/drm/i915/selftests/i915_buddy.c 	prev = NULL;
prev              135 drivers/gpu/drm/i915/selftests/i915_buddy.c 		if (is_contiguous && prev) {
prev              140 drivers/gpu/drm/i915/selftests/i915_buddy.c 			prev_offset = i915_buddy_block_offset(prev);
prev              141 drivers/gpu/drm/i915/selftests/i915_buddy.c 			prev_block_size = i915_buddy_block_size(mm, prev);
prev              154 drivers/gpu/drm/i915/selftests/i915_buddy.c 		prev = block;
prev              166 drivers/gpu/drm/i915/selftests/i915_buddy.c 	if (prev) {
prev              168 drivers/gpu/drm/i915/selftests/i915_buddy.c 		igt_dump_block(mm, prev);
prev              182 drivers/gpu/drm/i915/selftests/i915_buddy.c 	struct i915_buddy_block *prev;
prev              199 drivers/gpu/drm/i915/selftests/i915_buddy.c 	prev = NULL;
prev              229 drivers/gpu/drm/i915/selftests/i915_buddy.c 		if (prev) {
prev              234 drivers/gpu/drm/i915/selftests/i915_buddy.c 			prev_offset = i915_buddy_block_offset(prev);
prev              235 drivers/gpu/drm/i915/selftests/i915_buddy.c 			prev_block_size = i915_buddy_block_size(mm, prev);
prev              255 drivers/gpu/drm/i915/selftests/i915_buddy.c 		prev = root;
prev              268 drivers/gpu/drm/i915/selftests/i915_buddy.c 	if (prev) {
prev              270 drivers/gpu/drm/i915/selftests/i915_buddy.c 		igt_dump_block(mm, prev);
prev              939 drivers/gpu/drm/i915/selftests/i915_request.c 	struct i915_request *prev = NULL;
prev              978 drivers/gpu/drm/i915/selftests/i915_request.c 		if (prev) {
prev              980 drivers/gpu/drm/i915/selftests/i915_request.c 							   &prev->fence);
prev             1006 drivers/gpu/drm/i915/selftests/i915_request.c 		prev = request[id];
prev               32 drivers/gpu/drm/i915/selftests/intel_uncore.c 	s32 prev;
prev               34 drivers/gpu/drm/i915/selftests/intel_uncore.c 	for (i = 0, prev = -1; i < num_ranges; i++, ranges++) {
prev               36 drivers/gpu/drm/i915/selftests/intel_uncore.c 		if (is_watertight && (prev + 1) != (s32)ranges->start) {
prev               38 drivers/gpu/drm/i915/selftests/intel_uncore.c 			       __func__, i, ranges->start, ranges->end, prev);
prev               43 drivers/gpu/drm/i915/selftests/intel_uncore.c 		if (prev >= (s32)ranges->start) {
prev               45 drivers/gpu/drm/i915/selftests/intel_uncore.c 			       __func__, i, ranges->start, ranges->end, prev);
prev               56 drivers/gpu/drm/i915/selftests/intel_uncore.c 		prev = ranges->end;
prev               73 drivers/gpu/drm/i915/selftests/intel_uncore.c 	s32 prev;
prev               77 drivers/gpu/drm/i915/selftests/intel_uncore.c 		for (i = 0, prev = -1; i < reg_lists[j].size; i++, reg++) {
prev               80 drivers/gpu/drm/i915/selftests/intel_uncore.c 			if (prev >= (s32)offset) {
prev               82 drivers/gpu/drm/i915/selftests/intel_uncore.c 				       __func__, i, offset, prev);
prev               86 drivers/gpu/drm/i915/selftests/intel_uncore.c 			prev = offset;
prev              268 drivers/gpu/drm/mga/mga_dma.c 		entry->prev = dev_priv->head;
prev              273 drivers/gpu/drm/mga/mga_dma.c 			dev_priv->head->next->prev = entry;
prev              326 drivers/gpu/drm/mga/mga_dma.c 	drm_mga_freelist_t *prev;
prev              342 drivers/gpu/drm/mga/mga_dma.c 		prev = dev_priv->tail->prev;
prev              344 drivers/gpu/drm/mga/mga_dma.c 		prev->next = NULL;
prev              345 drivers/gpu/drm/mga/mga_dma.c 		next->prev = next->next = NULL;
prev              346 drivers/gpu/drm/mga/mga_dma.c 		dev_priv->tail = prev;
prev              359 drivers/gpu/drm/mga/mga_dma.c 	drm_mga_freelist_t *head, *entry, *prev;
prev              371 drivers/gpu/drm/mga/mga_dma.c 		prev = dev_priv->tail;
prev              372 drivers/gpu/drm/mga/mga_dma.c 		prev->next = entry;
prev              373 drivers/gpu/drm/mga/mga_dma.c 		entry->prev = prev;
prev              376 drivers/gpu/drm/mga/mga_dma.c 		prev = head->next;
prev              378 drivers/gpu/drm/mga/mga_dma.c 		prev->prev = entry;
prev              379 drivers/gpu/drm/mga/mga_dma.c 		entry->prev = head;
prev              380 drivers/gpu/drm/mga/mga_dma.c 		entry->next = prev;
prev               81 drivers/gpu/drm/mga/mga_drv.h 	struct drm_mga_freelist *prev;
prev              111 drivers/gpu/drm/nouveau/include/nvif/list.h     struct list_head *next, *prev;
prev              130 drivers/gpu/drm/nouveau/include/nvif/list.h     list->next = list->prev = list;
prev              135 drivers/gpu/drm/nouveau/include/nvif/list.h                 struct list_head *prev, struct list_head *next)
prev              137 drivers/gpu/drm/nouveau/include/nvif/list.h     next->prev = entry;
prev              139 drivers/gpu/drm/nouveau/include/nvif/list.h     entry->prev = prev;
prev              140 drivers/gpu/drm/nouveau/include/nvif/list.h     prev->next = entry;
prev              182 drivers/gpu/drm/nouveau/include/nvif/list.h     __list_add(entry, head->prev, head);
prev              186 drivers/gpu/drm/nouveau/include/nvif/list.h __list_del(struct list_head *prev, struct list_head *next)
prev              188 drivers/gpu/drm/nouveau/include/nvif/list.h     next->prev = prev;
prev              189 drivers/gpu/drm/nouveau/include/nvif/list.h     prev->next = next;
prev              209 drivers/gpu/drm/nouveau/include/nvif/list.h     __list_del(entry->prev, entry->next);
prev              215 drivers/gpu/drm/nouveau/include/nvif/list.h     __list_del(entry->prev, entry->next);
prev              222 drivers/gpu/drm/nouveau/include/nvif/list.h 	__list_del(list->prev, list->next);
prev              292 drivers/gpu/drm/nouveau/include/nvif/list.h     list_entry((ptr)->prev, type, member)
prev              334 drivers/gpu/drm/nouveau/include/nvif/list.h 	for (pos = __container_of((head)->prev, pos, member);		\
prev              336 drivers/gpu/drm/nouveau/include/nvif/list.h 	     pos = __container_of(pos->member.prev, pos, member))
prev              344 drivers/gpu/drm/nouveau/include/nvif/list.h 	for (pos = __container_of(pos->member.prev, pos, member);	\
prev              346 drivers/gpu/drm/nouveau/include/nvif/list.h 	     pos = __container_of(pos->member.prev, pos, member))
prev               53 drivers/gpu/drm/nouveau/nouveau_bo.h 	struct nouveau_bo *prev;
prev               57 drivers/gpu/drm/nouveau/nouveau_bo.h 	prev = *pnvbo;
prev               65 drivers/gpu/drm/nouveau/nouveau_bo.h 	if (prev)
prev               66 drivers/gpu/drm/nouveau/nouveau_bo.h 		ttm_bo_put(&prev->bo);
prev              353 drivers/gpu/drm/nouveau/nouveau_fence.c 		struct nouveau_channel *prev = NULL;
prev              359 drivers/gpu/drm/nouveau/nouveau_fence.c 			prev = rcu_dereference(f->channel);
prev              360 drivers/gpu/drm/nouveau/nouveau_fence.c 			if (prev && (prev == chan || fctx->sync(f, prev, chan) == 0))
prev              375 drivers/gpu/drm/nouveau/nouveau_fence.c 		struct nouveau_channel *prev = NULL;
prev              384 drivers/gpu/drm/nouveau/nouveau_fence.c 			prev = rcu_dereference(f->channel);
prev              385 drivers/gpu/drm/nouveau/nouveau_fence.c 			if (prev && (prev == chan || fctx->sync(f, prev, chan) == 0))
prev               54 drivers/gpu/drm/nouveau/nv04_fence.c 		struct nouveau_channel *prev, struct nouveau_channel *chan)
prev               45 drivers/gpu/drm/nouveau/nv10_fence.c 		struct nouveau_channel *prev, struct nouveau_channel *chan)
prev               35 drivers/gpu/drm/nouveau/nv17_fence.c 		struct nouveau_channel *prev, struct nouveau_channel *chan)
prev               37 drivers/gpu/drm/nouveau/nv17_fence.c 	struct nouveau_cli *cli = (void *)prev->user.client;
prev               51 drivers/gpu/drm/nouveau/nv17_fence.c 	ret = RING_SPACE(prev, 5);
prev               53 drivers/gpu/drm/nouveau/nv17_fence.c 		BEGIN_NV04(prev, 0, NV11_SUBCHAN_DMA_SEMAPHORE, 4);
prev               54 drivers/gpu/drm/nouveau/nv17_fence.c 		OUT_RING  (prev, fctx->sema.handle);
prev               55 drivers/gpu/drm/nouveau/nv17_fence.c 		OUT_RING  (prev, 0);
prev               56 drivers/gpu/drm/nouveau/nv17_fence.c 		OUT_RING  (prev, value + 0);
prev               57 drivers/gpu/drm/nouveau/nv17_fence.c 		OUT_RING  (prev, value + 1);
prev               58 drivers/gpu/drm/nouveau/nv17_fence.c 		FIRE_RING (prev);
prev               79 drivers/gpu/drm/nouveau/nv84_fence.c 		struct nouveau_channel *prev, struct nouveau_channel *chan)
prev               82 drivers/gpu/drm/nouveau/nv84_fence.c 	u64 addr = fctx->vma->addr + prev->chid * 16;
prev               53 drivers/gpu/drm/nouveau/nvkm/core/mm.c 		struct nvkm_mm_node *prev = node(this, prev);
prev               56 drivers/gpu/drm/nouveau/nvkm/core/mm.c 		if (prev && prev->type == NVKM_MM_TYPE_NONE) {
prev               57 drivers/gpu/drm/nouveau/nvkm/core/mm.c 			prev->length += this->length;
prev               59 drivers/gpu/drm/nouveau/nvkm/core/mm.c 			kfree(this); this = prev;
prev               72 drivers/gpu/drm/nouveau/nvkm/core/mm.c 			list_for_each_entry(prev, &mm->free, fl_entry) {
prev               73 drivers/gpu/drm/nouveau/nvkm/core/mm.c 				if (this->offset < prev->offset)
prev               77 drivers/gpu/drm/nouveau/nvkm/core/mm.c 			list_add_tail(&this->fl_entry, &prev->fl_entry);
prev              114 drivers/gpu/drm/nouveau/nvkm/core/mm.c 	struct nvkm_mm_node *prev, *this, *next;
prev              129 drivers/gpu/drm/nouveau/nvkm/core/mm.c 		prev = node(this, prev);
prev              130 drivers/gpu/drm/nouveau/nvkm/core/mm.c 		if (prev && prev->type != type)
prev              189 drivers/gpu/drm/nouveau/nvkm/core/mm.c 	struct nvkm_mm_node *prev, *this, *next;
prev              203 drivers/gpu/drm/nouveau/nvkm/core/mm.c 		prev = node(this, prev);
prev              204 drivers/gpu/drm/nouveau/nvkm/core/mm.c 		if (prev && prev->type != type)
prev              242 drivers/gpu/drm/nouveau/nvkm/core/mm.c 	struct nvkm_mm_node *node, *prev;
prev              246 drivers/gpu/drm/nouveau/nvkm/core/mm.c 		prev = list_last_entry(&mm->nodes, typeof(*node), nl_entry);
prev              247 drivers/gpu/drm/nouveau/nvkm/core/mm.c 		next = prev->offset + prev->length;
prev               60 drivers/gpu/drm/nouveau/nvkm/engine/disp/baseg84.c 	.prev = 0x000004,
prev               94 drivers/gpu/drm/nouveau/nvkm/engine/disp/basegf119.c 	.prev = -0x020000,
prev              105 drivers/gpu/drm/nouveau/nvkm/engine/disp/basenv50.c 	.prev = 0x000004,
prev               48 drivers/gpu/drm/nouveau/nvkm/engine/disp/channv50.c 			u32 prev = nvkm_rd32(device, list->data[i].addr + base + c);
prev               53 drivers/gpu/drm/nouveau/nvkm/engine/disp/channv50.c 			if (prev != next)
prev               60 drivers/gpu/drm/nouveau/nvkm/engine/disp/channv50.c 				     mthd, prev, mods, name ? " // " : "",
prev              100 drivers/gpu/drm/nouveau/nvkm/engine/disp/channv50.c 			nv50_disp_mthd_list(disp, debug, base, mthd->prev,
prev              162 drivers/gpu/drm/nouveau/nvkm/engine/disp/channv50.h 	s32 prev;
prev               94 drivers/gpu/drm/nouveau/nvkm/engine/disp/coreg84.c 	.prev = 0x000004,
prev               40 drivers/gpu/drm/nouveau/nvkm/engine/disp/coreg94.c 	.prev = 0x000004,
prev              158 drivers/gpu/drm/nouveau/nvkm/engine/disp/coregf119.c 	.prev = -0x020000,
prev              109 drivers/gpu/drm/nouveau/nvkm/engine/disp/coregk104.c 	.prev = -0x020000,
prev              125 drivers/gpu/drm/nouveau/nvkm/engine/disp/coregv100.c 	.prev = 0x008000,
prev              155 drivers/gpu/drm/nouveau/nvkm/engine/disp/corenv50.c 	.prev = 0x000004,
prev               58 drivers/gpu/drm/nouveau/nvkm/engine/disp/ovlyg84.c 	.prev = 0x000004,
prev               82 drivers/gpu/drm/nouveau/nvkm/engine/disp/ovlygf119.c 	.prev = -0x020000,
prev               84 drivers/gpu/drm/nouveau/nvkm/engine/disp/ovlygk104.c 	.prev = -0x020000,
prev               61 drivers/gpu/drm/nouveau/nvkm/engine/disp/ovlygt200.c 	.prev = 0x000004,
prev               94 drivers/gpu/drm/nouveau/nvkm/engine/disp/ovlynv50.c 	.prev = 0x000004,
prev              123 drivers/gpu/drm/nouveau/nvkm/engine/disp/wndwgv100.c 	.prev = 0x000800,
prev               48 drivers/gpu/drm/nouveau/nvkm/engine/fifo/gk104.c 	} prev, next, *chan;
prev               67 drivers/gpu/drm/nouveau/nvkm/engine/fifo/gk104.c 	status->prev.tsg = !!(stat & 0x00001000);
prev               68 drivers/gpu/drm/nouveau/nvkm/engine/fifo/gk104.c 	status->prev.id  =   (stat & 0x00000fff);
prev               76 drivers/gpu/drm/nouveau/nvkm/engine/fifo/gk104.c 				status->chan = &status->prev;
prev               81 drivers/gpu/drm/nouveau/nvkm/engine/fifo/gk104.c 			status->chan = &status->prev;
prev               85 drivers/gpu/drm/nouveau/nvkm/engine/fifo/gk104.c 		status->chan = &status->prev;
prev               92 drivers/gpu/drm/nouveau/nvkm/engine/fifo/gk104.c 		   status->prev.tsg ? "tsg" : "ch", status->prev.id,
prev               93 drivers/gpu/drm/nouveau/nvkm/engine/fifo/gk104.c 		   status->chan == &status->prev ? "*" : " ",
prev             1653 drivers/gpu/drm/nouveau/nvkm/engine/gr/gf100.c 	u32 addr = ~0, prev = ~0, xfer = 0;
prev             1667 drivers/gpu/drm/nouveau/nvkm/engine/gr/gf100.c 			if (head != prev + 4 || xfer >= 32) {
prev             1676 drivers/gpu/drm/nouveau/nvkm/engine/gr/gf100.c 			prev = head;
prev             1117 drivers/gpu/drm/nouveau/nvkm/engine/gr/nv04.c 	struct nv04_gr_chan *prev = NULL;
prev             1124 drivers/gpu/drm/nouveau/nvkm/engine/gr/nv04.c 	prev = nv04_gr_channel(gr);
prev             1125 drivers/gpu/drm/nouveau/nvkm/engine/gr/nv04.c 	if (prev)
prev             1126 drivers/gpu/drm/nouveau/nvkm/engine/gr/nv04.c 		nv04_gr_unload_context(prev);
prev              935 drivers/gpu/drm/nouveau/nvkm/engine/gr/nv10.c 	struct nv10_gr_chan *prev = NULL;
prev              942 drivers/gpu/drm/nouveau/nvkm/engine/gr/nv10.c 	prev = nv10_gr_channel(gr);
prev              943 drivers/gpu/drm/nouveau/nvkm/engine/gr/nv10.c 	if (prev)
prev              944 drivers/gpu/drm/nouveau/nvkm/engine/gr/nv10.c 		nv10_gr_unload_context(prev);
prev              116 drivers/gpu/drm/nouveau/nvkm/subdev/bios/base.c 		u8 *prev = bios->data;
prev              118 drivers/gpu/drm/nouveau/nvkm/subdev/bios/base.c 			bios->data = prev;
prev              121 drivers/gpu/drm/nouveau/nvkm/subdev/bios/base.c 		memcpy(bios->data, prev, bios->size);
prev              123 drivers/gpu/drm/nouveau/nvkm/subdev/bios/base.c 		kfree(prev);
prev              242 drivers/gpu/drm/nouveau/nvkm/subdev/fb/ramgk104.c 			u32 prev = nvkm_rd32(device, addr);
prev              243 drivers/gpu/drm/nouveau/nvkm/subdev/fb/ramgk104.c 			u32 next = (prev & ~mask) | data;
prev              883 drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmm.c nvkm_vmm_node_merge(struct nvkm_vmm *vmm, struct nvkm_vma *prev,
prev              890 drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmm.c 			if (prev) {
prev              891 drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmm.c 				prev->size += vma->size;
prev              893 drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmm.c 				return prev;
prev              897 drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmm.c 		BUG_ON(prev);
prev              907 drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmm.c 	if (prev) {
prev              910 drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmm.c 			prev->size += size;
prev              915 drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmm.c 			prev->size += vma->size;
prev              918 drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmm.c 		return prev;
prev              928 drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmm.c 	struct nvkm_vma *prev = NULL;
prev              931 drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmm.c 		prev = vma;
prev              941 drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmm.c 			nvkm_vmm_node_merge(vmm, prev, vma, NULL, vma->size);
prev             1154 drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmm.c 	struct nvkm_vma *prev = NULL;
prev             1157 drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmm.c 	if (vma->addr == addr && vma->part && (prev = node(vma, prev))) {
prev             1158 drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmm.c 		if (prev->memory || prev->mapped != map)
prev             1159 drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmm.c 			prev = NULL;
prev             1168 drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmm.c 	if (prev || next)
prev             1169 drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmm.c 		return nvkm_vmm_node_merge(vmm, prev, vma, next, size);
prev             1336 drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmm.c 	struct nvkm_vma *prev = NULL;
prev             1343 drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmm.c 	if (vma->part && (prev = node(vma, prev)) && prev->mapped)
prev             1344 drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmm.c 		prev = NULL;
prev             1347 drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmm.c 	nvkm_vmm_node_merge(vmm, prev, vma, next, vma->size);
prev             1524 drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmm.c 	struct nvkm_vma *prev, *next;
prev             1526 drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmm.c 	if ((prev = node(vma, prev)) && !prev->used) {
prev             1527 drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmm.c 		vma->addr  = prev->addr;
prev             1528 drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmm.c 		vma->size += prev->size;
prev             1529 drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmm.c 		nvkm_vmm_free_delete(vmm, prev);
prev             1706 drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmm.c 		struct nvkm_vma *prev = node(this, prev);
prev             1711 drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmm.c 		if (vmm->func->page_block && prev && prev->page != p)
prev              129 drivers/gpu/drm/nouveau/nvkm/subdev/therm/fan.c 	u32 cycles, cur, prev;
prev              143 drivers/gpu/drm/nouveau/nvkm/subdev/therm/fan.c 	prev = nvkm_gpio_get(gpio, 0, therm->fan->tach.func,
prev              151 drivers/gpu/drm/nouveau/nvkm/subdev/therm/fan.c 		if (prev != cur) {
prev              155 drivers/gpu/drm/nouveau/nvkm/subdev/therm/fan.c 			prev = cur;
prev              170 drivers/gpu/drm/nouveau/nvkm/subdev/volt/base.c 		int prev = nvkm_volt_get(volt);
prev              171 drivers/gpu/drm/nouveau/nvkm/subdev/volt/base.c 		if (!condition || prev < 0 ||
prev              172 drivers/gpu/drm/nouveau/nvkm/subdev/volt/base.c 		    (condition < 0 && ret < prev) ||
prev              173 drivers/gpu/drm/nouveau/nvkm/subdev/volt/base.c 		    (condition > 0 && ret > prev)) {
prev              798 drivers/gpu/drm/r128/r128_cce.c 		entry->prev = dev_priv->head;
prev              810 drivers/gpu/drm/r128/r128_cce.c 			dev_priv->head->next->prev = entry;
prev               71 drivers/gpu/drm/r128/r128_drv.h 	struct drm_r128_freelist *prev;
prev               43 drivers/gpu/drm/radeon/mkregtable.c 	struct list_head *next, *prev;
prev               50 drivers/gpu/drm/radeon/mkregtable.c 	list->prev = list;
prev               61 drivers/gpu/drm/radeon/mkregtable.c 			      struct list_head *prev, struct list_head *next)
prev               63 drivers/gpu/drm/radeon/mkregtable.c 	next->prev = new;
prev               65 drivers/gpu/drm/radeon/mkregtable.c 	new->prev = prev;
prev               66 drivers/gpu/drm/radeon/mkregtable.c 	prev->next = new;
prev               70 drivers/gpu/drm/radeon/mkregtable.c 		       struct list_head *prev, struct list_head *next);
prev               83 drivers/gpu/drm/radeon/mkregtable.c 	__list_add(new, head->prev, head);
prev              146 drivers/gpu/drm/radeon/radeon_sa.c 		sa_manager->hole = sa_bo->olist.prev;
prev              302 drivers/gpu/drm/radeon/radeon_sa.c 		sa_manager->hole = best_bo->olist.prev;
prev              225 drivers/gpu/drm/savage/savage_bci.c 	dev_priv->head.prev = NULL;
prev              229 drivers/gpu/drm/savage/savage_bci.c 	dev_priv->tail.prev = &dev_priv->head;
prev              240 drivers/gpu/drm/savage/savage_bci.c 		entry->prev = &dev_priv->head;
prev              241 drivers/gpu/drm/savage/savage_bci.c 		dev_priv->head.next->prev = entry;
prev              251 drivers/gpu/drm/savage/savage_bci.c 	drm_savage_buf_priv_t *tail = dev_priv->tail.prev;
prev              270 drivers/gpu/drm/savage/savage_bci.c 		drm_savage_buf_priv_t *prev = tail->prev;
prev              271 drivers/gpu/drm/savage/savage_bci.c 		prev->next = next;
prev              272 drivers/gpu/drm/savage/savage_bci.c 		next->prev = prev;
prev              273 drivers/gpu/drm/savage/savage_bci.c 		tail->next = tail->prev = NULL;
prev              284 drivers/gpu/drm/savage/savage_bci.c 	drm_savage_buf_priv_t *entry = buf->dev_private, *prev, *next;
prev              288 drivers/gpu/drm/savage/savage_bci.c 	if (entry->next != NULL || entry->prev != NULL) {
prev              293 drivers/gpu/drm/savage/savage_bci.c 	prev = &dev_priv->head;
prev              294 drivers/gpu/drm/savage/savage_bci.c 	next = prev->next;
prev              295 drivers/gpu/drm/savage/savage_bci.c 	prev->next = entry;
prev              296 drivers/gpu/drm/savage/savage_bci.c 	next->prev = entry;
prev              297 drivers/gpu/drm/savage/savage_bci.c 	entry->prev = prev;
prev             1071 drivers/gpu/drm/savage/savage_bci.c 		    buf_priv->next == NULL && buf_priv->prev == NULL) {
prev               65 drivers/gpu/drm/savage/savage_drv.h 	struct drm_savage_buf_priv *prev;
prev              318 drivers/gpu/drm/ttm/ttm_page_alloc.c 			__list_del(p->lru.prev, &pool->list);
prev              594 drivers/gpu/drm/vmwgfx/vmwgfx_cmdbuf.c 		list_splice_init(&ctx->preempted, restart_head[i].prev);
prev              733 drivers/gpu/drm/vmwgfx/vmwgfx_fence.c 			list_entry(fman->fence_list.prev, struct vmw_fence_obj,
prev               60 drivers/gpu/host1x/intr.c 	struct host1x_waitlist *waiter, *next, *prev;
prev               71 drivers/gpu/host1x/intr.c 			prev = list_entry(dest->prev,
prev               73 drivers/gpu/host1x/intr.c 			if (prev->data == waiter->data) {
prev               74 drivers/gpu/host1x/intr.c 				prev->count++;
prev              137 drivers/greybus/operation.c 	int prev;
prev              148 drivers/greybus/operation.c 		prev = operation->errno;
prev              149 drivers/greybus/operation.c 		if (prev == -EBADR)
prev              154 drivers/greybus/operation.c 		WARN_ON(prev != -EBADR);
prev              172 drivers/greybus/operation.c 	prev = operation->errno;
prev              173 drivers/greybus/operation.c 	if (prev == -EINPROGRESS)
prev              177 drivers/greybus/operation.c 	return prev == -EINPROGRESS;
prev              206 drivers/hid/hid-input.c 	__s32 prev;
prev              231 drivers/hid/hid-input.c 			prev = physical_extents;
prev              233 drivers/hid/hid-input.c 			if (physical_extents < prev)
prev              249 drivers/hid/hid-input.c 			prev = logical_extents;
prev              251 drivers/hid/hid-input.c 			if (logical_extents < prev)
prev              265 drivers/hid/hid-input.c 		prev = logical_extents;
prev              267 drivers/hid/hid-input.c 		if (logical_extents < prev)
prev              272 drivers/hid/hid-input.c 		prev = physical_extents;
prev              274 drivers/hid/hid-input.c 		if (physical_extents < prev)
prev             1080 drivers/hwtracing/intel_th/msu.c 		struct msc_window *prev = list_last_entry(&msc->win_list,
prev             1084 drivers/hwtracing/intel_th/msu.c 		win->pgoff = prev->pgoff + prev->nr_blocks;
prev              122 drivers/infiniband/core/rw.c 	struct rdma_rw_reg_ctx *prev = NULL;
prev              144 drivers/infiniband/core/rw.c 		if (prev) {
prev              146 drivers/infiniband/core/rw.c 				prev->wr.wr.next = &reg->inv_wr;
prev              148 drivers/infiniband/core/rw.c 				prev->wr.wr.next = &reg->reg_wr.wr;
prev              171 drivers/infiniband/core/rw.c 		prev = reg;
prev              175 drivers/infiniband/core/rw.c 	if (prev)
prev              176 drivers/infiniband/core/rw.c 		prev->wr.wr.next = NULL;
prev             4281 drivers/infiniband/hw/cxgb4/cm.c 		tmp->prev = NULL;
prev              134 drivers/infiniband/hw/hfi1/aspm.c 	ktime_t now, prev;
prev              141 drivers/infiniband/hw/hfi1/aspm.c 	prev = rcd->aspm_ts_last_intr;
prev              146 drivers/infiniband/hw/hfi1/aspm.c 	close_interrupts = ktime_to_ns(ktime_sub(now, prev)) < ASPM_TRIGGER_NS;
prev               58 drivers/infiniband/hw/hfi1/rc.c struct rvt_ack_entry *find_prev_entry(struct rvt_qp *qp, u32 psn, u8 *prev,
prev               89 drivers/infiniband/hw/hfi1/rc.c 	if (prev)
prev               90 drivers/infiniband/hw/hfi1/rc.c 		*prev = p;
prev             2559 drivers/infiniband/hw/hfi1/rc.c 	u8 prev;
prev             2607 drivers/infiniband/hw/hfi1/rc.c 	e = find_prev_entry(qp, psn, &prev, &mra, &old_req);
prev             2653 drivers/infiniband/hw/hfi1/rc.c 			qp->s_acked_ack_queue = prev;
prev             2654 drivers/infiniband/hw/hfi1/rc.c 		qp->s_tail_ack_queue = prev;
prev             2668 drivers/infiniband/hw/hfi1/rc.c 			qp->s_acked_ack_queue = prev;
prev             2669 drivers/infiniband/hw/hfi1/rc.c 		qp->s_tail_ack_queue = prev;
prev               52 drivers/infiniband/hw/hfi1/rc.h struct rvt_ack_entry *find_prev_entry(struct rvt_qp *qp, u32 psn, u8 *prev,
prev             1989 drivers/infiniband/hw/hfi1/tid_rdma.c 	u8 prev;
prev             2008 drivers/infiniband/hw/hfi1/tid_rdma.c 	e = find_prev_entry(qp, psn, &prev, NULL, &old_req);
prev             2086 drivers/infiniband/hw/hfi1/tid_rdma.c 			for (i = prev + 1; ; i++) {
prev             2144 drivers/infiniband/hw/hfi1/tid_rdma.c 		for (i = prev + 1; ; i++) {
prev             2177 drivers/infiniband/hw/hfi1/tid_rdma.c 		qp->s_acked_ack_queue = prev;
prev             2178 drivers/infiniband/hw/hfi1/tid_rdma.c 	qp->s_tail_ack_queue = prev;
prev             3203 drivers/infiniband/hw/hfi1/tid_rdma.c 	struct rvt_swqe *prev;
prev             3209 drivers/infiniband/hw/hfi1/tid_rdma.c 	prev = rvt_get_swqe_ptr(qp, s_prev);
prev             3218 drivers/infiniband/hw/hfi1/tid_rdma.c 		switch (prev->wr.opcode) {
prev             3220 drivers/infiniband/hw/hfi1/tid_rdma.c 			req = wqe_to_tid_req(prev);
prev             3228 drivers/infiniband/hw/hfi1/tid_rdma.c 		if (prev->wr.opcode != IB_WR_TID_RDMA_WRITE)
prev             3232 drivers/infiniband/hw/hfi1/tid_rdma.c 		switch (prev->wr.opcode) {
prev             3238 drivers/infiniband/hw/hfi1/tid_rdma.c 			req = wqe_to_tid_req(prev);
prev             5451 drivers/infiniband/hw/hfi1/tid_rdma.c 	struct rvt_ack_entry *prev;
prev             5459 drivers/infiniband/hw/hfi1/tid_rdma.c 	prev = &qp->s_ack_queue[s_prev];
prev             5463 drivers/infiniband/hw/hfi1/tid_rdma.c 	    prev->opcode == TID_OP(WRITE_REQ)) {
prev             5464 drivers/infiniband/hw/hfi1/tid_rdma.c 		req = ack_to_tid_req(prev);
prev             1222 drivers/infiniband/hw/i40iw/i40iw_puda.c 		buf = (struct i40iw_puda_buf *)(pbufl.prev);
prev             1209 drivers/infiniband/hw/mlx4/mcg.c 		pend_req = list_entry(group->func[slave].pending.prev, struct mcast_req, group_list);
prev               65 drivers/infiniband/hw/mthca/mthca_mcg.c 		    u16 *hash, int *prev, int *index)
prev               89 drivers/infiniband/hw/mthca/mthca_mcg.c 	*prev  = -1;
prev              109 drivers/infiniband/hw/mthca/mthca_mcg.c 		*prev = *index;
prev              126 drivers/infiniband/hw/mthca/mthca_mcg.c 	int index, prev;
prev              138 drivers/infiniband/hw/mthca/mthca_mcg.c 	err = find_mgm(dev, gid->raw, mailbox, &hash, &prev, &index);
prev              191 drivers/infiniband/hw/mthca/mthca_mcg.c 	err = mthca_READ_MGM(dev, prev, mailbox);
prev              199 drivers/infiniband/hw/mthca/mthca_mcg.c 	err = mthca_WRITE_MGM(dev, prev, mailbox);
prev              220 drivers/infiniband/hw/mthca/mthca_mcg.c 	int prev, index;
prev              231 drivers/infiniband/hw/mthca/mthca_mcg.c 	err = find_mgm(dev, gid->raw, mailbox, &hash, &prev, &index);
prev              266 drivers/infiniband/hw/mthca/mthca_mcg.c 	if (prev == -1) {
prev              291 drivers/infiniband/hw/mthca/mthca_mcg.c 		err = mthca_READ_MGM(dev, prev, mailbox);
prev              299 drivers/infiniband/hw/mthca/mthca_mcg.c 		err = mthca_WRITE_MGM(dev, prev, mailbox);
prev             1501 drivers/infiniband/hw/qib/qib_rc.c 	u8 i, prev;
prev             1551 drivers/infiniband/hw/qib/qib_rc.c 	for (i = qp->r_head_ack_queue; ; i = prev) {
prev             1555 drivers/infiniband/hw/qib/qib_rc.c 			prev = i - 1;
prev             1557 drivers/infiniband/hw/qib/qib_rc.c 			prev = QIB_MAX_RDMA_ATOMIC;
prev             1558 drivers/infiniband/hw/qib/qib_rc.c 		if (prev == qp->r_head_ack_queue) {
prev             1562 drivers/infiniband/hw/qib/qib_rc.c 		e = &qp->s_ack_queue[prev];
prev             1568 drivers/infiniband/hw/qib/qib_rc.c 			if (prev == qp->s_tail_ack_queue &&
prev             1621 drivers/infiniband/hw/qib/qib_rc.c 		qp->s_tail_ack_queue = prev;
prev             1634 drivers/infiniband/hw/qib/qib_rc.c 		qp->s_tail_ack_queue = prev;
prev             1085 drivers/infiniband/hw/qib/qib_user_sdma.c 		pkt = list_entry(free_list.prev,
prev             1492 drivers/infiniband/ulp/ipoib/ipoib_cm.c 		p = list_entry(priv->cm.passive_ids.prev, typeof(*p), list);
prev              111 drivers/infiniband/ulp/srpt/ib_srpt.c 	enum rdma_ch_state prev;
prev              115 drivers/infiniband/ulp/srpt/ib_srpt.c 	prev = ch->state;
prev              116 drivers/infiniband/ulp/srpt/ib_srpt.c 	if (new > prev) {
prev              894 drivers/infiniband/ulp/srpt/ib_srpt.c 	struct scatterlist *prev = NULL;
prev              928 drivers/infiniband/ulp/srpt/ib_srpt.c 		if (prev) {
prev              929 drivers/infiniband/ulp/srpt/ib_srpt.c 			sg_unmark_end(&prev[prev_nents - 1]);
prev              930 drivers/infiniband/ulp/srpt/ib_srpt.c 			sg_chain(prev, prev_nents + 1, ctx->sg);
prev              935 drivers/infiniband/ulp/srpt/ib_srpt.c 		prev = ctx->sg;
prev               36 drivers/input/joystick/joydump.c 	struct joydump *dump, *prev;	/* one entry each */
prev              107 drivers/input/joystick/joydump.c 	prev = dump;
prev              116 drivers/input/joystick/joydump.c 	for (i = 1; i < t; i++, dump++, prev++) {
prev              118 drivers/input/joystick/joydump.c 			i, dump->time - prev->time);
prev              106 drivers/input/mouse/hgpk.c static int approx_half(int curr, int prev)
prev              110 drivers/input/mouse/hgpk.c 	if (curr < 5 || prev < 5)
prev              113 drivers/input/mouse/hgpk.c 	belowhalf = (prev * 8) / 20;
prev              114 drivers/input/mouse/hgpk.c 	abovehalf = (prev * 12) / 20;
prev              671 drivers/input/touchscreen/ads7846.c 		list_entry(m->transfers.prev, struct spi_transfer, transfer_list);
prev              690 drivers/input/touchscreen/ads7846.c 		list_entry(m->transfers.prev, struct spi_transfer, transfer_list);
prev              220 drivers/interconnect/core.c 	struct icc_node *next, *prev = NULL;
prev              231 drivers/interconnect/core.c 		if (!prev || next->provider != prev->provider) {
prev              232 drivers/interconnect/core.c 			prev = next;
prev              237 drivers/interconnect/core.c 		ret = next->provider->set(prev, next);
prev              241 drivers/interconnect/core.c 		prev = next;
prev              816 drivers/iommu/dma-iommu.c 	struct scatterlist *s, *prev = NULL;
prev              857 drivers/iommu/dma-iommu.c 			prev->length += pad_len;
prev              862 drivers/iommu/dma-iommu.c 		prev = s;
prev              184 drivers/iommu/iova.c 	struct rb_node *curr, *prev;
prev              204 drivers/iommu/iova.c 		prev = curr;
prev              219 drivers/iommu/iova.c 	iova_insert_rbtree(&iovad->rbroot, new, prev);
prev              733 drivers/iommu/iova.c 	struct iova *prev = NULL, *next = NULL;
prev              737 drivers/iommu/iova.c 		prev = alloc_and_init_iova(iova->pfn_lo, pfn_lo - 1);
prev              738 drivers/iommu/iova.c 		if (prev == NULL)
prev              750 drivers/iommu/iova.c 	if (prev) {
prev              751 drivers/iommu/iova.c 		iova_insert_rbtree(&iovad->rbroot, prev, NULL);
prev              764 drivers/iommu/iova.c 	if (prev)
prev              765 drivers/iommu/iova.c 		free_iova_mem(prev);
prev              787 drivers/iommu/iova.c 	struct iova_magazine *prev;
prev              878 drivers/iommu/iova.c 			cpu_rcache->prev = iova_magazine_alloc(GFP_KERNEL);
prev              903 drivers/iommu/iova.c 	} else if (!iova_magazine_full(cpu_rcache->prev)) {
prev              904 drivers/iommu/iova.c 		swap(cpu_rcache->prev, cpu_rcache->loaded);
prev              966 drivers/iommu/iova.c 	} else if (!iova_magazine_empty(cpu_rcache->prev)) {
prev              967 drivers/iommu/iova.c 		swap(cpu_rcache->prev, cpu_rcache->loaded);
prev             1019 drivers/iommu/iova.c 			iova_magazine_free(cpu_rcache->prev);
prev             1042 drivers/iommu/iova.c 		iova_magazine_free_pfns(cpu_rcache->prev, iovad);
prev             1463 drivers/isdn/mISDN/dsp_cmx.c 			other = (list_entry(conf->mlist.prev,
prev             1475 drivers/lightnvm/pblk-core.c 	list_cut_position(&list, &l_mg->emeta_list, l_mg->emeta_list.prev);
prev               43 drivers/lightnvm/pblk-gc.c 	list_cut_position(&w_list, &gc->w_list, gc->w_list.prev);
prev              540 drivers/lightnvm/pblk-recovery.c 	__list_add(&line->list, t->list.prev, &t->list);
prev              295 drivers/md/bcache/bset.c 		kfree(t->prev);
prev              297 drivers/md/bcache/bset.c 		free_pages((unsigned long) t->prev,
prev              308 drivers/md/bcache/bset.c 	t->prev = NULL;
prev              334 drivers/md/bcache/bset.c 	t->prev = bset_prev_bytes(b) < PAGE_SIZE
prev              337 drivers/md/bcache/bset.c 	if (!t->prev)
prev              552 drivers/md/bcache/bset.c 	return (void *) (((uint64_t *) tree_to_bkey(t, j)) - t->prev[j]);
prev              561 drivers/md/bcache/bset.c 	return cacheline_to_bkey(t, cacheline, t->prev[cacheline]);
prev              645 drivers/md/bcache/bset.c 		t->prev = t[-1].prev + j;
prev              662 drivers/md/bcache/bset.c 		t->prev[0] = bkey_to_cacheline_offset(t, 0, t->data->start);
prev              695 drivers/md/bcache/bset.c 	struct bkey *prev = NULL, *k = t->data->start;
prev              718 drivers/md/bcache/bset.c 			prev = k, k = bkey_next(k);
prev              720 drivers/md/bcache/bset.c 		t->prev[j] = bkey_u64s(prev);
prev              810 drivers/md/bcache/bset.c 		t->prev[j] += shift;
prev              812 drivers/md/bcache/bset.c 		if (t->prev[j] > 7) {
prev              818 drivers/md/bcache/bset.c 			t->prev[j] = bkey_to_cacheline_offset(t, j, k);
prev              831 drivers/md/bcache/bset.c 			t->prev[t->size] =
prev              885 drivers/md/bcache/bset.c 	struct bkey *m, *prev = NULL;
prev              911 drivers/md/bcache/bset.c 		prev = m, m = bkey_next(m);
prev              915 drivers/md/bcache/bset.c 	if (prev &&
prev              916 drivers/md/bcache/bset.c 	    bch_bkey_try_merge(b, prev, k))
prev              183 drivers/md/bcache/bset.h 	uint8_t			*prev;
prev              239 drivers/md/bcache/journal.c 			seq = list_entry(list->prev, struct journal_replay,
prev              245 drivers/md/bcache/journal.c 			if (seq != list_entry(list->prev, struct journal_replay,
prev              292 drivers/md/bcache/journal.c 		c->journal.seq = list_entry(list->prev,
prev              363 drivers/md/bcache/journal.c 		list_entry(list->prev, struct journal_replay, list);
prev             1876 drivers/md/bcache/super.c 		j = &list_entry(journal.prev, struct journal_replay, list)->j;
prev             1848 drivers/md/dm-bufio.c 		b = list_entry(global_queue.prev, struct dm_buffer, global_list);
prev               40 drivers/md/dm-cache-policy-smq.c 	unsigned prev:28;
prev              139 drivers/md/dm-cache-policy-smq.c 	return to_entry(es, e->prev);
prev              152 drivers/md/dm-cache-policy-smq.c 	e->prev = INDEXER_NULL;
prev              155 drivers/md/dm-cache-policy-smq.c 		head->prev = l->head = to_index(es, e);
prev              168 drivers/md/dm-cache-policy-smq.c 	e->prev = l->tail;
prev              182 drivers/md/dm-cache-policy-smq.c 	struct entry *prev = l_prev(es, old);
prev              184 drivers/md/dm-cache-policy-smq.c 	if (!prev)
prev              188 drivers/md/dm-cache-policy-smq.c 		e->prev = old->prev;
prev              190 drivers/md/dm-cache-policy-smq.c 		prev->next = old->prev = to_index(es, e);
prev              199 drivers/md/dm-cache-policy-smq.c 	struct entry *prev = l_prev(es, e);
prev              202 drivers/md/dm-cache-policy-smq.c 	if (prev)
prev              203 drivers/md/dm-cache-policy-smq.c 		prev->next = e->next;
prev              208 drivers/md/dm-cache-policy-smq.c 		next->prev = e->prev;
prev              210 drivers/md/dm-cache-policy-smq.c 		l->tail = e->prev;
prev              629 drivers/md/dm-cache-policy-smq.c 				struct entry **prev)
prev              633 drivers/md/dm-cache-policy-smq.c 	*prev = NULL;
prev              638 drivers/md/dm-cache-policy-smq.c 		*prev = e;
prev              645 drivers/md/dm-cache-policy-smq.c 		       struct entry *e, struct entry *prev)
prev              647 drivers/md/dm-cache-policy-smq.c 	if (prev)
prev              648 drivers/md/dm-cache-policy-smq.c 		prev->hash_next = e->hash_next;
prev              658 drivers/md/dm-cache-policy-smq.c 	struct entry *e, *prev;
prev              661 drivers/md/dm-cache-policy-smq.c 	e = __h_lookup(ht, h, oblock, &prev);
prev              662 drivers/md/dm-cache-policy-smq.c 	if (e && prev) {
prev              667 drivers/md/dm-cache-policy-smq.c 		__h_unlink(ht, h, e, prev);
prev              677 drivers/md/dm-cache-policy-smq.c 	struct entry *prev;
prev              683 drivers/md/dm-cache-policy-smq.c 	e = __h_lookup(ht, h, e->oblock, &prev);
prev              685 drivers/md/dm-cache-policy-smq.c 		__h_unlink(ht, h, e, prev);
prev              720 drivers/md/dm-cache-policy-smq.c 	e->prev = INDEXER_NULL;
prev              529 drivers/md/dm-table.c 	struct dm_target *prev;
prev              534 drivers/md/dm-table.c 	prev = &table->targets[table->num_targets - 1];
prev              535 drivers/md/dm-table.c 	return (ti->begin == (prev->begin + prev->len));
prev              689 drivers/md/dm-writecache.c 	wc->lru.prev = LIST_POISON2;
prev              691 drivers/md/dm-writecache.c 	wc->freelist.prev = LIST_POISON2;
prev              758 drivers/md/dm-writecache.c 		if (unlikely(e->lru.prev == &wc->lru))
prev              760 drivers/md/dm-writecache.c 		e = container_of(e->lru.prev, struct wc_entry, lru);
prev             1448 drivers/md/dm-writecache.c 		list.next->prev = list.prev->next = &list;
prev             1510 drivers/md/dm-writecache.c 		e = container_of(wbl->list.prev, struct wc_entry, lru);
prev             1535 drivers/md/dm-writecache.c 			f = container_of(wbl->list.prev, struct wc_entry, lru);
prev             1568 drivers/md/dm-writecache.c 		e = container_of(wbl->list.prev, struct wc_entry, lru);
prev             1587 drivers/md/dm-writecache.c 			f = container_of(wbl->list.prev, struct wc_entry, lru);
prev             1647 drivers/md/dm-writecache.c 			e = container_of(wc->lru.prev, struct wc_entry, lru);
prev              328 drivers/md/md-multipath.c 		mp_bh = list_entry(head->prev, struct multipath_bh, retry_list);
prev              329 drivers/md/md-multipath.c 		list_del(head->prev);
prev             2564 drivers/md/raid1.c 		r1_bio = list_entry(head->prev, struct r1bio, retry_list);
prev             2565 drivers/md/raid1.c 		list_del(head->prev);
prev              632 drivers/md/raid10.c 		geo = &conf->prev;
prev              862 drivers/md/raid10.c 	     (i < conf->geo.raid_disks || i < conf->prev.raid_disks)
prev             1524 drivers/md/raid10.c 	sector_t chunk_mask = (conf->geo.chunk_mask & conf->prev.chunk_mask);
prev             1542 drivers/md/raid10.c 			 || conf->prev.near_copies <
prev             1543 drivers/md/raid10.c 			 conf->prev.raid_disks)))
prev             1593 drivers/md/raid10.c 		disks = conf->prev.raid_disks;
prev             1594 drivers/md/raid10.c 		ncopies = conf->prev.near_copies;
prev             2728 drivers/md/raid10.c 				list_move(conf->bio_end_io_list.prev, &tmp);
prev             2757 drivers/md/raid10.c 		r10_bio = list_entry(head->prev, struct r10bio, retry_list);
prev             2758 drivers/md/raid10.c 		list_del(head->prev);
prev             3532 drivers/md/raid10.c 				 conf->prev.raid_disks);
prev             3681 drivers/md/raid10.c 		conf->prev = conf->geo;
prev             3684 drivers/md/raid10.c 		if (setup_geo(&conf->prev, mddev, geo_old) != conf->copies) {
prev             3689 drivers/md/raid10.c 		if (conf->prev.far_offset)
prev             3690 drivers/md/raid10.c 			conf->prev.stride = 1 << conf->prev.chunk_shift;
prev             3693 drivers/md/raid10.c 			conf->prev.stride = conf->dev_sectors;
prev             3783 drivers/md/raid10.c 		    disk_idx >= conf->prev.raid_disks)
prev             3835 drivers/md/raid10.c 		if (conf->prev.far_copies != 1 &&
prev             3836 drivers/md/raid10.c 		    conf->prev.far_offset == 0)
prev             3843 drivers/md/raid10.c 		     || i < conf->prev.raid_disks;
prev             3907 drivers/md/raid10.c 		before_length = ((1 << conf->prev.chunk_shift) *
prev             3908 drivers/md/raid10.c 				 conf->prev.far_copies);
prev             4139 drivers/md/raid10.c 	for (i = 0; i < conf->prev.raid_disks; i++) {
prev             4151 drivers/md/raid10.c 	if (conf->geo.raid_disks == conf->prev.raid_disks)
prev             4165 drivers/md/raid10.c 			if (conf->geo.raid_disks <= conf->prev.raid_disks)
prev             4202 drivers/md/raid10.c 	before_length = ((1 << conf->prev.chunk_shift) *
prev             4203 drivers/md/raid10.c 			 conf->prev.far_copies);
prev             4234 drivers/md/raid10.c 		       sizeof(struct raid10_info)*conf->prev.raid_disks);
prev             4305 drivers/md/raid10.c 					    conf->prev.raid_disks)
prev             4313 drivers/md/raid10.c 			} else if (rdev->raid_disk >= conf->prev.raid_disks
prev             4350 drivers/md/raid10.c 	conf->geo = conf->prev;
prev             4478 drivers/md/raid10.c 					&conf->prev);
prev             4485 drivers/md/raid10.c 					       & conf->prev.chunk_mask);
prev             4497 drivers/md/raid10.c 		safe = first_dev_address(conf->reshape_safe, &conf->prev);
prev             4507 drivers/md/raid10.c 				     & conf->prev.chunk_mask);
prev             4731 drivers/md/raid10.c 	conf->prev = conf->geo;
prev             4785 drivers/md/raid10.c 	__raid10_find_phys(&conf->prev, r10b);
prev               57 drivers/md/raid10.h 	} prev, geo;
prev              497 drivers/md/raid5-ppl.c 			struct bio *prev = bio;
prev              501 drivers/md/raid5-ppl.c 			bio->bi_opf = prev->bi_opf;
prev              502 drivers/md/raid5-ppl.c 			bio->bi_write_hint = prev->bi_write_hint;
prev              503 drivers/md/raid5-ppl.c 			bio_copy_dev(bio, prev);
prev              504 drivers/md/raid5-ppl.c 			bio->bi_iter.bi_sector = bio_end_sector(prev);
prev              507 drivers/md/raid5-ppl.c 			bio_chain(bio, prev);
prev              508 drivers/md/raid5-ppl.c 			ppl_submit_iounit_bio(io, prev);
prev              133 drivers/media/common/saa7146/saa7146_fops.c 			buf, q->queue.prev, q->queue.next);
prev              310 drivers/media/common/videobuf2/videobuf2-core.c 		struct vb2_buffer *prev = q->bufs[vb->index - 1];
prev              311 drivers/media/common/videobuf2/videobuf2-core.c 		struct vb2_plane *p = &prev->planes[prev->num_planes - 1];
prev             1614 drivers/media/pci/cx23885/cx23885-core.c 	struct cx23885_buffer    *prev;
prev             1631 drivers/media/pci/cx23885/cx23885-core.c 		prev = list_entry(cx88q->active.prev, struct cx23885_buffer,
prev             1634 drivers/media/pci/cx23885/cx23885-core.c 		prev->risc.jmp[1] = cpu_to_le32(buf->risc.dma);
prev              188 drivers/media/pci/cx23885/cx23885-vbi.c 	struct cx23885_buffer *prev;
prev              206 drivers/media/pci/cx23885/cx23885-vbi.c 		prev = list_entry(q->active.prev, struct cx23885_buffer,
prev              211 drivers/media/pci/cx23885/cx23885-vbi.c 		prev->risc.jmp[1] = cpu_to_le32(buf->risc.dma);
prev              457 drivers/media/pci/cx23885/cx23885-video.c 	struct cx23885_buffer   *prev;
prev              474 drivers/media/pci/cx23885/cx23885-video.c 		prev = list_entry(q->active.prev, struct cx23885_buffer,
prev              477 drivers/media/pci/cx23885/cx23885-video.c 		prev->risc.jmp[1] = cpu_to_le32(buf->risc.dma);
prev              242 drivers/media/pci/cx25821/cx25821-video.c 	struct cx25821_buffer *prev;
prev              254 drivers/media/pci/cx25821/cx25821-video.c 		prev = list_entry(q->active.prev, struct cx25821_buffer,
prev              257 drivers/media/pci/cx25821/cx25821-video.c 		prev->risc.jmp[1] = cpu_to_le32(buf->risc.dma);
prev              240 drivers/media/pci/cx88/cx88-mpeg.c 	struct cx88_buffer    *prev;
prev              258 drivers/media/pci/cx88/cx88-mpeg.c 		prev = list_entry(cx88q->active.prev, struct cx88_buffer, list);
prev              260 drivers/media/pci/cx88/cx88-mpeg.c 		prev->risc.jmp[1] = cpu_to_le32(buf->risc.dma);
prev              171 drivers/media/pci/cx88/cx88-vbi.c 	struct cx88_buffer    *prev;
prev              186 drivers/media/pci/cx88/cx88-vbi.c 		prev = list_entry(q->active.prev, struct cx88_buffer, list);
prev              188 drivers/media/pci/cx88/cx88-vbi.c 		prev->risc.jmp[1] = cpu_to_le32(buf->risc.dma);
prev              506 drivers/media/pci/cx88/cx88-video.c 	struct cx88_buffer    *prev;
prev              521 drivers/media/pci/cx88/cx88-video.c 		prev = list_entry(q->active.prev, struct cx88_buffer, list);
prev              523 drivers/media/pci/cx88/cx88-video.c 		prev->risc.jmp[1] = cpu_to_le32(buf->risc.dma);
prev              133 drivers/media/pci/ivtv/ivtv-queue.c 		struct ivtv_buffer *buf = list_entry(steal->list.prev, struct ivtv_buffer, list);
prev              141 drivers/media/pci/ivtv/ivtv-queue.c 			list_move_tail(steal->list.prev, &from->list);
prev              152 drivers/media/pci/ivtv/ivtv-queue.c 			buf = list_entry(steal->list.prev, struct ivtv_buffer, list);
prev              316 drivers/media/pci/saa7134/saa7134-core.c 			buf, q->queue.prev, q->queue.next);
prev              323 drivers/media/pci/saa7134/saa7134-core.c 			q->queue.prev, q->queue.next);
prev              409 drivers/media/pci/tw68/tw68-video.c 	struct tw68_buf *prev;
prev              419 drivers/media/pci/tw68/tw68-video.c 		prev = list_entry(dev->active.prev, struct tw68_buf, list);
prev              421 drivers/media/pci/tw68/tw68-video.c 		prev->jmp[1] = cpu_to_le32(buf->dma);
prev              292 drivers/media/platform/fsl-viu.c 	struct viu_buf *buf, *prev;
prev              303 drivers/media/platform/fsl-viu.c 		list_for_each_entry_safe(buf, prev, &vidq->active, vb.queue) {
prev              312 drivers/media/platform/fsl-viu.c 	prev = NULL;
prev              317 drivers/media/platform/fsl-viu.c 		if (prev == NULL) {
prev              329 drivers/media/platform/fsl-viu.c 		} else if (prev->vb.width  == buf->vb.width  &&
prev              330 drivers/media/platform/fsl-viu.c 			   prev->vb.height == buf->vb.height &&
prev              331 drivers/media/platform/fsl-viu.c 			   prev->fmt       == buf->fmt) {
prev              339 drivers/media/platform/fsl-viu.c 		prev = buf;
prev              498 drivers/media/platform/fsl-viu.c 	struct viu_buf       *prev;
prev              506 drivers/media/platform/fsl-viu.c 			vidq->queued.prev);
prev              522 drivers/media/platform/fsl-viu.c 		prev = list_entry(vidq->active.prev, struct viu_buf, vb.queue);
prev              523 drivers/media/platform/fsl-viu.c 		if (prev->vb.width  == buf->vb.width  &&
prev              524 drivers/media/platform/fsl-viu.c 		    prev->vb.height == buf->vb.height &&
prev              525 drivers/media/platform/fsl-viu.c 		    prev->fmt       == buf->fmt) {
prev              149 drivers/media/platform/omap3isp/isppreview.c preview_config_luma_enhancement(struct isp_prev_device *prev,
prev              152 drivers/media/platform/omap3isp/isppreview.c 	struct isp_device *isp = to_isp_device(prev);
prev              168 drivers/media/platform/omap3isp/isppreview.c preview_enable_luma_enhancement(struct isp_prev_device *prev, bool enable)
prev              170 drivers/media/platform/omap3isp/isppreview.c 	struct isp_device *isp = to_isp_device(prev);
prev              183 drivers/media/platform/omap3isp/isppreview.c static void preview_enable_invalaw(struct isp_prev_device *prev, bool enable)
prev              185 drivers/media/platform/omap3isp/isppreview.c 	struct isp_device *isp = to_isp_device(prev);
prev              198 drivers/media/platform/omap3isp/isppreview.c static void preview_config_hmed(struct isp_prev_device *prev,
prev              201 drivers/media/platform/omap3isp/isppreview.c 	struct isp_device *isp = to_isp_device(prev);
prev              213 drivers/media/platform/omap3isp/isppreview.c static void preview_enable_hmed(struct isp_prev_device *prev, bool enable)
prev              215 drivers/media/platform/omap3isp/isppreview.c 	struct isp_device *isp = to_isp_device(prev);
prev              233 drivers/media/platform/omap3isp/isppreview.c static void preview_config_cfa(struct isp_prev_device *prev,
prev              242 drivers/media/platform/omap3isp/isppreview.c 	const unsigned int *order = cfa_coef_order[prev->params.cfa_order];
prev              244 drivers/media/platform/omap3isp/isppreview.c 	struct isp_device *isp = to_isp_device(prev);
prev              269 drivers/media/platform/omap3isp/isppreview.c preview_config_chroma_suppression(struct isp_prev_device *prev,
prev              272 drivers/media/platform/omap3isp/isppreview.c 	struct isp_device *isp = to_isp_device(prev);
prev              285 drivers/media/platform/omap3isp/isppreview.c preview_enable_chroma_suppression(struct isp_prev_device *prev, bool enable)
prev              287 drivers/media/platform/omap3isp/isppreview.c 	struct isp_device *isp = to_isp_device(prev);
prev              303 drivers/media/platform/omap3isp/isppreview.c preview_config_whitebalance(struct isp_prev_device *prev,
prev              306 drivers/media/platform/omap3isp/isppreview.c 	struct isp_device *isp = to_isp_device(prev);
prev              342 drivers/media/platform/omap3isp/isppreview.c preview_config_blkadj(struct isp_prev_device *prev,
prev              345 drivers/media/platform/omap3isp/isppreview.c 	struct isp_device *isp = to_isp_device(prev);
prev              358 drivers/media/platform/omap3isp/isppreview.c preview_config_rgb_blending(struct isp_prev_device *prev,
prev              361 drivers/media/platform/omap3isp/isppreview.c 	struct isp_device *isp = to_isp_device(prev);
prev              396 drivers/media/platform/omap3isp/isppreview.c preview_config_csc(struct isp_prev_device *prev,
prev              399 drivers/media/platform/omap3isp/isppreview.c 	struct isp_device *isp = to_isp_device(prev);
prev              428 drivers/media/platform/omap3isp/isppreview.c preview_config_yc_range(struct isp_prev_device *prev,
prev              431 drivers/media/platform/omap3isp/isppreview.c 	struct isp_device *isp = to_isp_device(prev);
prev              446 drivers/media/platform/omap3isp/isppreview.c preview_config_dcor(struct isp_prev_device *prev,
prev              449 drivers/media/platform/omap3isp/isppreview.c 	struct isp_device *isp = to_isp_device(prev);
prev              468 drivers/media/platform/omap3isp/isppreview.c static void preview_enable_dcor(struct isp_prev_device *prev, bool enable)
prev              470 drivers/media/platform/omap3isp/isppreview.c 	struct isp_device *isp = to_isp_device(prev);
prev              484 drivers/media/platform/omap3isp/isppreview.c preview_enable_drkframe_capture(struct isp_prev_device *prev, bool enable)
prev              486 drivers/media/platform/omap3isp/isppreview.c 	struct isp_device *isp = to_isp_device(prev);
prev              499 drivers/media/platform/omap3isp/isppreview.c static void preview_enable_drkframe(struct isp_prev_device *prev, bool enable)
prev              501 drivers/media/platform/omap3isp/isppreview.c 	struct isp_device *isp = to_isp_device(prev);
prev              515 drivers/media/platform/omap3isp/isppreview.c preview_config_noisefilter(struct isp_prev_device *prev,
prev              518 drivers/media/platform/omap3isp/isppreview.c 	struct isp_device *isp = to_isp_device(prev);
prev              535 drivers/media/platform/omap3isp/isppreview.c preview_enable_noisefilter(struct isp_prev_device *prev, bool enable)
prev              537 drivers/media/platform/omap3isp/isppreview.c 	struct isp_device *isp = to_isp_device(prev);
prev              551 drivers/media/platform/omap3isp/isppreview.c preview_config_gammacorrn(struct isp_prev_device *prev,
prev              554 drivers/media/platform/omap3isp/isppreview.c 	struct isp_device *isp = to_isp_device(prev);
prev              584 drivers/media/platform/omap3isp/isppreview.c preview_enable_gammacorrn(struct isp_prev_device *prev, bool enable)
prev              586 drivers/media/platform/omap3isp/isppreview.c 	struct isp_device *isp = to_isp_device(prev);
prev              602 drivers/media/platform/omap3isp/isppreview.c preview_config_contrast(struct isp_prev_device *prev,
prev              605 drivers/media/platform/omap3isp/isppreview.c 	struct isp_device *isp = to_isp_device(prev);
prev              616 drivers/media/platform/omap3isp/isppreview.c preview_config_brightness(struct isp_prev_device *prev,
prev              619 drivers/media/platform/omap3isp/isppreview.c 	struct isp_device *isp = to_isp_device(prev);
prev              633 drivers/media/platform/omap3isp/isppreview.c preview_update_contrast(struct isp_prev_device *prev, u8 contrast)
prev              638 drivers/media/platform/omap3isp/isppreview.c 	spin_lock_irqsave(&prev->params.lock, flags);
prev              639 drivers/media/platform/omap3isp/isppreview.c 	params = (prev->params.active & OMAP3ISP_PREV_CONTRAST)
prev              640 drivers/media/platform/omap3isp/isppreview.c 	       ? &prev->params.params[0] : &prev->params.params[1];
prev              646 drivers/media/platform/omap3isp/isppreview.c 	spin_unlock_irqrestore(&prev->params.lock, flags);
prev              655 drivers/media/platform/omap3isp/isppreview.c preview_update_brightness(struct isp_prev_device *prev, u8 brightness)
prev              660 drivers/media/platform/omap3isp/isppreview.c 	spin_lock_irqsave(&prev->params.lock, flags);
prev              661 drivers/media/platform/omap3isp/isppreview.c 	params = (prev->params.active & OMAP3ISP_PREV_BRIGHTNESS)
prev              662 drivers/media/platform/omap3isp/isppreview.c 	       ? &prev->params.params[0] : &prev->params.params[1];
prev              668 drivers/media/platform/omap3isp/isppreview.c 	spin_unlock_irqrestore(&prev->params.lock, flags);
prev              672 drivers/media/platform/omap3isp/isppreview.c preview_params_lock(struct isp_prev_device *prev, u32 update, bool shadow)
prev              674 drivers/media/platform/omap3isp/isppreview.c 	u32 active = prev->params.active;
prev              678 drivers/media/platform/omap3isp/isppreview.c 		prev->params.params[0].busy |= ~active & update;
prev              679 drivers/media/platform/omap3isp/isppreview.c 		prev->params.params[1].busy |= active & update;
prev              682 drivers/media/platform/omap3isp/isppreview.c 		update = (prev->params.params[0].update & active)
prev              683 drivers/media/platform/omap3isp/isppreview.c 		       | (prev->params.params[1].update & ~active);
prev              685 drivers/media/platform/omap3isp/isppreview.c 		prev->params.params[0].busy |= active & update;
prev              686 drivers/media/platform/omap3isp/isppreview.c 		prev->params.params[1].busy |= ~active & update;
prev              693 drivers/media/platform/omap3isp/isppreview.c preview_params_unlock(struct isp_prev_device *prev, u32 update, bool shadow)
prev              695 drivers/media/platform/omap3isp/isppreview.c 	u32 active = prev->params.active;
prev              701 drivers/media/platform/omap3isp/isppreview.c 		prev->params.params[0].update |= (~active & update);
prev              702 drivers/media/platform/omap3isp/isppreview.c 		prev->params.params[1].update |= (active & update);
prev              703 drivers/media/platform/omap3isp/isppreview.c 		prev->params.params[0].busy &= active;
prev              704 drivers/media/platform/omap3isp/isppreview.c 		prev->params.params[1].busy &= ~active;
prev              709 drivers/media/platform/omap3isp/isppreview.c 		prev->params.params[0].update &= ~(active & update);
prev              710 drivers/media/platform/omap3isp/isppreview.c 		prev->params.params[1].update &= ~(~active & update);
prev              711 drivers/media/platform/omap3isp/isppreview.c 		prev->params.params[0].busy &= ~active;
prev              712 drivers/media/platform/omap3isp/isppreview.c 		prev->params.params[1].busy &= active;
prev              716 drivers/media/platform/omap3isp/isppreview.c static void preview_params_switch(struct isp_prev_device *prev)
prev              724 drivers/media/platform/omap3isp/isppreview.c 	to_switch = (prev->params.params[0].update & ~prev->params.active)
prev              725 drivers/media/platform/omap3isp/isppreview.c 		  | (prev->params.params[1].update & prev->params.active);
prev              726 drivers/media/platform/omap3isp/isppreview.c 	to_switch &= ~(prev->params.params[0].busy |
prev              727 drivers/media/platform/omap3isp/isppreview.c 		       prev->params.params[1].busy);
prev              731 drivers/media/platform/omap3isp/isppreview.c 	prev->params.active ^= to_switch;
prev              736 drivers/media/platform/omap3isp/isppreview.c 	prev->params.params[0].update &= ~(~prev->params.active & to_switch);
prev              737 drivers/media/platform/omap3isp/isppreview.c 	prev->params.params[1].update &= ~(prev->params.active & to_switch);
prev              859 drivers/media/platform/omap3isp/isppreview.c static int preview_config(struct isp_prev_device *prev,
prev              872 drivers/media/platform/omap3isp/isppreview.c 	spin_lock_irqsave(&prev->params.lock, flags);
prev              873 drivers/media/platform/omap3isp/isppreview.c 	preview_params_lock(prev, cfg->update, true);
prev              874 drivers/media/platform/omap3isp/isppreview.c 	active = prev->params.active;
prev              875 drivers/media/platform/omap3isp/isppreview.c 	spin_unlock_irqrestore(&prev->params.lock, flags);
prev              887 drivers/media/platform/omap3isp/isppreview.c 		params = &prev->params.params[!!(active & bit)];
prev              909 drivers/media/platform/omap3isp/isppreview.c 	spin_lock_irqsave(&prev->params.lock, flags);
prev              910 drivers/media/platform/omap3isp/isppreview.c 	preview_params_unlock(prev, update, true);
prev              911 drivers/media/platform/omap3isp/isppreview.c 	preview_params_switch(prev);
prev              912 drivers/media/platform/omap3isp/isppreview.c 	spin_unlock_irqrestore(&prev->params.lock, flags);
prev              925 drivers/media/platform/omap3isp/isppreview.c static void preview_setup_hw(struct isp_prev_device *prev, u32 update,
prev              941 drivers/media/platform/omap3isp/isppreview.c 		params = &prev->params.params[!(active & bit)];
prev              945 drivers/media/platform/omap3isp/isppreview.c 				attr->config(prev, params);
prev              947 drivers/media/platform/omap3isp/isppreview.c 				attr->enable(prev, true);
prev              950 drivers/media/platform/omap3isp/isppreview.c 				attr->enable(prev, false);
prev              960 drivers/media/platform/omap3isp/isppreview.c static void preview_config_ycpos(struct isp_prev_device *prev, u32 pixelcode)
prev              962 drivers/media/platform/omap3isp/isppreview.c 	struct isp_device *isp = to_isp_device(prev);
prev              985 drivers/media/platform/omap3isp/isppreview.c static void preview_config_averager(struct isp_prev_device *prev, u8 average)
prev              987 drivers/media/platform/omap3isp/isppreview.c 	struct isp_device *isp = to_isp_device(prev);
prev             1008 drivers/media/platform/omap3isp/isppreview.c static void preview_config_input_format(struct isp_prev_device *prev,
prev             1011 drivers/media/platform/omap3isp/isppreview.c 	struct isp_device *isp = to_isp_device(prev);
prev             1023 drivers/media/platform/omap3isp/isppreview.c 		prev->params.cfa_order = 0;
prev             1026 drivers/media/platform/omap3isp/isppreview.c 		prev->params.cfa_order = 1;
prev             1029 drivers/media/platform/omap3isp/isppreview.c 		prev->params.cfa_order = 2;
prev             1032 drivers/media/platform/omap3isp/isppreview.c 		prev->params.cfa_order = 3;
prev             1045 drivers/media/platform/omap3isp/isppreview.c 	params = (prev->params.active & OMAP3ISP_PREV_CFA)
prev             1046 drivers/media/platform/omap3isp/isppreview.c 	       ? &prev->params.params[0] : &prev->params.params[1];
prev             1048 drivers/media/platform/omap3isp/isppreview.c 	preview_config_cfa(prev, params);
prev             1062 drivers/media/platform/omap3isp/isppreview.c static void preview_config_input_size(struct isp_prev_device *prev, u32 active)
prev             1064 drivers/media/platform/omap3isp/isppreview.c 	const struct v4l2_mbus_framefmt *format = &prev->formats[PREV_PAD_SINK];
prev             1065 drivers/media/platform/omap3isp/isppreview.c 	struct isp_device *isp = to_isp_device(prev);
prev             1066 drivers/media/platform/omap3isp/isppreview.c 	unsigned int sph = prev->crop.left;
prev             1067 drivers/media/platform/omap3isp/isppreview.c 	unsigned int eph = prev->crop.left + prev->crop.width - 1;
prev             1068 drivers/media/platform/omap3isp/isppreview.c 	unsigned int slv = prev->crop.top;
prev             1069 drivers/media/platform/omap3isp/isppreview.c 	unsigned int elv = prev->crop.top + prev->crop.height - 1;
prev             1080 drivers/media/platform/omap3isp/isppreview.c 	features = (prev->params.params[0].features & active)
prev             1081 drivers/media/platform/omap3isp/isppreview.c 		 | (prev->params.params[1].features & ~active);
prev             1113 drivers/media/platform/omap3isp/isppreview.c preview_config_inlineoffset(struct isp_prev_device *prev, u32 offset)
prev             1115 drivers/media/platform/omap3isp/isppreview.c 	struct isp_device *isp = to_isp_device(prev);
prev             1127 drivers/media/platform/omap3isp/isppreview.c static void preview_set_inaddr(struct isp_prev_device *prev, u32 addr)
prev             1129 drivers/media/platform/omap3isp/isppreview.c 	struct isp_device *isp = to_isp_device(prev);
prev             1140 drivers/media/platform/omap3isp/isppreview.c static void preview_config_outlineoffset(struct isp_prev_device *prev,
prev             1143 drivers/media/platform/omap3isp/isppreview.c 	struct isp_device *isp = to_isp_device(prev);
prev             1155 drivers/media/platform/omap3isp/isppreview.c static void preview_set_outaddr(struct isp_prev_device *prev, u32 addr)
prev             1157 drivers/media/platform/omap3isp/isppreview.c 	struct isp_device *isp = to_isp_device(prev);
prev             1162 drivers/media/platform/omap3isp/isppreview.c static void preview_adjust_bandwidth(struct isp_prev_device *prev)
prev             1164 drivers/media/platform/omap3isp/isppreview.c 	struct isp_pipeline *pipe = to_isp_pipeline(&prev->subdev.entity);
prev             1165 drivers/media/platform/omap3isp/isppreview.c 	struct isp_device *isp = to_isp_device(prev);
prev             1166 drivers/media/platform/omap3isp/isppreview.c 	const struct v4l2_mbus_framefmt *ifmt = &prev->formats[PREV_PAD_SINK];
prev             1176 drivers/media/platform/omap3isp/isppreview.c 	if (prev->input != PREVIEW_INPUT_MEMORY) {
prev             1215 drivers/media/platform/omap3isp/isppreview.c int omap3isp_preview_busy(struct isp_prev_device *prev)
prev             1217 drivers/media/platform/omap3isp/isppreview.c 	struct isp_device *isp = to_isp_device(prev);
prev             1228 drivers/media/platform/omap3isp/isppreview.c 	struct isp_prev_device *prev = &isp->isp_prev;
prev             1231 drivers/media/platform/omap3isp/isppreview.c 	prev->params.params[0].update = prev->params.active & update;
prev             1232 drivers/media/platform/omap3isp/isppreview.c 	prev->params.params[1].update = ~prev->params.active & update;
prev             1234 drivers/media/platform/omap3isp/isppreview.c 	preview_setup_hw(prev, update, prev->params.active);
prev             1236 drivers/media/platform/omap3isp/isppreview.c 	prev->params.params[0].update = 0;
prev             1237 drivers/media/platform/omap3isp/isppreview.c 	prev->params.params[1].update = 0;
prev             1247 drivers/media/platform/omap3isp/isppreview.c static void preview_print_status(struct isp_prev_device *prev)
prev             1249 drivers/media/platform/omap3isp/isppreview.c 	struct isp_device *isp = to_isp_device(prev);
prev             1297 drivers/media/platform/omap3isp/isppreview.c static void preview_init_params(struct isp_prev_device *prev)
prev             1302 drivers/media/platform/omap3isp/isppreview.c 	spin_lock_init(&prev->params.lock);
prev             1304 drivers/media/platform/omap3isp/isppreview.c 	prev->params.active = ~0;
prev             1305 drivers/media/platform/omap3isp/isppreview.c 	prev->params.params[0].busy = 0;
prev             1306 drivers/media/platform/omap3isp/isppreview.c 	prev->params.params[0].update = OMAP3ISP_PREV_FEATURES_END - 1;
prev             1307 drivers/media/platform/omap3isp/isppreview.c 	prev->params.params[1].busy = 0;
prev             1308 drivers/media/platform/omap3isp/isppreview.c 	prev->params.params[1].update = 0;
prev             1310 drivers/media/platform/omap3isp/isppreview.c 	params = &prev->params.params[0];
prev             1361 drivers/media/platform/omap3isp/isppreview.c static unsigned int preview_max_out_width(struct isp_prev_device *prev)
prev             1363 drivers/media/platform/omap3isp/isppreview.c 	struct isp_device *isp = to_isp_device(prev);
prev             1378 drivers/media/platform/omap3isp/isppreview.c static void preview_configure(struct isp_prev_device *prev)
prev             1380 drivers/media/platform/omap3isp/isppreview.c 	struct isp_device *isp = to_isp_device(prev);
prev             1387 drivers/media/platform/omap3isp/isppreview.c 	spin_lock_irqsave(&prev->params.lock, flags);
prev             1389 drivers/media/platform/omap3isp/isppreview.c 	update = preview_params_lock(prev, 0, false);
prev             1390 drivers/media/platform/omap3isp/isppreview.c 	active = prev->params.active;
prev             1391 drivers/media/platform/omap3isp/isppreview.c 	spin_unlock_irqrestore(&prev->params.lock, flags);
prev             1394 drivers/media/platform/omap3isp/isppreview.c 	format = &prev->formats[PREV_PAD_SINK];
prev             1397 drivers/media/platform/omap3isp/isppreview.c 	preview_adjust_bandwidth(prev);
prev             1399 drivers/media/platform/omap3isp/isppreview.c 	preview_config_input_format(prev, info);
prev             1400 drivers/media/platform/omap3isp/isppreview.c 	preview_config_input_size(prev, active);
prev             1402 drivers/media/platform/omap3isp/isppreview.c 	if (prev->input == PREVIEW_INPUT_CCDC)
prev             1403 drivers/media/platform/omap3isp/isppreview.c 		preview_config_inlineoffset(prev, 0);
prev             1405 drivers/media/platform/omap3isp/isppreview.c 		preview_config_inlineoffset(prev, ALIGN(format->width, 0x20) *
prev             1408 drivers/media/platform/omap3isp/isppreview.c 	preview_setup_hw(prev, update, active);
prev             1411 drivers/media/platform/omap3isp/isppreview.c 	format = &prev->formats[PREV_PAD_SOURCE];
prev             1413 drivers/media/platform/omap3isp/isppreview.c 	if (prev->output & PREVIEW_OUTPUT_MEMORY)
prev             1420 drivers/media/platform/omap3isp/isppreview.c 	if (prev->output & PREVIEW_OUTPUT_RESIZER)
prev             1427 drivers/media/platform/omap3isp/isppreview.c 	if (prev->output & PREVIEW_OUTPUT_MEMORY)
prev             1428 drivers/media/platform/omap3isp/isppreview.c 		preview_config_outlineoffset(prev,
prev             1431 drivers/media/platform/omap3isp/isppreview.c 	preview_config_averager(prev, 0);
prev             1432 drivers/media/platform/omap3isp/isppreview.c 	preview_config_ycpos(prev, format->code);
prev             1434 drivers/media/platform/omap3isp/isppreview.c 	spin_lock_irqsave(&prev->params.lock, flags);
prev             1435 drivers/media/platform/omap3isp/isppreview.c 	preview_params_unlock(prev, update, false);
prev             1436 drivers/media/platform/omap3isp/isppreview.c 	spin_unlock_irqrestore(&prev->params.lock, flags);
prev             1443 drivers/media/platform/omap3isp/isppreview.c static void preview_enable_oneshot(struct isp_prev_device *prev)
prev             1445 drivers/media/platform/omap3isp/isppreview.c 	struct isp_device *isp = to_isp_device(prev);
prev             1451 drivers/media/platform/omap3isp/isppreview.c 	if (prev->input == PREVIEW_INPUT_MEMORY)
prev             1459 drivers/media/platform/omap3isp/isppreview.c void omap3isp_preview_isr_frame_sync(struct isp_prev_device *prev)
prev             1467 drivers/media/platform/omap3isp/isppreview.c 	if (prev->state == ISP_PIPELINE_STREAM_CONTINUOUS &&
prev             1468 drivers/media/platform/omap3isp/isppreview.c 	    prev->video_out.dmaqueue_flags & ISP_VIDEO_DMAQUEUE_QUEUED) {
prev             1469 drivers/media/platform/omap3isp/isppreview.c 		preview_enable_oneshot(prev);
prev             1470 drivers/media/platform/omap3isp/isppreview.c 		isp_video_dmaqueue_flags_clr(&prev->video_out);
prev             1474 drivers/media/platform/omap3isp/isppreview.c static void preview_isr_buffer(struct isp_prev_device *prev)
prev             1476 drivers/media/platform/omap3isp/isppreview.c 	struct isp_pipeline *pipe = to_isp_pipeline(&prev->subdev.entity);
prev             1480 drivers/media/platform/omap3isp/isppreview.c 	if (prev->output & PREVIEW_OUTPUT_MEMORY) {
prev             1481 drivers/media/platform/omap3isp/isppreview.c 		buffer = omap3isp_video_buffer_next(&prev->video_out);
prev             1483 drivers/media/platform/omap3isp/isppreview.c 			preview_set_outaddr(prev, buffer->dma);
prev             1489 drivers/media/platform/omap3isp/isppreview.c 	if (prev->input == PREVIEW_INPUT_MEMORY) {
prev             1490 drivers/media/platform/omap3isp/isppreview.c 		buffer = omap3isp_video_buffer_next(&prev->video_in);
prev             1492 drivers/media/platform/omap3isp/isppreview.c 			preview_set_inaddr(prev, buffer->dma);
prev             1496 drivers/media/platform/omap3isp/isppreview.c 	switch (prev->state) {
prev             1508 drivers/media/platform/omap3isp/isppreview.c 			preview_enable_oneshot(prev);
prev             1522 drivers/media/platform/omap3isp/isppreview.c void omap3isp_preview_isr(struct isp_prev_device *prev)
prev             1528 drivers/media/platform/omap3isp/isppreview.c 	if (omap3isp_module_sync_is_stopping(&prev->wait, &prev->stopping))
prev             1531 drivers/media/platform/omap3isp/isppreview.c 	spin_lock_irqsave(&prev->params.lock, flags);
prev             1532 drivers/media/platform/omap3isp/isppreview.c 	preview_params_switch(prev);
prev             1533 drivers/media/platform/omap3isp/isppreview.c 	update = preview_params_lock(prev, 0, false);
prev             1534 drivers/media/platform/omap3isp/isppreview.c 	active = prev->params.active;
prev             1535 drivers/media/platform/omap3isp/isppreview.c 	spin_unlock_irqrestore(&prev->params.lock, flags);
prev             1537 drivers/media/platform/omap3isp/isppreview.c 	preview_setup_hw(prev, update, active);
prev             1538 drivers/media/platform/omap3isp/isppreview.c 	preview_config_input_size(prev, active);
prev             1540 drivers/media/platform/omap3isp/isppreview.c 	if (prev->input == PREVIEW_INPUT_MEMORY ||
prev             1541 drivers/media/platform/omap3isp/isppreview.c 	    prev->output & PREVIEW_OUTPUT_MEMORY)
prev             1542 drivers/media/platform/omap3isp/isppreview.c 		preview_isr_buffer(prev);
prev             1543 drivers/media/platform/omap3isp/isppreview.c 	else if (prev->state == ISP_PIPELINE_STREAM_CONTINUOUS)
prev             1544 drivers/media/platform/omap3isp/isppreview.c 		preview_enable_oneshot(prev);
prev             1546 drivers/media/platform/omap3isp/isppreview.c 	spin_lock_irqsave(&prev->params.lock, flags);
prev             1547 drivers/media/platform/omap3isp/isppreview.c 	preview_params_unlock(prev, update, false);
prev             1548 drivers/media/platform/omap3isp/isppreview.c 	spin_unlock_irqrestore(&prev->params.lock, flags);
prev             1558 drivers/media/platform/omap3isp/isppreview.c 	struct isp_prev_device *prev = &video->isp->isp_prev;
prev             1561 drivers/media/platform/omap3isp/isppreview.c 		preview_set_inaddr(prev, buffer->dma);
prev             1564 drivers/media/platform/omap3isp/isppreview.c 		preview_set_outaddr(prev, buffer->dma);
prev             1583 drivers/media/platform/omap3isp/isppreview.c 	struct isp_prev_device *prev =
prev             1588 drivers/media/platform/omap3isp/isppreview.c 		preview_update_brightness(prev, ctrl->val);
prev             1591 drivers/media/platform/omap3isp/isppreview.c 		preview_update_contrast(prev, ctrl->val);
prev             1611 drivers/media/platform/omap3isp/isppreview.c 	struct isp_prev_device *prev = v4l2_get_subdevdata(sd);
prev             1615 drivers/media/platform/omap3isp/isppreview.c 		return preview_config(prev, arg);
prev             1630 drivers/media/platform/omap3isp/isppreview.c 	struct isp_prev_device *prev = v4l2_get_subdevdata(sd);
prev             1631 drivers/media/platform/omap3isp/isppreview.c 	struct isp_video *video_out = &prev->video_out;
prev             1632 drivers/media/platform/omap3isp/isppreview.c 	struct isp_device *isp = to_isp_device(prev);
prev             1633 drivers/media/platform/omap3isp/isppreview.c 	struct device *dev = to_device(prev);
prev             1635 drivers/media/platform/omap3isp/isppreview.c 	if (prev->state == ISP_PIPELINE_STREAM_STOPPED) {
prev             1640 drivers/media/platform/omap3isp/isppreview.c 		preview_configure(prev);
prev             1641 drivers/media/platform/omap3isp/isppreview.c 		atomic_set(&prev->stopping, 0);
prev             1642 drivers/media/platform/omap3isp/isppreview.c 		preview_print_status(prev);
prev             1647 drivers/media/platform/omap3isp/isppreview.c 		if (prev->output & PREVIEW_OUTPUT_MEMORY)
prev             1651 drivers/media/platform/omap3isp/isppreview.c 		    !(prev->output & PREVIEW_OUTPUT_MEMORY))
prev             1652 drivers/media/platform/omap3isp/isppreview.c 			preview_enable_oneshot(prev);
prev             1658 drivers/media/platform/omap3isp/isppreview.c 		if (prev->input == PREVIEW_INPUT_MEMORY)
prev             1660 drivers/media/platform/omap3isp/isppreview.c 		if (prev->output & PREVIEW_OUTPUT_MEMORY)
prev             1663 drivers/media/platform/omap3isp/isppreview.c 		preview_enable_oneshot(prev);
prev             1667 drivers/media/platform/omap3isp/isppreview.c 		if (omap3isp_module_sync_idle(&sd->entity, &prev->wait,
prev             1668 drivers/media/platform/omap3isp/isppreview.c 					      &prev->stopping))
prev             1677 drivers/media/platform/omap3isp/isppreview.c 	prev->state = enable;
prev             1682 drivers/media/platform/omap3isp/isppreview.c __preview_get_format(struct isp_prev_device *prev, struct v4l2_subdev_pad_config *cfg,
prev             1686 drivers/media/platform/omap3isp/isppreview.c 		return v4l2_subdev_get_try_format(&prev->subdev, cfg, pad);
prev             1688 drivers/media/platform/omap3isp/isppreview.c 		return &prev->formats[pad];
prev             1692 drivers/media/platform/omap3isp/isppreview.c __preview_get_crop(struct isp_prev_device *prev, struct v4l2_subdev_pad_config *cfg,
prev             1696 drivers/media/platform/omap3isp/isppreview.c 		return v4l2_subdev_get_try_crop(&prev->subdev, cfg, PREV_PAD_SINK);
prev             1698 drivers/media/platform/omap3isp/isppreview.c 		return &prev->crop;
prev             1731 drivers/media/platform/omap3isp/isppreview.c static void preview_try_format(struct isp_prev_device *prev,
prev             1751 drivers/media/platform/omap3isp/isppreview.c 		if (prev->input == PREVIEW_INPUT_MEMORY) {
prev             1753 drivers/media/platform/omap3isp/isppreview.c 					     preview_max_out_width(prev));
prev             1773 drivers/media/platform/omap3isp/isppreview.c 		*fmt = *__preview_get_format(prev, cfg, PREV_PAD_SINK, which);
prev             1791 drivers/media/platform/omap3isp/isppreview.c 		crop = __preview_get_crop(prev, cfg, which);
prev             1814 drivers/media/platform/omap3isp/isppreview.c static void preview_try_crop(struct isp_prev_device *prev,
prev             1828 drivers/media/platform/omap3isp/isppreview.c 	if (prev->input == PREVIEW_INPUT_CCDC) {
prev             1892 drivers/media/platform/omap3isp/isppreview.c 	struct isp_prev_device *prev = v4l2_get_subdevdata(sd);
prev             1901 drivers/media/platform/omap3isp/isppreview.c 	preview_try_format(prev, cfg, fse->pad, &format, fse->which);
prev             1911 drivers/media/platform/omap3isp/isppreview.c 	preview_try_format(prev, cfg, fse->pad, &format, fse->which);
prev             1932 drivers/media/platform/omap3isp/isppreview.c 	struct isp_prev_device *prev = v4l2_get_subdevdata(sd);
prev             1945 drivers/media/platform/omap3isp/isppreview.c 		format = __preview_get_format(prev, cfg, PREV_PAD_SINK,
prev             1947 drivers/media/platform/omap3isp/isppreview.c 		preview_try_crop(prev, format, &sel->r);
prev             1951 drivers/media/platform/omap3isp/isppreview.c 		sel->r = *__preview_get_crop(prev, cfg, sel->which);
prev             1975 drivers/media/platform/omap3isp/isppreview.c 	struct isp_prev_device *prev = v4l2_get_subdevdata(sd);
prev             1983 drivers/media/platform/omap3isp/isppreview.c 	if (prev->state != ISP_PIPELINE_STREAM_STOPPED)
prev             1991 drivers/media/platform/omap3isp/isppreview.c 		sel->r = *__preview_get_crop(prev, cfg, sel->which);
prev             1995 drivers/media/platform/omap3isp/isppreview.c 	format = __preview_get_format(prev, cfg, PREV_PAD_SINK, sel->which);
prev             1996 drivers/media/platform/omap3isp/isppreview.c 	preview_try_crop(prev, format, &sel->r);
prev             1997 drivers/media/platform/omap3isp/isppreview.c 	*__preview_get_crop(prev, cfg, sel->which) = sel->r;
prev             2000 drivers/media/platform/omap3isp/isppreview.c 	format = __preview_get_format(prev, cfg, PREV_PAD_SOURCE, sel->which);
prev             2001 drivers/media/platform/omap3isp/isppreview.c 	preview_try_format(prev, cfg, PREV_PAD_SOURCE, format, sel->which);
prev             2016 drivers/media/platform/omap3isp/isppreview.c 	struct isp_prev_device *prev = v4l2_get_subdevdata(sd);
prev             2019 drivers/media/platform/omap3isp/isppreview.c 	format = __preview_get_format(prev, cfg, fmt->pad, fmt->which);
prev             2037 drivers/media/platform/omap3isp/isppreview.c 	struct isp_prev_device *prev = v4l2_get_subdevdata(sd);
prev             2041 drivers/media/platform/omap3isp/isppreview.c 	format = __preview_get_format(prev, cfg, fmt->pad, fmt->which);
prev             2045 drivers/media/platform/omap3isp/isppreview.c 	preview_try_format(prev, cfg, fmt->pad, &fmt->format, fmt->which);
prev             2051 drivers/media/platform/omap3isp/isppreview.c 		crop = __preview_get_crop(prev, cfg, fmt->which);
prev             2057 drivers/media/platform/omap3isp/isppreview.c 		preview_try_crop(prev, &fmt->format, crop);
prev             2060 drivers/media/platform/omap3isp/isppreview.c 		format = __preview_get_format(prev, cfg, PREV_PAD_SOURCE,
prev             2062 drivers/media/platform/omap3isp/isppreview.c 		preview_try_format(prev, cfg, PREV_PAD_SOURCE, format,
prev             2143 drivers/media/platform/omap3isp/isppreview.c 	struct isp_prev_device *prev = v4l2_get_subdevdata(sd);
prev             2154 drivers/media/platform/omap3isp/isppreview.c 			if (prev->input == PREVIEW_INPUT_CCDC)
prev             2156 drivers/media/platform/omap3isp/isppreview.c 			prev->input = PREVIEW_INPUT_MEMORY;
prev             2158 drivers/media/platform/omap3isp/isppreview.c 			if (prev->input == PREVIEW_INPUT_MEMORY)
prev             2159 drivers/media/platform/omap3isp/isppreview.c 				prev->input = PREVIEW_INPUT_NONE;
prev             2166 drivers/media/platform/omap3isp/isppreview.c 			if (prev->input == PREVIEW_INPUT_MEMORY)
prev             2168 drivers/media/platform/omap3isp/isppreview.c 			prev->input = PREVIEW_INPUT_CCDC;
prev             2170 drivers/media/platform/omap3isp/isppreview.c 			if (prev->input == PREVIEW_INPUT_CCDC)
prev             2171 drivers/media/platform/omap3isp/isppreview.c 				prev->input = PREVIEW_INPUT_NONE;
prev             2183 drivers/media/platform/omap3isp/isppreview.c 			if (prev->output & ~PREVIEW_OUTPUT_MEMORY)
prev             2185 drivers/media/platform/omap3isp/isppreview.c 			prev->output |= PREVIEW_OUTPUT_MEMORY;
prev             2187 drivers/media/platform/omap3isp/isppreview.c 			prev->output &= ~PREVIEW_OUTPUT_MEMORY;
prev             2194 drivers/media/platform/omap3isp/isppreview.c 			if (prev->output & ~PREVIEW_OUTPUT_RESIZER)
prev             2196 drivers/media/platform/omap3isp/isppreview.c 			prev->output |= PREVIEW_OUTPUT_RESIZER;
prev             2198 drivers/media/platform/omap3isp/isppreview.c 			prev->output &= ~PREVIEW_OUTPUT_RESIZER;
prev             2215 drivers/media/platform/omap3isp/isppreview.c void omap3isp_preview_unregister_entities(struct isp_prev_device *prev)
prev             2217 drivers/media/platform/omap3isp/isppreview.c 	v4l2_device_unregister_subdev(&prev->subdev);
prev             2218 drivers/media/platform/omap3isp/isppreview.c 	omap3isp_video_unregister(&prev->video_in);
prev             2219 drivers/media/platform/omap3isp/isppreview.c 	omap3isp_video_unregister(&prev->video_out);
prev             2222 drivers/media/platform/omap3isp/isppreview.c int omap3isp_preview_register_entities(struct isp_prev_device *prev,
prev             2228 drivers/media/platform/omap3isp/isppreview.c 	prev->subdev.dev = vdev->mdev->dev;
prev             2229 drivers/media/platform/omap3isp/isppreview.c 	ret = v4l2_device_register_subdev(vdev, &prev->subdev);
prev             2233 drivers/media/platform/omap3isp/isppreview.c 	ret = omap3isp_video_register(&prev->video_in, vdev);
prev             2237 drivers/media/platform/omap3isp/isppreview.c 	ret = omap3isp_video_register(&prev->video_out, vdev);
prev             2244 drivers/media/platform/omap3isp/isppreview.c 	omap3isp_preview_unregister_entities(prev);
prev             2257 drivers/media/platform/omap3isp/isppreview.c static int preview_init_entities(struct isp_prev_device *prev)
prev             2259 drivers/media/platform/omap3isp/isppreview.c 	struct v4l2_subdev *sd = &prev->subdev;
prev             2260 drivers/media/platform/omap3isp/isppreview.c 	struct media_pad *pads = prev->pads;
prev             2264 drivers/media/platform/omap3isp/isppreview.c 	prev->input = PREVIEW_INPUT_NONE;
prev             2270 drivers/media/platform/omap3isp/isppreview.c 	v4l2_set_subdevdata(sd, prev);
prev             2273 drivers/media/platform/omap3isp/isppreview.c 	v4l2_ctrl_handler_init(&prev->ctrls, 2);
prev             2274 drivers/media/platform/omap3isp/isppreview.c 	v4l2_ctrl_new_std(&prev->ctrls, &preview_ctrl_ops, V4L2_CID_BRIGHTNESS,
prev             2277 drivers/media/platform/omap3isp/isppreview.c 	v4l2_ctrl_new_std(&prev->ctrls, &preview_ctrl_ops, V4L2_CID_CONTRAST,
prev             2280 drivers/media/platform/omap3isp/isppreview.c 	v4l2_ctrl_handler_setup(&prev->ctrls);
prev             2281 drivers/media/platform/omap3isp/isppreview.c 	sd->ctrl_handler = &prev->ctrls;
prev             2298 drivers/media/platform/omap3isp/isppreview.c 	prev->video_in.type = V4L2_BUF_TYPE_VIDEO_OUTPUT;
prev             2299 drivers/media/platform/omap3isp/isppreview.c 	prev->video_in.ops = &preview_video_ops;
prev             2300 drivers/media/platform/omap3isp/isppreview.c 	prev->video_in.isp = to_isp_device(prev);
prev             2301 drivers/media/platform/omap3isp/isppreview.c 	prev->video_in.capture_mem = PAGE_ALIGN(4096 * 4096) * 2 * 3;
prev             2302 drivers/media/platform/omap3isp/isppreview.c 	prev->video_in.bpl_alignment = 64;
prev             2303 drivers/media/platform/omap3isp/isppreview.c 	prev->video_out.type = V4L2_BUF_TYPE_VIDEO_CAPTURE;
prev             2304 drivers/media/platform/omap3isp/isppreview.c 	prev->video_out.ops = &preview_video_ops;
prev             2305 drivers/media/platform/omap3isp/isppreview.c 	prev->video_out.isp = to_isp_device(prev);
prev             2306 drivers/media/platform/omap3isp/isppreview.c 	prev->video_out.capture_mem = PAGE_ALIGN(4096 * 4096) * 2 * 3;
prev             2307 drivers/media/platform/omap3isp/isppreview.c 	prev->video_out.bpl_alignment = 32;
prev             2309 drivers/media/platform/omap3isp/isppreview.c 	ret = omap3isp_video_init(&prev->video_in, "preview");
prev             2313 drivers/media/platform/omap3isp/isppreview.c 	ret = omap3isp_video_init(&prev->video_out, "preview");
prev             2320 drivers/media/platform/omap3isp/isppreview.c 	omap3isp_video_cleanup(&prev->video_in);
prev             2322 drivers/media/platform/omap3isp/isppreview.c 	media_entity_cleanup(&prev->subdev.entity);
prev             2333 drivers/media/platform/omap3isp/isppreview.c 	struct isp_prev_device *prev = &isp->isp_prev;
prev             2335 drivers/media/platform/omap3isp/isppreview.c 	init_waitqueue_head(&prev->wait);
prev             2337 drivers/media/platform/omap3isp/isppreview.c 	preview_init_params(prev);
prev             2339 drivers/media/platform/omap3isp/isppreview.c 	return preview_init_entities(prev);
prev             2344 drivers/media/platform/omap3isp/isppreview.c 	struct isp_prev_device *prev = &isp->isp_prev;
prev             2346 drivers/media/platform/omap3isp/isppreview.c 	v4l2_ctrl_handler_free(&prev->ctrls);
prev             2347 drivers/media/platform/omap3isp/isppreview.c 	omap3isp_video_cleanup(&prev->video_in);
prev             2348 drivers/media/platform/omap3isp/isppreview.c 	omap3isp_video_cleanup(&prev->video_out);
prev             2349 drivers/media/platform/omap3isp/isppreview.c 	media_entity_cleanup(&prev->subdev.entity);
prev              154 drivers/media/platform/omap3isp/isppreview.h void omap3isp_preview_isr_frame_sync(struct isp_prev_device *prev);
prev              155 drivers/media/platform/omap3isp/isppreview.h void omap3isp_preview_isr(struct isp_prev_device *prev);
prev             1089 drivers/media/platform/pxa_camera.c 	last_buf = list_entry(pcdev->capture.prev,
prev              830 drivers/media/platform/s5p-mfc/s5p_mfc_dec.c 			buf = list_entry(ctx->src_queue.prev,
prev             2294 drivers/media/platform/s5p-mfc/s5p_mfc_enc.c 			buf = list_entry(ctx->src_queue.prev,
prev              125 drivers/media/platform/sti/hva/hva-debugfs.c 	ktime_t prev = dbg->begin;
prev              131 drivers/media/platform/sti/hva/hva-debugfs.c 		div = (u64)ktime_us_delta(dbg->begin, prev);
prev             1577 drivers/media/platform/ti-vpe/cal.c 		 struct device_node *prev)
prev             1584 drivers/media/platform/ti-vpe/cal.c 	if (!prev) {
prev             1601 drivers/media/platform/ti-vpe/cal.c 		ports = of_get_parent(prev);
prev             1606 drivers/media/platform/ti-vpe/cal.c 			port = of_get_next_child(ports, prev);
prev             1611 drivers/media/platform/ti-vpe/cal.c 			prev = port;
prev             1621 drivers/media/platform/ti-vpe/cal.c 		     struct device_node *prev)
prev             1629 drivers/media/platform/ti-vpe/cal.c 		ep = of_get_next_child(parent, prev);
prev             1632 drivers/media/platform/ti-vpe/cal.c 		prev = ep;
prev               66 drivers/media/platform/vsp1/vsp1_drm.c 			      struct vsp1_entity *prev, unsigned int prev_pad,
prev               77 drivers/media/platform/vsp1/vsp1_drm.c 		prev->sink = next;
prev               78 drivers/media/platform/vsp1/vsp1_drm.c 		prev->sink_pad = next_pad;
prev               82 drivers/media/platform/vsp1/vsp1_drm.c 	prev->sink = uif;
prev               83 drivers/media/platform/vsp1/vsp1_drm.c 	prev->sink_pad = UIF_PAD_SINK;
prev               89 drivers/media/platform/vsp1/vsp1_drm.c 	ret = v4l2_subdev_call(&prev->subdev, pad, get_fmt, NULL, &format);
prev               92 drivers/media/usb/hdpvr/hdpvr-video.c 	list_splice_init(&dev->rec_buff_list, dev->free_buff_list.prev);
prev             1496 drivers/media/usb/uvc/uvc_driver.c 	struct uvc_entity *entity, struct uvc_entity *prev)
prev             1510 drivers/media/usb/uvc/uvc_driver.c 		if (forward == prev)
prev             1512 drivers/media/usb/uvc/uvc_driver.c 		if (forward->chain.next || forward->chain.prev) {
prev             1598 drivers/media/usb/uvc/uvc_driver.c 			if (term->chain.next || term->chain.prev) {
prev             1648 drivers/media/usb/uvc/uvc_driver.c 	struct uvc_entity *entity, *prev;
prev             1653 drivers/media/usb/uvc/uvc_driver.c 	prev = NULL;
prev             1657 drivers/media/usb/uvc/uvc_driver.c 		if (entity->chain.next || entity->chain.prev) {
prev             1668 drivers/media/usb/uvc/uvc_driver.c 		if (uvc_scan_chain_forward(chain, entity, prev) < 0)
prev             1672 drivers/media/usb/uvc/uvc_driver.c 		prev = entity;
prev             1751 drivers/media/usb/uvc/uvc_driver.c 	struct uvc_entity *prev;
prev             1782 drivers/media/usb/uvc/uvc_driver.c 	prev = oterm;
prev             1802 drivers/media/usb/uvc/uvc_driver.c 		prev->baSourceID[0] = entity->id;
prev             1803 drivers/media/usb/uvc/uvc_driver.c 		prev = entity;
prev             1809 drivers/media/usb/uvc/uvc_driver.c 	prev->baSourceID[0] = iterm->id;
prev             1843 drivers/media/usb/uvc/uvc_driver.c 		if (term->chain.next || term->chain.prev)
prev             2310 drivers/media/v4l2-core/v4l2-ctrls.c 	if (list_empty(&hdl->ctrl_refs) || id > node2id(hdl->ctrl_refs.prev)) {
prev             2324 drivers/media/v4l2-core/v4l2-ctrls.c 		list_add(&new_ref->node, ref->node.prev);
prev             2966 drivers/media/v4l2-core/v4l2-ctrls.c 		if (id >= node2id(hdl->ctrl_refs.prev)) {
prev              407 drivers/misc/habanalabs/memory.c 	struct hl_vm_va_block *prev, *next;
prev              409 drivers/misc/habanalabs/memory.c 	prev = list_prev_entry(va_block, node);
prev              410 drivers/misc/habanalabs/memory.c 	if (&prev->node != va_list && prev->end + 1 == va_block->start) {
prev              411 drivers/misc/habanalabs/memory.c 		prev->end = va_block->end;
prev              412 drivers/misc/habanalabs/memory.c 		prev->size = prev->end - prev->start;
prev              415 drivers/misc/habanalabs/memory.c 		va_block = prev;
prev               22 drivers/misc/mic/scif/scif_rma_list.c 	struct scif_window *prev = list_entry(head, struct scif_window, list);
prev               28 drivers/misc/mic/scif/scif_rma_list.c 		curr = list_entry(head->prev, struct scif_window, list);
prev               38 drivers/misc/mic/scif/scif_rma_list.c 		prev = curr;
prev               40 drivers/misc/mic/scif/scif_rma_list.c 	list_add(&window->list, &prev->list);
prev               51 drivers/misc/mic/scif/scif_rma_list.c 	struct scif_window *curr = NULL, *prev = NULL;
prev               59 drivers/misc/mic/scif/scif_rma_list.c 		prev = curr;
prev               61 drivers/misc/mic/scif/scif_rma_list.c 	if (!prev)
prev               64 drivers/misc/mic/scif/scif_rma_list.c 		list_add(&window->list, &prev->list);
prev              233 drivers/mmc/core/sdio_cis.c 	struct sdio_func_tuple *this, **prev;
prev              257 drivers/mmc/core/sdio_cis.c 		prev = &func->tuples;
prev              259 drivers/mmc/core/sdio_cis.c 		prev = &card->tuples;
prev              261 drivers/mmc/core/sdio_cis.c 	if (*prev)
prev              314 drivers/mmc/core/sdio_cis.c 			*prev = this;
prev              315 drivers/mmc/core/sdio_cis.c 			prev = &this->next;
prev              344 drivers/mmc/core/sdio_cis.c 		*prev = card->tuples;
prev             1795 drivers/mtd/nand/onenand/onenand_base.c 	int prev = 0, prevlen = 0, prev_subpage = 0, first = 1;
prev             1879 drivers/mtd/nand/onenand/onenand_base.c 			onenand_update_bufferram(mtd, prev, !ret && !prev_subpage);
prev             1945 drivers/mtd/nand/onenand/onenand_base.c 		prev = to;
prev              189 drivers/mtd/parsers/redboot.c 		struct fis_list *new_fl, **prev;
prev              216 drivers/mtd/parsers/redboot.c 		prev = &fl;
prev              217 drivers/mtd/parsers/redboot.c 		while(*prev && (*prev)->img->flash_base < new_fl->img->flash_base)
prev              218 drivers/mtd/parsers/redboot.c 			prev = &(*prev)->next;
prev              219 drivers/mtd/parsers/redboot.c 		new_fl->next = *prev;
prev              220 drivers/mtd/parsers/redboot.c 		*prev = new_fl;
prev              566 drivers/mtd/spi-nor/aspeed-smc.c 		u32 prev = readl(SEGMENT_ADDR_REG(controller, chip->cs - 1));
prev              568 drivers/mtd/spi-nor/aspeed-smc.c 		start = SEGMENT_ADDR_END(prev);
prev               84 drivers/net/bonding/bond_alb.c 	entry->prev = TLB_NULL_INDEX;
prev              209 drivers/net/bonding/bond_alb.c 			hash_table[hash_index].prev = TLB_NULL_INDEX;
prev              212 drivers/net/bonding/bond_alb.c 				hash_table[next_index].prev = hash_index;
prev              343 drivers/net/ethernet/brocade/bna/bnad.c 	struct bnad_rx_unmap *unmap, *prev;
prev              363 drivers/net/ethernet/brocade/bna/bnad.c 			prev = &unmap_q->unmap[unmap_q->reuse_pi];
prev              364 drivers/net/ethernet/brocade/bna/bnad.c 			page = prev->page;
prev              365 drivers/net/ethernet/brocade/bna/bnad.c 			page_offset = prev->page_offset + unmap_q->map_size;
prev              615 drivers/net/ethernet/cavium/liquidio/octeon_network.h 	if (root->prev == root && root->next == root)
prev              840 drivers/net/ethernet/cisco/enic/enic_main.c 		buf = wq->to_use->prev;
prev              847 drivers/net/ethernet/cisco/enic/enic_main.c 			buf = buf->prev;
prev               51 drivers/net/ethernet/cisco/enic/vnic_wq.c 				buf->next->prev = buf;
prev               55 drivers/net/ethernet/cisco/enic/vnic_wq.c 				buf->next->prev = buf;
prev               58 drivers/net/ethernet/cisco/enic/vnic_wq.c 				buf->next->prev = buf;
prev               65 drivers/net/ethernet/cisco/enic/vnic_wq.h 	struct vnic_wq_buf *prev;
prev               56 drivers/net/ethernet/freescale/dpaa/dpaa_eth_sysfs.c 	struct dpaa_fq *prev = NULL;
prev               93 drivers/net/ethernet/freescale/dpaa/dpaa_eth_sysfs.c 		if (prev && (abs(fq->fqid - prev->fqid) != 1 ||
prev               97 drivers/net/ethernet/freescale/dpaa/dpaa_eth_sysfs.c 					"%s: %d\n", prevstr, prev->fqid);
prev              104 drivers/net/ethernet/freescale/dpaa/dpaa_eth_sysfs.c 		if (prev && abs(fq->fqid - prev->fqid) == 1 &&
prev              112 drivers/net/ethernet/freescale/dpaa/dpaa_eth_sysfs.c 		prev = fq;
prev              117 drivers/net/ethernet/freescale/dpaa/dpaa_eth_sysfs.c 	if (prev) {
prev              120 drivers/net/ethernet/freescale/dpaa/dpaa_eth_sysfs.c 					prev->fqid);
prev             1250 drivers/net/ethernet/i825xx/lib82596.c 			struct i596_cmd *prev = ptr;
prev             1254 drivers/net/ethernet/i825xx/lib82596.c 			DMA_WBACK_INV(dev, prev, sizeof(struct i596_cmd));
prev              420 drivers/net/ethernet/intel/e100.c 	struct rx *next, *prev;
prev              499 drivers/net/ethernet/intel/e100.c 	struct cb *next, *prev;
prev              878 drivers/net/ethernet/intel/e100.c 	cb->prev->command &= cpu_to_le16(~cb_s);
prev             1672 drivers/net/ethernet/intel/e100.c 		u32 prev = nic->adaptive_ifs;
prev             1683 drivers/net/ethernet/intel/e100.c 		if (nic->adaptive_ifs != prev)
prev             1895 drivers/net/ethernet/intel/e100.c 		cb->prev = (i == 0) ? nic->cbs + count - 1 : cb - 1;
prev             1944 drivers/net/ethernet/intel/e100.c 	if (rx->prev->skb) {
prev             1945 drivers/net/ethernet/intel/e100.c 		struct rfd *prev_rfd = (struct rfd *)rx->prev->skb->data;
prev             1947 drivers/net/ethernet/intel/e100.c 		pci_dma_sync_single_for_device(nic->pdev, rx->prev->dma_addr,
prev             2079 drivers/net/ethernet/intel/e100.c 	old_before_last_rx = nic->rx_to_use->prev->prev;
prev             2088 drivers/net/ethernet/intel/e100.c 	new_before_last_rx = nic->rx_to_use->prev->prev;
prev             2166 drivers/net/ethernet/intel/e100.c 		rx->prev = (i == 0) ? nic->rxs + count - 1 : rx - 1;
prev             2179 drivers/net/ethernet/intel/e100.c 	rx = nic->rxs->prev->prev;
prev              661 drivers/net/ethernet/intel/ice/ice_sched.c 	struct ice_sched_node *prev, *new_node;
prev              726 drivers/net/ethernet/intel/ice/ice_sched.c 		prev = ice_sched_get_first_node(pi, tc_node, layer);
prev              727 drivers/net/ethernet/intel/ice/ice_sched.c 		if (prev && prev != new_node) {
prev              728 drivers/net/ethernet/intel/ice/ice_sched.c 			while (prev->sibling)
prev              729 drivers/net/ethernet/intel/ice/ice_sched.c 				prev = prev->sibling;
prev              730 drivers/net/ethernet/intel/ice/ice_sched.c 			prev->sibling = new_node;
prev              697 drivers/net/ethernet/mellanox/mlx4/mcg.c 		      int *prev, int *index)
prev              723 drivers/net/ethernet/mellanox/mlx4/mcg.c 	*prev  = -1;
prev              742 drivers/net/ethernet/mellanox/mlx4/mcg.c 		*prev = *index;
prev             1112 drivers/net/ethernet/mellanox/mlx4/mcg.c 	int index = -1, prev;
prev             1126 drivers/net/ethernet/mellanox/mlx4/mcg.c 			 mailbox, &prev, &index);
prev             1180 drivers/net/ethernet/mellanox/mlx4/mcg.c 	err = mlx4_READ_ENTRY(dev, prev, mailbox);
prev             1186 drivers/net/ethernet/mellanox/mlx4/mcg.c 	err = mlx4_WRITE_ENTRY(dev, prev, mailbox);
prev             1221 drivers/net/ethernet/mellanox/mlx4/mcg.c 	int prev, index;
prev             1235 drivers/net/ethernet/mellanox/mlx4/mcg.c 			 mailbox, &prev, &index);
prev             1282 drivers/net/ethernet/mellanox/mlx4/mcg.c 	if (prev == -1) {
prev             1307 drivers/net/ethernet/mellanox/mlx4/mcg.c 		err = mlx4_READ_ENTRY(dev, prev, mailbox);
prev             1313 drivers/net/ethernet/mellanox/mlx4/mcg.c 		err = mlx4_WRITE_ENTRY(dev, prev, mailbox);
prev             1319 drivers/net/ethernet/mellanox/mlx4/mcg.c 				  prev, index, dev->caps.num_mgms);
prev              316 drivers/net/ethernet/mellanox/mlx5/core/en_main.c 	struct mlx5e_wqe_frag_info *prev = NULL;
prev              331 drivers/net/ethernet/mellanox/mlx5/core/en_main.c 				if (prev)
prev              332 drivers/net/ethernet/mellanox/mlx5/core/en_main.c 					prev->last_in_page = true;
prev              338 drivers/net/ethernet/mellanox/mlx5/core/en_main.c 			prev = frag;
prev              342 drivers/net/ethernet/mellanox/mlx5/core/en_main.c 	if (prev)
prev              343 drivers/net/ethernet/mellanox/mlx5/core/en_main.c 		prev->last_in_page = true;
prev              679 drivers/net/ethernet/mellanox/mlx5/core/fs_core.c 						       struct list_head *prev)
prev              702 drivers/net/ethernet/mellanox/mlx5/core/fs_core.c 	list_add(&fg->node.list, prev);
prev              992 drivers/net/ethernet/mellanox/mlx5/core/fs_core.c 	struct list_head *prev = &prio->node.children;
prev              998 drivers/net/ethernet/mellanox/mlx5/core/fs_core.c 		prev = &iter->node.list;
prev             1000 drivers/net/ethernet/mellanox/mlx5/core/fs_core.c 	list_add(&ft->node.list, prev);
prev             1158 drivers/net/ethernet/mellanox/mlx5/core/fs_core.c 				     ft->node.children.prev);
prev             1324 drivers/net/ethernet/mellanox/mlx5/core/fs_core.c 	struct list_head *prev = &ft->node.children;
prev             1345 drivers/net/ethernet/mellanox/mlx5/core/fs_core.c 		prev = &fg->node.list;
prev             1356 drivers/net/ethernet/mellanox/mlx5/core/fs_core.c 				     prev);
prev              723 drivers/net/ethernet/mellanox/mlx5/core/health.c 	if (count == health->prev)
prev              728 drivers/net/ethernet/mellanox/mlx5/core/health.c 	health->prev = count;
prev              722 drivers/net/ethernet/mellanox/mlx5/core/steering/dr_matcher.c 	if (matcher->matcher_list.prev == &tbl->matcher_list)
prev               91 drivers/net/ethernet/mellanox/mlxsw/core_acl_flex_actions.c 	struct mlxsw_afa_set *prev; /* Pointer to the previous set,
prev              218 drivers/net/ethernet/mellanox/mlxsw/core_acl_flex_actions.c 	set->prev = NULL;
prev              335 drivers/net/ethernet/mellanox/mlxsw/core_acl_flex_actions.c 		block->cur_set->prev = block->first_set;
prev              380 drivers/net/ethernet/mellanox/mlxsw/core_acl_flex_actions.c 		prev_set = set->prev;
prev              651 drivers/net/ethernet/mellanox/mlxsw/core_acl_flex_actions.c 		set->prev = block->cur_set;
prev              612 drivers/net/ethernet/mscc/ocelot_ace.c 	list_add(&rule->list, pos->prev);
prev               44 drivers/net/ethernet/netronome/nfp/bpf/jit.c 	return meta->l.prev != &nfp_prog->insns;
prev             2581 drivers/net/ethernet/netronome/nfp/bpf/jit.c 	struct nfp_insn_meta *prev = nfp_meta_prev(meta);
prev             2585 drivers/net/ethernet/netronome/nfp/bpf/jit.c 	dst = prev->insn.dst_reg * 2;
prev             2586 drivers/net/ethernet/netronome/nfp/bpf/jit.c 	imm_lo = prev->insn.imm;
prev              857 drivers/net/ethernet/netronome/nfp/nfpcore/nfp_cppcore.c 	cache = list_entry(cpp->area_cache_list.prev,
prev              146 drivers/net/ethernet/qlogic/qlcnic/qlcnic_83xx_init.c 	u32 cur, prev;
prev              148 drivers/net/ethernet/qlogic/qlcnic/qlcnic_83xx_init.c 	prev = adapter->ahw->idc.prev_state;
prev              153 drivers/net/ethernet/qlogic/qlcnic/qlcnic_83xx_init.c 		 adapter->ahw->idc.name[prev]);
prev              984 drivers/net/ethernet/qlogic/qlcnic/qlcnic_83xx_init.c 	u32 cur, prev, next;
prev              987 drivers/net/ethernet/qlogic/qlcnic/qlcnic_83xx_init.c 	prev = adapter->ahw->idc.prev_state;
prev              994 drivers/net/ethernet/qlogic/qlcnic/qlcnic_83xx_init.c 			__func__, cur, prev, state);
prev              999 drivers/net/ethernet/qlogic/qlcnic/qlcnic_83xx_init.c 	    (prev == QLC_83XX_IDC_DEV_UNKNOWN)) {
prev             1004 drivers/net/ethernet/qlogic/qlcnic/qlcnic_83xx_init.c 				__func__, cur, prev, next);
prev             1010 drivers/net/ethernet/qlogic/qlcnic/qlcnic_83xx_init.c 		if ((prev != QLC_83XX_IDC_DEV_INIT) &&
prev             1011 drivers/net/ethernet/qlogic/qlcnic/qlcnic_83xx_init.c 		    (prev != QLC_83XX_IDC_DEV_COLD) &&
prev             1012 drivers/net/ethernet/qlogic/qlcnic/qlcnic_83xx_init.c 		    (prev != QLC_83XX_IDC_DEV_NEED_RESET)) {
prev             1015 drivers/net/ethernet/qlogic/qlcnic/qlcnic_83xx_init.c 				__func__, cur, prev, next);
prev              233 drivers/net/ethernet/sfc/falcon/mdio_10g.c 	struct ethtool_link_ksettings prev = {
prev              239 drivers/net/ethernet/sfc/falcon/mdio_10g.c 	efx->phy_op->get_link_ksettings(efx, &prev);
prev              244 drivers/net/ethernet/sfc/falcon/mdio_10g.c 						prev.link_modes.advertising);
prev              246 drivers/net/ethernet/sfc/falcon/mdio_10g.c 						prev.link_modes.supported);
prev              249 drivers/net/ethernet/sfc/falcon/mdio_10g.c 	    cmd->base.speed == prev.base.speed &&
prev              250 drivers/net/ethernet/sfc/falcon/mdio_10g.c 	    cmd->base.duplex == prev.base.duplex &&
prev              251 drivers/net/ethernet/sfc/falcon/mdio_10g.c 	    cmd->base.port == prev.base.port &&
prev              252 drivers/net/ethernet/sfc/falcon/mdio_10g.c 	    cmd->base.autoneg == prev.base.autoneg)
prev              256 drivers/net/ethernet/sfc/falcon/mdio_10g.c 	if (prev.base.port != PORT_TP || cmd->base.port != PORT_TP)
prev              182 drivers/net/ethernet/ti/cpmac.c 	struct cpmac_desc *prev;
prev              452 drivers/net/ethernet/ti/cpmac.c 		desc->prev->hw_next = (u32)0;
prev              453 drivers/net/ethernet/ti/cpmac.c 		priv->rx_head->prev->hw_next = priv->rx_head->mapping;
prev              465 drivers/net/ethernet/ti/cpmac.c 	     (priv->rx_head->prev->dataflags & (CPMAC_OWN|CPMAC_EOQ))
prev              471 drivers/net/ethernet/ti/cpmac.c 		priv->rx_head->prev->dataflags &= ~CPMAC_EOQ;
prev              696 drivers/net/ethernet/ti/cpmac.c 	priv->rx_head->prev->hw_next = 0;
prev              963 drivers/net/ethernet/ti/cpmac.c 		desc->next->prev = desc;
prev              967 drivers/net/ethernet/ti/cpmac.c 	priv->rx_head->prev->hw_next = (u32)0;
prev              984 drivers/net/ethernet/ti/davinci_cpdma.c 	struct cpdma_desc __iomem	*prev = chan->tail;
prev             1002 drivers/net/ethernet/ti/davinci_cpdma.c 	desc_write(prev, hw_next, desc_dma);
prev             1007 drivers/net/ethernet/ti/davinci_cpdma.c 	mode = desc_read(prev, hw_mode);
prev             1010 drivers/net/ethernet/ti/davinci_cpdma.c 		desc_write(prev, hw_mode, mode & ~CPDMA_DESC_EOQ);
prev              491 drivers/net/ethernet/ti/netcp_core.c 	__list_add(&entry->list, next->list.prev, &next->list);
prev              540 drivers/net/ethernet/ti/netcp_core.c 	__list_add(&entry->list, next->list.prev, &next->list);
prev              330 drivers/net/ethernet/toshiba/ps3_gelic_net.c 		descr->prev = descr - 1;
prev              334 drivers/net/ethernet/toshiba/ps3_gelic_net.c 	start_descr->prev = (descr - 1);
prev              482 drivers/net/ethernet/toshiba/ps3_gelic_net.c 	chain->tail = card->rx_top->prev; /* point to the last */
prev              873 drivers/net/ethernet/toshiba/ps3_gelic_net.c 	descr->prev->next_descr_addr = cpu_to_be32(descr->bus_addr);
prev              891 drivers/net/ethernet/toshiba/ps3_gelic_net.c 		descr->prev->next_descr_addr = 0;
prev             1065 drivers/net/ethernet/toshiba/ps3_gelic_net.c 	descr->prev->next_descr_addr = cpu_to_be32(descr->bus_addr);
prev              247 drivers/net/ethernet/toshiba/ps3_gelic_net.h 	struct gelic_descr *prev;
prev              330 drivers/net/ethernet/toshiba/spider_net.c 		descr->prev = descr - 1;
prev              336 drivers/net/ethernet/toshiba/spider_net.c 	chain->ring->prev = descr-1;
prev              522 drivers/net/ethernet/toshiba/spider_net.c 		descr->prev->hwdescr->next_descr_addr = descr->bus_addr;
prev              663 drivers/net/ethernet/toshiba/spider_net.c 	if (descr->next == chain->tail->prev) {
prev              693 drivers/net/ethernet/toshiba/spider_net.c 	descr->prev->hwdescr->next_descr_addr = descr->bus_addr;
prev              389 drivers/net/ethernet/toshiba/spider_net.h 	struct spider_net_descr *prev;
prev             2521 drivers/net/ethernet/via/via-velocity.c 	int index, prev;
prev             2593 drivers/net/ethernet/via/via-velocity.c 	prev = index - 1;
prev             2594 drivers/net/ethernet/via/via-velocity.c 	if (prev < 0)
prev             2595 drivers/net/ethernet/via/via-velocity.c 		prev = vptr->options.numtx - 1;
prev             2603 drivers/net/ethernet/via/via-velocity.c 	td_ptr = &(vptr->tx.rings[qnum][prev]);
prev               37 drivers/net/fddi/skfp/smttimer.c 	struct smt_timer	**prev ;
prev               47 drivers/net/fddi/skfp/smttimer.c 	for (prev = &smc->t.st_queue ; (tm = *prev) ; prev = &tm->tm_next ) {
prev               49 drivers/net/fddi/skfp/smttimer.c 			*prev = tm->tm_next ;
prev               61 drivers/net/fddi/skfp/smttimer.c 	struct smt_timer	**prev ;
prev               88 drivers/net/fddi/skfp/smttimer.c 	for (prev = &smc->t.st_queue ; (tm = *prev) ; prev = &tm->tm_next ) {
prev               95 drivers/net/fddi/skfp/smttimer.c 	*prev = timer ;
prev             1506 drivers/net/wan/cosa.c 	int i=0, id=0, prev=0, curr=0;
prev             1524 drivers/net/wan/cosa.c 	for (i=0; i<COSA_MAX_ID_STRING-1; i++, prev=curr) {
prev             1531 drivers/net/wan/cosa.c 		if (curr == 0x2e && prev == '\n')
prev              336 drivers/net/wireless/ath/ath9k/channel.c 	struct ath_chanctx *prev, *cur;
prev              344 drivers/net/wireless/ath/ath9k/channel.c 	prev = ath_chanctx_get_next(sc, cur);
prev              346 drivers/net/wireless/ath/ath9k/channel.c 	if (!prev->switch_after_beacon)
prev              353 drivers/net/wireless/ath/ath9k/channel.c 	prev_tsf = prev->last_beacon - (u32) prev->tsf_val + cur_tsf;
prev              354 drivers/net/wireless/ath/ath9k/channel.c 	prev_tsf -= ath9k_hw_get_tsf_offset(&prev->tsf_ts, &ts);
prev              366 drivers/net/wireless/ath/ath9k/channel.c 	prev->tsf_val += offset;
prev              350 drivers/net/wireless/ath/ath9k/xmit.c 	int prev = fi->retries;
prev              355 drivers/net/wireless/ath/ath9k/xmit.c 	if (prev > 0)
prev             1995 drivers/net/wireless/ath/ath9k/xmit.c 	bf_last = list_entry(head->prev, struct ath_buf, list);
prev             2616 drivers/net/wireless/ath/ath9k/xmit.c 				&txq->axq_q, lastbf->list.prev);
prev             2718 drivers/net/wireless/ath/ath9k/xmit.c 						  lastbf->list.prev);
prev              166 drivers/net/wireless/ath/dfs_pri_detector.c 	return list_entry(l->prev, struct pulse_elem, head);
prev              437 drivers/net/wireless/broadcom/brcm80211/brcmfmac/flowring.c 	struct brcmf_flowring_tdls_entry *prev;
prev              446 drivers/net/wireless/broadcom/brcm80211/brcmfmac/flowring.c 	prev = NULL;
prev              452 drivers/net/wireless/broadcom/brcm80211/brcmfmac/flowring.c 		prev = search;
prev              469 drivers/net/wireless/broadcom/brcm80211/brcmfmac/flowring.c 		if (prev)
prev              470 drivers/net/wireless/broadcom/brcm80211/brcmfmac/flowring.c 			prev->next = search->next;
prev             1312 drivers/net/wireless/broadcom/brcm80211/brcmfmac/fwsignal.c 				__skb_insert(p, p_tail->prev, p_tail, queue);
prev             1676 drivers/net/wireless/broadcom/brcm80211/brcmfmac/sdio.c 					   pfirst->prev);
prev             2240 drivers/net/wireless/broadcom/brcm80211/brcmfmac/sdio.c 				pkt_prev = pkt_next->prev;
prev              287 drivers/net/wireless/intel/iwlwifi/dvm/rx.c static void accum_stats(__le32 *prev, __le32 *cur, __le32 *delta,
prev              294 drivers/net/wireless/intel/iwlwifi/dvm/rx.c 	     i++, prev++, cur++, delta++, max_delta++, accum++) {
prev              295 drivers/net/wireless/intel/iwlwifi/dvm/rx.c 		if (le32_to_cpu(*cur) > le32_to_cpu(*prev)) {
prev              297 drivers/net/wireless/intel/iwlwifi/dvm/rx.c 				le32_to_cpu(*cur) - le32_to_cpu(*prev));
prev              875 drivers/net/wireless/intel/iwlwifi/mvm/tx.c 		tmp->prev = NULL;
prev              803 drivers/net/wireless/intel/iwlwifi/mvm/utils.c 	bool prev;
prev              807 drivers/net/wireless/intel/iwlwifi/mvm/utils.c 	prev = iwl_mvm_vif_low_latency(mvmvif);
prev              812 drivers/net/wireless/intel/iwlwifi/mvm/utils.c 	if (low_latency == prev)
prev             1069 drivers/net/wireless/intel/iwlwifi/mvm/utils.c 	bool low_latency, prev = mvmvif->low_latency & LOW_LATENCY_TRAFFIC;
prev             1077 drivers/net/wireless/intel/iwlwifi/mvm/utils.c 	    prev == low_latency) {
prev             1082 drivers/net/wireless/intel/iwlwifi/mvm/utils.c 	if (prev != low_latency) {
prev             1707 drivers/net/wireless/intel/iwlwifi/pcie/rx.c 	bool hw_rfkill, prev, report;
prev             1710 drivers/net/wireless/intel/iwlwifi/pcie/rx.c 	prev = test_bit(STATUS_RFKILL_OPMODE, &trans->status);
prev             1726 drivers/net/wireless/intel/iwlwifi/pcie/rx.c 	if (prev != report)
prev             1072 drivers/net/wireless/intel/iwlwifi/pcie/trans.c 	bool prev = test_bit(STATUS_RFKILL_OPMODE, &trans->status);
prev             1086 drivers/net/wireless/intel/iwlwifi/pcie/trans.c 	if (prev != report)
prev              353 drivers/net/wireless/intersil/hostap/hostap_80211_rx.c 		bss = list_entry(local->bss_list.prev,
prev              378 drivers/net/wireless/intersil/hostap/hostap_80211_rx.c 		bss = list_entry(local->bss_list.prev,
prev             1069 drivers/net/wireless/intersil/hostap/hostap_ap.c 	struct add_sta_proc_data *entry, *prev;
prev             1090 drivers/net/wireless/intersil/hostap/hostap_ap.c 		prev = entry;
prev             1092 drivers/net/wireless/intersil/hostap/hostap_ap.c 		kfree(prev);
prev             1956 drivers/net/wireless/intersil/hostap/hostap_ap.c 	struct wds_oper_data *entry, *prev;
prev             1974 drivers/net/wireless/intersil/hostap/hostap_ap.c 		prev = entry;
prev             1976 drivers/net/wireless/intersil/hostap/hostap_ap.c 		kfree(prev);
prev              260 drivers/net/wireless/intersil/hostap/hostap_info.c 	struct hfa384x_hostscan_result *results, *prev;
prev              290 drivers/net/wireless/intersil/hostap/hostap_info.c 	prev = local->last_scan_results;
prev              294 drivers/net/wireless/intersil/hostap/hostap_info.c 	kfree(prev);
prev              309 drivers/net/wireless/intersil/hostap/hostap_info.c 	struct hfa384x_hostscan_result *results, *prev;
prev              356 drivers/net/wireless/intersil/hostap/hostap_info.c 	prev = local->last_scan_results;
prev              360 drivers/net/wireless/intersil/hostap/hostap_info.c 	kfree(prev);
prev              261 drivers/net/wireless/intersil/hostap/hostap_main.c 	struct hostap_tx_callback_info *cb, *prev = NULL;
prev              266 drivers/net/wireless/intersil/hostap/hostap_main.c 		prev = cb;
prev              270 drivers/net/wireless/intersil/hostap/hostap_main.c 		if (prev == NULL)
prev              273 drivers/net/wireless/intersil/hostap/hostap_main.c 			prev->next = cb->next;
prev              112 drivers/net/wireless/intersil/p54/txrx.c 			target_skb = entry->prev;
prev             2137 drivers/net/wireless/intersil/prism54/isl_ioctl.c 			bss = list_entry(priv->bss_wpa_list.prev,
prev             2162 drivers/net/wireless/intersil/prism54/isl_ioctl.c 		bss = list_entry(priv->bss_wpa_list.prev,
prev              303 drivers/net/wireless/mediatek/mt76/dma.c 	skb->prev = skb->next = NULL;
prev              236 drivers/net/wireless/mediatek/mt76/tx.c 	if (!skb->prev) {
prev              782 drivers/net/wireless/mediatek/mt76/usb.c 	skb->prev = skb->next = NULL;
prev              568 drivers/net/wireless/realtek/rtw88/pci.c 	struct sk_buff *prev = skb_dequeue(&ring->queue);
prev              572 drivers/net/wireless/realtek/rtw88/pci.c 	if (!prev)
prev              575 drivers/net/wireless/realtek/rtw88/pci.c 	tx_data = rtw_pci_get_tx_data(prev);
prev              577 drivers/net/wireless/realtek/rtw88/pci.c 	pci_unmap_single(rtwpci->pdev, dma, prev->len,
prev              579 drivers/net/wireless/realtek/rtw88/pci.c 	dev_kfree_skb_any(prev);
prev             1599 drivers/net/wireless/st/cw1200/sta.c 	u32 bit, prev;
prev             1608 drivers/net/wireless/st/cw1200/sta.c 	prev = priv->sta_asleep_mask & bit;
prev             1612 drivers/net/wireless/st/cw1200/sta.c 		if (!prev) {
prev             1621 drivers/net/wireless/st/cw1200/sta.c 		if (prev) {
prev              321 drivers/net/wireless/st/cw1200/txrx.c 		entry = list_entry(cache->free.prev,
prev              564 drivers/nvdimm/namespace_devs.c 		struct nd_label_id *label_id, struct resource *prev,
prev             3650 drivers/nvme/host/core.c 	unsigned i, j, nsid, prev = 0;
prev             3659 drivers/nvme/host/core.c 		ret = nvme_identify_ns_list(ctrl, prev, ns_list);
prev             3670 drivers/nvme/host/core.c 			while (++prev < nsid) {
prev             3671 drivers/nvme/host/core.c 				ns = nvme_find_get_ns(ctrl, prev);
prev             3681 drivers/nvme/host/core.c 	nvme_remove_invalid_namespaces(ctrl, prev);
prev              188 drivers/nvme/target/io-cmd-bdev.c 			struct bio *prev = bio;
prev              195 drivers/nvme/target/io-cmd-bdev.c 			bio_chain(bio, prev);
prev              196 drivers/nvme/target/io-cmd-bdev.c 			submit_bio(prev);
prev              287 drivers/of/base.c struct device_node *__of_find_all_nodes(struct device_node *prev)
prev              290 drivers/of/base.c 	if (!prev) {
prev              292 drivers/of/base.c 	} else if (prev->child) {
prev              293 drivers/of/base.c 		np = prev->child;
prev              296 drivers/of/base.c 		np = prev;
prev              312 drivers/of/base.c struct device_node *of_find_all_nodes(struct device_node *prev)
prev              318 drivers/of/base.c 	np = __of_find_all_nodes(prev);
prev              320 drivers/of/base.c 	of_node_put(prev);
prev              729 drivers/of/base.c 						struct device_node *prev)
prev              736 drivers/of/base.c 	next = prev ? prev->sibling : node->child;
prev              740 drivers/of/base.c 	of_node_put(prev);
prev              757 drivers/of/base.c 	struct device_node *prev)
prev              763 drivers/of/base.c 	next = __of_get_next_child(node, prev);
prev              778 drivers/of/base.c 	struct device_node *prev)
prev              787 drivers/of/base.c 	next = prev ? prev->sibling : node->child;
prev              794 drivers/of/base.c 	of_node_put(prev);
prev              808 drivers/of/base.c struct device_node *of_get_next_cpu_node(struct device_node *prev)
prev              814 drivers/of/base.c 	if (!prev)
prev              818 drivers/of/base.c 	if (prev)
prev              819 drivers/of/base.c 		next = prev->sibling;
prev              831 drivers/of/base.c 	of_node_put(prev);
prev             2228 drivers/of/base.c 	struct device_node *prev = NULL, *np = of_cpu_device_node_get(cpu);
prev             2231 drivers/of/base.c 		prev = np;
prev             2236 drivers/of/base.c 	of_property_read_u32(prev, "cache-level", &cache_level);
prev               65 drivers/of/pdt.c static struct property * __init of_pdt_build_one_prop(phandle node, char *prev,
prev               90 drivers/of/pdt.c 		err = of_pdt_prom_ops->nextprop(node, prev, p->name);
prev              596 drivers/of/property.c 					struct device_node *prev)
prev              609 drivers/of/property.c 	if (!prev) {
prev              624 drivers/of/property.c 		port = of_get_parent(prev);
prev              626 drivers/of/property.c 			      __func__, prev))
prev              636 drivers/of/property.c 		endpoint = of_get_next_child(port, prev);
prev              643 drivers/of/property.c 		prev = NULL;
prev              935 drivers/of/property.c 				  struct fwnode_handle *prev)
prev              938 drivers/of/property.c 							   to_of_node(prev)));
prev              782 drivers/parport/share.c 	tmp->prev = NULL;
prev              807 drivers/parport/share.c 		port->physport->devices->prev = tmp;
prev              925 drivers/parport/share.c 	par_dev->prev = NULL;
prev              951 drivers/parport/share.c 		port->physport->devices->prev = par_dev;
prev             1021 drivers/parport/share.c 		dev->next->prev = dev->prev;
prev             1022 drivers/parport/share.c 	if (dev->prev)
prev             1023 drivers/parport/share.c 		dev->prev->next = dev->next;
prev              545 drivers/pci/hotplug/acpiphp_glue.c 	struct pci_dev *dev, *prev;
prev              554 drivers/pci/hotplug/acpiphp_glue.c 	list_for_each_entry_safe_reverse(dev, prev, &bus->devices, bus_list)
prev             1739 drivers/pci/hotplug/ibmphp_res.c static struct bus_node *find_bus_wprev(u8 bus_number, struct bus_node **prev, u8 flag)
prev             1745 drivers/pci/hotplug/ibmphp_res.c 			*prev = list_prev_entry(bus_cur, bus_list);
prev              571 drivers/pci/pci-driver.c 		pci_power_t prev = pci_dev->current_state;
prev              581 drivers/pci/pci-driver.c 			WARN_ONCE(pci_dev->current_state != prev,
prev              598 drivers/pci/pci-driver.c 		pci_power_t prev = pci_dev->current_state;
prev              608 drivers/pci/pci-driver.c 			WARN_ONCE(pci_dev->current_state != prev,
prev              786 drivers/pci/pci-driver.c 		pci_power_t prev = pci_dev->current_state;
prev              796 drivers/pci/pci-driver.c 			WARN_ONCE(pci_dev->current_state != prev,
prev              834 drivers/pci/pci-driver.c 		pci_power_t prev = pci_dev->current_state;
prev              844 drivers/pci/pci-driver.c 			WARN_ONCE(pci_dev->current_state != prev,
prev             1285 drivers/pci/pci-driver.c 	pci_power_t prev = pci_dev->current_state;
prev             1322 drivers/pci/pci-driver.c 		WARN_ONCE(pci_dev->current_state != prev,
prev              206 drivers/perf/arm_smmuv3_pmu.c 	u64 delta, prev, now;
prev              210 drivers/perf/arm_smmuv3_pmu.c 		prev = local64_read(&hwc->prev_count);
prev              212 drivers/perf/arm_smmuv3_pmu.c 	} while (local64_cmpxchg(&hwc->prev_count, prev, now) != prev);
prev              215 drivers/perf/arm_smmuv3_pmu.c 	delta = now - prev;
prev              342 drivers/perf/qcom_l2_pmu.c 	u64 delta, prev, now;
prev              346 drivers/perf/qcom_l2_pmu.c 		prev = local64_read(&hwc->prev_count);
prev              348 drivers/perf/qcom_l2_pmu.c 	} while (local64_cmpxchg(&hwc->prev_count, prev, now) != prev);
prev              354 drivers/perf/qcom_l2_pmu.c 	delta = now - prev;
prev              244 drivers/perf/qcom_l3_pmu.c 	u64 prev, new;
prev              247 drivers/perf/qcom_l3_pmu.c 		prev = local64_read(&event->hw.prev_count);
prev              253 drivers/perf/qcom_l3_pmu.c 	} while (local64_cmpxchg(&event->hw.prev_count, prev, new) != prev);
prev              255 drivers/perf/qcom_l3_pmu.c 	local64_add(new - prev, &event->count);
prev              320 drivers/perf/qcom_l3_pmu.c 	u32 prev, new;
prev              323 drivers/perf/qcom_l3_pmu.c 		prev = local64_read(&event->hw.prev_count);
prev              325 drivers/perf/qcom_l3_pmu.c 	} while (local64_cmpxchg(&event->hw.prev_count, prev, new) != prev);
prev              327 drivers/perf/qcom_l3_pmu.c 	local64_add(new - prev, &event->count);
prev              317 drivers/perf/thunderx2_pmu.c 	s64 prev, delta, new = 0;
prev              328 drivers/perf/thunderx2_pmu.c 	prev = local64_xchg(&hwc->prev_count, new);
prev              331 drivers/perf/thunderx2_pmu.c 	delta = (u32)(((1UL << 32) - prev) + new);
prev             1723 drivers/pinctrl/pinctrl-at91.c 	struct at91_gpio_chip   *prev = NULL;
prev             1775 drivers/pinctrl/pinctrl-at91.c 	prev = gpiochip_get_data(gpiochip_prev);
prev             1779 drivers/pinctrl/pinctrl-at91.c 		if (prev->next) {
prev             1780 drivers/pinctrl/pinctrl-at91.c 			prev = prev->next;
prev             1782 drivers/pinctrl/pinctrl-at91.c 			prev->next = at91_gpio;
prev              181 drivers/power/supply/ab8500_btemp.c 	static int prev;
prev              188 drivers/power/supply/ab8500_btemp.c 		return prev;
prev              190 drivers/power/supply/ab8500_btemp.c 	prev = vbtemp;
prev              459 drivers/power/supply/ab8500_btemp.c 	static int prev;
prev              488 drivers/power/supply/ab8500_btemp.c 			return prev;
prev              499 drivers/power/supply/ab8500_btemp.c 		prev = temp;
prev              833 drivers/power/supply/ab8500_fg.c 	static int prev;
prev              840 drivers/power/supply/ab8500_fg.c 		return prev;
prev              843 drivers/power/supply/ab8500_fg.c 	prev = vbat;
prev             1081 drivers/power/supply/bq27xxx_battery.c 	u16 *prev = bq27xxx_dm_reg_ptr(buf, reg);
prev             1083 drivers/power/supply/bq27xxx_battery.c 	if (prev == NULL) {
prev             1096 drivers/power/supply/bq27xxx_battery.c 	if (be16_to_cpup(prev) == val) {
prev             1113 drivers/power/supply/bq27xxx_battery.c 			 "\n", str, be16_to_cpup(prev), val);
prev             1119 drivers/power/supply/bq27xxx_battery.c 	*prev = cpu_to_be16(val);
prev              525 drivers/rapidio/rio-scan.c 			 u8 hopcount, struct rio_dev *prev, int prev_port)
prev              549 drivers/rapidio/rio-scan.c 			if (rdev && prev && rio_is_switch(prev)) {
prev              552 drivers/rapidio/rio-scan.c 				prev->rswitch->nextdev[prev_port] = rdev;
prev              584 drivers/rapidio/rio-scan.c 		rdev->prev = prev;
prev              585 drivers/rapidio/rio-scan.c 		if (prev && rio_is_switch(prev))
prev              586 drivers/rapidio/rio-scan.c 			prev->rswitch->nextdev[prev_port] = rdev;
prev              729 drivers/rapidio/rio-scan.c 	      u8 hopcount, struct rio_dev *prev, int prev_port)
prev              737 drivers/rapidio/rio-scan.c 		rdev->prev = prev;
prev              738 drivers/rapidio/rio-scan.c 		if (prev && rio_is_switch(prev))
prev              739 drivers/rapidio/rio-scan.c 			prev->rswitch->nextdev[prev_port] = rdev;
prev              958 drivers/rapidio/rio-scan.c 	if (mport->nnode.next || mport->nnode.prev)
prev               62 drivers/rapidio/rio-sysfs.c 			(rdev->prev) ? rio_name(rdev->prev) : "root");
prev              958 drivers/rapidio/rio.c 	struct rio_dev *prev = NULL;
prev              961 drivers/rapidio/rio.c 	while (rdev->prev && (rdev->prev->pef & RIO_PEF_SWITCH)) {
prev              962 drivers/rapidio/rio.c 		if (!rio_read_config_32(rdev->prev, RIO_DEV_ID_CAR, &result)) {
prev              963 drivers/rapidio/rio.c 			prev = rdev->prev;
prev              966 drivers/rapidio/rio.c 		rdev = rdev->prev;
prev              969 drivers/rapidio/rio.c 	if (!prev)
prev              972 drivers/rapidio/rio.c 	p_port = prev->rswitch->route_table[rdev->destid];
prev              976 drivers/rapidio/rio.c 			 rio_name(prev), p_port);
prev              977 drivers/rapidio/rio.c 		*nrdev = prev;
prev              286 drivers/ras/cec.c 	u64 prev = 0;
prev              292 drivers/ras/cec.c 		if (WARN(prev > this, "prev: 0x%016llx <-> this: 0x%016llx\n", prev, this))
prev              295 drivers/ras/cec.c 		prev = this;
prev              734 drivers/s390/char/raw3270.c 			__list_add(&rp->list, l->prev, l);
prev              379 drivers/s390/char/sclp.c 	    req->list.prev == &sclp_req_queue) {
prev              428 drivers/s390/char/sclp_cmd.c 	struct list_head *prev;
prev              437 drivers/s390/char/sclp_cmd.c 	prev = &sclp_mem_list;
prev              444 drivers/s390/char/sclp_cmd.c 		prev = &incr->list;
prev              452 drivers/s390/char/sclp_cmd.c 	list_add(&new_incr->list, prev);
prev              940 drivers/s390/char/tape_34xx.c 			tape_34xx_append_new_sbid(bid, l->prev);
prev              946 drivers/s390/char/tape_34xx.c 		tape_34xx_append_new_sbid(bid, l->prev);
prev              484 drivers/s390/char/tty3270.c 		if (tp->rcl_walk && tp->rcl_walk->prev != &tp->rcl_lines)
prev              485 drivers/s390/char/tty3270.c 			tp->rcl_walk = tp->rcl_walk->prev;
prev              487 drivers/s390/char/tty3270.c 			tp->rcl_walk = tp->rcl_lines.prev;
prev              189 drivers/s390/cio/qdio_setup.c 	struct qdio_q *prev;
prev              201 drivers/s390/cio/qdio_setup.c 		prev = (q->is_input_q) ? irq_ptr->input_qs[i - 1]
prev              203 drivers/s390/cio/qdio_setup.c 		prev->slib->nsliba = (unsigned long)q->slib;
prev              644 drivers/s390/net/lcs.c 	int prev, next;
prev              647 drivers/s390/net/lcs.c 	prev = (index - 1) & (LCS_NUM_BUFFS - 1);
prev              652 drivers/s390/net/lcs.c 		if (!(channel->ccws[prev].flags & CCW_FLAG_SUSPEND))
prev              689 drivers/s390/net/lcs.c 	int index, prev, next;
prev              695 drivers/s390/net/lcs.c 	prev = (index - 1) & (LCS_NUM_BUFFS - 1);
prev              701 drivers/s390/net/lcs.c 	if (channel->iob[prev].state == LCS_BUF_STATE_READY) {
prev              708 drivers/s390/net/lcs.c 		__lcs_ready_buffer_bits(channel, prev);
prev             4001 drivers/scsi/aacraid/aachba.c 	psg->sg[0].prev = 0;
prev             4015 drivers/scsi/aacraid/aachba.c 		psg->sg[i].prev = 0;
prev              477 drivers/scsi/aacraid/aacraid.h 	__le32		prev;	/* reserved for F/W use */
prev              485 drivers/scsi/aacraid/aacraid.h 	u32		prev;	/* reserved for F/W use */
prev              663 drivers/scsi/aha152x.c 	struct scsi_cmnd *ptr, *prev;
prev              665 drivers/scsi/aha152x.c 	for (ptr = *SC, prev = NULL;
prev              667 drivers/scsi/aha152x.c 	     prev = ptr, ptr = SCNEXT(ptr))
prev              671 drivers/scsi/aha152x.c 		if (prev)
prev              672 drivers/scsi/aha152x.c 			SCNEXT(prev) = SCNEXT(ptr);
prev              685 drivers/scsi/aha152x.c 	struct scsi_cmnd *ptr, *prev;
prev              687 drivers/scsi/aha152x.c 	for (ptr = *SC, prev = NULL;
prev              689 drivers/scsi/aha152x.c 	     prev = ptr, ptr = SCNEXT(ptr))
prev              693 drivers/scsi/aha152x.c 		if (prev)
prev              694 drivers/scsi/aha152x.c 			SCNEXT(prev) = SCNEXT(ptr);
prev              200 drivers/scsi/aic7xxx/aic79xx_core.c 				     u_int prev, u_int next, u_int tid);
prev             8383 drivers/scsi/aic7xxx/aic79xx_core.c 	u_int	prev;
prev             8388 drivers/scsi/aic7xxx/aic79xx_core.c 	prev = SCB_LIST_NULL;
prev             8410 drivers/scsi/aic7xxx/aic79xx_core.c 			prev = scbid;
prev             8421 drivers/scsi/aic7xxx/aic79xx_core.c 			ahd_rem_wscb(ahd, scbid, prev, next, tid);
prev             8422 drivers/scsi/aic7xxx/aic79xx_core.c 			*list_tail = prev;
prev             8423 drivers/scsi/aic7xxx/aic79xx_core.c 			if (SCBID_IS_NULL(prev))
prev             8430 drivers/scsi/aic7xxx/aic79xx_core.c 			prev = scbid;
prev             8482 drivers/scsi/aic7xxx/aic79xx_core.c 	     u_int prev, u_int next, u_int tid)
prev             8487 drivers/scsi/aic7xxx/aic79xx_core.c 	if (!SCBID_IS_NULL(prev)) {
prev             8488 drivers/scsi/aic7xxx/aic79xx_core.c 		ahd_set_scbptr(ahd, prev);
prev             8502 drivers/scsi/aic7xxx/aic79xx_core.c 		ahd_outw(ahd, tail_offset, prev);
prev              207 drivers/scsi/aic7xxx/aic7xxx_core.c 						   u_int prev, u_int scbptr);
prev              210 drivers/scsi/aic7xxx/aic7xxx_core.c 				     u_int scbpos, u_int prev);
prev             5839 drivers/scsi/aic7xxx/aic7xxx_core.c 	uint8_t prev;
prev             5969 drivers/scsi/aic7xxx/aic7xxx_core.c 	prev = SCB_LIST_NULL;
prev             6014 drivers/scsi/aic7xxx/aic7xxx_core.c 				next = ahc_rem_wscb(ahc, next, prev);
prev             6017 drivers/scsi/aic7xxx/aic7xxx_core.c 				prev = next;
prev             6023 drivers/scsi/aic7xxx/aic7xxx_core.c 			prev = next;
prev             6142 drivers/scsi/aic7xxx/aic7xxx_core.c 	u_int	prev;
prev             6148 drivers/scsi/aic7xxx/aic7xxx_core.c 	prev = SCB_LIST_NULL;
prev             6170 drivers/scsi/aic7xxx/aic7xxx_core.c 		if (next == prev) {
prev             6173 drivers/scsi/aic7xxx/aic7xxx_core.c 			      next, prev);
prev             6181 drivers/scsi/aic7xxx/aic7xxx_core.c 				    ahc_rem_scb_from_disc_list(ahc, prev, next);
prev             6183 drivers/scsi/aic7xxx/aic7xxx_core.c 				prev = next;
prev             6189 drivers/scsi/aic7xxx/aic7xxx_core.c 			prev = next;
prev             6203 drivers/scsi/aic7xxx/aic7xxx_core.c ahc_rem_scb_from_disc_list(struct ahc_softc *ahc, u_int prev, u_int scbptr)
prev             6214 drivers/scsi/aic7xxx/aic7xxx_core.c 	if (prev != SCB_LIST_NULL) {
prev             6215 drivers/scsi/aic7xxx/aic7xxx_core.c 		ahc_outb(ahc, SCBPTR, prev);
prev             6248 drivers/scsi/aic7xxx/aic7xxx_core.c ahc_rem_wscb(struct ahc_softc *ahc, u_int scbpos, u_int prev)
prev             6266 drivers/scsi/aic7xxx/aic7xxx_core.c 	if (prev == SCB_LIST_NULL) {
prev             6280 drivers/scsi/aic7xxx/aic7xxx_core.c 		ahc_outb(ahc, SCBPTR, prev);
prev             1099 drivers/scsi/aic94xx/aic94xx_hwi.c 			struct asd_ascb *last = list_entry(first->list.prev,
prev             1133 drivers/scsi/aic94xx/aic94xx_hwi.c 	struct asd_ascb *last = list_entry(ascb->list.prev,
prev             1209 drivers/scsi/aic94xx/aic94xx_hwi.c 	__list_add(&list, ascb->list.prev, &ascb->list);
prev             1214 drivers/scsi/aic94xx/aic94xx_hwi.c 	list_splice_init(&list, asd_ha->seq.pend_q.prev);
prev              348 drivers/scsi/aic94xx/aic94xx_hwi.h 	__list_add(&list, ascb_list->list.prev, &ascb_list->list);
prev              548 drivers/scsi/aic94xx/aic94xx_task.c 	__list_add(&alist, ascb->list.prev, &ascb->list);
prev              588 drivers/scsi/aic94xx/aic94xx_task.c 		__list_add(&alist, ascb->list.prev, &ascb->list);
prev              497 drivers/scsi/arm/acornscsi.c 	unsigned long prev;
prev              507 drivers/scsi/arm/acornscsi.c 	prev = host->status[target][ptr].when;
prev              526 drivers/scsi/arm/acornscsi.c 			time_diff = host->status[target][ptr].when - prev;
prev              527 drivers/scsi/arm/acornscsi.c 			prev = host->status[target][ptr].when;
prev             2832 drivers/scsi/arm/acornscsi.c 	unsigned int statptr, prev;
prev             2840 drivers/scsi/arm/acornscsi.c 	prev = host->status[devidx][statptr].when;
prev             2848 drivers/scsi/arm/acornscsi.c 			(host->status[devidx][statptr].when - prev) < 100 ?
prev             2849 drivers/scsi/arm/acornscsi.c 				(host->status[devidx][statptr].when - prev) : 99);
prev             2850 drivers/scsi/arm/acornscsi.c 		prev = host->status[devidx][statptr].when;
prev              118 drivers/scsi/bfa/bfa_cs.h #define bfa_q_prev(_qe) (((struct list_head *) (_qe))->prev)
prev               69 drivers/scsi/csiostor/csio_defs.h 	return ((list->next == list) && (list->prev == list));
prev               73 drivers/scsi/csiostor/csio_defs.h #define csio_list_prev(elem)	(((struct list_head *)(elem))->prev)
prev               69 drivers/scsi/dpt/dpti_i2o.h 	struct i2o_device *prev;
prev             1582 drivers/scsi/dpt_i2o.c 	d->prev=NULL;
prev             1584 drivers/scsi/dpt_i2o.c 		pHba->devices->prev=d;
prev              767 drivers/scsi/initio.c 	struct scsi_ctrl_blk *tmp, *prev;
prev              773 drivers/scsi/initio.c 	prev = tmp = host->first_pending;
prev              780 drivers/scsi/initio.c 				prev->next = tmp->next;
prev              782 drivers/scsi/initio.c 					host->last_pending = prev;
prev              787 drivers/scsi/initio.c 		prev = tmp;
prev              837 drivers/scsi/initio.c 	struct scsi_ctrl_blk *tmp, *prev;
prev              843 drivers/scsi/initio.c 	prev = tmp = host->first_busy;
prev              850 drivers/scsi/initio.c 				prev->next = tmp->next;
prev              852 drivers/scsi/initio.c 					host->last_busy = prev;
prev              861 drivers/scsi/initio.c 		prev = tmp;
prev              869 drivers/scsi/initio.c 	struct scsi_ctrl_blk *tmp, *prev;
prev              873 drivers/scsi/initio.c 	prev = tmp = host->first_busy;
prev              879 drivers/scsi/initio.c 		prev = tmp;
prev              923 drivers/scsi/initio.c 	struct scsi_ctrl_blk *tmp, *prev;
prev              939 drivers/scsi/initio.c 	prev = tmp = host->first_pending;	/* Check Pend queue */
prev              950 drivers/scsi/initio.c 				prev->next = tmp->next;
prev              952 drivers/scsi/initio.c 					host->last_pending = prev;
prev              961 drivers/scsi/initio.c 		prev = tmp;
prev              965 drivers/scsi/initio.c 	prev = tmp = host->first_busy;	/* Check Busy queue */
prev              980 drivers/scsi/initio.c 					prev->next = tmp->next;
prev              982 drivers/scsi/initio.c 						host->last_busy = prev;
prev              995 drivers/scsi/initio.c 		prev = tmp;
prev             2358 drivers/scsi/initio.c 	struct scsi_ctrl_blk *tmp, *prev;
prev             2373 drivers/scsi/initio.c 	prev = tmp = host->first_busy;	/* Check Busy queue */
prev             2381 drivers/scsi/initio.c 				prev->next = tmp->next;
prev             2383 drivers/scsi/initio.c 					host->last_busy = prev;
prev             2390 drivers/scsi/initio.c 			prev = tmp;
prev             2815 drivers/scsi/initio.c 	struct scsi_ctrl_blk *scb, *tmp, *prev = NULL /* silence gcc */;
prev             2878 drivers/scsi/initio.c 			prev->next = tmp;
prev             2879 drivers/scsi/initio.c 		prev = tmp;
prev             2881 drivers/scsi/initio.c 	prev->next = NULL;
prev             2884 drivers/scsi/initio.c 	host->last_avail = prev;
prev             2570 drivers/scsi/libfc/fc_exch.c 			return list_entry(lport->ema_list.prev,
prev              985 drivers/scsi/lpfc/lpfc_bsg.c 			iocbq = list_entry(head.prev, typeof(*iocbq), list);
prev             1130 drivers/scsi/lpfc/lpfc_bsg.c 		list_move(evt->events_to_see.prev, &evt->events_to_get);
prev             1323 drivers/scsi/lpfc/lpfc_bsg.c 			evt_dat = list_entry(evt->events_to_get.prev,
prev             2757 drivers/scsi/lpfc/lpfc_bsg.c 		list_move(evt->events_to_see.prev, &evt->events_to_get);
prev             2759 drivers/scsi/lpfc/lpfc_bsg.c 		*rxxri = (list_entry(evt->events_to_get.prev,
prev             3337 drivers/scsi/lpfc/lpfc_bsg.c 		list_move(evt->events_to_see.prev, &evt->events_to_get);
prev             3338 drivers/scsi/lpfc/lpfc_bsg.c 		evdat = list_entry(evt->events_to_get.prev,
prev             2183 drivers/scsi/lpfc/lpfc_hbadisc.c 			if (fcf_pri->list.prev == &phba->fcf.fcf_pri_list)
prev             2189 drivers/scsi/lpfc/lpfc_hbadisc.c 					fcf_pri->list.prev)->list);
prev             10964 drivers/scsi/lpfc/lpfc_sli.c 			slp->next, slp->prev, pring->postbufq_cnt);
prev             11008 drivers/scsi/lpfc/lpfc_sli.c 			slp->next, slp->prev, pring->postbufq_cnt);
prev             17338 drivers/scsi/lpfc/lpfc_sli.c 	d_buf = list_entry(seq_dmabuf->dbuf.list.prev, typeof(*d_buf), list);
prev             17355 drivers/scsi/lpfc/lpfc_sli.c 		d_buf = list_entry(d_buf->list.prev, typeof(*d_buf), list);
prev              560 drivers/scsi/mesh.c 	struct scsi_cmnd *cmd, *prev, *next;
prev              569 drivers/scsi/mesh.c 		prev = NULL;
prev              575 drivers/scsi/mesh.c 			prev = cmd;
prev              578 drivers/scsi/mesh.c 		if (prev == NULL)
prev              581 drivers/scsi/mesh.c 			prev->host_scribble = (void *) next;
prev              583 drivers/scsi/mesh.c 			ms->request_qtail = prev;
prev              861 drivers/scsi/mesh.c 	int b, t, prev;
prev              899 drivers/scsi/mesh.c 	prev = ms->conn_tgt;
prev              964 drivers/scsi/mesh.c 	dlog(ms, "resel prev tgt=%d", prev);
prev             6551 drivers/scsi/ncr53c8xx.c 		qp = lp->busy_ccbq.prev;
prev             6554 drivers/scsi/ncr53c8xx.c 			qp  = qp->prev;
prev              565 drivers/scsi/scsi.c 					   struct scsi_device *prev)
prev              567 drivers/scsi/scsi.c 	struct list_head *list = (prev ? &prev->siblings : &shost->__devices);
prev              582 drivers/scsi/scsi.c 	if (prev)
prev              583 drivers/scsi/scsi.c 		scsi_device_put(prev);
prev             5528 drivers/scsi/scsi_debug.c 		sdbg_host = list_entry(sdebug_host_list.prev,
prev              639 drivers/scsi/ses.c 		struct enclosure_device *prev = NULL;
prev              641 drivers/scsi/ses.c 		while ((edev = enclosure_find(&sdev->host->shost_gendev, prev)) != NULL) {
prev              643 drivers/scsi/ses.c 			prev = edev;
prev              787 drivers/scsi/ses.c 	struct enclosure_device *edev, *prev = NULL;
prev              789 drivers/scsi/ses.c 	while ((edev = enclosure_find(&sdev->host->shost_gendev, prev)) != NULL) {
prev              790 drivers/scsi/ses.c 		prev = edev;
prev              480 drivers/scsi/wd33c93.c 	struct scsi_cmnd *cmd, *prev;
prev              494 drivers/scsi/wd33c93.c 	prev = NULL;
prev              499 drivers/scsi/wd33c93.c 		prev = cmd;
prev              512 drivers/scsi/wd33c93.c 	if (prev)
prev              513 drivers/scsi/wd33c93.c 		prev->host_scribble = cmd->host_scribble;
prev              564 drivers/scsi/wd33c93.c 	for (prev = (struct scsi_cmnd *) hostdata->input_Q; prev;
prev              565 drivers/scsi/wd33c93.c 	     prev = (struct scsi_cmnd *) prev->host_scribble) {
prev              566 drivers/scsi/wd33c93.c 		if ((prev->device->id != cmd->device->id) ||
prev              567 drivers/scsi/wd33c93.c 		    (prev->device->lun != cmd->device->lun)) {
prev              568 drivers/scsi/wd33c93.c 			for (prev = (struct scsi_cmnd *) hostdata->input_Q; prev;
prev              569 drivers/scsi/wd33c93.c 			     prev = (struct scsi_cmnd *) prev->host_scribble)
prev              570 drivers/scsi/wd33c93.c 				prev->SCp.phase = 1;
prev             1608 drivers/scsi/wd33c93.c 	struct scsi_cmnd *tmp, *prev;
prev             1622 drivers/scsi/wd33c93.c 	prev = NULL;
prev             1625 drivers/scsi/wd33c93.c 			if (prev)
prev             1626 drivers/scsi/wd33c93.c 				prev->host_scribble = cmd->host_scribble;
prev             1639 drivers/scsi/wd33c93.c 		prev = tmp;
prev              421 drivers/sh/clk/core.c 	if (clk->node.next || clk->node.prev)
prev              749 drivers/siox/siox-core.c 		sdevice = container_of(smaster->devices.prev,
prev              858 drivers/siox/siox-core.c 	sdevice = container_of(smaster->devices.prev, struct siox_device, node);
prev              207 drivers/soc/fsl/qbman/qman_test_stash.c static inline u32 do_lfsr(u32 prev)
prev              209 drivers/soc/fsl/qbman/qman_test_stash.c 	return (prev >> 1) ^ (-(prev & 1u) & 0xd0000001u);
prev              385 drivers/spi/spi-mxs.c 		flag = (&t->transfer_list == m->transfers.prev) ^ t->cs_change ?
prev             1401 drivers/spi/spi-pl022.c 		previous = list_entry(transfer->transfer_list.prev,
prev             1521 drivers/spi/spi-pl022.c 			    list_entry(transfer->transfer_list.prev,
prev             2832 drivers/spi/spi.c 	rxfer->replaced_after = xfer_first->transfer_list.prev;
prev              366 drivers/staging/android/vsoc.c 	int prev = 0;
prev              372 drivers/staging/android/vsoc.c 	prev = atomic_xchg(owner_ptr, VSOC_REGION_FREE);
prev              373 drivers/staging/android/vsoc.c 	if (prev != perm->owned_value)
prev              378 drivers/staging/android/vsoc.c 			perm->owned_value, prev);
prev             1148 drivers/staging/comedi/drivers.c 	struct comedi_driver *prev;
prev             1156 drivers/staging/comedi/drivers.c 		for (prev = comedi_drivers; prev->next; prev = prev->next) {
prev             1157 drivers/staging/comedi/drivers.c 			if (prev->next == driver) {
prev             1158 drivers/staging/comedi/drivers.c 				prev->next = driver->next;
prev              530 drivers/staging/exfat/exfat.h 	struct buf_cache_t *prev;
prev               43 drivers/staging/exfat/exfat_cache.c 	bp->prev = list;
prev               44 drivers/staging/exfat/exfat_cache.c 	list->next->prev = bp;
prev               50 drivers/staging/exfat/exfat_cache.c 	bp->prev = list->prev;
prev               52 drivers/staging/exfat/exfat_cache.c 	list->prev->next = bp;
prev               53 drivers/staging/exfat/exfat_cache.c 	list->prev = bp;
prev               58 drivers/staging/exfat/exfat_cache.c 	bp->prev->next = bp->next;
prev               59 drivers/staging/exfat/exfat_cache.c 	bp->next->prev = bp->prev;
prev               65 drivers/staging/exfat/exfat_cache.c 	bp->prev->next = bp->next;
prev               66 drivers/staging/exfat/exfat_cache.c 	bp->next->prev = bp->prev;
prev               75 drivers/staging/exfat/exfat_cache.c 	bp = p_fs->FAT_cache_lru_list.prev;
prev              139 drivers/staging/exfat/exfat_cache.c 	p_fs->FAT_cache_lru_list.prev = &p_fs->FAT_cache_lru_list;
prev              146 drivers/staging/exfat/exfat_cache.c 		p_fs->FAT_cache_array[i].prev = NULL;
prev              153 drivers/staging/exfat/exfat_cache.c 	p_fs->buf_cache_lru_list.prev = &p_fs->buf_cache_lru_list;
prev              160 drivers/staging/exfat/exfat_cache.c 		p_fs->buf_cache_array[i].prev = NULL;
prev              556 drivers/staging/exfat/exfat_cache.c 	bp = p_fs->buf_cache_lru_list.prev;
prev              558 drivers/staging/exfat/exfat_cache.c 		bp = bp->prev;
prev              307 drivers/staging/exfat/exfat_core.c 	u32 clu, prev;
prev              330 drivers/staging/exfat/exfat_core.c 		prev = clu;
prev              334 drivers/staging/exfat/exfat_core.c 		if (FAT_write(sb, prev, CLUSTER_32(0)) < 0)
prev             1188 drivers/staging/fwserial/fwserial.c 			   struct async_icount *prev)
prev             1195 drivers/staging/fwserial/fwserial.c 	delta = ((mask & TIOCM_RNG && prev->rng != now.rng) ||
prev             1196 drivers/staging/fwserial/fwserial.c 		 (mask & TIOCM_DSR && prev->dsr != now.dsr) ||
prev             1197 drivers/staging/fwserial/fwserial.c 		 (mask & TIOCM_CAR && prev->dcd != now.dcd) ||
prev             1198 drivers/staging/fwserial/fwserial.c 		 (mask & TIOCM_CTS && prev->cts != now.cts));
prev             1200 drivers/staging/fwserial/fwserial.c 	*prev = now;
prev             1207 drivers/staging/fwserial/fwserial.c 	struct async_icount prev;
prev             1209 drivers/staging/fwserial/fwserial.c 	prev = port->icount;
prev             1212 drivers/staging/fwserial/fwserial.c 					check_msr_delta(port, mask, &prev));
prev              123 drivers/staging/gdm724x/gdm_mux.c 	r = list_entry(rx->rx_free_list.prev, struct mux_rx, free_list);
prev             1753 drivers/staging/isdn/gigaset/bas-gigaset.c 		cs->cmdbuf->prev = NULL;
prev             2014 drivers/staging/isdn/gigaset/bas-gigaset.c 	cb->prev = cs->lastcmdbuf;
prev              519 drivers/staging/isdn/gigaset/gigaset.h 	struct cmdbuf_t *next, *prev;
prev              135 drivers/staging/isdn/gigaset/ser-gigaset.c 			cb->prev = NULL;
prev              248 drivers/staging/isdn/gigaset/ser-gigaset.c 	cb->prev = cs->lastcmdbuf;
prev              439 drivers/staging/isdn/gigaset/usb-gigaset.c 				cs->cmdbuf->prev = NULL;
prev              499 drivers/staging/isdn/gigaset/usb-gigaset.c 	cb->prev = cs->lastcmdbuf;
prev             1703 drivers/staging/ks7010/ks_wlan_net.c 			pmk = list_entry(priv->pmklist.head.prev, struct pmk,
prev              193 drivers/staging/media/imx/imx-media-vdic.c 	struct imx_media_buffer *prev;
prev              201 drivers/staging/media/imx/imx-media-vdic.c 	prev = priv->prev_in_buf ? priv->prev_in_buf : curr;
prev              203 drivers/staging/media/imx/imx-media-vdic.c 	prev_vb = &prev->vbuf.vb2_buf;
prev               36 drivers/staging/rtl8192e/rtl819x_TSProc.c 					list_entry(pRxTs->RxPendingPktList.prev,
prev              413 drivers/staging/rtl8192e/rtl819x_TSProc.c 					list_entry(pRxTS->RxPendingPktList.prev,
prev              467 drivers/staging/rtl8192e/rtllib_rx.c 	pReorderEntry->List.next->prev = &pReorderEntry->List;
prev              468 drivers/staging/rtl8192e/rtllib_rx.c 	pReorderEntry->List.prev = pList;
prev              549 drivers/staging/rtl8192e/rtllib_rx.c 				  list_entry(pTS->RxPendingPktList.prev,
prev              701 drivers/staging/rtl8192e/rtllib_rx.c 					list_entry(pTS->RxPendingPktList.prev,
prev              512 drivers/staging/rtl8192e/rtllib_softmac_wx.c 	short prev = ieee->raw_tx;
prev              525 drivers/staging/rtl8192e/rtllib_softmac_wx.c 		if (prev == 0 && ieee->raw_tx) {
prev              532 drivers/staging/rtl8192e/rtllib_softmac_wx.c 		if (prev && ieee->raw_tx == 1)
prev              517 drivers/staging/rtl8192u/ieee80211/ieee80211_rx.c 	pReorderEntry->List.next->prev = &pReorderEntry->List;
prev              518 drivers/staging/rtl8192u/ieee80211/ieee80211_rx.c 	pReorderEntry->List.prev = pList;
prev              696 drivers/staging/rtl8192u/ieee80211/ieee80211_rx.c 		pReorderEntry = list_entry(pTS->rx_pending_pkt_list.prev, struct rx_reorder_entry, List);
prev              453 drivers/staging/rtl8192u/ieee80211/ieee80211_softmac_wx.c 	short prev = ieee->raw_tx;
prev              466 drivers/staging/rtl8192u/ieee80211/ieee80211_softmac_wx.c 		if (prev == 0 && ieee->raw_tx) {
prev              473 drivers/staging/rtl8192u/ieee80211/ieee80211_softmac_wx.c 		if (prev && ieee->raw_tx == 1)
prev               43 drivers/staging/rtl8192u/ieee80211/rtl819x_TSProc.c 			pReorderEntry = list_entry(pRxTs->rx_pending_pkt_list.prev, struct rx_reorder_entry, List);
prev              426 drivers/staging/rtl8192u/ieee80211/rtl819x_TSProc.c 			pRxReorderEntry = list_entry(pRxTS->rx_pending_pkt_list.prev, struct rx_reorder_entry, List);
prev              868 drivers/staging/speakup/main.c static int say_sentence_num(int num, int prev)
prev              872 drivers/staging/speakup/main.c 	if (prev && --bn == -1)
prev             1195 drivers/staging/unisys/visornic/visornic_main.c 	struct sk_buff *skb, *prev, *curr;
prev             1293 drivers/staging/unisys/visornic/visornic_main.c 		for (cc = 1, prev = NULL;
prev             1298 drivers/staging/unisys/visornic/visornic_main.c 			if (!prev)
prev             1301 drivers/staging/unisys/visornic/visornic_main.c 				prev->next = curr;
prev             1302 drivers/staging/unisys/visornic/visornic_main.c 			prev = curr;
prev               18 drivers/staging/wusbcore/host/whci/asl.c 			       struct whc_qset **next, struct whc_qset **prev)
prev               27 drivers/staging/wusbcore/host/whci/asl.c 	p = qset->list_node.prev;
prev               29 drivers/staging/wusbcore/host/whci/asl.c 		p = p->prev;
prev               32 drivers/staging/wusbcore/host/whci/asl.c 	*prev = container_of(p, struct whc_qset, list_node);
prev               44 drivers/staging/wusbcore/host/whci/asl.c 	struct whc_qset *next, *prev;
prev               49 drivers/staging/wusbcore/host/whci/asl.c 	qset_get_next_prev(whc, qset, &next, &prev);
prev               51 drivers/staging/wusbcore/host/whci/asl.c 	whc_qset_set_link_ptr(&prev->qh.link, qset->qset_dma);
prev               57 drivers/staging/wusbcore/host/whci/asl.c 	struct whc_qset *prev, *next;
prev               59 drivers/staging/wusbcore/host/whci/asl.c 	qset_get_next_prev(whc, qset, &next, &prev);
prev               72 drivers/staging/wusbcore/host/whci/asl.c 	whc_qset_set_link_ptr(&prev->qh.link, next->qset_dma);
prev               32 drivers/staging/wusbcore/host/whci/debug.c 	if (&qset->list_node == qset->whc->async_list.prev) {
prev             1747 drivers/staging/wusbcore/wa-xfer.c 			wa->xfer_delayed_list.prev);
prev             1779 drivers/staging/wusbcore/wa-xfer.c 			wa->xfer_errored_list.prev);
prev              119 drivers/target/iscsi/cxgbit/cxgbit_cm.c 	struct np_info *p, **prev = &cdev->np_hash_tab[bucket];
prev              122 drivers/target/iscsi/cxgbit/cxgbit_cm.c 	for (p = *prev; p; prev = &p->next, p = p->next) {
prev              125 drivers/target/iscsi/cxgbit/cxgbit_cm.c 			*prev = p->next;
prev              787 drivers/target/iscsi/iscsi_target_erl1.c 		ooo_tail = list_entry(sess->sess_ooo_cmdsn_list.prev,
prev              807 drivers/target/iscsi/iscsi_target_erl1.c 					ooo_tmp->ooo_list.prev);
prev             1028 drivers/target/iscsi/iscsi_target_erl1.c 		ooo_tail = list_entry(sess->sess_ooo_cmdsn_list.prev,
prev              532 drivers/target/target_core_transport.c 					se_nacl->acl_sess_list.prev,
prev             2860 drivers/target/target_core_transport.c 	char *prev = *str;
prev             2864 drivers/target/target_core_transport.c 	kfree(prev);
prev              671 drivers/thunderbolt/property.c 					 struct tb_property *prev)
prev              673 drivers/thunderbolt/property.c 	if (prev) {
prev              674 drivers/thunderbolt/property.c 		if (list_is_last(&prev->list, &dir->properties))
prev              676 drivers/thunderbolt/property.c 		return list_next_entry(prev, list);
prev              766 drivers/thunderbolt/switch.c 				     struct tb_port *prev)
prev              770 drivers/thunderbolt/switch.c 	if (!prev)
prev              773 drivers/thunderbolt/switch.c 	if (prev->sw == end->sw) {
prev              774 drivers/thunderbolt/switch.c 		if (prev == end)
prev              780 drivers/thunderbolt/switch.c 		if (prev->remote &&
prev              781 drivers/thunderbolt/switch.c 		    prev->remote->sw->config.depth > prev->sw->config.depth)
prev              782 drivers/thunderbolt/switch.c 			next = prev->remote;
prev              784 drivers/thunderbolt/switch.c 			next = tb_port_at(tb_route(end->sw), prev->sw);
prev              786 drivers/thunderbolt/switch.c 		if (tb_is_upstream_port(prev)) {
prev              787 drivers/thunderbolt/switch.c 			next = prev->remote;
prev              789 drivers/thunderbolt/switch.c 			next = tb_upstream_port(prev->sw);
prev              795 drivers/thunderbolt/switch.c 			    next->link_nr != prev->link_nr) {
prev              603 drivers/thunderbolt/tb.h 				     struct tb_port *prev);
prev             2313 drivers/tty/synclink_gt.c 	struct cond_wait *w, *prev;
prev             2316 drivers/tty/synclink_gt.c 	for (w = info->gpio_wait_q, prev = NULL ; w != NULL ; w = w->next) {
prev             2320 drivers/tty/synclink_gt.c 			if (prev != NULL)
prev             2321 drivers/tty/synclink_gt.c 				prev->next = w->next;
prev             2325 drivers/tty/synclink_gt.c 			prev = w;
prev             3007 drivers/tty/synclink_gt.c 	struct cond_wait *w, *prev;
prev             3010 drivers/tty/synclink_gt.c 	for (w = *head, prev = NULL ; w != NULL ; prev = w, w = w->next) {
prev             3012 drivers/tty/synclink_gt.c 			if (prev != NULL)
prev             3013 drivers/tty/synclink_gt.c 				prev->next = w->next;
prev             1408 drivers/tty/vt/vt_ioctl.c 	int prev;
prev             1416 drivers/tty/vt/vt_ioctl.c 	prev = fg_console;
prev             1439 drivers/tty/vt/vt_ioctl.c 	return prev;
prev              271 drivers/usb/c67x00/c67x00-sched.c 		struct c67x00_ep_data *prev;
prev              273 drivers/usb/c67x00/c67x00-sched.c 		list_for_each_entry(prev, &c67x00->list[type], node) {
prev              274 drivers/usb/c67x00/c67x00-sched.c 			if (prev->hep->desc.bEndpointAddress >
prev              276 drivers/usb/c67x00/c67x00-sched.c 				list_add(&ep_data->node, prev->node.prev);
prev              407 drivers/usb/c67x00/c67x00-sched.c 			last_urb = list_entry(urbp->ep_data->queue.prev,
prev              383 drivers/usb/chipidea/udc.c 		lastnode = list_entry(hwreq->tds.prev,
prev              461 drivers/usb/chipidea/udc.c 	lastnode = list_entry(hwreq->tds.prev,
prev              477 drivers/usb/chipidea/udc.c 		hwreqprev = list_entry(hwep->qh.queue.prev,
prev              479 drivers/usb/chipidea/udc.c 		prevlastnode = list_entry(hwreqprev->tds.prev,
prev             1796 drivers/usb/core/hcd.c 			urb = list_entry (ep->urb_list.prev, struct urb,
prev              790 drivers/usb/core/urb.c 		victim = list_entry(anchor->urb_list.prev, struct urb,
prev              823 drivers/usb/core/urb.c 		victim = list_entry(anchor->urb_list.prev, struct urb,
prev              976 drivers/usb/core/urb.c 		victim = list_entry(anchor->urb_list.prev, struct urb,
prev              259 drivers/usb/dwc2/debugfs.c 		   ep->queue.next, ep->queue.prev);
prev              433 drivers/usb/gadget/udc/aspeed-vhub/hub.c 	u16 prev;
prev              436 drivers/usb/gadget/udc/aspeed-vhub/hub.c 	prev = p->status;
prev              437 drivers/usb/gadget/udc/aspeed-vhub/hub.c 	p->status = (prev & ~clr_flags) | set_flags;
prev              439 drivers/usb/gadget/udc/aspeed-vhub/hub.c 	     port + 1, prev, p->status, set_c);
prev              443 drivers/usb/gadget/udc/aspeed-vhub/hub.c 		u16 chg = p->status ^ prev;
prev              738 drivers/usb/gadget/udc/fsl_udc_core.c 		lastreq = list_entry(ep->queue.prev, struct fsl_req, queue);
prev              975 drivers/usb/gadget/udc/fsl_udc_core.c 		prev_req = list_entry(req->queue.prev, struct fsl_req, queue);
prev             1222 drivers/usb/gadget/udc/goku_udc.c 			if (ep->dma && req->queue.prev == &ep->queue) {
prev              269 drivers/usb/gadget/udc/mv_udc_core.c 		lastreq = list_entry(ep->queue.prev, struct mv_req, queue);
prev              832 drivers/usb/gadget/udc/mv_udc_core.c 		prev_req = list_entry(req->queue.prev, struct mv_req, queue);
prev              495 drivers/usb/host/ehci-q.c 		if (stopped && qtd->qtd_list.prev != &qh->qtd_list) {
prev              496 drivers/usb/host/ehci-q.c 			last = list_entry (qtd->qtd_list.prev,
prev             1085 drivers/usb/host/ehci-q.c 			qtd = list_entry (qh->qtd_list.prev,
prev             1259 drivers/usb/host/ehci-q.c 	struct ehci_qh		*prev;
prev             1266 drivers/usb/host/ehci-q.c 	prev = ehci->async;
prev             1267 drivers/usb/host/ehci-q.c 	while (prev->qh_next.qh != qh)
prev             1268 drivers/usb/host/ehci-q.c 		prev = prev->qh_next.qh;
prev             1270 drivers/usb/host/ehci-q.c 	prev->hw->hw_next = qh->hw->hw_next;
prev             1271 drivers/usb/host/ehci-q.c 	prev->qh_next = qh->qh_next;
prev              550 drivers/usb/host/ehci-sched.c 		union ehci_shadow	*prev = &ehci->pshadow[i];
prev              552 drivers/usb/host/ehci-sched.c 		union ehci_shadow	here = *prev;
prev              560 drivers/usb/host/ehci-sched.c 			prev = periodic_next_shadow(ehci, prev, type);
prev              562 drivers/usb/host/ehci-sched.c 			here = *prev;
prev              571 drivers/usb/host/ehci-sched.c 			prev = &here.qh->qh_next;
prev              573 drivers/usb/host/ehci-sched.c 			here = *prev;
prev              581 drivers/usb/host/ehci-sched.c 			prev->qh = qh;
prev             1726 drivers/usb/host/ehci-sched.c 	union ehci_shadow	*prev = &ehci->pshadow[frame];
prev             1728 drivers/usb/host/ehci-sched.c 	union ehci_shadow	here = *prev;
prev             1736 drivers/usb/host/ehci-sched.c 		prev = periodic_next_shadow(ehci, prev, type);
prev             1738 drivers/usb/host/ehci-sched.c 		here = *prev;
prev             1743 drivers/usb/host/ehci-sched.c 	prev->itd = itd;
prev              280 drivers/usb/host/ehci-timer.c 				ehci->cached_itd_list.prev,
prev              283 drivers/usb/host/ehci-timer.c 				ehci->cached_sitd_list.prev,
prev             1232 drivers/usb/host/fotg210-hcd.c 				fotg210->cached_itd_list.prev,
prev             2443 drivers/usb/host/fotg210-hcd.c 		if (stopped && qtd->qtd_list.prev != &qh->qtd_list) {
prev             2444 drivers/usb/host/fotg210-hcd.c 			last = list_entry(qtd->qtd_list.prev,
prev             3007 drivers/usb/host/fotg210-hcd.c 			qtd = list_entry(qh->qtd_list.prev,
prev             3077 drivers/usb/host/fotg210-hcd.c 	struct fotg210_qh *prev;
prev             3088 drivers/usb/host/fotg210-hcd.c 	prev = fotg210->async;
prev             3089 drivers/usb/host/fotg210-hcd.c 	while (prev->qh_next.qh != qh)
prev             3090 drivers/usb/host/fotg210-hcd.c 		prev = prev->qh_next.qh;
prev             3092 drivers/usb/host/fotg210-hcd.c 	prev->hw->hw_next = qh->hw->hw_next;
prev             3093 drivers/usb/host/fotg210-hcd.c 	prev->qh_next = qh->qh_next;
prev             3500 drivers/usb/host/fotg210-hcd.c 		union fotg210_shadow *prev = &fotg210->pshadow[i];
prev             3502 drivers/usb/host/fotg210-hcd.c 		union fotg210_shadow here = *prev;
prev             3510 drivers/usb/host/fotg210-hcd.c 			prev = periodic_next_shadow(fotg210, prev, type);
prev             3512 drivers/usb/host/fotg210-hcd.c 			here = *prev;
prev             3521 drivers/usb/host/fotg210-hcd.c 			prev = &here.qh->qh_next;
prev             3523 drivers/usb/host/fotg210-hcd.c 			here = *prev;
prev             3531 drivers/usb/host/fotg210-hcd.c 			prev->qh = qh;
prev             4331 drivers/usb/host/fotg210-hcd.c 	union fotg210_shadow *prev = &fotg210->pshadow[frame];
prev             4333 drivers/usb/host/fotg210-hcd.c 	union fotg210_shadow here = *prev;
prev             4341 drivers/usb/host/fotg210-hcd.c 		prev = periodic_next_shadow(fotg210, prev, type);
prev             4343 drivers/usb/host/fotg210-hcd.c 		here = *prev;
prev             4348 drivers/usb/host/fotg210-hcd.c 	prev->itd = itd;
prev              807 drivers/usb/host/imx21-hcd.c 		urb->start_frame = wrap_frame(list_entry(ep_priv->td_list.prev,
prev              326 drivers/usb/host/isp116x-hcd.c 		struct isp116x_ep **prev = &isp116x->periodic[i];
prev              328 drivers/usb/host/isp116x-hcd.c 		while (*prev && ((temp = *prev) != ep))
prev              329 drivers/usb/host/isp116x-hcd.c 			prev = &temp->next;
prev              330 drivers/usb/host/isp116x-hcd.c 		if (*prev)
prev              331 drivers/usb/host/isp116x-hcd.c 			*prev = ep->next;
prev              565 drivers/usb/host/isp116x-hcd.c 	if ((&isp116x->async)->next != (&isp116x->async)->prev)
prev              797 drivers/usb/host/isp116x-hcd.c 			struct isp116x_ep **prev = &isp116x->periodic[i];
prev              798 drivers/usb/host/isp116x-hcd.c 			struct isp116x_ep *here = *prev;
prev              803 drivers/usb/host/isp116x-hcd.c 				prev = &here->next;
prev              804 drivers/usb/host/isp116x-hcd.c 				here = *prev;
prev              808 drivers/usb/host/isp116x-hcd.c 				*prev = ep;
prev              781 drivers/usb/host/isp1362-hcd.c 	if (isp1362_hcd->async.next != isp1362_hcd->async.prev) {
prev             1952 drivers/usb/host/max3421-hcd.c 	struct max3421_hcd *max3421_hcd = NULL, **prev;
prev             1956 drivers/usb/host/max3421-hcd.c 	for (prev = &max3421_hcd_list; *prev; prev = &(*prev)->next) {
prev             1957 drivers/usb/host/max3421-hcd.c 		max3421_hcd = *prev;
prev             1973 drivers/usb/host/max3421-hcd.c 	*prev = max3421_hcd->next;
prev              113 drivers/usb/host/ohci-mem.c 	struct td	**prev = &hc->td_hash [TD_HASH_FUNC (td->td_dma)];
prev              116 drivers/usb/host/ohci-mem.c 	while (*prev && *prev != td)
prev              117 drivers/usb/host/ohci-mem.c 		prev = &(*prev)->td_hash;
prev              118 drivers/usb/host/ohci-mem.c 	if (*prev)
prev              119 drivers/usb/host/ohci-mem.c 		*prev = td->td_hash;
prev              152 drivers/usb/host/ohci-q.c 		struct ed	**prev = &ohci->periodic [i];
prev              154 drivers/usb/host/ohci-q.c 		struct ed	*here = *prev;
prev              163 drivers/usb/host/ohci-q.c 			prev = &here->ed_next;
prev              165 drivers/usb/host/ohci-q.c 			here = *prev;
prev              172 drivers/usb/host/ohci-q.c 			*prev = ed;
prev              276 drivers/usb/host/ohci-q.c 		struct ed	**prev = &ohci->periodic [i];
prev              279 drivers/usb/host/ohci-q.c 		while (*prev && (temp = *prev) != ed) {
prev              281 drivers/usb/host/ohci-q.c 			prev = &temp->ed_next;
prev              283 drivers/usb/host/ohci-q.c 		if (*prev) {
prev              285 drivers/usb/host/ohci-q.c 			*prev = ed->ed_next;
prev              984 drivers/usb/host/ohci-q.c 		__hc32			*prev;
prev             1036 drivers/usb/host/ohci-q.c 		prev = &ed->hwHeadP;
prev             1049 drivers/usb/host/ohci-q.c 				prev = &td->hwNextTD;
prev             1054 drivers/usb/host/ohci-q.c 			savebits = *prev & ~cpu_to_hc32 (ohci, TD_MASK);
prev             1055 drivers/usb/host/ohci-q.c 			*prev = td->hwNextTD | savebits;
prev             1536 drivers/usb/host/oxu210hp-hcd.c 		if (stopped && qtd->qtd_list.prev != &qh->qtd_list) {
prev             1537 drivers/usb/host/oxu210hp-hcd.c 			last = list_entry(qtd->qtd_list.prev,
prev             2015 drivers/usb/host/oxu210hp-hcd.c 			list_splice(qtd_list, qh->qtd_list.prev);
prev             2022 drivers/usb/host/oxu210hp-hcd.c 			qtd = list_entry(qh->qtd_list.prev,
prev             2128 drivers/usb/host/oxu210hp-hcd.c 	struct ehci_qh *prev;
prev             2153 drivers/usb/host/oxu210hp-hcd.c 	prev = oxu->async;
prev             2154 drivers/usb/host/oxu210hp-hcd.c 	while (prev->qh_next.qh != qh)
prev             2155 drivers/usb/host/oxu210hp-hcd.c 		prev = prev->qh_next.qh;
prev             2157 drivers/usb/host/oxu210hp-hcd.c 	prev->hw_next = qh->hw_next;
prev             2158 drivers/usb/host/oxu210hp-hcd.c 	prev->qh_next = qh->qh_next;
prev             2367 drivers/usb/host/oxu210hp-hcd.c 		union ehci_shadow	*prev = &oxu->pshadow[i];
prev             2369 drivers/usb/host/oxu210hp-hcd.c 		union ehci_shadow	here = *prev;
prev             2377 drivers/usb/host/oxu210hp-hcd.c 			prev = periodic_next_shadow(prev, type);
prev             2379 drivers/usb/host/oxu210hp-hcd.c 			here = *prev;
prev             2388 drivers/usb/host/oxu210hp-hcd.c 			prev = &here.qh->qh_next;
prev             2390 drivers/usb/host/oxu210hp-hcd.c 			here = *prev;
prev             2398 drivers/usb/host/oxu210hp-hcd.c 			prev->qh = qh;
prev              454 drivers/usb/host/sl811-hcd.c 		struct sl811h_ep	**prev = &sl811->periodic[i];
prev              456 drivers/usb/host/sl811-hcd.c 		while (*prev && ((temp = *prev) != ep))
prev              457 drivers/usb/host/sl811-hcd.c 			prev = &temp->next;
prev              458 drivers/usb/host/sl811-hcd.c 		if (*prev)
prev              459 drivers/usb/host/sl811-hcd.c 			*prev = ep->next;
prev              927 drivers/usb/host/sl811-hcd.c 			struct sl811h_ep	**prev = &sl811->periodic[i];
prev              928 drivers/usb/host/sl811-hcd.c 			struct sl811h_ep	*here = *prev;
prev              933 drivers/usb/host/sl811-hcd.c 				prev = &here->next;
prev              934 drivers/usb/host/sl811-hcd.c 				here = *prev;
prev              938 drivers/usb/host/sl811-hcd.c 				*prev = ep;
prev               55 drivers/usb/host/uhci-q.c 	lqh = list_entry(uhci->skel_async_qh->node.prev,
prev               67 drivers/usb/host/uhci-q.c 	lqh = list_entry(uhci->skel_async_qh->node.prev,
prev              167 drivers/usb/host/uhci-q.c 		ltd = list_entry(ftd->fl_list.prev, struct uhci_td, fl_list);
prev              207 drivers/usb/host/uhci-q.c 		ptd = list_entry(td->fl_list.prev, struct uhci_td, fl_list);
prev              224 drivers/usb/host/uhci-q.c 		ltd = list_entry(ftd->fl_list.prev, struct uhci_td, fl_list);
prev              229 drivers/usb/host/uhci-q.c 			list_del_init(ftd->fl_list.prev);
prev              338 drivers/usb/host/uhci-q.c 		purbp = list_entry(urbp->node.prev, struct urb_priv, node);
prev              340 drivers/usb/host/uhci-q.c 		ptd = list_entry(purbp->td_list.prev, struct uhci_td,
prev              342 drivers/usb/host/uhci-q.c 		td = list_entry(urbp->td_list.prev, struct uhci_td,
prev              400 drivers/usb/host/uhci-q.c 			td = list_entry(urbp->td_list.prev, struct uhci_td,
prev              441 drivers/usb/host/uhci-q.c 	pqh = list_entry(qh->node.prev, struct uhci_qh, node);
prev              525 drivers/usb/host/uhci-q.c 	pqh = list_entry(qh->node.prev, struct uhci_qh, node);
prev              538 drivers/usb/host/uhci-q.c 	pqh = list_entry(qh->node.prev, struct uhci_qh, node);
prev             1129 drivers/usb/host/uhci-q.c 	td = list_entry(urbp->td_list.prev, struct uhci_td, list);
prev             1137 drivers/usb/host/uhci-q.c 		tmp = td->list.prev;
prev             1152 drivers/usb/host/uhci-q.c 		tmp = urbp->td_list.prev;
prev             1159 drivers/usb/host/uhci-q.c 		tmp = tmp->prev;
prev             1213 drivers/usb/host/uhci-q.c 				if (td->list.next != urbp->td_list.prev)
prev             1222 drivers/usb/host/uhci-q.c 			else if (&td->list != urbp->td_list.prev)
prev             1297 drivers/usb/host/uhci-q.c 			lurb = list_entry(qh->queue.prev,
prev             1534 drivers/usb/host/uhci-q.c 			urbp->node.prev == &qh->queue &&
prev               99 drivers/usb/host/xhci-mem.c static void xhci_link_segments(struct xhci_hcd *xhci, struct xhci_segment *prev,
prev              104 drivers/usb/host/xhci-mem.c 	if (!prev || !next)
prev              106 drivers/usb/host/xhci-mem.c 	prev->next = next;
prev              108 drivers/usb/host/xhci-mem.c 		prev->trbs[TRBS_PER_SEGMENT-1].link.segment_ptr =
prev              112 drivers/usb/host/xhci-mem.c 		val = le32_to_cpu(prev->trbs[TRBS_PER_SEGMENT-1].link.control);
prev              121 drivers/usb/host/xhci-mem.c 		prev->trbs[TRBS_PER_SEGMENT-1].link.control = cpu_to_le32(val);
prev              323 drivers/usb/host/xhci-mem.c 	struct xhci_segment *prev;
prev              325 drivers/usb/host/xhci-mem.c 	prev = xhci_segment_alloc(xhci, cycle_state, max_packet, flags);
prev              326 drivers/usb/host/xhci-mem.c 	if (!prev)
prev              330 drivers/usb/host/xhci-mem.c 	*first = prev;
prev              336 drivers/usb/host/xhci-mem.c 			prev = *first;
prev              337 drivers/usb/host/xhci-mem.c 			while (prev) {
prev              338 drivers/usb/host/xhci-mem.c 				next = prev->next;
prev              339 drivers/usb/host/xhci-mem.c 				xhci_segment_free(xhci, prev);
prev              340 drivers/usb/host/xhci-mem.c 				prev = next;
prev              344 drivers/usb/host/xhci-mem.c 		xhci_link_segments(xhci, prev, next, type);
prev              346 drivers/usb/host/xhci-mem.c 		prev = next;
prev              349 drivers/usb/host/xhci-mem.c 	xhci_link_segments(xhci, prev, *first, type);
prev              350 drivers/usb/host/xhci-mem.c 	*last = prev;
prev               32 drivers/usb/image/microtek.h 	struct mts_desc *prev;
prev              375 drivers/usb/musb/musb_host.c 				head = qh->ring.prev;
prev             2415 drivers/usb/musb/musb_host.c 			|| urb->urb_list.prev != &qh->hep->urb_list
prev              130 drivers/usb/usbip/vhci_hcd.c 		u32 prev = prev_status & bit;
prev              134 drivers/usb/usbip/vhci_hcd.c 		if (!prev && new)
prev              136 drivers/usb/usbip/vhci_hcd.c 		else if (prev && !new)
prev              141 drivers/usb/usbip/vhci_hcd.c 		if (prev || new) {
prev             1430 drivers/vfio/pci/vfio_pci_config.c 	u8 pos, *prev, cap;
prev             1446 drivers/vfio/pci/vfio_pci_config.c 	prev = &vdev->vconfig[PCI_CAPABILITY_LIST];
prev             1475 drivers/vfio/pci/vfio_pci_config.c 			*prev = next;
prev             1496 drivers/vfio/pci/vfio_pci_config.c 		prev = &vdev->vconfig[pos + PCI_CAP_LIST_NEXT];
prev             1515 drivers/vfio/pci/vfio_pci_config.c 	__le32 *prev = NULL;
prev             1551 drivers/vfio/pci/vfio_pci_config.c 			if (prev) {
prev             1553 drivers/vfio/pci/vfio_pci_config.c 				*prev &= cpu_to_le32(~(0xffcU << 20));
prev             1554 drivers/vfio/pci/vfio_pci_config.c 				*prev |= cpu_to_le32(val << 20);
prev             1598 drivers/vfio/pci/vfio_pci_config.c 		prev = (__le32 *)&vdev->vconfig[epos];
prev             1039 drivers/video/fbdev/core/fbmem.c 	if (info->modelist.prev && info->modelist.next &&
prev             1627 drivers/video/fbdev/core/fbmem.c 	if (!fb_info->modelist.prev || !fb_info->modelist.next)
prev             1151 drivers/video/fbdev/core/modedb.c 	if (!head->prev || !head->next || list_empty(head))
prev              303 drivers/video/fbdev/omap/hwa742.c 	list_splice_init(head, hwa742.pending_req_list.prev);
prev              470 drivers/video/fbdev/omap/hwa742.c 	last = list_entry(req_list.prev, struct hwa742_request, entry);
prev              501 drivers/video/fbdev/omap/hwa742.c 	last = list_entry(req_list.prev, struct hwa742_request, entry);
prev               20 drivers/video/fbdev/omap2/omapfb/dss/dss-of.c 			 struct device_node *prev)
prev               27 drivers/video/fbdev/omap2/omapfb/dss/dss-of.c 	if (!prev) {
prev               44 drivers/video/fbdev/omap2/omapfb/dss/dss-of.c 		ports = of_get_parent(prev);
prev               49 drivers/video/fbdev/omap2/omapfb/dss/dss-of.c 			port = of_get_next_child(ports, prev);
prev               54 drivers/video/fbdev/omap2/omapfb/dss/dss-of.c 			prev = port;
prev               66 drivers/video/fbdev/omap2/omapfb/dss/dss-of.c 			     struct device_node *prev)
prev               74 drivers/video/fbdev/omap2/omapfb/dss/dss-of.c 		ep = of_get_next_child(parent, prev);
prev               77 drivers/video/fbdev/omap2/omapfb/dss/dss-of.c 		prev = ep;
prev              427 drivers/virtio/virtio_ring.c 	unsigned int i, n, avail, descs_used, uninitialized_var(prev), err_idx;
prev              490 drivers/virtio/virtio_ring.c 			prev = i;
prev              503 drivers/virtio/virtio_ring.c 			prev = i;
prev              508 drivers/virtio/virtio_ring.c 	desc[prev].flags &= cpu_to_virtio16(_vq->vdev, ~VRING_DESC_F_NEXT);
prev             1105 drivers/virtio/virtio_ring.c 	u16 head, id, uninitialized_var(prev), curr, avail_used_flags;
prev             1171 drivers/virtio/virtio_ring.c 			prev = curr;
prev             1197 drivers/virtio/virtio_ring.c 	vq->packed.desc_state[id].last = prev;
prev             1021 drivers/vme/bridges/vme_ca91cx42.c 	struct ca91cx42_dma_entry *entry, *prev;
prev             1143 drivers/vme/bridges/vme_ca91cx42.c 	if (entry->list.prev != &list->entries) {
prev             1144 drivers/vme/bridges/vme_ca91cx42.c 		prev = list_entry(entry->list.prev, struct ca91cx42_dma_entry,
prev             1148 drivers/vme/bridges/vme_ca91cx42.c 		prev->descriptor.dcpp = desc_ptr & ~CA91CX42_DCPP_M;
prev             1627 drivers/vme/bridges/vme_tsi148.c 	struct tsi148_dma_entry *entry, *prev;
prev             1761 drivers/vme/bridges/vme_tsi148.c 	if (entry->list.prev != &list->entries) {
prev             1764 drivers/vme/bridges/vme_tsi148.c 		prev = list_entry(entry->list.prev, struct tsi148_dma_entry,
prev             1766 drivers/vme/bridges/vme_tsi148.c 		prev->descriptor.dnlau = cpu_to_be32(address_high);
prev             1767 drivers/vme/bridges/vme_tsi148.c 		prev->descriptor.dnlal = cpu_to_be32(address_low);
prev              457 fs/afs/file.c  	for (p = first->lru.prev; p != pages; p = p->prev) {
prev               97 fs/autofs/expire.c static struct dentry *get_next_positive_subdir(struct dentry *prev,
prev              105 fs/autofs/expire.c 	q = positive_after(root, prev);
prev              108 fs/autofs/expire.c 	dput(prev);
prev              115 fs/autofs/expire.c static struct dentry *get_next_positive_dentry(struct dentry *prev,
prev              119 fs/autofs/expire.c 	struct dentry *p = prev, *ret = NULL, *d = NULL;
prev              121 fs/autofs/expire.c 	if (prev == NULL)
prev              140 fs/autofs/expire.c 	dput(prev);
prev              679 fs/autofs/root.c 	    d_child->prev == &parent->d_subdirs)
prev              233 fs/btrfs/check-integrity.c 	struct btrfsic_stack_frame *prev;
prev              963 fs/btrfs/check-integrity.c 	sf->prev = NULL;
prev             1073 fs/btrfs/check-integrity.c 					next_stack->prev = sf;
prev             1166 fs/btrfs/check-integrity.c 				next_stack->prev = sf;
prev             1176 fs/btrfs/check-integrity.c 	if (NULL != sf->prev) {
prev             1177 fs/btrfs/check-integrity.c 		struct btrfsic_stack_frame *const prev = sf->prev;
prev             1183 fs/btrfs/check-integrity.c 			prev->error = sf->error;
prev             1185 fs/btrfs/check-integrity.c 			sf = prev;
prev             1190 fs/btrfs/check-integrity.c 		sf = prev;
prev              343 fs/btrfs/delayed-inode.c 				struct btrfs_delayed_item **prev,
prev              365 fs/btrfs/delayed-inode.c 	if (prev) {
prev              367 fs/btrfs/delayed-inode.c 			*prev = NULL;
prev              369 fs/btrfs/delayed-inode.c 			*prev = delayed_item;
prev              371 fs/btrfs/delayed-inode.c 			*prev = rb_entry(node, struct btrfs_delayed_item,
prev              374 fs/btrfs/delayed-inode.c 			*prev = NULL;
prev              840 fs/btrfs/delayed-inode.c 	struct btrfs_delayed_item *curr, *prev;
prev              855 fs/btrfs/delayed-inode.c 	prev = curr;
prev              856 fs/btrfs/delayed-inode.c 	curr = __btrfs_next_delayed_item(prev);
prev              857 fs/btrfs/delayed-inode.c 	if (curr && btrfs_is_continuous_delayed_item(prev, curr)) {
prev              862 fs/btrfs/delayed-inode.c 	btrfs_release_delayed_item(prev);
prev              942 fs/btrfs/delayed-inode.c 	struct btrfs_delayed_item *curr, *prev;
prev              962 fs/btrfs/delayed-inode.c 		prev = curr;
prev              963 fs/btrfs/delayed-inode.c 		curr = __btrfs_next_delayed_item(prev);
prev              964 fs/btrfs/delayed-inode.c 		btrfs_release_delayed_item(prev);
prev              388 fs/btrfs/extent_io.c 	struct rb_node *prev = NULL;
prev              394 fs/btrfs/extent_io.c 		prev = *n;
prev              395 fs/btrfs/extent_io.c 		entry = rb_entry(prev, struct tree_entry, rb_node);
prev              409 fs/btrfs/extent_io.c 		*parent_ret = prev;
prev              412 fs/btrfs/extent_io.c 		orig_prev = prev;
prev              413 fs/btrfs/extent_io.c 		while (prev && offset > prev_entry->end) {
prev              414 fs/btrfs/extent_io.c 			prev = rb_next(prev);
prev              415 fs/btrfs/extent_io.c 			prev_entry = rb_entry(prev, struct tree_entry, rb_node);
prev              417 fs/btrfs/extent_io.c 		*next_ret = prev;
prev              418 fs/btrfs/extent_io.c 		prev = orig_prev;
prev              422 fs/btrfs/extent_io.c 		prev_entry = rb_entry(prev, struct tree_entry, rb_node);
prev              423 fs/btrfs/extent_io.c 		while (prev && offset < prev_entry->start) {
prev              424 fs/btrfs/extent_io.c 			prev = rb_prev(prev);
prev              425 fs/btrfs/extent_io.c 			prev_entry = rb_entry(prev, struct tree_entry, rb_node);
prev              427 fs/btrfs/extent_io.c 		*prev_ret = prev;
prev             1579 fs/btrfs/extent_io.c 	struct rb_node *node, *prev = NULL, *next;
prev             1585 fs/btrfs/extent_io.c 		node = __etree_search(tree, start, &next, &prev, NULL, NULL);
prev             1586 fs/btrfs/extent_io.c 		if (!node && !next && !prev) {
prev             1599 fs/btrfs/extent_io.c 			state = rb_entry(prev, struct extent_state, rb_node);
prev             1645 fs/btrfs/extent_io.c 			if (prev) {
prev             1646 fs/btrfs/extent_io.c 				state = rb_entry(prev, struct extent_state,
prev              150 fs/btrfs/extent_map.c 	struct rb_node *prev = NULL;
prev              157 fs/btrfs/extent_map.c 		prev = n;
prev              169 fs/btrfs/extent_map.c 		orig_prev = prev;
prev              170 fs/btrfs/extent_map.c 		while (prev && offset >= extent_map_end(prev_entry)) {
prev              171 fs/btrfs/extent_map.c 			prev = rb_next(prev);
prev              172 fs/btrfs/extent_map.c 			prev_entry = rb_entry(prev, struct extent_map, rb_node);
prev              174 fs/btrfs/extent_map.c 		*prev_ret = prev;
prev              175 fs/btrfs/extent_map.c 		prev = orig_prev;
prev              179 fs/btrfs/extent_map.c 		prev_entry = rb_entry(prev, struct extent_map, rb_node);
prev              180 fs/btrfs/extent_map.c 		while (prev && offset < prev_entry->start) {
prev              181 fs/btrfs/extent_map.c 			prev = rb_prev(prev);
prev              182 fs/btrfs/extent_map.c 			prev_entry = rb_entry(prev, struct extent_map, rb_node);
prev              184 fs/btrfs/extent_map.c 		*next_ret = prev;
prev              190 fs/btrfs/extent_map.c static int mergable_maps(struct extent_map *prev, struct extent_map *next)
prev              192 fs/btrfs/extent_map.c 	if (test_bit(EXTENT_FLAG_PINNED, &prev->flags))
prev              199 fs/btrfs/extent_map.c 	if (test_bit(EXTENT_FLAG_COMPRESSED, &prev->flags))
prev              202 fs/btrfs/extent_map.c 	if (test_bit(EXTENT_FLAG_LOGGING, &prev->flags) ||
prev              211 fs/btrfs/extent_map.c 	if (!list_empty(&prev->list) || !list_empty(&next->list))
prev              215 fs/btrfs/extent_map.c 	       prev->block_start != EXTENT_MAP_DELALLOC);
prev              217 fs/btrfs/extent_map.c 	if (extent_map_end(prev) == next->start &&
prev              218 fs/btrfs/extent_map.c 	    prev->flags == next->flags &&
prev              219 fs/btrfs/extent_map.c 	    prev->bdev == next->bdev &&
prev              221 fs/btrfs/extent_map.c 	      prev->block_start == EXTENT_MAP_HOLE) ||
prev              223 fs/btrfs/extent_map.c 	      prev->block_start == EXTENT_MAP_INLINE) ||
prev              225 fs/btrfs/extent_map.c 	      next->block_start == extent_map_block_end(prev)))) {
prev              419 fs/btrfs/extent_map.c 	struct rb_node *prev = NULL;
prev              423 fs/btrfs/extent_map.c 	rb_node = __tree_search(&tree->map.rb_root, start, &prev, &next);
prev              425 fs/btrfs/extent_map.c 		if (prev)
prev              426 fs/btrfs/extent_map.c 			rb_node = prev;
prev              522 fs/btrfs/extent_map.c 	struct rb_node *prev;
prev              524 fs/btrfs/extent_map.c 	prev = rb_prev(&em->rb_node);
prev              525 fs/btrfs/extent_map.c 	if (!prev)
prev              527 fs/btrfs/extent_map.c 	return container_of(prev, struct extent_map, rb_node);
prev              541 fs/btrfs/extent_map.c 	struct extent_map *prev;
prev              551 fs/btrfs/extent_map.c 		prev = prev_extent_map(next);
prev              553 fs/btrfs/extent_map.c 		prev = existing;
prev              554 fs/btrfs/extent_map.c 		next = next_extent_map(prev);
prev              557 fs/btrfs/extent_map.c 	start = prev ? extent_map_end(prev) : em->start;
prev             2910 fs/btrfs/file.c 	struct falloc_range *prev = NULL;
prev             2920 fs/btrfs/file.c 	prev = list_entry(head->prev, struct falloc_range, list);
prev             2921 fs/btrfs/file.c 	if (prev->start + prev->len == start) {
prev             2922 fs/btrfs/file.c 		prev->len += len;
prev              639 fs/btrfs/free-space-cache.c 	struct btrfs_free_space *e, *prev = NULL;
prev              646 fs/btrfs/free-space-cache.c 		if (!prev)
prev              648 fs/btrfs/free-space-cache.c 		if (e->bitmap || prev->bitmap)
prev              650 fs/btrfs/free-space-cache.c 		if (prev->offset + prev->bytes == e->offset) {
prev              651 fs/btrfs/free-space-cache.c 			unlink_free_space(ctl, prev);
prev              653 fs/btrfs/free-space-cache.c 			prev->bytes += e->bytes;
prev              655 fs/btrfs/free-space-cache.c 			link_free_space(ctl, prev);
prev              656 fs/btrfs/free-space-cache.c 			prev = NULL;
prev              661 fs/btrfs/free-space-cache.c 		prev = e;
prev             1512 fs/btrfs/free-space-cache.c 	struct btrfs_free_space *entry, *prev = NULL;
prev             1522 fs/btrfs/free-space-cache.c 		prev = entry;
prev             1559 fs/btrfs/free-space-cache.c 				prev = rb_entry(n, struct btrfs_free_space,
prev             1561 fs/btrfs/free-space-cache.c 				if (!prev->bitmap &&
prev             1562 fs/btrfs/free-space-cache.c 				    prev->offset + prev->bytes > offset)
prev             1563 fs/btrfs/free-space-cache.c 					entry = prev;
prev             1569 fs/btrfs/free-space-cache.c 	if (!prev)
prev             1573 fs/btrfs/free-space-cache.c 	entry = prev;
prev             1591 fs/btrfs/free-space-cache.c 			prev = rb_entry(n, struct btrfs_free_space,
prev             1593 fs/btrfs/free-space-cache.c 			if (!prev->bitmap &&
prev             1594 fs/btrfs/free-space-cache.c 			    prev->offset + prev->bytes > offset)
prev             1595 fs/btrfs/free-space-cache.c 				return prev;
prev             2713 fs/btrfs/inode.c 				 struct sa_defrag_extent_backref *prev,
prev             2737 fs/btrfs/inode.c 	if (prev && prev->root_id == backref->root_id &&
prev             2738 fs/btrfs/inode.c 	    prev->inum == backref->inum &&
prev             2739 fs/btrfs/inode.c 	    prev->file_pos + prev->num_bytes == backref->file_pos)
prev             2931 fs/btrfs/inode.c 	struct sa_defrag_extent_backref *prev = NULL;
prev             2953 fs/btrfs/inode.c 		ret = relink_extent_backref(path, prev, backref);
prev             2956 fs/btrfs/inode.c 		kfree(prev);
prev             2959 fs/btrfs/inode.c 			prev = backref;
prev             2961 fs/btrfs/inode.c 			prev = NULL;
prev             2964 fs/btrfs/inode.c 	kfree(prev);
prev             4446 fs/btrfs/inode.c 	struct rb_node *prev;
prev             4457 fs/btrfs/inode.c 	prev = NULL;
prev             4459 fs/btrfs/inode.c 		prev = node;
prev             4470 fs/btrfs/inode.c 		while (prev) {
prev             4471 fs/btrfs/inode.c 			entry = rb_entry(prev, struct btrfs_inode, rb_node);
prev             4473 fs/btrfs/inode.c 				node = prev;
prev             4476 fs/btrfs/inode.c 			prev = rb_next(prev);
prev               71 fs/btrfs/ordered-data.c 	struct rb_node *prev = NULL;
prev               78 fs/btrfs/ordered-data.c 		prev = n;
prev               91 fs/btrfs/ordered-data.c 	while (prev && file_offset >= entry_end(prev_entry)) {
prev               92 fs/btrfs/ordered-data.c 		test = rb_next(prev);
prev              100 fs/btrfs/ordered-data.c 		prev = test;
prev              102 fs/btrfs/ordered-data.c 	if (prev)
prev              103 fs/btrfs/ordered-data.c 		prev_entry = rb_entry(prev, struct btrfs_ordered_extent,
prev              105 fs/btrfs/ordered-data.c 	while (prev && file_offset < entry_end(prev_entry)) {
prev              106 fs/btrfs/ordered-data.c 		test = rb_prev(prev);
prev              111 fs/btrfs/ordered-data.c 		prev = test;
prev              113 fs/btrfs/ordered-data.c 	*prev_ret = prev;
prev              145 fs/btrfs/ordered-data.c 	struct rb_node *prev = NULL;
prev              155 fs/btrfs/ordered-data.c 	ret = __tree_search(root, file_offset, &prev);
prev              157 fs/btrfs/ordered-data.c 		ret = prev;
prev              811 fs/btrfs/ordered-data.c 	struct rb_node *prev = NULL;
prev              868 fs/btrfs/ordered-data.c 		prev = tree_search(tree, offset);
prev              873 fs/btrfs/ordered-data.c 		if (prev) {
prev              874 fs/btrfs/ordered-data.c 			test = rb_entry(prev, struct btrfs_ordered_extent,
prev              878 fs/btrfs/ordered-data.c 		node = prev;
prev              494 fs/btrfs/raid56.c 		found = list_entry(table->stripe_cache.prev,
prev             1554 fs/btrfs/relocation.c 	struct rb_node *prev;
prev             1561 fs/btrfs/relocation.c 	prev = NULL;
prev             1563 fs/btrfs/relocation.c 		prev = node;
prev             1574 fs/btrfs/relocation.c 		while (prev) {
prev             1575 fs/btrfs/relocation.c 			entry = rb_entry(prev, struct btrfs_inode, rb_node);
prev             1577 fs/btrfs/relocation.c 				node = prev;
prev             1580 fs/btrfs/relocation.c 			prev = rb_next(prev);
prev             2059 fs/btrfs/transaction.c 	if (cur_trans->list.prev != &fs_info->trans_list) {
prev             2060 fs/btrfs/transaction.c 		prev_trans = list_entry(cur_trans->list.prev,
prev             2394 fs/btrfs/transaction.c 	unsigned long prev;
prev             2397 fs/btrfs/transaction.c 	prev = xchg(&fs_info->pending_changes, 0);
prev             2398 fs/btrfs/transaction.c 	if (!prev)
prev             2402 fs/btrfs/transaction.c 	if (prev & bit)
prev             2404 fs/btrfs/transaction.c 	prev &= ~bit;
prev             2407 fs/btrfs/transaction.c 	if (prev & bit)
prev             2409 fs/btrfs/transaction.c 	prev &= ~bit;
prev             2412 fs/btrfs/transaction.c 	if (prev & bit)
prev             2414 fs/btrfs/transaction.c 	prev &= ~bit;
prev             2416 fs/btrfs/transaction.c 	if (prev)
prev             2418 fs/btrfs/transaction.c 			"unknown pending changes left 0x%lx, ignoring", prev);
prev              724 fs/buffer.c    		bh = BH_ENTRY(tmp.prev);
prev              383 fs/ceph/addr.c 		page = list_entry(page_list->prev, struct page, lru);
prev             1694 fs/ceph/caps.c 	struct ceph_cap_flush *prev;
prev             1698 fs/ceph/caps.c 		if (wake && cf->g_list.prev != &mdsc->cap_flush_list) {
prev             1699 fs/ceph/caps.c 			prev = list_prev_entry(cf, g_list);
prev             1700 fs/ceph/caps.c 			prev->wake = true;
prev             1705 fs/ceph/caps.c 		if (wake && cf->i_list.prev != &ci->i_cap_flush_list) {
prev             1706 fs/ceph/caps.c 			prev = list_prev_entry(cf, i_list);
prev             1707 fs/ceph/caps.c 			prev->wake = true;
prev             3042 fs/ceph/caps.c 	struct dentry *dn, *prev = NULL;
prev             3056 fs/ceph/caps.c 		if (dn == prev) {
prev             3061 fs/ceph/caps.c 		if (prev)
prev             3062 fs/ceph/caps.c 			dput(prev);
prev             3063 fs/ceph/caps.c 		prev = dn;
prev             3065 fs/ceph/caps.c 	if (prev)
prev             3066 fs/ceph/caps.c 		dput(prev);
prev             1496 fs/ceph/mds_client.c 		struct ceph_cap *cap, *prev = NULL;
prev             1508 fs/ceph/mds_client.c 			if (cap == prev)
prev             1510 fs/ceph/mds_client.c 			prev = cap;
prev              100 fs/char_dev.c  	struct char_device_struct *cd, *curr, *prev = NULL;
prev              134 fs/char_dev.c  	for (curr = chrdevs[i]; curr; prev = curr, curr = curr->next) {
prev              155 fs/char_dev.c  	if (!prev) {
prev              159 fs/char_dev.c  		cd->next = prev->next;
prev              160 fs/char_dev.c  		prev->next = cd;
prev             1110 fs/cifs/file.c 					(lock->blist.prev == &lock->blist) &&
prev              123 fs/crypto/hkdf.c 	const u8 *prev = NULL;
prev              141 fs/crypto/hkdf.c 		if (prev) {
prev              142 fs/crypto/hkdf.c 			err = crypto_shash_update(desc, prev, HKDF_HASHLEN);
prev              168 fs/crypto/hkdf.c 		prev = &okm[i];
prev             1101 fs/dcache.c    		dentry = list_entry(list->prev, struct dentry, d_lru);
prev             1307 fs/dlm/lock.c  	__list_add(new, lkb->lkb_statequeue.prev, &lkb->lkb_statequeue);
prev             1434 fs/dlm/lowcomms.c 	e = list_entry(con->writequeue.prev, struct writequeue_entry, list);
prev              307 fs/dlm/member.c 		newlist->prev = tmp->prev;
prev              309 fs/dlm/member.c 		tmp->prev->next = newlist;
prev              310 fs/dlm/member.c 		tmp->prev = newlist;
prev              296 fs/erofs/data.c 		page = list_entry(pages->prev, struct page, lru);
prev             1137 fs/eventpoll.c 	struct list_head *prev;
prev             1156 fs/eventpoll.c 	prev = xchg(&head->prev, new);
prev             1163 fs/eventpoll.c 	prev->next = new;
prev             1164 fs/eventpoll.c 	new->prev = prev;
prev              699 fs/exec.c      	struct vm_area_struct *prev = NULL;
prev              759 fs/exec.c      	ret = mprotect_fixup(vma, &prev, vma->vm_start, vma->vm_end,
prev              763 fs/exec.c      	BUG_ON(prev != vma);
prev              205 fs/ext2/balloc.c 	struct ext2_reserve_window_node *rsv, *prev;
prev              211 fs/ext2/balloc.c 	prev = NULL;
prev              225 fs/ext2/balloc.c 		if (prev && prev->rsv_end >= rsv->rsv_start) {
prev              238 fs/ext2/balloc.c 		prev = rsv;
prev              784 fs/ext2/balloc.c 	struct ext2_reserve_window_node *rsv, *prev;
prev              811 fs/ext2/balloc.c 		prev = rsv;
prev              841 fs/ext2/balloc.c 	if ((prev != my_rsv) && (!rsv_is_empty(&my_rsv->rsv_window)))
prev              855 fs/ext2/balloc.c 	if (prev != my_rsv)
prev              407 fs/ext4/extents.c 		ext4_lblk_t prev = 0;
prev              416 fs/ext4/extents.c 			if ((lblock <= prev) && prev) {
prev              423 fs/ext4/extents.c 			prev = lblock + len - 1;
prev              505 fs/ext4/extents.c 	ext4_lblk_t prev = 0;
prev              513 fs/ext4/extents.c 		if (prev && (prev != lblk))
prev              514 fs/ext4/extents.c 			ext4_es_cache_extent(inode, prev, lblk - prev, ~0,
prev              521 fs/ext4/extents.c 		prev = lblk + len;
prev              373 fs/ext4/fsmap.c 	struct ext4_fsmap *prev = NULL;
prev              377 fs/ext4/fsmap.c 		if (!prev) {
prev              378 fs/ext4/fsmap.c 			prev = p;
prev              382 fs/ext4/fsmap.c 		if (prev->fmr_owner == p->fmr_owner &&
prev              383 fs/ext4/fsmap.c 		    prev->fmr_physical + prev->fmr_length == p->fmr_physical) {
prev              384 fs/ext4/fsmap.c 			prev->fmr_length += p->fmr_length;
prev              388 fs/ext4/fsmap.c 			prev = p;
prev             1787 fs/ext4/namei.c 	struct ext4_dir_entry_2 *next, *to, *prev, *de = (struct ext4_dir_entry_2 *) base;
prev             1790 fs/ext4/namei.c 	prev = to = de;
prev             1798 fs/ext4/namei.c 			prev = to;
prev             1803 fs/ext4/namei.c 	return prev;
prev             2988 fs/ext4/namei.c 	struct list_head *prev;
prev             3012 fs/ext4/namei.c 	prev = ei->i_orphan.prev;
prev             3025 fs/ext4/namei.c 	if (prev == &sbi->s_orphan) {
prev             3039 fs/ext4/namei.c 			&list_entry(prev, struct ext4_inode_info, i_orphan)->vfs_inode;
prev              149 fs/ext4/page-io.c 		   io, inode->i_ino, io->list.next, io->list.prev);
prev              177 fs/ext4/page-io.c 		before = cur->prev;
prev              366 fs/f2fs/checkpoint.c 	pgoff_t index = 0, prev = ULONG_MAX;
prev              386 fs/f2fs/checkpoint.c 			if (prev == ULONG_MAX)
prev              387 fs/f2fs/checkpoint.c 				prev = page->index - 1;
prev              388 fs/f2fs/checkpoint.c 			if (nr_to_write != LONG_MAX && page->index != prev + 1) {
prev              415 fs/f2fs/checkpoint.c 			prev = page->index;
prev              503 fs/f2fs/extent_cache.c 	struct extent_info ei, dei, prev;
prev              522 fs/f2fs/extent_cache.c 	prev = et->largest;
prev              609 fs/f2fs/extent_cache.c 				prev.len < F2FS_MIN_EXTENT_LEN &&
prev             2749 fs/f2fs/node.c 			list_add(&nes->set_list, cur->set_list.prev);
prev              169 fs/fat/cache.c 			struct list_head *p = MSDOS_I(inode)->cache_lru.prev;
prev             1290 fs/fat/dir.c   	struct buffer_head *bh, *prev, *bhs[3]; /* 32*slots (672bytes) */
prev             1299 fs/fat/dir.c   	bh = prev = NULL;
prev             1308 fs/fat/dir.c   			if (prev != bh) {
prev             1310 fs/fat/dir.c   				bhs[nr_bhs] = prev = bh;
prev             1319 fs/fat/dir.c   			prev = NULL;
prev              299 fs/fat/namei_vfat.c 		unsigned char prev = buf[0];
prev              308 fs/fat/namei_vfat.c 			if (buf[0] == prev)
prev              435 fs/fs-writeback.c 		inode_io_list_move_locked(inode, new_wb, pos->i_io_list.prev);
prev             1247 fs/fs-writeback.c 		inode = wb_inode(delaying_queue->prev);
prev             1270 fs/fs-writeback.c 		sb = wb_inode(tmp.prev)->i_sb;
prev             1643 fs/fs-writeback.c 		struct inode *inode = wb_inode(wb->b_io.prev);
prev             1773 fs/fs-writeback.c 		struct inode *inode = wb_inode(wb->b_io.prev);
prev             1920 fs/fs-writeback.c 		inode = wb_inode(wb->b_more_io.prev);
prev             2248 fs/gfs2/bmap.c 		jext = list_entry(jd->extent_list.prev, struct gfs2_journal_extent, list);
prev              563 fs/gfs2/dir.c  	struct gfs2_dirent *dent, *prev;
prev              573 fs/gfs2/dir.c  	prev = NULL;
prev              585 fs/gfs2/dir.c  		prev = dent;
prev              599 fs/gfs2/dir.c  		return prev ? prev : dent;
prev              673 fs/gfs2/dir.c  		       struct gfs2_dirent *prev, struct gfs2_dirent *cur)
prev              688 fs/gfs2/dir.c  	if (!prev) {
prev              696 fs/gfs2/dir.c  	prev_rec_len = be16_to_cpu(prev->de_rec_len);
prev              699 fs/gfs2/dir.c  	if ((char *)prev + prev_rec_len != (char *)cur)
prev              705 fs/gfs2/dir.c  	prev->de_rec_len = cpu_to_be16(prev_rec_len);
prev             1009 fs/gfs2/dir.c  	struct gfs2_dirent *dent = NULL, *prev = NULL, *next = NULL, *new;
prev             1108 fs/gfs2/dir.c  			dirent_del(dip, obh, prev, dent);
prev             1114 fs/gfs2/dir.c  			if (!prev)
prev             1115 fs/gfs2/dir.c  				prev = dent;
prev             1117 fs/gfs2/dir.c  			prev = dent;
prev             1878 fs/gfs2/dir.c  	struct gfs2_dirent *dent, *prev = NULL;
prev             1895 fs/gfs2/dir.c  		prev = dent;
prev             1896 fs/gfs2/dir.c  		dent = (struct gfs2_dirent *)((char *)dent + be16_to_cpu(prev->de_rec_len));
prev             1899 fs/gfs2/dir.c  	dirent_del(dip, bh, prev, dent);
prev              360 fs/gfs2/glock.c 			if (gh->gh_list.prev == &gl->gl_holders &&
prev              385 fs/gfs2/glock.c 		if (gh->gh_list.prev == &gl->gl_holders)
prev              283 fs/gfs2/log.c  		bd = list_entry(head->prev, struct gfs2_bufdata,
prev              491 fs/gfs2/log.c  		tr = list_entry(sdp->sd_ail1_list.prev, struct gfs2_trans,
prev              474 fs/gfs2/lops.c static struct bio *gfs2_chain_bio(struct bio *prev, unsigned int nr_iovecs)
prev              479 fs/gfs2/lops.c 	bio_copy_dev(new, prev);
prev              480 fs/gfs2/lops.c 	new->bi_iter.bi_sector = bio_end_sector(prev);
prev              481 fs/gfs2/lops.c 	new->bi_opf = prev->bi_opf;
prev              482 fs/gfs2/lops.c 	new->bi_write_hint = prev->bi_write_hint;
prev              483 fs/gfs2/lops.c 	bio_chain(new, prev);
prev              484 fs/gfs2/lops.c 	submit_bio(prev);
prev             1443 fs/gfs2/quota.c 		qd = list_entry(head->prev, struct gfs2_quota_data, qd_list);
prev               75 fs/gfs2/xattr.c 			  struct gfs2_ea_header *prev, void *private);
prev               80 fs/gfs2/xattr.c 	struct gfs2_ea_header *ea, *prev = NULL;
prev               86 fs/gfs2/xattr.c 	for (ea = GFS2_EA_BH2FIRST(bh);; prev = ea, ea = GFS2_EA2NEXT(ea)) {
prev               95 fs/gfs2/xattr.c 		error = ea_call(ip, bh, ea, prev, data);
prev              165 fs/gfs2/xattr.c 		     struct gfs2_ea_header *ea, struct gfs2_ea_header *prev,
prev              180 fs/gfs2/xattr.c 			el->el_prev = prev;
prev              225 fs/gfs2/xattr.c 				struct gfs2_ea_header *prev, void *private)
prev              294 fs/gfs2/xattr.c 	if (prev && !leave) {
prev              297 fs/gfs2/xattr.c 		len = GFS2_EA_REC_LEN(prev) + GFS2_EA_REC_LEN(ea);
prev              298 fs/gfs2/xattr.c 		prev->ea_rec_len = cpu_to_be32(len);
prev              301 fs/gfs2/xattr.c 			prev->ea_flags |= GFS2_EAFLAG_LAST;
prev              319 fs/gfs2/xattr.c 			       struct gfs2_ea_header *prev, int leave)
prev              331 fs/gfs2/xattr.c 	error = ea_dealloc_unstuffed(ip, bh, ea, prev, (leave) ? &error : NULL);
prev              344 fs/gfs2/xattr.c 		     struct gfs2_ea_header *ea, struct gfs2_ea_header *prev,
prev              830 fs/gfs2/xattr.c 	struct gfs2_ea_header *prev = el->el_prev;
prev              835 fs/gfs2/xattr.c 	if (!prev || !GFS2_EA_IS_STUFFED(ea)) {
prev              838 fs/gfs2/xattr.c 	} else if (GFS2_EA2NEXT(prev) != ea) {
prev              839 fs/gfs2/xattr.c 		prev = GFS2_EA2NEXT(prev);
prev              840 fs/gfs2/xattr.c 		gfs2_assert_withdraw(GFS2_SB(&ip->i_inode), GFS2_EA2NEXT(prev) == ea);
prev              843 fs/gfs2/xattr.c 	len = GFS2_EA_REC_LEN(prev) + GFS2_EA_REC_LEN(ea);
prev              844 fs/gfs2/xattr.c 	prev->ea_rec_len = cpu_to_be32(len);
prev              847 fs/gfs2/xattr.c 		prev->ea_flags |= GFS2_EAFLAG_LAST;
prev              910 fs/gfs2/xattr.c 			 struct gfs2_ea_header *ea, struct gfs2_ea_header *prev,
prev              925 fs/gfs2/xattr.c 			error = ea_remove_unstuffed(ip, bh, ea, prev, 1);
prev             1077 fs/gfs2/xattr.c 	struct gfs2_ea_header *prev = el->el_prev;
prev             1086 fs/gfs2/xattr.c 	if (prev) {
prev             1089 fs/gfs2/xattr.c 		len = GFS2_EA_REC_LEN(prev) + GFS2_EA_REC_LEN(ea);
prev             1090 fs/gfs2/xattr.c 		prev->ea_rec_len = cpu_to_be32(len);
prev             1093 fs/gfs2/xattr.c 			prev->ea_flags |= GFS2_EAFLAG_LAST;
prev              177 fs/hfs/bfind.c 			idx = bnode->prev;
prev              143 fs/hfs/bnode.c 		be32_to_cpu(desc.next), be32_to_cpu(desc.prev),
prev              178 fs/hfs/bnode.c 	if (node->prev) {
prev              179 fs/hfs/bnode.c 		tmp = hfs_bnode_find(tree, node->prev);
prev              193 fs/hfs/bnode.c 		tmp->prev = node->prev;
prev              194 fs/hfs/bnode.c 		cnid = cpu_to_be32(tmp->prev);
prev              195 fs/hfs/bnode.c 		hfs_bnode_write(tmp, &cnid, offsetof(struct hfs_bnode_desc, prev), 4);
prev              198 fs/hfs/bnode.c 		tree->leaf_tail = node->prev;
prev              201 fs/hfs/bnode.c 	if (!node->prev && !node->next) {
prev              339 fs/hfs/bnode.c 	node->prev = be32_to_cpu(desc->prev);
prev              248 fs/hfs/brec.c  	new_node->prev = node->this;
prev              318 fs/hfs/brec.c  	node_desc.prev = cpu_to_be32(new_node->prev);
prev              334 fs/hfs/brec.c  		next_node->prev = new_node->this;
prev              336 fs/hfs/brec.c  		node_desc.prev = cpu_to_be32(next_node->prev);
prev              488 fs/hfs/brec.c  	new_node->prev = 0;
prev              492 fs/hfs/brec.c  	node_desc.prev = cpu_to_be32(new_node->prev);
prev              188 fs/hfs/btree.c static struct hfs_bnode *hfs_bmap_new_bmap(struct hfs_bnode *prev, u32 idx)
prev              190 fs/hfs/btree.c 	struct hfs_btree *tree = prev->tree;
prev              202 fs/hfs/btree.c 	prev->next = idx;
prev              204 fs/hfs/btree.c 	hfs_bnode_write(prev, &cnid, offsetof(struct hfs_bnode_desc, next), 4);
prev              210 fs/hfs/btree.c 	desc.prev = 0;
prev               49 fs/hfs/btree.h 	u32 prev;
prev              128 fs/hfs/btree.h 	__be32 prev;		/* (V) Number of the prev node at this level */
prev              246 fs/hfsplus/bfind.c 			idx = bnode->prev;
prev              307 fs/hfsplus/bnode.c 		be32_to_cpu(desc.next), be32_to_cpu(desc.prev),
prev              342 fs/hfsplus/bnode.c 	if (node->prev) {
prev              343 fs/hfsplus/bnode.c 		tmp = hfs_bnode_find(tree, node->prev);
prev              358 fs/hfsplus/bnode.c 		tmp->prev = node->prev;
prev              359 fs/hfsplus/bnode.c 		cnid = cpu_to_be32(tmp->prev);
prev              361 fs/hfsplus/bnode.c 			offsetof(struct hfs_bnode_desc, prev), 4);
prev              364 fs/hfsplus/bnode.c 		tree->leaf_tail = node->prev;
prev              367 fs/hfsplus/bnode.c 	if (!node->prev && !node->next)
prev              507 fs/hfsplus/bnode.c 	node->prev = be32_to_cpu(desc->prev);
prev              252 fs/hfsplus/brec.c 	new_node->prev = node->this;
prev              322 fs/hfsplus/brec.c 	node_desc.prev = cpu_to_be32(new_node->prev);
prev              338 fs/hfsplus/brec.c 		next_node->prev = new_node->this;
prev              340 fs/hfsplus/brec.c 		node_desc.prev = cpu_to_be32(next_node->prev);
prev              492 fs/hfsplus/brec.c 	new_node->prev = 0;
prev              496 fs/hfsplus/brec.c 	node_desc.prev = cpu_to_be32(new_node->prev);
prev              312 fs/hfsplus/btree.c static struct hfs_bnode *hfs_bmap_new_bmap(struct hfs_bnode *prev, u32 idx)
prev              314 fs/hfsplus/btree.c 	struct hfs_btree *tree = prev->tree;
prev              324 fs/hfsplus/btree.c 	prev->next = idx;
prev              326 fs/hfsplus/btree.c 	hfs_bnode_write(prev, &cnid, offsetof(struct hfs_bnode_desc, next), 4);
prev              332 fs/hfsplus/btree.c 	desc.prev = 0;
prev              106 fs/hfsplus/hfsplus_fs.h 	u32 prev;
prev              153 fs/hfsplus/hfsplus_raw.h 	__be32 prev;
prev             1948 fs/io_uring.c  	struct io_kiocb *req, *prev;
prev             1962 fs/io_uring.c  	prev = req;
prev             1963 fs/io_uring.c  	list_for_each_entry_continue_reverse(prev, &ctx->timeout_list, list)
prev             1964 fs/io_uring.c  		prev->sequence++;
prev             2005 fs/io_uring.c  		entry = ctx->timeout_list.prev;
prev             2577 fs/io_uring.c  		struct io_kiocb *prev = *link;
prev             2587 fs/io_uring.c  		list_add_tail(&req->list, &prev->link_list);
prev              388 fs/iomap/buffered-io.c 	loff_t pos = page_offset(list_entry(pages->prev, struct page, lru));
prev               12 fs/iomap/fiemap.c 	struct iomap prev;
prev               55 fs/iomap/fiemap.c 	ret = iomap_to_fiemap(ctx->fi, &ctx->prev, 0);
prev               56 fs/iomap/fiemap.c 	ctx->prev = *iomap;
prev               75 fs/iomap/fiemap.c 	ctx.prev.type = IOMAP_HOLE;
prev              102 fs/iomap/fiemap.c 	if (ctx.prev.type != IOMAP_HOLE) {
prev              103 fs/iomap/fiemap.c 		ret = iomap_to_fiemap(fi, &ctx.prev, FIEMAP_EXTENT_LAST);
prev              806 fs/jbd2/commit.c 		struct buffer_head *bh = list_entry(io_bufs.prev,
prev              852 fs/jbd2/commit.c 		bh = list_entry(log_bufs.prev, struct buffer_head, b_assoc_buffers);
prev              322 fs/jffs2/compr.c 			list_add(&comp->list, this->list.prev);
prev              212 fs/jffs2/erase.c 	struct jffs2_raw_node_ref **prev;
prev              214 fs/jffs2/erase.c 	prev = &ref->next_in_ino;
prev              218 fs/jffs2/erase.c 		if (!(*prev)->next_in_ino) {
prev              222 fs/jffs2/erase.c 			ic = (struct jffs2_inode_cache *)(*prev);
prev              223 fs/jffs2/erase.c 			prev = &ic->nodes;
prev              227 fs/jffs2/erase.c 		if (SECTOR_ADDR((*prev)->flash_offset) == jeb->offset) {
prev              231 fs/jffs2/erase.c 			this = *prev;
prev              232 fs/jffs2/erase.c 			*prev = this->next_in_ino;
prev              241 fs/jffs2/erase.c 		prev = &((*prev)->next_in_ino);
prev               28 fs/jffs2/nodelist.c 	struct jffs2_full_dirent **prev = list;
prev               32 fs/jffs2/nodelist.c 	while ((*prev) && (*prev)->nhash <= new->nhash) {
prev               33 fs/jffs2/nodelist.c 		if ((*prev)->nhash == new->nhash && !strcmp((*prev)->name, new->name)) {
prev               35 fs/jffs2/nodelist.c 			if (new->version < (*prev)->version) {
prev               37 fs/jffs2/nodelist.c 					(*prev)->name, (*prev)->ino);
prev               42 fs/jffs2/nodelist.c 					(*prev)->name, (*prev)->ino);
prev               43 fs/jffs2/nodelist.c 				new->next = (*prev)->next;
prev               46 fs/jffs2/nodelist.c 				if ((*prev)->raw)
prev               47 fs/jffs2/nodelist.c 					jffs2_mark_node_obsolete(c, ((*prev)->raw));
prev               48 fs/jffs2/nodelist.c 				jffs2_free_full_dirent(*prev);
prev               49 fs/jffs2/nodelist.c 				*prev = new;
prev               53 fs/jffs2/nodelist.c 		prev = &((*prev)->next);
prev               55 fs/jffs2/nodelist.c 	new->next = *prev;
prev               56 fs/jffs2/nodelist.c 	*prev = new;
prev              386 fs/jffs2/nodelist.c 		struct jffs2_node_frag *prev = frag_prev(newfrag);
prev              390 fs/jffs2/nodelist.c 		if (prev->node)
prev              391 fs/jffs2/nodelist.c 			mark_ref_normal(prev->node->raw);
prev              438 fs/jffs2/nodelist.c 	struct jffs2_inode_cache **prev;
prev              446 fs/jffs2/nodelist.c 	prev = &c->inocache_list[new->ino % c->inocache_hashsize];
prev              448 fs/jffs2/nodelist.c 	while ((*prev) && (*prev)->ino < new->ino) {
prev              449 fs/jffs2/nodelist.c 		prev = &(*prev)->next;
prev              451 fs/jffs2/nodelist.c 	new->next = *prev;
prev              452 fs/jffs2/nodelist.c 	*prev = new;
prev              459 fs/jffs2/nodelist.c 	struct jffs2_inode_cache **prev;
prev              467 fs/jffs2/nodelist.c 	prev = &c->inocache_list[old->ino % c->inocache_hashsize];
prev              469 fs/jffs2/nodelist.c 	while ((*prev) && (*prev)->ino < old->ino) {
prev              470 fs/jffs2/nodelist.c 		prev = &(*prev)->next;
prev              472 fs/jffs2/nodelist.c 	if ((*prev) == old) {
prev              473 fs/jffs2/nodelist.c 		*prev = old->next;
prev              529 fs/jffs2/nodelist.c 	struct jffs2_node_frag *prev = NULL;
prev              541 fs/jffs2/nodelist.c 			if (!prev || frag->ofs > prev->ofs)
prev              542 fs/jffs2/nodelist.c 				prev = frag;
prev              554 fs/jffs2/nodelist.c 	if (prev)
prev              556 fs/jffs2/nodelist.c 			  prev->ofs, prev->ofs+prev->size);
prev              560 fs/jffs2/nodelist.c 	return prev;
prev             1172 fs/jfs/jfs_dtree.c 			    sp->header.prev != 0 || skip > 1) {
prev             1400 fs/jfs/jfs_dtree.c 	rp->header.prev = cpu_to_le64(addressPXD(&sp->header.self));
prev             1484 fs/jfs/jfs_dtree.c 		p->header.prev = cpu_to_le64(rbn);
prev             1928 fs/jfs/jfs_dtree.c 	rp->header.prev = 0;
prev             2410 fs/jfs/jfs_dtree.c 		    ((p->header.flag & BT_ROOT) || p->header.prev == 0))
prev             2496 fs/jfs/jfs_dtree.c 	if (p->header.prev) {
prev             2497 fs/jfs/jfs_dtree.c 		prevbn = le64_to_cpu(p->header.prev);
prev             2537 fs/jfs/jfs_dtree.c 		rp->header.prev = cpu_to_le64(nxaddr);
prev             2745 fs/jfs/jfs_dtree.c 	prevbn = le64_to_cpu(p->header.prev);
prev             2772 fs/jfs/jfs_dtree.c 		p->header.prev = cpu_to_le64(prevbn);
prev              183 fs/jfs/jfs_dtree.h 		__le64 prev;	/* 8: previous sibling */
prev              355 fs/jfs/jfs_extent.c 	s64 prev;
prev              367 fs/jfs/jfs_extent.c 	prev = ((offset & ~POFFSET) >> JFS_SBI(sb)->l2bsize) - nbperpage;
prev              371 fs/jfs/jfs_extent.c 	if (prev < 0)
prev              374 fs/jfs/jfs_extent.c 	rc = xtLookup(ip, prev, nbperpage, &xflag, &xaddr, &xlen, 0);
prev              383 fs/jfs/jfs_extent.c 		XADoffset(xp, prev);
prev              584 fs/jfs/jfs_logmgr.c 		tblk = list_entry(log->cqueue.prev, struct tblock, cqueue);
prev             1534 fs/jfs/jfs_logmgr.c 		target = list_entry(log->cqueue.prev, struct tblock, cqueue);
prev             1025 fs/jfs/jfs_xtree.c 	rp->header.prev = cpu_to_le64(addressPXD(&sp->header.self));
prev             1091 fs/jfs/jfs_xtree.c 		p->header.prev = cpu_to_le64(rbn);
prev             1267 fs/jfs/jfs_xtree.c 	rp->header.prev = 0;
prev             2727 fs/jfs/jfs_xtree.c 		if (p->header.prev) {
prev             2728 fs/jfs/jfs_xtree.c 			prevbn = le64_to_cpu(p->header.prev);
prev             2754 fs/jfs/jfs_xtree.c 			rp->header.prev = cpu_to_le64(nxaddr);
prev             3006 fs/jfs/jfs_xtree.c 	prevbn = le64_to_cpu(p->header.prev);
prev             3024 fs/jfs/jfs_xtree.c 		p->header.prev = cpu_to_le64(prevbn);
prev             3535 fs/jfs/jfs_xtree.c 		struct tlock *prev;
prev             3542 fs/jfs/jfs_xtree.c 			for (prev = lid_to_tlock(tblk->next);
prev             3543 fs/jfs/jfs_xtree.c 			     prev->next != lid;
prev             3544 fs/jfs/jfs_xtree.c 			     prev = lid_to_tlock(prev->next)) {
prev             3545 fs/jfs/jfs_xtree.c 				assert(prev->next);
prev             3547 fs/jfs/jfs_xtree.c 			prev->next = tlck->next;
prev               74 fs/jfs/jfs_xtree.h 		__le64 prev;	/* 8: */
prev              889 fs/namespace.c 	list_splice(&head, n->list.prev);
prev              916 fs/namespace.c 	struct list_head *prev = p->mnt_mounts.prev;
prev              917 fs/namespace.c 	while (prev != &p->mnt_mounts) {
prev              918 fs/namespace.c 		p = list_entry(prev, struct mount, mnt_child);
prev              919 fs/namespace.c 		prev = p->mnt_mounts.prev;
prev              878 fs/nfs/blocklayout/blocklayout.c bl_pg_test_read(struct nfs_pageio_descriptor *pgio, struct nfs_page *prev,
prev              883 fs/nfs/blocklayout/blocklayout.c 	return pnfs_generic_pg_test(pgio, prev, req);
prev              941 fs/nfs/blocklayout/blocklayout.c bl_pg_test_write(struct nfs_pageio_descriptor *pgio, struct nfs_page *prev,
prev              946 fs/nfs/blocklayout/blocklayout.c 	return pnfs_generic_pg_test(pgio, prev, req);
prev              502 fs/nfs/delegation.c 	struct nfs_delegation *prev;
prev              521 fs/nfs/delegation.c 	prev = NULL;
prev              538 fs/nfs/delegation.c 				prev = delegation;
prev              544 fs/nfs/delegation.c 			if (prev) {
prev              547 fs/nfs/delegation.c 				tmp = nfs_delegation_grab_inode(prev);
prev              551 fs/nfs/delegation.c 					place_holder_deleg = prev;
prev             2374 fs/nfs/dir.c   	lh = rcu_dereference(nfsi->access_cache_entry_lru.prev);
prev              847 fs/nfs/filelayout/filelayout.c filelayout_pg_test(struct nfs_pageio_descriptor *pgio, struct nfs_page *prev,
prev              857 fs/nfs/filelayout/filelayout.c 	size = pnfs_generic_pg_test(pgio, prev, req);
prev              862 fs/nfs/filelayout/filelayout.c 	if (prev) {
prev              863 fs/nfs/filelayout/filelayout.c 		p_stripe = (u64)req_offset(prev) - segment_offset;
prev              493 fs/nfs/nfs4client.c 			     struct nfs_client **prev, struct nfs_net *nn)
prev              511 fs/nfs/nfs4client.c 		nfs_put_client(*prev);
prev              512 fs/nfs/nfs4client.c 		*prev = pos;
prev              554 fs/nfs/nfs4client.c 	struct nfs_client *pos, *prev = NULL;
prev              567 fs/nfs/nfs4client.c 		status = nfs4_match_client(pos, new, &prev, nn);
prev              591 fs/nfs/nfs4client.c 		nfs_put_client(prev);
prev              592 fs/nfs/nfs4client.c 		prev = pos;
prev              603 fs/nfs/nfs4client.c 			prev = NULL;
prev              623 fs/nfs/nfs4client.c 	nfs_put_client(prev);
prev              720 fs/nfs/nfs4client.c 	struct nfs_client *pos, *prev = NULL;
prev              729 fs/nfs/nfs4client.c 		status = nfs4_match_client(pos, new, &prev, nn);
prev              753 fs/nfs/nfs4client.c 	nfs_put_client(prev);
prev             2827 fs/nfs/nfs4proc.c 	struct nfs4_lock_state *lsp, *prev = NULL;
prev             2841 fs/nfs/nfs4proc.c 			nfs4_put_lock_state(prev);
prev             2842 fs/nfs/nfs4proc.c 			prev = lsp;
prev             2856 fs/nfs/nfs4proc.c 				nfs4_put_lock_state(prev);
prev             2863 fs/nfs/nfs4proc.c 	nfs4_put_lock_state(prev);
prev              234 fs/nfs/pagelist.c nfs_page_group_init(struct nfs_page *req, struct nfs_page *prev)
prev              237 fs/nfs/pagelist.c 	WARN_ON_ONCE(prev == req);
prev              239 fs/nfs/pagelist.c 	if (!prev) {
prev              245 fs/nfs/pagelist.c 		WARN_ON_ONCE(prev->wb_this_page != prev->wb_head);
prev              246 fs/nfs/pagelist.c 		WARN_ON_ONCE(!test_bit(PG_HEADLOCK, &prev->wb_head->wb_flags));
prev              247 fs/nfs/pagelist.c 		req->wb_head = prev->wb_head;
prev              248 fs/nfs/pagelist.c 		req->wb_this_page = prev->wb_this_page;
prev              249 fs/nfs/pagelist.c 		prev->wb_this_page = req;
prev              258 fs/nfs/pagelist.c 		if (test_bit(PG_INODE_REF, &prev->wb_head->wb_flags)) {
prev              493 fs/nfs/pagelist.c 			   struct nfs_page *prev, struct nfs_page *req)
prev              916 fs/nfs/pagelist.c static bool nfs_can_coalesce_requests(struct nfs_page *prev,
prev              923 fs/nfs/pagelist.c 	if (prev) {
prev              924 fs/nfs/pagelist.c 		if (!nfs_match_open_context(nfs_req_openctx(req), nfs_req_openctx(prev)))
prev              931 fs/nfs/pagelist.c 					    prev->wb_lock_context))
prev              933 fs/nfs/pagelist.c 		if (req_offset(req) != req_offset(prev) + prev->wb_bytes)
prev              935 fs/nfs/pagelist.c 		if (req->wb_page == prev->wb_page) {
prev              936 fs/nfs/pagelist.c 			if (req->wb_pgbase != prev->wb_pgbase + prev->wb_bytes)
prev              940 fs/nfs/pagelist.c 			    prev->wb_pgbase + prev->wb_bytes != PAGE_SIZE)
prev              944 fs/nfs/pagelist.c 	size = pgio->pg_ops->pg_test(pgio, prev, req);
prev              964 fs/nfs/pagelist.c 	struct nfs_page *prev = NULL;
prev              967 fs/nfs/pagelist.c 		prev = nfs_list_entry(mirror->pg_list.prev);
prev              984 fs/nfs/pagelist.c 	if (!nfs_can_coalesce_requests(prev, req, desc))
prev             1299 fs/nfs/pagelist.c 	struct nfs_page *prev;
prev             1305 fs/nfs/pagelist.c 			prev = nfs_list_entry(mirror->pg_list.prev);
prev             1306 fs/nfs/pagelist.c 			if (index != prev->wb_index + 1) {
prev             2566 fs/nfs/pnfs.c  		     struct nfs_page *prev, struct nfs_page *req)
prev             2571 fs/nfs/pnfs.c  	size = nfs_generic_pg_test(pgio, prev, req);
prev              252 fs/nfs/pnfs.h  			    struct nfs_page *prev, struct nfs_page *req);
prev              202 fs/nilfs2/alloc.c 				  struct nilfs_bh_assoc *prev,
prev              208 fs/nilfs2/alloc.c 	if (prev->bh && blkoff == prev->blkoff) {
prev              209 fs/nilfs2/alloc.c 		get_bh(prev->bh);
prev              210 fs/nilfs2/alloc.c 		*bhp = prev->bh;
prev              223 fs/nilfs2/alloc.c 		brelse(prev->bh);
prev              225 fs/nilfs2/alloc.c 		prev->bh = *bhp;
prev              226 fs/nilfs2/alloc.c 		prev->blkoff = blkoff;
prev              240 fs/nilfs2/alloc.c 				     struct nilfs_bh_assoc *prev,
prev              244 fs/nilfs2/alloc.c 	if (prev->bh && blkoff == prev->blkoff) {
prev              245 fs/nilfs2/alloc.c 		brelse(prev->bh);
prev              246 fs/nilfs2/alloc.c 		prev->bh = NULL;
prev              623 fs/nilfs2/cpfile.c 	__u64 curr, prev;
prev              659 fs/nilfs2/cpfile.c 	prev = le64_to_cpu(list->ssl_prev);
prev              660 fs/nilfs2/cpfile.c 	while (prev > cno) {
prev              661 fs/nilfs2/cpfile.c 		prev_blkoff = nilfs_cpfile_get_blkoff(cpfile, prev);
prev              662 fs/nilfs2/cpfile.c 		curr = prev;
prev              676 fs/nilfs2/cpfile.c 		prev = le64_to_cpu(list->ssl_prev);
prev              680 fs/nilfs2/cpfile.c 	if (prev != 0) {
prev              681 fs/nilfs2/cpfile.c 		ret = nilfs_cpfile_get_checkpoint_block(cpfile, prev, 0,
prev              699 fs/nilfs2/cpfile.c 	cp->cp_snapshot_list.ssl_prev = cpu_to_le64(prev);
prev              705 fs/nilfs2/cpfile.c 		cpfile, prev, prev_bh, kaddr);
prev              742 fs/nilfs2/cpfile.c 	__u64 next, prev;
prev              768 fs/nilfs2/cpfile.c 	prev = le64_to_cpu(list->ssl_prev);
prev              783 fs/nilfs2/cpfile.c 	if (prev != 0) {
prev              784 fs/nilfs2/cpfile.c 		ret = nilfs_cpfile_get_checkpoint_block(cpfile, prev, 0,
prev              796 fs/nilfs2/cpfile.c 	list->ssl_prev = cpu_to_le64(prev);
prev              801 fs/nilfs2/cpfile.c 		cpfile, prev, prev_bh, kaddr);
prev               78 fs/nilfs2/segbuf.c 			   struct nilfs_segment_buffer *prev)
prev               80 fs/nilfs2/segbuf.c 	segbuf->sb_segnum = prev->sb_segnum;
prev               81 fs/nilfs2/segbuf.c 	segbuf->sb_fseg_start = prev->sb_fseg_start;
prev               82 fs/nilfs2/segbuf.c 	segbuf->sb_fseg_end = prev->sb_fseg_end;
prev               83 fs/nilfs2/segbuf.c 	segbuf->sb_pseg_start = prev->sb_pseg_start + prev->sb_sum.nblocks;
prev               88 fs/nilfs2/segbuf.h #define NILFS_PREV_SEGBUF(segbuf)  NILFS_LIST_SEGBUF((segbuf)->sb_list.prev)
prev               89 fs/nilfs2/segbuf.h #define NILFS_LAST_SEGBUF(head)    NILFS_LIST_SEGBUF((head)->prev)
prev              111 fs/nilfs2/segbuf.h 			   struct nilfs_segment_buffer *prev);
prev             1153 fs/nilfs2/segment.c 						ii->i_dirty.prev,
prev             1174 fs/nilfs2/segment.c 					list_entry(ii->i_dirty.prev,
prev             1276 fs/nilfs2/segment.c 	struct nilfs_segment_buffer *segbuf, *prev;
prev             1300 fs/nilfs2/segment.c 		prev = NILFS_LAST_SEGBUF(&sci->sc_write_logs);
prev             1301 fs/nilfs2/segment.c 		nilfs_segbuf_map_cont(segbuf, prev);
prev             1302 fs/nilfs2/segment.c 		segbuf->sb_sum.seg_seq = prev->sb_sum.seg_seq;
prev             1303 fs/nilfs2/segment.c 		nextnum = prev->sb_nextnum;
prev             1306 fs/nilfs2/segment.c 			nilfs_segbuf_map(segbuf, prev->sb_nextnum, 0, nilfs);
prev             1336 fs/nilfs2/segment.c 	struct nilfs_segment_buffer *segbuf, *prev;
prev             1342 fs/nilfs2/segment.c 	prev = NILFS_LAST_SEGBUF(&sci->sc_segbufs);
prev             1349 fs/nilfs2/segment.c 	err = nilfs_sufile_mark_dirty(sufile, prev->sb_nextnum);
prev             1361 fs/nilfs2/segment.c 		nilfs_segbuf_map(segbuf, prev->sb_nextnum, 0, nilfs);
prev             1369 fs/nilfs2/segment.c 		segbuf->sb_sum.seg_seq = prev->sb_sum.seg_seq + 1;
prev             1373 fs/nilfs2/segment.c 		prev = segbuf;
prev             1392 fs/nilfs2/segment.c 	struct nilfs_segment_buffer *segbuf, *prev;
prev             1414 fs/nilfs2/segment.c 	prev = segbuf;
prev             1416 fs/nilfs2/segment.c 		if (prev->sb_nextnum != segbuf->sb_nextnum) {
prev             1424 fs/nilfs2/segment.c 		prev = segbuf;
prev               82 fs/notify/dnotify/dnotify.c 	struct dnotify_struct **prev;
prev               96 fs/notify/dnotify/dnotify.c 	prev = &dn_mark->dn;
prev               97 fs/notify/dnotify/dnotify.c 	while ((dn = *prev) != NULL) {
prev               99 fs/notify/dnotify/dnotify.c 			prev = &dn->dn_next;
prev              105 fs/notify/dnotify/dnotify.c 			prev = &dn->dn_next;
prev              107 fs/notify/dnotify/dnotify.c 			*prev = dn->dn_next;
prev              146 fs/notify/dnotify/dnotify.c 	struct dnotify_struct **prev;
prev              162 fs/notify/dnotify/dnotify.c 	prev = &dn_mark->dn;
prev              163 fs/notify/dnotify/dnotify.c 	while ((dn = *prev) != NULL) {
prev              165 fs/notify/dnotify/dnotify.c 			*prev = dn->dn_next;
prev              170 fs/notify/dnotify/dnotify.c 		prev = &dn->dn_next;
prev               54 fs/notify/inotify/inotify_fsnotify.c 	last_event = list_entry(list->prev, struct fsnotify_event, list);
prev             1526 fs/ocfs2/dir.c 	struct ocfs2_dir_block_trailer *trailer, *prev;
prev             1538 fs/ocfs2/dir.c 		prev = ocfs2_trailer_from_bh(bh, dir->i_sb);
prev             1539 fs/ocfs2/dir.c 		prev->db_free_next = trailer->db_free_next;
prev              264 fs/ocfs2/extent_map.c 		emi = list_entry(em->em_list.prev,
prev             1075 fs/ocfs2/quota_local.c 	chunk->qc_num = list_entry(chunk->qc_chunk.prev,
prev             1111 fs/ocfs2/quota_local.c 	chunk = list_entry(oinfo->dqi_chunk.prev,
prev             1359 fs/ocfs2/refcounttree.c static int ocfs2_refcount_rec_no_intersect(struct ocfs2_refcount_rec *prev,
prev             1362 fs/ocfs2/refcounttree.c 	if (ocfs2_get_ref_rec_low_cpos(prev) + le32_to_cpu(prev->r_clusters) <=
prev              469 fs/ocfs2/reservations.c 	struct rb_node *prev, *next;
prev              561 fs/ocfs2/reservations.c 	prev = &prev_resv->r_node;
prev              565 fs/ocfs2/reservations.c 		next = rb_next(prev);
prev              609 fs/ocfs2/reservations.c 		prev = next;
prev              610 fs/ocfs2/reservations.c 		prev_resv = rb_entry(prev, struct ocfs2_alloc_reservation,
prev              165 fs/omfs/dir.c  	u64 block, prev;
prev              177 fs/omfs/dir.c  	bh2 = omfs_scan_list(dir, block, name, namelen, &prev);
prev              187 fs/omfs/dir.c  	if (prev != ~0) {
prev              190 fs/omfs/dir.c  		bh = omfs_bread(dir->i_sb, prev);
prev              201 fs/omfs/dir.c  	if (prev != ~0) {
prev              202 fs/omfs/dir.c  		dirty = omfs_iget(dir->i_sb, prev);
prev              653 fs/orangefs/devorangefs-req.c 			if (!orangefs_sb->list.prev)
prev              625 fs/orangefs/super.c 		ORANGEFS_SB(sb)->list.prev = NULL;
prev               29 fs/pnode.c     	return list_entry(p->mnt_slave_list.prev, struct mount, mnt_slave);
prev              113 fs/pnode.c     	list_splice(&mnt->mnt_slave_list, master->mnt_slave_list.prev);
prev              221 fs/readdir.c   	struct linux_dirent __user *dirent, *prev;
prev              244 fs/readdir.c   	prev = (void __user *) dirent - prev_reclen;
prev              245 fs/readdir.c   	if (!user_access_begin(prev, reclen + prev_reclen))
prev              249 fs/readdir.c   	unsafe_put_user(offset, &prev->d_off, efault_end);
prev              312 fs/readdir.c   	struct linux_dirent64 __user *dirent, *prev;
prev              329 fs/readdir.c   	prev = (void __user *)dirent - prev_reclen;
prev              330 fs/readdir.c   	if (!user_access_begin(prev, reclen + prev_reclen))
prev              334 fs/readdir.c   	unsafe_put_user(offset, &prev->d_off, efault_end);
prev              264 fs/reiserfs/item_ops.c 	__u32 prev = INT_MAX;
prev              274 fs/reiserfs/item_ops.c 		if (sequence_finished(prev, &num, get_block_num(unp, j))) {
prev              275 fs/reiserfs/item_ops.c 			print_sequence(prev, num);
prev              276 fs/reiserfs/item_ops.c 			start_new_sequence(&prev, &num, get_block_num(unp, j));
prev              279 fs/reiserfs/item_ops.c 	print_sequence(prev, num);
prev              358 fs/reiserfs/journal.c 	head[0].prev = NULL;
prev              361 fs/reiserfs/journal.c 		head[i].prev = head + (i - 1);
prev              386 fs/reiserfs/journal.c 		cn->next->prev = NULL;
prev              408 fs/reiserfs/journal.c 		journal->j_cnode_free_list->prev = cn;
prev              410 fs/reiserfs/journal.c 	cn->prev = NULL;	/* not needed with the memset, but I might kill the memset, and forget to do this */
prev              851 fs/reiserfs/journal.c 		jh = JH_ENTRY(tmp.prev);
prev              900 fs/reiserfs/journal.c 	entry = jl->j_list.prev;
prev              908 fs/reiserfs/journal.c 		entry = other_jl->j_list.prev;
prev             3361 fs/reiserfs/journal.c 	cn->prev = journal->j_last;
prev             3428 fs/reiserfs/journal.c 	if (cn->prev) {
prev             3429 fs/reiserfs/journal.c 		cn->prev->next = cn->next;
prev             3432 fs/reiserfs/journal.c 		cn->next->prev = cn->prev;
prev             3438 fs/reiserfs/journal.c 		journal->j_last = cn->prev;
prev             3539 fs/reiserfs/journal.c 		entry = journal->j_journal_list.prev;
prev             4126 fs/reiserfs/journal.c 			jl_cn->prev = last_cn;
prev             1400 fs/reiserfs/lbalance.c 		int prev, next;
prev             1408 fs/reiserfs/lbalance.c 			prev = (i != 0) ? deh_location(&deh[i - 1]) : 0;
prev             1410 fs/reiserfs/lbalance.c 			if (prev && prev <= deh_location(&deh[i]))
prev              239 fs/reiserfs/reiserfs.h 	struct reiserfs_journal_cnode *prev;	/* prev in transaction list */
prev              113 fs/squashfs/decompressor_multi.c 			decomp_strm = list_entry(stream->strm_list.prev,
prev              137 fs/squashfs/decompressor_multi.c 			decomp_strm = list_entry(stream->strm_list.prev,
prev              664 fs/ubifs/commit.c 				i = list_entry(list.prev, struct idx_node,
prev              708 fs/ubifs/commit.c 		i = list_entry(list.prev, struct idx_node, list);
prev             1482 fs/ubifs/debug.c 		struct ubifs_znode *prev;
prev             1501 fs/ubifs/debug.c 		prev = znode;
prev             1510 fs/ubifs/debug.c 		last = prev->child_cnt - 1;
prev             1511 fs/ubifs/debug.c 		if (prev->level == 0 && znode->level == 0 && !c->replaying &&
prev             1512 fs/ubifs/debug.c 		    !keys_cmp(c, &prev->zbranch[last].key,
prev             1514 fs/ubifs/debug.c 			err = dbg_check_key_order(c, &prev->zbranch[last],
prev             1520 fs/ubifs/debug.c 				ubifs_dump_znode(c, prev);
prev              106 fs/ubifs/master.c 		snod = list_entry(sleb->nodes.prev, struct ubifs_scan_node,
prev              124 fs/ubifs/master.c 	snod = list_entry(sleb->nodes.prev, struct ubifs_scan_node, list);
prev              500 fs/ubifs/recovery.c 		snod = list_entry(sleb->nodes.prev,
prev              568 fs/ubifs/recovery.c 		snod = list_entry(sleb->nodes.prev, struct ubifs_scan_node,
prev              596 fs/ubifs/recovery.c 		snod = list_entry(sleb->nodes.prev, struct ubifs_scan_node,
prev              879 fs/userfaultfd.c 	struct vm_area_struct *vma, *prev;
prev              900 fs/userfaultfd.c 	prev = NULL;
prev              906 fs/userfaultfd.c 			prev = vma;
prev              911 fs/userfaultfd.c 			prev = vma_merge(mm, prev, vma->vm_start, vma->vm_end,
prev              916 fs/userfaultfd.c 			if (prev)
prev              917 fs/userfaultfd.c 				vma = prev;
prev              919 fs/userfaultfd.c 				prev = vma;
prev             1306 fs/userfaultfd.c 	struct vm_area_struct *vma, *prev, *cur;
prev             1356 fs/userfaultfd.c 	vma = find_vma_prev(mm, start, &prev);
prev             1440 fs/userfaultfd.c 		prev = vma;
prev             1464 fs/userfaultfd.c 		prev = vma_merge(mm, prev, start, vma_end, new_flags,
prev             1468 fs/userfaultfd.c 		if (prev) {
prev             1469 fs/userfaultfd.c 			vma = prev;
prev             1492 fs/userfaultfd.c 		prev = vma;
prev             1518 fs/userfaultfd.c 	struct vm_area_struct *vma, *prev, *cur;
prev             1545 fs/userfaultfd.c 	vma = find_vma_prev(mm, start, &prev);
prev             1591 fs/userfaultfd.c 		prev = vma;
prev             1626 fs/userfaultfd.c 		prev = vma_merge(mm, prev, start, vma_end, new_flags,
prev             1630 fs/userfaultfd.c 		if (prev) {
prev             1631 fs/userfaultfd.c 			vma = prev;
prev             1654 fs/userfaultfd.c 		prev = vma;
prev             3068 fs/xfs/libxfs/xfs_bmap.c 	if (ap->eof && ap->prev.br_startoff != NULLFILEOFF &&
prev             3069 fs/xfs/libxfs/xfs_bmap.c 	    !isnullstartblock(ap->prev.br_startblock) &&
prev             3070 fs/xfs/libxfs/xfs_bmap.c 	    ISVALID(ap->prev.br_startblock + ap->prev.br_blockcount,
prev             3071 fs/xfs/libxfs/xfs_bmap.c 		    ap->prev.br_startblock)) {
prev             3072 fs/xfs/libxfs/xfs_bmap.c 		ap->blkno = ap->prev.br_startblock + ap->prev.br_blockcount;
prev             3077 fs/xfs/libxfs/xfs_bmap.c 			(ap->prev.br_startoff + ap->prev.br_blockcount);
prev             3079 fs/xfs/libxfs/xfs_bmap.c 		    ISVALID(ap->blkno + adjust, ap->prev.br_startblock))
prev             3097 fs/xfs/libxfs/xfs_bmap.c 		if (ap->prev.br_startoff != NULLFILEOFF &&
prev             3098 fs/xfs/libxfs/xfs_bmap.c 		    !isnullstartblock(ap->prev.br_startblock) &&
prev             3099 fs/xfs/libxfs/xfs_bmap.c 		    (prevbno = ap->prev.br_startblock +
prev             3100 fs/xfs/libxfs/xfs_bmap.c 			       ap->prev.br_blockcount) &&
prev             3101 fs/xfs/libxfs/xfs_bmap.c 		    ISVALID(prevbno, ap->prev.br_startblock)) {
prev             3106 fs/xfs/libxfs/xfs_bmap.c 				(ap->prev.br_startoff +
prev             3107 fs/xfs/libxfs/xfs_bmap.c 				 ap->prev.br_blockcount);
prev             3118 fs/xfs/libxfs/xfs_bmap.c 				    ap->prev.br_startblock))
prev             3417 fs/xfs/libxfs/xfs_bmap.c 		error = xfs_bmap_extsize_align(mp, &ap->got, &ap->prev,
prev             3944 fs/xfs/libxfs/xfs_bmap.c 		struct xfs_bmbt_irec	prev;
prev             3947 fs/xfs/libxfs/xfs_bmap.c 		if (!xfs_iext_peek_prev_extent(ifp, icur, &prev))
prev             3948 fs/xfs/libxfs/xfs_bmap.c 			prev.br_startoff = NULLFILEOFF;
prev             3950 fs/xfs/libxfs/xfs_bmap.c 		error = xfs_bmap_extsize_align(mp, got, &prev, extsz, 0, eof,
prev             4031 fs/xfs/libxfs/xfs_bmap.c 		xfs_iext_peek_prev_extent(ifp, &bma->icur, &bma->prev);
prev             4338 fs/xfs/libxfs/xfs_bmap.c 	if (!xfs_iext_peek_prev_extent(ifp, &bma.icur, &bma.prev))
prev             4339 fs/xfs/libxfs/xfs_bmap.c 		bma.prev.br_startoff = NULLFILEOFF;
prev             4426 fs/xfs/libxfs/xfs_bmap.c 		bma.prev = bma.got;
prev             4513 fs/xfs/libxfs/xfs_bmap.c 	if (!xfs_iext_peek_prev_extent(ifp, &bma.icur, &bma.prev))
prev             4514 fs/xfs/libxfs/xfs_bmap.c 		bma.prev.br_startoff = NULLFILEOFF;
prev             5405 fs/xfs/libxfs/xfs_bmap.c 				struct xfs_bmbt_irec	prev;
prev             5413 fs/xfs/libxfs/xfs_bmap.c 				if (!xfs_iext_prev_extent(ifp, &icur, &prev))
prev             5415 fs/xfs/libxfs/xfs_bmap.c 				ASSERT(prev.br_state == XFS_EXT_NORM);
prev             5416 fs/xfs/libxfs/xfs_bmap.c 				ASSERT(!isnullstartblock(prev.br_startblock));
prev             5418 fs/xfs/libxfs/xfs_bmap.c 				       prev.br_startblock + prev.br_blockcount);
prev             5419 fs/xfs/libxfs/xfs_bmap.c 				if (prev.br_startoff < start) {
prev             5420 fs/xfs/libxfs/xfs_bmap.c 					mod = start - prev.br_startoff;
prev             5421 fs/xfs/libxfs/xfs_bmap.c 					prev.br_blockcount -= mod;
prev             5422 fs/xfs/libxfs/xfs_bmap.c 					prev.br_startblock += mod;
prev             5423 fs/xfs/libxfs/xfs_bmap.c 					prev.br_startoff = start;
prev             5425 fs/xfs/libxfs/xfs_bmap.c 				prev.br_state = XFS_EXT_UNWRITTEN;
prev             5428 fs/xfs/libxfs/xfs_bmap.c 						&prev, &logflags);
prev             5661 fs/xfs/libxfs/xfs_bmap.c 	struct xfs_bmbt_irec	prev = *got;
prev             5669 fs/xfs/libxfs/xfs_bmap.c 		error = xfs_bmbt_lookup_eq(cur, &prev, &i);
prev             5685 fs/xfs/libxfs/xfs_bmap.c 	xfs_rmap_unmap_extent(tp, ip, whichfork, &prev);
prev             5702 fs/xfs/libxfs/xfs_bmap.c 	struct xfs_bmbt_irec	got, prev;
prev             5740 fs/xfs/libxfs/xfs_bmap.c 	if (xfs_iext_peek_prev_extent(ifp, &icur, &prev)) {
prev             5741 fs/xfs/libxfs/xfs_bmap.c 		if (new_startoff < prev.br_startoff + prev.br_blockcount) {
prev             5746 fs/xfs/libxfs/xfs_bmap.c 		if (xfs_bmse_can_merge(&prev, &got, offset_shift_fsb)) {
prev             5748 fs/xfs/libxfs/xfs_bmap.c 					offset_shift_fsb, &icur, &got, &prev,
prev               24 fs/xfs/libxfs/xfs_bmap.h 	struct xfs_bmbt_irec	prev;	/* extent before the new one */
prev              124 fs/xfs/libxfs/xfs_iext_tree.c 	struct xfs_iext_leaf	*prev;
prev              268 fs/xfs/libxfs/xfs_iext_tree.c 	if (ifp->if_height > 1 && cur->leaf->prev) {
prev              269 fs/xfs/libxfs/xfs_iext_tree.c 		cur->leaf = cur->leaf->prev;
prev              405 fs/xfs/libxfs/xfs_iext_tree.c 		struct xfs_iext_leaf *prev = ifp->if_u1.if_root;
prev              407 fs/xfs/libxfs/xfs_iext_tree.c 		node->keys[0] = xfs_iext_leaf_key(prev, 0);
prev              408 fs/xfs/libxfs/xfs_iext_tree.c 		node->ptrs[0] = prev;
prev              410 fs/xfs/libxfs/xfs_iext_tree.c 		struct xfs_iext_node *prev = ifp->if_u1.if_root;
prev              414 fs/xfs/libxfs/xfs_iext_tree.c 		node->keys[0] = prev->keys[0];
prev              415 fs/xfs/libxfs/xfs_iext_tree.c 		node->ptrs[0] = prev;
prev              572 fs/xfs/libxfs/xfs_iext_tree.c 		leaf->next->prev = new;
prev              574 fs/xfs/libxfs/xfs_iext_tree.c 	new->prev = leaf;
prev              687 fs/xfs/libxfs/xfs_iext_tree.c 		struct xfs_iext_node *prev = parent->ptrs[*pos - 1];
prev              688 fs/xfs/libxfs/xfs_iext_tree.c 		int nr_prev = xfs_iext_node_nr_entries(prev, 0), i;
prev              692 fs/xfs/libxfs/xfs_iext_tree.c 				prev->keys[nr_prev + i] = node->keys[i];
prev              693 fs/xfs/libxfs/xfs_iext_tree.c 				prev->ptrs[nr_prev + i] = node->ptrs[i];
prev              803 fs/xfs/libxfs/xfs_iext_tree.c 	if (leaf->prev) {
prev              804 fs/xfs/libxfs/xfs_iext_tree.c 		int nr_prev = xfs_iext_leaf_nr_entries(ifp, leaf->prev, 0), i;
prev              808 fs/xfs/libxfs/xfs_iext_tree.c 				leaf->prev->recs[nr_prev + i] = leaf->recs[i];
prev              811 fs/xfs/libxfs/xfs_iext_tree.c 				cur->leaf = leaf->prev;
prev              845 fs/xfs/libxfs/xfs_iext_tree.c 	if (leaf->prev)
prev              846 fs/xfs/libxfs/xfs_iext_tree.c 		leaf->prev->next = leaf->next;
prev              848 fs/xfs/libxfs/xfs_iext_tree.c 		leaf->next->prev = leaf->prev;
prev              734 fs/xfs/xfs_aops.c 	struct bio		*prev)
prev              739 fs/xfs/xfs_aops.c 	bio_copy_dev(new, prev);/* also copies over blkcg information */
prev              740 fs/xfs/xfs_aops.c 	new->bi_iter.bi_sector = bio_end_sector(prev);
prev              741 fs/xfs/xfs_aops.c 	new->bi_opf = prev->bi_opf;
prev              742 fs/xfs/xfs_aops.c 	new->bi_write_hint = prev->bi_write_hint;
prev              744 fs/xfs/xfs_aops.c 	bio_chain(prev, new);
prev              745 fs/xfs/xfs_aops.c 	bio_get(prev);		/* for xfs_destroy_ioend */
prev              746 fs/xfs/xfs_aops.c 	submit_bio(prev);
prev               40 fs/xfs/xfs_bio_io.c 			struct bio	*prev = bio;
prev               43 fs/xfs/xfs_bio_io.c 			bio_copy_dev(bio, prev);
prev               44 fs/xfs/xfs_bio_io.c 			bio->bi_iter.bi_sector = bio_end_sector(prev);
prev               45 fs/xfs/xfs_bio_io.c 			bio->bi_opf = prev->bi_opf;
prev               46 fs/xfs/xfs_bio_io.c 			bio_chain(prev, bio);
prev               48 fs/xfs/xfs_bio_io.c 			submit_bio(prev);
prev               86 fs/xfs/xfs_bmap_util.c 	error = xfs_bmap_extsize_align(mp, &ap->got, &ap->prev,
prev              403 fs/xfs/xfs_iomap.c 	struct xfs_bmbt_irec	prev;
prev              423 fs/xfs/xfs_iomap.c 	    !xfs_iext_peek_prev_extent(ifp, icur, &prev) ||
prev              424 fs/xfs/xfs_iomap.c 	    prev.br_startoff + prev.br_blockcount < offset_fsb)
prev              443 fs/xfs/xfs_iomap.c 	if (prev.br_blockcount <= (MAXEXTLEN >> 1))
prev              444 fs/xfs/xfs_iomap.c 		alloc_blocks = prev.br_blockcount << 1;
prev             4201 fs/xfs/xfs_log_recover.c 	item = list_entry(trans->r_itemq.prev, xlog_recover_item_t, ri_list);
prev             4271 fs/xfs/xfs_log_recover.c 	item = list_entry(trans->r_itemq.prev, xlog_recover_item_t, ri_list);
prev             4276 fs/xfs/xfs_log_recover.c 		item = list_entry(trans->r_itemq.prev,
prev              147 fs/xfs/xfs_mru_cache.c 			list_splice_init(lru_list, mru->reap_list.prev);
prev               51 fs/xfs/xfs_trans_ail.c 	prev_lip = list_entry(lip->li_ail.prev, struct xfs_log_item, li_ail);
prev               85 fs/xfs/xfs_trans_ail.c 	return list_entry(ailp->ail_head.prev, struct xfs_log_item, li_ail);
prev              312 fs/xfs/xfs_trans_ail.c 		cur->item = list_entry(list->prev, struct xfs_log_item, li_ail);
prev               18 include/asm-generic/cmpxchg-local.h 	unsigned long flags, prev;
prev               28 include/asm-generic/cmpxchg-local.h 	case 1: prev = *(u8 *)ptr;
prev               29 include/asm-generic/cmpxchg-local.h 		if (prev == old)
prev               32 include/asm-generic/cmpxchg-local.h 	case 2: prev = *(u16 *)ptr;
prev               33 include/asm-generic/cmpxchg-local.h 		if (prev == old)
prev               36 include/asm-generic/cmpxchg-local.h 	case 4: prev = *(u32 *)ptr;
prev               37 include/asm-generic/cmpxchg-local.h 		if (prev == old)
prev               40 include/asm-generic/cmpxchg-local.h 	case 8: prev = *(u64 *)ptr;
prev               41 include/asm-generic/cmpxchg-local.h 		if (prev == old)
prev               48 include/asm-generic/cmpxchg-local.h 	return prev;
prev               57 include/asm-generic/cmpxchg-local.h 	u64 prev;
prev               61 include/asm-generic/cmpxchg-local.h 	prev = *(u64 *)ptr;
prev               62 include/asm-generic/cmpxchg-local.h 	if (prev == old)
prev               65 include/asm-generic/cmpxchg-local.h 	return prev;
prev               35 include/asm-generic/mmu_context.h static inline void switch_mm(struct mm_struct *prev,
prev              712 include/asm-generic/pgtable.h #define arch_start_context_switch(prev)	do {} while (0)
prev               21 include/asm-generic/switch_to.h #define switch_to(prev, next, last)					\
prev               23 include/asm-generic/switch_to.h 		((last) = __switch_to((prev), (next)));			\
prev              133 include/crypto/drbg.h 	unsigned char *prev;	/* FIPS 140-2 continuous test value */
prev             1201 include/linux/acpi.h 			     struct fwnode_handle *prev)
prev               71 include/linux/agp_backend.h 	struct agp_memory *prev;
prev               93 include/linux/agpgart.h 	struct agp_client *prev;
prev              101 include/linux/agpgart.h 	struct agp_controller *prev;
prev              116 include/linux/agpgart.h 	struct agp_file_private *prev;
prev               29 include/linux/dma-fence-chain.h 	struct dma_fence __rcu *prev;
prev               69 include/linux/dma-fence-chain.h 			  struct dma_fence *prev,
prev             1512 include/linux/efi.h 			struct efivar_entry **prev);
prev               98 include/linux/fwnode.h 				   struct fwnode_handle *prev);
prev              154 include/linux/interval_tree_generic.h 	struct rb_node *rb = node->ITRB.rb_right, *prev;		      \
prev              176 include/linux/interval_tree_generic.h 			prev = &node->ITRB;				      \
prev              179 include/linux/interval_tree_generic.h 		} while (prev == rb);					      \
prev               29 include/linux/list.h 	list->prev = list;
prev               34 include/linux/list.h 			      struct list_head *prev,
prev               39 include/linux/list.h 				struct list_head *prev,
prev               57 include/linux/list.h 			      struct list_head *prev,
prev               60 include/linux/list.h 	if (!__list_add_valid(new, prev, next))
prev               63 include/linux/list.h 	next->prev = new;
prev               65 include/linux/list.h 	new->prev = prev;
prev               66 include/linux/list.h 	WRITE_ONCE(prev->next, new);
prev               93 include/linux/list.h 	__list_add(new, head->prev, head);
prev              103 include/linux/list.h static inline void __list_del(struct list_head * prev, struct list_head * next)
prev              105 include/linux/list.h 	next->prev = prev;
prev              106 include/linux/list.h 	WRITE_ONCE(prev->next, next);
prev              119 include/linux/list.h 	__list_del(entry->prev, entry->next);
prev              120 include/linux/list.h 	entry->prev = NULL;
prev              134 include/linux/list.h 	__list_del(entry->prev, entry->next);
prev              141 include/linux/list.h 	entry->prev = LIST_POISON2;
prev              155 include/linux/list.h 	new->next->prev = new;
prev              156 include/linux/list.h 	new->prev = old->prev;
prev              157 include/linux/list.h 	new->prev->next = new;
prev              175 include/linux/list.h 	struct list_head *pos = entry2->prev;
prev              230 include/linux/list.h 	first->prev->next = last->next;
prev              231 include/linux/list.h 	last->next->prev = first->prev;
prev              233 include/linux/list.h 	head->prev->next = first;
prev              234 include/linux/list.h 	first->prev = head->prev;
prev              237 include/linux/list.h 	head->prev = last;
prev              248 include/linux/list.h 	return list->prev == head;
prev              287 include/linux/list.h 	return (next == head) && (next == head->prev);
prev              328 include/linux/list.h 	return !list_empty(head) && (head->next == head->prev);
prev              336 include/linux/list.h 	list->next->prev = list;
prev              337 include/linux/list.h 	list->prev = entry;
prev              340 include/linux/list.h 	new_first->prev = head;
prev              394 include/linux/list.h 	list->next->prev = list;
prev              395 include/linux/list.h 	list->prev = entry->prev;
prev              396 include/linux/list.h 	list->prev->next = list;
prev              398 include/linux/list.h 	entry->prev = head;
prev              402 include/linux/list.h 				 struct list_head *prev,
prev              406 include/linux/list.h 	struct list_head *last = list->prev;
prev              408 include/linux/list.h 	first->prev = prev;
prev              409 include/linux/list.h 	prev->next = first;
prev              412 include/linux/list.h 	next->prev = last;
prev              436 include/linux/list.h 		__list_splice(list, head->prev, head);
prev              467 include/linux/list.h 		__list_splice(list, head->prev, head);
prev              501 include/linux/list.h 	list_entry((ptr)->prev, type, member)
prev              531 include/linux/list.h 	list_entry((pos)->member.prev, typeof(*(pos)), member)
prev              547 include/linux/list.h 	for (pos = (head)->prev; pos != (head); pos = pos->prev)
prev              566 include/linux/list.h 	for (pos = (head)->prev, n = pos->prev; \
prev              568 include/linux/list.h 	     pos = n, n = pos->prev)
prev              803 include/linux/list.h 				    struct hlist_node *prev)
prev              805 include/linux/list.h 	n->next = prev->next;
prev              806 include/linux/list.h 	prev->next = n;
prev              807 include/linux/list.h 	n->pprev = &prev->next;
prev              105 include/linux/list_bl.h 				       struct hlist_bl_node *prev)
prev              107 include/linux/list_bl.h 	n->next = prev->next;
prev              108 include/linux/list_bl.h 	n->pprev = &prev->next;
prev              109 include/linux/list_bl.h 	prev->next = n;
prev              939 include/linux/memcontrol.h 		struct mem_cgroup *prev,
prev              946 include/linux/memcontrol.h 					 struct mem_cgroup *prev)
prev              435 include/linux/mlx5/driver.h 	u32				prev;
prev              221 include/linux/mm.h #define lru_to_page(head) (list_entry((head)->prev, struct page, lru))
prev             2245 include/linux/mm.h 				    struct vm_area_struct *prev,
prev             2286 include/linux/mm.h 	struct vm_area_struct *prev, unsigned long addr, unsigned long end,
prev              137 include/linux/nfs_page.h 				struct nfs_page *prev,
prev              226 include/linux/of.h extern struct device_node *__of_find_all_nodes(struct device_node *prev);
prev              227 include/linux/of.h extern struct device_node *of_find_all_nodes(struct device_node *prev);
prev              289 include/linux/of.h 					     struct device_node *prev);
prev              291 include/linux/of.h 	const struct device_node *node, struct device_node *prev);
prev              353 include/linux/of.h extern struct device_node *of_get_next_cpu_node(struct device_node *prev);
prev              632 include/linux/of.h 	const struct device_node *node, struct device_node *prev)
prev              638 include/linux/of.h 	const struct device_node *node, struct device_node *prev)
prev              763 include/linux/of.h static inline struct device_node *of_get_next_cpu_node(struct device_node *prev)
prev               18 include/linux/of_pdt.h 	int (*nextprop)(phandle node, char *prev, char *buf);
prev               10 include/linux/osq_lock.h 	struct optimistic_spin_node *next, *prev;
prev              148 include/linux/parport.h 	struct pardevice *prev;
prev              894 include/linux/perf_event.h extern void __perf_event_task_sched_in(struct task_struct *prev,
prev              896 include/linux/perf_event.h extern void __perf_event_task_sched_out(struct task_struct *prev,
prev             1146 include/linux/perf_event.h static inline void perf_event_task_sched_in(struct task_struct *prev,
prev             1150 include/linux/perf_event.h 		__perf_event_task_sched_in(prev, task);
prev             1161 include/linux/perf_event.h static inline void perf_event_task_sched_out(struct task_struct *prev,
prev             1167 include/linux/perf_event.h 		__perf_event_task_sched_out(prev, next);
prev             1354 include/linux/perf_event.h perf_event_task_sched_in(struct task_struct *prev,
prev             1357 include/linux/perf_event.h perf_event_task_sched_out(struct task_struct *prev,
prev              294 include/linux/plist.h 	return list_entry(head->node_list.prev,
prev              340 include/linux/property.h 	const struct fwnode_handle *fwnode, struct fwnode_handle *prev);
prev               34 include/linux/rculist.h 	WRITE_ONCE(list->prev, list);
prev               68 include/linux/rculist.h 		struct list_head *prev, struct list_head *next)
prev               70 include/linux/rculist.h 	if (!__list_add_valid(new, prev, next))
prev               74 include/linux/rculist.h 	new->prev = prev;
prev               75 include/linux/rculist.h 	rcu_assign_pointer(list_next_rcu(prev), new);
prev               76 include/linux/rculist.h 	next->prev = new;
prev              119 include/linux/rculist.h 	__list_add_rcu(new, head->prev, head);
prev              149 include/linux/rculist.h 	entry->prev = LIST_POISON2;
prev              192 include/linux/rculist.h 	new->prev = old->prev;
prev              193 include/linux/rculist.h 	rcu_assign_pointer(list_next_rcu(new->prev), new);
prev              194 include/linux/rculist.h 	new->next->prev = new;
prev              195 include/linux/rculist.h 	old->prev = LIST_POISON2;
prev              218 include/linux/rculist.h 					  struct list_head *prev,
prev              223 include/linux/rculist.h 	struct list_head *last = list->prev;
prev              251 include/linux/rculist.h 	rcu_assign_pointer(list_next_rcu(prev), first);
prev              252 include/linux/rculist.h 	first->prev = prev;
prev              253 include/linux/rculist.h 	next->prev = last;
prev              283 include/linux/rculist.h 		__list_splice_init_rcu(list, head->prev, head, sync);
prev              620 include/linux/rculist.h 					struct hlist_node *prev)
prev              622 include/linux/rculist.h 	n->next = prev->next;
prev              623 include/linux/rculist.h 	n->pprev = &prev->next;
prev              624 include/linux/rculist.h 	rcu_assign_pointer(hlist_next_rcu(prev), n);
prev              202 include/linux/rio.h 	struct rio_dev *prev;
prev               57 include/linux/sched/cputime.h extern void cputime_adjust(struct task_cputime *curr, struct prev_cputime *prev,
prev              180 include/linux/sched/cputime.h static inline void prev_cputime_init(struct prev_cputime *prev)
prev              183 include/linux/sched/cputime.h 	prev->utime = prev->stime = 0;
prev              184 include/linux/sched/cputime.h 	raw_spin_lock_init(&prev->lock);
prev              379 include/linux/sched/mm.h static inline void membarrier_arch_switch_mm(struct mm_struct *prev,
prev               47 include/linux/sched/task.h extern asmlinkage void schedule_tail(struct task_struct *prev);
prev              292 include/linux/skbuff.h 	struct sk_buff	*prev;
prev              690 include/linux/skbuff.h 			struct sk_buff		*prev;
prev             1536 include/linux/skbuff.h 	return skb->prev == (const struct sk_buff *) list;
prev             1572 include/linux/skbuff.h 	return skb->prev;
prev             1800 include/linux/skbuff.h 	struct sk_buff *skb = READ_ONCE(list_->prev);
prev             1831 include/linux/skbuff.h 	list->prev = list->next = (struct sk_buff *)list;
prev             1863 include/linux/skbuff.h 				struct sk_buff *prev, struct sk_buff *next,
prev             1870 include/linux/skbuff.h 	WRITE_ONCE(newsk->prev, prev);
prev             1871 include/linux/skbuff.h 	WRITE_ONCE(next->prev, newsk);
prev             1872 include/linux/skbuff.h 	WRITE_ONCE(prev->next, newsk);
prev             1877 include/linux/skbuff.h 				      struct sk_buff *prev,
prev             1881 include/linux/skbuff.h 	struct sk_buff *last = list->prev;
prev             1883 include/linux/skbuff.h 	WRITE_ONCE(first->prev, prev);
prev             1884 include/linux/skbuff.h 	WRITE_ONCE(prev->next, first);
prev             1887 include/linux/skbuff.h 	WRITE_ONCE(next->prev, last);
prev             1930 include/linux/skbuff.h 		__skb_queue_splice(list, head->prev, (struct sk_buff *) head);
prev             1947 include/linux/skbuff.h 		__skb_queue_splice(list, head->prev, (struct sk_buff *) head);
prev             1965 include/linux/skbuff.h 				     struct sk_buff *prev,
prev             1968 include/linux/skbuff.h 	__skb_insert(newsk, prev, prev->next, list);
prev             1978 include/linux/skbuff.h 	__skb_insert(newsk, next->prev, next, list);
prev             2022 include/linux/skbuff.h 	struct sk_buff *next, *prev;
prev             2026 include/linux/skbuff.h 	prev	   = skb->prev;
prev             2027 include/linux/skbuff.h 	skb->next  = skb->prev = NULL;
prev             2028 include/linux/skbuff.h 	WRITE_ONCE(next->prev, prev);
prev             2029 include/linux/skbuff.h 	WRITE_ONCE(prev->next, next);
prev             3436 include/linux/skbuff.h 		for (skb = (queue)->prev;					\
prev             3438 include/linux/skbuff.h 		     skb = skb->prev)
prev             3441 include/linux/skbuff.h 		for (skb = (queue)->prev, tmp = skb->prev;			\
prev             3443 include/linux/skbuff.h 		     skb = tmp, tmp = skb->prev)
prev             3446 include/linux/skbuff.h 		for (tmp = skb->prev;						\
prev             3448 include/linux/skbuff.h 		     skb = tmp, tmp = skb->prev)
prev              160 include/linux/thunderbolt.h 					 struct tb_property *prev);
prev              182 include/linux/types.h 	struct list_head *next, *prev;
prev               19 include/linux/user-return-notifier.h static inline void propagate_user_return_notify(struct task_struct *prev,
prev               22 include/linux/user-return-notifier.h 	if (test_tsk_thread_flag(prev, TIF_USER_RETURN_NOTIFY)) {
prev               23 include/linux/user-return-notifier.h 		clear_tsk_thread_flag(prev, TIF_USER_RETURN_NOTIFY);
prev               39 include/linux/user-return-notifier.h static inline void propagate_user_return_notify(struct task_struct *prev,
prev               50 include/linux/vtime.h extern void vtime_task_switch(struct task_struct *prev);
prev               52 include/linux/vtime.h extern void vtime_common_task_switch(struct task_struct *prev);
prev               53 include/linux/vtime.h static inline void vtime_task_switch(struct task_struct *prev)
prev               56 include/linux/vtime.h 		vtime_common_task_switch(prev);
prev               65 include/linux/vtime.h static inline void vtime_task_switch(struct task_struct *prev) { }
prev               73 include/net/bond_alb.h 	u32 prev;		/* The previous Hash table entry index,
prev               64 include/net/bonding.h 		netdev_adjacent_get_private(bond_slave_list(bond)->prev) : \
prev             1026 include/net/sch_generic.h 	if (skb->prev)
prev             1027 include/net/sch_generic.h 		skb->prev->next = *to_free;
prev              415 include/net/sctp/sctp.h 	return (head->next != head) && (head->next == head->prev);
prev              184 include/sound/gus.h 	struct snd_gf1_mem_block *prev;
prev               49 include/sound/util_mem.h 					      struct list_head *prev);
prev              831 include/trace/events/rxrpc.h 		     rxrpc_seq_t first, rxrpc_seq_t prev, u8 reason, u8 n_acks),
prev              833 include/trace/events/rxrpc.h 	    TP_ARGS(call, serial, ack_serial, first, prev, reason, n_acks),
prev              840 include/trace/events/rxrpc.h 		    __field(rxrpc_seq_t,		prev		)
prev              850 include/trace/events/rxrpc.h 		    __entry->prev = prev;
prev              861 include/trace/events/rxrpc.h 		      __entry->prev,
prev              141 include/trace/events/sched.h 		 struct task_struct *prev,
prev              144 include/trace/events/sched.h 	TP_ARGS(preempt, prev, next),
prev              158 include/trace/events/sched.h 		__entry->prev_pid	= prev->pid;
prev              159 include/trace/events/sched.h 		__entry->prev_prio	= prev->prio;
prev              160 include/trace/events/sched.h 		__entry->prev_state	= __trace_sched_switch_state(preempt, prev);
prev              161 include/trace/events/sched.h 		memcpy(__entry->prev_comm, prev->comm, TASK_COMM_LEN);
prev              117 include/uapi/drm/drm.h 	unsigned char prev;
prev              151 include/uapi/drm/i810_drm.h 	unsigned char next, prev;	/* indices to form a circular LRU  */
prev              180 include/uapi/drm/via_drm.h 	unsigned char next, prev;	/* indices to form a circular LRU  */
prev               19 include/uapi/linux/netfilter/xt_limit.h 	unsigned long prev; /* moved to xt_limit_priv */
prev               20 include/uapi/linux/netfilter_bridge/ebt_limit.h 	unsigned long prev;
prev              816 include/video/omapfb_dss.h 			 struct device_node *prev);
prev              820 include/video/omapfb_dss.h 			     struct device_node *prev);
prev              689 ipc/mqueue.c   	ptr = info->e_wait_q[sr].list.prev;
prev              937 kernel/audit_tree.c 	while (barrier.prev != &tree_list) {
prev              940 kernel/audit_tree.c 		tree = container_of(barrier.prev, struct audit_tree, list);
prev               73 kernel/bpf/bpf_lru_list.c 		l->next_inactive_rotation = l->next_inactive_rotation->prev;
prev              119 kernel/bpf/bpf_lru_list.c 		l->next_inactive_rotation = l->next_inactive_rotation->prev;
prev              184 kernel/bpf/bpf_lru_list.c 			cur = cur->prev;
prev              189 kernel/bpf/bpf_lru_list.c 		next = cur->prev;
prev              643 kernel/bpf/core.c 	       fp->aux->ksym_lnode.prev == LIST_POISON2;
prev              662 kernel/bpf/cpumap.c 	if (!bq->flush_node.prev)
prev              455 kernel/bpf/devmap.c 	if (!bq->flush_node.prev)
prev             7543 kernel/bpf/verifier.c static bool reg_type_mismatch(enum bpf_reg_type src, enum bpf_reg_type prev)
prev             7545 kernel/bpf/verifier.c 	return src != prev && (!reg_type_mismatch_ok(src) ||
prev             7546 kernel/bpf/verifier.c 			       !reg_type_mismatch_ok(prev));
prev              186 kernel/bpf/xskmap.c 	if (!xs->flush_node.prev)
prev              445 kernel/crash_core.c 	VMCOREINFO_OFFSET(list_head, prev);
prev              748 kernel/debug/kdb/kdb_support.c 	struct debug_alloc_header *best, *bestprev, *prev, *h;
prev              760 kernel/debug/kdb/kdb_support.c 	prev = best = bestprev = NULL;
prev              764 kernel/debug/kdb/kdb_support.c 			bestprev = prev;
prev              770 kernel/debug/kdb/kdb_support.c 		prev = h;
prev              828 kernel/debug/kdb/kdb_support.c 		struct debug_alloc_header *prev;
prev              830 kernel/debug/kdb/kdb_support.c 		prev = (struct debug_alloc_header *)(debug_alloc_pool +
prev              833 kernel/debug/kdb/kdb_support.c 			if (!prev->next || prev->next > h_offset)
prev              835 kernel/debug/kdb/kdb_support.c 			prev = (struct debug_alloc_header *)
prev              836 kernel/debug/kdb/kdb_support.c 				(debug_alloc_pool + prev->next);
prev              838 kernel/debug/kdb/kdb_support.c 		prev_offset = (char *)prev - debug_alloc_pool;
prev              839 kernel/debug/kdb/kdb_support.c 		if (prev_offset + dah_overhead + prev->size == h_offset) {
prev              840 kernel/debug/kdb/kdb_support.c 			prev->size += dah_overhead + h->size;
prev              843 kernel/debug/kdb/kdb_support.c 			h = prev;
prev              846 kernel/debug/kdb/kdb_support.c 			h->next = prev->next;
prev              847 kernel/debug/kdb/kdb_support.c 			prev->next = h_offset;
prev              873 kernel/events/core.c static inline void perf_cgroup_sched_in(struct task_struct *prev,
prev              886 kernel/events/core.c 	cgrp2 = perf_cgroup_from_task(prev, NULL);
prev             1021 kernel/events/core.c static inline void perf_cgroup_sched_in(struct task_struct *prev,
prev             3275 kernel/events/core.c static void perf_pmu_sched_task(struct task_struct *prev,
prev             3282 kernel/events/core.c 	if (prev == next)
prev             3563 kernel/events/core.c void __perf_event_task_sched_in(struct task_struct *prev,
prev             3577 kernel/events/core.c 		perf_cgroup_sched_in(prev, task);
prev             3588 kernel/events/core.c 		perf_event_switch(task, prev, true);
prev             3591 kernel/events/core.c 		perf_pmu_sched_task(prev, task, true);
prev             9671 kernel/events/core.c 	s64 prev;
prev             9675 kernel/events/core.c 	prev = local64_xchg(&event->hw.prev_count, now);
prev             9676 kernel/events/core.c 	local64_add(now - prev, &event->count);
prev             9748 kernel/events/core.c 	u64 prev;
prev             9751 kernel/events/core.c 	prev = local64_xchg(&event->hw.prev_count, now);
prev             9752 kernel/events/core.c 	delta = now - prev;
prev              977 kernel/events/uprobes.c 	struct map_info *prev = NULL;
prev              987 kernel/events/uprobes.c 		if (!prev && !more) {
prev              992 kernel/events/uprobes.c 			prev = kmalloc(sizeof(struct map_info),
prev              994 kernel/events/uprobes.c 			if (prev)
prev              995 kernel/events/uprobes.c 				prev->next = NULL;
prev              997 kernel/events/uprobes.c 		if (!prev) {
prev             1005 kernel/events/uprobes.c 		info = prev;
prev             1006 kernel/events/uprobes.c 		prev = prev->next;
prev             1018 kernel/events/uprobes.c 	prev = curr;
prev             1030 kernel/events/uprobes.c 		info->next = prev;
prev             1031 kernel/events/uprobes.c 		prev = info;
prev             1036 kernel/events/uprobes.c 	while (prev)
prev             1037 kernel/events/uprobes.c 		prev = free_map_info(prev);
prev              481 kernel/fork.c  	struct vm_area_struct *mpnt, *tmp, *prev, **pprev;
prev              517 kernel/fork.c  	prev = NULL;
prev              592 kernel/fork.c  		tmp->vm_prev = prev;
prev              593 kernel/fork.c  		prev = tmp;
prev               59 kernel/gcov/base.c 	struct gcov_info *prev = NULL;
prev               68 kernel/gcov/base.c 			gcov_info_unlink(prev, info);
prev               72 kernel/gcov/base.c 			prev = info;
prev              206 kernel/gcov/clang.c void gcov_info_unlink(struct gcov_info *prev, struct gcov_info *info)
prev              132 kernel/gcov/gcc_3_4.c void gcov_info_unlink(struct gcov_info *prev, struct gcov_info *info)
prev              134 kernel/gcov/gcc_3_4.c 	if (prev)
prev              135 kernel/gcov/gcc_3_4.c 		prev->next = info->next;
prev              145 kernel/gcov/gcc_4_7.c void gcov_info_unlink(struct gcov_info *prev, struct gcov_info *info)
prev              147 kernel/gcov/gcc_4_7.c 	if (prev)
prev              148 kernel/gcov/gcc_4_7.c 		prev->next = info->next;
prev               49 kernel/gcov/gcov.h void gcov_info_unlink(struct gcov_info *prev, struct gcov_info *info);
prev              674 kernel/jump_label.c 	struct static_key_mod *jlm, **prev;
prev              689 kernel/jump_label.c 		prev = &key->next;
prev              693 kernel/jump_label.c 			prev = &jlm->next;
prev              701 kernel/jump_label.c 		if (prev == &key->next)
prev              704 kernel/jump_label.c 			*prev = jlm->next;
prev              223 kernel/livepatch/transition.c 				struct klp_func *prev;
prev              225 kernel/livepatch/transition.c 				prev = list_next_entry(func, stack_node);
prev              226 kernel/livepatch/transition.c 				func_addr = (unsigned long)prev->new_func;
prev              227 kernel/livepatch/transition.c 				func_size = prev->new_size;
prev             2033 kernel/locking/lockdep.c 			 struct held_lock *prev,
prev             2057 kernel/locking/lockdep.c 	print_lock(prev);
prev             2059 kernel/locking/lockdep.c 	print_lock_name(hlock_class(prev));
prev             2080 kernel/locking/lockdep.c 				hlock_class(prev), hlock_class(next));
prev             2227 kernel/locking/lockdep.c static int check_irq_usage(struct task_struct *curr, struct held_lock *prev,
prev             2242 kernel/locking/lockdep.c 	this.class = hlock_class(prev);
prev             2298 kernel/locking/lockdep.c 				 prev, next,
prev             2320 kernel/locking/lockdep.c 				  struct held_lock *prev, struct held_lock *next)
prev             2336 kernel/locking/lockdep.c 	struct lock_class *prev = hlock_class(prv);
prev             2342 kernel/locking/lockdep.c 	__print_lock_name(prev);
prev             2352 kernel/locking/lockdep.c print_deadlock_bug(struct task_struct *curr, struct held_lock *prev,
prev             2367 kernel/locking/lockdep.c 	print_lock(prev);
prev             2370 kernel/locking/lockdep.c 	print_deadlock_scenario(next, prev);
prev             2388 kernel/locking/lockdep.c 	struct held_lock *prev;
prev             2393 kernel/locking/lockdep.c 		prev = curr->held_locks + i;
prev             2395 kernel/locking/lockdep.c 		if (prev->instance == next->nest_lock)
prev             2396 kernel/locking/lockdep.c 			nest = prev;
prev             2398 kernel/locking/lockdep.c 		if (hlock_class(prev) != hlock_class(next))
prev             2405 kernel/locking/lockdep.c 		if ((next->read == 2) && prev->read)
prev             2415 kernel/locking/lockdep.c 		print_deadlock_bug(curr, prev, next);
prev             2444 kernel/locking/lockdep.c check_prev_add(struct task_struct *curr, struct held_lock *prev,
prev             2451 kernel/locking/lockdep.c 	if (!hlock_class(prev)->key || !hlock_class(next)->key) {
prev             2458 kernel/locking/lockdep.c 		WARN_ONCE(!debug_locks_silent && !hlock_class(prev)->key,
prev             2460 kernel/locking/lockdep.c 			  hlock_class(prev),
prev             2461 kernel/locking/lockdep.c 			  hlock_class(prev)->name);
prev             2479 kernel/locking/lockdep.c 	ret = check_noncircular(next, prev, trace);
prev             2483 kernel/locking/lockdep.c 	if (!check_irq_usage(curr, prev, next))
prev             2494 kernel/locking/lockdep.c 	if (next->read == 2 || prev->read == 2)
prev             2504 kernel/locking/lockdep.c 	list_for_each_entry(entry, &hlock_class(prev)->locks_after, entry) {
prev             2516 kernel/locking/lockdep.c 	ret = check_redundant(prev, next);
prev             2531 kernel/locking/lockdep.c 	ret = add_lock_to_list(hlock_class(next), hlock_class(prev),
prev             2532 kernel/locking/lockdep.c 			       &hlock_class(prev)->locks_after,
prev             2538 kernel/locking/lockdep.c 	ret = add_lock_to_list(hlock_class(prev), hlock_class(next),
prev               67 kernel/locking/mcs_spinlock.h 	struct mcs_spinlock *prev;
prev               79 kernel/locking/mcs_spinlock.h 	prev = xchg(lock, node);
prev               80 kernel/locking/mcs_spinlock.h 	if (likely(prev == NULL)) {
prev               91 kernel/locking/mcs_spinlock.h 	WRITE_ONCE(prev->next, node);
prev               76 kernel/locking/mutex-debug.c 		DEBUG_LOCKS_WARN_ON(!lock->wait_list.prev && !lock->wait_list.next);
prev               14 kernel/locking/mutex.h 		__list_del((waiter)->list.prev, (waiter)->list.next)
prev               44 kernel/locking/osq_lock.c 	      struct optimistic_spin_node *prev)
prev               55 kernel/locking/osq_lock.c 	old = prev ? prev->cpu : OSQ_UNLOCKED_VAL;
prev               93 kernel/locking/osq_lock.c 	struct optimistic_spin_node *prev, *next;
prev              111 kernel/locking/osq_lock.c 	prev = decode_cpu(old);
prev              112 kernel/locking/osq_lock.c 	node->prev = prev;
prev              126 kernel/locking/osq_lock.c 	WRITE_ONCE(prev->next, node);
prev              143 kernel/locking/osq_lock.c 		if (need_resched() || vcpu_is_preempted(node_cpu(node->prev)))
prev              160 kernel/locking/osq_lock.c 		if (prev->next == node &&
prev              161 kernel/locking/osq_lock.c 		    cmpxchg(&prev->next, node, NULL) == node)
prev              178 kernel/locking/osq_lock.c 		prev = READ_ONCE(node->prev);
prev              188 kernel/locking/osq_lock.c 	next = osq_wait_next(lock, node, prev);
prev              200 kernel/locking/osq_lock.c 	WRITE_ONCE(next->prev, prev);
prev              201 kernel/locking/osq_lock.c 	WRITE_ONCE(prev->next, next);
prev              273 kernel/locking/qspinlock.c 					   struct mcs_spinlock *prev) { }
prev              316 kernel/locking/qspinlock.c 	struct mcs_spinlock *prev, *next, *node;
prev              467 kernel/locking/qspinlock.c 		prev = decode_tail(old);
prev              470 kernel/locking/qspinlock.c 		WRITE_ONCE(prev->next, node);
prev              472 kernel/locking/qspinlock.c 		pv_wait_node(node, prev);
prev              267 kernel/locking/qspinlock_paravirt.h pv_wait_early(struct pv_node *prev, int loop)
prev              272 kernel/locking/qspinlock_paravirt.h 	return READ_ONCE(prev->state) != vcpu_running;
prev              293 kernel/locking/qspinlock_paravirt.h static void pv_wait_node(struct mcs_spinlock *node, struct mcs_spinlock *prev)
prev              296 kernel/locking/qspinlock_paravirt.h 	struct pv_node *pp = (struct pv_node *)prev;
prev             1398 kernel/module.c 			      char *prev);
prev             2542 kernel/module.c 			      char *prev)
prev             2555 kernel/module.c 	if (prev) {
prev             2556 kernel/module.c 		size -= prev - modinfo;
prev             2557 kernel/module.c 		modinfo = next_string(prev, &size);
prev              958 kernel/power/snapshot.c 		region = list_entry(nosave_regions.prev,
prev              688 kernel/rcu/tree_exp.h 	t = list_entry(rnp->exp_tasks->prev,
prev              182 kernel/rcu/tree_stall.h 	t = list_entry(rnp->gp_tasks->prev,
prev              208 kernel/rcu/tree_stall.h 	t = list_entry(rnp->gp_tasks->prev,
prev              138 kernel/sched/autogroup.c 	struct autogroup *prev;
prev              144 kernel/sched/autogroup.c 	prev = p->signal->autogroup;
prev              145 kernel/sched/autogroup.c 	if (prev == ag) {
prev              166 kernel/sched/autogroup.c 	autogroup_kref_put(prev);
prev             3077 kernel/sched/core.c static inline void finish_task(struct task_struct *prev)
prev             3090 kernel/sched/core.c 	smp_store_release(&prev->on_cpu, 0);
prev             3148 kernel/sched/core.c prepare_task_switch(struct rq *rq, struct task_struct *prev,
prev             3151 kernel/sched/core.c 	kcov_prepare_switch(prev);
prev             3152 kernel/sched/core.c 	sched_info_switch(rq, prev, next);
prev             3153 kernel/sched/core.c 	perf_event_task_sched_out(prev, next);
prev             3154 kernel/sched/core.c 	rseq_preempt(prev);
prev             3155 kernel/sched/core.c 	fire_sched_out_preempt_notifiers(prev, next);
prev             3179 kernel/sched/core.c static struct rq *finish_task_switch(struct task_struct *prev)
prev             3215 kernel/sched/core.c 	prev_state = prev->state;
prev             3216 kernel/sched/core.c 	vtime_task_switch(prev);
prev             3217 kernel/sched/core.c 	perf_event_task_sched_in(prev, current);
prev             3218 kernel/sched/core.c 	finish_task(prev);
prev             3241 kernel/sched/core.c 		if (prev->sched_class->task_dead)
prev             3242 kernel/sched/core.c 			prev->sched_class->task_dead(prev);
prev             3248 kernel/sched/core.c 		kprobe_flush_task(prev);
prev             3251 kernel/sched/core.c 		put_task_stack(prev);
prev             3253 kernel/sched/core.c 		put_task_struct_rcu_user(prev);
prev             3301 kernel/sched/core.c asmlinkage __visible void schedule_tail(struct task_struct *prev)
prev             3315 kernel/sched/core.c 	rq = finish_task_switch(prev);
prev             3329 kernel/sched/core.c context_switch(struct rq *rq, struct task_struct *prev,
prev             3332 kernel/sched/core.c 	prepare_task_switch(rq, prev, next);
prev             3339 kernel/sched/core.c 	arch_start_context_switch(prev);
prev             3349 kernel/sched/core.c 		enter_lazy_tlb(prev->active_mm, next);
prev             3351 kernel/sched/core.c 		next->active_mm = prev->active_mm;
prev             3352 kernel/sched/core.c 		if (prev->mm)                           // from user
prev             3353 kernel/sched/core.c 			mmgrab(prev->active_mm);
prev             3355 kernel/sched/core.c 			prev->active_mm = NULL;
prev             3357 kernel/sched/core.c 		membarrier_switch_mm(rq, prev->active_mm, next->mm);
prev             3366 kernel/sched/core.c 		switch_mm_irqs_off(prev->active_mm, next->mm, next);
prev             3368 kernel/sched/core.c 		if (!prev->mm) {                        // from kernel
prev             3370 kernel/sched/core.c 			rq->prev_mm = prev->active_mm;
prev             3371 kernel/sched/core.c 			prev->active_mm = NULL;
prev             3380 kernel/sched/core.c 	switch_to(prev, next, prev);
prev             3383 kernel/sched/core.c 	return finish_task_switch(prev);
prev             3842 kernel/sched/core.c static noinline void __schedule_bug(struct task_struct *prev)
prev             3851 kernel/sched/core.c 		prev->comm, prev->pid, preempt_count());
prev             3853 kernel/sched/core.c 	debug_show_held_locks(prev);
prev             3856 kernel/sched/core.c 		print_irqtrace_events(prev);
prev             3873 kernel/sched/core.c static inline void schedule_debug(struct task_struct *prev, bool preempt)
prev             3876 kernel/sched/core.c 	if (task_stack_end_corrupted(prev))
prev             3881 kernel/sched/core.c 	if (!preempt && prev->state && prev->non_block_count) {
prev             3883 kernel/sched/core.c 			prev->comm, prev->pid, prev->non_block_count);
prev             3890 kernel/sched/core.c 		__schedule_bug(prev);
prev             3904 kernel/sched/core.c pick_next_task(struct rq *rq, struct task_struct *prev, struct rq_flags *rf)
prev             3915 kernel/sched/core.c 	if (likely((prev->sched_class == &idle_sched_class ||
prev             3916 kernel/sched/core.c 		    prev->sched_class == &fair_sched_class) &&
prev             3919 kernel/sched/core.c 		p = fair_sched_class.pick_next_task(rq, prev, rf);
prev             3925 kernel/sched/core.c 			p = idle_sched_class.pick_next_task(rq, prev, rf);
prev             3940 kernel/sched/core.c 	for_class_range(class, prev->sched_class, &idle_sched_class) {
prev             3941 kernel/sched/core.c 		if (class->balance(rq, prev, rf))
prev             3946 kernel/sched/core.c 	put_prev_task(rq, prev);
prev             3999 kernel/sched/core.c 	struct task_struct *prev, *next;
prev             4007 kernel/sched/core.c 	prev = rq->curr;
prev             4009 kernel/sched/core.c 	schedule_debug(prev, preempt);
prev             4032 kernel/sched/core.c 	switch_count = &prev->nivcsw;
prev             4033 kernel/sched/core.c 	if (!preempt && prev->state) {
prev             4034 kernel/sched/core.c 		if (signal_pending_state(prev->state, prev)) {
prev             4035 kernel/sched/core.c 			prev->state = TASK_RUNNING;
prev             4037 kernel/sched/core.c 			deactivate_task(rq, prev, DEQUEUE_SLEEP | DEQUEUE_NOCLOCK);
prev             4039 kernel/sched/core.c 			if (prev->in_iowait) {
prev             4044 kernel/sched/core.c 		switch_count = &prev->nvcsw;
prev             4047 kernel/sched/core.c 	next = pick_next_task(rq, prev, &rf);
prev             4048 kernel/sched/core.c 	clear_tsk_need_resched(prev);
prev             4051 kernel/sched/core.c 	if (likely(prev != next)) {
prev             4074 kernel/sched/core.c 		trace_sched_switch(preempt, prev, next);
prev             4077 kernel/sched/core.c 		rq = context_switch(rq, prev, next, &rf);
prev              410 kernel/sched/cputime.c void vtime_common_task_switch(struct task_struct *prev)
prev              412 kernel/sched/cputime.c 	if (is_idle_task(prev))
prev              413 kernel/sched/cputime.c 		vtime_account_idle(prev);
prev              415 kernel/sched/cputime.c 		vtime_account_system(prev);
prev              417 kernel/sched/cputime.c 	vtime_flush(prev);
prev              418 kernel/sched/cputime.c 	arch_vtime_task_switch(prev);
prev              444 kernel/sched/cputime.c void cputime_adjust(struct task_cputime *curr, struct prev_cputime *prev,
prev              591 kernel/sched/cputime.c void cputime_adjust(struct task_cputime *curr, struct prev_cputime *prev,
prev              598 kernel/sched/cputime.c 	raw_spin_lock_irqsave(&prev->lock, flags);
prev              609 kernel/sched/cputime.c 	if (prev->stime + prev->utime >= rtime)
prev              642 kernel/sched/cputime.c 	if (stime < prev->stime)
prev              643 kernel/sched/cputime.c 		stime = prev->stime;
prev              650 kernel/sched/cputime.c 	if (utime < prev->utime) {
prev              651 kernel/sched/cputime.c 		utime = prev->utime;
prev              655 kernel/sched/cputime.c 	prev->stime = stime;
prev              656 kernel/sched/cputime.c 	prev->utime = utime;
prev              658 kernel/sched/cputime.c 	*ut = prev->utime;
prev              659 kernel/sched/cputime.c 	*st = prev->stime;
prev              660 kernel/sched/cputime.c 	raw_spin_unlock_irqrestore(&prev->lock, flags);
prev              807 kernel/sched/cputime.c void arch_vtime_task_switch(struct task_struct *prev)
prev              809 kernel/sched/cputime.c 	struct vtime *vtime = &prev->vtime;
prev              503 kernel/sched/deadline.c static inline bool need_pull_dl_task(struct rq *rq, struct task_struct *prev)
prev              505 kernel/sched/deadline.c 	return dl_task(prev);
prev              621 kernel/sched/deadline.c static inline bool need_pull_dl_task(struct rq *rq, struct task_struct *prev)
prev             1777 kernel/sched/deadline.c pick_next_task_dl(struct rq *rq, struct task_struct *prev, struct rq_flags *rf)
prev             1783 kernel/sched/deadline.c 	WARN_ON_ONCE(prev || rf);
prev              372 kernel/sched/fair.c 			rq->tmp_alone_branch = cfs_rq->leaf_cfs_rq_list.prev;
prev             3171 kernel/sched/fair.c 		      struct cfs_rq *prev, struct cfs_rq *next)
prev             3186 kernel/sched/fair.c 	if (!(se->avg.last_update_time && prev))
prev             3195 kernel/sched/fair.c 			p_last_update_time_copy = prev->load_last_update_time_copy;
prev             3200 kernel/sched/fair.c 			p_last_update_time = prev->avg.last_update_time;
prev             3207 kernel/sched/fair.c 	p_last_update_time = prev->avg.last_update_time;
prev             4256 kernel/sched/fair.c static void put_prev_entity(struct cfs_rq *cfs_rq, struct sched_entity *prev)
prev             4262 kernel/sched/fair.c 	if (prev->on_rq)
prev             4268 kernel/sched/fair.c 	check_spread(cfs_rq, prev);
prev             4270 kernel/sched/fair.c 	if (prev->on_rq) {
prev             4271 kernel/sched/fair.c 		update_stats_wait_start(cfs_rq, prev);
prev             4273 kernel/sched/fair.c 		__enqueue_entity(cfs_rq, prev);
prev             4275 kernel/sched/fair.c 		update_load_avg(cfs_rq, prev, 0);
prev             6014 kernel/sched/fair.c static int select_idle_sibling(struct task_struct *p, int prev, int target)
prev             6025 kernel/sched/fair.c 	if (prev != target && cpus_share_cache(prev, target) &&
prev             6026 kernel/sched/fair.c 	    (available_idle_cpu(prev) || sched_idle_cpu(prev)))
prev             6027 kernel/sched/fair.c 		return prev;
prev             6031 kernel/sched/fair.c 	if (recent_used_cpu != prev &&
prev             6040 kernel/sched/fair.c 		p->recent_used_cpu = prev;
prev             6601 kernel/sched/fair.c balance_fair(struct rq *rq, struct task_struct *prev, struct rq_flags *rf)
prev             6776 kernel/sched/fair.c pick_next_task_fair(struct rq *rq, struct task_struct *prev, struct rq_flags *rf)
prev             6788 kernel/sched/fair.c 	if (!prev || prev->sched_class != &fair_sched_class)
prev             6841 kernel/sched/fair.c 	if (prev != p) {
prev             6842 kernel/sched/fair.c 		struct sched_entity *pse = &prev->se;
prev             6865 kernel/sched/fair.c 	if (prev)
prev             6866 kernel/sched/fair.c 		put_prev_task(rq, prev);
prev             6922 kernel/sched/fair.c static void put_prev_task_fair(struct rq *rq, struct task_struct *prev)
prev             6924 kernel/sched/fair.c 	struct sched_entity *se = &prev->se;
prev              370 kernel/sched/idle.c balance_idle(struct rq *rq, struct task_struct *prev, struct rq_flags *rf)
prev              384 kernel/sched/idle.c static void put_prev_task_idle(struct rq *rq, struct task_struct *prev)
prev              395 kernel/sched/idle.c pick_next_task_idle(struct rq *rq, struct task_struct *prev, struct rq_flags *rf)
prev              399 kernel/sched/idle.c 	if (prev)
prev              400 kernel/sched/idle.c 		put_prev_task(rq, prev);
prev              263 kernel/sched/rt.c static inline bool need_pull_rt_task(struct rq *rq, struct task_struct *prev)
prev              266 kernel/sched/rt.c 	return rq->rt.highest_prio.curr > prev->prio;
prev              418 kernel/sched/rt.c static inline bool need_pull_rt_task(struct rq *rq, struct task_struct *prev)
prev             1571 kernel/sched/rt.c pick_next_task_rt(struct rq *rq, struct task_struct *prev, struct rq_flags *rf)
prev             1575 kernel/sched/rt.c 	WARN_ON_ONCE(prev || rf);
prev             2341 kernel/sched/rt.c 		if (rt_se->run_list.prev != rt_se->run_list.next) {
prev              482 kernel/sched/sched.h 			     struct cfs_rq *prev, struct cfs_rq *next);
prev              485 kernel/sched/sched.h 			     struct cfs_rq *prev, struct cfs_rq *next) { }
prev             1734 kernel/sched/sched.h 					       struct task_struct *prev,
prev             1740 kernel/sched/sched.h 	int (*balance)(struct rq *rq, struct task_struct *prev, struct rq_flags *rf);
prev             1780 kernel/sched/sched.h static inline void put_prev_task(struct rq *rq, struct task_struct *prev)
prev             1782 kernel/sched/sched.h 	WARN_ON_ONCE(rq->curr != prev);
prev             1783 kernel/sched/sched.h 	prev->sched_class->put_prev_task(rq, prev);
prev              226 kernel/sched/stats.h __sched_info_switch(struct rq *rq, struct task_struct *prev, struct task_struct *next)
prev              233 kernel/sched/stats.h 	if (prev != rq->idle)
prev              234 kernel/sched/stats.h 		sched_info_depart(rq, prev);
prev              241 kernel/sched/stats.h sched_info_switch(struct rq *rq, struct task_struct *prev, struct task_struct *next)
prev              244 kernel/sched/stats.h 		__sched_info_switch(rq, prev, next);
prev               20 kernel/sched/stop_task.c balance_stop(struct rq *rq, struct task_struct *prev, struct rq_flags *rf)
prev               38 kernel/sched/stop_task.c pick_next_task_stop(struct rq *rq, struct task_struct *prev, struct rq_flags *rf)
prev               40 kernel/sched/stop_task.c 	WARN_ON_ONCE(prev || rf);
prev               66 kernel/sched/stop_task.c static void put_prev_task_stop(struct rq *rq, struct task_struct *prev)
prev              131 kernel/seccomp.c 	struct seccomp_filter *prev;
prev              271 kernel/seccomp.c 	for (; f; f = f->prev) {
prev              322 kernel/seccomp.c 	for (; child; child = child->prev)
prev              520 kernel/seccomp.c 	for (walker = current->seccomp.filter; walker; walker = walker->prev)
prev              542 kernel/seccomp.c 	filter->prev = current->seccomp.filter;
prev              579 kernel/seccomp.c 		orig = orig->prev;
prev             1216 kernel/seccomp.c 	for (cur = current->seccomp.filter; cur; cur = cur->prev) {
prev             1476 kernel/seccomp.c 	for (filter = orig; filter; filter = filter->prev)
prev             1485 kernel/seccomp.c 	for (filter = orig; filter && count > 1; filter = filter->prev)
prev              285 kernel/time/tick-sched.c 	int prev;
prev              287 kernel/time/tick-sched.c 	prev = atomic_fetch_or(BIT(bit), dep);
prev              288 kernel/time/tick-sched.c 	if (!prev)
prev              312 kernel/time/tick-sched.c 	int prev;
prev              317 kernel/time/tick-sched.c 	prev = atomic_fetch_or(BIT(bit), &ts->tick_dep_mask);
prev              318 kernel/time/tick-sched.c 	if (!prev) {
prev              390 kernel/trace/fgraph.c 			struct task_struct *prev, struct task_struct *next)
prev              404 kernel/trace/fgraph.c 	prev->ftrace_timestamp = timestamp;
prev             6424 kernel/trace/ftrace.c 		    struct task_struct *prev, struct task_struct *next)
prev              891 kernel/trace/ring_buffer.c 	struct list_head *list = page->list.prev;
prev              923 kernel/trace/ring_buffer.c 	rb_set_list_to_head(cpu_buffer, head->list.prev);
prev              950 kernel/trace/ring_buffer.c 			    struct buffer_page *prev,
prev              957 kernel/trace/ring_buffer.c 	list = &prev->list;
prev              973 kernel/trace/ring_buffer.c 				   struct buffer_page *prev,
prev              976 kernel/trace/ring_buffer.c 	return rb_head_page_set(cpu_buffer, head, prev,
prev              982 kernel/trace/ring_buffer.c 				 struct buffer_page *prev,
prev              985 kernel/trace/ring_buffer.c 	return rb_head_page_set(cpu_buffer, head, prev,
prev              991 kernel/trace/ring_buffer.c 				   struct buffer_page *prev,
prev              994 kernel/trace/ring_buffer.c 	return rb_head_page_set(cpu_buffer, head, prev,
prev             1019 kernel/trace/ring_buffer.c 	if (RB_WARN_ON(cpu_buffer, rb_list_head(list->prev->next) != list))
prev             1031 kernel/trace/ring_buffer.c 			if (rb_is_head_page(cpu_buffer, page, page->list.prev)) {
prev             1047 kernel/trace/ring_buffer.c 	unsigned long *ptr = (unsigned long *)&old->list.prev->next;
prev             1140 kernel/trace/ring_buffer.c 	if (RB_WARN_ON(cpu_buffer, rb_list_head(list->prev) != list->prev))
prev             1165 kernel/trace/ring_buffer.c 	if (RB_WARN_ON(cpu_buffer, head->next->prev != head))
prev             1167 kernel/trace/ring_buffer.c 	if (RB_WARN_ON(cpu_buffer, head->prev->next != head))
prev             1175 kernel/trace/ring_buffer.c 			       bpage->list.next->prev != &bpage->list))
prev             1178 kernel/trace/ring_buffer.c 			       bpage->list.prev->next != &bpage->list))
prev             1546 kernel/trace/ring_buffer.c 	next_page->prev = tail_page;
prev             1637 kernel/trace/ring_buffer.c 		prev_page = head_page->prev;
prev             1640 kernel/trace/ring_buffer.c 		last_page  = pages->prev;
prev             1646 kernel/trace/ring_buffer.c 		first_page->prev = prev_page;
prev             1656 kernel/trace/ring_buffer.c 			head_page->prev = last_page;
prev             3731 kernel/trace/ring_buffer.c 	cpu_buffer->reader_page->list.prev = reader->list.prev;
prev             3738 kernel/trace/ring_buffer.c 	cpu_buffer->pages = reader->list.prev;
prev             3779 kernel/trace/ring_buffer.c 	rb_list_head(reader->list.next)->prev = &cpu_buffer->reader_page->list;
prev             2292 kernel/trace/trace.c void tracing_record_taskinfo_sched_switch(struct task_struct *prev,
prev             2304 kernel/trace/trace.c 	done  = !(flags & TRACE_RECORD_CMDLINE) || trace_save_cmdline(prev);
prev             2306 kernel/trace/trace.c 	done &= !(flags & TRACE_RECORD_TGID) || trace_save_tgid(prev);
prev              359 kernel/trace/trace.h 	tr = list_entry(ftrace_trace_arrays.prev,
prev              542 kernel/trace/trace_events.c 		    struct task_struct *prev, struct task_struct *next)
prev              550 kernel/trace/trace_events.c 		       trace_ignore_this_task(pid_list, prev) &&
prev              556 kernel/trace/trace_events.c 		    struct task_struct *prev, struct task_struct *next)
prev             1206 kernel/trace/trace_events.c 	node = node->prev;
prev               25 kernel/trace/trace_sched_switch.c 		   struct task_struct *prev, struct task_struct *next)
prev               34 kernel/trace/trace_sched_switch.c 	tracing_record_taskinfo_sched_switch(prev, next, flags);
prev              376 kernel/trace/trace_sched_wakeup.c 			   struct task_struct *prev,
prev              390 kernel/trace/trace_sched_wakeup.c 	entry->prev_pid			= prev->pid;
prev              391 kernel/trace/trace_sched_wakeup.c 	entry->prev_prio		= prev->prio;
prev              392 kernel/trace/trace_sched_wakeup.c 	entry->prev_state		= task_state_index(prev);
prev              432 kernel/trace/trace_sched_wakeup.c 			  struct task_struct *prev, struct task_struct *next)
prev              441 kernel/trace/trace_sched_wakeup.c 	tracing_record_cmdline(prev);
prev              477 kernel/trace/trace_sched_wakeup.c 	tracing_sched_switch_trace(wakeup_trace, prev, next, flags, pc);
prev               17 kernel/trace/trace_stat.h 	void			*(*stat_next)(void *prev, int idx);
prev              723 kernel/user_namespace.c 		struct uid_gid_extent *prev;
prev              726 kernel/user_namespace.c 			prev = &new_map->extent[idx];
prev              728 kernel/user_namespace.c 			prev = &new_map->forward[idx];
prev              730 kernel/user_namespace.c 		prev_upper_first = prev->first;
prev              731 kernel/user_namespace.c 		prev_lower_first = prev->lower_first;
prev              732 kernel/user_namespace.c 		prev_upper_last = prev_upper_first + prev->count - 1;
prev              733 kernel/user_namespace.c 		prev_lower_last = prev_lower_first + prev->count - 1;
prev             2000 kernel/workqueue.c 		worker = list_entry(pool->idle_list.prev, struct worker, entry);
prev             4175 kernel/workqueue.c 			      wq->pwqs.prev != &wq->dfl_pwq->pwqs_node),
prev              138 lib/dim/net_dim.c 				 struct dim_stats *prev)
prev              140 lib/dim/net_dim.c 	if (!prev->bpms)
prev              143 lib/dim/net_dim.c 	if (IS_SIGNIFICANT_DIFF(curr->bpms, prev->bpms))
prev              144 lib/dim/net_dim.c 		return (curr->bpms > prev->bpms) ? DIM_STATS_BETTER :
prev              147 lib/dim/net_dim.c 	if (!prev->ppms)
prev              151 lib/dim/net_dim.c 	if (IS_SIGNIFICANT_DIFF(curr->ppms, prev->ppms))
prev              152 lib/dim/net_dim.c 		return (curr->ppms > prev->ppms) ? DIM_STATS_BETTER :
prev              155 lib/dim/net_dim.c 	if (!prev->epms)
prev              158 lib/dim/net_dim.c 	if (IS_SIGNIFICANT_DIFF(curr->epms, prev->epms))
prev              159 lib/dim/net_dim.c 		return (curr->epms < prev->epms) ? DIM_STATS_BETTER :
prev               27 lib/dim/rdma_dim.c 				  struct dim_stats *prev)
prev               30 lib/dim/rdma_dim.c 	if (!prev->cpms)
prev               33 lib/dim/rdma_dim.c 	if (IS_SIGNIFICANT_DIFF(curr->cpms, prev->cpms))
prev               34 lib/dim/rdma_dim.c 		return (curr->cpms > prev->cpms) ? DIM_STATS_BETTER :
prev               37 lib/dim/rdma_dim.c 	if (IS_SIGNIFICANT_DIFF(curr->cpe_ratio, prev->cpe_ratio))
prev               38 lib/dim/rdma_dim.c 		return (curr->cpe_ratio > prev->cpe_ratio) ? DIM_STATS_BETTER :
prev              337 lib/klist.c    	struct klist_node *prev;
prev              343 lib/klist.c    		prev = to_klist_node(last->n_node.prev);
prev              347 lib/klist.c    		prev = to_klist_node(i->i_klist->k_list.prev);
prev              350 lib/klist.c    	while (prev != to_klist_node(&i->i_klist->k_list)) {
prev              351 lib/klist.c    		if (likely(!knode_dead(prev))) {
prev              352 lib/klist.c    			kref_get(&prev->n_ref);
prev              353 lib/klist.c    			i->i_cur = prev;
prev              356 lib/klist.c    		prev = to_klist_node(prev->n_node.prev);
prev               20 lib/list_debug.c bool __list_add_valid(struct list_head *new, struct list_head *prev,
prev               23 lib/list_debug.c 	if (CHECK_DATA_CORRUPTION(next->prev != prev,
prev               25 lib/list_debug.c 			prev, next->prev, next) ||
prev               26 lib/list_debug.c 	    CHECK_DATA_CORRUPTION(prev->next != next,
prev               28 lib/list_debug.c 			next, prev->next, prev) ||
prev               29 lib/list_debug.c 	    CHECK_DATA_CORRUPTION(new == prev || new == next,
prev               31 lib/list_debug.c 			new, prev, next))
prev               40 lib/list_debug.c 	struct list_head *prev, *next;
prev               42 lib/list_debug.c 	prev = entry->prev;
prev               48 lib/list_debug.c 	    CHECK_DATA_CORRUPTION(prev == LIST_POISON2,
prev               51 lib/list_debug.c 	    CHECK_DATA_CORRUPTION(prev->next != entry,
prev               53 lib/list_debug.c 			entry, prev->next) ||
prev               54 lib/list_debug.c 	    CHECK_DATA_CORRUPTION(next->prev != entry,
prev               56 lib/list_debug.c 			entry, next->prev))
prev               65 lib/list_sort.c 			a->prev = tail;
prev               72 lib/list_sort.c 			b->prev = tail;
prev               94 lib/list_sort.c 		b->prev = tail;
prev              101 lib/list_sort.c 	head->prev = tail;
prev              195 lib/list_sort.c 	if (list == head->prev)	/* Zero or one elements */
prev              199 lib/list_sort.c 	head->prev->next = NULL;
prev              225 lib/list_sort.c 			tail = &(*tail)->prev;
prev              228 lib/list_sort.c 			struct list_head *a = *tail, *b = a->prev;
prev              232 lib/list_sort.c 			a->prev = b->prev;
prev              237 lib/list_sort.c 		list->prev = pending;
prev              246 lib/list_sort.c 	pending = pending->prev;
prev              248 lib/list_sort.c 		struct list_head *next = pending->prev;
prev               17 lib/lockref.c  		struct lockref new = old, prev = old;				\
prev               22 lib/lockref.c  		if (likely(old.lock_count == prev.lock_count)) {		\
prev              329 lib/lru_cache.c 		n = lc->lru.prev;
prev               35 lib/plist.c    	WARN(n->prev != p || p->next != n,
prev               39 lib/plist.c    			 t, t->next, t->prev,
prev               40 lib/plist.c    			p, p->next, p->prev,
prev               41 lib/plist.c    			n, n->next, n->prev);
prev               46 lib/plist.c    	struct list_head *prev = top, *next = top->next;
prev               48 lib/plist.c    	plist_check_prev_next(top, prev, next);
prev               50 lib/plist.c    		prev = next;
prev               51 lib/plist.c    		next = prev->next;
prev               52 lib/plist.c    		plist_check_prev_next(top, prev, next);
prev               75 lib/plist.c    	struct plist_node *first, *iter, *prev = NULL;
prev               93 lib/plist.c    		prev = iter;
prev               98 lib/plist.c    	if (!prev || prev->prio != node->prio)
prev              114 lib/rhashtable.c 					      union nested_table __rcu **prev,
prev              120 lib/rhashtable.c 	ntbl = rcu_dereference(*prev);
prev              131 lib/rhashtable.c 	if (cmpxchg((union nested_table **)prev, NULL, ntbl) == NULL)
prev              135 lib/rhashtable.c 	return rcu_dereference(*prev);
prev              104 lib/test_list_sort.c 		if (cur->next->prev != cur) {
prev              129 lib/test_list_sort.c 	if (head.prev != cur) {
prev             2022 lib/xarray.c   		node->array, node->private_list.prev, node->private_list.next);
prev              155 lib/zlib_deflate/deflate.c     s->prev[(str) & s->w_mask] = match_head = s->head[s->ins_h], \
prev              233 lib/zlib_deflate/deflate.c     s->prev   = (Pos *)  mem->prev_memory;
prev              570 lib/zlib_deflate/deflate.c     Pos *prev = s->prev;
prev              686 lib/zlib_deflate/deflate.c     } while ((cur_match = prev[cur_match & wmask]) > limit
prev              780 lib/zlib_deflate/deflate.c             p = &s->prev[n];
prev              104 lib/zlib_deflate/defutil.h     Pos *prev;
prev              290 mm/hugetlb.c   		list_add(&nrg->link, rg->link.prev);
prev              302 mm/hugetlb.c   	list_for_each_entry_safe(rg, trg, rg->link.prev, link) {
prev              415 mm/hugetlb.c   		list_add(&nrg->link, rg->link.prev);
prev              426 mm/hugetlb.c   	list_for_each_entry(rg, rg->link.prev, link) {
prev              293 mm/internal.h  		struct vm_area_struct *prev, struct rb_node *rb_parent);
prev               29 mm/interval_tree.c 				    struct vm_area_struct *prev,
prev               36 mm/interval_tree.c 	VM_BUG_ON_VMA(vma_start_pgoff(node) != vma_start_pgoff(prev), node);
prev               38 mm/interval_tree.c 	if (!prev->shared.rb.rb_right) {
prev               39 mm/interval_tree.c 		parent = prev;
prev               40 mm/interval_tree.c 		link = &prev->shared.rb.rb_right;
prev               42 mm/interval_tree.c 		parent = rb_entry(prev->shared.rb.rb_right,
prev              229 mm/ksm.c       #define STABLE_NODE_DUP_HEAD ((struct list_head *)&migrate_nodes.prev)
prev               66 mm/madvise.c   		     struct vm_area_struct **prev,
prev              130 mm/madvise.c   		*prev = vma;
prev              135 mm/madvise.c   	*prev = vma_merge(mm, *prev, start, end, new_flags, vma->anon_vma,
prev              138 mm/madvise.c   	if (*prev) {
prev              139 mm/madvise.c   		vma = *prev;
prev              143 mm/madvise.c   	*prev = vma;
prev              254 mm/madvise.c   			     struct vm_area_struct **prev,
prev              260 mm/madvise.c   	*prev = vma;
prev              289 mm/madvise.c   	*prev = NULL;	/* tell sys_madvise we drop mmap_sem */
prev              492 mm/madvise.c   			struct vm_area_struct **prev,
prev              498 mm/madvise.c   	*prev = vma;
prev              541 mm/madvise.c   			struct vm_area_struct **prev,
prev              547 mm/madvise.c   	*prev = vma;
prev              761 mm/madvise.c   				  struct vm_area_struct **prev,
prev              765 mm/madvise.c   	*prev = vma;
prev              770 mm/madvise.c   		*prev = NULL; /* mmap_sem has been dropped, prev is stale */
prev              821 mm/madvise.c   				struct vm_area_struct **prev,
prev              828 mm/madvise.c   	*prev = NULL;	/* tell sys_madvise we drop mmap_sem */
prev              934 mm/madvise.c   madvise_vma(struct vm_area_struct *vma, struct vm_area_struct **prev,
prev              939 mm/madvise.c   		return madvise_remove(vma, prev, start, end);
prev              941 mm/madvise.c   		return madvise_willneed(vma, prev, start, end);
prev              943 mm/madvise.c   		return madvise_cold(vma, prev, start, end);
prev              945 mm/madvise.c   		return madvise_pageout(vma, prev, start, end);
prev              948 mm/madvise.c   		return madvise_dontneed_free(vma, prev, start, end, behavior);
prev              950 mm/madvise.c   		return madvise_behavior(vma, prev, start, end, behavior);
prev             1056 mm/madvise.c   	struct vm_area_struct *vma, *prev;
prev             1102 mm/madvise.c   	vma = find_vma_prev(current->mm, start, &prev);
prev             1104 mm/madvise.c   		prev = vma;
prev             1127 mm/madvise.c   		error = madvise_vma(vma, &prev, start, tmp, behavior);
prev             1131 mm/madvise.c   		if (prev && start < prev->vm_end)
prev             1132 mm/madvise.c   			start = prev->vm_end;
prev             1136 mm/madvise.c   		if (prev)
prev             1137 mm/madvise.c   			vma = prev->vm_next;
prev             1039 mm/memcontrol.c 				   struct mem_cgroup *prev,
prev             1053 mm/memcontrol.c 	if (prev && !reclaim)
prev             1054 mm/memcontrol.c 		pos = prev;
prev             1057 mm/memcontrol.c 		if (prev)
prev             1070 mm/memcontrol.c 		if (prev && reclaim->generation != iter->generation)
prev             1101 mm/memcontrol.c 			if (!prev)
prev             1135 mm/memcontrol.c 		else if (!prev)
prev             1142 mm/memcontrol.c 	if (prev && prev != root)
prev             1143 mm/memcontrol.c 		css_put(&prev->css);
prev             1154 mm/memcontrol.c 			   struct mem_cgroup *prev)
prev             1158 mm/memcontrol.c 	if (prev && prev != root)
prev             1159 mm/memcontrol.c 		css_put(&prev->css);
prev              413 mm/mempolicy.c 	struct vm_area_struct *prev;
prev              637 mm/mempolicy.c 		if (qp->prev && qp->prev->vm_end < vma->vm_start)
prev              641 mm/mempolicy.c 	qp->prev = vma;
prev              688 mm/mempolicy.c 		.prev = NULL,
prev              735 mm/mempolicy.c 	struct vm_area_struct *prev;
prev              746 mm/mempolicy.c 	prev = vma->vm_prev;
prev              748 mm/mempolicy.c 		prev = vma;
prev              750 mm/mempolicy.c 	for (; vma && vma->vm_start < end; prev = vma, vma = next) {
prev              760 mm/mempolicy.c 		prev = vma_merge(mm, prev, vmstart, vmend, vma->vm_flags,
prev              763 mm/mempolicy.c 		if (prev) {
prev              764 mm/mempolicy.c 			vma = prev;
prev             2306 mm/mempolicy.c 		struct rb_node *prev = rb_prev(n);
prev             2307 mm/mempolicy.c 		if (!prev)
prev             2309 mm/mempolicy.c 		w = rb_entry(prev, struct sp_node, nd);
prev             2312 mm/mempolicy.c 		n = prev;
prev              519 mm/mlock.c     static int mlock_fixup(struct vm_area_struct *vma, struct vm_area_struct **prev,
prev              536 mm/mlock.c     	*prev = vma_merge(mm, *prev, start, end, newflags, vma->anon_vma,
prev              539 mm/mlock.c     	if (*prev) {
prev              540 mm/mlock.c     		vma = *prev;
prev              579 mm/mlock.c     	*prev = vma;
prev              587 mm/mlock.c     	struct vm_area_struct * vma, * prev;
prev              601 mm/mlock.c     	prev = vma->vm_prev;
prev              603 mm/mlock.c     		prev = vma;
prev              614 mm/mlock.c     		error = mlock_fixup(vma, &prev, nstart, tmp, newflags);
prev              618 mm/mlock.c     		if (nstart < prev->vm_end)
prev              619 mm/mlock.c     			nstart = prev->vm_end;
prev              623 mm/mlock.c     		vma = prev->vm_next;
prev              765 mm/mlock.c     	struct vm_area_struct * vma, * prev = NULL;
prev              785 mm/mlock.c     	for (vma = current->mm->mmap; vma ; vma = prev->vm_next) {
prev              792 mm/mlock.c     		mlock_fixup(vma, &prev, vma->vm_start, vma->vm_end, newflags);
prev               77 mm/mmap.c      		struct vm_area_struct *vma, struct vm_area_struct *prev,
prev              331 mm/mmap.c      	unsigned long prev = 0, pend = 0;
prev              336 mm/mmap.c      		if (vma->vm_start < prev) {
prev              338 mm/mmap.c      				  vma->vm_start, prev);
prev              361 mm/mmap.c      		prev = vma->vm_start;
prev              633 mm/mmap.c      	struct vm_area_struct *prev, struct rb_node **rb_link,
prev              636 mm/mmap.c      	__vma_link_list(mm, vma, prev, rb_parent);
prev              641 mm/mmap.c      			struct vm_area_struct *prev, struct rb_node **rb_link,
prev              651 mm/mmap.c      	__vma_link(mm, vma, prev, rb_link, rb_parent);
prev              667 mm/mmap.c      	struct vm_area_struct *prev;
prev              671 mm/mmap.c      			   &prev, &rb_link, &rb_parent))
prev              673 mm/mmap.c      	__vma_link(mm, vma, prev, rb_link, rb_parent);
prev              679 mm/mmap.c      						struct vm_area_struct *prev,
prev              688 mm/mmap.c      		prev->vm_next = next;
prev              690 mm/mmap.c      		prev = vma->vm_prev;
prev              691 mm/mmap.c      		if (prev)
prev              692 mm/mmap.c      			prev->vm_next = next;
prev              697 mm/mmap.c      		next->vm_prev = prev;
prev              705 mm/mmap.c      				     struct vm_area_struct *prev)
prev              707 mm/mmap.c      	__vma_unlink_common(mm, vma, prev, true, vma);
prev             1136 mm/mmap.c      			struct vm_area_struct *prev, unsigned long addr,
prev             1153 mm/mmap.c      	if (prev)
prev             1154 mm/mmap.c      		next = prev->vm_next;
prev             1162 mm/mmap.c      	VM_WARN_ON(prev && addr <= prev->vm_start);
prev             1169 mm/mmap.c      	if (prev && prev->vm_end == addr &&
prev             1170 mm/mmap.c      			mpol_equal(vma_policy(prev), policy) &&
prev             1171 mm/mmap.c      			can_vma_merge_after(prev, vm_flags,
prev             1183 mm/mmap.c      				is_mergeable_anon_vma(prev->anon_vma,
prev             1186 mm/mmap.c      			err = __vma_adjust(prev, prev->vm_start,
prev             1187 mm/mmap.c      					 next->vm_end, prev->vm_pgoff, NULL,
prev             1188 mm/mmap.c      					 prev);
prev             1190 mm/mmap.c      			err = __vma_adjust(prev, prev->vm_start,
prev             1191 mm/mmap.c      					 end, prev->vm_pgoff, NULL, prev);
prev             1194 mm/mmap.c      		khugepaged_enter_vma_merge(prev, vm_flags);
prev             1195 mm/mmap.c      		return prev;
prev             1206 mm/mmap.c      		if (prev && addr < prev->vm_end)	/* case 4 */
prev             1207 mm/mmap.c      			err = __vma_adjust(prev, prev->vm_start,
prev             1208 mm/mmap.c      					 addr, prev->vm_pgoff, NULL, next);
prev             1716 mm/mmap.c      	struct vm_area_struct *vma, *prev;
prev             1737 mm/mmap.c      	while (find_vma_links(mm, addr, addr + len, &prev, &rb_link,
prev             1756 mm/mmap.c      	vma = vma_merge(mm, prev, addr, addr + len, vm_flags,
prev             1819 mm/mmap.c      	vma_link(mm, vma, prev, rb_link, rb_parent);
prev             1862 mm/mmap.c      	unmap_region(mm, vma, prev, vma->vm_start, vma->vm_end);
prev             1947 mm/mmap.c      			struct rb_node *prev = &vma->vm_rb;
prev             1948 mm/mmap.c      			if (!rb_parent(prev))
prev             1950 mm/mmap.c      			vma = rb_entry(rb_parent(prev),
prev             1952 mm/mmap.c      			if (prev == vma->vm_rb.rb_left) {
prev             2051 mm/mmap.c      			struct rb_node *prev = &vma->vm_rb;
prev             2052 mm/mmap.c      			if (!rb_parent(prev))
prev             2054 mm/mmap.c      			vma = rb_entry(rb_parent(prev),
prev             2056 mm/mmap.c      			if (prev == vma->vm_rb.rb_right) {
prev             2105 mm/mmap.c      	struct vm_area_struct *vma, *prev;
prev             2117 mm/mmap.c      		vma = find_vma_prev(mm, addr, &prev);
prev             2120 mm/mmap.c      		    (!prev || addr >= vm_end_gap(prev)))
prev             2143 mm/mmap.c      	struct vm_area_struct *vma, *prev;
prev             2158 mm/mmap.c      		vma = find_vma_prev(mm, addr, &prev);
prev             2161 mm/mmap.c      				(!prev || addr >= vm_end_gap(prev)))
prev             2434 mm/mmap.c      	struct vm_area_struct *prev;
prev             2442 mm/mmap.c      	prev = vma->vm_prev;
prev             2444 mm/mmap.c      	if (prev && !(prev->vm_flags & VM_GROWSDOWN) &&
prev             2445 mm/mmap.c      			(prev->vm_flags & (VM_WRITE|VM_READ|VM_EXEC))) {
prev             2446 mm/mmap.c      		if (address - prev->vm_end < stack_guard_gap)
prev             2529 mm/mmap.c      	struct vm_area_struct *vma, *prev;
prev             2532 mm/mmap.c      	vma = find_vma_prev(mm, addr, &prev);
prev             2536 mm/mmap.c      	if (!prev || !mmget_still_valid(mm) || expand_stack(prev, addr))
prev             2538 mm/mmap.c      	if (prev->vm_flags & VM_LOCKED)
prev             2539 mm/mmap.c      		populate_vma_page_range(prev, addr, prev->vm_end, NULL);
prev             2540 mm/mmap.c      	return prev;
prev             2606 mm/mmap.c      		struct vm_area_struct *vma, struct vm_area_struct *prev,
prev             2609 mm/mmap.c      	struct vm_area_struct *next = prev ? prev->vm_next : mm->mmap;
prev             2616 mm/mmap.c      	free_pgtables(&tlb, vma, prev ? prev->vm_end : FIRST_USER_ADDRESS,
prev             2627 mm/mmap.c      	struct vm_area_struct *prev, unsigned long end)
prev             2632 mm/mmap.c      	insertion_point = (prev ? &prev->vm_next : &mm->mmap);
prev             2642 mm/mmap.c      		vma->vm_prev = prev;
prev             2645 mm/mmap.c      		mm->highest_vm_end = prev ? vm_end_gap(prev) : 0;
prev             2738 mm/mmap.c      	struct vm_area_struct *vma, *prev, *last;
prev             2759 mm/mmap.c      	prev = vma->vm_prev;
prev             2787 mm/mmap.c      		prev = vma;
prev             2797 mm/mmap.c      	vma = prev ? prev->vm_next : mm->mmap;
prev             2830 mm/mmap.c      	detach_vmas_to_be_unmapped(mm, vma, prev, end);
prev             2835 mm/mmap.c      	unmap_region(mm, vma, prev, start, end);
prev             2995 mm/mmap.c      	struct vm_area_struct *vma, *prev;
prev             3016 mm/mmap.c      	while (find_vma_links(mm, addr, addr + len, &prev, &rb_link,
prev             3033 mm/mmap.c      	vma = vma_merge(mm, prev, addr, addr + len, flags,
prev             3053 mm/mmap.c      	vma_link(mm, vma, prev, rb_link, rb_parent);
prev             3173 mm/mmap.c      	struct vm_area_struct *prev;
prev             3177 mm/mmap.c      			   &prev, &rb_link, &rb_parent))
prev             3200 mm/mmap.c      	vma_link(mm, vma, prev, rb_link, rb_parent);
prev             3215 mm/mmap.c      	struct vm_area_struct *new_vma, *prev;
prev             3228 mm/mmap.c      	if (find_vma_links(mm, addr, addr + len, &prev, &rb_link, &rb_parent))
prev             3230 mm/mmap.c      	new_vma = vma_merge(mm, prev, addr, addr + len, vma->vm_flags,
prev             3270 mm/mmap.c      		vma_link(mm, new_vma, prev, rb_link, rb_parent);
prev              490 mm/mprotect.c  	struct vm_area_struct *vma, *prev;
prev              530 mm/mprotect.c  	prev = vma->vm_prev;
prev              549 mm/mprotect.c  		prev = vma;
prev              587 mm/mprotect.c  		error = mprotect_fixup(vma, &prev, nstart, tmp, newflags);
prev              592 mm/mprotect.c  		if (nstart < prev->vm_end)
prev              593 mm/mprotect.c  			nstart = prev->vm_end;
prev              597 mm/mprotect.c  		vma = prev->vm_next;
prev              588 mm/nommu.c     	struct vm_area_struct *pvma, *prev;
prev              640 mm/nommu.c     	prev = NULL;
prev              642 mm/nommu.c     		prev = rb_entry(rb_prev, struct vm_area_struct, vm_rb);
prev              644 mm/nommu.c     	__vma_link_list(mm, vma, prev, parent);
prev             2192 mm/slab.c      		p = n->slabs_free.prev;
prev             3801 mm/slab.c      	struct array_cache __percpu *cpu_cache, *prev;
prev             3808 mm/slab.c      	prev = cachep->cpu_cache;
prev             3814 mm/slab.c      	if (prev)
prev             3822 mm/slab.c      	if (!prev)
prev             3829 mm/slab.c      		struct array_cache *ac = per_cpu_ptr(prev, cpu);
prev             3838 mm/slab.c      	free_percpu(prev);
prev              240 mm/slob.c      	slob_t *prev, *cur, *aligned = NULL;
prev              244 mm/slob.c      	for (prev = NULL, cur = sp->freelist; ; prev = cur, cur = slob_next(cur)) {
prev              267 mm/slob.c      				prev = cur;
prev              274 mm/slob.c      				if (prev)
prev              275 mm/slob.c      					set_slob(prev, slob_units(prev), next);
prev              279 mm/slob.c      				if (prev)
prev              280 mm/slob.c      					set_slob(prev, slob_units(prev), cur + units);
prev              385 mm/slob.c      	slob_t *prev, *next, *b = (slob_t *)block;
prev              441 mm/slob.c      		prev = sp->freelist;
prev              442 mm/slob.c      		next = slob_next(prev);
prev              444 mm/slob.c      			prev = next;
prev              445 mm/slob.c      			next = slob_next(prev);
prev              448 mm/slob.c      		if (!slob_last(prev) && b + units == next) {
prev              454 mm/slob.c      		if (prev + slob_units(prev) == b) {
prev              455 mm/slob.c      			units = slob_units(b) + slob_units(prev);
prev              456 mm/slob.c      			set_slob(prev, units, slob_next(b));
prev              458 mm/slob.c      			set_slob(prev, slob_units(prev), b);
prev             1398 mm/swapfile.c  	struct swap_info_struct *p, *prev;
prev             1404 mm/swapfile.c  	prev = NULL;
prev             1415 mm/swapfile.c  		p = swap_info_get_cont(entries[i], prev);
prev             1418 mm/swapfile.c  		prev = p;
prev             2093 mm/swapfile.c  					unsigned int prev, bool frontswap)
prev             2104 mm/swapfile.c  	for (i = prev + 1; i < si->max; i++) {
prev             3677 mm/swapfile.c  		page = list_entry(page->lru.prev, struct page, lru);
prev             3682 mm/swapfile.c  			page = list_entry(page->lru.prev, struct page, lru);
prev             3702 mm/swapfile.c  		page = list_entry(page->lru.prev, struct page, lru);
prev             3708 mm/swapfile.c  			page = list_entry(page->lru.prev, struct page, lru);
prev              274 mm/util.c      		struct vm_area_struct *prev, struct rb_node *rb_parent)
prev              278 mm/util.c      	vma->vm_prev = prev;
prev              279 mm/util.c      	if (prev) {
prev              280 mm/util.c      		next = prev->vm_next;
prev              281 mm/util.c      		prev->vm_next = vma;
prev              511 mm/vmalloc.c   			head = head->prev;
prev              740 mm/vmalloc.c   	if (next->prev != head) {
prev              741 mm/vmalloc.c   		sibling = list_entry(next->prev, struct vmap_area, list);
prev              139 mm/vmscan.c    		if ((_page)->lru.prev != _base) {			\
prev              140 mm/vmscan.c    			struct page *prev;				\
prev              142 mm/vmscan.c    			prev = lru_to_page(&(_page->lru));		\
prev              143 mm/vmscan.c    			prefetch(&prev->_field);			\
prev              153 mm/vmscan.c    		if ((_page)->lru.prev != _base) {			\
prev              154 mm/vmscan.c    			struct page *prev;				\
prev              156 mm/vmscan.c    			prev = lru_to_page(&(_page->lru));		\
prev              157 mm/vmscan.c    			prefetchw(&prev->_field);			\
prev             1255 mm/zsmalloc.c  static bool can_merge(struct size_class *prev, int pages_per_zspage,
prev             1258 mm/zsmalloc.c  	if (prev->pages_per_zspage == pages_per_zspage &&
prev             1259 mm/zsmalloc.c  		prev->objs_per_zspage == objs_per_zspage)
prev              860 net/9p/trans_fd.c 		 m, m->mux_list.prev, m->mux_list.next);
prev              376 net/atm/br2684.c 		    brdev->brvccs.next != brdev->brvccs.prev)	/* >1 VCC */
prev              709 net/atm/br2684.c 		brdev->number = BRPRIV(list_entry_brdev(br2684_devs.prev))->number + 1;
prev              113 net/atm/mpoa_caches.c 	entry->prev = NULL;
prev              115 net/atm/mpoa_caches.c 		client->in_cache->prev = entry;
prev              200 net/atm/mpoa_caches.c 	if (entry->prev != NULL)
prev              201 net/atm/mpoa_caches.c 		entry->prev->next = entry->next;
prev              205 net/atm/mpoa_caches.c 		entry->next->prev = entry->prev;
prev              433 net/atm/mpoa_caches.c 	if (entry->prev != NULL)
prev              434 net/atm/mpoa_caches.c 		entry->prev->next = entry->next;
prev              438 net/atm/mpoa_caches.c 		entry->next->prev = entry->prev;
prev              473 net/atm/mpoa_caches.c 	entry->prev = NULL;
prev              475 net/atm/mpoa_caches.c 		client->eg_cache->prev = entry;
prev               19 net/atm/mpoa_caches.h 	struct in_cache_entry *prev;
prev               56 net/atm/mpoa_caches.h 	struct               eg_cache_entry *prev;
prev              114 net/bridge/br_forward.c static int deliver_clone(const struct net_bridge_port *prev,
prev              125 net/bridge/br_forward.c 	__br_forward(prev, skb, local_orig);
prev              169 net/bridge/br_forward.c 	struct net_bridge_port *prev, struct net_bridge_port *p,
prev              176 net/bridge/br_forward.c 		return prev;
prev              178 net/bridge/br_forward.c 	if (!prev)
prev              181 net/bridge/br_forward.c 	err = deliver_clone(prev, skb, local_orig);
prev              194 net/bridge/br_forward.c 	struct net_bridge_port *prev = NULL;
prev              223 net/bridge/br_forward.c 		prev = maybe_deliver(prev, p, skb, local_orig);
prev              224 net/bridge/br_forward.c 		if (IS_ERR(prev))
prev              228 net/bridge/br_forward.c 	if (!prev)
prev              232 net/bridge/br_forward.c 		deliver_clone(prev, skb, local_orig);
prev              234 net/bridge/br_forward.c 		__br_forward(prev, skb, local_orig);
prev              275 net/bridge/br_forward.c 	struct net_bridge_port *prev = NULL;
prev              299 net/bridge/br_forward.c 		prev = maybe_deliver(prev, port, skb, local_orig);
prev              300 net/bridge/br_forward.c 		if (IS_ERR(prev))
prev              309 net/bridge/br_forward.c 	if (!prev)
prev              313 net/bridge/br_forward.c 		deliver_clone(prev, skb, local_orig);
prev              315 net/bridge/br_forward.c 		__br_forward(prev, skb, local_orig);
prev               42 net/bridge/netfilter/ebt_limit.c 	info->credit += (now - xchg(&info->prev, now)) * CREDITS_PER_JIFFY;
prev               82 net/bridge/netfilter/ebt_limit.c 	info->prev = jiffies;
prev               97 net/bridge/netfilter/ebt_limit.c 	compat_ulong_t prev;
prev              109 net/bridge/netfilter/ebt_limit.c 	.usersize	= offsetof(struct ebt_limit_info, prev),
prev              129 net/can/j1939/address-claim.c 	struct j1939_ecu *ecu, *prev;
prev              187 net/can/j1939/address-claim.c 	prev = j1939_ecu_get_by_addr_locked(priv, skcb->addr.sa);
prev              188 net/can/j1939/address-claim.c 	if (prev) {
prev              189 net/can/j1939/address-claim.c 		if (ecu->name > prev->name) {
prev              191 net/can/j1939/address-claim.c 			j1939_ecu_put(prev);
prev              195 net/can/j1939/address-claim.c 			j1939_ecu_unmap_locked(prev);
prev              196 net/can/j1939/address-claim.c 			j1939_ecu_put(prev);
prev               32 net/ceph/pagelist.c 		struct page *page = list_entry(pl->head.prev, struct page, lru);
prev              140 net/ceph/pagelist.c 	c->page_lru = pl->head.prev;
prev              158 net/ceph/pagelist.c 	while (pl->head.prev != c->page_lru) {
prev              159 net/ceph/pagelist.c 		page = list_entry(pl->head.prev, struct page, lru);
prev              166 net/ceph/pagelist.c 		page = list_entry(pl->head.prev, struct page, lru);
prev              101 net/core/datagram.c 	if (READ_ONCE(sk->sk_receive_queue.prev) != skb)
prev              152 net/core/datagram.c 	skb->prev->next = nskb;
prev              153 net/core/datagram.c 	skb->next->prev = nskb;
prev              154 net/core/datagram.c 	nskb->prev = skb->prev;
prev              183 net/core/datagram.c 	*last = queue->prev;
prev              282 net/core/datagram.c 	} while (READ_ONCE(sk->sk_receive_queue.prev) != *last);
prev              409 net/core/dev.c 	list_add_rcu(&po->list, elem->list.prev);
prev             3318 net/core/dev.c 		skb->prev = skb;
prev             3331 net/core/dev.c 		tail = skb->prev;
prev              205 net/core/dst.c 		unsigned long prev, new;
prev              211 net/core/dst.c 		prev = cmpxchg(&dst->_metrics, old, new);
prev              213 net/core/dst.c 		if (prev != old) {
prev              215 net/core/dst.c 			p = (struct dst_metrics *)__DST_METRICS_PTR(prev);
prev              216 net/core/dst.c 			if (prev & DST_METRICS_READ_ONLY)
prev              218 net/core/dst.c 		} else if (prev & DST_METRICS_REFCOUNTED) {
prev              231 net/core/dst.c 	unsigned long prev, new;
prev              234 net/core/dst.c 	prev = cmpxchg(&dst->_metrics, old, new);
prev              235 net/core/dst.c 	if (prev == old)
prev             2479 net/core/filter.c 	int prev;
prev             2482 net/core/filter.c 		prev = i;
prev             2484 net/core/filter.c 		msg->sg.data[prev] = msg->sg.data[i];
prev              985 net/core/skbuff.c 	n->next = n->prev = NULL;
prev             3942 net/core/skbuff.c 	segs->prev = tail;
prev               34 net/ieee802154/6lowpan/reassembly.c 			     struct sk_buff *prev,  struct net_device *ldev);
prev              792 net/ieee802154/socket.c 	struct sock *sk, *prev = NULL;
prev              808 net/ieee802154/socket.c 			if (prev) {
prev              813 net/ieee802154/socket.c 					dgram_rcv_skb(prev, clone);
prev              816 net/ieee802154/socket.c 			prev = sk;
prev              820 net/ieee802154/socket.c 	if (prev) {
prev              821 net/ieee802154/socket.c 		dgram_rcv_skb(prev, skb);
prev              268 net/ipv4/cipso_ipv4.c 				__list_del(entry->list.prev, entry->list.next);
prev              270 net/ipv4/cipso_ipv4.c 					   prev_entry->list.prev,
prev              330 net/ipv4/cipso_ipv4.c 		old_entry = list_entry(cipso_v4_cache[bkt].list.prev,
prev             2020 net/ipv4/igmp.c 				struct ip_sf_list *prev = NULL;
prev             2025 net/ipv4/igmp.c 					prev = dpsf;
prev             2028 net/ipv4/igmp.c 					if (prev)
prev             2029 net/ipv4/igmp.c 						prev->sf_next = dpsf->sf_next;
prev              298 net/ipv4/inet_fragment.c 						struct inet_frag_queue **prev)
prev              305 net/ipv4/inet_fragment.c 		*prev = ERR_PTR(-ENOMEM);
prev              310 net/ipv4/inet_fragment.c 	*prev = rhashtable_lookup_get_insert_key(&fqdir->rhashtable, &q->key,
prev              312 net/ipv4/inet_fragment.c 	if (*prev) {
prev              324 net/ipv4/inet_fragment.c 	struct inet_frag_queue *fq = NULL, *prev;
prev              331 net/ipv4/inet_fragment.c 	prev = rhashtable_lookup(&fqdir->rhashtable, key, fqdir->f->rhash_params);
prev              332 net/ipv4/inet_fragment.c 	if (!prev)
prev              333 net/ipv4/inet_fragment.c 		fq = inet_frag_create(fqdir, key, &prev);
prev              334 net/ipv4/inet_fragment.c 	if (!IS_ERR_OR_NULL(prev)) {
prev              335 net/ipv4/inet_fragment.c 		fq = prev;
prev              513 net/ipv4/inet_fragment.c 				fp->prev = NULL;
prev              540 net/ipv4/inet_fragment.c 	head->prev = NULL;
prev             1468 net/ipv4/route.c 	struct rtable *orig, *prev, **p;
prev             1482 net/ipv4/route.c 	prev = cmpxchg(p, orig, rt);
prev             1483 net/ipv4/route.c 	if (prev == orig) {
prev              218 net/ipv4/tcp_cong.c 	const struct tcp_congestion_ops *prev;
prev              228 net/ipv4/tcp_cong.c 		prev = xchg(&net->ipv4.tcp_congestion_control, ca);
prev              229 net/ipv4/tcp_cong.c 		if (prev)
prev              230 net/ipv4/tcp_cong.c 			module_put(prev->owner);
prev             1282 net/ipv4/tcp_input.c static bool tcp_shifted_skb(struct sock *sk, struct sk_buff *prev,
prev             1308 net/ipv4/tcp_input.c 	TCP_SKB_CB(prev)->end_seq += shifted;
prev             1311 net/ipv4/tcp_input.c 	tcp_skb_pcount_add(prev, pcount);
prev             1320 net/ipv4/tcp_input.c 	if (!TCP_SKB_CB(prev)->tcp_gso_size)
prev             1321 net/ipv4/tcp_input.c 		TCP_SKB_CB(prev)->tcp_gso_size = mss;
prev             1328 net/ipv4/tcp_input.c 	TCP_SKB_CB(prev)->sacked |= (TCP_SKB_CB(skb)->sacked & TCPCB_EVER_RETRANS);
prev             1339 net/ipv4/tcp_input.c 		tp->retransmit_skb_hint = prev;
prev             1341 net/ipv4/tcp_input.c 		tp->lost_skb_hint = prev;
prev             1342 net/ipv4/tcp_input.c 		tp->lost_cnt_hint -= tcp_skb_pcount(prev);
prev             1345 net/ipv4/tcp_input.c 	TCP_SKB_CB(prev)->tcp_flags |= TCP_SKB_CB(skb)->tcp_flags;
prev             1346 net/ipv4/tcp_input.c 	TCP_SKB_CB(prev)->eor = TCP_SKB_CB(skb)->eor;
prev             1348 net/ipv4/tcp_input.c 		TCP_SKB_CB(prev)->end_seq++;
prev             1353 net/ipv4/tcp_input.c 	tcp_skb_collapse_tstamp(prev, skb);
prev             1354 net/ipv4/tcp_input.c 	if (unlikely(TCP_SKB_CB(prev)->tx.delivered_mstamp))
prev             1355 net/ipv4/tcp_input.c 		TCP_SKB_CB(prev)->tx.delivered_mstamp = 0;
prev             1402 net/ipv4/tcp_input.c 	struct sk_buff *prev;
prev             1419 net/ipv4/tcp_input.c 	prev = skb_rb_prev(skb);
prev             1420 net/ipv4/tcp_input.c 	if (!prev)
prev             1423 net/ipv4/tcp_input.c 	if ((TCP_SKB_CB(prev)->sacked & TCPCB_TAGBITS) != TCPCB_SACKED_ACKED)
prev             1426 net/ipv4/tcp_input.c 	if (!tcp_skb_can_collapse_to(prev))
prev             1440 net/ipv4/tcp_input.c 		if (mss != tcp_skb_seglen(prev))
prev             1481 net/ipv4/tcp_input.c 		if (mss != tcp_skb_seglen(prev))
prev             1498 net/ipv4/tcp_input.c 	if (!tcp_skb_shift(prev, skb, pcount, len))
prev             1500 net/ipv4/tcp_input.c 	if (!tcp_shifted_skb(sk, prev, skb, state, pcount, len, mss, dup_sack))
prev             1506 net/ipv4/tcp_input.c 	skb = skb_rb_next(prev);
prev             1517 net/ipv4/tcp_input.c 	if (tcp_skb_shift(prev, skb, pcount, len))
prev             1518 net/ipv4/tcp_input.c 		tcp_shifted_skb(sk, prev, skb, state, pcount,
prev             1522 net/ipv4/tcp_input.c 	return prev;
prev             5077 net/ipv4/tcp_input.c 	struct rb_node *node, *prev;
prev             5087 net/ipv4/tcp_input.c 		prev = rb_prev(node);
prev             5091 net/ipv4/tcp_input.c 		if (!prev || goal <= 0) {
prev             5098 net/ipv4/tcp_input.c 		node = prev;
prev             5100 net/ipv4/tcp_input.c 	tp->ooo_last_skb = rb_to_skb(prev);
prev             3432 net/ipv6/addrconf.c 		struct fib6_info *f6i, *prev;
prev             3441 net/ipv6/addrconf.c 		prev = ifp->rt;
prev             3445 net/ipv6/addrconf.c 		fib6_info_release(prev);
prev              230 net/ipv6/calipso.c 				__list_del(entry->list.prev, entry->list.next);
prev              232 net/ipv6/calipso.c 					   prev_entry->list.prev,
prev              294 net/ipv6/calipso.c 		old_entry = list_entry(calipso_cache[bkt].list.prev,
prev              243 net/ipv6/ila/ila_xlat.c 		struct ila_map *tila = head, *prev = NULL;
prev              254 net/ipv6/ila/ila_xlat.c 			prev = tila;
prev              259 net/ipv6/ila/ila_xlat.c 		if (prev) {
prev              262 net/ipv6/ila/ila_xlat.c 			rcu_assign_pointer(prev->next, ila);
prev              286 net/ipv6/ila/ila_xlat.c 	struct ila_map *ila, *head, *prev;
prev              296 net/ipv6/ila/ila_xlat.c 	prev = NULL;
prev              300 net/ipv6/ila/ila_xlat.c 			prev = ila;
prev              308 net/ipv6/ila/ila_xlat.c 		if (prev) {
prev              310 net/ipv6/ila/ila_xlat.c 			rcu_assign_pointer(prev->next, ila->next);
prev             1590 net/ipv6/ip6_fib.c 	struct fib6_node *fn, *prev = NULL;
prev             1617 net/ipv6/ip6_fib.c 			prev = fn;
prev             1632 net/ipv6/ip6_fib.c 		return prev;
prev             2281 net/ipv6/mcast.c 				struct ip6_sf_list *prev = NULL;
prev             2288 net/ipv6/mcast.c 					prev = dpsf;
prev             2291 net/ipv6/mcast.c 					if (prev)
prev             2292 net/ipv6/mcast.c 						prev->sf_next = dpsf->sf_next;
prev              174 net/ipv6/netfilter/nf_conntrack_reasm.c 	struct sk_buff *prev;
prev              255 net/ipv6/netfilter/nf_conntrack_reasm.c 	prev = fq->q.fragments_tail;
prev              289 net/ipv6/netfilter/nf_conntrack_reasm.c 		err = nf_ct_frag6_reasm(fq, skb, prev, dev);
prev             1411 net/ipv6/route.c 		struct rt6_info *prev, **p;
prev             1414 net/ipv6/route.c 		prev = xchg(p, NULL);
prev             1415 net/ipv6/route.c 		if (prev) {
prev             1416 net/ipv6/route.c 			dst_dev_put(&prev->dst);
prev             1417 net/ipv6/route.c 			dst_release(&prev->dst);
prev             1429 net/ipv6/route.c 	struct rt6_info *pcpu_rt, *prev, **p;
prev             1436 net/ipv6/route.c 	prev = cmpxchg(p, NULL, pcpu_rt);
prev             1437 net/ipv6/route.c 	BUG_ON(prev);
prev             4403 net/mac80211/rx.c 	struct ieee80211_sub_if_data *prev;
prev             4475 net/mac80211/rx.c 	prev = NULL;
prev             4491 net/mac80211/rx.c 		if (!prev) {
prev             4492 net/mac80211/rx.c 			prev = sdata;
prev             4496 net/mac80211/rx.c 		rx.sta = sta_info_get_bss(prev, hdr->addr2);
prev             4497 net/mac80211/rx.c 		rx.sdata = prev;
prev             4500 net/mac80211/rx.c 		prev = sdata;
prev             4503 net/mac80211/rx.c 	if (prev) {
prev             4504 net/mac80211/rx.c 		rx.sta = sta_info_get_bss(prev, hdr->addr2);
prev             4505 net/mac80211/rx.c 		rx.sdata = prev;
prev             3913 net/mac80211/tx.c 		skb->prev = NULL;
prev              190 net/netfilter/ipset/ip_set_list_set.c 	struct set_elem *e, *next, *prev = NULL;
prev              198 net/netfilter/ipset/ip_set_list_set.c 			prev = e;
prev              209 net/netfilter/ipset/ip_set_list_set.c 			ret = prev && prev->id == d->refid;
prev              237 net/netfilter/ipset/ip_set_list_set.c 	struct set_elem *e, *n, *prev, *next;
prev              241 net/netfilter/ipset/ip_set_list_set.c 	n = prev = next = NULL;
prev              253 net/netfilter/ipset/ip_set_list_set.c 			prev = e;
prev              258 net/netfilter/ipset/ip_set_list_set.c 	    (d->before < 0 && !prev))
prev              284 net/netfilter/ipset/ip_set_list_set.c 		if (prev->list.prev != &map->members)
prev              285 net/netfilter/ipset/ip_set_list_set.c 			n = list_prev_entry(prev, list);
prev              304 net/netfilter/ipset/ip_set_list_set.c 	else if (prev)
prev              305 net/netfilter/ipset/ip_set_list_set.c 		list_add_rcu(&e->list, &prev->list);
prev              319 net/netfilter/ipset/ip_set_list_set.c 	struct set_elem *e, *next, *prev = NULL;
prev              326 net/netfilter/ipset/ip_set_list_set.c 			prev = e;
prev              336 net/netfilter/ipset/ip_set_list_set.c 			if (!prev || prev->id != d->refid)
prev               43 net/netfilter/ipvs/ip_vs_rr.c 		svc->sched_data = p->next->prev;
prev              122 net/netfilter/nft_set_hash.c 	struct nft_rhash_elem *he, *prev;
prev              137 net/netfilter/nft_set_hash.c 	prev = rhashtable_lookup_get_insert_key(&priv->ht, &arg, &he->node,
prev              139 net/netfilter/nft_set_hash.c 	if (IS_ERR(prev))
prev              143 net/netfilter/nft_set_hash.c 	if (prev) {
prev              145 net/netfilter/nft_set_hash.c 		he = prev;
prev              169 net/netfilter/nft_set_hash.c 	struct nft_rhash_elem *prev;
prev              171 net/netfilter/nft_set_hash.c 	prev = rhashtable_lookup_get_insert_key(&priv->ht, &arg, &he->node,
prev              173 net/netfilter/nft_set_hash.c 	if (IS_ERR(prev))
prev              174 net/netfilter/nft_set_hash.c 		return PTR_ERR(prev);
prev              175 net/netfilter/nft_set_hash.c 	if (prev) {
prev              176 net/netfilter/nft_set_hash.c 		*ext = &prev->ext;
prev              275 net/netfilter/x_tables.c 			  u8 __user *prev, u8 rev)
prev              281 net/netfilter/x_tables.c 	if (put_user(rev, prev))
prev               98 net/netfilter/xt_hashlimit.c 		unsigned long prev;	/* last modification */
prev              532 net/netfilter/xt_hashlimit.c 	unsigned long delta = now - dh->rateinfo.prev;
prev              544 net/netfilter/xt_hashlimit.c 		dh->rateinfo.prev = now;
prev              553 net/netfilter/xt_hashlimit.c 	dh->rateinfo.prev = now;
prev              576 net/netfilter/xt_hashlimit.c 	dh->rateinfo.prev = jiffies;
prev               19 net/netfilter/xt_limit.c 	unsigned long prev;
prev               72 net/netfilter/xt_limit.c 	priv->credit += (now - xchg(&priv->prev, now)) * CREDITS_PER_JIFFY;
prev              119 net/netfilter/xt_limit.c 	priv->prev = jiffies;
prev              142 net/netfilter/xt_limit.c 	compat_ulong_t prev;
prev              157 net/netfilter/xt_limit.c 		.prev		= cm->prev | (unsigned long)cm->master << 32,
prev              171 net/netfilter/xt_limit.c 		.prev		= m->prev,
prev              175 net/netfilter/xt_limit.c 		.master		= m->prev >> 32,
prev              194 net/netfilter/xt_limit.c 	.usersize         = offsetof(struct xt_rateinfo, prev),
prev              165 net/netlabel/netlabel_addrlist.c 				       iter->list.prev,
prev              203 net/netlabel/netlabel_addrlist.c 				       iter->list.prev,
prev              560 net/netlabel/netlabel_kapi.c 	struct netlbl_lsm_catmap *prev = NULL;
prev              567 net/netlabel/netlabel_kapi.c 		prev = iter;
prev              587 net/netlabel/netlabel_kapi.c 	if (prev == NULL) {
prev              591 net/netlabel/netlabel_kapi.c 		iter->next = prev->next;
prev              592 net/netlabel/netlabel_kapi.c 		prev->next = iter;
prev              666 net/netlabel/netlabel_kapi.c 	struct netlbl_lsm_catmap *prev = NULL;
prev              692 net/netlabel/netlabel_kapi.c 		if (prev && idx == 0 && bit == 0)
prev              693 net/netlabel/netlabel_kapi.c 			return prev->startbit + NETLBL_CATMAP_SIZE - 1;
prev              700 net/netlabel/netlabel_kapi.c 			prev = iter;
prev             1114 net/netlink/genetlink.c 	struct net *net, *prev = NULL;
prev             1119 net/netlink/genetlink.c 		if (prev) {
prev             1125 net/netlink/genetlink.c 			err = nlmsg_multicast(prev->genl_sock, tmp,
prev             1133 net/netlink/genetlink.c 		prev = net;
prev             1136 net/netlink/genetlink.c 	err = nlmsg_multicast(prev->genl_sock, skb, portid, group, flags);
prev              765 net/packet/af_packet.c 	last_pkt = (struct tpacket3_hdr *)pkc1->prev;
prev              834 net/packet/af_packet.c 	pkc1->prev = pkc1->nxt_offset;
prev              996 net/packet/af_packet.c 	pkc->prev = curr;
prev             1101 net/packet/af_packet.c 	unsigned int prev;
prev             1103 net/packet/af_packet.c 		prev = rb->prb_bdqc.kactive_blk_num-1;
prev             1105 net/packet/af_packet.c 		prev = rb->prb_bdqc.knum_blocks-1;
prev             1106 net/packet/af_packet.c 	return prev;
prev               38 net/packet/internal.h 	char		*prev;
prev              666 net/rds/af_rds.c 		 &rs->rs_item != rs->rs_item.prev));
prev               82 net/rds/ib_recv.c 	struct list_head *from_last = from->prev;
prev              490 net/rds/ib_send.c 	struct rds_ib_send_work *prev;
prev              620 net/rds/ib_send.c 	prev = NULL;
prev              686 net/rds/ib_send.c 		if (prev)
prev              687 net/rds/ib_send.c 			prev->s_wr.next = &send->s_wr;
prev              688 net/rds/ib_send.c 		prev = send;
prev              704 net/rds/ib_send.c 		prev->s_op = ic->i_data_op;
prev              705 net/rds/ib_send.c 		prev->s_wr.send_flags |= IB_SEND_SOLICITED;
prev              706 net/rds/ib_send.c 		if (!(prev->s_wr.send_flags & IB_SEND_SIGNALED))
prev              707 net/rds/ib_send.c 			nr_sig += rds_ib_set_wr_signal_state(ic, prev, true);
prev              733 net/rds/ib_send.c 		if (prev->s_op) {
prev              734 net/rds/ib_send.c 			ic->i_data_op = prev->s_op;
prev              735 net/rds/ib_send.c 			prev->s_op = NULL;
prev              844 net/rds/ib_send.c 	struct rds_ib_send_work *prev;
prev              890 net/rds/ib_send.c 	prev = NULL;
prev              917 net/rds/ib_send.c 		if (prev)
prev              918 net/rds/ib_send.c 			prev->s_rdma_wr.wr.next = &send->s_rdma_wr.wr;
prev              939 net/rds/ib_send.c 		prev = send;
prev              946 net/rds/ib_send.c 		prev->s_op = op;
prev              311 net/rfkill/core.c 	bool prev, curr;
prev              326 net/rfkill/core.c 	prev = rfkill->state & RFKILL_BLOCK_SW;
prev              328 net/rfkill/core.c 	if (prev)
prev              363 net/rfkill/core.c 	if (prev != curr)
prev              528 net/rfkill/core.c 	bool ret, prev;
prev              533 net/rfkill/core.c 	prev = !!(rfkill->state & RFKILL_BLOCK_HW);
prev              544 net/rfkill/core.c 	if (rfkill->registered && prev != blocked)
prev              568 net/rfkill/core.c 	bool prev, hwblock;
prev              573 net/rfkill/core.c 	prev = !!(rfkill->state & RFKILL_BLOCK_SW);
prev              582 net/rfkill/core.c 	if (prev != blocked && !hwblock)
prev              476 net/rxrpc/call_object.c 	call->recvmsg_link.prev = NULL;
prev              217 net/sched/sch_etf.c 		skb->prev = NULL;
prev              241 net/sched/sch_etf.c 	skb->prev = NULL;
prev              564 net/sched/sch_generic.c 		.prev = (struct sk_buff *)&noop_qdisc.gso_skb,
prev              570 net/sched/sch_generic.c 		.prev = (struct sk_buff *)&noop_qdisc.skb_bad_txq,
prev              191 net/sched/sch_hhf.c 		u32 prev = flow->hit_timestamp + q->hhf_evict_timeout;
prev              193 net/sched/sch_hhf.c 		if (hhf_time_before(prev, now)) {
prev              221 net/sched/sch_hhf.c 			u32 prev = flow->hit_timestamp + q->hhf_evict_timeout;
prev              223 net/sched/sch_hhf.c 			if (hhf_time_before(prev, now))
prev              255 net/sched/sch_hhf.c 	u32 prev;
prev              259 net/sched/sch_hhf.c 	prev = q->hhf_arrays_reset_timestamp + q->hhf_reset_timeout;
prev              260 net/sched/sch_hhf.c 	if (hhf_time_before(prev, now)) {
prev              448 net/sched/sch_netem.c 	skb->prev = NULL;
prev              704 net/sched/sch_netem.c 			skb->prev = NULL;
prev               97 net/sched/sch_sfq.c 	sfq_index	prev;
prev              213 net/sched/sch_sfq.c 	slot->dep.prev = p;
prev              216 net/sched/sch_sfq.c 	sfq_dep_head(q, n)->prev = x;
prev              222 net/sched/sch_sfq.c 		p = q->slots[x].dep.prev;	\
prev              224 net/sched/sch_sfq.c 		sfq_dep_head(q, n)->prev = p;	\
prev              261 net/sched/sch_sfq.c 	slot->skblist_prev = skb->prev;
prev              262 net/sched/sch_sfq.c 	skb->prev->next = (struct sk_buff *)slot;
prev              263 net/sched/sch_sfq.c 	skb->next = skb->prev = NULL;
prev              273 net/sched/sch_sfq.c 	skb->next->prev = (struct sk_buff *)slot;
prev              274 net/sched/sch_sfq.c 	skb->next = skb->prev = NULL;
prev              287 net/sched/sch_sfq.c 	skb->prev = slot->skblist_prev;
prev              747 net/sched/sch_sfq.c 		q->dep[i].prev = i + SFQ_MAX_FLOWS;
prev              133 net/sched/sch_teql.c 	struct Qdisc *q, *prev;
prev              137 net/sched/sch_teql.c 	prev = master->slaves;
prev              138 net/sched/sch_teql.c 	if (prev) {
prev              140 net/sched/sch_teql.c 			q = NEXT_SLAVE(prev);
prev              142 net/sched/sch_teql.c 				NEXT_SLAVE(prev) = NEXT_SLAVE(q);
prev              162 net/sched/sch_teql.c 		} while ((prev = q) != master->slaves);
prev              332 net/sctp/outqueue.c 			list_add(new, pos->prev);
prev              428 net/sctp/stream_interleave.c 	event_list = (struct sk_buff_head *)sctp_event2skb(event)->prev;
prev              133 net/sctp/stream_sched_prio.c 		list_add(&soute->prio_list, prio_head->next->prio_list.prev);
prev              142 net/sctp/stream_sched_prio.c 			list_add(&prio_head->prio_sched, prio->prio_sched.prev);
prev              758 net/sctp/ulpqueue.c 	event_list = (struct sk_buff_head *) sctp_event2skb(event)->prev;
prev              667 net/sunrpc/cache.c 			discard = list_entry(cache_defer_list.prev,
prev              150 net/sunrpc/sched.c 			task->u.tk_wait.list.prev = NULL;
prev              167 net/sunrpc/sched.c 	if (task->u.tk_wait.list.prev == NULL) {
prev              616 net/sunrpc/svc_xprt.c 			xprt = list_entry(serv->sv_tempsocks.prev,
prev              280 net/sunrpc/xprtmultipath.c 	struct rpc_xprt *pos, *prev = NULL;
prev              284 net/sunrpc/xprtmultipath.c 		if (cur == prev)
prev              288 net/sunrpc/xprtmultipath.c 		prev = pos;
prev              528 net/sunrpc/xprtrdma/frwr_ops.c 	struct ib_send_wr *first, **prev, *last;
prev              540 net/sunrpc/xprtrdma/frwr_ops.c 	prev = &first;
prev              557 net/sunrpc/xprtrdma/frwr_ops.c 		*prev = last;
prev              558 net/sunrpc/xprtrdma/frwr_ops.c 		prev = &last->next;
prev              633 net/sunrpc/xprtrdma/frwr_ops.c 	struct ib_send_wr *first, *last, **prev;
prev              643 net/sunrpc/xprtrdma/frwr_ops.c 	prev = &first;
prev              660 net/sunrpc/xprtrdma/frwr_ops.c 		*prev = last;
prev              661 net/sunrpc/xprtrdma/frwr_ops.c 		prev = &last->next;
prev              390 net/tipc/group.c 	u16 prev = grp->bc_snd_nxt - 1;
prev              399 net/tipc/group.c 			m->bc_acked = prev;
prev              326 net/tipc/monitor.c 	struct tipc_peer *peer, *prev, *head;
prev              332 net/tipc/monitor.c 	prev = peer_prev(peer);
prev              338 net/tipc/monitor.c 	head = peer_head(prev);
prev              341 net/tipc/monitor.c 	mon_update_neighbors(mon, prev);
prev              360 net/tipc/monitor.c 	struct tipc_peer *cur, *prev, *p;
prev              373 net/tipc/monitor.c 	prev = self;
prev              375 net/tipc/monitor.c 		if ((addr > prev->addr) && (addr < cur->addr))
prev              377 net/tipc/monitor.c 		if (((addr < cur->addr) || (addr > prev->addr)) &&
prev              378 net/tipc/monitor.c 		    (prev->addr > cur->addr))
prev              380 net/tipc/monitor.c 		prev = cur;
prev              359 net/tls/tls_device_fallback.c 	nskb->prev = nskb;
prev             10882 net/wireless/nl80211.c 	s32 prev = S32_MIN;
prev             10886 net/wireless/nl80211.c 		if (thresholds[i] > 0 || thresholds[i] <= prev)
prev             10889 net/wireless/nl80211.c 		prev = thresholds[i];
prev              694 net/xfrm/xfrm_policy.c 	struct xfrm_pol_inexact_bin *bin, *prev;
prev              721 net/xfrm/xfrm_policy.c 	prev = rhashtable_lookup_get_insert_key(&xfrm_policy_inexact_table,
prev              724 net/xfrm/xfrm_policy.c 	if (!prev) {
prev              731 net/xfrm/xfrm_policy.c 	return IS_ERR(prev) ? NULL : prev;
prev               37 samples/bpf/test_lru_dist.c 	struct list_head *next, *prev;
prev               43 samples/bpf/test_lru_dist.c 	list->prev = list;
prev               52 samples/bpf/test_lru_dist.c 			      struct list_head *prev,
prev               55 samples/bpf/test_lru_dist.c 	next->prev = new;
prev               57 samples/bpf/test_lru_dist.c 	new->prev = prev;
prev               58 samples/bpf/test_lru_dist.c 	prev->next = new;
prev               66 samples/bpf/test_lru_dist.c static inline void __list_del(struct list_head *prev, struct list_head *next)
prev               68 samples/bpf/test_lru_dist.c 	next->prev = prev;
prev               69 samples/bpf/test_lru_dist.c 	prev->next = next;
prev               74 samples/bpf/test_lru_dist.c 	__list_del(entry->prev, entry->next);
prev               87 samples/bpf/test_lru_dist.c 	list_entry((ptr)->prev, type, member)
prev               47 samples/bpf/xdp1_user.c 	__u64 values[nr_cpus], prev[UINT8_MAX] = { 0 };
prev               61 samples/bpf/xdp1_user.c 			if (sum > prev[key])
prev               63 samples/bpf/xdp1_user.c 				       key, (sum - prev[key]) / interval);
prev               64 samples/bpf/xdp1_user.c 			prev[key] = sum;
prev              311 samples/bpf/xdp_monitor_user.c 		struct record_u64 *rec, *prev;
prev              316 samples/bpf/xdp_monitor_user.c 		prev = &stats_prev->xdp_redirect[rec_i];
prev              317 samples/bpf/xdp_monitor_user.c 		t = calc_period_u64(rec, prev);
prev              321 samples/bpf/xdp_monitor_user.c 			struct u64rec *p = &prev->cpu[i];
prev              329 samples/bpf/xdp_monitor_user.c 		pps = calc_pps_u64(&rec->total, &prev->total, t);
prev              336 samples/bpf/xdp_monitor_user.c 		struct record_u64 *rec, *prev;
prev              341 samples/bpf/xdp_monitor_user.c 		prev = &stats_prev->xdp_exception[rec_i];
prev              342 samples/bpf/xdp_monitor_user.c 		t = calc_period_u64(rec, prev);
prev              346 samples/bpf/xdp_monitor_user.c 			struct u64rec *p = &prev->cpu[i];
prev              353 samples/bpf/xdp_monitor_user.c 		pps = calc_pps_u64(&rec->total, &prev->total, t);
prev              363 samples/bpf/xdp_monitor_user.c 		struct record *rec, *prev;
prev              368 samples/bpf/xdp_monitor_user.c 		prev = &stats_prev->xdp_cpumap_enqueue[to_cpu];
prev              369 samples/bpf/xdp_monitor_user.c 		t = calc_period(rec, prev);
prev              372 samples/bpf/xdp_monitor_user.c 			struct datarec *p = &prev->cpu[i];
prev              385 samples/bpf/xdp_monitor_user.c 		pps = calc_pps(&rec->total, &prev->total, t);
prev              387 samples/bpf/xdp_monitor_user.c 			drop = calc_drop(&rec->total, &prev->total, t);
prev              388 samples/bpf/xdp_monitor_user.c 			info = calc_info(&rec->total, &prev->total, t);
prev              402 samples/bpf/xdp_monitor_user.c 		struct record *rec, *prev;
prev              407 samples/bpf/xdp_monitor_user.c 		prev = &stats_prev->xdp_cpumap_kthread;
prev              408 samples/bpf/xdp_monitor_user.c 		t = calc_period(rec, prev);
prev              411 samples/bpf/xdp_monitor_user.c 			struct datarec *p = &prev->cpu[i];
prev              422 samples/bpf/xdp_monitor_user.c 		pps = calc_pps(&rec->total, &prev->total, t);
prev              423 samples/bpf/xdp_monitor_user.c 		drop = calc_drop(&rec->total, &prev->total, t);
prev              424 samples/bpf/xdp_monitor_user.c 		info = calc_info(&rec->total, &prev->total, t);
prev              434 samples/bpf/xdp_monitor_user.c 		struct record *rec, *prev;
prev              440 samples/bpf/xdp_monitor_user.c 		prev = &stats_prev->xdp_devmap_xmit;
prev              441 samples/bpf/xdp_monitor_user.c 		t = calc_period(rec, prev);
prev              444 samples/bpf/xdp_monitor_user.c 			struct datarec *p = &prev->cpu[i];
prev              460 samples/bpf/xdp_monitor_user.c 		pps = calc_pps(&rec->total, &prev->total, t);
prev              461 samples/bpf/xdp_monitor_user.c 		drop = calc_drop(&rec->total, &prev->total, t);
prev              462 samples/bpf/xdp_monitor_user.c 		info = calc_info(&rec->total, &prev->total, t);
prev              463 samples/bpf/xdp_monitor_user.c 		err  = calc_err(&rec->total, &prev->total, t);
prev              587 samples/bpf/xdp_monitor_user.c 	struct stats_record *rec, *prev;
prev              590 samples/bpf/xdp_monitor_user.c 	prev = alloc_stats_record();
prev              612 samples/bpf/xdp_monitor_user.c 		swap(&prev, &rec);
prev              614 samples/bpf/xdp_monitor_user.c 		stats_print(rec, prev, err_only);
prev              620 samples/bpf/xdp_monitor_user.c 	free_stats_record(prev);
prev              314 samples/bpf/xdp_redirect_cpu_user.c 	struct record *rec, *prev;
prev              331 samples/bpf/xdp_redirect_cpu_user.c 		prev = &stats_prev->rx_cnt;
prev              332 samples/bpf/xdp_redirect_cpu_user.c 		t = calc_period(rec, prev);
prev              335 samples/bpf/xdp_redirect_cpu_user.c 			struct datarec *p = &prev->cpu[i];
prev              346 samples/bpf/xdp_redirect_cpu_user.c 		pps  = calc_pps(&rec->total, &prev->total, t);
prev              347 samples/bpf/xdp_redirect_cpu_user.c 		drop = calc_drop_pps(&rec->total, &prev->total, t);
prev              348 samples/bpf/xdp_redirect_cpu_user.c 		err  = calc_errs_pps(&rec->total, &prev->total, t);
prev              359 samples/bpf/xdp_redirect_cpu_user.c 		prev = &stats_prev->enq[to_cpu];
prev              360 samples/bpf/xdp_redirect_cpu_user.c 		t = calc_period(rec, prev);
prev              363 samples/bpf/xdp_redirect_cpu_user.c 			struct datarec *p = &prev->cpu[i];
prev              376 samples/bpf/xdp_redirect_cpu_user.c 		pps = calc_pps(&rec->total, &prev->total, t);
prev              378 samples/bpf/xdp_redirect_cpu_user.c 			drop = calc_drop_pps(&rec->total, &prev->total, t);
prev              379 samples/bpf/xdp_redirect_cpu_user.c 			err  = calc_errs_pps(&rec->total, &prev->total, t);
prev              396 samples/bpf/xdp_redirect_cpu_user.c 		prev = &stats_prev->kthread;
prev              397 samples/bpf/xdp_redirect_cpu_user.c 		t = calc_period(rec, prev);
prev              400 samples/bpf/xdp_redirect_cpu_user.c 			struct datarec *p = &prev->cpu[i];
prev              411 samples/bpf/xdp_redirect_cpu_user.c 		pps = calc_pps(&rec->total, &prev->total, t);
prev              412 samples/bpf/xdp_redirect_cpu_user.c 		drop = calc_drop_pps(&rec->total, &prev->total, t);
prev              413 samples/bpf/xdp_redirect_cpu_user.c 		err  = calc_errs_pps(&rec->total, &prev->total, t);
prev              425 samples/bpf/xdp_redirect_cpu_user.c 		prev = &stats_prev->redir_err;
prev              426 samples/bpf/xdp_redirect_cpu_user.c 		t = calc_period(rec, prev);
prev              429 samples/bpf/xdp_redirect_cpu_user.c 			struct datarec *p = &prev->cpu[i];
prev              436 samples/bpf/xdp_redirect_cpu_user.c 		pps = calc_pps(&rec->total, &prev->total, t);
prev              437 samples/bpf/xdp_redirect_cpu_user.c 		drop = calc_drop_pps(&rec->total, &prev->total, t);
prev              447 samples/bpf/xdp_redirect_cpu_user.c 		prev = &stats_prev->exception;
prev              448 samples/bpf/xdp_redirect_cpu_user.c 		t = calc_period(rec, prev);
prev              451 samples/bpf/xdp_redirect_cpu_user.c 			struct datarec *p = &prev->cpu[i];
prev              458 samples/bpf/xdp_redirect_cpu_user.c 		pps = calc_pps(&rec->total, &prev->total, t);
prev              459 samples/bpf/xdp_redirect_cpu_user.c 		drop = calc_drop_pps(&rec->total, &prev->total, t);
prev              580 samples/bpf/xdp_redirect_cpu_user.c 	struct stats_record *record, *prev;
prev              583 samples/bpf/xdp_redirect_cpu_user.c 	prev   = alloc_stats_record();
prev              591 samples/bpf/xdp_redirect_cpu_user.c 		swap(&prev, &record);
prev              593 samples/bpf/xdp_redirect_cpu_user.c 		stats_print(record, prev, prog_name);
prev              600 samples/bpf/xdp_redirect_cpu_user.c 	free_stats_record(prev);
prev               66 samples/bpf/xdp_redirect_map_user.c 	__u64 values[nr_cpus], prev[nr_cpus];
prev               68 samples/bpf/xdp_redirect_map_user.c 	memset(prev, 0, sizeof(prev));
prev               78 samples/bpf/xdp_redirect_map_user.c 			sum += (values[i] - prev[i]);
prev               82 samples/bpf/xdp_redirect_map_user.c 		memcpy(prev, values, sizeof(values));
prev               66 samples/bpf/xdp_redirect_user.c 	__u64 values[nr_cpus], prev[nr_cpus];
prev               68 samples/bpf/xdp_redirect_user.c 	memset(prev, 0, sizeof(prev));
prev               78 samples/bpf/xdp_redirect_user.c 			sum += (values[i] - prev[i]);
prev               82 samples/bpf/xdp_redirect_user.c 		memcpy(prev, values, sizeof(values));
prev              520 samples/bpf/xdp_router_ipv4_user.c 	__u64 prev[nr_keys][nr_cpus];
prev              565 samples/bpf/xdp_router_ipv4_user.c 	memset(prev, 0, sizeof(prev));
prev              577 samples/bpf/xdp_router_ipv4_user.c 				sum += (values[i] - prev[key][i]);
prev              581 samples/bpf/xdp_router_ipv4_user.c 			memcpy(prev[key], values, sizeof(values));
prev              348 samples/bpf/xdp_rxq_info_user.c 	struct record *rec, *prev;
prev              367 samples/bpf/xdp_rxq_info_user.c 		prev = &stats_prev->stats;
prev              368 samples/bpf/xdp_rxq_info_user.c 		t = calc_period(rec, prev);
prev              371 samples/bpf/xdp_rxq_info_user.c 			struct datarec *p = &prev->cpu[i];
prev              381 samples/bpf/xdp_rxq_info_user.c 		pps  = calc_pps     (&rec->total, &prev->total, t);
prev              382 samples/bpf/xdp_rxq_info_user.c 		err  = calc_errs_pps(&rec->total, &prev->total, t);
prev              401 samples/bpf/xdp_rxq_info_user.c 		prev = &stats_prev->rxq[rxq];
prev              402 samples/bpf/xdp_rxq_info_user.c 		t = calc_period(rec, prev);
prev              405 samples/bpf/xdp_rxq_info_user.c 			struct datarec *p = &prev->cpu[i];
prev              419 samples/bpf/xdp_rxq_info_user.c 		pps  = calc_pps     (&rec->total, &prev->total, t);
prev              420 samples/bpf/xdp_rxq_info_user.c 		err  = calc_errs_pps(&rec->total, &prev->total, t);
prev              439 samples/bpf/xdp_rxq_info_user.c 	struct stats_record *record, *prev;
prev              442 samples/bpf/xdp_rxq_info_user.c 	prev   = alloc_stats_record();
prev              446 samples/bpf/xdp_rxq_info_user.c 		swap(&prev, &record);
prev              448 samples/bpf/xdp_rxq_info_user.c 		stats_print(record, prev, action, cfg_opt);
prev              453 samples/bpf/xdp_rxq_info_user.c 	free_stats_record(prev);
prev               56 samples/bpf/xdp_tx_iptunnel_user.c 	__u64 values[nr_cpus], prev[nr_protos][nr_cpus];
prev               60 samples/bpf/xdp_tx_iptunnel_user.c 	memset(prev, 0, sizeof(prev));
prev               71 samples/bpf/xdp_tx_iptunnel_user.c 				sum += (values[i] - prev[proto][i]);
prev               76 samples/bpf/xdp_tx_iptunnel_user.c 			memcpy(prev[proto], values, sizeof(values));
prev              183 scripts/dtc/srcpos.c 	srcfile->prev = current_srcfile;
prev              200 scripts/dtc/srcpos.c 	current_srcfile = srcfile->prev;
prev               18 scripts/dtc/srcpos.h 	struct srcfile_state *prev;
prev               25 scripts/kconfig/list.h 	struct list_head *next, *prev;
prev               83 scripts/kconfig/list.h 			      struct list_head *prev,
prev               86 scripts/kconfig/list.h 	next->prev = _new;
prev               88 scripts/kconfig/list.h 	_new->prev = prev;
prev               89 scripts/kconfig/list.h 	prev->next = _new;
prev              102 scripts/kconfig/list.h 	__list_add(_new, head->prev, head);
prev              112 scripts/kconfig/list.h static inline void __list_del(struct list_head *prev, struct list_head *next)
prev              114 scripts/kconfig/list.h 	next->prev = prev;
prev              115 scripts/kconfig/list.h 	prev->next = next;
prev              128 scripts/kconfig/list.h 	__list_del(entry->prev, entry->next);
prev              130 scripts/kconfig/list.h 	entry->prev = (struct list_head*)LIST_POISON2;
prev              461 scripts/kconfig/mconf.c 	list_del(trail.prev);
prev              751 scripts/kconfig/mconf.c 	list_del(trail.prev);
prev              734 scripts/kconfig/menu.c 			jump->index = list_entry(head->prev, struct jump_key,
prev              348 scripts/kconfig/preprocess.c 	char *tmp, *name, *res, *endptr, *prev, *p;
prev              368 scripts/kconfig/preprocess.c 	prev = p = tmp;
prev              389 scripts/kconfig/preprocess.c 			new_argv[new_argc++] = prev;
prev              390 scripts/kconfig/preprocess.c 			prev = p + 1;
prev              399 scripts/kconfig/preprocess.c 	new_argv[new_argc++] = prev;
prev             1008 scripts/kconfig/symbol.c 	struct dep_stack *prev, *next;
prev             1019 scripts/kconfig/symbol.c 	stack->prev = check_top;
prev             1026 scripts/kconfig/symbol.c 	check_top = check_top->prev;
prev             1049 scripts/kconfig/symbol.c 	for (stack = check_top; stack != NULL; stack = stack->prev)
prev              793 scripts/mod/modpost.c 			      char *prev)
prev              800 scripts/mod/modpost.c 	if (prev) {
prev              801 scripts/mod/modpost.c 		size -= prev - modinfo;
prev              802 scripts/mod/modpost.c 		modinfo = next_string(prev, &size);
prev               53 security/apparmor/include/policy.h #define on_list_rcu(X) (!list_empty(X) && (X)->prev != LIST_POISON2)
prev              226 security/apparmor/lib.c 	bool prev = false;
prev              231 security/apparmor/lib.c 			if (!prev) {
prev              232 security/apparmor/lib.c 				prev = true;
prev              108 security/selinux/ibpkey.c 				sel_ib_pkey_hash[idx].list.prev,
prev              168 security/selinux/netnode.c 			rcu_dereference_protected(sel_netnode_hash[idx].list.prev,
prev              117 security/selinux/netport.c 				sel_netport_hash[idx].list.prev,
prev               70 security/selinux/ss/avtab.c 		  struct avtab_node *prev, struct avtab_node *cur,
prev               92 security/selinux/ss/avtab.c 	if (prev) {
prev               93 security/selinux/ss/avtab.c 		newnode->next = prev->next;
prev               94 security/selinux/ss/avtab.c 		prev->next = newnode;
prev              109 security/selinux/ss/avtab.c 	struct avtab_node *prev, *cur, *newnode;
prev              116 security/selinux/ss/avtab.c 	for (prev = NULL, cur = h->htable[hvalue];
prev              118 security/selinux/ss/avtab.c 	     prev = cur, cur = cur->next) {
prev              139 security/selinux/ss/avtab.c 	newnode = avtab_insert_node(h, hvalue, prev, cur, key, datum);
prev              154 security/selinux/ss/avtab.c 	struct avtab_node *prev, *cur;
prev              160 security/selinux/ss/avtab.c 	for (prev = NULL, cur = h->htable[hvalue];
prev              162 security/selinux/ss/avtab.c 	     prev = cur, cur = cur->next) {
prev              178 security/selinux/ss/avtab.c 	return avtab_insert_node(h, hvalue, prev, cur, key, datum);
prev               54 security/selinux/ss/ebitmap.c 	struct ebitmap_node *n, *new, *prev;
prev               58 security/selinux/ss/ebitmap.c 	prev = NULL;
prev               68 security/selinux/ss/ebitmap.c 		if (prev)
prev               69 security/selinux/ss/ebitmap.c 			prev->next = new;
prev               72 security/selinux/ss/ebitmap.c 		prev = new;
prev              261 security/selinux/ss/ebitmap.c 	struct ebitmap_node *n, *prev, *new;
prev              263 security/selinux/ss/ebitmap.c 	prev = NULL;
prev              284 security/selinux/ss/ebitmap.c 					if (prev)
prev              285 security/selinux/ss/ebitmap.c 						e->highbit = prev->startbit
prev              290 security/selinux/ss/ebitmap.c 				if (prev)
prev              291 security/selinux/ss/ebitmap.c 					prev->next = n->next;
prev              298 security/selinux/ss/ebitmap.c 		prev = n;
prev              316 security/selinux/ss/ebitmap.c 	if (prev) {
prev              317 security/selinux/ss/ebitmap.c 		new->next = prev->next;
prev              318 security/selinux/ss/ebitmap.c 		prev->next = new;
prev               45 security/selinux/ss/hashtab.c 	struct hashtab_node *prev, *cur, *newnode;
prev               53 security/selinux/ss/hashtab.c 	prev = NULL;
prev               56 security/selinux/ss/hashtab.c 		prev = cur;
prev               68 security/selinux/ss/hashtab.c 	if (prev) {
prev               69 security/selinux/ss/hashtab.c 		newnode->next = prev->next;
prev               70 security/selinux/ss/hashtab.c 		prev->next = newnode;
prev               38 security/selinux/ss/mls.c 	int i, l, len, head, prev;
prev               53 security/selinux/ss/mls.c 		prev = -2;
prev               56 security/selinux/ss/mls.c 			if (i - prev > 1) {
prev               58 security/selinux/ss/mls.c 				if (head != prev) {
prev               59 security/selinux/ss/mls.c 					nm = sym_name(p, SYM_CATS, prev);
prev               66 security/selinux/ss/mls.c 			prev = i;
prev               68 security/selinux/ss/mls.c 		if (prev != head) {
prev               69 security/selinux/ss/mls.c 			nm = sym_name(p, SYM_CATS, prev);
prev               94 security/selinux/ss/mls.c 	int i, l, head, prev;
prev              113 security/selinux/ss/mls.c 		prev = -2;
prev              116 security/selinux/ss/mls.c 			if (i - prev > 1) {
prev              118 security/selinux/ss/mls.c 				if (prev != head) {
prev              119 security/selinux/ss/mls.c 					if (prev - head > 1)
prev              123 security/selinux/ss/mls.c 					nm = sym_name(p, SYM_CATS, prev);
prev              127 security/selinux/ss/mls.c 				if (prev < 0)
prev              136 security/selinux/ss/mls.c 			prev = i;
prev              139 security/selinux/ss/mls.c 		if (prev != head) {
prev              140 security/selinux/ss/mls.c 			if (prev - head > 1)
prev              144 security/selinux/ss/mls.c 			nm = sym_name(p, SYM_CATS, prev);
prev             1914 security/smack/smackfs.c 		last = public->prev;
prev             1917 security/smack/smackfs.c 		private->prev->next = public;
prev             1918 security/smack/smackfs.c 		private->next->prev = public;
prev             1920 security/smack/smackfs.c 		public->prev = private->prev;
prev             1927 security/smack/smackfs.c 		private->prev = last;
prev             1928 security/smack/smackfs.c 		first->prev = private;
prev              474 security/tomoyo/gc.c 	list_add_rcu(element, element->prev);
prev              558 sound/core/oss/pcm_oss.c 	plugin->prev = NULL;
prev              560 sound/core/oss/pcm_oss.c 		runtime->oss.plugin_first->prev = plugin;
prev              573 sound/core/oss/pcm_oss.c 	plugin->prev = runtime->oss.plugin_last;
prev              776 sound/core/oss/pcm_oss.c 	unsigned int rate, prev;
prev              800 sound/core/oss/pcm_oss.c 		prev = rate;
prev              802 sound/core/oss/pcm_oss.c 		if (rate <= prev)
prev              123 sound/core/oss/pcm_plugin.c 		while (plugin->prev) {
prev              128 sound/core/oss/pcm_plugin.c 			plugin = plugin->prev;
prev              214 sound/core/oss/pcm_plugin.c 			plugin_prev = plugin->prev;
prev              269 sound/core/oss/pcm_plugin.c 			plugin_prev = plugin->prev;
prev               59 sound/core/oss/pcm_plugin.h 	struct snd_pcm_plugin *prev;
prev               75 sound/core/pcm.c 			list_add(&newpcm->list, pcm->list.prev);
prev              640 sound/core/pcm.c 	struct snd_pcm_substream *substream, *prev;
prev              664 sound/core/pcm.c 	prev = NULL;
prev              665 sound/core/pcm.c 	for (idx = 0, prev = NULL; idx < substream_count; idx++) {
prev              675 sound/core/pcm.c 		if (prev == NULL)
prev              678 sound/core/pcm.c 			prev->next = substream;
prev              685 sound/core/pcm.c 				if (prev == NULL)
prev              688 sound/core/pcm.c 					prev->next = NULL;
prev              697 sound/core/pcm.c 		prev = substream;
prev               30 sound/core/pcm_lib.c #define trace_applptr(substream, prev, curr)
prev               34 sound/core/pcm_native.c #define trace_hw_mask_param(substream, type, index, prev, curr)
prev               35 sound/core/pcm_native.c #define trace_hw_interval_param(substream, type, index, prev, curr)
prev               29 sound/core/pcm_param_trace.h 	TP_PROTO(struct snd_pcm_substream *substream, snd_pcm_hw_param_t type, int index, const struct snd_mask *prev, const struct snd_mask *curr),
prev               30 sound/core/pcm_param_trace.h 	TP_ARGS(substream, type, index, prev, curr),
prev               50 sound/core/pcm_param_trace.h 		memcpy(__entry->prev_bits, prev->bits, sizeof(__u32) * 8);
prev               69 sound/core/pcm_param_trace.h 	TP_PROTO(struct snd_pcm_substream *substream, snd_pcm_hw_param_t type, int index, const struct snd_interval *prev, const struct snd_interval *curr),
prev               70 sound/core/pcm_param_trace.h 	TP_ARGS(substream, type, index, prev, curr),
prev              100 sound/core/pcm_param_trace.h 		__entry->prev_min = prev->min;
prev              101 sound/core/pcm_param_trace.h 		__entry->prev_max = prev->max;
prev              102 sound/core/pcm_param_trace.h 		__entry->prev_openmin = prev->openmin;
prev              103 sound/core/pcm_param_trace.h 		__entry->prev_openmax = prev->openmax;
prev              104 sound/core/pcm_param_trace.h 		__entry->prev_integer = prev->integer;
prev              105 sound/core/pcm_param_trace.h 		__entry->prev_empty = prev->empty;
prev              107 sound/core/pcm_trace.h 	TP_PROTO(struct snd_pcm_substream *substream, snd_pcm_uframes_t prev, snd_pcm_uframes_t curr),
prev              108 sound/core/pcm_trace.h 	TP_ARGS(substream, prev, curr),
prev              114 sound/core/pcm_trace.h 		__field( snd_pcm_uframes_t, prev )
prev              125 sound/core/pcm_trace.h 		__entry->prev = (prev);
prev              136 sound/core/pcm_trace.h 		__entry->prev,
prev              134 sound/core/seq/seq_prioq.c 	struct snd_seq_event_cell *cur, *prev;
prev              164 sound/core/seq/seq_prioq.c 	prev = NULL;		/* previous cell */
prev              179 sound/core/seq/seq_prioq.c 		prev = cur;
prev              189 sound/core/seq/seq_prioq.c 	if (prev != NULL)
prev              190 sound/core/seq/seq_prioq.c 		prev->next = cell;
prev              279 sound/core/seq/seq_prioq.c 	struct snd_seq_event_cell *prev = NULL;
prev              292 sound/core/seq/seq_prioq.c 				prev->next = cell->next;
prev              314 sound/core/seq/seq_prioq.c 			prev = cell;
prev              389 sound/core/seq/seq_prioq.c 	struct snd_seq_event_cell *prev = NULL;
prev              405 sound/core/seq/seq_prioq.c 				prev->next = cell->next;
prev              422 sound/core/seq/seq_prioq.c 			prev = cell;
prev             1298 sound/core/timer.c 	int prev;
prev             1302 sound/core/timer.c 		prev = tu->qtail == 0 ? tu->queue_size - 1 : tu->qtail - 1;
prev             1303 sound/core/timer.c 		r = &tu->queue[prev];
prev             1376 sound/core/timer.c 	int prev, append = 0;
prev             1406 sound/core/timer.c 		prev = tu->qtail == 0 ? tu->queue_size - 1 : tu->qtail - 1;
prev             1407 sound/core/timer.c 		r = &tu->tqueue[prev];
prev               39 sound/isa/gus/gus_mem.c 			nblock->prev = pblock->prev;
prev               41 sound/isa/gus/gus_mem.c 			pblock->prev = nblock;
prev               45 sound/isa/gus/gus_mem.c 				nblock->prev->next = nblock;
prev               53 sound/isa/gus/gus_mem.c 		nblock->prev = NULL;
prev               56 sound/isa/gus/gus_mem.c 		nblock->prev = alloc->last;
prev               73 sound/isa/gus/gus_mem.c 			block->next->prev = NULL;
prev               75 sound/isa/gus/gus_mem.c 		block->prev->next = block->next;
prev               77 sound/isa/gus/gus_mem.c 			block->next->prev = block->prev;
prev               80 sound/isa/gus/gus_mem.c 		alloc->last = block->prev;
prev               81 sound/isa/gus/gus_mem.c 		if (block->prev)
prev               82 sound/isa/gus/gus_mem.c 			block->prev->next = NULL;
prev               84 sound/isa/gus/gus_mem.c 		block->next->prev = block->prev;
prev               85 sound/isa/gus/gus_mem.c 		if (block->prev)
prev               86 sound/isa/gus/gus_mem.c 			block->prev->next = block->next;
prev              139 sound/isa/gus/gus_mem.c 	block->prev = block->next = NULL;
prev              261 sound/pci/au88x0/au88x0_core.c 	int temp, lifeboat = 0, prev;
prev              269 sound/pci/au88x0/au88x0_core.c 	prev = VORTEX_MIXER_CHNBASE + (ch << 2);
prev              270 sound/pci/au88x0/au88x0_core.c 	temp = hwread(vortex->mmio, prev);
prev              272 sound/pci/au88x0/au88x0_core.c 		prev = VORTEX_MIXER_RTBASE + ((temp & 0xf) << 2);
prev              273 sound/pci/au88x0/au88x0_core.c 		temp = hwread(vortex->mmio, prev);
prev              282 sound/pci/au88x0/au88x0_core.c 	hwwrite(vortex->mmio, prev, (temp & 0xf) | 0x10);
prev              657 sound/pci/au88x0/au88x0_core.c 	int temp, lifeboat = 0, prev;
prev              666 sound/pci/au88x0/au88x0_core.c 	prev = VORTEX_SRC_CHNBASE + (ch << 2);	/*ebp */
prev              667 sound/pci/au88x0/au88x0_core.c 	temp = hwread(vortex->mmio, prev);
prev              670 sound/pci/au88x0/au88x0_core.c 		prev = VORTEX_SRC_RTBASE + ((temp & 0xf) << 2);	/*esp12 */
prev              672 sound/pci/au88x0/au88x0_core.c 		temp = hwread(vortex->mmio, prev);
prev              682 sound/pci/au88x0/au88x0_core.c 	hwwrite(vortex->mmio, prev, (temp & 0xf) | 0x10);
prev             1632 sound/pci/au88x0/au88x0_core.c 	int temp, prev, lifeboat = 0;
prev             1660 sound/pci/au88x0/au88x0_core.c 		prev = temp;
prev             1672 sound/pci/au88x0/au88x0_core.c 	hwwrite(vortex->mmio, VORTEX_ADB_RTBASE + (prev << 2), route[0]);
prev             1679 sound/pci/au88x0/au88x0_core.c 	int temp, lifeboat = 0, prev;
prev             1696 sound/pci/au88x0/au88x0_core.c 		prev = temp;
prev             1699 sound/pci/au88x0/au88x0_core.c 			   VORTEX_ADB_RTBASE + (prev << 2)) & ADB_MASK;
prev             1712 sound/pci/au88x0/au88x0_core.c 	hwwrite(vortex->mmio, VORTEX_ADB_RTBASE + (prev << 2), temp);
prev              325 sound/pci/azt3328.c 	u8 prev = inb(reg), new;
prev              327 sound/pci/azt3328.c 	new = (do_set) ? (prev|mask) : (prev & ~mask);
prev              331 sound/pci/azt3328.c 	if (new != prev)
prev               42 sound/pci/ctxfi/ctimap.c 		pre = pos->prev;
prev               44 sound/pci/ctxfi/ctimap.c 			pre = head->prev;
prev               46 sound/pci/ctxfi/ctimap.c 		__list_add(&entry->list, pos->prev, pos);
prev               48 sound/pci/ctxfi/ctimap.c 		pre = head->prev;
prev               75 sound/pci/ctxfi/ctimap.c 	pre = (entry->list.prev == head) ? head->prev : entry->list.prev;
prev              100 sound/pci/ctxfi/ctvmem.c 			__list_add(&block->list, pos->prev, pos);
prev              106 sound/pci/ctxfi/ctvmem.c 	pre = pos->prev;
prev              117 sound/pci/ctxfi/ctvmem.c 		pre = pos->prev;
prev              172 sound/pci/emu10k1/memory.c 	if ((p = blk->mapped_link.prev) != &emu->mapped_link_head) {
prev              221 sound/pci/emu10k1/memory.c 	blk = (struct snd_emu10k1_memblk *)__snd_util_memblk_new(emu->memhdr, psize << PAGE_SHIFT, p->prev);
prev              457 sound/pci/emu10k1/memory.c 	if ((p = blk->mem.list.prev) != &hdr->block) {
prev             1385 sound/pci/es1968.c 	if (buf->list.prev != &chip->buf_list) {
prev             1386 sound/pci/es1968.c 		chunk = list_entry(buf->list.prev, struct esm_memory, list);
prev              944 sound/pci/hda/hda_intel.c 	int prev = power_save;
prev              947 sound/pci/hda/hda_intel.c 	if (ret || prev == power_save)
prev              144 sound/pci/trident/trident_memory.c 	blk = __snd_util_memblk_new(hdr, psize * ALIGN_PAGE_SIZE, p->prev);
prev              447 sound/soc/intel/atom/sst/sst.c 			stream->resume_prev = stream->prev;
prev              562 sound/soc/intel/atom/sst/sst.c 				stream->prev = stream->resume_prev;
prev              185 sound/soc/intel/atom/sst/sst.h 	unsigned int		prev;
prev              320 sound/soc/intel/atom/sst/sst_drv_interface.c 	str_info->prev = str_info->status;
prev              553 sound/soc/intel/atom/sst/sst_drv_interface.c 	str_info->prev = str_info->status;
prev              571 sound/soc/intel/atom/sst/sst_drv_interface.c 	str_info->prev = STREAM_UN_INIT;
prev              624 sound/soc/intel/atom/sst/sst_drv_interface.c 	stream->prev = stream->status;
prev              215 sound/soc/intel/atom/sst/sst_pvt.c 	stream->prev = STREAM_UN_INIT;
prev               73 sound/soc/intel/atom/sst/sst_stream.c 	sst_drv_ctx->streams[str_id].prev = STREAM_UN_INIT;
prev              255 sound/soc/intel/atom/sst/sst_stream.c 		if (str_info->prev == STREAM_UN_INIT)
prev              263 sound/soc/intel/atom/sst/sst_stream.c 			str_info->prev = str_info->status;
prev              306 sound/soc/intel/atom/sst/sst_stream.c 		str_info->prev = STREAM_PAUSED;
prev              317 sound/soc/intel/atom/sst/sst_stream.c 		str_info->prev = STREAM_PAUSED;
prev              326 sound/soc/intel/atom/sst/sst_stream.c 			if (str_info->prev == STREAM_RUNNING)
prev              330 sound/soc/intel/atom/sst/sst_stream.c 			str_info->prev = STREAM_PAUSED;
prev              364 sound/soc/intel/atom/sst/sst_stream.c 		str_info->prev = STREAM_UN_INIT;
prev              442 sound/soc/intel/atom/sst/sst_stream.c 		str_info->prev =  str_info->status;
prev              480 sound/synth/emux/soundfont.c 	struct snd_sf_zone *prev, *next, *p;
prev              483 sound/synth/emux/soundfont.c 	prev = NULL;
prev              489 sound/synth/emux/soundfont.c 			if (prev)
prev              490 sound/synth/emux/soundfont.c 				prev->next = next;
prev              496 sound/synth/emux/soundfont.c 			prev = p;
prev               88 sound/synth/util_mem.c 	return __snd_util_memblk_new(hdr, units, p->prev);
prev               98 sound/synth/util_mem.c 		      struct list_head *prev)
prev              107 sound/synth/util_mem.c 	if (prev == &hdr->block)
prev              110 sound/synth/util_mem.c 		struct snd_util_memblk *p = get_memblk(prev);
prev              114 sound/synth/util_mem.c 	list_add(&blk->list, prev);
prev              136 sound/usb/stream.c 	struct audioformat *prev = fp;
prev              138 sound/usb/stream.c 	list_for_each_entry_continue_reverse(prev, &subs->fmt_list, list) {
prev              139 sound/usb/stream.c 		if (prev->chmap &&
prev              140 sound/usb/stream.c 		    !memcmp(prev->chmap, fp->chmap, sizeof(*fp->chmap)))
prev                3 tools/firewire/list.h 	struct list *next, *prev;
prev               10 tools/firewire/list.h 	list->prev = list;
prev               22 tools/firewire/list.h 	new_link->prev		= link->prev;
prev               24 tools/firewire/list.h 	new_link->prev->next	= new_link;
prev               25 tools/firewire/list.h 	new_link->next->prev	= new_link;
prev               43 tools/firewire/list.h 	link->prev->next = link->next;
prev               44 tools/firewire/list.h 	link->next->prev = link->prev;
prev               54 tools/firewire/list.h 	list_entry((list)->prev, type, member)
prev              485 tools/firewire/nosy-dump.c 	struct subaction *sa, *prev;
prev              494 tools/firewire/nosy-dump.c 		prev = list_tail(&t->request_list,
prev              497 tools/firewire/nosy-dump.c 		if (!ACK_BUSY(prev->ack)) {
prev              506 tools/firewire/nosy-dump.c 		if (prev->packet.common.tcode != sa->packet.common.tcode ||
prev              507 tools/firewire/nosy-dump.c 		    prev->packet.common.tlabel != sa->packet.common.tlabel) {
prev              550 tools/firewire/nosy-dump.c 	struct subaction *sa, *prev;
prev              563 tools/firewire/nosy-dump.c 		prev = list_tail(&t->response_list, struct subaction, link);
prev              565 tools/firewire/nosy-dump.c 		if (!ACK_BUSY(prev->ack)) {
prev              572 tools/firewire/nosy-dump.c 		if (prev->packet.common.tcode != sa->packet.common.tcode ||
prev              573 tools/firewire/nosy-dump.c 		    prev->packet.common.tlabel != sa->packet.common.tlabel) {
prev              578 tools/firewire/nosy-dump.c 		prev = list_tail(&t->request_list, struct subaction, link);
prev              579 tools/firewire/nosy-dump.c 		if (prev->ack != ACK_PENDING) {
prev              586 tools/firewire/nosy-dump.c 		if (packet_info[prev->packet.common.tcode].response_tcode !=
prev               28 tools/include/linux/list.h 	list->prev = list;
prev               39 tools/include/linux/list.h 			      struct list_head *prev,
prev               42 tools/include/linux/list.h 	next->prev = new;
prev               44 tools/include/linux/list.h 	new->prev = prev;
prev               45 tools/include/linux/list.h 	prev->next = new;
prev               49 tools/include/linux/list.h 			      struct list_head *prev,
prev               77 tools/include/linux/list.h 	__list_add(new, head->prev, head);
prev               87 tools/include/linux/list.h static inline void __list_del(struct list_head * prev, struct list_head * next)
prev               89 tools/include/linux/list.h 	next->prev = prev;
prev               90 tools/include/linux/list.h 	WRITE_ONCE(prev->next, next);
prev              102 tools/include/linux/list.h 	__list_del(entry->prev, entry->next);
prev              107 tools/include/linux/list.h 	__list_del(entry->prev, entry->next);
prev              109 tools/include/linux/list.h 	entry->prev = LIST_POISON2;
prev              127 tools/include/linux/list.h 	new->next->prev = new;
prev              128 tools/include/linux/list.h 	new->prev = old->prev;
prev              129 tools/include/linux/list.h 	new->prev->next = new;
prev              208 tools/include/linux/list.h 	return (next == head) && (next == head->prev);
prev              231 tools/include/linux/list.h 	return !list_empty(head) && (head->next == head->prev);
prev              239 tools/include/linux/list.h 	list->next->prev = list;
prev              240 tools/include/linux/list.h 	list->prev = entry;
prev              243 tools/include/linux/list.h 	new_first->prev = head;
prev              275 tools/include/linux/list.h 				 struct list_head *prev,
prev              279 tools/include/linux/list.h 	struct list_head *last = list->prev;
prev              281 tools/include/linux/list.h 	first->prev = prev;
prev              282 tools/include/linux/list.h 	prev->next = first;
prev              285 tools/include/linux/list.h 	next->prev = last;
prev              309 tools/include/linux/list.h 		__list_splice(list, head->prev, head);
prev              340 tools/include/linux/list.h 		__list_splice(list, head->prev, head);
prev              374 tools/include/linux/list.h 	list_entry((ptr)->prev, type, member)
prev              401 tools/include/linux/list.h 	list_entry((pos)->member.prev, typeof(*(pos)), member)
prev              417 tools/include/linux/list.h 	for (pos = (head)->prev; pos != (head); pos = pos->prev)
prev              436 tools/include/linux/list.h 	for (pos = (head)->prev, n = pos->prev; \
prev              438 tools/include/linux/list.h 	     pos = n, n = pos->prev)
prev              660 tools/include/linux/list.h 				    struct hlist_node *prev)
prev              662 tools/include/linux/list.h 	n->next = prev->next;
prev              663 tools/include/linux/list.h 	prev->next = n;
prev              664 tools/include/linux/list.h 	n->pprev = &prev->next;
prev              760 tools/include/linux/list.h 	begin->prev->next = end->next;
prev              761 tools/include/linux/list.h 	end->next->prev = begin->prev;
prev               70 tools/include/linux/types.h 	struct list_head *next, *prev;
prev              117 tools/include/uapi/drm/drm.h 	unsigned char prev;
prev             4227 tools/lib/bpf/libbpf.c bpf_object__next(struct bpf_object *prev)
prev             4231 tools/lib/bpf/libbpf.c 	if (!prev)
prev             4236 tools/lib/bpf/libbpf.c 		next = list_next_entry(prev, list);
prev             4308 tools/lib/bpf/libbpf.c bpf_program__next(struct bpf_program *prev, const struct bpf_object *obj)
prev             4310 tools/lib/bpf/libbpf.c 	struct bpf_program *prog = prev;
prev             4750 tools/lib/bpf/libbpf.c bpf_map__next(const struct bpf_map *prev, const struct bpf_object *obj)
prev             4752 tools/lib/bpf/libbpf.c 	if (prev == NULL)
prev             4755 tools/lib/bpf/libbpf.c 	return __bpf_map__iter(prev, obj, 1);
prev              113 tools/lib/bpf/libbpf.h LIBBPF_API struct bpf_object *bpf_object__next(struct bpf_object *prev);
prev               51 tools/lib/traceevent/event-parse.h 	struct tep_record	*prev;
prev              132 tools/perf/builtin-annotate.c 	struct addr_map_symbol *prev = NULL;
prev              147 tools/perf/builtin-annotate.c 		if (prev)
prev              148 tools/perf/builtin-annotate.c 			process_basic_block(prev, &bi[i].from, &bi[i].flags);
prev              149 tools/perf/builtin-annotate.c 		prev = &bi[i].to;
prev              188 tools/perf/builtin-kvm.c 	void *prev;
prev              196 tools/perf/builtin-kvm.c 	prev = event->vcpu;
prev              200 tools/perf/builtin-kvm.c 		free(prev);
prev              838 tools/perf/builtin-sched.c 	struct task_desc *prev, __maybe_unused *next;
prev              863 tools/perf/builtin-sched.c 	prev = register_pid(sched, prev_pid, prev_comm);
prev              868 tools/perf/builtin-sched.c 	add_sched_event_run(sched, prev, timestamp, delta);
prev              869 tools/perf/builtin-sched.c 	add_sched_event_sleep(sched, prev, timestamp, prev_state);
prev             1064 tools/perf/builtin-sched.c 	atom = list_entry(atoms->work_list.prev, struct work_atom, list);
prev             1079 tools/perf/builtin-sched.c 	atom = list_entry(atoms->work_list.prev, struct work_atom, list);
prev             1236 tools/perf/builtin-sched.c 	atom = list_entry(atoms->work_list.prev, struct work_atom, list);
prev             1304 tools/perf/builtin-sched.c 	atom = list_entry(atoms->work_list.prev, struct work_atom, list);
prev              724 tools/perf/builtin-timechart.c 	struct io_sample *prev;
prev              735 tools/perf/builtin-timechart.c 	prev = c->io_samples;
prev              737 tools/perf/builtin-timechart.c 	if (prev && prev->start_time && !prev->end_time) {
prev              743 tools/perf/builtin-timechart.c 		c->io_samples = prev->next;
prev              744 tools/perf/builtin-timechart.c 		free(prev);
prev              768 tools/perf/builtin-timechart.c 	struct io_sample *sample, *prev;
prev              792 tools/perf/builtin-timechart.c 	prev = sample->next;
prev              798 tools/perf/builtin-timechart.c 	if (prev && sample->start_time < prev->end_time) {
prev              799 tools/perf/builtin-timechart.c 		if (prev->err) /* try to make errors more visible */
prev              800 tools/perf/builtin-timechart.c 			sample->start_time = prev->end_time;
prev              802 tools/perf/builtin-timechart.c 			prev->end_time = sample->start_time;
prev              819 tools/perf/builtin-timechart.c 	if (prev &&
prev              820 tools/perf/builtin-timechart.c 	    prev->type == sample->type &&
prev              821 tools/perf/builtin-timechart.c 	    prev->err == sample->err &&
prev              822 tools/perf/builtin-timechart.c 	    prev->fd == sample->fd &&
prev              823 tools/perf/builtin-timechart.c 	    prev->end_time + tchart->merge_dist >= sample->start_time) {
prev              825 tools/perf/builtin-timechart.c 		sample->bytes += prev->bytes;
prev              826 tools/perf/builtin-timechart.c 		sample->merges += prev->merges + 1;
prev              828 tools/perf/builtin-timechart.c 		sample->start_time = prev->start_time;
prev              829 tools/perf/builtin-timechart.c 		sample->next = prev->next;
prev              830 tools/perf/builtin-timechart.c 		free(prev);
prev              966 tools/perf/builtin-timechart.c 	struct per_pid *new_list, *p, *cursor, *prev;
prev              981 tools/perf/builtin-timechart.c 		prev = NULL;
prev              987 tools/perf/builtin-timechart.c 				if (prev) {
prev              988 tools/perf/builtin-timechart.c 					p->next = prev->next;
prev              989 tools/perf/builtin-timechart.c 					prev->next = p;
prev             1000 tools/perf/builtin-timechart.c 			prev = cursor;
prev             1003 tools/perf/builtin-timechart.c 				prev->next = p;
prev               91 tools/perf/lib/cpumap.c 	int n, cpu, prev;
prev               95 tools/perf/lib/cpumap.c 	prev = -1;
prev              100 tools/perf/lib/cpumap.c 		if (prev >= 0) {
prev              101 tools/perf/lib/cpumap.c 			int new_max = nr_cpus + cpu - prev - 1;
prev              114 tools/perf/lib/cpumap.c 			while (++prev < cpu)
prev              115 tools/perf/lib/cpumap.c 				tmp_cpus[nr_cpus++] = prev;
prev              127 tools/perf/lib/cpumap.c 			prev = cpu;
prev              129 tools/perf/lib/cpumap.c 			prev = -1;
prev               85 tools/perf/lib/evlist.c perf_evlist__next(struct perf_evlist *evlist, struct perf_evsel *prev)
prev               89 tools/perf/lib/evlist.c 	if (!prev) {
prev               94 tools/perf/lib/evlist.c 		next = list_next_entry(prev, node);
prev               70 tools/perf/lib/include/internal/evlist.h 	return list_entry(evlist->entries.prev, struct perf_evsel, node);
prev               24 tools/perf/lib/include/internal/mmap.h 	u64		 prev;
prev               97 tools/perf/ui/browser.c 		pos = pos->prev;
prev              119 tools/perf/ui/browser.c 		pos = ui_browser__list_head_filter_prev_entries(browser, head->prev);
prev              132 tools/perf/ui/browser.c 			pos = ui_browser__list_head_filter_prev_entries(browser, pos->prev);
prev              276 tools/perf/ui/browsers/annotate.c 		pos = list_entry(pos->node.prev, struct annotation_line, node);
prev              357 tools/perf/ui/browsers/hists.c 		chain = list_entry(node->val.prev, struct callchain_list, list);
prev             3328 tools/perf/ui/browsers/hists.c 				if (pos->core.node.prev == &evlist->core.entries)
prev              537 tools/perf/util/annotate.c 	char *s = strchr(ops->raw, ','), *target, *comment, prev;
prev              560 tools/perf/util/annotate.c 	prev = *s;
prev              564 tools/perf/util/annotate.c 	*s = prev;
prev              598 tools/perf/util/annotate.c 	char *target, *comment, *s, prev;
prev              604 tools/perf/util/annotate.c 	prev = *s;
prev              608 tools/perf/util/annotate.c 	*s = prev;
prev             1583 tools/perf/util/annotate.c 		dl = list_entry(list->prev, struct disasm_line, al.node);
prev             2934 tools/perf/util/annotate.c 					int prev;
prev             2938 tools/perf/util/annotate.c 					prev = obj__set_jumps_percent_color(obj, al->jump_sources,
prev             2941 tools/perf/util/annotate.c 					obj__set_color(obj, prev);
prev               77 tools/perf/util/auxtrace.c 	mm->prev = 0;
prev              644 tools/perf/util/auxtrace.c 	auxtrace_index = list_entry(head->prev, struct auxtrace_index, list);
prev              650 tools/perf/util/auxtrace.c 		auxtrace_index = list_entry(head->prev, struct auxtrace_index,
prev             1237 tools/perf/util/auxtrace.c 	u64 head, old = mm->prev, offset, ref;
prev             1326 tools/perf/util/auxtrace.c 	mm->prev = head;
prev              276 tools/perf/util/auxtrace.h 	u64		prev;
prev              804 tools/perf/util/callchain.c 	old_tail = parent->val.prev;
prev              807 tools/perf/util/callchain.c 	new->val.prev = old_tail;
prev              808 tools/perf/util/callchain.c 	to_split->list.prev = &new->val;
prev              194 tools/perf/util/evlist.c 	evsel = list_entry(list->prev, struct evsel, core.node);
prev              314 tools/perf/util/evsel.h 	return list_entry(evsel->core.node.prev, struct evsel, core.node);
prev             2582 tools/perf/util/hist.c 			struct addr_map_symbol *prev = NULL;
prev             2596 tools/perf/util/hist.c 					nonany_branch_mode ? NULL : prev,
prev             2598 tools/perf/util/hist.c 				prev = &bi[i].to;
prev              254 tools/perf/util/intel-bts.c 	if (b->list.prev == &queue->head)
prev              256 tools/perf/util/intel-bts.c 	a = list_entry(b->list.prev, struct auxtrace_buffer, list);
prev               37 tools/perf/util/intel-pt-decoder/intel-pt-decoder.c 	struct intel_pt_blk *prev;
prev              285 tools/perf/util/intel-pt-decoder/intel-pt-decoder.c 	stack->blk = blk->prev;
prev              318 tools/perf/util/intel-pt-decoder/intel-pt-decoder.c 	blk->prev = stack->blk;
prev               83 tools/perf/util/mem2node.c 				struct phys_entry *prev = &entries[j - 1];
prev               85 tools/perf/util/mem2node.c 				if ((prev->end == start) &&
prev               86 tools/perf/util/mem2node.c 				    (prev->node == n->node)) {
prev               87 tools/perf/util/mem2node.c 					prev->end += bsize;
prev              103 tools/perf/util/mmap.c 		map->core.prev = map->core.start;
prev              110 tools/perf/util/mmap.c 	return perf_mmap__read_head(map) == map->core.prev && !map->auxtrace_mmap.base;
prev              129 tools/perf/util/mmap.c 		u64 old = map->core.prev;
prev              372 tools/perf/util/mmap.c 	map->core.prev = 0;
prev              447 tools/perf/util/mmap.c 	u64 old = md->core.prev;
prev              462 tools/perf/util/mmap.c 			md->core.prev = head;
prev              524 tools/perf/util/mmap.c 	md->core.prev = head;
prev              544 tools/perf/util/mmap.c 	map->core.prev = perf_mmap__read_head(map);
prev               53 tools/perf/util/ordered-events.c 			p = last->list.prev;
prev              258 tools/perf/util/ordered-events.c 		oe->last = list_entry(head->prev, struct ordered_event, list);
prev             2075 tools/perf/util/parse-events.c 		if (last->core.node.prev == &evlist->core.entries)
prev             2077 tools/perf/util/parse-events.c 		last = list_entry(last->core.node.prev, struct evsel, core.node);
prev              117 tools/perf/util/srccode.c 		h = list_entry(srcfile_list.prev, struct srcfile, nd);
prev              222 tools/perf/util/symbol.c 	struct symbol *curr, *prev;
prev              230 tools/perf/util/symbol.c 		prev = curr;
prev              233 tools/perf/util/symbol.c 		if (prev->end == prev->start && prev->end != curr->start)
prev              234 tools/perf/util/symbol.c 			arch__symbols__fixup_end(prev, curr);
prev               33 tools/testing/selftests/bpf/progs/btf_dump_test_case_ordering.c 	struct list_head *prev;
prev              644 tools/testing/selftests/kselftest_harness.h 	struct __test_metadata *prev, *next;
prev              672 tools/testing/selftests/kselftest_harness.h 		t->prev = t;
prev              677 tools/testing/selftests/kselftest_harness.h 		t->prev = __test_list->prev;
prev              678 tools/testing/selftests/kselftest_harness.h 		t->prev->next = t;
prev              679 tools/testing/selftests/kselftest_harness.h 		__test_list->prev = t;
prev              682 tools/testing/selftests/kselftest_harness.h 		t->next->prev = t;
prev              683 tools/testing/selftests/kselftest_harness.h 		t->prev = t;
prev               48 tools/testing/selftests/kvm/include/sparsebit.h sparsebit_idx_t sparsebit_next_set(struct sparsebit *sbit, sparsebit_idx_t prev);
prev               49 tools/testing/selftests/kvm/include/sparsebit.h sparsebit_idx_t sparsebit_next_clear(struct sparsebit *sbit, sparsebit_idx_t prev);
prev              419 tools/testing/selftests/kvm/lib/kvm_util.c 		vcpu->next->prev = vcpu->prev;
prev              420 tools/testing/selftests/kvm/lib/kvm_util.c 	if (vcpu->prev)
prev              421 tools/testing/selftests/kvm/lib/kvm_util.c 		vcpu->prev->next = vcpu->next;
prev              688 tools/testing/selftests/kvm/lib/kvm_util.c 		vm->userspace_mem_region_head->prev = region;
prev              835 tools/testing/selftests/kvm/lib/kvm_util.c 		vm->vcpu_head->prev = vcpu;
prev               27 tools/testing/selftests/kvm/lib/kvm_util_internal.h 	struct userspace_mem_region *next, *prev;
prev               38 tools/testing/selftests/kvm/lib/kvm_util_internal.h 	struct vcpu *next, *prev;
prev              335 tools/testing/selftests/kvm/lib/sparsebit.c 	struct node *nodep, *parentp, *prev;
prev              381 tools/testing/selftests/kvm/lib/sparsebit.c 	prev = node_prev(s, nodep);
prev              382 tools/testing/selftests/kvm/lib/sparsebit.c 	while (prev && prev->idx + MASK_BITS + prev->num_after - 1 >= nodep->idx) {
prev              383 tools/testing/selftests/kvm/lib/sparsebit.c 		unsigned int n1 = (prev->idx + MASK_BITS + prev->num_after - 1)
prev              385 tools/testing/selftests/kvm/lib/sparsebit.c 		assert(prev->num_after > 0);
prev              389 tools/testing/selftests/kvm/lib/sparsebit.c 		prev->num_after--;
prev              605 tools/testing/selftests/kvm/lib/sparsebit.c 		struct node *prev, *next, *tmp;
prev              670 tools/testing/selftests/kvm/lib/sparsebit.c 		prev = node_prev(s, nodep);
prev              671 tools/testing/selftests/kvm/lib/sparsebit.c 		if (prev) {
prev              675 tools/testing/selftests/kvm/lib/sparsebit.c 			if (prev->mask == 0 && prev->num_after == 0) {
prev              676 tools/testing/selftests/kvm/lib/sparsebit.c 				node_rm(s, prev);
prev              687 tools/testing/selftests/kvm/lib/sparsebit.c 			    prev->idx + MASK_BITS == nodep->idx) {
prev              688 tools/testing/selftests/kvm/lib/sparsebit.c 				prev->num_after += MASK_BITS + nodep->num_after;
prev              701 tools/testing/selftests/kvm/lib/sparsebit.c 			prev_highest_bit = prev->idx + MASK_BITS - 1 + prev->num_after;
prev              716 tools/testing/selftests/kvm/lib/sparsebit.c 				prev->num_after += num_contiguous;
prev              733 tools/testing/selftests/kvm/lib/sparsebit.c 					prev->num_after += nodep->num_after;
prev             1156 tools/testing/selftests/kvm/lib/sparsebit.c 	sparsebit_idx_t prev)
prev             1158 tools/testing/selftests/kvm/lib/sparsebit.c 	sparsebit_idx_t lowest_possible = prev + 1;
prev             1249 tools/testing/selftests/kvm/lib/sparsebit.c 	sparsebit_idx_t prev)
prev             1251 tools/testing/selftests/kvm/lib/sparsebit.c 	sparsebit_idx_t lowest_possible = prev + 1;
prev             1688 tools/testing/selftests/kvm/lib/sparsebit.c 	struct node *nodep, *prev = NULL;
prev             1694 tools/testing/selftests/kvm/lib/sparsebit.c 		prev = nodep, nodep = node_next(s, nodep)) {
prev             1803 tools/testing/selftests/kvm/lib/sparsebit.c 		if (prev) {
prev             1808 tools/testing/selftests/kvm/lib/sparsebit.c 			if (prev->idx >= nodep->idx) {
prev             1813 tools/testing/selftests/kvm/lib/sparsebit.c 					prev, prev->idx, nodep, nodep->idx);
prev             1822 tools/testing/selftests/kvm/lib/sparsebit.c 			if ((prev->idx + MASK_BITS + prev->num_after - 1)
prev             1831 tools/testing/selftests/kvm/lib/sparsebit.c 					prev, prev->idx, prev->num_after,
prev             1844 tools/testing/selftests/kvm/lib/sparsebit.c 			    prev->idx + MASK_BITS + prev->num_after == nodep->idx) {
prev             1853 tools/testing/selftests/kvm/lib/sparsebit.c 					prev, prev->idx, prev->num_after,
prev              109 tools/testing/selftests/rcutorture/formal/srcu-cbmc/include/linux/types.h 	struct list_head *next, *prev;
prev               20 tools/usb/usbip/libsrc/list.h 	struct list_head *next, *prev;
prev               31 tools/usb/usbip/libsrc/list.h 	list->prev = list;
prev               41 tools/usb/usbip/libsrc/list.h 			      struct list_head *prev,
prev               44 tools/usb/usbip/libsrc/list.h 	next->prev = new;
prev               46 tools/usb/usbip/libsrc/list.h 	new->prev = prev;
prev               47 tools/usb/usbip/libsrc/list.h 	prev->next = new;
prev               70 tools/usb/usbip/libsrc/list.h static inline void __list_del(struct list_head * prev, struct list_head * next)
prev               72 tools/usb/usbip/libsrc/list.h 	next->prev = prev;
prev               73 tools/usb/usbip/libsrc/list.h 	prev->next = next;
prev               88 tools/usb/usbip/libsrc/list.h 	__list_del(entry->prev, entry->next);
prev               93 tools/usb/usbip/libsrc/list.h 	__list_del(entry->prev, entry->next);
prev               95 tools/usb/usbip/libsrc/list.h 	entry->prev = LIST_POISON2;
prev               21 tools/virtio/ringtest/ring.c 			      unsigned short prev)
prev               23 tools/virtio/ringtest/ring.c 	return (unsigned short)(next - event - 1) < (unsigned short)(next - prev);