entry              13 arch/alpha/include/asm/a.out.h 	set_personality (((BFPM->taso || EX.ah.entry < 0x100000000L \
entry              32 arch/alpha/include/uapi/asm/a.out.h 	__u64 entry;
entry              69 arch/alpha/include/uapi/asm/a.out.h #define a_entry		ah.entry
entry              34 arch/alpha/kernel/binfmt_loader.c 	bprm->taso = eh->ah.entry < 0x100000000UL;
entry             134 arch/alpha/kernel/srm_env.c 	srm_env_t	*entry;
entry             182 arch/alpha/kernel/srm_env.c 	entry = srm_named_entries;
entry             183 arch/alpha/kernel/srm_env.c 	while (entry->name && entry->id) {
entry             184 arch/alpha/kernel/srm_env.c 		if (!proc_create_data(entry->name, 0644, named_dir,
entry             185 arch/alpha/kernel/srm_env.c 			     &srm_env_proc_fops, (void *)entry->id))
entry             187 arch/alpha/kernel/srm_env.c 		entry++;
entry             148 arch/arc/include/asm/entry-arcv2.h 	; 2. Upon entry SP is always saved (for any inspection, unwinding etc),
entry             180 arch/arc/include/asm/entry-arcv2.h 	;  - Z flag set on K is inverse of what hardware does on interrupt entry
entry              65 arch/arc/kernel/perf_event.c 	struct perf_callchain_entry_ctx *entry = ctrl->perf_stuff;
entry              67 arch/arc/kernel/perf_event.c 	perf_callchain_store(entry, addr);
entry              75 arch/arc/kernel/perf_event.c void perf_callchain_kernel(struct perf_callchain_entry_ctx *entry,
entry              80 arch/arc/kernel/perf_event.c 		.perf_stuff = entry,
entry              86 arch/arc/kernel/perf_event.c void perf_callchain_user(struct perf_callchain_entry_ctx *entry,
entry              93 arch/arc/kernel/perf_event.c 	perf_callchain_store(entry, instruction_pointer(regs));
entry             254 arch/arc/mm/tlb.c 	unsigned int entry;
entry             267 arch/arc/mm/tlb.c 	for (entry = 0; entry < num_tlb; entry++) {
entry             269 arch/arc/mm/tlb.c 		write_aux_reg(ARC_REG_TLBINDEX, entry);
entry             279 arch/arc/mm/tlb.c 		for (entry = stlb_idx; entry < stlb_idx + 16; entry++) {
entry             280 arch/arc/mm/tlb.c 			write_aux_reg(ARC_REG_TLBINDEX, entry);
entry             489 arch/arm/include/asm/assembler.h #define _ASM_NOKPROBE(entry)				\
entry             492 arch/arm/include/asm/assembler.h 	.long entry;					\
entry             495 arch/arm/include/asm/assembler.h #define _ASM_NOKPROBE(entry)
entry              68 arch/arm/include/asm/kexec.h static inline phys_addr_t boot_phys_to_phys(unsigned long entry)
entry              70 arch/arm/include/asm/kexec.h 	return idmap_to_phys(entry);
entry              41 arch/arm/include/asm/tlb.h #define tlb_remove_table(tlb, entry) tlb_remove_page(tlb, entry)
entry               7 arch/arm/kernel/jump_label.c static void __arch_jump_label_transform(struct jump_entry *entry,
entry              11 arch/arm/kernel/jump_label.c 	void *addr = (void *)entry->code;
entry              15 arch/arm/kernel/jump_label.c 		insn = arm_gen_branch(entry->code, entry->target);
entry              25 arch/arm/kernel/jump_label.c void arch_jump_label_transform(struct jump_entry *entry,
entry              28 arch/arm/kernel/jump_label.c 	__arch_jump_label_transform(entry, type, false);
entry              31 arch/arm/kernel/jump_label.c void arch_jump_label_transform_static(struct jump_entry *entry,
entry              34 arch/arm/kernel/jump_label.c 	__arch_jump_label_transform(entry, type, true);
entry              35 arch/arm/kernel/perf_callchain.c 	       struct perf_callchain_entry_ctx *entry)
entry              50 arch/arm/kernel/perf_callchain.c 	perf_callchain_store(entry, buftail.lr);
entry              63 arch/arm/kernel/perf_callchain.c perf_callchain_user(struct perf_callchain_entry_ctx *entry, struct pt_regs *regs)
entry              72 arch/arm/kernel/perf_callchain.c 	perf_callchain_store(entry, regs->ARM_pc);
entry              79 arch/arm/kernel/perf_callchain.c 	while ((entry->nr < entry->max_stack) &&
entry              81 arch/arm/kernel/perf_callchain.c 		tail = user_backtrace(tail, entry);
entry              93 arch/arm/kernel/perf_callchain.c 	struct perf_callchain_entry_ctx *entry = data;
entry              94 arch/arm/kernel/perf_callchain.c 	perf_callchain_store(entry, fr->pc);
entry              99 arch/arm/kernel/perf_callchain.c perf_callchain_kernel(struct perf_callchain_entry_ctx *entry, struct pt_regs *regs)
entry             109 arch/arm/kernel/perf_callchain.c 	walk_stackframe(&fr, callchain_trace, entry);
entry              67 arch/arm/kernel/vmlinux.lds.h 		*(.entry.text)						\
entry              40 arch/arm/mm/fault-armv.c 	pte_t entry = *ptep;
entry              46 arch/arm/mm/fault-armv.c 	ret = pte_present(entry);
entry              52 arch/arm/mm/fault-armv.c 	if (ret && (pte_val(entry) & L_PTE_MT_MASK) != shared_pte_mask) {
entry              56 arch/arm/mm/fault-armv.c 		pte_val(entry) &= ~L_PTE_MT_MASK;
entry              57 arch/arm/mm/fault-armv.c 		pte_val(entry) |= shared_pte_mask;
entry              58 arch/arm/mm/fault-armv.c 		set_pte_at(vma->vm_mm, address, ptep, entry);
entry             738 arch/arm/probes/kprobes/test-core.c 	struct coverage_entry *entry = coverage->base + coverage->num_entries;
entry             747 arch/arm/probes/kprobes/test-core.c 	entry->header = h;
entry             748 arch/arm/probes/kprobes/test-core.c 	entry->regs = coverage_start_registers(h);
entry             749 arch/arm/probes/kprobes/test-core.c 	entry->nesting = coverage->nesting;
entry             750 arch/arm/probes/kprobes/test-core.c 	entry->matched = false;
entry             775 arch/arm/probes/kprobes/test-core.c coverage_add_registers(struct coverage_entry *entry, kprobe_opcode_t insn)
entry             777 arch/arm/probes/kprobes/test-core.c 	int regs = entry->header->type_regs.bits >> DECODE_TYPE_BITS;
entry             793 arch/arm/probes/kprobes/test-core.c 		entry->regs &= ~(flag << i);
entry             827 arch/arm/probes/kprobes/test-core.c 				entry->regs &= ~(COVERAGE_PCWB << i);
entry             844 arch/arm/probes/kprobes/test-core.c 	struct coverage_entry *entry = coverage.base;
entry             849 arch/arm/probes/kprobes/test-core.c 	for (; entry < end; ++entry) {
entry             850 arch/arm/probes/kprobes/test-core.c 		const struct decode_header *h = entry->header;
entry             853 arch/arm/probes/kprobes/test-core.c 		if (entry->nesting > nesting)
entry             856 arch/arm/probes/kprobes/test-core.c 		if (entry->nesting < nesting)
entry             862 arch/arm/probes/kprobes/test-core.c 			entry->matched = true;
entry             874 arch/arm/probes/kprobes/test-core.c 			coverage_add_registers(entry, insn);
entry             891 arch/arm/probes/kprobes/test-core.c 	struct coverage_entry *entry = coverage.base;
entry             894 arch/arm/probes/kprobes/test-core.c 	for (; entry < end; ++entry) {
entry             895 arch/arm/probes/kprobes/test-core.c 		u32 mask = entry->header->mask.bits;
entry             896 arch/arm/probes/kprobes/test-core.c 		u32 value = entry->header->value.bits;
entry             898 arch/arm/probes/kprobes/test-core.c 		if (entry->regs) {
entry             900 arch/arm/probes/kprobes/test-core.c 				mask, value, entry->regs);
entry             903 arch/arm/probes/kprobes/test-core.c 		if (!entry->matched) {
entry              36 arch/arm/xen/p2m.c 	struct xen_p2m_entry *entry;
entry              41 arch/arm/xen/p2m.c 		entry = rb_entry(parent, struct xen_p2m_entry, rbnode_phys);
entry              43 arch/arm/xen/p2m.c 		if (new->pfn == entry->pfn)
entry              46 arch/arm/xen/p2m.c 		if (new->pfn < entry->pfn)
entry              58 arch/arm/xen/p2m.c 			__func__, &new->pfn, &new->mfn, &entry->pfn, &entry->mfn);
entry              66 arch/arm/xen/p2m.c 	struct xen_p2m_entry *entry;
entry              71 arch/arm/xen/p2m.c 		entry = rb_entry(n, struct xen_p2m_entry, rbnode_phys);
entry              72 arch/arm/xen/p2m.c 		if (entry->pfn <= pfn &&
entry              73 arch/arm/xen/p2m.c 				entry->pfn + entry->nr_pages > pfn) {
entry              74 arch/arm/xen/p2m.c 			unsigned long mfn = entry->mfn + (pfn - entry->pfn);
entry              78 arch/arm/xen/p2m.c 		if (pfn < entry->pfn)
entry              38 arch/arm64/include/asm/acpi.h #define BAD_MADT_GICC_ENTRY(entry, end)					\
entry              39 arch/arm64/include/asm/acpi.h 	(!(entry) || (entry)->header.length < ACPI_MADT_GICC_MIN_LENGTH || \
entry              40 arch/arm64/include/asm/acpi.h 	(unsigned long)(entry) + (entry)->header.length > (end))
entry             361 arch/arm64/include/asm/cpufeature.h cpucap_multi_entry_cap_matches(const struct arm64_cpu_capabilities *entry,
entry             366 arch/arm64/include/asm/cpufeature.h 	for (caps = entry->match_list; caps->matches; caps++)
entry              37 arch/arm64/include/asm/hugetlb.h extern pte_t arch_make_huge_pte(pte_t entry, struct vm_area_struct *vma,
entry             677 arch/arm64/include/asm/pgtable.h 				 pte_t entry, int dirty);
entry             683 arch/arm64/include/asm/pgtable.h 					pmd_t entry, int dirty)
entry             685 arch/arm64/include/asm/pgtable.h 	return ptep_set_access_flags(vma, address, (pte_t *)pmdp, pmd_pte(entry), dirty);
entry              13 arch/arm64/kernel/cpu-reset.h void __cpu_soft_restart(unsigned long el2_switch, unsigned long entry,
entry              16 arch/arm64/kernel/cpu-reset.h static inline void __noreturn cpu_soft_restart(unsigned long entry,
entry              28 arch/arm64/kernel/cpu-reset.h 	restart(el2_switch, entry, arg0, arg1, arg2);
entry              18 arch/arm64/kernel/cpu_errata.c is_affected_midr_range(const struct arm64_cpu_capabilities *entry, int scope)
entry              24 arch/arm64/kernel/cpu_errata.c 	if (!is_midr_in_range(midr, &entry->midr_range))
entry              29 arch/arm64/kernel/cpu_errata.c 	for (fix = entry->fixed_revs; fix && fix->revidr_mask; fix++)
entry              37 arch/arm64/kernel/cpu_errata.c is_affected_midr_range_list(const struct arm64_cpu_capabilities *entry,
entry              41 arch/arm64/kernel/cpu_errata.c 	return is_midr_in_range_list(read_cpuid_id(), entry->midr_range_list);
entry              45 arch/arm64/kernel/cpu_errata.c is_kryo_midr(const struct arm64_cpu_capabilities *entry, int scope)
entry              55 arch/arm64/kernel/cpu_errata.c 	return model == entry->midr_range.model;
entry              59 arch/arm64/kernel/cpu_errata.c has_mismatched_cache_type(const struct arm64_cpu_capabilities *entry,
entry             378 arch/arm64/kernel/cpu_errata.c static bool has_ssbd_mitigation(const struct arm64_cpu_capabilities *entry,
entry             392 arch/arm64/kernel/cpu_errata.c 	if (is_midr_in_range_list(read_cpuid_id(), entry->midr_range_list))
entry             438 arch/arm64/kernel/cpu_errata.c 		pr_info_once("%s mitigation not required\n", entry->desc);
entry             484 arch/arm64/kernel/cpu_errata.c 		pr_info_once("%s disabled from command-line\n", entry->desc);
entry             488 arch/arm64/kernel/cpu_errata.c 		pr_info_once("%s forced from command-line\n", entry->desc);
entry             508 arch/arm64/kernel/cpu_errata.c has_cortex_a76_erratum_1463225(const struct arm64_cpu_capabilities *entry,
entry             595 arch/arm64/kernel/cpu_errata.c check_branch_predictor(const struct arm64_cpu_capabilities *entry, int scope)
entry             645 arch/arm64/kernel/cpu_errata.c needs_tx2_tvm_workaround(const struct arm64_cpu_capabilities *entry,
entry             650 arch/arm64/kernel/cpu_errata.c 	if (!is_affected_midr_range_list(entry, scope) ||
entry             663 arch/arm64/kernel/cpu_errata.c has_neoverse_n1_erratum_1542419(const struct arm64_cpu_capabilities *entry,
entry             111 arch/arm64/kernel/cpufeature.c cpufeature_pan_not_uao(const struct arm64_cpu_capabilities *entry, int __unused);
entry             860 arch/arm64/kernel/cpufeature.c feature_matches(u64 reg, const struct arm64_cpu_capabilities *entry)
entry             862 arch/arm64/kernel/cpufeature.c 	int val = cpuid_feature_extract_field(reg, entry->field_pos, entry->sign);
entry             864 arch/arm64/kernel/cpufeature.c 	return val >= entry->min_field_value;
entry             868 arch/arm64/kernel/cpufeature.c has_cpuid_feature(const struct arm64_cpu_capabilities *entry, int scope)
entry             874 arch/arm64/kernel/cpufeature.c 		val = read_sanitised_ftr_reg(entry->sys_reg);
entry             876 arch/arm64/kernel/cpufeature.c 		val = __read_sysreg_by_encoding(entry->sys_reg);
entry             878 arch/arm64/kernel/cpufeature.c 	return feature_matches(val, entry);
entry             881 arch/arm64/kernel/cpufeature.c static bool has_useable_gicv3_cpuif(const struct arm64_cpu_capabilities *entry, int scope)
entry             885 arch/arm64/kernel/cpufeature.c 	if (!has_cpuid_feature(entry, scope))
entry             891 arch/arm64/kernel/cpufeature.c 			     entry->desc);
entry             896 arch/arm64/kernel/cpufeature.c static bool has_no_hw_prefetch(const struct arm64_cpu_capabilities *entry, int __unused)
entry             906 arch/arm64/kernel/cpufeature.c static bool has_no_fpsimd(const struct arm64_cpu_capabilities *entry, int __unused)
entry             914 arch/arm64/kernel/cpufeature.c static bool has_cache_idc(const struct arm64_cpu_capabilities *entry,
entry             939 arch/arm64/kernel/cpufeature.c static bool has_cache_dic(const struct arm64_cpu_capabilities *entry,
entry             953 arch/arm64/kernel/cpufeature.c has_useable_cnp(const struct arm64_cpu_capabilities *entry, int scope)
entry             963 arch/arm64/kernel/cpufeature.c 	return has_cpuid_feature(entry, scope);
entry             969 arch/arm64/kernel/cpufeature.c static bool unmap_kernel_at_el0(const struct arm64_cpu_capabilities *entry,
entry             991 arch/arm64/kernel/cpufeature.c 	if (has_cpuid_feature(entry, scope))
entry            1151 arch/arm64/kernel/cpufeature.c static bool runs_at_el2(const struct arm64_cpu_capabilities *entry, int __unused)
entry            1260 arch/arm64/kernel/cpufeature.c static bool can_use_gic_priorities(const struct arm64_cpu_capabilities *entry,
entry            1263 arch/arm64/kernel/cpufeature.c 	return enable_pseudo_nmi && has_useable_gicv3_cpuif(entry, scope);
entry            2112 arch/arm64/kernel/cpufeature.c cpufeature_pan_not_uao(const struct arm64_cpu_capabilities *entry, int __unused)
entry              12 arch/arm64/kernel/jump_label.c void arch_jump_label_transform(struct jump_entry *entry,
entry              15 arch/arm64/kernel/jump_label.c 	void *addr = (void *)jump_entry_code(entry);
entry              19 arch/arm64/kernel/jump_label.c 		insn = aarch64_insn_gen_branch_imm(jump_entry_code(entry),
entry              20 arch/arm64/kernel/jump_label.c 						   jump_entry_target(entry),
entry              29 arch/arm64/kernel/jump_label.c void arch_jump_label_transform_static(struct jump_entry *entry,
entry              85 arch/arm64/kernel/machine_kexec.c 	kimage_entry_t *entry;
entry              87 arch/arm64/kernel/machine_kexec.c 	for (entry = &kimage->head; ; entry++) {
entry              92 arch/arm64/kernel/machine_kexec.c 		__flush_dcache_area(entry, sizeof(kimage_entry_t));
entry              94 arch/arm64/kernel/machine_kexec.c 		flag = *entry & IND_FLAGS;
entry              98 arch/arm64/kernel/machine_kexec.c 		addr = phys_to_virt(*entry & PAGE_MASK);
entry             103 arch/arm64/kernel/machine_kexec.c 			entry = (kimage_entry_t *)addr - 1;
entry              99 arch/arm64/kernel/pci.c 	struct resource_entry *entry, *tmp;
entry             103 arch/arm64/kernel/pci.c 	resource_list_for_each_entry_safe(entry, tmp, &ci->resources) {
entry             104 arch/arm64/kernel/pci.c 		if (!(entry->res->flags & IORESOURCE_WINDOW))
entry             105 arch/arm64/kernel/pci.c 			resource_list_destroy_entry(entry);
entry              24 arch/arm64/kernel/perf_callchain.c 	       struct perf_callchain_entry_ctx *entry)
entry              43 arch/arm64/kernel/perf_callchain.c 	perf_callchain_store(entry, lr);
entry              72 arch/arm64/kernel/perf_callchain.c 		      struct perf_callchain_entry_ctx *entry)
entry              88 arch/arm64/kernel/perf_callchain.c 	perf_callchain_store(entry, buftail.lr);
entry             102 arch/arm64/kernel/perf_callchain.c void perf_callchain_user(struct perf_callchain_entry_ctx *entry,
entry             110 arch/arm64/kernel/perf_callchain.c 	perf_callchain_store(entry, regs->pc);
entry             118 arch/arm64/kernel/perf_callchain.c 		while (entry->nr < entry->max_stack &&
entry             120 arch/arm64/kernel/perf_callchain.c 			tail = user_backtrace(tail, entry);
entry             128 arch/arm64/kernel/perf_callchain.c 		while ((entry->nr < entry->max_stack) &&
entry             130 arch/arm64/kernel/perf_callchain.c 			tail = compat_user_backtrace(tail, entry);
entry             142 arch/arm64/kernel/perf_callchain.c 	struct perf_callchain_entry_ctx *entry = data;
entry             143 arch/arm64/kernel/perf_callchain.c 	perf_callchain_store(entry, frame->pc);
entry             147 arch/arm64/kernel/perf_callchain.c void perf_callchain_kernel(struct perf_callchain_entry_ctx *entry,
entry             158 arch/arm64/kernel/perf_callchain.c 	walk_stackframe(current, &frame, callchain_trace, entry);
entry             199 arch/arm64/mm/fault.c 			  pte_t entry, int dirty)
entry             204 arch/arm64/mm/fault.c 	if (pte_same(pte, entry))
entry             208 arch/arm64/mm/fault.c 	pte_val(entry) &= PTE_RDONLY | PTE_AF | PTE_WRITE | PTE_DIRTY;
entry             216 arch/arm64/mm/fault.c 	pte_val(entry) ^= PTE_RDONLY;
entry             221 arch/arm64/mm/fault.c 		pteval |= pte_val(entry);
entry             297 arch/arm64/mm/hugetlbpage.c pte_t arch_make_huge_pte(pte_t entry, struct vm_area_struct *vma,
entry             303 arch/arm64/mm/hugetlbpage.c 		entry = pte_mkcont(entry);
entry             305 arch/arm64/mm/hugetlbpage.c 		entry = pmd_pte(pmd_mkcont(pte_pmd(entry)));
entry             310 arch/arm64/mm/hugetlbpage.c 	return entry;
entry              36 arch/csky/kernel/perf_callchain.c 			struct perf_callchain_entry_ctx *entry)
entry              39 arch/csky/kernel/perf_callchain.c 		perf_callchain_store(entry, fr->lr);
entry              47 arch/csky/kernel/perf_callchain.c static unsigned long user_backtrace(struct perf_callchain_entry_ctx *entry,
entry              67 arch/csky/kernel/perf_callchain.c 	perf_callchain_store(entry, lr);
entry              86 arch/csky/kernel/perf_callchain.c void perf_callchain_user(struct perf_callchain_entry_ctx *entry,
entry              96 arch/csky/kernel/perf_callchain.c 	perf_callchain_store(entry, regs->pc);
entry             104 arch/csky/kernel/perf_callchain.c 	fp = user_backtrace(entry, fp, regs->lr);
entry             106 arch/csky/kernel/perf_callchain.c 	while (fp && !(fp & 0x3) && entry->nr < entry->max_stack)
entry             107 arch/csky/kernel/perf_callchain.c 		fp = user_backtrace(entry, fp, 0);
entry             110 arch/csky/kernel/perf_callchain.c void perf_callchain_kernel(struct perf_callchain_entry_ctx *entry,
entry             123 arch/csky/kernel/perf_callchain.c 	walk_stackframe(&fr, entry);
entry              64 arch/ia64/hp/common/aml_nfw.c 	ia64_sal_handler entry;
entry              69 arch/ia64/hp/common/aml_nfw.c 	entry = (ia64_sal_handler) &virt_entry;
entry              71 arch/ia64/hp/common/aml_nfw.c 	IA64_FW_CALL(entry, c->ret,
entry             476 arch/ia64/include/asm/pgtable.h #define __swp_type(entry)		(((entry).val >> 1) & 0x7f)
entry             477 arch/ia64/include/asm/pgtable.h #define __swp_offset(entry)		(((entry).val << 1) >> 9)
entry              50 arch/ia64/include/asm/sal.h #define __IA64_FW_CALL(entry,result,a0,a1,a2,a3,a4,a5,a6,a7)	\
entry              51 arch/ia64/include/asm/sal.h 	result = (*entry)(a0,a1,a2,a3,a4,a5,a6,a7)
entry              53 arch/ia64/include/asm/sal.h # define IA64_FW_CALL(entry,result,args...) do {		\
entry              58 arch/ia64/include/asm/sal.h 	__IA64_FW_CALL(entry, result, args);			\
entry             507 arch/ia64/kernel/acpi.c 			    slit_table->entry[i * slit_table->locality_count + j];
entry             982 arch/ia64/kernel/kprobes.c unsigned long arch_deref_entry_point(void *entry)
entry             984 arch/ia64/kernel/kprobes.c 	return ((struct fnptr *)entry)->ip;
entry              51 arch/ia64/kernel/palinfo.c 	struct proc_dir_entry	*entry;		/* registered entry (removal) */
entry            1205 arch/ia64/kernel/perfmon.c 	pfm_buffer_fmt_t * entry;
entry            1208 arch/ia64/kernel/perfmon.c 		entry = list_entry(pos, pfm_buffer_fmt_t, fmt_list);
entry            1209 arch/ia64/kernel/perfmon.c 		if (pfm_uuid_cmp(uuid, entry->fmt_uuid) == 0)
entry            1210 arch/ia64/kernel/perfmon.c 			return entry;
entry            5544 arch/ia64/kernel/perfmon.c 	pfm_buffer_fmt_t * entry;
entry            5578 arch/ia64/kernel/perfmon.c 		entry = list_entry(pos, pfm_buffer_fmt_t, fmt_list);
entry            5580 arch/ia64/kernel/perfmon.c 			   entry->fmt_uuid, entry->fmt_name);
entry              61 arch/ia64/kernel/salinfo.c 	struct proc_dir_entry	*entry;		/* registered entry (removal) */
entry             596 arch/ia64/kernel/salinfo.c 	struct proc_dir_entry *dir, *entry;
entry             619 arch/ia64/kernel/salinfo.c 		entry = proc_create_data("event", S_IRUSR, dir,
entry             621 arch/ia64/kernel/salinfo.c 		if (!entry)
entry             623 arch/ia64/kernel/salinfo.c 		*sdir++ = entry;
entry             625 arch/ia64/kernel/salinfo.c 		entry = proc_create_data("data", S_IRUSR | S_IWUSR, dir,
entry             627 arch/ia64/kernel/salinfo.c 		if (!entry)
entry             629 arch/ia64/kernel/salinfo.c 		*sdir++ = entry;
entry            2189 arch/ia64/kernel/unwind.c 	const struct unw_table_entry *entry, *start, *end;
entry            2213 arch/ia64/kernel/unwind.c 	for (entry = start; entry < end; ++entry)
entry            2214 arch/ia64/kernel/unwind.c 		size += 3*8 + 8 + 8*UNW_LENGTH(*(u64 *) (segbase + entry->info_offset));
entry            2228 arch/ia64/kernel/unwind.c 	for (entry = start; entry < end; ++entry, lp += 3) {
entry            2229 arch/ia64/kernel/unwind.c 		info_size = 8 + 8*UNW_LENGTH(*(u64 *) (segbase + entry->info_offset));
entry            2231 arch/ia64/kernel/unwind.c 		memcpy(info, (char *) segbase + entry->info_offset, info_size);
entry            2233 arch/ia64/kernel/unwind.c 		lp[0] = segbase + entry->start_offset;		/* start */
entry            2234 arch/ia64/kernel/unwind.c 		lp[1] = segbase + entry->end_offset;		/* end */
entry              69 arch/ia64/mm/hugetlbpage.c #define mk_pte_huge(entry) { pte_val(entry) |= _PAGE_P; }
entry             152 arch/ia64/pci/pci.c 			struct resource_entry *entry)
entry             155 arch/ia64/pci/pci.c 	struct resource *resource, *res = entry->res;
entry             170 arch/ia64/pci/pci.c 	space_nr = new_space(entry->offset, sparse);
entry             175 arch/ia64/pci/pci.c 	min = res->start - entry->offset;
entry             176 arch/ia64/pci/pci.c 	max = res->end - entry->offset;
entry             202 arch/ia64/pci/pci.c 	entry->offset = base_port;
entry             239 arch/ia64/pci/pci.c 	struct resource_entry *entry, *tmp;
entry             245 arch/ia64/pci/pci.c 		resource_list_for_each_entry_safe(entry, tmp, &ci->resources) {
entry             246 arch/ia64/pci/pci.c 			res = entry->res;
entry             253 arch/ia64/pci/pci.c 					resource_list_del(entry);
entry             255 arch/ia64/pci/pci.c 							entry->res);
entry             256 arch/ia64/pci/pci.c 					resource_list_add_tail(entry,
entry             260 arch/ia64/pci/pci.c 				if (resource_is_pcicfg_ioport(entry->res))
entry             261 arch/ia64/pci/pci.c 					resource_list_destroy_entry(entry);
entry             262 arch/ia64/pci/pci.c 				else if (add_io_space(dev, info, entry))
entry             263 arch/ia64/pci/pci.c 					resource_list_destroy_entry(entry);
entry             274 arch/ia64/pci/pci.c 	struct resource_entry *entry, *tmp;
entry             277 arch/ia64/pci/pci.c 	resource_list_for_each_entry_safe(entry, tmp, &info->io_resources) {
entry             278 arch/ia64/pci/pci.c 		release_resource(entry->res);
entry             279 arch/ia64/pci/pci.c 		resource_list_destroy_entry(entry);
entry              88 arch/m68k/include/asm/sun3mmu.h         register unsigned long entry;
entry              95 arch/m68k/include/asm/sun3mmu.h         entry = c;
entry              97 arch/m68k/include/asm/sun3mmu.h         return entry;
entry             101 arch/m68k/include/asm/sun3mmu.h static inline void sun3_put_segmap(unsigned long addr, unsigned long entry)
entry             107 arch/m68k/include/asm/sun3mmu.h         SET_CONTROL_BYTE (AC_SEGMAP | (addr & SUN3_CONTROL_MASK), entry);
entry             116 arch/m68k/include/asm/sun3mmu.h         register unsigned long entry;
entry             121 arch/m68k/include/asm/sun3mmu.h         GET_CONTROL_WORD (AC_PAGEMAP | (addr & SUN3_CONTROL_MASK), entry);
entry             124 arch/m68k/include/asm/sun3mmu.h         return entry;
entry             128 arch/m68k/include/asm/sun3mmu.h static inline void sun3_put_pte(unsigned long addr, unsigned long entry)
entry             134 arch/m68k/include/asm/sun3mmu.h         SET_CONTROL_WORD (AC_PAGEMAP | (addr & SUN3_CONTROL_MASK), entry);
entry             506 arch/microblaze/include/asm/pgtable.h #define __swp_type(entry)		((entry).val & 0x3f)
entry             507 arch/microblaze/include/asm/pgtable.h #define __swp_offset(entry)	((entry).val >> 6)
entry              43 arch/mips/boot/ecoff.h 	int32_t	entry;		/* entry pt.				*/
entry             239 arch/mips/boot/elf2ecoff.c 	a->entry = swab32(a->entry);
entry             425 arch/mips/boot/elf2ecoff.c 	eah.entry = ex.e_entry;
entry             196 arch/mips/cavium-octeon/smp.c 	unsigned long entry = (unsigned long)kernel_entry;
entry             199 arch/mips/cavium-octeon/smp.c 	octeon_processor_relocated_kernel_entry = entry + offset;
entry             135 arch/mips/include/asm/mach-cavium-octeon/kernel-entry-init.h 	# Jump to the normal Linux SMP entry point
entry            2552 arch/mips/include/asm/octeon/cvmx-pip-defs.h 		uint64_t entry:62;
entry            2554 arch/mips/include/asm/octeon/cvmx-pip-defs.h 		uint64_t entry:62;
entry             790 arch/mips/include/asm/sn/sn0/hubio.h 	u64	entry;
entry             211 arch/mips/kernel/machine_kexec.c 	unsigned long entry;
entry             236 arch/mips/kernel/machine_kexec.c 	for (ptr = &image->head; (entry = *ptr) && !(entry &IND_DONE);
entry             237 arch/mips/kernel/machine_kexec.c 	     ptr = (entry & IND_INDIRECTION) ?
entry             238 arch/mips/kernel/machine_kexec.c 	       phys_to_virt(entry & PAGE_MASK) : ptr + 1) {
entry              26 arch/mips/kernel/perf_event.c static void save_raw_perf_callchain(struct perf_callchain_entry_ctx *entry,
entry              35 arch/mips/kernel/perf_event.c 			perf_callchain_store(entry, addr);
entry              36 arch/mips/kernel/perf_event.c 			if (entry->nr >= entry->max_stack)
entry              42 arch/mips/kernel/perf_event.c void perf_callchain_kernel(struct perf_callchain_entry_ctx *entry,
entry              55 arch/mips/kernel/perf_event.c 			save_raw_perf_callchain(entry, sp);
entry              59 arch/mips/kernel/perf_event.c 		perf_callchain_store(entry, pc);
entry              60 arch/mips/kernel/perf_event.c 		if (entry->nr >= entry->max_stack)
entry              65 arch/mips/kernel/perf_event.c 	save_raw_perf_callchain(entry, sp);
entry             118 arch/mips/kernel/pm-cps.c 	cps_nc_entry_fn entry;
entry             123 arch/mips/kernel/pm-cps.c 	entry = per_cpu(nc_asm_enter, core)[state];
entry             124 arch/mips/kernel/pm-cps.c 	if (!entry)
entry             170 arch/mips/kernel/pm-cps.c 	left = entry(online, nc_core_ready_count);
entry              78 arch/mips/kvm/mmu.c 	unsigned long entry;
entry              81 arch/mips/kvm/mmu.c 	entry = (unsigned long)invalid_pte_table;
entry              83 arch/mips/kvm/mmu.c 	entry = (unsigned long)invalid_pmd_table;
entry              90 arch/mips/kvm/mmu.c 		p[0] = entry;
entry              91 arch/mips/kvm/mmu.c 		p[1] = entry;
entry              92 arch/mips/kvm/mmu.c 		p[2] = entry;
entry              93 arch/mips/kvm/mmu.c 		p[3] = entry;
entry              94 arch/mips/kvm/mmu.c 		p[4] = entry;
entry              96 arch/mips/kvm/mmu.c 		p[-3] = entry;
entry              97 arch/mips/kvm/mmu.c 		p[-2] = entry;
entry              98 arch/mips/kvm/mmu.c 		p[-1] = entry;
entry             695 arch/mips/kvm/mmu.c 	pte_t *ptep, entry, old_pte;
entry             765 arch/mips/kvm/mmu.c 	entry = pfn_pte(pfn, __pgprot(prot_bits));
entry             769 arch/mips/kvm/mmu.c 	set_pte(ptep, entry);
entry             395 arch/mips/kvm/tlb.c 	int entry;
entry             412 arch/mips/kvm/tlb.c 	for (entry = 0; entry < current_cpu_data.tlbsize; entry++) {
entry             413 arch/mips/kvm/tlb.c 		write_c0_index(entry);
entry             423 arch/mips/kvm/tlb.c 		write_c0_entryhi(UNIQUE_ENTRYHI(entry));
entry             453 arch/mips/kvm/tlb.c 	int entry;
entry             478 arch/mips/kvm/tlb.c 	for (entry = 0; entry < current_cpu_data.guest.tlbsize; entry++) {
entry             480 arch/mips/kvm/tlb.c 		write_gc0_index(entry);
entry             481 arch/mips/kvm/tlb.c 		write_gc0_entryhi(UNIQUE_GUEST_ENTRYHI(entry));
entry              20 arch/mips/mm/pgtable-64.c 	unsigned long entry;
entry              23 arch/mips/mm/pgtable-64.c 	entry = (unsigned long)invalid_pud_table;
entry              25 arch/mips/mm/pgtable-64.c 	entry = (unsigned long)invalid_pmd_table;
entry              27 arch/mips/mm/pgtable-64.c 	entry = (unsigned long)invalid_pte_table;
entry              34 arch/mips/mm/pgtable-64.c 		p[0] = entry;
entry              35 arch/mips/mm/pgtable-64.c 		p[1] = entry;
entry              36 arch/mips/mm/pgtable-64.c 		p[2] = entry;
entry              37 arch/mips/mm/pgtable-64.c 		p[3] = entry;
entry              38 arch/mips/mm/pgtable-64.c 		p[4] = entry;
entry              40 arch/mips/mm/pgtable-64.c 		p[-3] = entry;
entry              41 arch/mips/mm/pgtable-64.c 		p[-2] = entry;
entry              42 arch/mips/mm/pgtable-64.c 		p[-1] = entry;
entry              43 arch/mips/mm/tlb-r3k.c static void local_flush_tlb_from(int entry)
entry              49 arch/mips/mm/tlb-r3k.c 	while (entry < current_cpu_data.tlbsize) {
entry              50 arch/mips/mm/tlb-r3k.c 		write_c0_index(entry << 8);
entry              51 arch/mips/mm/tlb-r3k.c 		write_c0_entryhi((entry | 0x80000) << 12);
entry              52 arch/mips/mm/tlb-r3k.c 		entry++;				/* BARRIER */
entry              59 arch/mips/mm/tlb-r4k.c 	int entry, ftlbhighset;
entry              68 arch/mips/mm/tlb-r4k.c 	entry = num_wired_entries();
entry              74 arch/mips/mm/tlb-r4k.c 	if (cpu_has_tlbinv && !entry) {
entry              82 arch/mips/mm/tlb-r4k.c 		for (entry = current_cpu_data.tlbsizevtlb;
entry              83 arch/mips/mm/tlb-r4k.c 		     entry < ftlbhighset;
entry              84 arch/mips/mm/tlb-r4k.c 		     entry++) {
entry              85 arch/mips/mm/tlb-r4k.c 			write_c0_index(entry);
entry              90 arch/mips/mm/tlb-r4k.c 		while (entry < current_cpu_data.tlbsize) {
entry              92 arch/mips/mm/tlb-r4k.c 			write_c0_entryhi(UNIQUE_ENTRYHI(entry));
entry              93 arch/mips/mm/tlb-r4k.c 			write_c0_index(entry);
entry              96 arch/mips/mm/tlb-r4k.c 			entry++;
entry            2568 arch/mips/mm/tlbex.c 	unsigned long entry;
entry            2582 arch/mips/mm/tlbex.c 	entry = read_c0_entrylo0();
entry            2585 arch/mips/mm/tlbex.c 	entry &= ~((1 << MIPS_ENTRYLO_PFN_SHIFT) - 1);
entry            2586 arch/mips/mm/tlbex.c 	entry &= ~(MIPS_ENTRYLO_RI | MIPS_ENTRYLO_XI);
entry            2589 arch/mips/mm/tlbex.c 	pabits = fls_long(entry) + 6;
entry             187 arch/mips/pci/msi-octeon.c 	struct msi_desc *entry;
entry             203 arch/mips/pci/msi-octeon.c 	for_each_pci_msi_entry(entry, dev) {
entry             204 arch/mips/pci/msi-octeon.c 		ret = arch_setup_msi_irq(dev, entry);
entry              72 arch/mips/pci/pci-alchemy.c static void mod_wired_entry(int entry, unsigned long entrylo0,
entry              82 arch/mips/pci/pci-alchemy.c 	write_c0_index(entry);
entry              34 arch/mips/tools/elf-entry.c 	uint64_t entry;
entry              64 arch/mips/tools/elf-entry.c 			entry = le32toh(hdr.ehdr32.e_entry);
entry              67 arch/mips/tools/elf-entry.c 			entry = be32toh(hdr.ehdr32.e_entry);
entry              74 arch/mips/tools/elf-entry.c 		entry = (int64_t)(int32_t)entry;
entry              80 arch/mips/tools/elf-entry.c 			entry = le64toh(hdr.ehdr64.e_entry);
entry              83 arch/mips/tools/elf-entry.c 			entry = be64toh(hdr.ehdr64.e_entry);
entry              94 arch/mips/tools/elf-entry.c 	printf("0x%016" PRIx64 "\n", entry);
entry            1290 arch/nds32/kernel/perf_event_cpu.c 	struct perf_callchain_entry_ctx *entry = data;
entry            1292 arch/nds32/kernel/perf_event_cpu.c 	perf_callchain_store(entry, fr->lp);
entry            1301 arch/nds32/kernel/perf_event_cpu.c user_backtrace(struct perf_callchain_entry_ctx *entry, unsigned long fp)
entry            1320 arch/nds32/kernel/perf_event_cpu.c 	perf_callchain_store(entry, lp);
entry            1325 arch/nds32/kernel/perf_event_cpu.c user_backtrace_opt_size(struct perf_callchain_entry_ctx *entry,
entry            1347 arch/nds32/kernel/perf_event_cpu.c 	perf_callchain_store(entry, lp);
entry            1363 arch/nds32/kernel/perf_event_cpu.c perf_callchain_user(struct perf_callchain_entry_ctx *entry,
entry            1379 arch/nds32/kernel/perf_event_cpu.c 	perf_callchain_store(entry, regs->ipc);
entry            1384 arch/nds32/kernel/perf_event_cpu.c 	if (entry->nr < PERF_MAX_STACK_DEPTH &&
entry            1435 arch/nds32/kernel/perf_event_cpu.c 				perf_callchain_store(entry, lp);
entry            1438 arch/nds32/kernel/perf_event_cpu.c 				while ((entry->nr < PERF_MAX_STACK_DEPTH) &&
entry            1443 arch/nds32/kernel/perf_event_cpu.c 					fp = user_backtrace_opt_size(entry, fp);
entry            1451 arch/nds32/kernel/perf_event_cpu.c 				perf_callchain_store(entry, lp);
entry            1452 arch/nds32/kernel/perf_event_cpu.c 				while ((entry->nr < PERF_MAX_STACK_DEPTH) &&
entry            1457 arch/nds32/kernel/perf_event_cpu.c 					fp = user_backtrace(entry, fp);
entry            1463 arch/nds32/kernel/perf_event_cpu.c 			perf_callchain_store(entry, lp);
entry            1466 arch/nds32/kernel/perf_event_cpu.c 			while ((entry->nr < PERF_MAX_STACK_DEPTH) &&
entry            1470 arch/nds32/kernel/perf_event_cpu.c 				fp = user_backtrace(entry, fp);
entry            1479 arch/nds32/kernel/perf_event_cpu.c perf_callchain_kernel(struct perf_callchain_entry_ctx *entry,
entry            1491 arch/nds32/kernel/perf_event_cpu.c 	walk_stackframe(&fr, callchain_trace, entry);
entry             269 arch/nds32/kernel/traps.c void do_debug_trap(unsigned long entry, unsigned long addr,
entry             295 arch/nds32/kernel/traps.c void unhandled_exceptions(unsigned long entry, unsigned long addr,
entry             298 arch/nds32/kernel/traps.c 	pr_emerg("Unhandled Exception: entry: %lx addr:%lx itype:%lx\n", entry,
entry             306 arch/nds32/kernel/traps.c extern int do_page_fault(unsigned long entry, unsigned long addr,
entry             313 arch/nds32/kernel/traps.c void do_dispatch_tlb_misc(unsigned long entry, unsigned long addr,
entry             319 arch/nds32/kernel/traps.c 		do_page_fault(entry, addr, type, regs);
entry             321 arch/nds32/kernel/traps.c 		unhandled_exceptions(entry, addr, type, regs);
entry             337 arch/nds32/kernel/traps.c void do_dispatch_general(unsigned long entry, unsigned long addr,
entry             358 arch/nds32/kernel/traps.c 		do_page_fault(entry, addr, type, regs);
entry             376 arch/nds32/kernel/traps.c 		unhandled_exceptions(entry, addr, type, regs);
entry             398 arch/nds32/kernel/traps.c 		do_debug_trap(entry, addr, itype, regs);
entry             401 arch/nds32/kernel/traps.c 		unhandled_exceptions(entry, addr, type, regs);
entry              70 arch/nds32/mm/fault.c void do_page_fault(unsigned long entry, unsigned long addr,
entry              99 arch/nds32/mm/fault.c 		    && (entry == ENTRY_PTE_NOT_PRESENT))
entry             106 arch/nds32/mm/fault.c 	if (entry == ENTRY_GENERAL_EXCPETION
entry             168 arch/nds32/mm/fault.c 	if (entry == ENTRY_PTE_NOT_PRESENT) {
entry             174 arch/nds32/mm/fault.c 	} else if (entry == ENTRY_TLB_MISC) {
entry             273 arch/nds32/mm/fault.c 		tsk->thread.trap_no = entry;
entry             290 arch/nds32/mm/fault.c 		const struct exception_table_entry *entry;
entry             292 arch/nds32/mm/fault.c 		if ((entry =
entry             296 arch/nds32/mm/fault.c 			instruction_pointer(regs) = entry->fixup;
entry             342 arch/nds32/mm/fault.c 	tsk->thread.trap_no = entry;
entry             357 arch/openrisc/kernel/traps.c 	const struct exception_table_entry *entry;
entry             378 arch/openrisc/kernel/traps.c 		if ((entry = search_exception_tables(orig_pc))) {
entry             379 arch/openrisc/kernel/traps.c 			regs->pc = entry->fixup;
entry             399 arch/openrisc/kernel/traps.c 	const struct exception_table_entry *entry;
entry             425 arch/openrisc/kernel/traps.c 		if ((entry = search_exception_tables(orig_pc))) {
entry             426 arch/openrisc/kernel/traps.c 			regs->pc = entry->fixup;
entry             228 arch/openrisc/mm/fault.c 		const struct exception_table_entry *entry;
entry             232 arch/openrisc/mm/fault.c 		if ((entry = search_exception_tables(regs->pc)) != NULL) {
entry             234 arch/openrisc/mm/fault.c 			regs->pc = entry->fixup;
entry              33 arch/parisc/boot/compressed/misc.c extern void startup_continue(void *entry, unsigned long cmdline,
entry              29 arch/parisc/include/asm/linkage.h 	.entry		ASM_NL\
entry              25 arch/parisc/kernel/alternative.c 	struct alt_instr *entry;
entry              29 arch/parisc/kernel/alternative.c 	for (entry = start; entry < end; entry++, index++) {
entry              34 arch/parisc/kernel/alternative.c 		from = (u32 *)((ulong)&entry->orig_offset + entry->orig_offset);
entry              35 arch/parisc/kernel/alternative.c 		len = entry->len;
entry              36 arch/parisc/kernel/alternative.c 		cond = entry->cond;
entry              37 arch/parisc/kernel/alternative.c 		replacement = entry->replacement;
entry              87 arch/parisc/kernel/alternative.c 			source = (u32 *)((ulong)&entry->replacement + entry->replacement);
entry              21 arch/parisc/kernel/jump_label.c void arch_jump_label_transform(struct jump_entry *entry,
entry              24 arch/parisc/kernel/jump_label.c 	void *addr = (void *)jump_entry_code(entry);
entry              28 arch/parisc/kernel/jump_label.c 		void *target = (void *)jump_entry_target(entry);
entry              46 arch/parisc/kernel/jump_label.c void arch_jump_label_transform_static(struct jump_entry *entry,
entry             869 arch/parisc/kernel/module.c 	Elf_Fdesc *entry;
entry             872 arch/parisc/kernel/module.c 	entry = (Elf_Fdesc *)me->init;
entry             873 arch/parisc/kernel/module.c 	printk("FINALIZE, ->init FPTR is %p, GP %lx ADDR %lx\n", entry,
entry             874 arch/parisc/kernel/module.c 	       entry->gp, entry->addr);
entry             875 arch/parisc/kernel/module.c 	addr = (u32 *)entry->addr;
entry             118 arch/parisc/mm/hugetlbpage.c 		     pte_t *ptep, pte_t entry)
entry             127 arch/parisc/mm/hugetlbpage.c 		set_pte(ptep, entry);
entry             131 arch/parisc/mm/hugetlbpage.c 		pte_val(entry) += PAGE_SIZE;
entry             138 arch/parisc/mm/hugetlbpage.c 		     pte_t *ptep, pte_t entry)
entry             143 arch/parisc/mm/hugetlbpage.c 	__set_huge_pte_at(mm, addr, ptep, entry);
entry             152 arch/parisc/mm/hugetlbpage.c 	pte_t entry;
entry             155 arch/parisc/mm/hugetlbpage.c 	entry = *ptep;
entry             159 arch/parisc/mm/hugetlbpage.c 	return entry;
entry              15 arch/powerpc/boot/opal.c 	u64 entry;
entry              75 arch/powerpc/boot/opal.c 	if (getprop(opal_node, "opal-entry-address", &opal.entry, sizeof(u64)) < 0)
entry              77 arch/powerpc/boot/opal.c 	opal.entry = be64_to_cpu(opal.entry);
entry              46 arch/powerpc/boot/rs6000.h   unsigned char	entry[4];	/* entry pt.			*/
entry             332 arch/powerpc/include/asm/book3s/32/pgtable.h 					   pte_t *ptep, pte_t entry,
entry             336 arch/powerpc/include/asm/book3s/32/pgtable.h 	unsigned long set = pte_val(entry) &
entry             375 arch/powerpc/include/asm/book3s/32/pgtable.h #define __swp_type(entry)		((entry).val & 0x1f)
entry             376 arch/powerpc/include/asm/book3s/32/pgtable.h #define __swp_offset(entry)		((entry).val >> 5)
entry             185 arch/powerpc/include/asm/book3s/64/hash.h static inline void hash__ptep_set_access_flags(pte_t *ptep, pte_t entry)
entry             192 arch/powerpc/include/asm/book3s/64/hash.h 	val = pte_raw(entry) & mask;
entry             790 arch/powerpc/include/asm/book3s/64/pgtable.h 					   pte_t *ptep, pte_t entry,
entry             795 arch/powerpc/include/asm/book3s/64/pgtable.h 		return radix__ptep_set_access_flags(vma, ptep, entry,
entry             797 arch/powerpc/include/asm/book3s/64/pgtable.h 	return hash__ptep_set_access_flags(ptep, entry);
entry            1239 arch/powerpc/include/asm/book3s/64/pgtable.h 				 pmd_t entry, int dirty);
entry             125 arch/powerpc/include/asm/book3s/64/radix.h 					 pte_t entry, unsigned long address,
entry              22 arch/powerpc/include/asm/book3s/pgtable.h 				 pte_t *ptep, pte_t entry, int dirty);
entry             112 arch/powerpc/include/asm/code-patching.h 	return ((func_descr_t *)func)->entry;
entry             107 arch/powerpc/include/asm/eeh.h 		list_for_each_entry_safe(edev, tmp, &pe->edevs, entry)
entry             146 arch/powerpc/include/asm/eeh.h 	struct list_head entry;		/* Membership in eeh_pe.edevs	*/
entry             119 arch/powerpc/include/asm/iommu.h #define IOMMU_TABLE_USERSPACE_ENTRY_RO(tbl, entry) \
entry             120 arch/powerpc/include/asm/iommu.h 		((tbl)->it_ops->useraddrptr((tbl), (entry), false))
entry             121 arch/powerpc/include/asm/iommu.h #define IOMMU_TABLE_USERSPACE_ENTRY(tbl, entry) \
entry             122 arch/powerpc/include/asm/iommu.h 		((tbl)->it_ops->useraddrptr((tbl), (entry), true))
entry             211 arch/powerpc/include/asm/iommu.h 		unsigned long entry, unsigned long *hpa,
entry             215 arch/powerpc/include/asm/iommu.h 		unsigned long entry, unsigned long *hpa,
entry             218 arch/powerpc/include/asm/iommu.h 		unsigned long entry, unsigned long pages);
entry             482 arch/powerpc/include/asm/kvm_host.h 	struct mmio_hpte_cache_entry entry[MMIO_HPTE_CACHE_SIZE];
entry             188 arch/powerpc/include/asm/lppaca.h extern void (*dtl_consumer)(struct dtl_entry *entry, u64 index);
entry             316 arch/powerpc/include/asm/nohash/32/pgtable.h 					   pte_t *ptep, pte_t entry,
entry             322 arch/powerpc/include/asm/nohash/32/pgtable.h 	unsigned long set = pte_val(entry) & pte_val(pte_set);
entry             323 arch/powerpc/include/asm/nohash/32/pgtable.h 	unsigned long clr = ~pte_val(entry) & ~pte_val(pte_clr);
entry             381 arch/powerpc/include/asm/nohash/32/pgtable.h #define __swp_type(entry)		((entry).val & 0x1f)
entry             382 arch/powerpc/include/asm/nohash/32/pgtable.h #define __swp_offset(entry)		((entry).val >> 5)
entry             306 arch/powerpc/include/asm/nohash/64/pgtable.h 					   pte_t *ptep, pte_t entry,
entry             310 arch/powerpc/include/asm/nohash/64/pgtable.h 	unsigned long bits = pte_val(entry) &
entry             231 arch/powerpc/include/asm/nohash/pgtable.h 				 pte_t *ptep, pte_t entry, int dirty);
entry             983 arch/powerpc/include/asm/opal-api.h 	struct opal_sg_entry entry[];
entry             266 arch/powerpc/include/asm/ppc_asm.h #define _ASM_NOKPROBE_SYMBOL(entry)			\
entry             268 arch/powerpc/include/asm/ppc_asm.h 	PPC_LONG (entry) ;				\
entry             271 arch/powerpc/include/asm/ppc_asm.h #define _ASM_NOKPROBE_SYMBOL(entry)
entry              56 arch/powerpc/include/asm/rtas.h 	unsigned long entry;		/* physical address pointer */
entry              27 arch/powerpc/include/asm/types.h 	unsigned long entry;
entry             284 arch/powerpc/kernel/asm-offsets.c 	OFFSET(RTASENTRY, rtas_t, entry);
entry             410 arch/powerpc/kernel/eeh_driver.c 		if (list_is_last(&edev->entry, &edev->pe->edevs))
entry             716 arch/powerpc/kernel/eeh_driver.c 		edev = list_first_entry(&pe->edevs, struct eeh_dev, entry);
entry             388 arch/powerpc/kernel/eeh_pe.c 			list_add_tail(&edev->entry, &pe->edevs);
entry             411 arch/powerpc/kernel/eeh_pe.c 			list_add_tail(&edev->entry, &pe->edevs);
entry             453 arch/powerpc/kernel/eeh_pe.c 	list_add_tail(&edev->entry, &pe->edevs);
entry             484 arch/powerpc/kernel/eeh_pe.c 	list_del(&edev->entry);
entry             609 arch/powerpc/kernel/eeh_pe.c 		list_for_each_entry(edev, &pe->edevs, entry) {
entry             934 arch/powerpc/kernel/eeh_pe.c 	edev = list_first_entry_or_null(&pe->edevs, struct eeh_dev, entry);
entry             299 arch/powerpc/kernel/iommu.c 	unsigned long entry;
entry             303 arch/powerpc/kernel/iommu.c 	entry = iommu_range_alloc(dev, tbl, npages, NULL, mask, align_order);
entry             305 arch/powerpc/kernel/iommu.c 	if (unlikely(entry == DMA_MAPPING_ERROR))
entry             308 arch/powerpc/kernel/iommu.c 	entry += tbl->it_offset;	/* Offset into real TCE table */
entry             309 arch/powerpc/kernel/iommu.c 	ret = entry << tbl->it_page_shift;	/* Set the return dma address */
entry             312 arch/powerpc/kernel/iommu.c 	build_fail = tbl->it_ops->set(tbl, entry, npages,
entry             339 arch/powerpc/kernel/iommu.c 	unsigned long entry, free_entry;
entry             341 arch/powerpc/kernel/iommu.c 	entry = dma_addr >> tbl->it_page_shift;
entry             342 arch/powerpc/kernel/iommu.c 	free_entry = entry - tbl->it_offset;
entry             345 arch/powerpc/kernel/iommu.c 	    (entry < tbl->it_offset)) {
entry             348 arch/powerpc/kernel/iommu.c 			printk(KERN_INFO "\tentry     = 0x%lx\n", entry); 
entry             365 arch/powerpc/kernel/iommu.c 				   unsigned long entry)
entry             371 arch/powerpc/kernel/iommu.c 	if (entry >= largepool_start) {
entry             374 arch/powerpc/kernel/iommu.c 		unsigned int pool_nr = entry / tbl->poolsize;
entry             386 arch/powerpc/kernel/iommu.c 	unsigned long entry, free_entry;
entry             390 arch/powerpc/kernel/iommu.c 	entry = dma_addr >> tbl->it_page_shift;
entry             391 arch/powerpc/kernel/iommu.c 	free_entry = entry - tbl->it_offset;
entry             398 arch/powerpc/kernel/iommu.c 	tbl->it_ops->clear(tbl, entry, npages);
entry             447 arch/powerpc/kernel/iommu.c 		unsigned long vaddr, npages, entry, slen;
entry             462 arch/powerpc/kernel/iommu.c 		entry = iommu_range_alloc(dev, tbl, npages, &handle,
entry             468 arch/powerpc/kernel/iommu.c 		if (unlikely(entry == DMA_MAPPING_ERROR)) {
entry             478 arch/powerpc/kernel/iommu.c 		entry += tbl->it_offset;
entry             479 arch/powerpc/kernel/iommu.c 		dma_addr = entry << tbl->it_page_shift;
entry             483 arch/powerpc/kernel/iommu.c 			    npages, entry, dma_addr);
entry             486 arch/powerpc/kernel/iommu.c 		build_fail = tbl->it_ops->set(tbl, entry, npages,
entry            1018 arch/powerpc/kernel/iommu.c 		unsigned long entry, unsigned long *hpa,
entry            1024 arch/powerpc/kernel/iommu.c 	ret = tbl->it_ops->xchg_no_kill(tbl, entry, hpa, direction, false);
entry            1036 arch/powerpc/kernel/iommu.c 		unsigned long entry, unsigned long pages)
entry            1039 arch/powerpc/kernel/iommu.c 		tbl->it_ops->tce_kill(tbl, entry, pages, false);
entry              10 arch/powerpc/kernel/jump_label.c void arch_jump_label_transform(struct jump_entry *entry,
entry              13 arch/powerpc/kernel/jump_label.c 	u32 *addr = (u32 *)(unsigned long)entry->code;
entry              16 arch/powerpc/kernel/jump_label.c 		patch_branch(addr, entry->target, 0);
entry             526 arch/powerpc/kernel/kprobes.c 	const struct exception_table_entry *entry;
entry             570 arch/powerpc/kernel/kprobes.c 		if ((entry = search_exception_tables(regs->nip)) != NULL) {
entry             571 arch/powerpc/kernel/kprobes.c 			regs->nip = extable_fixup(entry);
entry             587 arch/powerpc/kernel/kprobes.c unsigned long arch_deref_entry_point(void *entry)
entry             590 arch/powerpc/kernel/kprobes.c 	if (!kernel_text_address((unsigned long)entry))
entry             591 arch/powerpc/kernel/kprobes.c 		return ppc_global_function_entry(entry);
entry             594 arch/powerpc/kernel/kprobes.c 		return (unsigned long)entry;
entry              77 arch/powerpc/kernel/machine_kexec_64.c 	unsigned long entry;
entry              91 arch/powerpc/kernel/machine_kexec_64.c 	for (entry = ind; !(entry & IND_DONE); entry = *ptr++) {
entry              92 arch/powerpc/kernel/machine_kexec_64.c 		addr = __va(entry & PAGE_MASK);
entry              94 arch/powerpc/kernel/machine_kexec_64.c 		switch (entry & IND_FLAGS) {
entry             582 arch/powerpc/kernel/mce_power.c 	const struct exception_table_entry *entry;
entry             584 arch/powerpc/kernel/mce_power.c 	entry = search_kernel_exception_table(regs->nip);
entry             585 arch/powerpc/kernel/mce_power.c 	if (entry) {
entry             587 arch/powerpc/kernel/mce_power.c 		regs->nip = extable_fixup(entry);
entry             161 arch/powerpc/kernel/module_32.c static inline int entry_matches(struct ppc_plt_entry *entry, Elf32_Addr val)
entry             163 arch/powerpc/kernel/module_32.c 	if (entry->jump[0] != (PPC_INST_ADDIS | __PPC_RT(R12) | PPC_HA(val)))
entry             165 arch/powerpc/kernel/module_32.c 	if (entry->jump[1] != (PPC_INST_ADDI | __PPC_RT(R12) | __PPC_RA(R12) |
entry             177 arch/powerpc/kernel/module_32.c 	struct ppc_plt_entry *entry;
entry             183 arch/powerpc/kernel/module_32.c 		entry = (void *)sechdrs[mod->arch.core_plt_section].sh_addr;
entry             185 arch/powerpc/kernel/module_32.c 		entry = (void *)sechdrs[mod->arch.init_plt_section].sh_addr;
entry             188 arch/powerpc/kernel/module_32.c 	while (entry->jump[0]) {
entry             189 arch/powerpc/kernel/module_32.c 		if (entry_matches(entry, val)) return (uint32_t)entry;
entry             190 arch/powerpc/kernel/module_32.c 		entry++;
entry             199 arch/powerpc/kernel/module_32.c 	entry->jump[0] = PPC_INST_ADDIS | __PPC_RT(R12) | PPC_HA(val);
entry             200 arch/powerpc/kernel/module_32.c 	entry->jump[1] = PPC_INST_ADDI | __PPC_RT(R12) | __PPC_RA(R12) | PPC_LO(val);
entry             201 arch/powerpc/kernel/module_32.c 	entry->jump[2] = PPC_INST_MTCTR | __PPC_RS(R12);
entry             202 arch/powerpc/kernel/module_32.c 	entry->jump[3] = PPC_INST_BCTR;
entry             204 arch/powerpc/kernel/module_32.c 	pr_debug("Initialized plt for 0x%x at %p\n", val, entry);
entry             205 arch/powerpc/kernel/module_32.c 	return (uint32_t)entry;
entry             400 arch/powerpc/kernel/module_64.c 			      struct ppc64_stub_entry *entry,
entry             406 arch/powerpc/kernel/module_64.c 	memcpy(entry->jump, ppc64_stub_insns, sizeof(ppc64_stub_insns));
entry             409 arch/powerpc/kernel/module_64.c 	reladdr = (unsigned long)entry - my_r2(sechdrs, me);
entry             415 arch/powerpc/kernel/module_64.c 	pr_debug("Stub %p get data from reladdr %li\n", entry, reladdr);
entry             417 arch/powerpc/kernel/module_64.c 	entry->jump[0] |= PPC_HA(reladdr);
entry             418 arch/powerpc/kernel/module_64.c 	entry->jump[1] |= PPC_LO(reladdr);
entry             419 arch/powerpc/kernel/module_64.c 	entry->funcdata = func_desc(addr);
entry             420 arch/powerpc/kernel/module_64.c 	entry->magic = STUB_MAGIC;
entry             768 arch/powerpc/kernel/module_64.c 	struct ppc64_stub_entry *entry;
entry             786 arch/powerpc/kernel/module_64.c 	num_stubs = sechdrs[me->arch.stubs_section].sh_size / sizeof(*entry);
entry             789 arch/powerpc/kernel/module_64.c 	entry = (void *)sechdrs[me->arch.stubs_section].sh_addr;
entry             790 arch/powerpc/kernel/module_64.c 	for (i = 0; i < num_stubs && stub_func_addr(entry->funcdata); i++, entry++);
entry             797 arch/powerpc/kernel/module_64.c 	memcpy(entry->jump, stub_insns, sizeof(stub_insns));
entry             807 arch/powerpc/kernel/module_64.c 	entry->jump[1] |= PPC_HA(reladdr);
entry             808 arch/powerpc/kernel/module_64.c 	entry->jump[2] |= PPC_LO(reladdr);
entry             811 arch/powerpc/kernel/module_64.c 	entry->funcdata = func_desc(addr);
entry             812 arch/powerpc/kernel/module_64.c 	entry->magic = STUB_MAGIC;
entry             814 arch/powerpc/kernel/module_64.c 	return (unsigned long)entry;
entry            1740 arch/powerpc/kernel/process.c 		unsigned long entry;
entry            1744 arch/powerpc/kernel/process.c 			entry = start;
entry            1766 arch/powerpc/kernel/process.c 			__get_user(entry, (unsigned long __user *)start);
entry            1773 arch/powerpc/kernel/process.c 				entry += load_addr;
entry            1778 arch/powerpc/kernel/process.c 		regs->nip = entry;
entry             138 arch/powerpc/kernel/prom_init.c extern int enter_prom(struct prom_args *args, unsigned long entry);
entry             140 arch/powerpc/kernel/prom_init.c static inline int enter_prom(struct prom_args *args, unsigned long entry)
entry             142 arch/powerpc/kernel/prom_init.c 	return ((int (*)(struct prom_args *))entry)(args);
entry            1778 arch/powerpc/kernel/prom_init.c 	u32 base, entry = 0;
entry            1807 arch/powerpc/kernel/prom_init.c 	if (call_prom_ret("call-method", 3, 2, &entry,
entry            1810 arch/powerpc/kernel/prom_init.c 	    || entry == 0) {
entry            1821 arch/powerpc/kernel/prom_init.c 	val = cpu_to_be32(entry);
entry            1831 arch/powerpc/kernel/prom_init.c 	prom_debug("rtas entry    = 0x%x\n", entry);
entry            1845 arch/powerpc/kernel/prom_init.c 	u32 entry = 0, size = 0, succ = 0;
entry            1894 arch/powerpc/kernel/prom_init.c 	if (call_prom_ret("call-method", 4, 2, &entry,
entry            1896 arch/powerpc/kernel/prom_init.c 			  ibmvtpm_inst, size, base) != 0 || entry == 0) {
entry             453 arch/powerpc/kernel/rtas.c 	if (!rtas.entry || token == RTAS_UNKNOWN_SERVICE)
entry            1072 arch/powerpc/kernel/rtas.c 	if (!rtas.entry)
entry            1157 arch/powerpc/kernel/rtas.c 	u32 base, size, entry;
entry            1177 arch/powerpc/kernel/rtas.c 	no_entry = of_property_read_u32(rtas.dev, "linux,rtas-entry", &entry);
entry            1178 arch/powerpc/kernel/rtas.c 	rtas.entry = no_entry ? rtas.base : entry;
entry            1214 arch/powerpc/kernel/rtas.c 		rtas.entry = *entryp;
entry             566 arch/powerpc/kernel/rtasd.c 	struct proc_dir_entry *entry;
entry             574 arch/powerpc/kernel/rtasd.c 	entry = proc_create("powerpc/rtas/error_log", 0400, NULL,
entry             576 arch/powerpc/kernel/rtasd.c 	if (!entry)
entry             895 arch/powerpc/kernel/signal_64.c 		err |= get_user(regs->nip, &funct_desc_ptr->entry);
entry             126 arch/powerpc/kernel/trace/ftrace.c 	unsigned long entry, ptr, tramp;
entry             154 arch/powerpc/kernel/trace/ftrace.c 	entry = ppc_global_function_entry((void *)addr);
entry             156 arch/powerpc/kernel/trace/ftrace.c 	if (ptr != entry) {
entry             157 arch/powerpc/kernel/trace/ftrace.c 		pr_err("addr %lx does not match expected %lx\n", ptr, entry);
entry             515 arch/powerpc/kernel/trace/ftrace.c 	unsigned long entry, ptr, tramp;
entry             552 arch/powerpc/kernel/trace/ftrace.c 	entry = ppc_global_function_entry((void *)addr);
entry             554 arch/powerpc/kernel/trace/ftrace.c 	if (ptr != entry) {
entry             555 arch/powerpc/kernel/trace/ftrace.c 		pr_err("addr %lx does not match expected %lx\n", ptr, entry);
entry             618 arch/powerpc/kernel/trace/ftrace.c 	unsigned long tramp, entry, ptr;
entry             621 arch/powerpc/kernel/trace/ftrace.c 	entry = ppc_global_function_entry((void *)ftrace_caller);
entry             624 arch/powerpc/kernel/trace/ftrace.c 	if (ptr != entry) {
entry             626 arch/powerpc/kernel/trace/ftrace.c 		entry = ppc_global_function_entry((void *)ftrace_regs_caller);
entry             627 arch/powerpc/kernel/trace/ftrace.c 		if (ptr != entry) {
entry             705 arch/powerpc/kernel/trace/ftrace.c 	unsigned long entry, ptr, tramp;
entry             728 arch/powerpc/kernel/trace/ftrace.c 	entry = ppc_global_function_entry((void *)old_addr);
entry             732 arch/powerpc/kernel/trace/ftrace.c 	if (tramp != entry) {
entry             742 arch/powerpc/kernel/trace/ftrace.c 		if (ptr != entry) {
entry             743 arch/powerpc/kernel/trace/ftrace.c 			pr_err("addr %lx does not match expected %lx\n", ptr, entry);
entry             771 arch/powerpc/kernel/trace/ftrace.c 	entry = ppc_global_function_entry((void *)addr);
entry             773 arch/powerpc/kernel/trace/ftrace.c 	if (ptr != entry) {
entry             774 arch/powerpc/kernel/trace/ftrace.c 		pr_err("addr %lx does not match expected %lx\n", ptr, entry);
entry             544 arch/powerpc/kernel/traps.c 	const struct exception_table_entry *entry;
entry             548 arch/powerpc/kernel/traps.c 	    && (entry = search_exception_tables(regs->nip)) != NULL) {
entry             570 arch/powerpc/kernel/traps.c 			regs->nip = extable_fixup(entry);
entry             227 arch/powerpc/kvm/book3s_64_mmu_radix.c 	struct prtb_entry entry;
entry             234 arch/powerpc/kvm/book3s_64_mmu_radix.c 	if ((table_index * sizeof(entry)) >= size)
entry             238 arch/powerpc/kvm/book3s_64_mmu_radix.c 	ptbl = (table & PRTB_MASK) + (table_index * sizeof(entry));
entry             239 arch/powerpc/kvm/book3s_64_mmu_radix.c 	ret = kvm_read_guest(kvm, ptbl, &entry, sizeof(entry));
entry             244 arch/powerpc/kvm/book3s_64_mmu_radix.c 	root = be64_to_cpu(entry.prtb0);
entry             414 arch/powerpc/kvm/book3s_64_vio.c 		unsigned long entry)
entry             419 arch/powerpc/kvm/book3s_64_vio.c 	iommu_tce_xchg_no_kill(mm, tbl, entry, &hpa, &dir);
entry             423 arch/powerpc/kvm/book3s_64_vio.c 		struct iommu_table *tbl, unsigned long entry)
entry             427 arch/powerpc/kvm/book3s_64_vio.c 	__be64 *pua = IOMMU_TABLE_USERSPACE_ENTRY_RO(tbl, entry);
entry             444 arch/powerpc/kvm/book3s_64_vio.c 		struct iommu_table *tbl, unsigned long entry)
entry             450 arch/powerpc/kvm/book3s_64_vio.c 	if (WARN_ON_ONCE(iommu_tce_xchg_no_kill(kvm->mm, tbl, entry, &hpa,
entry             457 arch/powerpc/kvm/book3s_64_vio.c 	ret = kvmppc_tce_iommu_mapped_dec(kvm, tbl, entry);
entry             459 arch/powerpc/kvm/book3s_64_vio.c 		iommu_tce_xchg_no_kill(kvm->mm, tbl, entry, &hpa, &dir);
entry             466 arch/powerpc/kvm/book3s_64_vio.c 		unsigned long entry)
entry             470 arch/powerpc/kvm/book3s_64_vio.c 	unsigned long io_entry = entry * subpages;
entry             482 arch/powerpc/kvm/book3s_64_vio.c 		unsigned long entry, unsigned long ua,
entry             487 arch/powerpc/kvm/book3s_64_vio.c 	__be64 *pua = IOMMU_TABLE_USERSPACE_ENTRY(tbl, entry);
entry             505 arch/powerpc/kvm/book3s_64_vio.c 	ret = iommu_tce_xchg_no_kill(kvm->mm, tbl, entry, &hpa, &dir);
entry             512 arch/powerpc/kvm/book3s_64_vio.c 		kvmppc_tce_iommu_mapped_dec(kvm, tbl, entry);
entry             521 arch/powerpc/kvm/book3s_64_vio.c 		unsigned long entry, unsigned long ua,
entry             526 arch/powerpc/kvm/book3s_64_vio.c 	unsigned long io_entry = entry * subpages;
entry             546 arch/powerpc/kvm/book3s_64_vio.c 	unsigned long entry, ua = 0;
entry             573 arch/powerpc/kvm/book3s_64_vio.c 	entry = ioba >> stt->page_shift;
entry             578 arch/powerpc/kvm/book3s_64_vio.c 					stit->tbl, entry);
entry             581 arch/powerpc/kvm/book3s_64_vio.c 					entry, ua, dir);
entry             583 arch/powerpc/kvm/book3s_64_vio.c 		iommu_tce_kill(stit->tbl, entry, 1);
entry             586 arch/powerpc/kvm/book3s_64_vio.c 			kvmppc_clear_tce(vcpu->kvm->mm, stit->tbl, entry);
entry             591 arch/powerpc/kvm/book3s_64_vio.c 	kvmppc_tce_put(stt, entry, tce);
entry             606 arch/powerpc/kvm/book3s_64_vio.c 	unsigned long entry, ua = 0;
entry             615 arch/powerpc/kvm/book3s_64_vio.c 	entry = ioba >> stt->page_shift;
entry             673 arch/powerpc/kvm/book3s_64_vio.c 					stit->tbl, entry + i, ua,
entry             678 arch/powerpc/kvm/book3s_64_vio.c 						entry);
entry             683 arch/powerpc/kvm/book3s_64_vio.c 		kvmppc_tce_put(stt, entry + i, tce);
entry             688 arch/powerpc/kvm/book3s_64_vio.c 		iommu_tce_kill(stit->tbl, entry, npages);
entry             718 arch/powerpc/kvm/book3s_64_vio.c 		unsigned long entry = ioba >> stt->page_shift;
entry             722 arch/powerpc/kvm/book3s_64_vio.c 					stit->tbl, entry + i);
entry             731 arch/powerpc/kvm/book3s_64_vio.c 			kvmppc_clear_tce(vcpu->kvm->mm, stit->tbl, entry);
entry             223 arch/powerpc/kvm/book3s_64_vio_hv.c 		unsigned long entry, unsigned long *hpa,
entry             228 arch/powerpc/kvm/book3s_64_vio_hv.c 	ret = tbl->it_ops->xchg_no_kill(tbl, entry, hpa, direction, true);
entry             232 arch/powerpc/kvm/book3s_64_vio_hv.c 		__be64 *pua = IOMMU_TABLE_USERSPACE_ENTRY_RO(tbl, entry);
entry             245 arch/powerpc/kvm/book3s_64_vio_hv.c 		unsigned long entry, unsigned long pages)
entry             248 arch/powerpc/kvm/book3s_64_vio_hv.c 		tbl->it_ops->tce_kill(tbl, entry, pages, true);
entry             252 arch/powerpc/kvm/book3s_64_vio_hv.c 		unsigned long entry)
entry             257 arch/powerpc/kvm/book3s_64_vio_hv.c 	iommu_tce_xchg_no_kill_rm(kvm->mm, tbl, entry, &hpa, &dir);
entry             261 arch/powerpc/kvm/book3s_64_vio_hv.c 		struct iommu_table *tbl, unsigned long entry)
entry             265 arch/powerpc/kvm/book3s_64_vio_hv.c 	__be64 *pua = IOMMU_TABLE_USERSPACE_ENTRY_RO(tbl, entry);
entry             283 arch/powerpc/kvm/book3s_64_vio_hv.c 		struct iommu_table *tbl, unsigned long entry)
entry             289 arch/powerpc/kvm/book3s_64_vio_hv.c 	if (iommu_tce_xchg_no_kill_rm(kvm->mm, tbl, entry, &hpa, &dir))
entry             299 arch/powerpc/kvm/book3s_64_vio_hv.c 	ret = kvmppc_rm_tce_iommu_mapped_dec(kvm, tbl, entry);
entry             301 arch/powerpc/kvm/book3s_64_vio_hv.c 		iommu_tce_xchg_no_kill_rm(kvm->mm, tbl, entry, &hpa, &dir);
entry             308 arch/powerpc/kvm/book3s_64_vio_hv.c 		unsigned long entry)
entry             312 arch/powerpc/kvm/book3s_64_vio_hv.c 	unsigned long io_entry = entry * subpages;
entry             324 arch/powerpc/kvm/book3s_64_vio_hv.c 		unsigned long entry, unsigned long ua,
entry             329 arch/powerpc/kvm/book3s_64_vio_hv.c 	__be64 *pua = IOMMU_TABLE_USERSPACE_ENTRY_RO(tbl, entry);
entry             347 arch/powerpc/kvm/book3s_64_vio_hv.c 	ret = iommu_tce_xchg_no_kill_rm(kvm->mm, tbl, entry, &hpa, &dir);
entry             358 arch/powerpc/kvm/book3s_64_vio_hv.c 		kvmppc_rm_tce_iommu_mapped_dec(kvm, tbl, entry);
entry             367 arch/powerpc/kvm/book3s_64_vio_hv.c 		unsigned long entry, unsigned long ua,
entry             372 arch/powerpc/kvm/book3s_64_vio_hv.c 	unsigned long io_entry = entry * subpages;
entry             392 arch/powerpc/kvm/book3s_64_vio_hv.c 	unsigned long entry, ua = 0;
entry             418 arch/powerpc/kvm/book3s_64_vio_hv.c 	entry = ioba >> stt->page_shift;
entry             423 arch/powerpc/kvm/book3s_64_vio_hv.c 					stit->tbl, entry);
entry             426 arch/powerpc/kvm/book3s_64_vio_hv.c 					stit->tbl, entry, ua, dir);
entry             428 arch/powerpc/kvm/book3s_64_vio_hv.c 		iommu_tce_kill_rm(stit->tbl, entry, 1);
entry             431 arch/powerpc/kvm/book3s_64_vio_hv.c 			kvmppc_rm_clear_tce(vcpu->kvm, stit->tbl, entry);
entry             436 arch/powerpc/kvm/book3s_64_vio_hv.c 	kvmppc_rm_tce_put(stt, entry, tce);
entry             483 arch/powerpc/kvm/book3s_64_vio_hv.c 	unsigned long tces, entry, ua = 0;
entry             496 arch/powerpc/kvm/book3s_64_vio_hv.c 	entry = ioba >> stt->page_shift;
entry             576 arch/powerpc/kvm/book3s_64_vio_hv.c 					stit->tbl, entry + i, ua,
entry             581 arch/powerpc/kvm/book3s_64_vio_hv.c 						entry);
entry             586 arch/powerpc/kvm/book3s_64_vio_hv.c 		kvmppc_rm_tce_put(stt, entry + i, tce);
entry             591 arch/powerpc/kvm/book3s_64_vio_hv.c 		iommu_tce_kill_rm(stit->tbl, entry, npages);
entry             625 arch/powerpc/kvm/book3s_64_vio_hv.c 		unsigned long entry = ioba >> stt->page_shift;
entry             629 arch/powerpc/kvm/book3s_64_vio_hv.c 					stit->tbl, entry + i);
entry             638 arch/powerpc/kvm/book3s_64_vio_hv.c 			kvmppc_rm_clear_tce(vcpu->kvm, stit->tbl, entry);
entry             762 arch/powerpc/kvm/book3s_hv_nested.c 	struct llist_node *entry = ((struct llist_head *) rmapp)->first;
entry             774 arch/powerpc/kvm/book3s_hv_nested.c 	for_each_nest_rmap_safe(cursor, entry, &rmap) {
entry             828 arch/powerpc/kvm/book3s_hv_nested.c 	struct llist_node *entry = ((struct llist_head *) rmapp)->first;
entry             838 arch/powerpc/kvm/book3s_hv_nested.c 	for_each_nest_rmap_safe(cursor, entry, &rmap)
entry             866 arch/powerpc/kvm/book3s_hv_nested.c 	struct llist_node *entry = llist_del_all((struct llist_head *) rmapp);
entry             870 arch/powerpc/kvm/book3s_hv_nested.c 	for_each_nest_rmap_safe(cursor, entry, &rmap) {
entry             906 arch/powerpc/kvm/book3s_hv_nested.c 		struct llist_node *entry;
entry             908 arch/powerpc/kvm/book3s_hv_nested.c 		entry = llist_del_all((struct llist_head *) rmapp);
entry             909 arch/powerpc/kvm/book3s_hv_nested.c 		for_each_nest_rmap_safe(cursor, entry, &rmap)
entry            1083 arch/powerpc/kvm/book3s_hv_rm_mmu.c 	struct mmio_hpte_cache_entry *entry = NULL;
entry            1088 arch/powerpc/kvm/book3s_hv_rm_mmu.c 		entry = &vcpu->arch.mmio_cache.entry[i];
entry            1089 arch/powerpc/kvm/book3s_hv_rm_mmu.c 		if (entry->mmio_update == mmio_update) {
entry            1090 arch/powerpc/kvm/book3s_hv_rm_mmu.c 			pshift = entry->slb_base_pshift;
entry            1091 arch/powerpc/kvm/book3s_hv_rm_mmu.c 			if ((entry->eaddr >> pshift) == (eaddr >> pshift) &&
entry            1092 arch/powerpc/kvm/book3s_hv_rm_mmu.c 			    entry->slb_v == slb_v)
entry            1093 arch/powerpc/kvm/book3s_hv_rm_mmu.c 				return entry;
entry            1108 arch/powerpc/kvm/book3s_hv_rm_mmu.c 	return &vcpu->arch.mmio_cache.entry[index];
entry             383 arch/powerpc/kvm/book3s_xive_template.c 		u32 idx, toggle, entry, irq, hw_num;
entry             396 arch/powerpc/kvm/book3s_xive_template.c 			entry = be32_to_cpup(qpage + idx);
entry             399 arch/powerpc/kvm/book3s_xive_template.c 			if ((entry >> 31) == toggle)
entry             401 arch/powerpc/kvm/book3s_xive_template.c 			irq = entry & 0x7fffffff;
entry             419 arch/powerpc/kvm/book3s_xive_template.c 			qpage[idx] = cpu_to_be32((entry & 0x80000000) | XICS_DUMMY);
entry              53 arch/powerpc/kvm/e500.c 	struct id *entry[NUM_TIDS];
entry              70 arch/powerpc/kvm/e500.c static inline int local_sid_setup_one(struct id *entry)
entry              77 arch/powerpc/kvm/e500.c 		__this_cpu_write(pcpu_sids.entry[sid], entry);
entry              78 arch/powerpc/kvm/e500.c 		entry->val = sid;
entry              79 arch/powerpc/kvm/e500.c 		entry->pentry = this_cpu_ptr(&pcpu_sids.entry[sid]);
entry             104 arch/powerpc/kvm/e500.c static inline int local_sid_lookup(struct id *entry)
entry             106 arch/powerpc/kvm/e500.c 	if (entry && entry->val != 0 &&
entry             107 arch/powerpc/kvm/e500.c 	    __this_cpu_read(pcpu_sids.entry[entry->val]) == entry &&
entry             108 arch/powerpc/kvm/e500.c 	    entry->pentry == this_cpu_ptr(&pcpu_sids.entry[entry->val]))
entry             109 arch/powerpc/kvm/e500.c 		return entry->val;
entry             276 arch/powerpc/kvm/e500.h 	struct kvmppc_vcpu_e500 *vcpu_e500, int tlbsel, int entry)
entry             279 arch/powerpc/kvm/e500.h 	return &vcpu_e500->gtlb_arch[offset + entry];
entry             481 arch/powerpc/lib/feature-fixups.c static long calc_offset(struct fixup_entry *entry, unsigned int *p)
entry             483 arch/powerpc/lib/feature-fixups.c 	return (unsigned long)p - (unsigned long)entry;
entry              63 arch/powerpc/mm/book3s64/iommu_api.c 	unsigned long entry, chunk;
entry             103 arch/powerpc/mm/book3s64/iommu_api.c 	for (entry = 0; entry < entries; entry += chunk) {
entry             104 arch/powerpc/mm/book3s64/iommu_api.c 		unsigned long n = min(entries - entry, chunk);
entry             106 arch/powerpc/mm/book3s64/iommu_api.c 		ret = get_user_pages(ua + (entry << PAGE_SHIFT), n,
entry             108 arch/powerpc/mm/book3s64/iommu_api.c 				mem->hpages + entry, NULL);
entry             344 arch/powerpc/mm/book3s64/iommu_api.c 	const long entry = (ua - mem->ua) >> PAGE_SHIFT;
entry             347 arch/powerpc/mm/book3s64/iommu_api.c 	if (entry >= mem->entries)
entry             358 arch/powerpc/mm/book3s64/iommu_api.c 	va = &mem->hpas[entry];
entry             368 arch/powerpc/mm/book3s64/iommu_api.c 	const long entry = (ua - mem->ua) >> PAGE_SHIFT;
entry             371 arch/powerpc/mm/book3s64/iommu_api.c 	if (entry >= mem->entries)
entry             382 arch/powerpc/mm/book3s64/iommu_api.c 	pa = (void *) vmalloc_to_phys(&mem->hpas[entry]);
entry             394 arch/powerpc/mm/book3s64/iommu_api.c 	long entry;
entry             405 arch/powerpc/mm/book3s64/iommu_api.c 	entry = (ua - mem->ua) >> PAGE_SHIFT;
entry             406 arch/powerpc/mm/book3s64/iommu_api.c 	va = &mem->hpas[entry];
entry              36 arch/powerpc/mm/book3s64/pgtable.c 			  pmd_t *pmdp, pmd_t entry, int dirty)
entry              43 arch/powerpc/mm/book3s64/pgtable.c 	changed = !pmd_same(*(pmdp), entry);
entry              50 arch/powerpc/mm/book3s64/pgtable.c 					pmd_pte(entry), address, MMU_PAGE_2M);
entry            1033 arch/powerpc/mm/book3s64/radix_pgtable.c 				  pte_t entry, unsigned long address, int psize)
entry            1036 arch/powerpc/mm/book3s64/radix_pgtable.c 	unsigned long set = pte_val(entry) & (_PAGE_DIRTY | _PAGE_ACCESSED |
entry            1039 arch/powerpc/mm/book3s64/radix_pgtable.c 	unsigned long change = pte_val(entry) ^ pte_val(*ptep);
entry             650 arch/powerpc/mm/fault.c 	const struct exception_table_entry *entry;
entry             653 arch/powerpc/mm/fault.c 	if ((entry = search_exception_tables(regs->nip)) != NULL) {
entry             654 arch/powerpc/mm/fault.c 		regs->nip = extable_fixup(entry);
entry              56 arch/powerpc/mm/nohash/44x.c 	unsigned int entry = tlb_44x_hwater--;
entry              70 arch/powerpc/mm/nohash/44x.c 	  "r" (entry),
entry             217 arch/powerpc/mm/numa.c 		const __be32 *entry;
entry             219 arch/powerpc/mm/numa.c 		entry = &associativity[be32_to_cpu(distance_ref_points[i]) - 1];
entry             220 arch/powerpc/mm/numa.c 		distance_lookup_table[nid][i] = of_read_number(entry, 1);
entry             209 arch/powerpc/mm/pgtable.c 			  pte_t *ptep, pte_t entry, int dirty)
entry             212 arch/powerpc/mm/pgtable.c 	entry = set_access_flags_filter(entry, vma, dirty);
entry             213 arch/powerpc/mm/pgtable.c 	changed = !pte_same(*(ptep), entry);
entry             216 arch/powerpc/mm/pgtable.c 		__ptep_set_access_flags(vma, ptep, entry,
entry             280 arch/powerpc/mm/ptdump/hashpagetable.c 	struct mmu_psize_def entry;
entry             297 arch/powerpc/mm/ptdump/hashpagetable.c 	entry = mmu_psize_defs[bps];
entry             299 arch/powerpc/mm/ptdump/hashpagetable.c 		penc = entry.penc[idx];
entry              73 arch/powerpc/oprofile/cell/spu_profiler.c static void spu_pc_extract(int cpu, int entry)
entry              96 arch/powerpc/oprofile/cell/spu_profiler.c 		samples[spu * TRACE_ARRAY_SIZE + entry]
entry              98 arch/powerpc/oprofile/cell/spu_profiler.c 		samples[(spu + SPUS_PER_TB_ENTRY) * TRACE_ARRAY_SIZE + entry]
entry             109 arch/powerpc/oprofile/cell/spu_profiler.c 	int entry;
entry             113 arch/powerpc/oprofile/cell/spu_profiler.c 	entry = 0;
entry             118 arch/powerpc/oprofile/cell/spu_profiler.c 		spu_pc_extract(cpu, entry);
entry             120 arch/powerpc/oprofile/cell/spu_profiler.c 		entry++;
entry             122 arch/powerpc/oprofile/cell/spu_profiler.c 		if (entry >= TRACE_ARRAY_SIZE)
entry             129 arch/powerpc/oprofile/cell/spu_profiler.c 	return entry;
entry              47 arch/powerpc/perf/callchain.c perf_callchain_kernel(struct perf_callchain_entry_ctx *entry, struct pt_regs *regs)
entry              57 arch/powerpc/perf/callchain.c 	perf_callchain_store(entry, perf_instruction_pointer(regs));
entry              76 arch/powerpc/perf/callchain.c 			perf_callchain_store_context(entry, PERF_CONTEXT_KERNEL);
entry              98 arch/powerpc/perf/callchain.c 		perf_callchain_store(entry, next_ip);
entry             232 arch/powerpc/perf/callchain.c static void perf_callchain_user_64(struct perf_callchain_entry_ctx *entry,
entry             245 arch/powerpc/perf/callchain.c 	perf_callchain_store(entry, next_ip);
entry             247 arch/powerpc/perf/callchain.c 	while (entry->nr < entry->max_stack) {
entry             274 arch/powerpc/perf/callchain.c 			perf_callchain_store_context(entry, PERF_CONTEXT_USER);
entry             275 arch/powerpc/perf/callchain.c 			perf_callchain_store(entry, next_ip);
entry             281 arch/powerpc/perf/callchain.c 		perf_callchain_store(entry, next_ip);
entry             319 arch/powerpc/perf/callchain.c static inline void perf_callchain_user_64(struct perf_callchain_entry_ctx *entry,
entry             439 arch/powerpc/perf/callchain.c static void perf_callchain_user_32(struct perf_callchain_entry_ctx *entry,
entry             451 arch/powerpc/perf/callchain.c 	perf_callchain_store(entry, next_ip);
entry             453 arch/powerpc/perf/callchain.c 	while (entry->nr < entry->max_stack) {
entry             473 arch/powerpc/perf/callchain.c 			perf_callchain_store_context(entry, PERF_CONTEXT_USER);
entry             474 arch/powerpc/perf/callchain.c 			perf_callchain_store(entry, next_ip);
entry             480 arch/powerpc/perf/callchain.c 		perf_callchain_store(entry, next_ip);
entry             487 arch/powerpc/perf/callchain.c perf_callchain_user(struct perf_callchain_entry_ctx *entry, struct pt_regs *regs)
entry             490 arch/powerpc/perf/callchain.c 		perf_callchain_user_64(entry, regs);
entry             492 arch/powerpc/perf/callchain.c 		perf_callchain_user_32(entry, regs);
entry              40 arch/powerpc/platforms/4xx/hsta_msi.c 	struct msi_desc *entry;
entry              50 arch/powerpc/platforms/4xx/hsta_msi.c 	for_each_pci_msi_entry(entry, dev) {
entry              78 arch/powerpc/platforms/4xx/hsta_msi.c 		if (irq_set_msi_desc(hwirq, entry)) {
entry             105 arch/powerpc/platforms/4xx/hsta_msi.c 	struct msi_desc *entry;
entry             108 arch/powerpc/platforms/4xx/hsta_msi.c 	for_each_pci_msi_entry(entry, dev) {
entry             109 arch/powerpc/platforms/4xx/hsta_msi.c 		if (!entry->irq)
entry             112 arch/powerpc/platforms/4xx/hsta_msi.c 		irq = hsta_find_hwirq_offset(entry->irq);
entry             116 arch/powerpc/platforms/4xx/hsta_msi.c 		irq_set_msi_desc(entry->irq, NULL);
entry             119 arch/powerpc/platforms/4xx/hsta_msi.c 			 entry->irq, irq);
entry              70 arch/powerpc/platforms/4xx/msi.c 	struct msi_desc *entry;
entry              82 arch/powerpc/platforms/4xx/msi.c 	for_each_pci_msi_entry(entry, dev) {
entry             102 arch/powerpc/platforms/4xx/msi.c 		irq_set_msi_desc(virq, entry);
entry             111 arch/powerpc/platforms/4xx/msi.c 	struct msi_desc *entry;
entry             117 arch/powerpc/platforms/4xx/msi.c 	for_each_pci_msi_entry(entry, dev) {
entry             118 arch/powerpc/platforms/4xx/msi.c 		if (!entry->irq)
entry             120 arch/powerpc/platforms/4xx/msi.c 		hwirq = virq_to_hw(entry->irq);
entry             121 arch/powerpc/platforms/4xx/msi.c 		irq_set_msi_desc(entry->irq, NULL);
entry             122 arch/powerpc/platforms/4xx/msi.c 		irq_dispose_mapping(entry->irq);
entry             595 arch/powerpc/platforms/512x/clock-commonclk.c static void mpc512x_clk_setup_mclk(struct mclk_setup_data *entry, size_t idx)
entry             602 arch/powerpc/platforms/512x/clock-commonclk.c 	switch (entry->type) {
entry             675 arch/powerpc/platforms/512x/clock-commonclk.c 			entry->name_mux0,
entry             682 arch/powerpc/platforms/512x/clock-commonclk.c 			entry->name_en0, entry->name_mux0,
entry             685 arch/powerpc/platforms/512x/clock-commonclk.c 			entry->name_div0,
entry             686 arch/powerpc/platforms/512x/clock-commonclk.c 			entry->name_en0, CLK_SET_RATE_GATE,
entry             688 arch/powerpc/platforms/512x/clock-commonclk.c 	if (entry->has_mclk1) {
entry             690 arch/powerpc/platforms/512x/clock-commonclk.c 				entry->name_mclk,
entry             691 arch/powerpc/platforms/512x/clock-commonclk.c 				&entry->parent_names_mux1[0],
entry             692 arch/powerpc/platforms/512x/clock-commonclk.c 				ARRAY_SIZE(entry->parent_names_mux1),
entry             696 arch/powerpc/platforms/512x/clock-commonclk.c 				entry->name_mclk,
entry             697 arch/powerpc/platforms/512x/clock-commonclk.c 				entry->parent_names_mux1[0],
entry             202 arch/powerpc/platforms/cell/axon_msi.c 	struct msi_desc *entry;
entry             212 arch/powerpc/platforms/cell/axon_msi.c 	entry = first_pci_msi_entry(dev);
entry             215 arch/powerpc/platforms/cell/axon_msi.c 		if (entry->msi_attrib.is_64) {
entry             256 arch/powerpc/platforms/cell/axon_msi.c 	struct msi_desc *entry;
entry             268 arch/powerpc/platforms/cell/axon_msi.c 	for_each_pci_msi_entry(entry, dev) {
entry             277 arch/powerpc/platforms/cell/axon_msi.c 		irq_set_msi_desc(virq, entry);
entry             287 arch/powerpc/platforms/cell/axon_msi.c 	struct msi_desc *entry;
entry             291 arch/powerpc/platforms/cell/axon_msi.c 	for_each_pci_msi_entry(entry, dev) {
entry             292 arch/powerpc/platforms/cell/axon_msi.c 		if (!entry->irq)
entry             295 arch/powerpc/platforms/cell/axon_msi.c 		irq_set_msi_desc(entry->irq, NULL);
entry             296 arch/powerpc/platforms/cell/axon_msi.c 		irq_dispose_mapping(entry->irq);
entry              38 arch/powerpc/platforms/cell/spu_callbacks.c #define __SYSCALL(nr, entry)	entry,
entry            1086 arch/powerpc/platforms/cell/spufs/sched.c 	struct proc_dir_entry *entry;
entry            1110 arch/powerpc/platforms/cell/spufs/sched.c 	entry = proc_create_single("spu_loadavg", 0, NULL, show_spu_loadavg);
entry            1111 arch/powerpc/platforms/cell/spufs/sched.c 	if (!entry)
entry             246 arch/powerpc/platforms/embedded6xx/holly.c 	const struct exception_table_entry *entry;
entry             249 arch/powerpc/platforms/embedded6xx/holly.c 	if ((entry = search_exception_tables(regs->nip)) != NULL) {
entry             252 arch/powerpc/platforms/embedded6xx/holly.c 		regs->nip = extable_fixup(entry);
entry             167 arch/powerpc/platforms/embedded6xx/mpc7448_hpc2.c 	const struct exception_table_entry *entry;
entry             170 arch/powerpc/platforms/embedded6xx/mpc7448_hpc2.c 	if ((entry = search_exception_tables(regs->nip)) != NULL) {
entry             173 arch/powerpc/platforms/embedded6xx/mpc7448_hpc2.c 		regs->nip = extable_fixup(entry);
entry              22 arch/powerpc/platforms/pasemi/idle.c 	void (*entry)(void);
entry              26 arch/powerpc/platforms/pasemi/idle.c 	{ .name = "spin", .entry = &idle_spin },
entry              27 arch/powerpc/platforms/pasemi/idle.c 	{ .name = "doze", .entry = &idle_doze },
entry              73 arch/powerpc/platforms/pasemi/idle.c 	ppc_md.power_save = modes[current_mode].entry;
entry              60 arch/powerpc/platforms/pasemi/msi.c 	struct msi_desc *entry;
entry              65 arch/powerpc/platforms/pasemi/msi.c 	for_each_pci_msi_entry(entry, pdev) {
entry              66 arch/powerpc/platforms/pasemi/msi.c 		if (!entry->irq)
entry              69 arch/powerpc/platforms/pasemi/msi.c 		hwirq = virq_to_hw(entry->irq);
entry              70 arch/powerpc/platforms/pasemi/msi.c 		irq_set_msi_desc(entry->irq, NULL);
entry              71 arch/powerpc/platforms/pasemi/msi.c 		irq_dispose_mapping(entry->irq);
entry              81 arch/powerpc/platforms/pasemi/msi.c 	struct msi_desc *entry;
entry              93 arch/powerpc/platforms/pasemi/msi.c 	for_each_pci_msi_entry(entry, pdev) {
entry             121 arch/powerpc/platforms/pasemi/msi.c 		irq_set_msi_desc(virq, entry);
entry            1054 arch/powerpc/platforms/powernv/eeh-powernv.c 	edev = list_first_entry_or_null(&pe->edevs, struct eeh_dev, entry);
entry             357 arch/powerpc/platforms/powernv/opal-lpc.c 	struct lpc_debugfs_entry *entry;
entry             358 arch/powerpc/platforms/powernv/opal-lpc.c 	entry = kzalloc(sizeof(*entry), GFP_KERNEL);
entry             359 arch/powerpc/platforms/powernv/opal-lpc.c 	if (!entry)
entry             361 arch/powerpc/platforms/powernv/opal-lpc.c 	entry->lpc_type = type;
entry             362 arch/powerpc/platforms/powernv/opal-lpc.c 	debugfs_create_file(fname, 0600, folder, entry, &lpc_fops);
entry              43 arch/powerpc/platforms/powernv/opal.c 	u64 entry;
entry             115 arch/powerpc/platforms/powernv/opal.c 	opal.entry = of_read_number(entryp, entrysz/4);
entry             121 arch/powerpc/platforms/powernv/opal.c 		 opal.entry, entryp, entrysz);
entry             350 arch/powerpc/platforms/powernv/opal.c 	if (!opal.entry)
entry             369 arch/powerpc/platforms/powernv/opal.c 	if (!opal.entry)
entry            1062 arch/powerpc/platforms/powernv/opal.c 		sg->entry[i].data = cpu_to_be64(data);
entry            1063 arch/powerpc/platforms/powernv/opal.c 		sg->entry[i].length = cpu_to_be64(length);
entry             163 arch/powerpc/platforms/powernv/pci.c 	struct msi_desc *entry;
entry             175 arch/powerpc/platforms/powernv/pci.c 	for_each_pci_msi_entry(entry, pdev) {
entry             176 arch/powerpc/platforms/powernv/pci.c 		if (!entry->msi_attrib.is_64 && !phb->msi32_support) {
entry             195 arch/powerpc/platforms/powernv/pci.c 				    virq, entry->msi_attrib.is_64, &msg);
entry             202 arch/powerpc/platforms/powernv/pci.c 		irq_set_msi_desc(virq, entry);
entry             212 arch/powerpc/platforms/powernv/pci.c 	struct msi_desc *entry;
entry             218 arch/powerpc/platforms/powernv/pci.c 	for_each_pci_msi_entry(entry, pdev) {
entry             219 arch/powerpc/platforms/powernv/pci.c 		if (!entry->irq)
entry             221 arch/powerpc/platforms/powernv/pci.c 		hwirq = virq_to_hw(entry->irq);
entry             222 arch/powerpc/platforms/powernv/pci.c 		irq_set_msi_desc(entry->irq, NULL);
entry             223 arch/powerpc/platforms/powernv/pci.c 		irq_dispose_mapping(entry->irq);
entry             114 arch/powerpc/platforms/pseries/msi.c 	struct msi_desc *entry;
entry             116 arch/powerpc/platforms/pseries/msi.c 	for_each_pci_msi_entry(entry, pdev) {
entry             117 arch/powerpc/platforms/pseries/msi.c 		if (!entry->irq)
entry             120 arch/powerpc/platforms/pseries/msi.c 		irq_set_msi_desc(entry->irq, NULL);
entry             121 arch/powerpc/platforms/pseries/msi.c 		irq_dispose_mapping(entry->irq);
entry             202 arch/powerpc/platforms/pseries/msi.c 					entry);
entry             334 arch/powerpc/platforms/pseries/msi.c 	struct msi_desc *entry;
entry             342 arch/powerpc/platforms/pseries/msi.c 	for_each_pci_msi_entry(entry, pdev) {
entry             343 arch/powerpc/platforms/pseries/msi.c 		if (entry->msi_attrib.entry_nr != expected) {
entry             374 arch/powerpc/platforms/pseries/msi.c 	struct msi_desc *entry;
entry             454 arch/powerpc/platforms/pseries/msi.c 	for_each_pci_msi_entry(entry, pdev) {
entry             469 arch/powerpc/platforms/pseries/msi.c 		irq_set_msi_desc(virq, entry);
entry             472 arch/powerpc/platforms/pseries/msi.c 		__pci_read_msi_msg(entry, &msg);
entry             473 arch/powerpc/platforms/pseries/msi.c 		entry->msg = msg;
entry             124 arch/powerpc/sysdev/fsl_msi.c 	struct msi_desc *entry;
entry             128 arch/powerpc/sysdev/fsl_msi.c 	for_each_pci_msi_entry(entry, pdev) {
entry             129 arch/powerpc/sysdev/fsl_msi.c 		if (!entry->irq)
entry             131 arch/powerpc/sysdev/fsl_msi.c 		hwirq = virq_to_hw(entry->irq);
entry             132 arch/powerpc/sysdev/fsl_msi.c 		msi_data = irq_get_chip_data(entry->irq);
entry             133 arch/powerpc/sysdev/fsl_msi.c 		irq_set_msi_desc(entry->irq, NULL);
entry             134 arch/powerpc/sysdev/fsl_msi.c 		irq_dispose_mapping(entry->irq);
entry             185 arch/powerpc/sysdev/fsl_msi.c 	struct msi_desc *entry;
entry             218 arch/powerpc/sysdev/fsl_msi.c 	for_each_pci_msi_entry(entry, pdev) {
entry             255 arch/powerpc/sysdev/fsl_msi.c 		irq_set_msi_desc(virq, entry);
entry              96 arch/powerpc/sysdev/fsl_rio.c 	const struct exception_table_entry *entry;
entry             105 arch/powerpc/sysdev/fsl_rio.c 		entry = search_exception_tables(regs->nip);
entry             106 arch/powerpc/sysdev/fsl_rio.c 		if (entry) {
entry             112 arch/powerpc/sysdev/fsl_rio.c 			regs->nip = extable_fixup(entry);
entry             104 arch/powerpc/sysdev/mpic_u3msi.c 	struct msi_desc *entry;
entry             107 arch/powerpc/sysdev/mpic_u3msi.c 	for_each_pci_msi_entry(entry, pdev) {
entry             108 arch/powerpc/sysdev/mpic_u3msi.c 		if (!entry->irq)
entry             111 arch/powerpc/sysdev/mpic_u3msi.c 		hwirq = virq_to_hw(entry->irq);
entry             112 arch/powerpc/sysdev/mpic_u3msi.c 		irq_set_msi_desc(entry->irq, NULL);
entry             113 arch/powerpc/sysdev/mpic_u3msi.c 		irq_dispose_mapping(entry->irq);
entry             123 arch/powerpc/sysdev/mpic_u3msi.c 	struct msi_desc *entry;
entry             139 arch/powerpc/sysdev/mpic_u3msi.c 	for_each_pci_msi_entry(entry, pdev) {
entry             159 arch/powerpc/sysdev/mpic_u3msi.c 		irq_set_msi_desc(virq, entry);
entry            3105 arch/powerpc/xmon/xmon.c 	pte_t entry = __pte(pte);
entry            3111 arch/powerpc/xmon/xmon.c 	       pte_young(entry) ? "Accessed " : "",
entry            3112 arch/powerpc/xmon/xmon.c 	       pte_dirty(entry) ? "Dirty " : "",
entry            3113 arch/powerpc/xmon/xmon.c 	       pte_read(entry)  ? "Read " : "",
entry            3114 arch/powerpc/xmon/xmon.c 	       pte_write(entry) ? "Write " : "",
entry            3115 arch/powerpc/xmon/xmon.c 	       pte_exec(entry)  ? "Exec " : "");
entry             351 arch/riscv/include/asm/pgtable.h 					pte_t entry, int dirty)
entry             353 arch/riscv/include/asm/pgtable.h 	if (!pte_same(*ptep, entry))
entry             354 arch/riscv/include/asm/pgtable.h 		set_pte_at(vma->vm_mm, address, ptep, entry);
entry              17 arch/riscv/kernel/perf_callchain.c static unsigned long user_backtrace(struct perf_callchain_entry_ctx *entry,
entry              39 arch/riscv/kernel/perf_callchain.c 		perf_callchain_store(entry, ra);
entry              60 arch/riscv/kernel/perf_callchain.c void perf_callchain_user(struct perf_callchain_entry_ctx *entry,
entry              70 arch/riscv/kernel/perf_callchain.c 	perf_callchain_store(entry, regs->sepc);
entry              72 arch/riscv/kernel/perf_callchain.c 	fp = user_backtrace(entry, fp, regs->ra);
entry              73 arch/riscv/kernel/perf_callchain.c 	while (fp && !(fp & 0x3) && entry->nr < entry->max_stack)
entry              74 arch/riscv/kernel/perf_callchain.c 		fp = user_backtrace(entry, fp, 0);
entry              77 arch/riscv/kernel/perf_callchain.c bool fill_callchain(unsigned long pc, void *entry)
entry              79 arch/riscv/kernel/perf_callchain.c 	return perf_callchain_store(entry, pc);
entry              84 arch/riscv/kernel/perf_callchain.c void perf_callchain_kernel(struct perf_callchain_entry_ctx *entry,
entry              93 arch/riscv/kernel/perf_callchain.c 	walk_stackframe(NULL, regs, fill_callchain, entry);
entry              15 arch/s390/boot/compressed/decompressor.h 	void (*entry)(void);
entry              19 arch/s390/boot/ipl_report.c #define for_each_rb_entry(entry, rb) \
entry              20 arch/s390/boot/ipl_report.c 	for (entry = rb->entries; \
entry              21 arch/s390/boot/ipl_report.c 	     (void *) entry + sizeof(*entry) <= (void *) rb + rb->len; \
entry              22 arch/s390/boot/ipl_report.c 	     entry++)
entry              40 arch/s390/boot/startup.c void _diag0c_dma(struct hypfs_diag0c_entry *entry);
entry             152 arch/s390/boot/startup.c 			vmlinux.entry += __kaslr_offset;
entry             182 arch/s390/boot/startup.c 	vmlinux.entry();
entry              43 arch/s390/hypfs/hypfs_diag0c.c 	diag0c_data = kzalloc(struct_size(diag0c_data, entry, cpu_count),
entry              50 arch/s390/hypfs/hypfs_diag0c.c 		diag0c_data->entry[i].cpu = cpu;
entry              51 arch/s390/hypfs/hypfs_diag0c.c 		cpu_vec[cpu] = &diag0c_data->entry[i++];
entry              26 arch/s390/include/asm/debug.h #define DEBUG_DATA(entry) (char *)(entry + 1) /* data is stored behind */
entry              57 arch/s390/include/asm/debug.h 				   debug_entry_t *entry,
entry              73 arch/s390/include/asm/debug.h 			 int area, debug_entry_t *entry, char *out_buf);
entry             318 arch/s390/include/asm/diag.h 	void (*diag0c)(struct hypfs_diag0c_entry *entry);
entry              95 arch/s390/include/asm/ipl.h 	struct ipl_rb_component_entry entry;
entry             100 arch/s390/include/asm/ipl.h 	struct ipl_rb_certificate_entry entry;
entry              31 arch/s390/include/asm/os_info.h 	struct os_info_entry entry[2];
entry             101 arch/s390/include/asm/pci.h 	struct list_head entry;		/* list of all zpci_devices, needed for hotplug, etc. */
entry             100 arch/s390/include/asm/pci_dma.h static inline void set_pt_pfaa(unsigned long *entry, void *pfaa)
entry             102 arch/s390/include/asm/pci_dma.h 	*entry &= ZPCI_PTE_FLAG_MASK;
entry             103 arch/s390/include/asm/pci_dma.h 	*entry |= ((unsigned long) pfaa & ZPCI_PTE_ADDR_MASK);
entry             106 arch/s390/include/asm/pci_dma.h static inline void set_rt_sto(unsigned long *entry, void *sto)
entry             108 arch/s390/include/asm/pci_dma.h 	*entry &= ZPCI_RTE_FLAG_MASK;
entry             109 arch/s390/include/asm/pci_dma.h 	*entry |= ((unsigned long) sto & ZPCI_RTE_ADDR_MASK);
entry             110 arch/s390/include/asm/pci_dma.h 	*entry |= ZPCI_TABLE_TYPE_RTX;
entry             113 arch/s390/include/asm/pci_dma.h static inline void set_st_pto(unsigned long *entry, void *pto)
entry             115 arch/s390/include/asm/pci_dma.h 	*entry &= ZPCI_STE_FLAG_MASK;
entry             116 arch/s390/include/asm/pci_dma.h 	*entry |= ((unsigned long) pto & ZPCI_STE_ADDR_MASK);
entry             117 arch/s390/include/asm/pci_dma.h 	*entry |= ZPCI_TABLE_TYPE_SX;
entry             120 arch/s390/include/asm/pci_dma.h static inline void validate_rt_entry(unsigned long *entry)
entry             122 arch/s390/include/asm/pci_dma.h 	*entry &= ~ZPCI_TABLE_VALID_MASK;
entry             123 arch/s390/include/asm/pci_dma.h 	*entry &= ~ZPCI_TABLE_OFFSET_MASK;
entry             124 arch/s390/include/asm/pci_dma.h 	*entry |= ZPCI_TABLE_VALID;
entry             125 arch/s390/include/asm/pci_dma.h 	*entry |= ZPCI_TABLE_LEN_RTX;
entry             128 arch/s390/include/asm/pci_dma.h static inline void validate_st_entry(unsigned long *entry)
entry             130 arch/s390/include/asm/pci_dma.h 	*entry &= ~ZPCI_TABLE_VALID_MASK;
entry             131 arch/s390/include/asm/pci_dma.h 	*entry |= ZPCI_TABLE_VALID;
entry             134 arch/s390/include/asm/pci_dma.h static inline void invalidate_table_entry(unsigned long *entry)
entry             136 arch/s390/include/asm/pci_dma.h 	*entry &= ~ZPCI_TABLE_VALID_MASK;
entry             137 arch/s390/include/asm/pci_dma.h 	*entry |= ZPCI_TABLE_INVALID;
entry             140 arch/s390/include/asm/pci_dma.h static inline void invalidate_pt_entry(unsigned long *entry)
entry             142 arch/s390/include/asm/pci_dma.h 	WARN_ON_ONCE((*entry & ZPCI_PTE_VALID_MASK) == ZPCI_PTE_INVALID);
entry             143 arch/s390/include/asm/pci_dma.h 	*entry &= ~ZPCI_PTE_VALID_MASK;
entry             144 arch/s390/include/asm/pci_dma.h 	*entry |= ZPCI_PTE_INVALID;
entry             147 arch/s390/include/asm/pci_dma.h static inline void validate_pt_entry(unsigned long *entry)
entry             149 arch/s390/include/asm/pci_dma.h 	WARN_ON_ONCE((*entry & ZPCI_PTE_VALID_MASK) == ZPCI_PTE_VALID);
entry             150 arch/s390/include/asm/pci_dma.h 	*entry &= ~ZPCI_PTE_VALID_MASK;
entry             151 arch/s390/include/asm/pci_dma.h 	*entry |= ZPCI_PTE_VALID;
entry             154 arch/s390/include/asm/pci_dma.h static inline void entry_set_protected(unsigned long *entry)
entry             156 arch/s390/include/asm/pci_dma.h 	*entry &= ~ZPCI_TABLE_PROT_MASK;
entry             157 arch/s390/include/asm/pci_dma.h 	*entry |= ZPCI_TABLE_PROTECTED;
entry             160 arch/s390/include/asm/pci_dma.h static inline void entry_clr_protected(unsigned long *entry)
entry             162 arch/s390/include/asm/pci_dma.h 	*entry &= ~ZPCI_TABLE_PROT_MASK;
entry             163 arch/s390/include/asm/pci_dma.h 	*entry |= ZPCI_TABLE_UNPROTECTED;
entry             166 arch/s390/include/asm/pci_dma.h static inline int reg_entry_isvalid(unsigned long entry)
entry             168 arch/s390/include/asm/pci_dma.h 	return (entry & ZPCI_TABLE_VALID_MASK) == ZPCI_TABLE_VALID;
entry             171 arch/s390/include/asm/pci_dma.h static inline int pt_entry_isvalid(unsigned long entry)
entry             173 arch/s390/include/asm/pci_dma.h 	return (entry & ZPCI_PTE_VALID_MASK) == ZPCI_PTE_VALID;
entry             176 arch/s390/include/asm/pci_dma.h static inline int entry_isprotected(unsigned long entry)
entry             178 arch/s390/include/asm/pci_dma.h 	return (entry & ZPCI_TABLE_PROT_MASK) == ZPCI_TABLE_PROTECTED;
entry             181 arch/s390/include/asm/pci_dma.h static inline unsigned long *get_rt_sto(unsigned long entry)
entry             183 arch/s390/include/asm/pci_dma.h 	return ((entry & ZPCI_TABLE_TYPE_MASK) == ZPCI_TABLE_TYPE_RTX)
entry             184 arch/s390/include/asm/pci_dma.h 		? (unsigned long *) (entry & ZPCI_RTE_ADDR_MASK)
entry             188 arch/s390/include/asm/pci_dma.h static inline unsigned long *get_st_pto(unsigned long entry)
entry             190 arch/s390/include/asm/pci_dma.h 	return ((entry & ZPCI_TABLE_TYPE_MASK) == ZPCI_TABLE_TYPE_SX)
entry             191 arch/s390/include/asm/pci_dma.h 		? (unsigned long *) (entry & ZPCI_STE_ADDR_MASK)
entry             202 arch/s390/include/asm/pci_dma.h void dma_update_cpu_trans(unsigned long *entry, void *page_addr, int flags);
entry              32 arch/s390/include/asm/pgalloc.h static inline void crst_table_init(unsigned long *crst, unsigned long entry)
entry              34 arch/s390/include/asm/pgalloc.h 	memset64((u64 *)crst, entry, _CRST_ENTRIES);
entry            1128 arch/s390/include/asm/pgtable.h 					pte_t entry, int dirty)
entry            1130 arch/s390/include/asm/pgtable.h 	if (pte_same(*ptep, entry))
entry            1132 arch/s390/include/asm/pgtable.h 	ptep_xchg_direct(vma->vm_mm, addr, ptep, entry);
entry            1140 arch/s390/include/asm/pgtable.h 		     pte_t *ptep, pte_t entry);
entry            1180 arch/s390/include/asm/pgtable.h 			      pte_t *ptep, pte_t entry)
entry            1182 arch/s390/include/asm/pgtable.h 	if (pte_present(entry))
entry            1183 arch/s390/include/asm/pgtable.h 		pte_val(entry) &= ~_PAGE_UNUSED;
entry            1185 arch/s390/include/asm/pgtable.h 		ptep_set_pte_at(mm, addr, ptep, entry);
entry            1187 arch/s390/include/asm/pgtable.h 		*ptep = entry;
entry            1514 arch/s390/include/asm/pgtable.h 					pmd_t entry, int dirty)
entry            1518 arch/s390/include/asm/pgtable.h 	entry = pmd_mkyoung(entry);
entry            1520 arch/s390/include/asm/pgtable.h 		entry = pmd_mkdirty(entry);
entry            1521 arch/s390/include/asm/pgtable.h 	if (pmd_val(*pmdp) == pmd_val(entry))
entry            1523 arch/s390/include/asm/pgtable.h 	pmdp_xchg_direct(vma->vm_mm, addr, pmdp, entry);
entry            1546 arch/s390/include/asm/pgtable.h 			      pmd_t *pmdp, pmd_t entry)
entry            1549 arch/s390/include/asm/pgtable.h 		pmd_val(entry) &= ~_SEGMENT_ENTRY_NOEXEC;
entry            1550 arch/s390/include/asm/pgtable.h 	*pmdp = entry;
entry            1662 arch/s390/include/asm/pgtable.h static inline unsigned long __swp_type(swp_entry_t entry)
entry            1664 arch/s390/include/asm/pgtable.h 	return (entry.val >> __SWP_TYPE_SHIFT) & __SWP_TYPE_MASK;
entry            1667 arch/s390/include/asm/pgtable.h static inline unsigned long __swp_offset(swp_entry_t entry)
entry            1669 arch/s390/include/asm/pgtable.h 	return (entry.val >> __SWP_OFFSET_SHIFT) & __SWP_OFFSET_MASK;
entry             429 arch/s390/include/asm/qdio.h 				void *entry),
entry              15 arch/s390/include/asm/vtimer.h 	struct list_head entry;
entry              52 arch/s390/include/uapi/asm/hypfs.h 	struct hypfs_diag0c_entry	entry[];	/* diag0c entry array */
entry              97 arch/s390/kernel/debug.c 			       int area, debug_entry_t *entry, char *out_buf);
entry            1391 arch/s390/kernel/debug.c 			       int area, debug_entry_t *entry, char *out_buf)
entry            1396 arch/s390/kernel/debug.c 	memcpy(out_buf, entry, sizeof(debug_entry_t));
entry            1440 arch/s390/kernel/debug.c 			 int area, debug_entry_t *entry, char *out_buf)
entry            1448 arch/s390/kernel/debug.c 	level = entry->id.fields.level;
entry            1450 arch/s390/kernel/debug.c 	sec = (entry->id.stck >> 12) + base - (TOD_UNIX_EPOCH >> 12);
entry            1453 arch/s390/kernel/debug.c 	if (entry->id.fields.exception)
entry            1457 arch/s390/kernel/debug.c 	caller = (unsigned long) entry->caller;
entry            1460 arch/s390/kernel/debug.c 		      entry->id.fields.cpuid, (void *)caller);
entry             394 arch/s390/kernel/dis.c 	struct s390_opcode_offset *entry;
entry             405 arch/s390/kernel/dis.c 		entry = &opcode_offset[i];
entry             406 arch/s390/kernel/dis.c 		if (entry->opcode == code[0])
entry             410 arch/s390/kernel/dis.c 	opfrag = *(code + entry->byte) & entry->mask;
entry             412 arch/s390/kernel/dis.c 	insn = &opcode[entry->offset];
entry             413 arch/s390/kernel/dis.c 	for (i = 0; i < entry->count; i++) {
entry            1724 arch/s390/kernel/ipl.c 	comp->entry.addr = kbuf->mem;
entry            1725 arch/s390/kernel/ipl.c 	comp->entry.len = kbuf->memsz;
entry            1726 arch/s390/kernel/ipl.c 	comp->entry.flags = flags;
entry            1727 arch/s390/kernel/ipl.c 	comp->entry.certificate_index = cert;
entry            1729 arch/s390/kernel/ipl.c 	report->size += sizeof(comp->entry);
entry            1744 arch/s390/kernel/ipl.c 	cert->entry.addr = addr;
entry            1745 arch/s390/kernel/ipl.c 	cert->entry.len = len;
entry            1748 arch/s390/kernel/ipl.c 	report->size += sizeof(cert->entry);
entry            1749 arch/s390/kernel/ipl.c 	report->size += cert->entry.len;
entry            1804 arch/s390/kernel/ipl.c 		memcpy(ptr, &comp->entry, sizeof(comp->entry));
entry            1805 arch/s390/kernel/ipl.c 		ptr += sizeof(comp->entry);
entry            1813 arch/s390/kernel/ipl.c 		memcpy(ptr, &cert->entry, sizeof(cert->entry));
entry            1814 arch/s390/kernel/ipl.c 		ptr += sizeof(cert->entry);
entry            1820 arch/s390/kernel/ipl.c 		memcpy(ptr, cert->key, cert->entry.len);
entry            1821 arch/s390/kernel/ipl.c 		ptr += cert->entry.len;
entry             222 arch/s390/kernel/irq.c 	struct hlist_node entry;
entry             251 arch/s390/kernel/irq.c 	hlist_add_head_rcu(&p->entry, &ext_int_hash[index]);
entry             264 arch/s390/kernel/irq.c 	hlist_for_each_entry_rcu(p, &ext_int_hash[index], entry) {
entry             266 arch/s390/kernel/irq.c 			hlist_del_rcu(&p->entry);
entry             288 arch/s390/kernel/irq.c 	hlist_for_each_entry_rcu(p, &ext_int_hash[index], entry) {
entry              18 arch/s390/kernel/jump_label.c static void jump_label_make_nop(struct jump_entry *entry, struct insn *insn)
entry              22 arch/s390/kernel/jump_label.c 	insn->offset = (jump_entry_target(entry) - jump_entry_code(entry)) >> 1;
entry              25 arch/s390/kernel/jump_label.c static void jump_label_make_branch(struct jump_entry *entry, struct insn *insn)
entry              29 arch/s390/kernel/jump_label.c 	insn->offset = (jump_entry_target(entry) - jump_entry_code(entry)) >> 1;
entry              32 arch/s390/kernel/jump_label.c static void jump_label_bug(struct jump_entry *entry, struct insn *expected,
entry              35 arch/s390/kernel/jump_label.c 	unsigned char *ipc = (unsigned char *)jump_entry_code(entry);
entry              51 arch/s390/kernel/jump_label.c static void __jump_label_transform(struct jump_entry *entry,
entry              55 arch/s390/kernel/jump_label.c 	void *code = (void *)jump_entry_code(entry);
entry              59 arch/s390/kernel/jump_label.c 		jump_label_make_nop(entry, &old);
entry              60 arch/s390/kernel/jump_label.c 		jump_label_make_branch(entry, &new);
entry              62 arch/s390/kernel/jump_label.c 		jump_label_make_branch(entry, &old);
entry              63 arch/s390/kernel/jump_label.c 		jump_label_make_nop(entry, &new);
entry              67 arch/s390/kernel/jump_label.c 			jump_label_bug(entry, &orignop, &new);
entry              70 arch/s390/kernel/jump_label.c 			jump_label_bug(entry, &old, &new);
entry              79 arch/s390/kernel/jump_label.c void arch_jump_label_transform(struct jump_entry *entry,
entry              82 arch/s390/kernel/jump_label.c 	__jump_label_transform(entry, type, 0);
entry              86 arch/s390/kernel/jump_label.c void arch_jump_label_transform_static(struct jump_entry *entry,
entry              89 arch/s390/kernel/jump_label.c 	__jump_label_transform(entry, type, 1);
entry              22 arch/s390/kernel/kexec_elf.c 	Elf_Addr entry;
entry              30 arch/s390/kernel/kexec_elf.c 		entry = STARTUP_KDUMP_OFFSET;
entry              32 arch/s390/kernel/kexec_elf.c 		entry = ehdr->e_entry;
entry              48 arch/s390/kernel/kexec_elf.c 		if (entry - phdr->p_paddr < phdr->p_memsz) {
entry             537 arch/s390/kernel/kprobes.c 	const struct exception_table_entry *entry;
entry             576 arch/s390/kernel/kprobes.c 		entry = s390_search_extables(regs->psw.addr);
entry             577 arch/s390/kernel/kprobes.c 		if (entry) {
entry             578 arch/s390/kernel/kprobes.c 			regs->psw.addr = extable_fixup(entry);
entry              76 arch/s390/kernel/machine_kexec_file.c 	u64 entry, type;
entry              80 arch/s390/kernel/machine_kexec_file.c 		entry = STARTUP_KDUMP_OFFSET;
entry              83 arch/s390/kernel/machine_kexec_file.c 		entry = STARTUP_NORMAL_OFFSET;
entry              87 arch/s390/kernel/machine_kexec_file.c 	ret = kexec_purgatory_get_set_symbol(image, "kernel_entry", &entry,
entry              88 arch/s390/kernel/machine_kexec_file.c 					     sizeof(entry), false);
entry              48 arch/s390/kernel/os_info.c 	os_info.entry[nr].addr = (u64)(unsigned long)ptr;
entry              49 arch/s390/kernel/os_info.c 	os_info.entry[nr].size = size;
entry              50 arch/s390/kernel/os_info.c 	os_info.entry[nr].csum = (__force u32)csum_partial(ptr, size, 0);
entry              81 arch/s390/kernel/os_info.c 	addr = os_info_old->entry[nr].addr;
entry              86 arch/s390/kernel/os_info.c 	size = os_info_old->entry[nr].size;
entry              98 arch/s390/kernel/os_info.c 	if (csum != os_info_old->entry[nr].csum) {
entry             102 arch/s390/kernel/os_info.c 	os_info_old->entry[nr].addr = (u64)(unsigned long)buf_align;
entry             108 arch/s390/kernel/os_info.c 	os_info_old->entry[nr].addr = 0;
entry             165 arch/s390/kernel/os_info.c 	if (!os_info_old->entry[nr].addr)
entry             167 arch/s390/kernel/os_info.c 	*size = (unsigned long) os_info_old->entry[nr].size;
entry             168 arch/s390/kernel/os_info.c 	return (void *)(unsigned long)os_info_old->entry[nr].addr;
entry             223 arch/s390/kernel/perf_event.c void perf_callchain_kernel(struct perf_callchain_entry_ctx *entry,
entry             231 arch/s390/kernel/perf_event.c 		if (!addr || perf_callchain_store(entry, addr))
entry             262 arch/s390/kernel/vtime.c 	list_for_each_entry(tmp, head, entry) {
entry             264 arch/s390/kernel/vtime.c 			list_add_tail(&timer->entry, &tmp->entry);
entry             268 arch/s390/kernel/vtime.c 	list_add_tail(&timer->entry, head);
entry             283 arch/s390/kernel/vtime.c 	list_for_each_entry_safe(timer, tmp, &virt_timer_list, entry) {
entry             286 arch/s390/kernel/vtime.c 			list_move_tail(&timer->entry, &cb_list);
entry             292 arch/s390/kernel/vtime.c 					 struct vtimer_list, entry);
entry             299 arch/s390/kernel/vtime.c 	list_for_each_entry_safe(timer, tmp, &cb_list, entry) {
entry             300 arch/s390/kernel/vtime.c 		list_del_init(&timer->entry);
entry             316 arch/s390/kernel/vtime.c 	INIT_LIST_HEAD(&timer->entry);
entry             322 arch/s390/kernel/vtime.c 	return !list_empty(&timer->entry);
entry             331 arch/s390/kernel/vtime.c 		list_add(&timer->entry, &virt_timer_list);
entry             384 arch/s390/kernel/vtime.c 		list_del_init(&timer->entry);
entry             422 arch/s390/kernel/vtime.c 	list_del_init(&timer->entry);
entry             337 arch/s390/mm/gmap.c static unsigned long __gmap_segment_gaddr(unsigned long *entry)
entry             342 arch/s390/mm/gmap.c 	offset = (unsigned long) entry / sizeof(unsigned long);
entry             345 arch/s390/mm/gmap.c 	page = virt_to_page((void *)((unsigned long) entry & mask));
entry             358 arch/s390/mm/gmap.c 	unsigned long *entry;
entry             363 arch/s390/mm/gmap.c 	entry = radix_tree_delete(&gmap->host_to_guest, vmaddr >> PMD_SHIFT);
entry             364 arch/s390/mm/gmap.c 	if (entry) {
entry             365 arch/s390/mm/gmap.c 		flush = (*entry != _SEGMENT_ENTRY_EMPTY);
entry             366 arch/s390/mm/gmap.c 		*entry = _SEGMENT_ENTRY_EMPTY;
entry            2358 arch/s390/mm/gmap.c 	unsigned long *entry, gaddr;
entry            2365 arch/s390/mm/gmap.c 		entry = radix_tree_delete(&gmap->host_to_guest,
entry            2367 arch/s390/mm/gmap.c 		if (entry) {
entry            2368 arch/s390/mm/gmap.c 			pmdp = (pmd_t *)entry;
entry            2369 arch/s390/mm/gmap.c 			gaddr = __gmap_segment_gaddr(entry);
entry            2371 arch/s390/mm/gmap.c 			WARN_ON(*entry & ~(_SEGMENT_ENTRY_HARDWARE_BITS_LARGE |
entry            2378 arch/s390/mm/gmap.c 			*entry = _SEGMENT_ENTRY_EMPTY;
entry            2393 arch/s390/mm/gmap.c 	unsigned long *entry, gaddr;
entry            2400 arch/s390/mm/gmap.c 		entry = radix_tree_delete(&gmap->host_to_guest,
entry            2402 arch/s390/mm/gmap.c 		if (entry) {
entry            2403 arch/s390/mm/gmap.c 			pmdp = (pmd_t *)entry;
entry            2404 arch/s390/mm/gmap.c 			gaddr = __gmap_segment_gaddr(entry);
entry            2406 arch/s390/mm/gmap.c 			WARN_ON(*entry & ~(_SEGMENT_ENTRY_HARDWARE_BITS_LARGE |
entry            2415 arch/s390/mm/gmap.c 			*entry = _SEGMENT_ENTRY_EMPTY;
entry             176 arch/s390/mm/pgtable.c static inline void pgste_set_key(pte_t *ptep, pgste_t pgste, pte_t entry,
entry             183 arch/s390/mm/pgtable.c 	if (!mm_uses_skeys(mm) || pte_val(entry) & _PAGE_INVALID)
entry             186 arch/s390/mm/pgtable.c 	address = pte_val(entry) & PAGE_MASK;
entry             198 arch/s390/mm/pgtable.c static inline pgste_t pgste_set_pte(pte_t *ptep, pgste_t pgste, pte_t entry)
entry             201 arch/s390/mm/pgtable.c 	if ((pte_val(entry) & _PAGE_PRESENT) &&
entry             202 arch/s390/mm/pgtable.c 	    (pte_val(entry) & _PAGE_WRITE) &&
entry             203 arch/s390/mm/pgtable.c 	    !(pte_val(entry) & _PAGE_INVALID)) {
entry             209 arch/s390/mm/pgtable.c 			pte_val(entry) |= _PAGE_DIRTY;
entry             210 arch/s390/mm/pgtable.c 			pte_val(entry) &= ~_PAGE_PROTECT;
entry             212 arch/s390/mm/pgtable.c 		if (!(pte_val(entry) & _PAGE_PROTECT))
entry             217 arch/s390/mm/pgtable.c 	*ptep = entry;
entry             559 arch/s390/mm/pgtable.c 		     pte_t *ptep, pte_t entry)
entry             567 arch/s390/mm/pgtable.c 	pgste_set_key(ptep, pgste, entry, mm);
entry             568 arch/s390/mm/pgtable.c 	pgste = pgste_set_pte(ptep, pgste, entry);
entry             598 arch/s390/mm/pgtable.c 	pte_t entry;
entry             603 arch/s390/mm/pgtable.c 	entry = *ptep;
entry             605 arch/s390/mm/pgtable.c 	pte_i = pte_val(entry) & _PAGE_INVALID;
entry             606 arch/s390/mm/pgtable.c 	pte_p = pte_val(entry) & _PAGE_PROTECT;
entry             616 arch/s390/mm/pgtable.c 		pgste = pgste_update_all(entry, pgste, mm);
entry             617 arch/s390/mm/pgtable.c 		pte_val(entry) |= _PAGE_INVALID;
entry             621 arch/s390/mm/pgtable.c 		pte_val(entry) &= ~_PAGE_INVALID;
entry             622 arch/s390/mm/pgtable.c 		pte_val(entry) |= _PAGE_PROTECT;
entry             625 arch/s390/mm/pgtable.c 	pgste = pgste_set_pte(ptep, pgste, entry);
entry             671 arch/s390/mm/pgtable.c static void ptep_zap_swap_entry(struct mm_struct *mm, swp_entry_t entry)
entry             673 arch/s390/mm/pgtable.c 	if (!non_swap_entry(entry))
entry             675 arch/s390/mm/pgtable.c 	else if (is_migration_entry(entry)) {
entry             676 arch/s390/mm/pgtable.c 		struct page *page = migration_entry_to_page(entry);
entry             680 arch/s390/mm/pgtable.c 	free_swap_and_cache(entry);
entry              63 arch/s390/pci/pci.c 	list_for_each_entry(tmp, &zpci_list, entry) {
entry              80 arch/s390/pci/pci.c 	list_for_each_entry_safe(zdev, tmp, &zpci_list, entry) {
entry              84 arch/s390/pci/pci.c 			list_move_tail(&zdev->entry, &remove);
entry              88 arch/s390/pci/pci.c 	list_for_each_entry_safe(zdev, tmp, &remove, entry)
entry             465 arch/s390/pci/pci.c 	unsigned long entry;
entry             468 arch/s390/pci/pci.c 	entry = find_first_zero_bit(zpci_iomap_bitmap, ZPCI_IOMAP_ENTRIES);
entry             469 arch/s390/pci/pci.c 	if (entry == ZPCI_IOMAP_ENTRIES) {
entry             473 arch/s390/pci/pci.c 	set_bit(entry, zpci_iomap_bitmap);
entry             475 arch/s390/pci/pci.c 	return entry;
entry             478 arch/s390/pci/pci.c static void zpci_free_iomap(struct zpci_dev *zdev, int entry)
entry             481 arch/s390/pci/pci.c 	memset(&zpci_iomap_start[entry], 0, sizeof(struct zpci_iomap_entry));
entry             482 arch/s390/pci/pci.c 	clear_bit(entry, zpci_iomap_bitmap);
entry             512 arch/s390/pci/pci.c 	int i, entry;
entry             520 arch/s390/pci/pci.c 		entry = zpci_alloc_iomap(zdev);
entry             521 arch/s390/pci/pci.c 		if (entry < 0)
entry             522 arch/s390/pci/pci.c 			return entry;
entry             523 arch/s390/pci/pci.c 		zdev->bars[i].map_idx = entry;
entry             535 arch/s390/pci/pci.c 			addr = ZPCI_ADDR(entry);
entry             540 arch/s390/pci/pci.c 			zpci_free_iomap(zdev, entry);
entry             700 arch/s390/pci/pci.c 	list_del(&zdev->entry);
entry             784 arch/s390/pci/pci.c 	list_add_tail(&zdev->entry, &zpci_list);
entry             360 arch/s390/pci/pci_clp.c static void __clp_add(struct clp_fh_list_entry *entry, void *data)
entry             364 arch/s390/pci/pci_clp.c 	if (!entry->vendor_id)
entry             367 arch/s390/pci/pci_clp.c 	zdev = get_zdev_by_fid(entry->fid);
entry             369 arch/s390/pci/pci_clp.c 		clp_add_pci_device(entry->fid, entry->fh, entry->config_state);
entry             372 arch/s390/pci/pci_clp.c static void __clp_update(struct clp_fh_list_entry *entry, void *data)
entry             377 arch/s390/pci/pci_clp.c 	if (!entry->vendor_id)
entry             380 arch/s390/pci/pci_clp.c 	if (fid && *fid != entry->fid)
entry             383 arch/s390/pci/pci_clp.c 	zdev = get_zdev_by_fid(entry->fid);
entry             387 arch/s390/pci/pci_clp.c 	zdev->fh = entry->fh;
entry             445 arch/s390/pci/pci_clp.c static void __clp_get_state(struct clp_fh_list_entry *entry, void *data)
entry             449 arch/s390/pci/pci_clp.c 	if (entry->fid != sd->fid)
entry             452 arch/s390/pci/pci_clp.c 	sd->state = entry->config_state;
entry              30 arch/s390/pci/pci_dma.c 	unsigned long *table, *entry;
entry              36 arch/s390/pci/pci_dma.c 	for (entry = table; entry < table + ZPCI_TABLE_ENTRIES; entry++)
entry              37 arch/s390/pci/pci_dma.c 		*entry = ZPCI_TABLE_INVALID;
entry              48 arch/s390/pci/pci_dma.c 	unsigned long *table, *entry;
entry              54 arch/s390/pci/pci_dma.c 	for (entry = table; entry < table + ZPCI_PT_ENTRIES; entry++)
entry              55 arch/s390/pci/pci_dma.c 		*entry = ZPCI_PTE_INVALID;
entry              64 arch/s390/pci/pci_dma.c static unsigned long *dma_get_seg_table_origin(unsigned long *entry)
entry              68 arch/s390/pci/pci_dma.c 	if (reg_entry_isvalid(*entry))
entry              69 arch/s390/pci/pci_dma.c 		sto = get_rt_sto(*entry);
entry              75 arch/s390/pci/pci_dma.c 		set_rt_sto(entry, sto);
entry              76 arch/s390/pci/pci_dma.c 		validate_rt_entry(entry);
entry              77 arch/s390/pci/pci_dma.c 		entry_clr_protected(entry);
entry              82 arch/s390/pci/pci_dma.c static unsigned long *dma_get_page_table_origin(unsigned long *entry)
entry              86 arch/s390/pci/pci_dma.c 	if (reg_entry_isvalid(*entry))
entry              87 arch/s390/pci/pci_dma.c 		pto = get_st_pto(*entry);
entry              92 arch/s390/pci/pci_dma.c 		set_st_pto(entry, pto);
entry              93 arch/s390/pci/pci_dma.c 		validate_st_entry(entry);
entry              94 arch/s390/pci/pci_dma.c 		entry_clr_protected(entry);
entry             118 arch/s390/pci/pci_dma.c void dma_update_cpu_trans(unsigned long *entry, void *page_addr, int flags)
entry             121 arch/s390/pci/pci_dma.c 		invalidate_pt_entry(entry);
entry             123 arch/s390/pci/pci_dma.c 		set_pt_pfaa(entry, page_addr);
entry             124 arch/s390/pci/pci_dma.c 		validate_pt_entry(entry);
entry             128 arch/s390/pci/pci_dma.c 		entry_set_protected(entry);
entry             130 arch/s390/pci/pci_dma.c 		entry_clr_protected(entry);
entry             139 arch/s390/pci/pci_dma.c 	unsigned long *entry;
entry             152 arch/s390/pci/pci_dma.c 		entry = dma_walk_cpu_trans(zdev->dma_table, dma_addr);
entry             153 arch/s390/pci/pci_dma.c 		if (!entry) {
entry             157 arch/s390/pci/pci_dma.c 		dma_update_cpu_trans(entry, page_addr, flags);
entry             168 arch/s390/pci/pci_dma.c 			entry = dma_walk_cpu_trans(zdev->dma_table, dma_addr);
entry             169 arch/s390/pci/pci_dma.c 			if (!entry)
entry             171 arch/s390/pci/pci_dma.c 			dma_update_cpu_trans(entry, page_addr, flags);
entry             234 arch/s390/pci/pci_dma.c void dma_free_seg_table(unsigned long entry)
entry             236 arch/s390/pci/pci_dma.c 	unsigned long *sto = get_rt_sto(entry);
entry             168 arch/s390/pci/pci_insn.c 	struct zpci_iomap_entry *entry = &zpci_iomap_start[ZPCI_IDX(addr)];
entry             169 arch/s390/pci/pci_insn.c 	u64 req = ZPCI_CREATE_REQ(entry->fh, entry->bar, len);
entry             252 arch/s390/pci/pci_insn.c 	struct zpci_iomap_entry *entry = &zpci_iomap_start[ZPCI_IDX(addr)];
entry             253 arch/s390/pci/pci_insn.c 	u64 req = ZPCI_CREATE_REQ(entry->fh, entry->bar, len);
entry             332 arch/s390/pci/pci_insn.c 	struct zpci_iomap_entry *entry = &zpci_iomap_start[ZPCI_IDX(dst)];
entry             333 arch/s390/pci/pci_insn.c 	u64 req = ZPCI_CREATE_REQ(entry->fh, entry->bar, len);
entry             104 arch/s390/pci/pci_irq.c 	struct msi_desc *entry = irq_get_msi_desc(data->irq);
entry             105 arch/s390/pci/pci_irq.c 	struct msi_msg msg = entry->msg;
entry              55 arch/sh/include/asm/tlb_64.h int sh64_put_wired_dtlb_entry(unsigned long long entry);
entry              63 arch/sh/include/asm/tlb_64.h #define sh64_put_wired_dtlb_entry(entry)		do { } while (0)
entry             290 arch/sh/kernel/cpu/sh4/sq.c 	struct sq_mapping **list, *entry;
entry             293 arch/sh/kernel/cpu/sh4/sq.c 	for (list = &sq_mapping_list; (entry = *list); list = &entry->next)
entry             295 arch/sh/kernel/cpu/sh4/sq.c 			     entry->sq_addr, entry->sq_addr + entry->size,
entry             296 arch/sh/kernel/cpu/sh4/sq.c 			     entry->addr, entry->name);
entry             735 arch/sh/kernel/dwarf.c static int dwarf_parse_cie(void *entry, void *p, unsigned long len,
entry             756 arch/sh/kernel/dwarf.c 	cie->cie_pointer = (unsigned long)entry;
entry             865 arch/sh/kernel/dwarf.c static int dwarf_parse_fde(void *entry, u32 entry_type,
entry            1028 arch/sh/kernel/dwarf.c 	void *p, *entry;
entry            1036 arch/sh/kernel/dwarf.c 	entry = eh_frame_start;
entry            1038 arch/sh/kernel/dwarf.c 	while ((char *)entry < eh_frame_end) {
entry            1039 arch/sh/kernel/dwarf.c 		p = entry;
entry            1062 arch/sh/kernel/dwarf.c 			err = dwarf_parse_cie(entry, p, len, end, mod);
entry            1068 arch/sh/kernel/dwarf.c 			err = dwarf_parse_fde(entry, entry_type, p, len,
entry            1076 arch/sh/kernel/dwarf.c 		entry = (char *)entry + len + 4;
entry             142 arch/sh/kernel/io_trapped.c 	pte_t entry;
entry             157 arch/sh/kernel/io_trapped.c 	entry = *pte_k;
entry             159 arch/sh/kernel/io_trapped.c 	return pfn_to_kaddr(pte_pfn(entry));
entry             416 arch/sh/kernel/kprobes.c 	const struct exception_table_entry *entry;
entry             458 arch/sh/kernel/kprobes.c 		if ((entry = search_exception_tables(regs->pc)) != NULL) {
entry             459 arch/sh/kernel/kprobes.c 			regs->pc = entry->fixup;
entry              76 arch/sh/kernel/machine_kexec.c 	unsigned long entry;
entry              85 arch/sh/kernel/machine_kexec.c 	for (ptr = &image->head; (entry = *ptr) && !(entry & IND_DONE);
entry              86 arch/sh/kernel/machine_kexec.c 	     ptr = (entry & IND_INDIRECTION) ?
entry              87 arch/sh/kernel/machine_kexec.c 	       phys_to_virt(entry & PAGE_MASK) : ptr + 1) {
entry             130 arch/sh/kernel/machine_kexec.c 	for (ptr = &image->head; (entry = *ptr) && !(entry & IND_DONE);
entry              21 arch/sh/kernel/perf_callchain.c 	struct perf_callchain_entry_ctx *entry = data;
entry              24 arch/sh/kernel/perf_callchain.c 		perf_callchain_store(entry, addr);
entry              33 arch/sh/kernel/perf_callchain.c perf_callchain_kernel(struct perf_callchain_entry_ctx *entry, struct pt_regs *regs)
entry              35 arch/sh/kernel/perf_callchain.c 	perf_callchain_store(entry, regs->pc);
entry              37 arch/sh/kernel/perf_callchain.c 	unwind_stack(NULL, regs, NULL, &callchain_ops, entry);
entry              85 arch/sh/kernel/unwinder.c 	struct list_head *tmp, *entry = &unwinder_list;
entry              95 arch/sh/kernel/unwinder.c 			entry = tmp;
entry              97 arch/sh/kernel/unwinder.c 	list_add(&ops->list, entry);
entry             389 arch/sh/mm/cache-sh5.c 	pte_t entry;
entry             410 arch/sh/mm/cache-sh5.c 		entry = *pte;
entry             411 arch/sh/mm/cache-sh5.c 		if (pte_none(entry) || !pte_present(entry))
entry             413 arch/sh/mm/cache-sh5.c 		paddr = pte_val(entry) & PAGE_MASK;
entry              48 arch/sh/mm/pmb.c 	int entry;
entry              72 arch/sh/mm/pmb.c static __always_inline unsigned long mk_pmb_entry(unsigned int entry)
entry              74 arch/sh/mm/pmb.c 	return (entry & PMB_E_MASK) << PMB_E_SHIFT;
entry              77 arch/sh/mm/pmb.c static __always_inline unsigned long mk_pmb_addr(unsigned int entry)
entry              79 arch/sh/mm/pmb.c 	return mk_pmb_entry(entry) | PMB_ADDR;
entry              82 arch/sh/mm/pmb.c static __always_inline unsigned long mk_pmb_data(unsigned int entry)
entry              84 arch/sh/mm/pmb.c 	return mk_pmb_entry(entry) | PMB_DATA;
entry             237 arch/sh/mm/pmb.c 				   unsigned long flags, int entry)
entry             246 arch/sh/mm/pmb.c 	if (entry == PMB_NO_ENTRY) {
entry             253 arch/sh/mm/pmb.c 		if (__test_and_set_bit(entry, pmb_map)) {
entry             258 arch/sh/mm/pmb.c 		pos = entry;
entry             272 arch/sh/mm/pmb.c 	pmbe->entry	= pos;
entry             283 arch/sh/mm/pmb.c 	__clear_bit(pmbe->entry, pmb_map);
entry             285 arch/sh/mm/pmb.c 	pmbe->entry	= PMB_NO_ENTRY;
entry             296 arch/sh/mm/pmb.c 	addr = mk_pmb_addr(pmbe->entry);
entry             297 arch/sh/mm/pmb.c 	data = mk_pmb_data(pmbe->entry);
entry             313 arch/sh/mm/pmb.c 	addr = mk_pmb_addr(pmbe->entry);
entry             314 arch/sh/mm/pmb.c 	data = mk_pmb_data(pmbe->entry);
entry              45 arch/sh/mm/tlb-debugfs.c 	unsigned int nentries, entry;
entry              93 arch/sh/mm/tlb-debugfs.c 	for (entry = 0; entry < nentries; entry++) {
entry             100 arch/sh/mm/tlb-debugfs.c 		val = __raw_readl(addr1 | (entry << MMU_TLB_ENTRY_SHIFT));
entry             105 arch/sh/mm/tlb-debugfs.c 		val = __raw_readl(addr2 | (entry << MMU_TLB_ENTRY_SHIFT));
entry             109 arch/sh/mm/tlb-debugfs.c 		val = __raw_readl(data1 | (entry << MMU_TLB_ENTRY_SHIFT));
entry             113 arch/sh/mm/tlb-debugfs.c 		val = __raw_readl(data2 | (entry << MMU_TLB_ENTRY_SHIFT));
entry             126 arch/sh/mm/tlb-debugfs.c 			   entry, vpn, ppn, asid,
entry             128 arch/sh/mm/tlb-debugfs.c 			   (urb <= entry) ? "W" : "-");
entry              59 arch/sh/mm/tlb-sh5.c 	unsigned long long entry = sh64_next_free_dtlb_entry();
entry              64 arch/sh/mm/tlb-sh5.c 	return entry;
entry              74 arch/sh/mm/tlb-sh5.c int sh64_put_wired_dtlb_entry(unsigned long long entry)
entry              76 arch/sh/mm/tlb-sh5.c 	__flush_tlb_slot(entry);
entry              92 arch/sh/mm/tlb-sh5.c 	if (entry <= DTLB_FIXED)
entry              99 arch/sh/mm/tlb-sh5.c 	if (entry < (cpu_data->dtlb.first - cpu_data->dtlb.step))
entry             104 arch/sh/mm/tlb-sh5.c 	cpu_data->dtlb.next	= entry;
entry             152 arch/sh/mm/tlb-sh5.c 	unsigned long long entry;
entry             159 arch/sh/mm/tlb-sh5.c 	entry = sh64_get_wired_dtlb_entry();
entry             160 arch/sh/mm/tlb-sh5.c 	dtlb_entries[dtlb_entry++] = entry;
entry             165 arch/sh/mm/tlb-sh5.c 	sh64_setup_tlb_slot(entry, addr, get_asid(), paddr);
entry             172 arch/sh/mm/tlb-sh5.c 	unsigned long long entry;
entry             178 arch/sh/mm/tlb-sh5.c 	entry = dtlb_entries[dtlb_entry--];
entry             180 arch/sh/mm/tlb-sh5.c 	sh64_teardown_tlb_slot(entry);
entry             181 arch/sh/mm/tlb-sh5.c 	sh64_put_wired_dtlb_entry(entry);
entry              29 arch/sh/mm/tlbex_32.c 	pte_t entry;
entry              52 arch/sh/mm/tlbex_32.c 	entry = *pte;
entry              53 arch/sh/mm/tlbex_32.c 	if (unlikely(pte_none(entry) || pte_not_present(entry)))
entry              55 arch/sh/mm/tlbex_32.c 	if (unlikely(error_code && !pte_write(entry)))
entry              59 arch/sh/mm/tlbex_32.c 		entry = pte_mkdirty(entry);
entry              60 arch/sh/mm/tlbex_32.c 	entry = pte_mkyoung(entry);
entry              62 arch/sh/mm/tlbex_32.c 	set_pte(pte, entry);
entry              50 arch/sh/mm/tlbex_64.c 	pte_t entry;
entry              70 arch/sh/mm/tlbex_64.c 	entry = *pte;
entry              71 arch/sh/mm/tlbex_64.c 	if (pte_none(entry) || !pte_present(entry))
entry              80 arch/sh/mm/tlbex_64.c 	if ((pte_val(entry) & protection_flags) != protection_flags)
entry              51 arch/sparc/include/asm/iommu-common.h 				 unsigned long entry);
entry             352 arch/sparc/include/asm/pgtable_32.h static inline unsigned long __swp_type(swp_entry_t entry)
entry             354 arch/sparc/include/asm/pgtable_32.h 	return (entry.val >> SRMMU_SWP_TYPE_SHIFT) & SRMMU_SWP_TYPE_MASK;
entry             357 arch/sparc/include/asm/pgtable_32.h static inline unsigned long __swp_offset(swp_entry_t entry)
entry             359 arch/sparc/include/asm/pgtable_32.h 	return (entry.val >> SRMMU_SWP_OFF_SHIFT) & SRMMU_SWP_OFF_MASK;
entry             380 arch/sparc/include/asm/pgtable_64.h extern pte_t arch_make_huge_pte(pte_t entry, struct vm_area_struct *vma,
entry             786 arch/sparc/include/asm/pgtable_64.h static inline pgprot_t pmd_pgprot(pmd_t entry)
entry             788 arch/sparc/include/asm/pgtable_64.h 	unsigned long val = pmd_val(entry);
entry            1010 arch/sparc/include/asm/pgtable_64.h #define __swp_type(entry)	(((entry).val >> PAGE_SHIFT) & 0xffUL)
entry            1011 arch/sparc/include/asm/pgtable_64.h #define __swp_offset(entry)	((entry).val >> (PAGE_SHIFT + 8UL))
entry             169 arch/sparc/include/asm/pgtsrmmu.h 	register unsigned long entry;
entry             172 arch/sparc/include/asm/pgtsrmmu.h 				"=r" (entry):
entry             174 arch/sparc/include/asm/pgtsrmmu.h 	return entry;
entry             119 arch/sparc/include/asm/spitfire.h static inline unsigned long spitfire_get_dtlb_data(int entry)
entry             125 arch/sparc/include/asm/spitfire.h 			     : "r" (entry << 3), "i" (ASI_DTLB_DATA_ACCESS));
entry             133 arch/sparc/include/asm/spitfire.h static inline unsigned long spitfire_get_dtlb_tag(int entry)
entry             139 arch/sparc/include/asm/spitfire.h 			     : "r" (entry << 3), "i" (ASI_DTLB_TAG_READ));
entry             143 arch/sparc/include/asm/spitfire.h static inline void spitfire_put_dtlb_data(int entry, unsigned long data)
entry             148 arch/sparc/include/asm/spitfire.h 			     : "r" (data), "r" (entry << 3),
entry             152 arch/sparc/include/asm/spitfire.h static inline unsigned long spitfire_get_itlb_data(int entry)
entry             158 arch/sparc/include/asm/spitfire.h 			     : "r" (entry << 3), "i" (ASI_ITLB_DATA_ACCESS));
entry             166 arch/sparc/include/asm/spitfire.h static inline unsigned long spitfire_get_itlb_tag(int entry)
entry             172 arch/sparc/include/asm/spitfire.h 			     : "r" (entry << 3), "i" (ASI_ITLB_TAG_READ));
entry             176 arch/sparc/include/asm/spitfire.h static inline void spitfire_put_itlb_data(int entry, unsigned long data)
entry             181 arch/sparc/include/asm/spitfire.h 			     : "r" (data), "r" (entry << 3),
entry             232 arch/sparc/include/asm/spitfire.h static inline unsigned long cheetah_get_ldtlb_data(int entry)
entry             239 arch/sparc/include/asm/spitfire.h 			     : "r" ((0 << 16) | (entry << 3)),
entry             245 arch/sparc/include/asm/spitfire.h static inline unsigned long cheetah_get_litlb_data(int entry)
entry             252 arch/sparc/include/asm/spitfire.h 			     : "r" ((0 << 16) | (entry << 3)),
entry             258 arch/sparc/include/asm/spitfire.h static inline unsigned long cheetah_get_ldtlb_tag(int entry)
entry             264 arch/sparc/include/asm/spitfire.h 			     : "r" ((0 << 16) | (entry << 3)),
entry             270 arch/sparc/include/asm/spitfire.h static inline unsigned long cheetah_get_litlb_tag(int entry)
entry             276 arch/sparc/include/asm/spitfire.h 			     : "r" ((0 << 16) | (entry << 3)),
entry             282 arch/sparc/include/asm/spitfire.h static inline void cheetah_put_ldtlb_data(int entry, unsigned long data)
entry             288 arch/sparc/include/asm/spitfire.h 			       "r" ((0 << 16) | (entry << 3)),
entry             292 arch/sparc/include/asm/spitfire.h static inline void cheetah_put_litlb_data(int entry, unsigned long data)
entry             298 arch/sparc/include/asm/spitfire.h 			       "r" ((0 << 16) | (entry << 3)),
entry             302 arch/sparc/include/asm/spitfire.h static inline unsigned long cheetah_get_dtlb_data(int entry, int tlb)
entry             309 arch/sparc/include/asm/spitfire.h 			     : "r" ((tlb << 16) | (entry << 3)), "i" (ASI_DTLB_DATA_ACCESS));
entry             314 arch/sparc/include/asm/spitfire.h static inline unsigned long cheetah_get_dtlb_tag(int entry, int tlb)
entry             320 arch/sparc/include/asm/spitfire.h 			     : "r" ((tlb << 16) | (entry << 3)), "i" (ASI_DTLB_TAG_READ));
entry             324 arch/sparc/include/asm/spitfire.h static inline void cheetah_put_dtlb_data(int entry, unsigned long data, int tlb)
entry             330 arch/sparc/include/asm/spitfire.h 			       "r" ((tlb << 16) | (entry << 3)),
entry             334 arch/sparc/include/asm/spitfire.h static inline unsigned long cheetah_get_itlb_data(int entry)
entry             341 arch/sparc/include/asm/spitfire.h 			     : "r" ((2 << 16) | (entry << 3)),
entry             347 arch/sparc/include/asm/spitfire.h static inline unsigned long cheetah_get_itlb_tag(int entry)
entry             353 arch/sparc/include/asm/spitfire.h 			     : "r" ((2 << 16) | (entry << 3)), "i" (ASI_ITLB_TAG_READ));
entry             357 arch/sparc/include/asm/spitfire.h static inline void cheetah_put_itlb_data(int entry, unsigned long data)
entry             362 arch/sparc/include/asm/spitfire.h 			     : "r" (data), "r" ((2 << 16) | (entry << 3)),
entry             228 arch/sparc/kernel/iommu-common.c 				   unsigned long entry)
entry             235 arch/sparc/kernel/iommu-common.c 	if (large_pool && entry >= largepool_start) {
entry             238 arch/sparc/kernel/iommu-common.c 		unsigned int pool_nr = entry / tbl->poolsize;
entry             251 arch/sparc/kernel/iommu-common.c 			  unsigned long npages, unsigned long entry)
entry             257 arch/sparc/kernel/iommu-common.c 	if (entry == IOMMU_ERROR_CODE) /* use default addr->entry mapping */
entry             258 arch/sparc/kernel/iommu-common.c 		entry = (dma_addr - iommu->table_map_base) >> shift;
entry             259 arch/sparc/kernel/iommu-common.c 	pool = get_pool(iommu, entry);
entry             262 arch/sparc/kernel/iommu-common.c 	bitmap_clear(iommu->map, entry, npages);
entry              57 arch/sparc/kernel/iommu.c 		int entry;
entry              60 arch/sparc/kernel/iommu.c 		for (entry = 0; entry < 16; entry++) {
entry             160 arch/sparc/kernel/iommu.c 	unsigned long entry;
entry             162 arch/sparc/kernel/iommu.c 	entry = iommu_tbl_range_alloc(dev, &iommu->tbl, npages, NULL,
entry             164 arch/sparc/kernel/iommu.c 	if (unlikely(entry == IOMMU_ERROR_CODE))
entry             167 arch/sparc/kernel/iommu.c 	return iommu->page_table + entry;
entry             479 arch/sparc/kernel/iommu.c 		unsigned long paddr, npages, entry, out_entry = 0, slen;
entry             491 arch/sparc/kernel/iommu.c 		entry = iommu_tbl_range_alloc(dev, &iommu->tbl, npages,
entry             495 arch/sparc/kernel/iommu.c 		if (unlikely(entry == IOMMU_ERROR_CODE)) {
entry             502 arch/sparc/kernel/iommu.c 		base = iommu->page_table + entry;
entry             506 arch/sparc/kernel/iommu.c 			(entry << IO_PAGE_SHIFT);
entry             539 arch/sparc/kernel/iommu.c 			out_entry = entry;
entry             559 arch/sparc/kernel/iommu.c 			unsigned long vaddr, npages, entry, j;
entry             566 arch/sparc/kernel/iommu.c 			entry = (vaddr - iommu->tbl.table_map_base)
entry             568 arch/sparc/kernel/iommu.c 			base = iommu->page_table + entry;
entry             630 arch/sparc/kernel/iommu.c 		unsigned long npages, entry;
entry             638 arch/sparc/kernel/iommu.c 		entry = ((dma_handle - iommu->tbl.table_map_base)
entry             640 arch/sparc/kernel/iommu.c 		base = iommu->page_table + entry;
entry              38 arch/sparc/kernel/iommu_common.h static inline int is_span_boundary(unsigned long entry,
entry              48 arch/sparc/kernel/iommu_common.h 	return iommu_is_span_boundary(entry, nr, shift, boundary_size);
entry              12 arch/sparc/kernel/jump_label.c void arch_jump_label_transform(struct jump_entry *entry,
entry              15 arch/sparc/kernel/jump_label.c 	u32 *insn = (u32 *) (unsigned long) entry->code;
entry              19 arch/sparc/kernel/jump_label.c 		s32 off = (s32)entry->target - (s32)entry->code;
entry             324 arch/sparc/kernel/kprobes.c 	const struct exception_table_entry *entry;
entry             370 arch/sparc/kernel/kprobes.c 		entry = search_exception_tables(regs->tpc);
entry             371 arch/sparc/kernel/kprobes.c 		if (entry) {
entry             372 arch/sparc/kernel/kprobes.c 			regs->tpc = entry->fixup;
entry            1023 arch/sparc/kernel/ldc.c 		      unsigned long entry, unsigned long npages)
entry            1029 arch/sparc/kernel/ldc.c 	base = iommu->page_table + entry;
entry            2020 arch/sparc/kernel/ldc.c 	long entry;
entry            2022 arch/sparc/kernel/ldc.c 	entry = iommu_tbl_range_alloc(NULL, &iommu->iommu_map_table,
entry            2024 arch/sparc/kernel/ldc.c 	if (unlikely(entry == IOMMU_ERROR_CODE))
entry            2027 arch/sparc/kernel/ldc.c 	return iommu->page_table + entry;
entry            2236 arch/sparc/kernel/ldc.c 	unsigned long npages, entry;
entry            2240 arch/sparc/kernel/ldc.c 	entry = ldc_cookie_to_index(cookie, iommu);
entry            2241 arch/sparc/kernel/ldc.c 	ldc_demap(iommu, id, cookie, entry, npages);
entry            2242 arch/sparc/kernel/ldc.c 	iommu_tbl_range_free(&iommu->iommu_map_table, cookie, npages, entry);
entry             950 arch/sparc/kernel/pci.c 	struct msi_desc *entry = irq_get_msi_desc(irq);
entry             951 arch/sparc/kernel/pci.c 	struct pci_dev *pdev = msi_desc_to_pci_dev(entry);
entry             142 arch/sparc/kernel/pci_impl.h 			     struct msi_desc *entry);
entry             124 arch/sparc/kernel/pci_msi.c 				 struct msi_desc *entry)
entry             149 arch/sparc/kernel/pci_msi.c 			     (entry->msi_attrib.is_64 ? 1 : 0));
entry             155 arch/sparc/kernel/pci_msi.c 	if (entry->msi_attrib.is_64) {
entry             164 arch/sparc/kernel/pci_msi.c 	irq_set_msi_desc(*irq_p, entry);
entry              57 arch/sparc/kernel/pci_sun4v.c 	unsigned long	entry;		/* Index into IOTSB.		*/
entry              66 arch/sparc/kernel/pci_sun4v.c static inline void iommu_batch_start(struct device *dev, unsigned long prot, unsigned long entry)
entry              72 arch/sparc/kernel/pci_sun4v.c 	p->entry	= entry;
entry              89 arch/sparc/kernel/pci_sun4v.c 	unsigned long entry = p->entry;
entry             102 arch/sparc/kernel/pci_sun4v.c 						  HV_PCI_TSBID(0, entry),
entry             110 arch/sparc/kernel/pci_sun4v.c 						   HV_PCI_TSBID(0, entry),
entry             116 arch/sparc/kernel/pci_sun4v.c 			index_count = HV_PCI_IOTSB_INDEX_COUNT(npages, entry),
entry             133 arch/sparc/kernel/pci_sun4v.c 		entry += num;
entry             138 arch/sparc/kernel/pci_sun4v.c 	p->entry = entry;
entry             144 arch/sparc/kernel/pci_sun4v.c static inline void iommu_batch_new_entry(unsigned long entry, u64 mask)
entry             148 arch/sparc/kernel/pci_sun4v.c 	if (p->entry + p->npages == entry)
entry             150 arch/sparc/kernel/pci_sun4v.c 	if (p->entry != ~0UL)
entry             152 arch/sparc/kernel/pci_sun4v.c 	p->entry = entry;
entry             190 arch/sparc/kernel/pci_sun4v.c 	long entry;
entry             218 arch/sparc/kernel/pci_sun4v.c 	entry = iommu_tbl_range_alloc(dev, tbl, npages, NULL,
entry             221 arch/sparc/kernel/pci_sun4v.c 	if (unlikely(entry == IOMMU_ERROR_CODE))
entry             224 arch/sparc/kernel/pci_sun4v.c 	*dma_addrp = (tbl->table_map_base + (entry << IO_PAGE_SHIFT));
entry             233 arch/sparc/kernel/pci_sun4v.c 			  entry);
entry             296 arch/sparc/kernel/pci_sun4v.c 			       unsigned long entry, unsigned long npages)
entry             305 arch/sparc/kernel/pci_sun4v.c 						    HV_PCI_TSBID(0, entry),
entry             309 arch/sparc/kernel/pci_sun4v.c 						    entry, npages, &num);
entry             315 arch/sparc/kernel/pci_sun4v.c 		entry += num;
entry             328 arch/sparc/kernel/pci_sun4v.c 	unsigned long order, npages, entry;
entry             345 arch/sparc/kernel/pci_sun4v.c 	entry = ((dvma - tbl->table_map_base) >> IO_PAGE_SHIFT);
entry             346 arch/sparc/kernel/pci_sun4v.c 	dma_4v_iommu_demap(dev, devhandle, dvma, iotsb_num, entry, npages);
entry             366 arch/sparc/kernel/pci_sun4v.c 	long entry;
entry             384 arch/sparc/kernel/pci_sun4v.c 	entry = iommu_tbl_range_alloc(dev, tbl, npages, NULL,
entry             387 arch/sparc/kernel/pci_sun4v.c 	if (unlikely(entry == IOMMU_ERROR_CODE))
entry             390 arch/sparc/kernel/pci_sun4v.c 	bus_addr = (tbl->table_map_base + (entry << IO_PAGE_SHIFT));
entry             402 arch/sparc/kernel/pci_sun4v.c 	iommu_batch_start(dev, prot, entry);
entry             437 arch/sparc/kernel/pci_sun4v.c 	long entry;
entry             462 arch/sparc/kernel/pci_sun4v.c 	entry = (bus_addr - tbl->table_map_base) >> IO_PAGE_SHIFT;
entry             463 arch/sparc/kernel/pci_sun4v.c 	dma_4v_iommu_demap(dev, devhandle, bus_addr, iotsb_num, entry, npages);
entry             523 arch/sparc/kernel/pci_sun4v.c 		unsigned long paddr, npages, entry, out_entry = 0, slen;
entry             534 arch/sparc/kernel/pci_sun4v.c 		entry = iommu_tbl_range_alloc(dev, tbl, npages,
entry             538 arch/sparc/kernel/pci_sun4v.c 		if (unlikely(entry == IOMMU_ERROR_CODE)) {
entry             544 arch/sparc/kernel/pci_sun4v.c 		iommu_batch_new_entry(entry, mask);
entry             547 arch/sparc/kernel/pci_sun4v.c 		dma_addr = tbl->table_map_base + (entry << IO_PAGE_SHIFT);
entry             581 arch/sparc/kernel/pci_sun4v.c 			out_entry = entry;
entry             633 arch/sparc/kernel/pci_sun4v.c 	unsigned long flags, entry;
entry             665 arch/sparc/kernel/pci_sun4v.c 		entry = ((dma_handle - tbl->table_map_base) >> shift);
entry             667 arch/sparc/kernel/pci_sun4v.c 				   entry, npages);
entry            1737 arch/sparc/kernel/perf_event.c void perf_callchain_kernel(struct perf_callchain_entry_ctx *entry,
entry            1747 arch/sparc/kernel/perf_event.c 	perf_callchain_store(entry, regs->tpc);
entry            1771 arch/sparc/kernel/perf_event.c 		perf_callchain_store(entry, pc);
entry            1779 arch/sparc/kernel/perf_event.c 				perf_callchain_store(entry, pc);
entry            1784 arch/sparc/kernel/perf_event.c 	} while (entry->nr < entry->max_stack);
entry            1797 arch/sparc/kernel/perf_event.c static void perf_callchain_user_64(struct perf_callchain_entry_ctx *entry,
entry            1817 arch/sparc/kernel/perf_event.c 		perf_callchain_store(entry, pc);
entry            1818 arch/sparc/kernel/perf_event.c 	} while (entry->nr < entry->max_stack);
entry            1821 arch/sparc/kernel/perf_event.c static void perf_callchain_user_32(struct perf_callchain_entry_ctx *entry,
entry            1849 arch/sparc/kernel/perf_event.c 		perf_callchain_store(entry, pc);
entry            1850 arch/sparc/kernel/perf_event.c 	} while (entry->nr < entry->max_stack);
entry            1854 arch/sparc/kernel/perf_event.c perf_callchain_user(struct perf_callchain_entry_ctx *entry, struct pt_regs *regs)
entry            1859 arch/sparc/kernel/perf_event.c 	perf_callchain_store(entry, regs->tpc);
entry            1869 arch/sparc/kernel/perf_event.c 		perf_callchain_user_32(entry, regs);
entry            1871 arch/sparc/kernel/perf_event.c 		perf_callchain_user_64(entry, regs);
entry             355 arch/sparc/kernel/smp_64.c 	unsigned long entry =
entry             373 arch/sparc/kernel/smp_64.c 			prom_startcpu_cpuid(cpu, entry, cookie);
entry             377 arch/sparc/kernel/smp_64.c 		prom_startcpu(dp->phandle, entry, cookie);
entry             124 arch/sparc/kernel/sun4d_smp.c 	unsigned long *entry = &sun4d_cpu_startup;
entry             140 arch/sparc/kernel/sun4d_smp.c 	printk(KERN_INFO "Starting CPU %d at %p\n", i, entry);
entry             143 arch/sparc/kernel/sun4d_smp.c 		      &smp_penguin_ctable, 0, (char *)entry);
entry              81 arch/sparc/kernel/sun4m_smp.c 	unsigned long *entry = &sun4m_cpu_startup;
entry              89 arch/sparc/kernel/sun4m_smp.c 	entry += ((i - 1) * 3);
entry             101 arch/sparc/kernel/sun4m_smp.c 	printk(KERN_INFO "Starting CPU %d at %p\n", i, entry);
entry             103 arch/sparc/kernel/sun4m_smp.c 	prom_startcpu(cpu_node, &smp_penguin_ctable, 0, (char *)entry);
entry             302 arch/sparc/kernel/traps_64.c 		const struct exception_table_entry *entry;
entry             304 arch/sparc/kernel/traps_64.c 		entry = search_exception_tables(regs->tpc);
entry             305 arch/sparc/kernel/traps_64.c 		if (entry) {
entry             310 arch/sparc/kernel/traps_64.c 			       regs->tpc, entry->fixup);
entry             312 arch/sparc/kernel/traps_64.c 			regs->tpc = entry->fixup;
entry             351 arch/sparc/kernel/traps_64.c 		const struct exception_table_entry *entry;
entry             353 arch/sparc/kernel/traps_64.c 		entry = search_exception_tables(regs->tpc);
entry             354 arch/sparc/kernel/traps_64.c 		if (entry) {
entry             359 arch/sparc/kernel/traps_64.c 			       regs->tpc, entry->fixup);
entry             361 arch/sparc/kernel/traps_64.c 			regs->tpc = entry->fixup;
entry            1726 arch/sparc/kernel/traps_64.c 			const struct exception_table_entry *entry;
entry            1728 arch/sparc/kernel/traps_64.c 			entry = search_exception_tables(regs->tpc);
entry            1729 arch/sparc/kernel/traps_64.c 			if (entry) {
entry            1748 arch/sparc/kernel/traps_64.c 					regs->tpc = entry->fixup;
entry            2053 arch/sparc/kernel/traps_64.c 		const struct exception_table_entry *entry;
entry            2055 arch/sparc/kernel/traps_64.c 		entry = search_exception_tables(regs->tpc);
entry            2056 arch/sparc/kernel/traps_64.c 		if (entry) {
entry            2062 arch/sparc/kernel/traps_64.c 				 ent.err_raddr, entry->fixup);
entry            2064 arch/sparc/kernel/traps_64.c 			regs->tpc = entry->fixup;
entry            2677 arch/sparc/kernel/traps_64.c 		const struct exception_table_entry *entry;
entry            2679 arch/sparc/kernel/traps_64.c 		entry = search_exception_tables(regs->tpc);
entry            2680 arch/sparc/kernel/traps_64.c 		if (entry) {
entry            2686 arch/sparc/kernel/traps_64.c 				 regs->tpc, entry->fixup);
entry            2688 arch/sparc/kernel/traps_64.c 			regs->tpc = entry->fixup;
entry             263 arch/sparc/kernel/unaligned_64.c 	const struct exception_table_entry *entry;
entry             265 arch/sparc/kernel/unaligned_64.c 	entry = search_exception_tables(regs->tpc);
entry             266 arch/sparc/kernel/unaligned_64.c 	if (!entry) {
entry             287 arch/sparc/kernel/unaligned_64.c 	regs->tpc = entry->fixup;
entry              94 arch/sparc/mm/extable.c 	const struct exception_table_entry *entry;
entry              96 arch/sparc/mm/extable.c 	entry = search_exception_tables(addr);
entry              97 arch/sparc/mm/extable.c 	if (!entry)
entry             101 arch/sparc/mm/extable.c 	if (!entry->fixup) {
entry             102 arch/sparc/mm/extable.c 		*g2 = (addr - entry->insn) / 4;
entry             103 arch/sparc/mm/extable.c 		return (entry + 1)->fixup;
entry             106 arch/sparc/mm/extable.c 	return entry->fixup;
entry             229 arch/sparc/mm/fault_64.c 		const struct exception_table_entry *entry;
entry             231 arch/sparc/mm/fault_64.c 		entry = search_exception_tables(regs->tpc);
entry             232 arch/sparc/mm/fault_64.c 		if (entry) {
entry             233 arch/sparc/mm/fault_64.c 			regs->tpc = entry->fixup;
entry             135 arch/sparc/mm/hugetlbpage.c static pte_t sun4u_hugepage_shift_to_tte(pte_t entry, unsigned int shift)
entry             137 arch/sparc/mm/hugetlbpage.c 	return entry;
entry             140 arch/sparc/mm/hugetlbpage.c static pte_t sun4v_hugepage_shift_to_tte(pte_t entry, unsigned int shift)
entry             144 arch/sparc/mm/hugetlbpage.c 	pte_val(entry) = pte_val(entry) & ~_PAGE_SZALL_4V;
entry             149 arch/sparc/mm/hugetlbpage.c 		pte_val(entry) |= _PAGE_PUD_HUGE;
entry             153 arch/sparc/mm/hugetlbpage.c 		pte_val(entry) |= _PAGE_PMD_HUGE;
entry             157 arch/sparc/mm/hugetlbpage.c 		pte_val(entry) |= _PAGE_PMD_HUGE;
entry             160 arch/sparc/mm/hugetlbpage.c 		pte_val(entry) |= _PAGE_PMD_HUGE;
entry             169 arch/sparc/mm/hugetlbpage.c 	pte_val(entry) = pte_val(entry) | hugepage_size;
entry             170 arch/sparc/mm/hugetlbpage.c 	return entry;
entry             173 arch/sparc/mm/hugetlbpage.c static pte_t hugepage_shift_to_tte(pte_t entry, unsigned int shift)
entry             176 arch/sparc/mm/hugetlbpage.c 		return sun4v_hugepage_shift_to_tte(entry, shift);
entry             178 arch/sparc/mm/hugetlbpage.c 		return sun4u_hugepage_shift_to_tte(entry, shift);
entry             181 arch/sparc/mm/hugetlbpage.c pte_t arch_make_huge_pte(pte_t entry, struct vm_area_struct *vma,
entry             187 arch/sparc/mm/hugetlbpage.c 	pte = hugepage_shift_to_tte(entry, shift);
entry             201 arch/sparc/mm/hugetlbpage.c static unsigned int sun4v_huge_tte_to_shift(pte_t entry)
entry             203 arch/sparc/mm/hugetlbpage.c 	unsigned long tte_szbits = pte_val(entry) & _PAGE_SZALL_4V;
entry             229 arch/sparc/mm/hugetlbpage.c static unsigned int sun4u_huge_tte_to_shift(pte_t entry)
entry             231 arch/sparc/mm/hugetlbpage.c 	unsigned long tte_szbits = pte_val(entry) & _PAGE_SZALL_4U;
entry             251 arch/sparc/mm/hugetlbpage.c static unsigned int huge_tte_to_shift(pte_t entry)
entry             256 arch/sparc/mm/hugetlbpage.c 		shift = sun4v_huge_tte_to_shift(entry);
entry             258 arch/sparc/mm/hugetlbpage.c 		shift = sun4u_huge_tte_to_shift(entry);
entry             262 arch/sparc/mm/hugetlbpage.c 			  pte_val(entry));
entry             321 arch/sparc/mm/hugetlbpage.c 		     pte_t *ptep, pte_t entry)
entry             327 arch/sparc/mm/hugetlbpage.c 	size = huge_tte_to_size(entry);
entry             339 arch/sparc/mm/hugetlbpage.c 	if (!pte_present(*ptep) && pte_present(entry))
entry             347 arch/sparc/mm/hugetlbpage.c 		ptep[i] = __pte(pte_val(entry) + (i << shift));
entry             361 arch/sparc/mm/hugetlbpage.c 	pte_t entry;
entry             363 arch/sparc/mm/hugetlbpage.c 	entry = *ptep;
entry             364 arch/sparc/mm/hugetlbpage.c 	size = huge_tte_to_size(entry);
entry             375 arch/sparc/mm/hugetlbpage.c 	orig_shift = pte_none(entry) ? PAGE_SHIFT : huge_tte_to_shift(entry);
entry             377 arch/sparc/mm/hugetlbpage.c 	if (pte_present(entry))
entry             384 arch/sparc/mm/hugetlbpage.c 	maybe_tlb_batch_add(mm, addr, ptep, entry, 0, orig_shift);
entry             387 arch/sparc/mm/hugetlbpage.c 		maybe_tlb_batch_add(mm, addr + REAL_HPAGE_SIZE, ptep, entry, 0,
entry             390 arch/sparc/mm/hugetlbpage.c 	return entry;
entry            2945 arch/sparc/mm/init_64.c 	pmd_t entry = *pmd;
entry            2947 arch/sparc/mm/init_64.c 	if (!pmd_large(entry) || !pmd_young(entry))
entry            2950 arch/sparc/mm/init_64.c 	pte = pmd_val(entry);
entry            2986 arch/sparc/mm/init_64.c 		const struct exception_table_entry *entry;
entry            2988 arch/sparc/mm/init_64.c 		entry = search_exception_tables(regs->tpc);
entry            2989 arch/sparc/mm/init_64.c 		if (entry) {
entry            2990 arch/sparc/mm/init_64.c 			regs->tpc = entry->fixup;
entry             419 arch/sparc/mm/srmmu.c static inline void remove_from_ctx_list(struct ctx_list *entry)
entry             421 arch/sparc/mm/srmmu.c 	entry->next->prev = entry->prev;
entry             422 arch/sparc/mm/srmmu.c 	entry->prev->next = entry->next;
entry             425 arch/sparc/mm/srmmu.c static inline void add_to_ctx_list(struct ctx_list *head, struct ctx_list *entry)
entry             427 arch/sparc/mm/srmmu.c 	entry->next = head;
entry             428 arch/sparc/mm/srmmu.c 	(entry->prev = head->prev)->next = entry;
entry             429 arch/sparc/mm/srmmu.c 	head->prev = entry;
entry             431 arch/sparc/mm/srmmu.c #define add_to_free_ctxlist(entry) add_to_ctx_list(&ctx_free, entry)
entry             432 arch/sparc/mm/srmmu.c #define add_to_used_ctxlist(entry) add_to_ctx_list(&ctx_used, entry)
entry             247 arch/sparc/mm/tlb.c 	pmd_t old, entry;
entry             249 arch/sparc/mm/tlb.c 	entry = __pmd(pmd_val(*pmdp) & ~_PAGE_VALID);
entry             250 arch/sparc/mm/tlb.c 	old = pmdp_establish(vma, address, pmdp, entry);
entry             258 arch/sparc/mm/tlb.c 	if ((pmd_val(entry) & _PAGE_PMD_HUGE) &&
entry             259 arch/sparc/mm/tlb.c 	    !is_huge_zero_page(pmd_page(entry)))
entry             441 arch/um/drivers/chan_kern.c 	const struct chan_type *entry;
entry             450 arch/um/drivers/chan_kern.c 		entry = &chan_table[i];
entry             451 arch/um/drivers/chan_kern.c 		if (!strncmp(str, entry->key, strlen(entry->key))) {
entry             452 arch/um/drivers/chan_kern.c 			ops = entry->ops;
entry             453 arch/um/drivers/chan_kern.c 			str += strlen(entry->key);
entry             348 arch/um/drivers/mconsole_kern.c 				struct list_head *entry = unplugged_pages.next;
entry             351 arch/um/drivers/mconsole_kern.c 				unplugged = list_entry(entry,
entry             567 arch/um/drivers/mconsole_kern.c 			struct mconsole_output *entry;
entry             569 arch/um/drivers/mconsole_kern.c 			entry = list_entry(ele, struct mconsole_output, list);
entry             570 arch/um/drivers/mconsole_kern.c 			mconsole_reply_len(entry->req, console_buf, n, 0, 1);
entry             591 arch/um/drivers/mconsole_kern.c 	struct mconsole_output entry;
entry             594 arch/um/drivers/mconsole_kern.c 	entry.req = req;
entry             596 arch/um/drivers/mconsole_kern.c 	list_add(&entry.list, &clients);
entry             604 arch/um/drivers/mconsole_kern.c 	list_del(&entry.list);
entry              87 arch/um/include/asm/processor-generic.h extern void start_thread(struct pt_regs *regs, unsigned long entry, 
entry              50 arch/um/include/shared/mem_user.h extern void setup_memory(void *entry);
entry             309 arch/x86/boot/compressed/acpi.c 	u8 *entry;
entry             346 arch/x86/boot/compressed/acpi.c 	entry = (u8 *)(root_table + sizeof(struct acpi_table_header));
entry             350 arch/x86/boot/compressed/acpi.c 			acpi_table = *(u32 *)entry;
entry             352 arch/x86/boot/compressed/acpi.c 			acpi_table = *(u64 *)entry;
entry             360 arch/x86/boot/compressed/acpi.c 		entry += size;
entry             514 arch/x86/boot/compressed/eboot.c 	struct boot_e820_entry *entry = params->e820_table;
entry             587 arch/x86/boot/compressed/eboot.c 			entry = (struct boot_e820_entry *)e820ext->data;
entry             590 arch/x86/boot/compressed/eboot.c 		entry->addr = d->phys_addr;
entry             591 arch/x86/boot/compressed/eboot.c 		entry->size = d->num_pages << PAGE_SHIFT;
entry             592 arch/x86/boot/compressed/eboot.c 		entry->type = e820_type;
entry             593 arch/x86/boot/compressed/eboot.c 		prev = entry++;
entry             576 arch/x86/boot/compressed/kaslr.c static void __process_mem_region(struct mem_vector *entry,
entry             585 arch/x86/boot/compressed/kaslr.c 	if (IS_ENABLED(CONFIG_X86_32) && entry->start >= KERNEL_IMAGE_SIZE)
entry             589 arch/x86/boot/compressed/kaslr.c 	if (entry->start + entry->size < minimum)
entry             593 arch/x86/boot/compressed/kaslr.c 	end = min(entry->size + entry->start, mem_limit);
entry             594 arch/x86/boot/compressed/kaslr.c 	if (entry->start >= end)
entry             596 arch/x86/boot/compressed/kaslr.c 	cur_entry.start = entry->start;
entry             597 arch/x86/boot/compressed/kaslr.c 	cur_entry.size = end - entry->start;
entry             680 arch/x86/boot/compressed/kaslr.c 		struct mem_vector entry;
entry             689 arch/x86/boot/compressed/kaslr.c 		entry.start = clamp(region->start, start, end);
entry             691 arch/x86/boot/compressed/kaslr.c 		entry.size = entry_end - entry.start;
entry             693 arch/x86/boot/compressed/kaslr.c 		__process_mem_region(&entry, minimum, image_size);
entry             787 arch/x86/boot/compressed/kaslr.c 	struct boot_e820_entry *entry;
entry             791 arch/x86/boot/compressed/kaslr.c 		entry = &boot_params->e820_table[i];
entry             793 arch/x86/boot/compressed/kaslr.c 		if (entry->type != E820_TYPE_RAM)
entry             795 arch/x86/boot/compressed/kaslr.c 		region.start = entry->addr;
entry             796 arch/x86/boot/compressed/kaslr.c 		region.size = entry->size;
entry              47 arch/x86/boot/compressed/kaslr_64.c 	unsigned char *entry;
entry              57 arch/x86/boot/compressed/kaslr_64.c 	entry = pages->pgt_buf + pages->pgt_buf_offset;
entry              60 arch/x86/boot/compressed/kaslr_64.c 	return entry;
entry              43 arch/x86/boot/compressed/pgtable_64.c 	struct boot_e820_entry *entry;
entry              77 arch/x86/boot/compressed/pgtable_64.c 		entry = &boot_params->e820_table[i];
entry              80 arch/x86/boot/compressed/pgtable_64.c 		if (bios_start <= entry->addr)
entry              84 arch/x86/boot/compressed/pgtable_64.c 		if (entry->type != E820_TYPE_RAM)
entry              88 arch/x86/boot/compressed/pgtable_64.c 		if (bios_start > entry->addr + entry->size)
entry              89 arch/x86/boot/compressed/pgtable_64.c 			new = entry->addr + entry->size;
entry              95 arch/x86/boot/compressed/pgtable_64.c 		if (new - TRAMPOLINE_32BIT_SIZE < entry->addr)
entry              28 arch/x86/entry/vsyscall/vsyscall_trace.h #define TRACE_INCLUDE_PATH ../../arch/x86/entry/vsyscall/
entry            2366 arch/x86/events/core.c perf_callchain_kernel(struct perf_callchain_entry_ctx *entry, struct pt_regs *regs)
entry            2376 arch/x86/events/core.c 	if (perf_callchain_store(entry, regs->ip))
entry            2386 arch/x86/events/core.c 		if (!addr || perf_callchain_store(entry, addr))
entry            2430 arch/x86/events/core.c perf_callchain_user32(struct pt_regs *regs, struct perf_callchain_entry_ctx *entry)
entry            2445 arch/x86/events/core.c 	while (entry->nr < entry->max_stack) {
entry            2460 arch/x86/events/core.c 		perf_callchain_store(entry, cs_base + frame.return_address);
entry            2468 arch/x86/events/core.c perf_callchain_user32(struct pt_regs *regs, struct perf_callchain_entry_ctx *entry)
entry            2475 arch/x86/events/core.c perf_callchain_user(struct perf_callchain_entry_ctx *entry, struct pt_regs *regs)
entry            2493 arch/x86/events/core.c 	perf_callchain_store(entry, regs->ip);
entry            2498 arch/x86/events/core.c 	if (perf_callchain_user32(regs, entry))
entry            2502 arch/x86/events/core.c 	while (entry->nr < entry->max_stack) {
entry            2518 arch/x86/events/core.c 		perf_callchain_store(entry, frame.return_address);
entry            1274 arch/x86/events/intel/pt.c 	list_for_each_entry(filter, filters, entry) {
entry            1310 arch/x86/events/intel/pt.c 	list_for_each_entry(filter, &head->list, entry) {
entry             145 arch/x86/include/asm/asm.h # define _ASM_NOKPROBE(entry)					\
entry             148 arch/x86/include/asm/asm.h 	_ASM_PTR (entry);					\
entry             126 arch/x86/include/asm/desc.h #define write_ldt_entry(dt, entry, desc)	native_write_ldt_entry(dt, entry, desc)
entry             127 arch/x86/include/asm/desc.h #define write_gdt_entry(dt, entry, desc, type)	native_write_gdt_entry(dt, entry, desc, type)
entry             128 arch/x86/include/asm/desc.h #define write_idt_entry(dt, entry, g)		native_write_idt_entry(dt, entry, g)
entry             141 arch/x86/include/asm/desc.h static inline void native_write_idt_entry(gate_desc *idt, int entry, const gate_desc *gate)
entry             143 arch/x86/include/asm/desc.h 	memcpy(&idt[entry], gate, sizeof(*gate));
entry             146 arch/x86/include/asm/desc.h static inline void native_write_ldt_entry(struct desc_struct *ldt, int entry, const void *desc)
entry             148 arch/x86/include/asm/desc.h 	memcpy(&ldt[entry], desc, 8);
entry             152 arch/x86/include/asm/desc.h native_write_gdt_entry(struct desc_struct *gdt, int entry, const void *desc, int type)
entry             162 arch/x86/include/asm/desc.h 	memcpy(&gdt[entry], desc, size);
entry             184 arch/x86/include/asm/desc.h static inline void __set_tss_desc(unsigned cpu, unsigned int entry, struct x86_hw_tss *addr)
entry             191 arch/x86/include/asm/desc.h 	write_gdt_entry(d, entry, &tss, DESC_TSS);
entry            1114 arch/x86/include/asm/kvm_host.h 	void (*set_supported_cpuid)(u32 func, struct kvm_cpuid_entry2 *entry);
entry            1211 arch/x86/include/asm/kvm_host.h 	int (*get_msr_feature)(struct kvm_msr_entry *entry);
entry             142 arch/x86/include/asm/mce.h 	struct mce entry[MCE_LOG_LEN];
entry             281 arch/x86/include/asm/paravirt.h static inline void write_ldt_entry(struct desc_struct *dt, int entry,
entry             284 arch/x86/include/asm/paravirt.h 	PVOP_VCALL3(cpu.write_ldt_entry, dt, entry, desc);
entry             287 arch/x86/include/asm/paravirt.h static inline void write_gdt_entry(struct desc_struct *dt, int entry,
entry             290 arch/x86/include/asm/paravirt.h 	PVOP_VCALL4(cpu.write_gdt_entry, dt, entry, desc, type);
entry             293 arch/x86/include/asm/paravirt.h static inline void write_idt_entry(gate_desc *dt, int entry, const gate_desc *g)
entry             295 arch/x86/include/asm/paravirt.h 	PVOP_VCALL3(cpu.write_idt_entry, dt, entry, g);
entry            1113 arch/x86/include/asm/pgtable.h 				 pte_t entry, int dirty);
entry            1163 arch/x86/include/asm/pgtable.h 				 pmd_t entry, int dirty);
entry            1166 arch/x86/include/asm/pgtable.h 				 pud_t entry, int dirty);
entry              99 arch/x86/include/asm/uv/bios.h 	} entry[1];		/* additional entries follow */
entry              78 arch/x86/kernel/apic/io_apic.c #define for_each_irq_pin(entry, head) \
entry              79 arch/x86/kernel/apic/io_apic.c 	list_for_each_entry(entry, &head, list)
entry              93 arch/x86/kernel/apic/io_apic.c 	struct IO_APIC_route_entry entry;
entry             304 arch/x86/kernel/apic/io_apic.c 	struct IO_APIC_route_entry entry;
entry             314 arch/x86/kernel/apic/io_apic.c 	return eu.entry;
entry             323 arch/x86/kernel/apic/io_apic.c 	eu.entry = __ioapic_read_entry(apic, pin);
entry             326 arch/x86/kernel/apic/io_apic.c 	return eu.entry;
entry             339 arch/x86/kernel/apic/io_apic.c 	eu.entry = e;
entry             361 arch/x86/kernel/apic/io_apic.c 	union entry_union eu = { .entry.mask = IOAPIC_MASKED };
entry             377 arch/x86/kernel/apic/io_apic.c 	struct irq_pin_list *entry;
entry             380 arch/x86/kernel/apic/io_apic.c 	for_each_irq_pin(entry, data->irq_2_pin)
entry             381 arch/x86/kernel/apic/io_apic.c 		if (entry->apic == apic && entry->pin == pin)
entry             384 arch/x86/kernel/apic/io_apic.c 	entry = kzalloc_node(sizeof(struct irq_pin_list), GFP_ATOMIC, node);
entry             385 arch/x86/kernel/apic/io_apic.c 	if (!entry) {
entry             390 arch/x86/kernel/apic/io_apic.c 	entry->apic = apic;
entry             391 arch/x86/kernel/apic/io_apic.c 	entry->pin = pin;
entry             392 arch/x86/kernel/apic/io_apic.c 	list_add_tail(&entry->list, &data->irq_2_pin);
entry             399 arch/x86/kernel/apic/io_apic.c 	struct irq_pin_list *tmp, *entry;
entry             401 arch/x86/kernel/apic/io_apic.c 	list_for_each_entry_safe(entry, tmp, &data->irq_2_pin, list)
entry             402 arch/x86/kernel/apic/io_apic.c 		if (entry->apic == apic && entry->pin == pin) {
entry             403 arch/x86/kernel/apic/io_apic.c 			list_del(&entry->list);
entry             404 arch/x86/kernel/apic/io_apic.c 			kfree(entry);
entry             423 arch/x86/kernel/apic/io_apic.c 	struct irq_pin_list *entry;
entry             425 arch/x86/kernel/apic/io_apic.c 	for_each_irq_pin(entry, data->irq_2_pin) {
entry             426 arch/x86/kernel/apic/io_apic.c 		if (entry->apic == oldapic && entry->pin == oldpin) {
entry             427 arch/x86/kernel/apic/io_apic.c 			entry->apic = newapic;
entry             428 arch/x86/kernel/apic/io_apic.c 			entry->pin = newpin;
entry             440 arch/x86/kernel/apic/io_apic.c 			       void (*final)(struct irq_pin_list *entry))
entry             443 arch/x86/kernel/apic/io_apic.c 	struct irq_pin_list *entry;
entry             445 arch/x86/kernel/apic/io_apic.c 	eu.entry = data->entry;
entry             448 arch/x86/kernel/apic/io_apic.c 	data->entry = eu.entry;
entry             450 arch/x86/kernel/apic/io_apic.c 	for_each_irq_pin(entry, data->irq_2_pin) {
entry             451 arch/x86/kernel/apic/io_apic.c 		io_apic_write(entry->apic, 0x10 + 2 * entry->pin, eu.w1);
entry             453 arch/x86/kernel/apic/io_apic.c 			final(entry);
entry             457 arch/x86/kernel/apic/io_apic.c static void io_apic_sync(struct irq_pin_list *entry)
entry             465 arch/x86/kernel/apic/io_apic.c 	io_apic = io_apic_base(entry->apic);
entry             515 arch/x86/kernel/apic/io_apic.c 		struct IO_APIC_route_entry entry, entry1;
entry             517 arch/x86/kernel/apic/io_apic.c 		entry = entry1 = __ioapic_read_entry(apic, pin);
entry             530 arch/x86/kernel/apic/io_apic.c 		__ioapic_write_entry(apic, pin, entry);
entry             537 arch/x86/kernel/apic/io_apic.c 	struct irq_pin_list *entry;
entry             540 arch/x86/kernel/apic/io_apic.c 	for_each_irq_pin(entry, data->irq_2_pin)
entry             541 arch/x86/kernel/apic/io_apic.c 		__eoi_ioapic_pin(entry->apic, entry->pin, vector);
entry             547 arch/x86/kernel/apic/io_apic.c 	struct IO_APIC_route_entry entry;
entry             550 arch/x86/kernel/apic/io_apic.c 	entry = ioapic_read_entry(apic, pin);
entry             551 arch/x86/kernel/apic/io_apic.c 	if (entry.delivery_mode == dest_SMI)
entry             558 arch/x86/kernel/apic/io_apic.c 	if (entry.mask == IOAPIC_UNMASKED) {
entry             559 arch/x86/kernel/apic/io_apic.c 		entry.mask = IOAPIC_MASKED;
entry             560 arch/x86/kernel/apic/io_apic.c 		ioapic_write_entry(apic, pin, entry);
entry             561 arch/x86/kernel/apic/io_apic.c 		entry = ioapic_read_entry(apic, pin);
entry             564 arch/x86/kernel/apic/io_apic.c 	if (entry.irr) {
entry             572 arch/x86/kernel/apic/io_apic.c 		if (entry.trigger == IOAPIC_EDGE) {
entry             573 arch/x86/kernel/apic/io_apic.c 			entry.trigger = IOAPIC_LEVEL;
entry             574 arch/x86/kernel/apic/io_apic.c 			ioapic_write_entry(apic, pin, entry);
entry             577 arch/x86/kernel/apic/io_apic.c 		__eoi_ioapic_pin(apic, pin, entry.vector);
entry             586 arch/x86/kernel/apic/io_apic.c 	entry = ioapic_read_entry(apic, pin);
entry             587 arch/x86/kernel/apic/io_apic.c 	if (entry.irr)
entry             672 arch/x86/kernel/apic/io_apic.c 			struct IO_APIC_route_entry entry;
entry             674 arch/x86/kernel/apic/io_apic.c 			entry = ioapics[apic].saved_registers[pin];
entry             675 arch/x86/kernel/apic/io_apic.c 			if (entry.mask == IOAPIC_UNMASKED) {
entry             676 arch/x86/kernel/apic/io_apic.c 				entry.mask = IOAPIC_MASKED;
entry             677 arch/x86/kernel/apic/io_apic.c 				ioapic_write_entry(apic, pin, entry);
entry             951 arch/x86/kernel/apic/io_apic.c 		data->entry.trigger = data->trigger = info->ioapic_trigger;
entry             952 arch/x86/kernel/apic/io_apic.c 		data->entry.polarity = data->polarity = info->ioapic_polarity;
entry            1237 arch/x86/kernel/apic/io_apic.c 	struct IO_APIC_route_entry entry;
entry            1238 arch/x86/kernel/apic/io_apic.c 	struct IR_IO_APIC_route_entry *ir_entry = (void *)&entry;
entry            1242 arch/x86/kernel/apic/io_apic.c 		entry = ioapic_read_entry(apic, i);
entry            1246 arch/x86/kernel/apic/io_apic.c 			 entry.mask == IOAPIC_MASKED ? "disabled" : "enabled ",
entry            1247 arch/x86/kernel/apic/io_apic.c 			 entry.trigger == IOAPIC_LEVEL ? "level" : "edge ",
entry            1248 arch/x86/kernel/apic/io_apic.c 			 entry.polarity == IOAPIC_POL_LOW ? "low " : "high",
entry            1249 arch/x86/kernel/apic/io_apic.c 			 entry.vector, entry.irr, entry.delivery_status);
entry            1257 arch/x86/kernel/apic/io_apic.c 			       entry.dest_mode == IOAPIC_DEST_MODE_LOGICAL ?
entry            1259 arch/x86/kernel/apic/io_apic.c 			       entry.dest, entry.delivery_mode);
entry            1341 arch/x86/kernel/apic/io_apic.c 		struct irq_pin_list *entry;
entry            1355 arch/x86/kernel/apic/io_apic.c 		for_each_irq_pin(entry, data->irq_2_pin)
entry            1356 arch/x86/kernel/apic/io_apic.c 			pr_cont("-> %d:%d", entry->apic, entry->pin);
entry            1379 arch/x86/kernel/apic/io_apic.c 		struct IO_APIC_route_entry entry = ioapic_read_entry(apic, pin);
entry            1384 arch/x86/kernel/apic/io_apic.c 		if ((entry.mask == 0) && (entry.delivery_mode == dest_ExtINT)) {
entry            1425 arch/x86/kernel/apic/io_apic.c 		struct IO_APIC_route_entry entry;
entry            1427 arch/x86/kernel/apic/io_apic.c 		memset(&entry, 0, sizeof(entry));
entry            1428 arch/x86/kernel/apic/io_apic.c 		entry.mask		= IOAPIC_UNMASKED;
entry            1429 arch/x86/kernel/apic/io_apic.c 		entry.trigger		= IOAPIC_EDGE;
entry            1430 arch/x86/kernel/apic/io_apic.c 		entry.polarity		= IOAPIC_POL_HIGH;
entry            1431 arch/x86/kernel/apic/io_apic.c 		entry.dest_mode		= IOAPIC_DEST_MODE_PHYSICAL;
entry            1432 arch/x86/kernel/apic/io_apic.c 		entry.delivery_mode	= dest_ExtINT;
entry            1433 arch/x86/kernel/apic/io_apic.c 		entry.dest		= read_apic_id();
entry            1438 arch/x86/kernel/apic/io_apic.c 		ioapic_write_entry(ioapic_i8259.apic, ioapic_i8259.pin, entry);
entry            1707 arch/x86/kernel/apic/io_apic.c 	struct irq_pin_list *entry;
entry            1711 arch/x86/kernel/apic/io_apic.c 	for_each_irq_pin(entry, data->irq_2_pin) {
entry            1715 arch/x86/kernel/apic/io_apic.c 		pin = entry->pin;
entry            1716 arch/x86/kernel/apic/io_apic.c 		reg = io_apic_read(entry->apic, 0x10 + pin*2);
entry            1862 arch/x86/kernel/apic/io_apic.c 	eoi_ioapic_pin(data->entry.vector, data);
entry            1869 arch/x86/kernel/apic/io_apic.c 	struct irq_pin_list *entry;
entry            1877 arch/x86/kernel/apic/io_apic.c 		mpd->entry.dest = cfg->dest_apicid;
entry            1878 arch/x86/kernel/apic/io_apic.c 		mpd->entry.vector = cfg->vector;
entry            1880 arch/x86/kernel/apic/io_apic.c 	for_each_irq_pin(entry, mpd->irq_2_pin)
entry            1881 arch/x86/kernel/apic/io_apic.c 		__ioapic_write_entry(entry->apic, entry->pin, mpd->entry);
entry            2955 arch/x86/kernel/apic/io_apic.c 			   struct IO_APIC_route_entry *entry)
entry            2957 arch/x86/kernel/apic/io_apic.c 	memset(entry, 0, sizeof(*entry));
entry            2958 arch/x86/kernel/apic/io_apic.c 	entry->delivery_mode = apic->irq_delivery_mode;
entry            2959 arch/x86/kernel/apic/io_apic.c 	entry->dest_mode     = apic->irq_dest_mode;
entry            2960 arch/x86/kernel/apic/io_apic.c 	entry->dest	     = cfg->dest_apicid;
entry            2961 arch/x86/kernel/apic/io_apic.c 	entry->vector	     = cfg->vector;
entry            2962 arch/x86/kernel/apic/io_apic.c 	entry->trigger	     = data->trigger;
entry            2963 arch/x86/kernel/apic/io_apic.c 	entry->polarity	     = data->polarity;
entry            2969 arch/x86/kernel/apic/io_apic.c 		entry->mask = IOAPIC_MASKED;
entry            2971 arch/x86/kernel/apic/io_apic.c 		entry->mask = IOAPIC_UNMASKED;
entry            2999 arch/x86/kernel/apic/io_apic.c 	info->ioapic_entry = &data->entry;
entry            1272 arch/x86/kernel/apic/x2apic_uv_x.c 	for (i = 0; st->entry[i].type != UV_SYSTAB_TYPE_UNUSED; i++) {
entry            1273 arch/x86/kernel/apic/x2apic_uv_x.c 		unsigned long ptr = st->entry[i].offset;
entry            1280 arch/x86/kernel/apic/x2apic_uv_x.c 		switch (st->entry[i].type) {
entry              55 arch/x86/kernel/asm-offsets_64.c #define ENTRY(entry) OFFSET(pt_regs_ ## entry, pt_regs, entry)
entry              75 arch/x86/kernel/asm-offsets_64.c #define ENTRY(entry) OFFSET(saved_context_ ## entry, saved_context, entry)
entry             475 arch/x86/kernel/cpu/mce/core.c static void mce_irq_work_cb(struct irq_work *entry)
entry              44 arch/x86/kernel/cpu/mce/dev-mcelog.c 	unsigned int entry;
entry              48 arch/x86/kernel/cpu/mce/dev-mcelog.c 	entry = mcelog.next;
entry              54 arch/x86/kernel/cpu/mce/dev-mcelog.c 	if (entry >= MCE_LOG_LEN) {
entry              59 arch/x86/kernel/cpu/mce/dev-mcelog.c 	mcelog.next = entry + 1;
entry              61 arch/x86/kernel/cpu/mce/dev-mcelog.c 	memcpy(mcelog.entry + entry, mce, sizeof(struct mce));
entry              62 arch/x86/kernel/cpu/mce/dev-mcelog.c 	mcelog.entry[entry].finished = 1;
entry             224 arch/x86/kernel/cpu/mce/dev-mcelog.c 		struct mce *m = &mcelog.entry[i];
entry             230 arch/x86/kernel/cpu/mce/dev-mcelog.c 	memset(mcelog.entry, 0, next * sizeof(struct mce));
entry              41 arch/x86/kernel/cpu/microcode/amd.c 	struct equiv_cpu_entry *entry;
entry              75 arch/x86/kernel/cpu/microcode/amd.c 		struct equiv_cpu_entry *e = &et->entry[i];
entry             309 arch/x86/kernel/cpu/microcode/amd.c 	table.entry = (struct equiv_cpu_entry *)(buf + CONTAINER_HDR_SZ);
entry             730 arch/x86/kernel/cpu/microcode/amd.c 	equiv_table.entry = vmalloc(equiv_tbl_len);
entry             731 arch/x86/kernel/cpu/microcode/amd.c 	if (!equiv_table.entry) {
entry             736 arch/x86/kernel/cpu/microcode/amd.c 	memcpy(equiv_table.entry, buf + CONTAINER_HDR_SZ, equiv_tbl_len);
entry             745 arch/x86/kernel/cpu/microcode/amd.c 	vfree(equiv_table.entry);
entry              69 arch/x86/kernel/cpu/resctrl/monitor.c 	struct rmid_entry *entry;
entry              71 arch/x86/kernel/cpu/resctrl/monitor.c 	entry = &rmid_ptrs[rmid];
entry              72 arch/x86/kernel/cpu/resctrl/monitor.c 	WARN_ON(entry->rmid != rmid);
entry              74 arch/x86/kernel/cpu/resctrl/monitor.c 	return entry;
entry              95 arch/x86/kernel/cpu/resctrl/monitor.c static bool rmid_dirty(struct rmid_entry *entry)
entry              97 arch/x86/kernel/cpu/resctrl/monitor.c 	u64 val = __rmid_read(entry->rmid, QOS_L3_OCCUP_EVENT_ID);
entry             110 arch/x86/kernel/cpu/resctrl/monitor.c 	struct rmid_entry *entry;
entry             127 arch/x86/kernel/cpu/resctrl/monitor.c 		entry = __rmid_entry(nrmid);
entry             128 arch/x86/kernel/cpu/resctrl/monitor.c 		if (force_free || !rmid_dirty(entry)) {
entry             129 arch/x86/kernel/cpu/resctrl/monitor.c 			clear_bit(entry->rmid, d->rmid_busy_llc);
entry             130 arch/x86/kernel/cpu/resctrl/monitor.c 			if (!--entry->busy) {
entry             132 arch/x86/kernel/cpu/resctrl/monitor.c 				list_add_tail(&entry->list, &rmid_free_lru);
entry             151 arch/x86/kernel/cpu/resctrl/monitor.c 	struct rmid_entry *entry;
entry             158 arch/x86/kernel/cpu/resctrl/monitor.c 	entry = list_first_entry(&rmid_free_lru,
entry             160 arch/x86/kernel/cpu/resctrl/monitor.c 	list_del(&entry->list);
entry             162 arch/x86/kernel/cpu/resctrl/monitor.c 	return entry->rmid;
entry             165 arch/x86/kernel/cpu/resctrl/monitor.c static void add_rmid_to_limbo(struct rmid_entry *entry)
entry             174 arch/x86/kernel/cpu/resctrl/monitor.c 	entry->busy = 0;
entry             178 arch/x86/kernel/cpu/resctrl/monitor.c 			val = __rmid_read(entry->rmid, QOS_L3_OCCUP_EVENT_ID);
entry             189 arch/x86/kernel/cpu/resctrl/monitor.c 		set_bit(entry->rmid, d->rmid_busy_llc);
entry             190 arch/x86/kernel/cpu/resctrl/monitor.c 		entry->busy++;
entry             194 arch/x86/kernel/cpu/resctrl/monitor.c 	if (entry->busy)
entry             197 arch/x86/kernel/cpu/resctrl/monitor.c 		list_add_tail(&entry->list, &rmid_free_lru);
entry             202 arch/x86/kernel/cpu/resctrl/monitor.c 	struct rmid_entry *entry;
entry             209 arch/x86/kernel/cpu/resctrl/monitor.c 	entry = __rmid_entry(rmid);
entry             212 arch/x86/kernel/cpu/resctrl/monitor.c 		add_rmid_to_limbo(entry);
entry             214 arch/x86/kernel/cpu/resctrl/monitor.c 		list_add_tail(&entry->list, &rmid_free_lru);
entry             299 arch/x86/kernel/cpu/resctrl/monitor.c 	struct rdtgroup *rdtgrp, *entry;
entry             314 arch/x86/kernel/cpu/resctrl/monitor.c 		list_for_each_entry(entry, head, mon.crdtgrp_list) {
entry             315 arch/x86/kernel/cpu/resctrl/monitor.c 			if (__mon_event_count(entry->mon.rmid, rr))
entry             361 arch/x86/kernel/cpu/resctrl/monitor.c 	struct rdtgroup *entry;
entry             386 arch/x86/kernel/cpu/resctrl/monitor.c 	list_for_each_entry(entry, head, mon.crdtgrp_list) {
entry             387 arch/x86/kernel/cpu/resctrl/monitor.c 		cmbm_data = &dom_mbm->mbm_local[entry->mon.rmid];
entry             430 arch/x86/kernel/cpu/resctrl/monitor.c 	list_for_each_entry(entry, head, mon.crdtgrp_list) {
entry             431 arch/x86/kernel/cpu/resctrl/monitor.c 		cmbm_data = &dom_mbm->mbm_local[entry->mon.rmid];
entry             555 arch/x86/kernel/cpu/resctrl/monitor.c 	struct rmid_entry *entry = NULL;
entry             564 arch/x86/kernel/cpu/resctrl/monitor.c 		entry = &rmid_ptrs[i];
entry             565 arch/x86/kernel/cpu/resctrl/monitor.c 		INIT_LIST_HEAD(&entry->list);
entry             567 arch/x86/kernel/cpu/resctrl/monitor.c 		entry->rmid = i;
entry             568 arch/x86/kernel/cpu/resctrl/monitor.c 		list_add_tail(&entry->list, &rmid_free_lru);
entry             575 arch/x86/kernel/cpu/resctrl/monitor.c 	entry = __rmid_entry(0);
entry             576 arch/x86/kernel/cpu/resctrl/monitor.c 	list_del(&entry->list);
entry             291 arch/x86/kernel/crash.c static int add_e820_entry(struct boot_params *params, struct e820_entry *entry)
entry             299 arch/x86/kernel/crash.c 	memcpy(&params->e820_table[nr_e820_entries], entry,
entry              83 arch/x86/kernel/e820.c 		struct e820_entry *entry = &table->entries[i];
entry              85 arch/x86/kernel/e820.c 		if (type && entry->type != type)
entry              87 arch/x86/kernel/e820.c 		if (entry->addr >= end || entry->addr + entry->size <= start)
entry             118 arch/x86/kernel/e820.c 		struct e820_entry *entry = &e820_table->entries[i];
entry             120 arch/x86/kernel/e820.c 		if (type && entry->type != type)
entry             124 arch/x86/kernel/e820.c 		if (entry->addr >= end || entry->addr + entry->size <= start)
entry             131 arch/x86/kernel/e820.c 		if (entry->addr <= start)
entry             132 arch/x86/kernel/e820.c 			start = entry->addr + entry->size;
entry             139 arch/x86/kernel/e820.c 			return entry;
entry             158 arch/x86/kernel/e820.c 	struct e820_entry *entry = __e820__mapped_all(start, end, 0);
entry             160 arch/x86/kernel/e820.c 	return entry ? entry->type : -EINVAL;
entry             280 arch/x86/kernel/e820.c 	struct e820_entry	*entry;
entry             304 arch/x86/kernel/e820.c 	return (ap->addr != ap->entry->addr) - (bp->addr != bp->entry->addr);
entry             340 arch/x86/kernel/e820.c 			change_point[chg_idx++]->entry	= &entries[i];
entry             342 arch/x86/kernel/e820.c 			change_point[chg_idx++]->entry	= &entries[i];
entry             359 arch/x86/kernel/e820.c 		if (change_point[chg_idx]->addr == change_point[chg_idx]->entry->addr) {
entry             361 arch/x86/kernel/e820.c 			overlap_list[overlap_entries++] = change_point[chg_idx]->entry;
entry             365 arch/x86/kernel/e820.c 				if (overlap_list[i] == change_point[chg_idx]->entry)
entry             409 arch/x86/kernel/e820.c 	struct boot_e820_entry *entry = entries;
entry             412 arch/x86/kernel/e820.c 		u64 start = entry->addr;
entry             413 arch/x86/kernel/e820.c 		u64 size = entry->size;
entry             415 arch/x86/kernel/e820.c 		u32 type = entry->type;
entry             423 arch/x86/kernel/e820.c 		entry++;
entry             467 arch/x86/kernel/e820.c 		struct e820_entry *entry = &table->entries[i];
entry             471 arch/x86/kernel/e820.c 		if (entry->type != old_type)
entry             474 arch/x86/kernel/e820.c 		entry_end = entry->addr + entry->size;
entry             477 arch/x86/kernel/e820.c 		if (entry->addr >= start && entry_end <= end) {
entry             478 arch/x86/kernel/e820.c 			entry->type = new_type;
entry             479 arch/x86/kernel/e820.c 			real_updated_size += entry->size;
entry             484 arch/x86/kernel/e820.c 		if (entry->addr < start && entry_end > end) {
entry             486 arch/x86/kernel/e820.c 			__e820__range_add(table, end, entry_end - end, entry->type);
entry             487 arch/x86/kernel/e820.c 			entry->size = start - entry->addr;
entry             493 arch/x86/kernel/e820.c 		final_start = max(start, entry->addr);
entry             506 arch/x86/kernel/e820.c 		entry->size -= final_end - final_start;
entry             507 arch/x86/kernel/e820.c 		if (entry->addr < final_start)
entry             510 arch/x86/kernel/e820.c 		entry->addr = final_end;
entry             542 arch/x86/kernel/e820.c 		struct e820_entry *entry = &e820_table->entries[i];
entry             546 arch/x86/kernel/e820.c 		if (check_type && entry->type != old_type)
entry             549 arch/x86/kernel/e820.c 		entry_end = entry->addr + entry->size;
entry             552 arch/x86/kernel/e820.c 		if (entry->addr >= start && entry_end <= end) {
entry             553 arch/x86/kernel/e820.c 			real_removed_size += entry->size;
entry             554 arch/x86/kernel/e820.c 			memset(entry, 0, sizeof(*entry));
entry             559 arch/x86/kernel/e820.c 		if (entry->addr < start && entry_end > end) {
entry             560 arch/x86/kernel/e820.c 			e820__range_add(end, entry_end - end, entry->type);
entry             561 arch/x86/kernel/e820.c 			entry->size = start - entry->addr;
entry             567 arch/x86/kernel/e820.c 		final_start = max(start, entry->addr);
entry             578 arch/x86/kernel/e820.c 		entry->size -= final_end - final_start;
entry             579 arch/x86/kernel/e820.c 		if (entry->addr < final_start)
entry             582 arch/x86/kernel/e820.c 		entry->addr = final_end;
entry             744 arch/x86/kernel/e820.c 		struct e820_entry *entry = &e820_table->entries[i];
entry             746 arch/x86/kernel/e820.c 		if (pfn < PFN_UP(entry->addr))
entry             747 arch/x86/kernel/e820.c 			register_nosave_region(pfn, PFN_UP(entry->addr));
entry             749 arch/x86/kernel/e820.c 		pfn = PFN_DOWN(entry->addr + entry->size);
entry             751 arch/x86/kernel/e820.c 		if (entry->type != E820_TYPE_RAM && entry->type != E820_TYPE_RESERVED_KERN)
entry             752 arch/x86/kernel/e820.c 			register_nosave_region(PFN_UP(entry->addr), pfn);
entry             769 arch/x86/kernel/e820.c 		struct e820_entry *entry = &e820_table->entries[i];
entry             771 arch/x86/kernel/e820.c 		if (entry->type == E820_TYPE_NVS)
entry             772 arch/x86/kernel/e820.c 			acpi_nvs_register(entry->addr, entry->size);
entry             822 arch/x86/kernel/e820.c 		struct e820_entry *entry = &e820_table->entries[i];
entry             826 arch/x86/kernel/e820.c 		if (entry->type != type)
entry             829 arch/x86/kernel/e820.c 		start_pfn = entry->addr >> PAGE_SHIFT;
entry             830 arch/x86/kernel/e820.c 		end_pfn = (entry->addr + entry->size) >> PAGE_SHIFT;
entry            1029 arch/x86/kernel/e820.c static const char *__init e820_type_to_string(struct e820_entry *entry)
entry            1031 arch/x86/kernel/e820.c 	switch (entry->type) {
entry            1044 arch/x86/kernel/e820.c static unsigned long __init e820_type_to_iomem_type(struct e820_entry *entry)
entry            1046 arch/x86/kernel/e820.c 	switch (entry->type) {
entry            1059 arch/x86/kernel/e820.c static unsigned long __init e820_type_to_iores_desc(struct e820_entry *entry)
entry            1061 arch/x86/kernel/e820.c 	switch (entry->type) {
entry            1119 arch/x86/kernel/e820.c 		struct e820_entry *entry = e820_table->entries + i;
entry            1121 arch/x86/kernel/e820.c 		end = entry->addr + entry->size - 1;
entry            1126 arch/x86/kernel/e820.c 		res->start = entry->addr;
entry            1128 arch/x86/kernel/e820.c 		res->name  = e820_type_to_string(entry);
entry            1129 arch/x86/kernel/e820.c 		res->flags = e820_type_to_iomem_type(entry);
entry            1130 arch/x86/kernel/e820.c 		res->desc  = e820_type_to_iores_desc(entry);
entry            1137 arch/x86/kernel/e820.c 		if (do_mark_busy(entry->type, res)) {
entry            1146 arch/x86/kernel/e820.c 		struct e820_entry *entry = e820_table_firmware->entries + i;
entry            1148 arch/x86/kernel/e820.c 		firmware_map_add_early(entry->addr, entry->addr + entry->size, e820_type_to_string(entry));
entry            1190 arch/x86/kernel/e820.c 		struct e820_entry *entry = &e820_table->entries[i];
entry            1193 arch/x86/kernel/e820.c 		if (entry->type != E820_TYPE_RAM)
entry            1196 arch/x86/kernel/e820.c 		start = entry->addr + entry->size;
entry            1282 arch/x86/kernel/e820.c 		struct e820_entry *entry = &e820_table->entries[i];
entry            1284 arch/x86/kernel/e820.c 		end = entry->addr + entry->size;
entry            1288 arch/x86/kernel/e820.c 		if (entry->type != E820_TYPE_RAM && entry->type != E820_TYPE_RESERVED_KERN)
entry            1291 arch/x86/kernel/e820.c 		memblock_add(entry->addr, entry->size);
entry             310 arch/x86/kernel/idt.c 	void *entry;
entry             315 arch/x86/kernel/idt.c 		entry = irq_entries_start + 8 * (i - FIRST_EXTERNAL_VECTOR);
entry             316 arch/x86/kernel/idt.c 		set_intr_gate(i, entry);
entry             322 arch/x86/kernel/idt.c 		entry = spurious_entries_start + 8 * (i - FIRST_SYSTEM_VECTOR);
entry             323 arch/x86/kernel/idt.c 		set_intr_gate(i, entry);
entry              38 arch/x86/kernel/jump_label.c static void __jump_label_set_jump_code(struct jump_entry *entry,
entry              49 arch/x86/kernel/jump_label.c 	code->offset = jump_entry_target(entry) -
entry              50 arch/x86/kernel/jump_label.c 		       (jump_entry_code(entry) + JUMP_LABEL_NOP_SIZE);
entry              60 arch/x86/kernel/jump_label.c 	if (memcmp((void *)jump_entry_code(entry), expect, JUMP_LABEL_NOP_SIZE))
entry              61 arch/x86/kernel/jump_label.c 		bug_at((void *)jump_entry_code(entry), line);
entry              67 arch/x86/kernel/jump_label.c static void __ref __jump_label_transform(struct jump_entry *entry,
entry              73 arch/x86/kernel/jump_label.c 	__jump_label_set_jump_code(entry, type, &code, init);
entry              87 arch/x86/kernel/jump_label.c 		text_poke_early((void *)jump_entry_code(entry), &code,
entry              92 arch/x86/kernel/jump_label.c 	text_poke_bp((void *)jump_entry_code(entry), &code, JUMP_LABEL_NOP_SIZE,
entry              93 arch/x86/kernel/jump_label.c 		     (void *)jump_entry_code(entry) + JUMP_LABEL_NOP_SIZE);
entry              96 arch/x86/kernel/jump_label.c void arch_jump_label_transform(struct jump_entry *entry,
entry             100 arch/x86/kernel/jump_label.c 	__jump_label_transform(entry, type, 0);
entry             108 arch/x86/kernel/jump_label.c bool arch_jump_label_transform_queue(struct jump_entry *entry,
entry             118 arch/x86/kernel/jump_label.c 		arch_jump_label_transform(entry, type);
entry             131 arch/x86/kernel/jump_label.c 	entry_code = (void *)jump_entry_code(entry);
entry             149 arch/x86/kernel/jump_label.c 	__jump_label_set_jump_code(entry, type,
entry             179 arch/x86/kernel/jump_label.c __init_or_module void arch_jump_label_transform_static(struct jump_entry *entry,
entry             199 arch/x86/kernel/jump_label.c 		__jump_label_transform(entry, type, 1);
entry             257 arch/x86/kernel/pci-calgary_64.c 	unsigned long entry;
entry             260 arch/x86/kernel/pci-calgary_64.c 	entry = iommu_range_alloc(dev, tbl, npages);
entry             261 arch/x86/kernel/pci-calgary_64.c 	if (unlikely(entry == DMA_MAPPING_ERROR)) {
entry             268 arch/x86/kernel/pci-calgary_64.c 	ret = (entry << PAGE_SHIFT) | ((unsigned long)vaddr & ~PAGE_MASK);
entry             271 arch/x86/kernel/pci-calgary_64.c 	tce_build(tbl, entry, npages, (unsigned long)vaddr & PAGE_MASK,
entry             279 arch/x86/kernel/pci-calgary_64.c 	unsigned long entry;
entry             289 arch/x86/kernel/pci-calgary_64.c 	entry = dma_addr >> PAGE_SHIFT;
entry             291 arch/x86/kernel/pci-calgary_64.c 	BUG_ON(entry + npages > tbl->it_size);
entry             293 arch/x86/kernel/pci-calgary_64.c 	tce_free(tbl, entry, npages);
entry             297 arch/x86/kernel/pci-calgary_64.c 	bitmap_clear(tbl->it_map, entry, npages);
entry             357 arch/x86/kernel/pci-calgary_64.c 	unsigned long entry;
entry             366 arch/x86/kernel/pci-calgary_64.c 		entry = iommu_range_alloc(dev, tbl, npages);
entry             367 arch/x86/kernel/pci-calgary_64.c 		if (entry == DMA_MAPPING_ERROR) {
entry             373 arch/x86/kernel/pci-calgary_64.c 		s->dma_address = (entry << PAGE_SHIFT) | s->offset;
entry             376 arch/x86/kernel/pci-calgary_64.c 		tce_build(tbl, entry, npages, vaddr & PAGE_MASK, dir);
entry              29 arch/x86/kernel/resource.c 	struct e820_entry *entry;
entry              32 arch/x86/kernel/resource.c 		entry = &e820_table->entries[i];
entry              34 arch/x86/kernel/resource.c 		resource_clip(avail, entry->addr,
entry              35 arch/x86/kernel/resource.c 			      entry->addr + entry->size - 1);
entry             166 arch/x86/kvm/cpuid.c 	struct kvm_cpuid_entry2 *e, *entry;
entry             168 arch/x86/kvm/cpuid.c 	entry = NULL;
entry             172 arch/x86/kvm/cpuid.c 			entry = e;
entry             176 arch/x86/kvm/cpuid.c 	if (entry && (entry->edx & F(NX)) && !is_efer_nx()) {
entry             177 arch/x86/kvm/cpuid.c 		entry->edx &= ~F(NX);
entry             289 arch/x86/kvm/cpuid.c static void do_host_cpuid(struct kvm_cpuid_entry2 *entry, u32 function,
entry             292 arch/x86/kvm/cpuid.c 	entry->function = function;
entry             293 arch/x86/kvm/cpuid.c 	entry->index = index;
entry             294 arch/x86/kvm/cpuid.c 	entry->flags = 0;
entry             296 arch/x86/kvm/cpuid.c 	cpuid_count(entry->function, entry->index,
entry             297 arch/x86/kvm/cpuid.c 		    &entry->eax, &entry->ebx, &entry->ecx, &entry->edx);
entry             301 arch/x86/kvm/cpuid.c 		entry->flags |= KVM_CPUID_FLAG_STATEFUL_FUNC;
entry             315 arch/x86/kvm/cpuid.c 		entry->flags |= KVM_CPUID_FLAG_SIGNIFCANT_INDEX;
entry             320 arch/x86/kvm/cpuid.c static int __do_cpuid_func_emulated(struct kvm_cpuid_entry2 *entry,
entry             323 arch/x86/kvm/cpuid.c 	entry->function = func;
entry             324 arch/x86/kvm/cpuid.c 	entry->index = 0;
entry             325 arch/x86/kvm/cpuid.c 	entry->flags = 0;
entry             329 arch/x86/kvm/cpuid.c 		entry->eax = 7;
entry             333 arch/x86/kvm/cpuid.c 		entry->ecx = F(MOVBE);
entry             337 arch/x86/kvm/cpuid.c 		entry->flags |= KVM_CPUID_FLAG_SIGNIFCANT_INDEX;
entry             338 arch/x86/kvm/cpuid.c 		entry->eax = 0;
entry             339 arch/x86/kvm/cpuid.c 		entry->ecx = F(RDPID);
entry             348 arch/x86/kvm/cpuid.c static inline void do_cpuid_7_mask(struct kvm_cpuid_entry2 *entry, int index)
entry             384 arch/x86/kvm/cpuid.c 		entry->eax = min(entry->eax, 1u);
entry             385 arch/x86/kvm/cpuid.c 		entry->ebx &= kvm_cpuid_7_0_ebx_x86_features;
entry             386 arch/x86/kvm/cpuid.c 		cpuid_mask(&entry->ebx, CPUID_7_0_EBX);
entry             388 arch/x86/kvm/cpuid.c 		entry->ebx |= F(TSC_ADJUST);
entry             390 arch/x86/kvm/cpuid.c 		entry->ecx &= kvm_cpuid_7_0_ecx_x86_features;
entry             391 arch/x86/kvm/cpuid.c 		f_la57 = entry->ecx & F(LA57);
entry             392 arch/x86/kvm/cpuid.c 		cpuid_mask(&entry->ecx, CPUID_7_ECX);
entry             394 arch/x86/kvm/cpuid.c 		entry->ecx |= f_la57;
entry             395 arch/x86/kvm/cpuid.c 		entry->ecx |= f_umip;
entry             396 arch/x86/kvm/cpuid.c 		entry->ecx |= f_pku;
entry             399 arch/x86/kvm/cpuid.c 			entry->ecx &= ~F(PKU);
entry             401 arch/x86/kvm/cpuid.c 		entry->edx &= kvm_cpuid_7_0_edx_x86_features;
entry             402 arch/x86/kvm/cpuid.c 		cpuid_mask(&entry->edx, CPUID_7_EDX);
entry             404 arch/x86/kvm/cpuid.c 			entry->edx |= F(SPEC_CTRL);
entry             406 arch/x86/kvm/cpuid.c 			entry->edx |= F(INTEL_STIBP);
entry             409 arch/x86/kvm/cpuid.c 			entry->edx |= F(SPEC_CTRL_SSBD);
entry             414 arch/x86/kvm/cpuid.c 		entry->edx |= F(ARCH_CAPABILITIES);
entry             417 arch/x86/kvm/cpuid.c 		entry->eax &= kvm_cpuid_7_1_eax_x86_features;
entry             418 arch/x86/kvm/cpuid.c 		entry->ebx = 0;
entry             419 arch/x86/kvm/cpuid.c 		entry->ecx = 0;
entry             420 arch/x86/kvm/cpuid.c 		entry->edx = 0;
entry             424 arch/x86/kvm/cpuid.c 		entry->eax = 0;
entry             425 arch/x86/kvm/cpuid.c 		entry->ebx = 0;
entry             426 arch/x86/kvm/cpuid.c 		entry->ecx = 0;
entry             427 arch/x86/kvm/cpuid.c 		entry->edx = 0;
entry             432 arch/x86/kvm/cpuid.c static inline int __do_cpuid_func(struct kvm_cpuid_entry2 *entry, u32 function,
entry             513 arch/x86/kvm/cpuid.c 	do_host_cpuid(entry, function, 0);
entry             519 arch/x86/kvm/cpuid.c 		entry->eax = min(entry->eax, 0x1fU);
entry             522 arch/x86/kvm/cpuid.c 		entry->edx &= kvm_cpuid_1_edx_x86_features;
entry             523 arch/x86/kvm/cpuid.c 		cpuid_mask(&entry->edx, CPUID_1_EDX);
entry             524 arch/x86/kvm/cpuid.c 		entry->ecx &= kvm_cpuid_1_ecx_x86_features;
entry             525 arch/x86/kvm/cpuid.c 		cpuid_mask(&entry->ecx, CPUID_1_ECX);
entry             528 arch/x86/kvm/cpuid.c 		entry->ecx |= F(X2APIC);
entry             535 arch/x86/kvm/cpuid.c 		int t, times = entry->eax & 0xff;
entry             537 arch/x86/kvm/cpuid.c 		entry->flags |= KVM_CPUID_FLAG_STATE_READ_NEXT;
entry             542 arch/x86/kvm/cpuid.c 			do_host_cpuid(&entry[t], function, 0);
entry             557 arch/x86/kvm/cpuid.c 			cache_type = entry[i - 1].eax & 0x1f;
entry             560 arch/x86/kvm/cpuid.c 			do_host_cpuid(&entry[i], function, i);
entry             566 arch/x86/kvm/cpuid.c 		entry->eax = 0x4; /* allow ARAT */
entry             567 arch/x86/kvm/cpuid.c 		entry->ebx = 0;
entry             568 arch/x86/kvm/cpuid.c 		entry->ecx = 0;
entry             569 arch/x86/kvm/cpuid.c 		entry->edx = 0;
entry             576 arch/x86/kvm/cpuid.c 			do_cpuid_7_mask(&entry[i], i);
entry             577 arch/x86/kvm/cpuid.c 			if (i == entry->eax)
entry             583 arch/x86/kvm/cpuid.c 			do_host_cpuid(&entry[i], function, i);
entry             613 arch/x86/kvm/cpuid.c 		entry->eax = eax.full;
entry             614 arch/x86/kvm/cpuid.c 		entry->ebx = cap.events_mask;
entry             615 arch/x86/kvm/cpuid.c 		entry->ecx = 0;
entry             616 arch/x86/kvm/cpuid.c 		entry->edx = edx.full;
entry             635 arch/x86/kvm/cpuid.c 		for (i = 1; entry[i - 1].ecx & 0xff00; ++i) {
entry             639 arch/x86/kvm/cpuid.c 			do_host_cpuid(&entry[i], function, i);
entry             648 arch/x86/kvm/cpuid.c 		entry->eax &= supported;
entry             649 arch/x86/kvm/cpuid.c 		entry->ebx = xstate_required_size(supported, false);
entry             650 arch/x86/kvm/cpuid.c 		entry->ecx = entry->ebx;
entry             651 arch/x86/kvm/cpuid.c 		entry->edx &= supported >> 32;
entry             660 arch/x86/kvm/cpuid.c 			do_host_cpuid(&entry[i], function, idx);
entry             662 arch/x86/kvm/cpuid.c 				entry[i].eax &= kvm_cpuid_D_1_eax_x86_features;
entry             663 arch/x86/kvm/cpuid.c 				cpuid_mask(&entry[i].eax, CPUID_D_1_EAX);
entry             664 arch/x86/kvm/cpuid.c 				entry[i].ebx = 0;
entry             665 arch/x86/kvm/cpuid.c 				if (entry[i].eax & (F(XSAVES)|F(XSAVEC)))
entry             666 arch/x86/kvm/cpuid.c 					entry[i].ebx =
entry             670 arch/x86/kvm/cpuid.c 				if (entry[i].eax == 0 || !(supported & mask))
entry             672 arch/x86/kvm/cpuid.c 				if (WARN_ON_ONCE(entry[i].ecx & 1))
entry             675 arch/x86/kvm/cpuid.c 			entry[i].ecx = 0;
entry             676 arch/x86/kvm/cpuid.c 			entry[i].edx = 0;
entry             684 arch/x86/kvm/cpuid.c 		int t, times = entry->eax;
entry             692 arch/x86/kvm/cpuid.c 			do_host_cpuid(&entry[t], function, t);
entry             700 arch/x86/kvm/cpuid.c 		entry->eax = KVM_CPUID_FEATURES;
entry             701 arch/x86/kvm/cpuid.c 		entry->ebx = sigptr[0];
entry             702 arch/x86/kvm/cpuid.c 		entry->ecx = sigptr[1];
entry             703 arch/x86/kvm/cpuid.c 		entry->edx = sigptr[2];
entry             707 arch/x86/kvm/cpuid.c 		entry->eax = (1 << KVM_FEATURE_CLOCKSOURCE) |
entry             721 arch/x86/kvm/cpuid.c 			entry->eax |= (1 << KVM_FEATURE_STEAL_TIME);
entry             723 arch/x86/kvm/cpuid.c 		entry->ebx = 0;
entry             724 arch/x86/kvm/cpuid.c 		entry->ecx = 0;
entry             725 arch/x86/kvm/cpuid.c 		entry->edx = 0;
entry             728 arch/x86/kvm/cpuid.c 		entry->eax = min(entry->eax, 0x8000001f);
entry             731 arch/x86/kvm/cpuid.c 		entry->edx &= kvm_cpuid_8000_0001_edx_x86_features;
entry             732 arch/x86/kvm/cpuid.c 		cpuid_mask(&entry->edx, CPUID_8000_0001_EDX);
entry             733 arch/x86/kvm/cpuid.c 		entry->ecx &= kvm_cpuid_8000_0001_ecx_x86_features;
entry             734 arch/x86/kvm/cpuid.c 		cpuid_mask(&entry->ecx, CPUID_8000_0001_ECX);
entry             738 arch/x86/kvm/cpuid.c 		entry->edx &= (1 << 8);
entry             740 arch/x86/kvm/cpuid.c 		entry->edx &= boot_cpu_data.x86_power;
entry             741 arch/x86/kvm/cpuid.c 		entry->eax = entry->ebx = entry->ecx = 0;
entry             744 arch/x86/kvm/cpuid.c 		unsigned g_phys_as = (entry->eax >> 16) & 0xff;
entry             745 arch/x86/kvm/cpuid.c 		unsigned virt_as = max((entry->eax >> 8) & 0xff, 48U);
entry             746 arch/x86/kvm/cpuid.c 		unsigned phys_as = entry->eax & 0xff;
entry             750 arch/x86/kvm/cpuid.c 		entry->eax = g_phys_as | (virt_as << 8);
entry             751 arch/x86/kvm/cpuid.c 		entry->edx = 0;
entry             752 arch/x86/kvm/cpuid.c 		entry->ebx &= kvm_cpuid_8000_0008_ebx_x86_features;
entry             753 arch/x86/kvm/cpuid.c 		cpuid_mask(&entry->ebx, CPUID_8000_0008_EBX);
entry             760 arch/x86/kvm/cpuid.c 			entry->ebx |= F(AMD_IBPB);
entry             762 arch/x86/kvm/cpuid.c 			entry->ebx |= F(AMD_IBRS);
entry             764 arch/x86/kvm/cpuid.c 			entry->ebx |= F(AMD_STIBP);
entry             767 arch/x86/kvm/cpuid.c 			entry->ebx |= F(AMD_SSBD);
entry             769 arch/x86/kvm/cpuid.c 			entry->ebx |= F(AMD_SSB_NO);
entry             776 arch/x86/kvm/cpuid.c 			entry->ebx |= F(VIRT_SSBD);
entry             780 arch/x86/kvm/cpuid.c 		entry->ecx = entry->edx = 0;
entry             788 arch/x86/kvm/cpuid.c 		entry->eax = min(entry->eax, 0xC0000004);
entry             791 arch/x86/kvm/cpuid.c 		entry->edx &= kvm_cpuid_C000_0001_edx_x86_features;
entry             792 arch/x86/kvm/cpuid.c 		cpuid_mask(&entry->edx, CPUID_C000_0001_EDX);
entry             800 arch/x86/kvm/cpuid.c 		entry->eax = entry->ebx = entry->ecx = entry->edx = 0;
entry             804 arch/x86/kvm/cpuid.c 	kvm_x86_ops->set_supported_cpuid(function, entry);
entry             814 arch/x86/kvm/cpuid.c static int do_cpuid_func(struct kvm_cpuid_entry2 *entry, u32 func,
entry             821 arch/x86/kvm/cpuid.c 		return __do_cpuid_func_emulated(entry, func, nent, maxnent);
entry             823 arch/x86/kvm/cpuid.c 	return __do_cpuid_func(entry, func, nent, maxnent);
entry            1000 arch/x86/kvm/cpuid.c 	struct kvm_cpuid_entry2 *entry;
entry            1004 arch/x86/kvm/cpuid.c 	entry = kvm_find_cpuid_entry(vcpu, function, index);
entry            1005 arch/x86/kvm/cpuid.c 	found = entry;
entry            1012 arch/x86/kvm/cpuid.c 	if (!entry && check_limit && !guest_cpuid_is_amd(vcpu) &&
entry            1017 arch/x86/kvm/cpuid.c 			entry = kvm_find_cpuid_entry(vcpu, function, index);
entry            1020 arch/x86/kvm/cpuid.c 	if (entry) {
entry            1021 arch/x86/kvm/cpuid.c 		*eax = entry->eax;
entry            1022 arch/x86/kvm/cpuid.c 		*ebx = entry->ebx;
entry            1023 arch/x86/kvm/cpuid.c 		*ecx = entry->ecx;
entry            1024 arch/x86/kvm/cpuid.c 		*edx = entry->edx;
entry            1035 arch/x86/kvm/cpuid.c 			entry = kvm_find_cpuid_entry(vcpu, function, 1);
entry            1036 arch/x86/kvm/cpuid.c 			if (entry) {
entry            1038 arch/x86/kvm/cpuid.c 				*edx = entry->edx;
entry              70 arch/x86/kvm/cpuid.h 	struct kvm_cpuid_entry2 *entry;
entry              73 arch/x86/kvm/cpuid.h 	entry = kvm_find_cpuid_entry(vcpu, cpuid.function, cpuid.index);
entry              74 arch/x86/kvm/cpuid.h 	if (!entry)
entry              79 arch/x86/kvm/cpuid.h 		return &entry->eax;
entry              81 arch/x86/kvm/cpuid.h 		return &entry->ebx;
entry              83 arch/x86/kvm/cpuid.h 		return &entry->ecx;
entry              85 arch/x86/kvm/cpuid.h 		return &entry->edx;
entry             176 arch/x86/kvm/ioapic.c 	union kvm_ioapic_redirect_entry entry;
entry             181 arch/x86/kvm/ioapic.c 	entry = ioapic->redirtbl[irq];
entry             182 arch/x86/kvm/ioapic.c 	edge = (entry.fields.trig_mode == IOAPIC_EDGE_TRIG);
entry             220 arch/x86/kvm/ioapic.c 	trace_kvm_ioapic_set_irq(entry.bits, irq, ret == 0);
entry             334 arch/x86/kvm/ioapic.c 	union kvm_ioapic_redirect_entry *entry = &ioapic->redirtbl[irq];
entry             338 arch/x86/kvm/ioapic.c 	if (entry->fields.mask ||
entry             339 arch/x86/kvm/ioapic.c 	    (entry->fields.trig_mode == IOAPIC_LEVEL_TRIG &&
entry             340 arch/x86/kvm/ioapic.c 	    entry->fields.remote_irr))
entry             343 arch/x86/kvm/ioapic.c 	irqe.dest_id = entry->fields.dest_id;
entry             344 arch/x86/kvm/ioapic.c 	irqe.vector = entry->fields.vector;
entry             345 arch/x86/kvm/ioapic.c 	irqe.dest_mode = entry->fields.dest_mode;
entry             346 arch/x86/kvm/ioapic.c 	irqe.trig_mode = entry->fields.trig_mode;
entry             347 arch/x86/kvm/ioapic.c 	irqe.delivery_mode = entry->fields.delivery_mode << 8;
entry             370 arch/x86/kvm/ioapic.c 		entry->fields.remote_irr = 1;
entry             401 arch/x86/kvm/irq_comm.c 	struct kvm_kernel_irq_routing_entry *entry;
entry             411 arch/x86/kvm/irq_comm.c 		hlist_for_each_entry(entry, &table->map[i], link) {
entry             414 arch/x86/kvm/irq_comm.c 			if (entry->type != KVM_IRQ_ROUTING_MSI)
entry             417 arch/x86/kvm/irq_comm.c 			kvm_set_msi_irq(vcpu->kvm, entry, &irq);
entry            5438 arch/x86/kvm/mmu.c 	u64 entry, gentry, *spte;
entry            5483 arch/x86/kvm/mmu.c 			entry = *spte;
entry            5489 arch/x86/kvm/mmu.c 			if (need_remote_flush(entry, *spte))
entry             493 arch/x86/kvm/svm.c 	u64 *entry = svm->avic_physical_id_cache;
entry             495 arch/x86/kvm/svm.c 	if (!entry)
entry             498 arch/x86/kvm/svm.c 	return (READ_ONCE(*entry) & AVIC_PHYSICAL_ID_ENTRY_IS_RUNNING_MASK);
entry            1743 arch/x86/kvm/svm.c 	u64 *entry, new_entry;
entry            1760 arch/x86/kvm/svm.c 	entry = avic_get_physical_id_entry(vcpu, id);
entry            1761 arch/x86/kvm/svm.c 	if (!entry)
entry            1767 arch/x86/kvm/svm.c 	WRITE_ONCE(*entry, new_entry);
entry            1769 arch/x86/kvm/svm.c 	svm->avic_physical_id_cache = entry;
entry            2078 arch/x86/kvm/svm.c 	u64 entry;
entry            2093 arch/x86/kvm/svm.c 	entry = READ_ONCE(*(svm->avic_physical_id_cache));
entry            2094 arch/x86/kvm/svm.c 	WARN_ON(entry & AVIC_PHYSICAL_ID_ENTRY_IS_RUNNING_MASK);
entry            2096 arch/x86/kvm/svm.c 	entry &= ~AVIC_PHYSICAL_ID_ENTRY_HOST_PHYSICAL_ID_MASK;
entry            2097 arch/x86/kvm/svm.c 	entry |= (h_physical_id & AVIC_PHYSICAL_ID_ENTRY_HOST_PHYSICAL_ID_MASK);
entry            2099 arch/x86/kvm/svm.c 	entry &= ~AVIC_PHYSICAL_ID_ENTRY_IS_RUNNING_MASK;
entry            2101 arch/x86/kvm/svm.c 		entry |= AVIC_PHYSICAL_ID_ENTRY_IS_RUNNING_MASK;
entry            2103 arch/x86/kvm/svm.c 	WRITE_ONCE(*(svm->avic_physical_id_cache), entry);
entry            2110 arch/x86/kvm/svm.c 	u64 entry;
entry            2116 arch/x86/kvm/svm.c 	entry = READ_ONCE(*(svm->avic_physical_id_cache));
entry            2117 arch/x86/kvm/svm.c 	if (entry & AVIC_PHYSICAL_ID_ENTRY_IS_RUNNING_MASK)
entry            2120 arch/x86/kvm/svm.c 	entry &= ~AVIC_PHYSICAL_ID_ENTRY_IS_RUNNING_MASK;
entry            2121 arch/x86/kvm/svm.c 	WRITE_ONCE(*(svm->avic_physical_id_cache), entry);
entry            4615 arch/x86/kvm/svm.c 	u32 *entry, new_entry;
entry            4618 arch/x86/kvm/svm.c 	entry = avic_get_logical_id_entry(vcpu, ldr, flat);
entry            4619 arch/x86/kvm/svm.c 	if (!entry)
entry            4622 arch/x86/kvm/svm.c 	new_entry = READ_ONCE(*entry);
entry            4626 arch/x86/kvm/svm.c 	WRITE_ONCE(*entry, new_entry);
entry            4635 arch/x86/kvm/svm.c 	u32 *entry = avic_get_logical_id_entry(vcpu, svm->ldr_reg, flat);
entry            4637 arch/x86/kvm/svm.c 	if (entry)
entry            4638 arch/x86/kvm/svm.c 		clear_bit(AVIC_LOGICAL_ID_ENTRY_VALID_BIT, (unsigned long *)entry);
entry            5958 arch/x86/kvm/svm.c static void svm_set_supported_cpuid(u32 func, struct kvm_cpuid_entry2 *entry)
entry            5963 arch/x86/kvm/svm.c 			entry->ecx &= ~bit(X86_FEATURE_X2APIC);
entry            5967 arch/x86/kvm/svm.c 			entry->ecx |= (1 << 2); /* Set SVM bit */
entry            5972 arch/x86/kvm/svm.c 			entry->ebx |= F(VIRT_SSBD);
entry            5975 arch/x86/kvm/svm.c 		entry->eax = 1; /* SVM revision 1 */
entry            5976 arch/x86/kvm/svm.c 		entry->ebx = 8; /* Lets support 8 ASIDs in case we add proper
entry            5978 arch/x86/kvm/svm.c 		entry->ecx = 0; /* Reserved */
entry            5979 arch/x86/kvm/svm.c 		entry->edx = 0; /* Per default do not support any
entry            5984 arch/x86/kvm/svm.c 			entry->edx |= F(NRIPS);
entry            5988 arch/x86/kvm/svm.c 			entry->edx |= F(NPT);
entry            5994 arch/x86/kvm/svm.c 			cpuid(0x8000001f, &entry->eax, &entry->ebx,
entry            5995 arch/x86/kvm/svm.c 				&entry->ecx, &entry->edx);
entry              78 arch/x86/kvm/vmx/nested.c 		struct shadow_vmcs_field entry = shadow_read_only_fields[i];
entry              79 arch/x86/kvm/vmx/nested.c 		u16 field = entry.encoding;
entry              92 arch/x86/kvm/vmx/nested.c 			entry.offset += sizeof(u32);
entry              94 arch/x86/kvm/vmx/nested.c 		shadow_read_only_fields[j++] = entry;
entry              99 arch/x86/kvm/vmx/nested.c 		struct shadow_vmcs_field entry = shadow_read_write_fields[i];
entry             100 arch/x86/kvm/vmx/nested.c 		u16 field = entry.encoding;
entry             140 arch/x86/kvm/vmx/nested.c 			entry.offset += sizeof(u32);
entry             142 arch/x86/kvm/vmx/nested.c 		shadow_read_write_fields[j++] = entry;
entry             274 arch/x86/kvm/vmx/pmu_intel.c 	struct kvm_cpuid_entry2 *entry;
entry             285 arch/x86/kvm/vmx/pmu_intel.c 	entry = kvm_find_cpuid_entry(vcpu, 0xa, 0);
entry             286 arch/x86/kvm/vmx/pmu_intel.c 	if (!entry)
entry             288 arch/x86/kvm/vmx/pmu_intel.c 	eax.full = entry->eax;
entry             289 arch/x86/kvm/vmx/pmu_intel.c 	edx.full = entry->edx;
entry             300 arch/x86/kvm/vmx/pmu_intel.c 	pmu->available_event_types = ~entry->ebx &
entry             323 arch/x86/kvm/vmx/pmu_intel.c 	entry = kvm_find_cpuid_entry(vcpu, 7, 0);
entry             324 arch/x86/kvm/vmx/pmu_intel.c 	if (entry &&
entry             326 arch/x86/kvm/vmx/pmu_intel.c 	    (entry->ebx & (X86_FEATURE_HLE|X86_FEATURE_RTM)))
entry             809 arch/x86/kvm/vmx/vmx.c 		unsigned long entry, unsigned long exit)
entry             811 arch/x86/kvm/vmx/vmx.c 	vm_entry_controls_clearbit(vmx, entry);
entry             867 arch/x86/kvm/vmx/vmx.c 		unsigned long entry, unsigned long exit,
entry             874 arch/x86/kvm/vmx/vmx.c 	vm_entry_controls_setbit(vmx, entry);
entry            6221 arch/x86/kvm/vmx/vmx.c 	unsigned long entry;
entry            6235 arch/x86/kvm/vmx/vmx.c 	entry = gate_offset(desc);
entry            6255 arch/x86/kvm/vmx/vmx.c 		THUNK_TARGET(entry),
entry            6941 arch/x86/kvm/vmx/vmx.c 	struct kvm_cpuid_entry2 *entry;
entry            6947 arch/x86/kvm/vmx/vmx.c 	if (entry && (entry->_reg & (_cpuid_mask)))			\
entry            6951 arch/x86/kvm/vmx/vmx.c 	entry = kvm_find_cpuid_entry(vcpu, 0x1, 0);
entry            6967 arch/x86/kvm/vmx/vmx.c 	entry = kvm_find_cpuid_entry(vcpu, 0x7, 0);
entry            7089 arch/x86/kvm/vmx/vmx.c static void vmx_set_supported_cpuid(u32 func, struct kvm_cpuid_entry2 *entry)
entry            7092 arch/x86/kvm/vmx/vmx.c 		entry->ecx |= bit(X86_FEATURE_VMX);
entry             961 arch/x86/mm/init.c void update_cache_mode_entry(unsigned entry, enum page_cache_mode cache)
entry             964 arch/x86/mm/init.c 	BUG_ON(!entry && cache != _PAGE_CACHE_MODE_WB);
entry             966 arch/x86/mm/init.c 	__cachemode2pte_tbl[cache] = __cm_idx2pte(entry);
entry             967 arch/x86/mm/init.c 	__pte2cachemode_tbl[entry] = cache;
entry            1483 arch/x86/mm/init_64.c 				pte_t entry;
entry            1485 arch/x86/mm/init_64.c 				entry = pfn_pte(__pa(p) >> PAGE_SHIFT,
entry            1487 arch/x86/mm/init_64.c 				set_pmd(pmd, __pmd(pte_val(entry)));
entry              63 arch/x86/mm/kasan_init_64.c 		pte_t entry;
entry              70 arch/x86/mm/kasan_init_64.c 		entry = pfn_pte(PFN_DOWN(__pa(p)), PAGE_KERNEL);
entry              71 arch/x86/mm/kasan_init_64.c 		set_pte_at(&init_mm, addr, pte, entry);
entry              23 arch/x86/mm/mm_internal.h void update_cache_mode_entry(unsigned entry, enum page_cache_mode cache);
entry             630 arch/x86/mm/pat.c 	struct memtype *entry;
entry             653 arch/x86/mm/pat.c 	entry = rbt_memtype_erase(start, end);
entry             656 arch/x86/mm/pat.c 	if (IS_ERR(entry)) {
entry             662 arch/x86/mm/pat.c 	kfree(entry);
entry             682 arch/x86/mm/pat.c 	struct memtype *entry;
entry             696 arch/x86/mm/pat.c 	entry = rbt_memtype_lookup(paddr);
entry             697 arch/x86/mm/pat.c 	if (entry != NULL)
entry             698 arch/x86/mm/pat.c 		rettype = entry->type;
entry             481 arch/x86/mm/pgtable.c 			  pte_t entry, int dirty)
entry             483 arch/x86/mm/pgtable.c 	int changed = !pte_same(*ptep, entry);
entry             486 arch/x86/mm/pgtable.c 		set_pte(ptep, entry);
entry             494 arch/x86/mm/pgtable.c 			  pmd_t entry, int dirty)
entry             496 arch/x86/mm/pgtable.c 	int changed = !pmd_same(*pmdp, entry);
entry             501 arch/x86/mm/pgtable.c 		set_pmd(pmdp, entry);
entry             514 arch/x86/mm/pgtable.c 			  pud_t *pudp, pud_t entry, int dirty)
entry             516 arch/x86/mm/pgtable.c 	int changed = !pud_same(*pudp, entry);
entry             521 arch/x86/mm/pgtable.c 		set_pud(pudp, entry);
entry             136 arch/x86/oprofile/op_model_amd.c 	struct op_entry entry;
entry             145 arch/x86/oprofile/op_model_amd.c 			oprofile_write_reserve(&entry, regs, val,
entry             147 arch/x86/oprofile/op_model_amd.c 			oprofile_add_data64(&entry, val);
entry             148 arch/x86/oprofile/op_model_amd.c 			oprofile_add_data64(&entry, ctl);
entry             150 arch/x86/oprofile/op_model_amd.c 			oprofile_add_data64(&entry, val);
entry             151 arch/x86/oprofile/op_model_amd.c 			oprofile_write_commit(&entry);
entry             164 arch/x86/oprofile/op_model_amd.c 			oprofile_write_reserve(&entry, regs, val, IBS_OP_CODE,
entry             166 arch/x86/oprofile/op_model_amd.c 			oprofile_add_data64(&entry, val);
entry             168 arch/x86/oprofile/op_model_amd.c 			oprofile_add_data64(&entry, val);
entry             170 arch/x86/oprofile/op_model_amd.c 			oprofile_add_data64(&entry, val);
entry             172 arch/x86/oprofile/op_model_amd.c 			oprofile_add_data64(&entry, val);
entry             174 arch/x86/oprofile/op_model_amd.c 			oprofile_add_data64(&entry, val);
entry             176 arch/x86/oprofile/op_model_amd.c 			oprofile_add_data64(&entry, val);
entry             179 arch/x86/oprofile/op_model_amd.c 				oprofile_add_data(&entry, (unsigned long)val);
entry             181 arch/x86/oprofile/op_model_amd.c 			oprofile_write_commit(&entry);
entry             298 arch/x86/pci/acpi.c 	struct resource_entry *entry, *tmp;
entry             303 arch/x86/pci/acpi.c 		resource_list_for_each_entry_safe(entry, tmp, &ci->resources)
entry             304 arch/x86/pci/acpi.c 			if (resource_is_pcicfg_ioport(entry->res))
entry             305 arch/x86/pci/acpi.c 				resource_list_destroy_entry(entry);
entry             309 arch/x86/pci/acpi.c 	resource_list_for_each_entry_safe(entry, tmp, &ci->resources) {
entry             311 arch/x86/pci/acpi.c 			   "host bridge window %pR (ignored)\n", entry->res);
entry             312 arch/x86/pci/acpi.c 		resource_list_destroy_entry(entry);
entry             120 arch/x86/pci/i386.c 	struct pcibios_fwaddrmap *entry, *next;
entry             123 arch/x86/pci/i386.c 	list_for_each_entry_safe(entry, next, &pcibios_fwaddrmappings, list) {
entry             124 arch/x86/pci/i386.c 		list_del(&entry->list);
entry             125 arch/x86/pci/i386.c 		pci_dev_put(entry->dev);
entry             126 arch/x86/pci/i386.c 		kfree(entry);
entry              66 arch/x86/pci/pcbios.c 		unsigned long entry;		/* 32 bit physical address */
entry              96 arch/x86/pci/pcbios.c 	unsigned long entry;		/* %edx */
entry             104 arch/x86/pci/pcbios.c 		  "=d" (entry)
entry             112 arch/x86/pci/pcbios.c 			return address + entry;
entry             324 arch/x86/pci/pcbios.c 		if (check->fields.entry >= 0x100000) {
entry             329 arch/x86/pci/pcbios.c 			unsigned long bios32_entry = check->fields.entry;
entry             710 arch/x86/platform/efi/efi.c static inline void *efi_map_next_entry_reverse(void *entry)
entry             713 arch/x86/platform/efi/efi.c 	if (!entry)
entry             716 arch/x86/platform/efi/efi.c 	entry -= efi.memmap.desc_size;
entry             717 arch/x86/platform/efi/efi.c 	if (entry < efi.memmap.map)
entry             720 arch/x86/platform/efi/efi.c 	return entry;
entry             734 arch/x86/platform/efi/efi.c static void *efi_map_next_entry(void *entry)
entry             754 arch/x86/platform/efi/efi.c 		return efi_map_next_entry_reverse(entry);
entry             758 arch/x86/platform/efi/efi.c 	if (!entry)
entry             761 arch/x86/platform/efi/efi.c 	entry += efi.memmap.desc_size;
entry             762 arch/x86/platform/efi/efi.c 	if (entry >= efi.memmap.map_end)
entry             765 arch/x86/platform/efi/efi.c 	return entry;
entry              77 arch/x86/platform/intel-mid/device_libs/platform_msic.c 	struct sfi_device_table_entry *entry = info;
entry              80 arch/x86/platform/intel-mid/device_libs/platform_msic.c 	msic_pdata.irq[block] = entry->irq;
entry              30 arch/x86/platform/uv/uv_irq.c 	struct uv_IO_APIC_route_entry *entry;
entry              36 arch/x86/platform/uv/uv_irq.c 	entry = (struct uv_IO_APIC_route_entry *)&mmr_value;
entry              37 arch/x86/platform/uv/uv_irq.c 	entry->vector		= cfg->vector;
entry              38 arch/x86/platform/uv/uv_irq.c 	entry->delivery_mode	= apic->irq_delivery_mode;
entry              39 arch/x86/platform/uv/uv_irq.c 	entry->dest_mode	= apic->irq_dest_mode;
entry              40 arch/x86/platform/uv/uv_irq.c 	entry->polarity		= 0;
entry              41 arch/x86/platform/uv/uv_irq.c 	entry->trigger		= 0;
entry              42 arch/x86/platform/uv/uv_irq.c 	entry->mask		= 0;
entry              43 arch/x86/platform/uv/uv_irq.c 	entry->dest		= cfg->dest_apicid;
entry             140 arch/x86/platform/uv/uv_irq.c 	struct uv_IO_APIC_route_entry *entry;
entry             143 arch/x86/platform/uv/uv_irq.c 	entry = (struct uv_IO_APIC_route_entry *)&mmr_value;
entry             144 arch/x86/platform/uv/uv_irq.c 	entry->mask = 1;
entry             585 arch/x86/xen/enlighten_pv.c 	u64 entry = *(u64 *)ptr;
entry             587 arch/x86/xen/enlighten_pv.c 	trace_xen_cpu_write_ldt_entry(dt, entrynum, entry);
entry             592 arch/x86/xen/enlighten_pv.c 	if (HYPERVISOR_update_descriptor(mach_lp.maddr, entry))
entry             646 arch/x86/xen/enlighten_pv.c 		struct trap_array_entry *entry = trap_array + nr;
entry             648 arch/x86/xen/enlighten_pv.c 		if (*addr == entry->orig) {
entry             649 arch/x86/xen/enlighten_pv.c 			*addr = entry->xen;
entry             650 arch/x86/xen/enlighten_pv.c 			ist_okay = entry->ist_okay;
entry             739 arch/x86/xen/enlighten_pv.c 		gate_desc *entry = (gate_desc *)(desc->address) + in;
entry             741 arch/x86/xen/enlighten_pv.c 		if (cvt_gate_to_trap(in, entry, &traps[out]))
entry             779 arch/x86/xen/enlighten_pv.c static void xen_write_gdt_entry(struct desc_struct *dt, int entry,
entry             782 arch/x86/xen/enlighten_pv.c 	trace_xen_cpu_write_gdt_entry(dt, entry, desc, type);
entry             793 arch/x86/xen/enlighten_pv.c 		xmaddr_t maddr = arbitrary_virt_to_machine(&dt[entry]);
entry             809 arch/x86/xen/enlighten_pv.c static void __init xen_write_gdt_entry_boot(struct desc_struct *dt, int entry,
entry             812 arch/x86/xen/enlighten_pv.c 	trace_xen_cpu_write_gdt_entry(dt, entry, desc, type);
entry             821 arch/x86/xen/enlighten_pv.c 		xmaddr_t maddr = virt_to_machine(&dt[entry]);
entry             824 arch/x86/xen/enlighten_pv.c 			dt[entry] = *(struct desc_struct *)desc;
entry             202 arch/x86/xen/setup.c 	const struct e820_entry *entry = xen_e820_table.entries;
entry             206 arch/x86/xen/setup.c 	for (i = 0; i < xen_e820_table.nr_entries; i++, entry++) {
entry             210 arch/x86/xen/setup.c 		if (entry->type != E820_TYPE_RAM)
entry             213 arch/x86/xen/setup.c 		e_pfn = PFN_DOWN(entry->addr + entry->size);
entry             219 arch/x86/xen/setup.c 		s_pfn = PFN_UP(entry->addr);
entry             459 arch/x86/xen/setup.c 	const struct e820_entry *entry = xen_e820_table.entries;
entry             473 arch/x86/xen/setup.c 	for (i = 0; i < xen_e820_table.nr_entries; i++, entry++) {
entry             474 arch/x86/xen/setup.c 		phys_addr_t end = entry->addr + entry->size;
entry             475 arch/x86/xen/setup.c 		if (entry->type == E820_TYPE_RAM || i == xen_e820_table.nr_entries - 1) {
entry             479 arch/x86/xen/setup.c 			if (entry->type == E820_TYPE_RAM)
entry             480 arch/x86/xen/setup.c 				end_pfn = PFN_UP(entry->addr);
entry             608 arch/x86/xen/setup.c 	struct e820_entry *entry = xen_e820_table.entries;
entry             611 arch/x86/xen/setup.c 	for (i = 0; i < xen_e820_table.nr_entries; i++, entry++) {
entry             612 arch/x86/xen/setup.c 		if (entry->type == E820_TYPE_UNUSABLE)
entry             613 arch/x86/xen/setup.c 			entry->type = E820_TYPE_RAM;
entry             619 arch/x86/xen/setup.c 	struct e820_entry *entry;
entry             627 arch/x86/xen/setup.c 	entry = xen_e820_table.entries;
entry             630 arch/x86/xen/setup.c 		if (entry->type == E820_TYPE_RAM && entry->addr <= start &&
entry             631 arch/x86/xen/setup.c 		    (entry->addr + entry->size) >= end)
entry             634 arch/x86/xen/setup.c 		entry++;
entry             652 arch/x86/xen/setup.c 	struct e820_entry *entry = xen_e820_table.entries;
entry             654 arch/x86/xen/setup.c 	for (mapcnt = 0; mapcnt < xen_e820_table.nr_entries; mapcnt++, entry++) {
entry             655 arch/x86/xen/setup.c 		if (entry->type != E820_TYPE_RAM || entry->size < size)
entry             657 arch/x86/xen/setup.c 		start = entry->addr;
entry             662 arch/x86/xen/setup.c 			if (start + size > entry->addr + entry->size)
entry             201 arch/xtensa/include/asm/asmmacro.h 	entry sp, (XTENSA_FRAME_SIZE_RESERVE + \
entry             391 arch/xtensa/include/asm/pgtable.h #define __swp_type(entry)	(((entry).val >> 6) & 0x1f)
entry             392 arch/xtensa/include/asm/pgtable.h #define __swp_offset(entry)	((entry).val >> 11)
entry              93 arch/xtensa/include/asm/tlbflush.h static inline void invalidate_itlb_entry_no_isync (unsigned entry)
entry              96 arch/xtensa/include/asm/tlbflush.h 	__asm__ __volatile__ ("iitlb  %0\n" : : "a" (entry) );
entry              99 arch/xtensa/include/asm/tlbflush.h static inline void invalidate_dtlb_entry_no_isync (unsigned entry)
entry             102 arch/xtensa/include/asm/tlbflush.h 	__asm__ __volatile__ ("idtlb  %0\n" : : "a" (entry) );
entry             130 arch/xtensa/include/asm/tlbflush.h static inline void write_dtlb_entry (pte_t entry, int way)
entry             133 arch/xtensa/include/asm/tlbflush.h 			     : : "r" (way), "r" (entry) );
entry             136 arch/xtensa/include/asm/tlbflush.h static inline void write_itlb_entry (pte_t entry, int way)
entry             139 arch/xtensa/include/asm/tlbflush.h 	                     : : "r" (way), "r" (entry) );
entry             323 arch/xtensa/kernel/perf_event.c 	struct perf_callchain_entry_ctx *entry = data;
entry             325 arch/xtensa/kernel/perf_event.c 	perf_callchain_store(entry, frame->pc);
entry             329 arch/xtensa/kernel/perf_event.c void perf_callchain_kernel(struct perf_callchain_entry_ctx *entry,
entry             332 arch/xtensa/kernel/perf_event.c 	xtensa_backtrace_kernel(regs, entry->max_stack,
entry             333 arch/xtensa/kernel/perf_event.c 				callchain_trace, NULL, entry);
entry             336 arch/xtensa/kernel/perf_event.c void perf_callchain_user(struct perf_callchain_entry_ctx *entry,
entry             339 arch/xtensa/kernel/perf_event.c 	xtensa_backtrace_user(regs, entry->max_stack,
entry             340 arch/xtensa/kernel/perf_event.c 			      callchain_trace, entry);
entry              34 arch/xtensa/kernel/syscall.c #define __SYSCALL(nr, entry, nargs)[nr] = (syscall_t)entry,
entry             236 arch/xtensa/mm/fault.c 	const struct exception_table_entry *entry;
entry             239 arch/xtensa/mm/fault.c 	if ((entry = search_exception_tables(regs->pc)) != NULL) {
entry             241 arch/xtensa/mm/fault.c 			 current->comm, regs->pc, entry->fixup);
entry             243 arch/xtensa/mm/fault.c 		regs->pc = entry->fixup;
entry             376 block/bfq-wf2q.c 	struct bfq_entity *entry;
entry             382 block/bfq-wf2q.c 		entry = rb_entry(parent, struct bfq_entity, rb_node);
entry             384 block/bfq-wf2q.c 		if (bfq_gt(entry->finish, entity->finish))
entry            1361 block/bfq-wf2q.c 	struct bfq_entity *entry, *first = NULL;
entry            1365 block/bfq-wf2q.c 		entry = rb_entry(node, struct bfq_entity, rb_node);
entry            1367 block/bfq-wf2q.c 		if (!bfq_gt(entry->start, vtime))
entry            1368 block/bfq-wf2q.c 			first = entry;
entry            1371 block/bfq-wf2q.c 			entry = rb_entry(node->rb_left,
entry            1373 block/bfq-wf2q.c 			if (!bfq_gt(entry->min_start, vtime)) {
entry              68 block/bio.c    	unsigned int i, entry = -1;
entry              76 block/bio.c    		if (!bslab->slab && entry == -1)
entry              77 block/bio.c    			entry = i;
entry              89 block/bio.c    	if (bio_slab_nr == bio_slab_max && entry == -1) {
entry              99 block/bio.c    	if (entry == -1)
entry             100 block/bio.c    		entry = bio_slab_nr++;
entry             102 block/bio.c    	bslab = &bio_slabs[entry];
entry             104 block/bio.c    	snprintf(bslab->name, sizeof(bslab->name), "bio-%d", entry);
entry             228 block/blk-integrity.c 	struct integrity_sysfs_entry *entry =
entry             231 block/blk-integrity.c 	return entry->show(bi, page);
entry             240 block/blk-integrity.c 	struct integrity_sysfs_entry *entry =
entry             244 block/blk-integrity.c 	if (entry->store)
entry             245 block/blk-integrity.c 		ret = entry->store(bi, page, count);
entry            1135 block/blk-iocost.c 	list_del_init(&wq_entry->entry);
entry              65 block/blk-mq-sysfs.c 	struct blk_mq_ctx_sysfs_entry *entry;
entry              70 block/blk-mq-sysfs.c 	entry = container_of(attr, struct blk_mq_ctx_sysfs_entry, attr);
entry              74 block/blk-mq-sysfs.c 	if (!entry->show)
entry              80 block/blk-mq-sysfs.c 		res = entry->show(ctx, page);
entry              88 block/blk-mq-sysfs.c 	struct blk_mq_ctx_sysfs_entry *entry;
entry              93 block/blk-mq-sysfs.c 	entry = container_of(attr, struct blk_mq_ctx_sysfs_entry, attr);
entry              97 block/blk-mq-sysfs.c 	if (!entry->store)
entry             103 block/blk-mq-sysfs.c 		res = entry->store(ctx, page, length);
entry             111 block/blk-mq-sysfs.c 	struct blk_mq_hw_ctx_sysfs_entry *entry;
entry             116 block/blk-mq-sysfs.c 	entry = container_of(attr, struct blk_mq_hw_ctx_sysfs_entry, attr);
entry             120 block/blk-mq-sysfs.c 	if (!entry->show)
entry             126 block/blk-mq-sysfs.c 		res = entry->show(hctx, page);
entry             135 block/blk-mq-sysfs.c 	struct blk_mq_hw_ctx_sysfs_entry *entry;
entry             140 block/blk-mq-sysfs.c 	entry = container_of(attr, struct blk_mq_hw_ctx_sysfs_entry, attr);
entry             144 block/blk-mq-sysfs.c 	if (!entry->store)
entry             150 block/blk-mq-sysfs.c 		res = entry->store(hctx, page, length);
entry            1094 block/blk-mq.c 	if (!list_empty(&wait->entry)) {
entry            1097 block/blk-mq.c 		list_del_init(&wait->entry);
entry            1136 block/blk-mq.c 	if (!list_empty_careful(&wait->entry))
entry            1143 block/blk-mq.c 	if (!list_empty(&wait->entry)) {
entry            1169 block/blk-mq.c 	list_del_init(&wait->entry);
entry            1347 block/blk-mq.c 		    (no_tag && list_empty_careful(&hctx->dispatch_wait.entry)))
entry            2393 block/blk-mq.c 	INIT_LIST_HEAD(&hctx->dispatch_wait.entry);
entry             228 block/blk-rq-qos.c 	list_del_init(&curr->entry);
entry             256 block/blk-rq-qos.c 			.entry	= LIST_HEAD_INIT(data.wq.entry),
entry             796 block/blk-sysfs.c 	struct queue_sysfs_entry *entry = to_queue(attr);
entry             801 block/blk-sysfs.c 	if (!entry->show)
entry             808 block/blk-sysfs.c 	res = entry->show(q, page);
entry             817 block/blk-sysfs.c 	struct queue_sysfs_entry *entry = to_queue(attr);
entry             821 block/blk-sysfs.c 	if (!entry->store)
entry             830 block/blk-sysfs.c 	res = entry->store(q, page, length);
entry             441 block/elevator.c 	struct elv_fs_entry *entry = to_elv(attr);
entry             445 block/elevator.c 	if (!entry->show)
entry             450 block/elevator.c 	error = e->type ? entry->show(e, page) : -ENOENT;
entry             459 block/elevator.c 	struct elv_fs_entry *entry = to_elv(attr);
entry             463 block/elevator.c 	if (!entry->store)
entry             468 block/elevator.c 	error = e->type ? entry->store(e, page, length) : -ENOENT;
entry             497 block/kyber-iosched.c 		INIT_LIST_HEAD(&khd->domain_wait[i].wait.entry);
entry             719 block/kyber-iosched.c 	if (nr < 0 && list_empty_careful(&wait->wait.entry)) {
entry             739 block/kyber-iosched.c 	if (nr >= 0 && !list_empty_careful(&wait->wait.entry)) {
entry             947 block/kyber-iosched.c 	seq_printf(m, "%d\n", !list_empty_careful(&wait->entry));	\
entry              25 crypto/ablkcipher.c 	struct list_head	entry;
entry              44 crypto/ablkcipher.c 	list_for_each_entry_safe(p, tmp, &walk->buffers, entry) {
entry              46 crypto/ablkcipher.c 		list_del(&p->entry);
entry              56 crypto/ablkcipher.c 	list_add_tail(&p->entry, &walk->buffers);
entry              35 crypto/skcipher.c 	struct list_head entry;
entry             185 crypto/skcipher.c 	list_for_each_entry_safe(p, tmp, &walk->buffers, entry) {
entry             204 crypto/skcipher.c 		list_del(&p->entry);
entry             221 crypto/skcipher.c 	list_add_tail(&p->entry, &walk->buffers);
entry             191 drivers/acpi/ac.c 	struct proc_dir_entry *entry = NULL;
entry             203 drivers/acpi/ac.c 	entry = proc_create_single_data(ACPI_AC_FILE_STATE, S_IRUGO,
entry             205 drivers/acpi/ac.c 	if (!entry)
entry             174 drivers/acpi/acpi_video.c 	struct list_head entry;
entry             204 drivers/acpi/acpi_video.c 	struct list_head entry;
entry            1217 drivers/acpi/acpi_video.c 	list_add_tail(&data->entry, &video->video_device_list);
entry            1241 drivers/acpi/acpi_video.c 	list_for_each_entry(dev, &video->video_device_list, entry)
entry            1824 drivers/acpi/acpi_video.c 	list_for_each_entry(dev, &video->video_device_list, entry) {
entry            1860 drivers/acpi/acpi_video.c 	list_for_each_entry(dev, &video->video_device_list, entry) {
entry            1903 drivers/acpi/acpi_video.c 	list_for_each_entry(dev, &video->video_device_list, entry)
entry            1964 drivers/acpi/acpi_video.c 	list_for_each_entry(dev, &video->video_device_list, entry)
entry            1993 drivers/acpi/acpi_video.c 	list_for_each_entry(dev, &video->video_device_list, entry)
entry            2007 drivers/acpi/acpi_video.c 	list_for_each_entry_safe(dev, next, &video->video_device_list, entry) {
entry            2008 drivers/acpi/acpi_video.c 		list_del(&dev->entry);
entry            2078 drivers/acpi/acpi_video.c 	list_add_tail(&video->entry, &video_bus_head);
entry            2111 drivers/acpi/acpi_video.c 	list_del(&video->entry);
entry            2241 drivers/acpi/acpi_video.c 		list_for_each_entry(video, &video_bus_head, entry)
entry              57 drivers/acpi/apei/apei-base.c int __apei_exec_read_register(struct acpi_whea_header *entry, u64 *val)
entry              61 drivers/acpi/apei/apei-base.c 	rc = apei_read(val, &entry->register_region);
entry              64 drivers/acpi/apei/apei-base.c 	*val >>= entry->register_region.bit_offset;
entry              65 drivers/acpi/apei/apei-base.c 	*val &= entry->mask;
entry              71 drivers/acpi/apei/apei-base.c 			    struct acpi_whea_header *entry)
entry              76 drivers/acpi/apei/apei-base.c 	rc = __apei_exec_read_register(entry, &val);
entry              86 drivers/acpi/apei/apei-base.c 				  struct acpi_whea_header *entry)
entry              90 drivers/acpi/apei/apei-base.c 	rc = apei_exec_read_register(ctx, entry);
entry              93 drivers/acpi/apei/apei-base.c 	ctx->value = (ctx->value == entry->value);
entry              99 drivers/acpi/apei/apei-base.c int __apei_exec_write_register(struct acpi_whea_header *entry, u64 val)
entry             103 drivers/acpi/apei/apei-base.c 	val &= entry->mask;
entry             104 drivers/acpi/apei/apei-base.c 	val <<= entry->register_region.bit_offset;
entry             105 drivers/acpi/apei/apei-base.c 	if (entry->flags & APEI_EXEC_PRESERVE_REGISTER) {
entry             107 drivers/acpi/apei/apei-base.c 		rc = apei_read(&valr, &entry->register_region);
entry             110 drivers/acpi/apei/apei-base.c 		valr &= ~(entry->mask << entry->register_region.bit_offset);
entry             113 drivers/acpi/apei/apei-base.c 	rc = apei_write(val, &entry->register_region);
entry             119 drivers/acpi/apei/apei-base.c 			     struct acpi_whea_header *entry)
entry             121 drivers/acpi/apei/apei-base.c 	return __apei_exec_write_register(entry, ctx->value);
entry             126 drivers/acpi/apei/apei-base.c 				   struct acpi_whea_header *entry)
entry             130 drivers/acpi/apei/apei-base.c 	ctx->value = entry->value;
entry             131 drivers/acpi/apei/apei-base.c 	rc = apei_exec_write_register(ctx, entry);
entry             138 drivers/acpi/apei/apei-base.c 		   struct acpi_whea_header *entry)
entry             153 drivers/acpi/apei/apei-base.c 	struct acpi_whea_header *entry;
entry             167 drivers/acpi/apei/apei-base.c 		entry = &ctx->action_table[i];
entry             168 drivers/acpi/apei/apei-base.c 		if (entry->action != action)
entry             171 drivers/acpi/apei/apei-base.c 			if (entry->instruction >= ctx->instructions ||
entry             172 drivers/acpi/apei/apei-base.c 			    !ctx->ins_table[entry->instruction].run) {
entry             175 drivers/acpi/apei/apei-base.c 					   entry->instruction);
entry             178 drivers/acpi/apei/apei-base.c 			run = ctx->ins_table[entry->instruction].run;
entry             179 drivers/acpi/apei/apei-base.c 			rc = run(ctx, entry);
entry             195 drivers/acpi/apei/apei-base.c 				      struct acpi_whea_header *entry,
entry             205 drivers/acpi/apei/apei-base.c 	struct acpi_whea_header *entry;
entry             209 drivers/acpi/apei/apei-base.c 		entry = ctx->action_table + i;
entry             210 drivers/acpi/apei/apei-base.c 		ins = entry->instruction;
entry             219 drivers/acpi/apei/apei-base.c 		rc = func(ctx, entry, data);
entry             228 drivers/acpi/apei/apei-base.c 				struct acpi_whea_header *entry,
entry             231 drivers/acpi/apei/apei-base.c 	u8 ins = entry->instruction;
entry             234 drivers/acpi/apei/apei-base.c 		return apei_map_generic_address(&entry->register_region);
entry             261 drivers/acpi/apei/apei-base.c 				   struct acpi_whea_header *entry,
entry             264 drivers/acpi/apei/apei-base.c 	u8 ins = entry->instruction;
entry             267 drivers/acpi/apei/apei-base.c 		apei_unmap_generic_address(&entry->register_region);
entry             706 drivers/acpi/apei/apei-base.c 				struct acpi_whea_header *entry,
entry             710 drivers/acpi/apei/apei-base.c 	struct acpi_generic_address *reg = &entry->register_region;
entry             711 drivers/acpi/apei/apei-base.c 	u8 ins = entry->instruction;
entry              16 drivers/acpi/apei/apei-internal.h 				    struct acpi_whea_header *entry);
entry              83 drivers/acpi/apei/apei-internal.h int __apei_exec_read_register(struct acpi_whea_header *entry, u64 *val);
entry              84 drivers/acpi/apei/apei-internal.h int __apei_exec_write_register(struct acpi_whea_header *entry, u64 val);
entry              86 drivers/acpi/apei/apei-internal.h 			    struct acpi_whea_header *entry);
entry              88 drivers/acpi/apei/apei-internal.h 				  struct acpi_whea_header *entry);
entry              90 drivers/acpi/apei/apei-internal.h 			     struct acpi_whea_header *entry);
entry              92 drivers/acpi/apei/apei-internal.h 				   struct acpi_whea_header *entry);
entry              94 drivers/acpi/apei/apei-internal.h 		   struct acpi_whea_header *entry);
entry             208 drivers/acpi/apei/einj.c 	struct acpi_whea_header *entry;
entry             210 drivers/acpi/apei/einj.c 	entry = EINJ_TAB_ENTRY(einj_tab);
entry             212 drivers/acpi/apei/einj.c 		if (entry->action == ACPI_EINJ_SET_ERROR_TYPE &&
entry             213 drivers/acpi/apei/einj.c 		    entry->instruction == ACPI_EINJ_WRITE_REGISTER &&
entry             214 drivers/acpi/apei/einj.c 		    entry->register_region.space_id ==
entry             216 drivers/acpi/apei/einj.c 			pa_v4 = get_unaligned(&entry->register_region.address);
entry             217 drivers/acpi/apei/einj.c 		if (entry->action == ACPI_EINJ_SET_ERROR_TYPE_WITH_ADDRESS &&
entry             218 drivers/acpi/apei/einj.c 		    entry->instruction == ACPI_EINJ_WRITE_REGISTER &&
entry             219 drivers/acpi/apei/einj.c 		    entry->register_region.space_id ==
entry             221 drivers/acpi/apei/einj.c 			pa_v5 = get_unaligned(&entry->register_region.address);
entry             222 drivers/acpi/apei/einj.c 		entry++;
entry             270 drivers/acpi/apei/einj.c 	struct acpi_whea_header *entry;
entry             272 drivers/acpi/apei/einj.c 	entry = (struct acpi_whea_header *)
entry             275 drivers/acpi/apei/einj.c 		if (entry->action == ACPI_EINJ_TRIGGER_ERROR &&
entry             276 drivers/acpi/apei/einj.c 		entry->instruction <= ACPI_EINJ_WRITE_REGISTER_VALUE &&
entry             277 drivers/acpi/apei/einj.c 		entry->register_region.space_id ==
entry             279 drivers/acpi/apei/einj.c 		(entry->register_region.address & param2) == (param1 & param2))
entry             280 drivers/acpi/apei/einj.c 			return &entry->register_region;
entry             281 drivers/acpi/apei/einj.c 		entry++;
entry             113 drivers/acpi/apei/erst.c 			       struct acpi_whea_header *entry)
entry             115 drivers/acpi/apei/erst.c 	return __apei_exec_read_register(entry, &ctx->var1);
entry             119 drivers/acpi/apei/erst.c 			       struct acpi_whea_header *entry)
entry             121 drivers/acpi/apei/erst.c 	return __apei_exec_read_register(entry, &ctx->var2);
entry             125 drivers/acpi/apei/erst.c 				struct acpi_whea_header *entry)
entry             127 drivers/acpi/apei/erst.c 	return __apei_exec_write_register(entry, ctx->var1);
entry             131 drivers/acpi/apei/erst.c 			 struct acpi_whea_header *entry)
entry             138 drivers/acpi/apei/erst.c 			      struct acpi_whea_header *entry)
entry             145 drivers/acpi/apei/erst.c 			       struct acpi_whea_header *entry)
entry             150 drivers/acpi/apei/erst.c 	rc = __apei_exec_read_register(entry, &val);
entry             154 drivers/acpi/apei/erst.c 	rc = __apei_exec_write_register(entry, val);
entry             159 drivers/acpi/apei/erst.c 				    struct acpi_whea_header *entry)
entry             164 drivers/acpi/apei/erst.c 	rc = __apei_exec_read_register(entry, &val);
entry             168 drivers/acpi/apei/erst.c 	rc = __apei_exec_write_register(entry, val);
entry             173 drivers/acpi/apei/erst.c 			   struct acpi_whea_header *entry)
entry             190 drivers/acpi/apei/erst.c 				      struct acpi_whea_header *entry)
entry             207 drivers/acpi/apei/erst.c 		rc = __apei_exec_read_register(entry, &val);
entry             220 drivers/acpi/apei/erst.c 	struct acpi_whea_header *entry)
entry             225 drivers/acpi/apei/erst.c 	rc = __apei_exec_read_register(entry, &val);
entry             237 drivers/acpi/apei/erst.c 			  struct acpi_whea_header *entry)
entry             244 drivers/acpi/apei/erst.c 					  struct acpi_whea_header *entry)
entry             246 drivers/acpi/apei/erst.c 	return __apei_exec_read_register(entry, &ctx->src_base);
entry             250 drivers/acpi/apei/erst.c 					  struct acpi_whea_header *entry)
entry             252 drivers/acpi/apei/erst.c 	return __apei_exec_read_register(entry, &ctx->dst_base);
entry             256 drivers/acpi/apei/erst.c 			       struct acpi_whea_header *entry)
entry             268 drivers/acpi/apei/erst.c 	rc = __apei_exec_read_register(entry, &offset);
entry             277 drivers/acpi/button.c 	struct proc_dir_entry *entry = NULL;
entry             309 drivers/acpi/button.c 	entry = proc_create_single_data(ACPI_BUTTON_FILE_STATE, S_IRUGO,
entry             312 drivers/acpi/button.c 	if (!entry) {
entry             268 drivers/acpi/glue.c 	struct acpi_device_physical_node *entry;
entry             275 drivers/acpi/glue.c 	list_for_each_entry(entry, &acpi_dev->physical_node_list, node)
entry             276 drivers/acpi/glue.c 		if (entry->dev == dev) {
entry             279 drivers/acpi/glue.c 			list_del(&entry->node);
entry             282 drivers/acpi/glue.c 			acpi_physnode_link_name(physnode_name, entry->node_id);
entry             289 drivers/acpi/glue.c 			kfree(entry);
entry             161 drivers/acpi/hmat/hmat.c static u32 hmat_normalize(u16 entry, u64 base, u8 type)
entry             168 drivers/acpi/hmat/hmat.c 	if (entry == 0xffff || !entry)
entry             170 drivers/acpi/hmat/hmat.c 	else if (base > (UINT_MAX / (entry)))
entry             177 drivers/acpi/hmat/hmat.c 	value = entry * base;
entry             192 drivers/acpi/numa.c 			u8 val = slit->entry[d*i + j];
entry             237 drivers/acpi/numa.c 				slit->entry[slit->locality_count * i + j]);
entry              95 drivers/acpi/nvs.c 	struct nvs_page *entry, *next;
entry             103 drivers/acpi/nvs.c 		entry = kzalloc(sizeof(struct nvs_page), GFP_KERNEL);
entry             104 drivers/acpi/nvs.c 		if (!entry)
entry             107 drivers/acpi/nvs.c 		list_add_tail(&entry->node, &nvs_list);
entry             108 drivers/acpi/nvs.c 		entry->phys_start = start;
entry             110 drivers/acpi/nvs.c 		entry->size = (size < nr_bytes) ? size : nr_bytes;
entry             112 drivers/acpi/nvs.c 		start += entry->size;
entry             113 drivers/acpi/nvs.c 		size -= entry->size;
entry             118 drivers/acpi/nvs.c 	list_for_each_entry_safe(entry, next, &nvs_list, node) {
entry             119 drivers/acpi/nvs.c 		list_del(&entry->node);
entry             120 drivers/acpi/nvs.c 		kfree(entry);
entry             130 drivers/acpi/nvs.c 	struct nvs_page *entry;
entry             132 drivers/acpi/nvs.c 	list_for_each_entry(entry, &nvs_list, node)
entry             133 drivers/acpi/nvs.c 		if (entry->data) {
entry             134 drivers/acpi/nvs.c 			free_page((unsigned long)entry->data);
entry             135 drivers/acpi/nvs.c 			entry->data = NULL;
entry             136 drivers/acpi/nvs.c 			if (entry->kaddr) {
entry             137 drivers/acpi/nvs.c 				if (entry->unmap) {
entry             138 drivers/acpi/nvs.c 					iounmap(entry->kaddr);
entry             139 drivers/acpi/nvs.c 					entry->unmap = false;
entry             141 drivers/acpi/nvs.c 					acpi_os_unmap_iomem(entry->kaddr,
entry             142 drivers/acpi/nvs.c 							    entry->size);
entry             144 drivers/acpi/nvs.c 				entry->kaddr = NULL;
entry             154 drivers/acpi/nvs.c 	struct nvs_page *entry;
entry             156 drivers/acpi/nvs.c 	list_for_each_entry(entry, &nvs_list, node) {
entry             157 drivers/acpi/nvs.c 		entry->data = (void *)__get_free_page(GFP_KERNEL);
entry             158 drivers/acpi/nvs.c 		if (!entry->data) {
entry             171 drivers/acpi/nvs.c 	struct nvs_page *entry;
entry             175 drivers/acpi/nvs.c 	list_for_each_entry(entry, &nvs_list, node)
entry             176 drivers/acpi/nvs.c 		if (entry->data) {
entry             177 drivers/acpi/nvs.c 			unsigned long phys = entry->phys_start;
entry             178 drivers/acpi/nvs.c 			unsigned int size = entry->size;
entry             180 drivers/acpi/nvs.c 			entry->kaddr = acpi_os_get_iomem(phys, size);
entry             181 drivers/acpi/nvs.c 			if (!entry->kaddr) {
entry             182 drivers/acpi/nvs.c 				entry->kaddr = acpi_os_ioremap(phys, size);
entry             183 drivers/acpi/nvs.c 				entry->unmap = !!entry->kaddr;
entry             185 drivers/acpi/nvs.c 			if (!entry->kaddr) {
entry             189 drivers/acpi/nvs.c 			memcpy(entry->data, entry->kaddr, entry->size);
entry             203 drivers/acpi/nvs.c 	struct nvs_page *entry;
entry             207 drivers/acpi/nvs.c 	list_for_each_entry(entry, &nvs_list, node)
entry             208 drivers/acpi/nvs.c 		if (entry->data)
entry             209 drivers/acpi/nvs.c 			memcpy(entry->kaddr, entry->data, entry->size);
entry             112 drivers/acpi/pci_irq.c static void do_prt_fixups(struct acpi_prt_entry *entry,
entry             123 drivers/acpi/pci_irq.c 		    entry->id.segment == quirk->segment &&
entry             124 drivers/acpi/pci_irq.c 		    entry->id.bus == quirk->bus &&
entry             125 drivers/acpi/pci_irq.c 		    entry->id.device == quirk->device &&
entry             126 drivers/acpi/pci_irq.c 		    entry->pin == quirk->pin &&
entry             132 drivers/acpi/pci_irq.c 				entry->id.segment, entry->id.bus,
entry             133 drivers/acpi/pci_irq.c 				entry->id.device, pin_name(entry->pin),
entry             147 drivers/acpi/pci_irq.c 	struct acpi_prt_entry *entry;
entry             153 drivers/acpi/pci_irq.c 	entry = kzalloc(sizeof(struct acpi_prt_entry), GFP_KERNEL);
entry             154 drivers/acpi/pci_irq.c 	if (!entry)
entry             162 drivers/acpi/pci_irq.c 	entry->id.segment = segment;
entry             163 drivers/acpi/pci_irq.c 	entry->id.bus = bus;
entry             164 drivers/acpi/pci_irq.c 	entry->id.device = (prt->address >> 16) & 0xFFFF;
entry             165 drivers/acpi/pci_irq.c 	entry->pin = prt->pin + 1;
entry             167 drivers/acpi/pci_irq.c 	do_prt_fixups(entry, prt);
entry             169 drivers/acpi/pci_irq.c 	entry->index = prt->source_index;
entry             185 drivers/acpi/pci_irq.c 		acpi_get_handle(handle, prt->source, &entry->link);
entry             197 drivers/acpi/pci_irq.c 			      entry->id.segment, entry->id.bus,
entry             198 drivers/acpi/pci_irq.c 			      entry->id.device, pin_name(entry->pin),
entry             199 drivers/acpi/pci_irq.c 			      prt->source, entry->index));
entry             201 drivers/acpi/pci_irq.c 	*entry_ptr = entry;
entry             211 drivers/acpi/pci_irq.c 	struct acpi_pci_routing_table *entry;
entry             227 drivers/acpi/pci_irq.c 	entry = buffer.pointer;
entry             228 drivers/acpi/pci_irq.c 	while (entry && (entry->length > 0)) {
entry             230 drivers/acpi/pci_irq.c 						 entry, entry_ptr))
entry             232 drivers/acpi/pci_irq.c 		entry = (struct acpi_pci_routing_table *)
entry             233 drivers/acpi/pci_irq.c 		    ((unsigned long)entry + entry->length);
entry             268 drivers/acpi/pci_irq.c 				       struct acpi_prt_entry *entry)
entry             285 drivers/acpi/pci_irq.c 				 "IRQ %d\n", entry->index,
entry             286 drivers/acpi/pci_irq.c 				 (entry->index % 4) + 16);
entry             287 drivers/acpi/pci_irq.c 			entry->index = (entry->index % 4) + 16;
entry             291 drivers/acpi/pci_irq.c 				 "IRQ: unknown mapping\n", entry->index);
entry             300 drivers/acpi/pci_irq.c 	struct acpi_prt_entry *entry = NULL;
entry             305 drivers/acpi/pci_irq.c 	ret = acpi_pci_irq_find_prt_entry(dev, pin, &entry);
entry             306 drivers/acpi/pci_irq.c 	if (!ret && entry) {
entry             308 drivers/acpi/pci_irq.c 		acpi_reroute_boot_interrupt(dev, entry);
entry             312 drivers/acpi/pci_irq.c 		return entry;
entry             335 drivers/acpi/pci_irq.c 		ret = acpi_pci_irq_find_prt_entry(bridge, pin, &entry);
entry             336 drivers/acpi/pci_irq.c 		if (!ret && entry) {
entry             341 drivers/acpi/pci_irq.c 			return entry;
entry             397 drivers/acpi/pci_irq.c 	struct acpi_prt_entry *entry;
entry             425 drivers/acpi/pci_irq.c 	entry = acpi_pci_irq_lookup(dev, pin);
entry             426 drivers/acpi/pci_irq.c 	if (!entry) {
entry             436 drivers/acpi/pci_irq.c 	if (entry) {
entry             437 drivers/acpi/pci_irq.c 		if (entry->link)
entry             438 drivers/acpi/pci_irq.c 			gsi = acpi_pci_link_allocate_irq(entry->link,
entry             439 drivers/acpi/pci_irq.c 							 entry->index,
entry             443 drivers/acpi/pci_irq.c 			gsi = entry->index;
entry             453 drivers/acpi/pci_irq.c 			kfree(entry);
entry             461 drivers/acpi/pci_irq.c 		kfree(entry);
entry             469 drivers/acpi/pci_irq.c 		kfree(entry);
entry             485 drivers/acpi/pci_irq.c 	kfree(entry);
entry             491 drivers/acpi/pci_irq.c 	struct acpi_prt_entry *entry;
entry             507 drivers/acpi/pci_irq.c 	entry = acpi_pci_irq_lookup(dev, pin);
entry             508 drivers/acpi/pci_irq.c 	if (!entry)
entry             511 drivers/acpi/pci_irq.c 	if (entry->link)
entry             512 drivers/acpi/pci_irq.c 		gsi = acpi_pci_link_free_irq(entry->link);
entry             514 drivers/acpi/pci_irq.c 		gsi = entry->index;
entry             516 drivers/acpi/pci_irq.c 	kfree(entry);
entry             151 drivers/acpi/pci_root.c 	struct pci_osc_bit_struct *entry;
entry             154 drivers/acpi/pci_root.c 	for (i = 0, entry = table; i < size; i++, entry++)
entry             155 drivers/acpi/pci_root.c 		if (word & entry->bit)
entry             157 drivers/acpi/pci_root.c 					len ? " " : "", entry->desc);
entry             681 drivers/acpi/pci_root.c 	struct resource_entry *tmp, *entry, *entry2;
entry             687 drivers/acpi/pci_root.c 	resource_list_for_each_entry_safe(entry, tmp, &list) {
entry             691 drivers/acpi/pci_root.c 		res1 = entry->res;
entry             730 drivers/acpi/pci_root.c 		resource_list_del(entry);
entry             732 drivers/acpi/pci_root.c 			resource_list_free_entry(entry);
entry             734 drivers/acpi/pci_root.c 			resource_list_add_tail(entry, resources);
entry             739 drivers/acpi/pci_root.c 			struct resource_entry *entry)
entry             742 drivers/acpi/pci_root.c 	struct resource *res = entry->res;
entry             744 drivers/acpi/pci_root.c 	resource_size_t pci_addr = cpu_addr - entry->offset;
entry             757 drivers/acpi/pci_root.c 	entry->offset = port - pci_addr;
entry             774 drivers/acpi/pci_root.c 	struct resource_entry *entry, *tmp;
entry             788 drivers/acpi/pci_root.c 		resource_list_for_each_entry_safe(entry, tmp, list) {
entry             789 drivers/acpi/pci_root.c 			if (entry->res->flags & IORESOURCE_IO)
entry             791 drivers/acpi/pci_root.c 						entry);
entry             793 drivers/acpi/pci_root.c 			if (entry->res->flags & IORESOURCE_DISABLED)
entry             794 drivers/acpi/pci_root.c 				resource_list_destroy_entry(entry);
entry             796 drivers/acpi/pci_root.c 				entry->res->name = info->name;
entry             809 drivers/acpi/pci_root.c 	struct resource_entry *entry, *tmp;
entry             812 drivers/acpi/pci_root.c 	resource_list_for_each_entry_safe(entry, tmp, &info->resources) {
entry             813 drivers/acpi/pci_root.c 		res = entry->res;
entry             833 drivers/acpi/pci_root.c 			resource_list_destroy_entry(entry);
entry             841 drivers/acpi/pci_root.c 	struct resource_entry *entry, *tmp;
entry             846 drivers/acpi/pci_root.c 	resource_list_for_each_entry_safe(entry, tmp, &info->resources) {
entry             847 drivers/acpi/pci_root.c 		res = entry->res;
entry             851 drivers/acpi/pci_root.c 		resource_list_destroy_entry(entry);
entry             860 drivers/acpi/pci_root.c 	struct resource_entry *entry;
entry             862 drivers/acpi/pci_root.c 	resource_list_for_each_entry(entry, &bridge->windows) {
entry             863 drivers/acpi/pci_root.c 		res = entry->res;
entry              94 drivers/acpi/power.c 	struct acpi_power_resource_entry *entry;
entry              99 drivers/acpi/power.c 	entry = kzalloc(sizeof(*entry), GFP_KERNEL);
entry             100 drivers/acpi/power.c 	if (!entry)
entry             103 drivers/acpi/power.c 	entry->resource = resource;
entry             109 drivers/acpi/power.c 				list_add_tail(&entry->node, &e->node);
entry             113 drivers/acpi/power.c 	list_add_tail(&entry->node, list);
entry             119 drivers/acpi/power.c 	struct acpi_power_resource_entry *entry, *e;
entry             121 drivers/acpi/power.c 	list_for_each_entry_safe(entry, e, list, node) {
entry             122 drivers/acpi/power.c 		list_del(&entry->node);
entry             123 drivers/acpi/power.c 		kfree(entry);
entry             211 drivers/acpi/power.c 	struct acpi_power_resource_entry *entry;
entry             219 drivers/acpi/power.c 	list_for_each_entry(entry, list, node) {
entry             220 drivers/acpi/power.c 		struct acpi_power_resource *resource = entry->resource;
entry             307 drivers/acpi/power.c 	struct acpi_power_resource_entry *entry;
entry             315 drivers/acpi/power.c 	list_for_each_entry(entry, resources, node) {
entry             316 drivers/acpi/power.c 		ret = acpi_power_resource_add_dependent(entry->resource, dev);
entry             324 drivers/acpi/power.c 	list_for_each_entry(entry, resources, node)
entry             325 drivers/acpi/power.c 		acpi_power_resource_remove_dependent(entry->resource, dev);
entry             342 drivers/acpi/power.c 	struct acpi_power_resource_entry *entry;
entry             349 drivers/acpi/power.c 	list_for_each_entry_reverse(entry, resources, node)
entry             350 drivers/acpi/power.c 		acpi_power_resource_remove_dependent(entry->resource, dev);
entry             458 drivers/acpi/power.c 	struct acpi_power_resource_entry *entry;
entry             461 drivers/acpi/power.c 	list_for_each_entry_reverse(entry, list, node) {
entry             462 drivers/acpi/power.c 		result = acpi_power_off(entry->resource);
entry             469 drivers/acpi/power.c 	list_for_each_entry_continue(entry, list, node)
entry             470 drivers/acpi/power.c 		acpi_power_on(entry->resource);
entry             477 drivers/acpi/power.c 	struct acpi_power_resource_entry *entry;
entry             480 drivers/acpi/power.c 	list_for_each_entry(entry, list, node) {
entry             481 drivers/acpi/power.c 		result = acpi_power_on(entry->resource);
entry             488 drivers/acpi/power.c 	list_for_each_entry_continue_reverse(entry, list, node)
entry             489 drivers/acpi/power.c 		acpi_power_off(entry->resource);
entry             526 drivers/acpi/power.c 	struct acpi_power_resource_entry *entry;
entry             531 drivers/acpi/power.c 	list_for_each_entry_reverse(entry, resources, node) {
entry             532 drivers/acpi/power.c 		struct acpi_device *res_dev = &entry->resource->device;
entry             545 drivers/acpi/power.c 	struct acpi_power_resource_entry *entry;
entry             555 drivers/acpi/power.c 	list_for_each_entry(entry, resources, node) {
entry             556 drivers/acpi/power.c 		struct acpi_device *res_dev = &entry->resource->device;
entry             599 drivers/acpi/power.c 	struct acpi_power_resource_entry *entry;
entry             602 drivers/acpi/power.c 	list_for_each_entry(entry, list, node) {
entry             603 drivers/acpi/power.c 		struct acpi_power_resource *resource = entry->resource;
entry             701 drivers/acpi/power.c 	struct acpi_power_resource_entry *entry;
entry             712 drivers/acpi/power.c 	list_for_each_entry(entry, &dev->wakeup.resources, node) {
entry             713 drivers/acpi/power.c 		struct acpi_power_resource *resource = entry->resource;
entry             753 drivers/acpi/power.c 	struct acpi_power_resource_entry *entry;
entry             775 drivers/acpi/power.c 	list_for_each_entry(entry, &dev->wakeup.resources, node) {
entry             776 drivers/acpi/power.c 		struct acpi_power_resource *resource = entry->resource;
entry              27 drivers/acpi/pptt.c 	struct acpi_subtable_header *entry;
entry              36 drivers/acpi/pptt.c 	entry = ACPI_ADD_PTR(struct acpi_subtable_header, table_hdr, pptt_ref);
entry              38 drivers/acpi/pptt.c 	if (entry->length == 0)
entry              41 drivers/acpi/pptt.c 	if (pptt_ref + entry->length > table_hdr->length)
entry              44 drivers/acpi/pptt.c 	return entry;
entry             206 drivers/acpi/pptt.c 	struct acpi_subtable_header *entry;
entry             217 drivers/acpi/pptt.c 	entry = ACPI_ADD_PTR(struct acpi_subtable_header, table_hdr,
entry             221 drivers/acpi/pptt.c 	while ((unsigned long)entry + proc_sz < table_end) {
entry             222 drivers/acpi/pptt.c 		cpu_node = (struct acpi_pptt_processor *)entry;
entry             223 drivers/acpi/pptt.c 		if (entry->type == ACPI_PPTT_TYPE_PROCESSOR &&
entry             226 drivers/acpi/pptt.c 		if (entry->length == 0)
entry             228 drivers/acpi/pptt.c 		entry = ACPI_ADD_PTR(struct acpi_subtable_header, entry,
entry             229 drivers/acpi/pptt.c 				     entry->length);
entry             252 drivers/acpi/pptt.c 	struct acpi_subtable_header *entry;
entry             258 drivers/acpi/pptt.c 	entry = ACPI_ADD_PTR(struct acpi_subtable_header, table_hdr,
entry             263 drivers/acpi/pptt.c 	while ((unsigned long)entry + proc_sz < table_end) {
entry             264 drivers/acpi/pptt.c 		cpu_node = (struct acpi_pptt_processor *)entry;
entry             266 drivers/acpi/pptt.c 		if (entry->length == 0) {
entry             270 drivers/acpi/pptt.c 		if (entry->type == ACPI_PPTT_TYPE_PROCESSOR &&
entry             273 drivers/acpi/pptt.c 			return (struct acpi_pptt_processor *)entry;
entry             276 drivers/acpi/pptt.c 		entry = ACPI_ADD_PTR(struct acpi_subtable_header, entry,
entry             277 drivers/acpi/pptt.c 				     entry->length);
entry              33 drivers/acpi/proc.c 		struct acpi_device_physical_node *entry;
entry              51 drivers/acpi/proc.c 			list_for_each_entry(entry, &dev->physical_node_list,
entry              53 drivers/acpi/proc.c 				ldev = get_device(entry->dev);
entry              57 drivers/acpi/proc.c 				if (&entry->node !=
entry              80 drivers/acpi/proc.c 	struct acpi_device_physical_node *entry;
entry              84 drivers/acpi/proc.c 	list_for_each_entry(entry,
entry              86 drivers/acpi/proc.c 		if (entry->dev && device_can_wakeup(entry->dev)) {
entry              87 drivers/acpi/proc.c 			bool enable = !device_may_wakeup(entry->dev);
entry              88 drivers/acpi/proc.c 			device_set_wakeup_enable(entry->dev, enable);
entry              35 drivers/acpi/processor_core.c static int map_lapic_id(struct acpi_subtable_header *entry,
entry              39 drivers/acpi/processor_core.c 		container_of(entry, struct acpi_madt_local_apic, header);
entry              51 drivers/acpi/processor_core.c static int map_x2apic_id(struct acpi_subtable_header *entry,
entry              55 drivers/acpi/processor_core.c 		container_of(entry, struct acpi_madt_local_x2apic, header);
entry              68 drivers/acpi/processor_core.c static int map_lsapic_id(struct acpi_subtable_header *entry,
entry              72 drivers/acpi/processor_core.c 		container_of(entry, struct acpi_madt_local_sapic, header);
entry              78 drivers/acpi/processor_core.c 		if ((entry->length < 16) || (lsapic->uid != acpi_id))
entry              90 drivers/acpi/processor_core.c static int map_gicc_mpidr(struct acpi_subtable_header *entry,
entry              94 drivers/acpi/processor_core.c 	    container_of(entry, struct acpi_madt_generic_interrupt, header);
entry             115 drivers/acpi/processor_core.c 	unsigned long madt_end, entry;
entry             121 drivers/acpi/processor_core.c 	entry = (unsigned long)madt;
entry             122 drivers/acpi/processor_core.c 	madt_end = entry + madt->header.length;
entry             126 drivers/acpi/processor_core.c 	entry += sizeof(struct acpi_table_madt);
entry             127 drivers/acpi/processor_core.c 	while (entry + sizeof(struct acpi_subtable_header) < madt_end) {
entry             129 drivers/acpi/processor_core.c 			(struct acpi_subtable_header *)entry;
entry             143 drivers/acpi/processor_core.c 		entry += header->length;
entry             268 drivers/acpi/processor_core.c static int get_ioapic_id(struct acpi_subtable_header *entry, u32 gsi_base,
entry             271 drivers/acpi/processor_core.c 	struct acpi_madt_io_apic *ioapic = (struct acpi_madt_io_apic *)entry;
entry             284 drivers/acpi/processor_core.c 	unsigned long madt_end, entry;
entry             292 drivers/acpi/processor_core.c 	entry = (unsigned long)madt;
entry             293 drivers/acpi/processor_core.c 	madt_end = entry + madt->header.length;
entry             296 drivers/acpi/processor_core.c 	entry += sizeof(struct acpi_table_madt);
entry             297 drivers/acpi/processor_core.c 	while (entry + sizeof(struct acpi_subtable_header) < madt_end) {
entry             298 drivers/acpi/processor_core.c 		hdr = (struct acpi_subtable_header *)entry;
entry             303 drivers/acpi/processor_core.c 			entry += hdr->length;
entry             218 drivers/acpi/tables.c acpi_get_entry_type(struct acpi_subtable_entry *entry)
entry             220 drivers/acpi/tables.c 	switch (entry->type) {
entry             222 drivers/acpi/tables.c 		return entry->hdr->common.type;
entry             224 drivers/acpi/tables.c 		return entry->hdr->hmat.type;
entry             230 drivers/acpi/tables.c acpi_get_entry_length(struct acpi_subtable_entry *entry)
entry             232 drivers/acpi/tables.c 	switch (entry->type) {
entry             234 drivers/acpi/tables.c 		return entry->hdr->common.length;
entry             236 drivers/acpi/tables.c 		return entry->hdr->hmat.length;
entry             242 drivers/acpi/tables.c acpi_get_subtable_header_length(struct acpi_subtable_entry *entry)
entry             244 drivers/acpi/tables.c 	switch (entry->type) {
entry             246 drivers/acpi/tables.c 		return sizeof(entry->hdr->common);
entry             248 drivers/acpi/tables.c 		return sizeof(entry->hdr->hmat);
entry             289 drivers/acpi/tables.c 	struct acpi_subtable_entry entry;
entry             313 drivers/acpi/tables.c 	entry.type = acpi_get_subtable_type(id);
entry             314 drivers/acpi/tables.c 	entry.hdr = (union acpi_subtable_headers *)
entry             316 drivers/acpi/tables.c 	subtable_len = acpi_get_subtable_header_length(&entry);
entry             318 drivers/acpi/tables.c 	while (((unsigned long)entry.hdr) + subtable_len  < table_end) {
entry             323 drivers/acpi/tables.c 			if (acpi_get_entry_type(&entry) != proc[i].id)
entry             326 drivers/acpi/tables.c 			     (!errs && proc[i].handler(entry.hdr, table_end))) {
entry             341 drivers/acpi/tables.c 		entry_len = acpi_get_entry_length(&entry);
entry             347 drivers/acpi/tables.c 		entry.hdr = (union acpi_subtable_headers *)
entry             348 drivers/acpi/tables.c 		    ((unsigned long)entry.hdr + entry_len);
entry             206 drivers/android/binder.c 	if (cur >= ARRAY_SIZE(log->entry))
entry             208 drivers/android/binder.c 	e = &log->entry[cur % ARRAY_SIZE(log->entry)];
entry             228 drivers/android/binder.c 	struct list_head entry;
entry             800 drivers/android/binder.c 	BUG_ON(work->entry.next && !list_empty(&work->entry));
entry             801 drivers/android/binder.c 	list_add_tail(&work->entry, target_list);
entry             862 drivers/android/binder.c 	list_del_init(&work->entry);
entry             886 drivers/android/binder.c 	w = list_first_entry_or_null(list, struct binder_work, entry);
entry             888 drivers/android/binder.c 		list_del_init(&w->entry);
entry            1150 drivers/android/binder.c 	INIT_LIST_HEAD(&node->work.entry);
entry            1220 drivers/android/binder.c 		if (!node->has_weak_ref && list_empty(&node->work.entry)) {
entry            1271 drivers/android/binder.c 		if (list_empty(&node->work.entry)) {
entry            1285 drivers/android/binder.c 				BUG_ON(!list_empty(&node->work.entry));
entry            2997 drivers/android/binder.c 					     struct binder_work, entry);
entry            3920 drivers/android/binder.c 				INIT_LIST_HEAD(&death->work.entry);
entry            3952 drivers/android/binder.c 				if (list_empty(&death->work.entry)) {
entry            3987 drivers/android/binder.c 					    entry) {
entry            5391 drivers/android/binder.c 		BUG_ON(!list_empty(&ref->death->work.entry));
entry            5657 drivers/android/binder.c 	list_for_each_entry(w, &thread->todo, entry) {
entry            5688 drivers/android/binder.c 		list_for_each_entry(w, &node->async_todo, entry)
entry            5762 drivers/android/binder.c 	list_for_each_entry(w, &proc->todo, entry)
entry            5765 drivers/android/binder.c 	list_for_each_entry(w, &proc->delivered_death, entry) {
entry            5923 drivers/android/binder.c 	list_for_each_entry(w, &proc->todo, entry) {
entry            6054 drivers/android/binder.c 	cur = count < ARRAY_SIZE(log->entry) && !log->full ?
entry            6055 drivers/android/binder.c 		0 : count % ARRAY_SIZE(log->entry);
entry            6056 drivers/android/binder.c 	if (count > ARRAY_SIZE(log->entry) || log->full)
entry            6057 drivers/android/binder.c 		count = ARRAY_SIZE(log->entry);
entry            6059 drivers/android/binder.c 		unsigned int index = cur++ % ARRAY_SIZE(log->entry);
entry            6061 drivers/android/binder.c 		print_binder_transaction_log_entry(m, &log->entry[index]);
entry              52 drivers/android/binder_alloc.c 	return list_entry(buffer->entry.next, struct binder_buffer, entry);
entry              57 drivers/android/binder_alloc.c 	return list_entry(buffer->entry.prev, struct binder_buffer, entry);
entry              63 drivers/android/binder_alloc.c 	if (list_is_last(&buffer->entry, &alloc->buffers))
entry             474 drivers/android/binder_alloc.c 		list_add(&new_buffer->entry, &buffer->entry);
entry             551 drivers/android/binder_alloc.c 	BUG_ON(alloc->buffers.next == &buffer->entry);
entry             562 drivers/android/binder_alloc.c 	if (!list_is_last(&buffer->entry, &alloc->buffers)) {
entry             590 drivers/android/binder_alloc.c 	list_del(&buffer->entry);
entry             630 drivers/android/binder_alloc.c 	if (!list_is_last(&buffer->entry, &alloc->buffers)) {
entry             638 drivers/android/binder_alloc.c 	if (alloc->buffers.next != &buffer->entry) {
entry             714 drivers/android/binder_alloc.c 	list_add(&buffer->entry, &alloc->buffers);
entry             762 drivers/android/binder_alloc.c 					  struct binder_buffer, entry);
entry             765 drivers/android/binder_alloc.c 		list_del(&buffer->entry);
entry              39 drivers/android/binder_alloc.h 	struct list_head entry; /* free and allocated entries by address */
entry             141 drivers/android/binder_internal.h 	struct binder_transaction_log_entry entry[32];
entry            3424 drivers/ata/libata-scsi.c 		u64 entry = sector |
entry            3426 drivers/ata/libata-scsi.c 		buf[i++] = __cpu_to_le64(entry);
entry              37 drivers/atm/adummy.c 	struct list_head entry;
entry             177 drivers/atm/adummy.c 	list_add(&adummy_dev->entry, &adummy_devs);
entry             192 drivers/atm/adummy.c 	list_for_each_entry_safe(adummy_dev, next, &adummy_devs, entry) {
entry             559 drivers/atm/fore200e.c     struct host_cmdq_entry* entry = &cmdq->host_entry[ cmdq->head ];
entry             574 drivers/atm/fore200e.c     fore200e->bus->write(prom_dma, &entry->cp_entry->cmd.prom_block.prom_haddr);
entry             576 drivers/atm/fore200e.c     *entry->status = STATUS_PENDING;
entry             578 drivers/atm/fore200e.c     fore200e->bus->write(*(u32*)&opcode, (u32 __iomem *)&entry->cp_entry->cmd.prom_block.opcode);
entry             580 drivers/atm/fore200e.c     ok = fore200e_poll(fore200e, entry->status, STATUS_COMPLETE, 400);
entry             582 drivers/atm/fore200e.c     *entry->status = STATUS_FREE;
entry             773 drivers/atm/fore200e.c     struct host_txq_entry*  entry;
entry             782 drivers/atm/fore200e.c 	entry = &txq->host_entry[ txq->tail ];
entry             784 drivers/atm/fore200e.c         if ((*entry->status & STATUS_COMPLETE) == 0) {
entry             789 drivers/atm/fore200e.c 		entry, txq->tail, entry->vc_map, entry->skb);
entry             792 drivers/atm/fore200e.c 	kfree(entry->data);
entry             795 drivers/atm/fore200e.c 	dma_unmap_single(fore200e->dev, entry->tpd->tsd[ 0 ].buffer, entry->tpd->tsd[ 0 ].length,
entry             798 drivers/atm/fore200e.c 	vc_map = entry->vc_map;
entry             807 drivers/atm/fore200e.c 	    dev_kfree_skb_any(entry->skb);
entry             813 drivers/atm/fore200e.c 	    if (vc_map->incarn != entry->incarn) {
entry             830 drivers/atm/fore200e.c 		dev_kfree_skb_any(entry->skb);
entry             838 drivers/atm/fore200e.c 		    vcc->pop(vcc, entry->skb);
entry             841 drivers/atm/fore200e.c 		    dev_kfree_skb_any(entry->skb);
entry             845 drivers/atm/fore200e.c 		if (*entry->status & STATUS_ERROR)
entry             852 drivers/atm/fore200e.c 	*entry->status = STATUS_FREE;
entry             909 drivers/atm/fore200e.c     struct host_bsq_entry* entry;
entry             925 drivers/atm/fore200e.c 		entry = &bsq->host_entry[ bsq->head ];
entry             944 drivers/atm/fore200e.c 		    entry->rbd_block->rbd[ i ].buffer_haddr = buffer->data.dma_addr;
entry             945 drivers/atm/fore200e.c 		    entry->rbd_block->rbd[ i ].handle       = FORE200E_BUF2HDL(buffer);
entry             953 drivers/atm/fore200e.c 		*entry->status = STATUS_PENDING;
entry             954 drivers/atm/fore200e.c 		fore200e->bus->write(entry->rbd_block_dma, &entry->cp_entry->rbd_block_haddr);
entry            1090 drivers/atm/fore200e.c     struct host_rxq_entry*  entry;
entry            1096 drivers/atm/fore200e.c 	entry = &rxq->host_entry[ rxq->head ];
entry            1099 drivers/atm/fore200e.c 	if ((*entry->status & STATUS_COMPLETE) == 0)
entry            1102 drivers/atm/fore200e.c 	vc_map = FORE200E_VC_MAP(fore200e, entry->rpd->atm_header.vpi, entry->rpd->atm_header.vci);
entry            1109 drivers/atm/fore200e.c 		    entry->rpd->atm_header.vpi, entry->rpd->atm_header.vci);
entry            1115 drivers/atm/fore200e.c 	    if ((*entry->status & STATUS_ERROR) == 0) {
entry            1117 drivers/atm/fore200e.c 		fore200e_push_rpd(fore200e, vcc, entry->rpd);
entry            1122 drivers/atm/fore200e.c 			entry->rpd->atm_header.vpi, entry->rpd->atm_header.vci);
entry            1129 drivers/atm/fore200e.c 	fore200e_collect_rpd(fore200e, entry->rpd);
entry            1132 drivers/atm/fore200e.c 	fore200e->bus->write(entry->rpd_dma, &entry->cp_entry->rpd_haddr);
entry            1133 drivers/atm/fore200e.c 	*entry->status = STATUS_FREE;
entry            1228 drivers/atm/fore200e.c     struct host_cmdq_entry*  entry = &cmdq->host_entry[ cmdq->head ];
entry            1253 drivers/atm/fore200e.c     *entry->status = STATUS_PENDING;
entry            1261 drivers/atm/fore200e.c 	fore200e->bus->write(mtu,                        &entry->cp_entry->cmd.activate_block.mtu);
entry            1262 drivers/atm/fore200e.c 	fore200e->bus->write(*(u32*)&vpvc,         (u32 __iomem *)&entry->cp_entry->cmd.activate_block.vpvc);
entry            1263 drivers/atm/fore200e.c 	fore200e->bus->write(*(u32*)&activ_opcode, (u32 __iomem *)&entry->cp_entry->cmd.activate_block.opcode);
entry            1266 drivers/atm/fore200e.c 	fore200e->bus->write(*(u32*)&vpvc,         (u32 __iomem *)&entry->cp_entry->cmd.deactivate_block.vpvc);
entry            1267 drivers/atm/fore200e.c 	fore200e->bus->write(*(u32*)&deactiv_opcode, (u32 __iomem *)&entry->cp_entry->cmd.deactivate_block.opcode);
entry            1270 drivers/atm/fore200e.c     ok = fore200e_poll(fore200e, entry->status, STATUS_COMPLETE, 400);
entry            1272 drivers/atm/fore200e.c     *entry->status = STATUS_FREE;
entry            1473 drivers/atm/fore200e.c     struct host_txq_entry*  entry;
entry            1561 drivers/atm/fore200e.c     entry = &txq->host_entry[ txq->head ];
entry            1563 drivers/atm/fore200e.c     if ((*entry->status != STATUS_FREE) || (txq->txing >= QUEUE_SIZE_TX - 2)) {
entry            1568 drivers/atm/fore200e.c 	if (*entry->status != STATUS_FREE) {
entry            1597 drivers/atm/fore200e.c     entry->incarn = vc_map->incarn;
entry            1598 drivers/atm/fore200e.c     entry->vc_map = vc_map;
entry            1599 drivers/atm/fore200e.c     entry->skb    = skb;
entry            1600 drivers/atm/fore200e.c     entry->data   = tx_copy ? data : NULL;
entry            1602 drivers/atm/fore200e.c     tpd = entry->tpd;
entry            1657 drivers/atm/fore200e.c     tpd_haddr.haddr = entry->tpd_dma >> TPD_HADDR_SHIFT;          /* shift the address, as we are in a bitfield */
entry            1659 drivers/atm/fore200e.c     *entry->status = STATUS_PENDING;
entry            1660 drivers/atm/fore200e.c     fore200e->bus->write(*(u32*)&tpd_haddr, (u32 __iomem *)&entry->cp_entry->tpd_haddr);
entry            1672 drivers/atm/fore200e.c     struct host_cmdq_entry* entry = &cmdq->host_entry[ cmdq->head ];
entry            1693 drivers/atm/fore200e.c     fore200e->bus->write(stats_dma_addr, &entry->cp_entry->cmd.stats_block.stats_haddr);
entry            1695 drivers/atm/fore200e.c     *entry->status = STATUS_PENDING;
entry            1697 drivers/atm/fore200e.c     fore200e->bus->write(*(u32*)&opcode, (u32 __iomem *)&entry->cp_entry->cmd.stats_block.opcode);
entry            1699 drivers/atm/fore200e.c     ok = fore200e_poll(fore200e, entry->status, STATUS_COMPLETE, 400);
entry            1701 drivers/atm/fore200e.c     *entry->status = STATUS_FREE;
entry            1743 drivers/atm/fore200e.c     struct host_cmdq_entry* entry = &cmdq->host_entry[ cmdq->head ];
entry            1757 drivers/atm/fore200e.c     fore200e->bus->write(oc3_regs_dma_addr, &entry->cp_entry->cmd.oc3_block.regs_haddr);
entry            1759 drivers/atm/fore200e.c     *entry->status = STATUS_PENDING;
entry            1761 drivers/atm/fore200e.c     fore200e->bus->write(*(u32*)&opcode, (u32*)&entry->cp_entry->cmd.oc3_block.opcode);
entry            1763 drivers/atm/fore200e.c     ok = fore200e_poll(fore200e, entry->status, STATUS_COMPLETE, 400);
entry            1765 drivers/atm/fore200e.c     *entry->status = STATUS_FREE;
entry            1783 drivers/atm/fore200e.c     struct host_cmdq_entry* entry = &cmdq->host_entry[ cmdq->head ];
entry            1796 drivers/atm/fore200e.c     fore200e->bus->write(0, &entry->cp_entry->cmd.oc3_block.regs_haddr);
entry            1798 drivers/atm/fore200e.c     *entry->status = STATUS_PENDING;
entry            1800 drivers/atm/fore200e.c     fore200e->bus->write(*(u32*)&opcode, (u32 __iomem *)&entry->cp_entry->cmd.oc3_block.opcode);
entry            1802 drivers/atm/fore200e.c     ok = fore200e_poll(fore200e, entry->status, STATUS_COMPLETE, 400);
entry            1804 drivers/atm/fore200e.c     *entry->status = STATUS_FREE;
entry             833 drivers/atm/fore200e.h     struct       list_head     entry;                  /* next device                        */
entry             825 drivers/atm/he.c 		list_add(&heb->entry, &he_dev->rbpl_outstanding);
entry             894 drivers/atm/he.c 	list_for_each_entry_safe(heb, next, &he_dev->rbpl_outstanding, entry)
entry            1573 drivers/atm/he.c 		list_for_each_entry_safe(heb, next, &he_dev->rbpl_outstanding, entry)
entry            1684 drivers/atm/he.c 				list_del(&heb->entry);
entry            1699 drivers/atm/he.c 		list_move_tail(&heb->entry, &he_vcc->buffers);
entry            1736 drivers/atm/he.c 		list_for_each_entry(heb, &he_vcc->buffers, entry)
entry            1777 drivers/atm/he.c 		list_for_each_entry_safe(heb, next, &he_vcc->buffers, entry)
entry            1822 drivers/atm/he.c 		list_for_each_entry(__tpd, &he_dev->outstanding_tpds, entry) {
entry            1825 drivers/atm/he.c 				list_del(&__tpd->entry);
entry            1913 drivers/atm/he.c 		list_add(&heb->entry, &he_dev->rbpl_outstanding);
entry            2106 drivers/atm/he.c 	list_add_tail(&tpd->entry, &he_dev->outstanding_tpds);
entry             136 drivers/atm/he.h 	struct list_head entry;
entry             224 drivers/atm/he.h 	struct list_head entry;
entry            1223 drivers/atm/horizon.c   rx_queue_entry = rd_mem (dev, &dev->rx_q_entry->entry);
entry             321 drivers/atm/horizon.h   HDW entry;
entry             115 drivers/atm/iphase.c    IARTN_Q *entry = kmalloc(sizeof(*entry), GFP_ATOMIC);
entry             116 drivers/atm/iphase.c    if (!entry)
entry             118 drivers/atm/iphase.c    entry->data = data;
entry             119 drivers/atm/iphase.c    entry->next = NULL;
entry             121 drivers/atm/iphase.c       que->next = que->tail = entry;
entry             123 drivers/atm/iphase.c       que->tail->next = entry;
entry            1545 drivers/base/core.c 	list_move_tail(&deva->kobj.entry, &devb->kobj.entry);
entry            1561 drivers/base/core.c 	list_move(&deva->kobj.entry, &devb->kobj.entry);
entry            1575 drivers/base/core.c 	list_move_tail(&dev->kobj.entry, &devices_kset->list);
entry            1815 drivers/base/core.c 		list_for_each_entry(k, &dev->class->p->glue_dirs.list, entry)
entry            3193 drivers/base/core.c 				kobj.entry);
entry            3206 drivers/base/core.c 		list_del_init(&dev->kobj.entry);
entry              19 drivers/base/devres.c 	struct list_head		entry;
entry             109 drivers/base/devres.c 	INIT_LIST_HEAD(&dr->node.entry);
entry             117 drivers/base/devres.c 	BUG_ON(!list_empty(&node->entry));
entry             118 drivers/base/devres.c 	list_add_tail(&node->entry, &dev->devres_head);
entry             190 drivers/base/devres.c 			&dev->devres_head, entry) {
entry             214 drivers/base/devres.c 		BUG_ON(!list_empty(&dr->node.entry));
entry             245 drivers/base/devres.c 	list_for_each_entry_reverse(node, &dev->devres_head, entry) {
entry             347 drivers/base/devres.c 		list_del_init(&dr->node.entry);
entry             436 drivers/base/devres.c 		node = list_entry(cur, struct devres_node, entry);
entry             446 drivers/base/devres.c 			if (&node->entry == first)
entry             448 drivers/base/devres.c 			list_move_tail(&node->entry, todo);
entry             467 drivers/base/devres.c 		node = list_entry(cur, struct devres_node, entry);
entry             471 drivers/base/devres.c 		BUG_ON(!grp || list_empty(&grp->node[0].entry));
entry             474 drivers/base/devres.c 		if (list_empty(&grp->node[1].entry))
entry             482 drivers/base/devres.c 			list_move_tail(&grp->node[0].entry, todo);
entry             483 drivers/base/devres.c 			list_del_init(&grp->node[1].entry);
entry             505 drivers/base/devres.c 	list_for_each_entry_safe_reverse(dr, tmp, &todo, node.entry) {
entry             557 drivers/base/devres.c 	INIT_LIST_HEAD(&grp->node[0].entry);
entry             558 drivers/base/devres.c 	INIT_LIST_HEAD(&grp->node[1].entry);
entry             577 drivers/base/devres.c 	list_for_each_entry_reverse(node, &dev->devres_head, entry) {
entry             588 drivers/base/devres.c 		} else if (list_empty(&grp->node[1].entry))
entry             638 drivers/base/devres.c 		list_del_init(&grp->node[0].entry);
entry             639 drivers/base/devres.c 		list_del_init(&grp->node[1].entry);
entry             672 drivers/base/devres.c 		struct list_head *first = &grp->node[0].entry;
entry             675 drivers/base/devres.c 		if (!list_empty(&grp->node[1].entry))
entry             676 drivers/base/devres.c 			end = grp->node[1].entry.next;
entry             102 drivers/base/power/main.c 	INIT_LIST_HEAD(&dev->power.entry);
entry             138 drivers/base/power/main.c 	list_add_tail(&dev->power.entry, &dpm_list);
entry             156 drivers/base/power/main.c 	list_del_init(&dev->power.entry);
entry             175 drivers/base/power/main.c 	list_move_tail(&deva->power.entry, &devb->power.entry);
entry             189 drivers/base/power/main.c 	list_move(&deva->power.entry, &devb->power.entry);
entry             200 drivers/base/power/main.c 	list_move_tail(&dev->power.entry, &dpm_list);
entry             762 drivers/base/power/main.c 	list_for_each_entry(dev, &dpm_noirq_list, power.entry)
entry             768 drivers/base/power/main.c 		list_move_tail(&dev->power.entry, &dpm_late_early_list);
entry             912 drivers/base/power/main.c 	list_for_each_entry(dev, &dpm_late_early_list, power.entry)
entry             918 drivers/base/power/main.c 		list_move_tail(&dev->power.entry, &dpm_suspended_list);
entry            1073 drivers/base/power/main.c 	list_for_each_entry(dev, &dpm_suspended_list, power.entry)
entry            1094 drivers/base/power/main.c 		if (!list_empty(&dev->power.entry))
entry            1095 drivers/base/power/main.c 			list_move_tail(&dev->power.entry, &dpm_prepared_list);
entry            1172 drivers/base/power/main.c 		list_move(&dev->power.entry, &list);
entry            1414 drivers/base/power/main.c 		if (!list_empty(&dev->power.entry))
entry            1415 drivers/base/power/main.c 			list_move(&dev->power.entry, &dpm_noirq_list);
entry            1608 drivers/base/power/main.c 		if (!list_empty(&dev->power.entry))
entry            1609 drivers/base/power/main.c 			list_move(&dev->power.entry, &dpm_late_early_list);
entry            1886 drivers/base/power/main.c 		if (!list_empty(&dev->power.entry))
entry            1887 drivers/base/power/main.c 			list_move(&dev->power.entry, &dpm_suspended_list);
entry            2030 drivers/base/power/main.c 		if (!list_empty(&dev->power.entry))
entry            2031 drivers/base/power/main.c 			list_move_tail(&dev->power.entry, &dpm_prepared_list);
entry            2097 drivers/base/power/main.c 	list_for_each_entry(dev, &dpm_list, power.entry)
entry             102 drivers/base/power/power.h static inline struct device *to_device(struct list_head *entry)
entry             104 drivers/base/power/power.h 	return container_of(entry, struct device, power.entry);
entry             197 drivers/base/power/trace.c 	struct list_head *entry;
entry             200 drivers/base/power/trace.c 	entry = dpm_list.prev;
entry             201 drivers/base/power/trace.c 	while (entry != &dpm_list) {
entry             202 drivers/base/power/trace.c 		struct device * dev = to_device(entry);
entry             208 drivers/base/power/trace.c 		entry = entry->prev;
entry             220 drivers/base/power/trace.c 	struct list_head *entry;
entry             227 drivers/base/power/trace.c 	entry = dpm_list.prev;
entry             228 drivers/base/power/trace.c 	while (size && entry != &dpm_list) {
entry             229 drivers/base/power/trace.c 		struct device *dev = to_device(entry);
entry             241 drivers/base/power/trace.c 		entry = entry->prev;
entry             180 drivers/base/power/wakeup.c 	list_add_rcu(&ws->entry, &wakeup_sources);
entry             197 drivers/base/power/wakeup.c 	list_del_rcu(&ws->entry);
entry             356 drivers/base/power/wakeup.c 	list_for_each_entry_rcu(ws, &wakeup_sources, entry)
entry             372 drivers/base/power/wakeup.c 	list_for_each_entry_rcu(ws, &wakeup_sources, entry)
entry             432 drivers/base/power/wakeup.c 	if (device_is_registered(dev) && !list_empty(&dev->power.entry)) {
entry             825 drivers/base/power/wakeup.c 	list_for_each_entry_rcu(ws, &wakeup_sources, entry) {
entry             976 drivers/base/power/wakeup.c 	list_for_each_entry_rcu(ws, &wakeup_sources, entry) {
entry            1055 drivers/base/power/wakeup.c 	list_for_each_entry_rcu(ws, &wakeup_sources, entry) {
entry            1071 drivers/base/power/wakeup.c 	list_for_each_entry_continue_rcu(ws, &wakeup_sources, entry) {
entry             367 drivers/base/regmap/regmap-debugfs.c 	char *entry;
entry             378 drivers/base/regmap/regmap-debugfs.c 	entry = kmalloc(PAGE_SIZE, GFP_KERNEL);
entry             379 drivers/base/regmap/regmap-debugfs.c 	if (!entry) {
entry             397 drivers/base/regmap/regmap-debugfs.c 		entry_len = snprintf(entry, PAGE_SIZE, "%x-%x\n",
entry             402 drivers/base/regmap/regmap-debugfs.c 			memcpy(buf + buf_pos, entry, entry_len);
entry             409 drivers/base/regmap/regmap-debugfs.c 	kfree(entry);
entry              22 drivers/base/swnode.c 	struct list_head entry;
entry              62 drivers/base/swnode.c 	list_for_each_entry(k, &swnode_kset->list, entry) {
entry             537 drivers/base/swnode.c 	    (c && list_is_last(&c->entry, &p->children)))
entry             541 drivers/base/swnode.c 		c = list_next_entry(c, entry);
entry             543 drivers/base/swnode.c 		c = list_first_entry(&p->children, struct swnode, entry);
entry             557 drivers/base/swnode.c 	list_for_each_entry(child, &swnode->children, entry) {
entry             647 drivers/base/swnode.c 	list_for_each_entry(k, &swnode_kset->list, entry) {
entry             684 drivers/base/swnode.c 		list_del(&swnode->entry);
entry             730 drivers/base/swnode.c 	INIT_LIST_HEAD(&swnode->entry);
entry             747 drivers/base/swnode.c 		list_add_tail(&swnode->entry, &parent->children);
entry             315 drivers/block/ps3vram.c static void ps3vram_cache_evict(struct ps3_system_bus_device *dev, int entry)
entry             320 drivers/block/ps3vram.c 	if (!(cache->tags[entry].flags & CACHE_PAGE_DIRTY))
entry             323 drivers/block/ps3vram.c 	dev_dbg(&dev->core, "Flushing %d: 0x%08x\n", entry,
entry             324 drivers/block/ps3vram.c 		cache->tags[entry].address);
entry             325 drivers/block/ps3vram.c 	if (ps3vram_upload(dev, CACHE_OFFSET + entry * cache->page_size,
entry             326 drivers/block/ps3vram.c 			   cache->tags[entry].address, DMA_PAGE_SIZE,
entry             330 drivers/block/ps3vram.c 			entry * cache->page_size, cache->tags[entry].address,
entry             333 drivers/block/ps3vram.c 	cache->tags[entry].flags &= ~CACHE_PAGE_DIRTY;
entry             336 drivers/block/ps3vram.c static void ps3vram_cache_load(struct ps3_system_bus_device *dev, int entry,
entry             342 drivers/block/ps3vram.c 	dev_dbg(&dev->core, "Fetching %d: 0x%08x\n", entry, address);
entry             344 drivers/block/ps3vram.c 			     CACHE_OFFSET + entry * cache->page_size,
entry             349 drivers/block/ps3vram.c 			address, entry * cache->page_size, cache->page_size);
entry             352 drivers/block/ps3vram.c 	cache->tags[entry].address = address;
entry             353 drivers/block/ps3vram.c 	cache->tags[entry].flags |= CACHE_PAGE_PRESENT;
entry             450 drivers/block/ps3vram.c 		unsigned int entry;
entry             455 drivers/block/ps3vram.c 		entry = ps3vram_cache_match(dev, from);
entry             456 drivers/block/ps3vram.c 		cached = CACHE_OFFSET + entry * priv->cache.page_size + offset;
entry             491 drivers/block/ps3vram.c 		unsigned int entry;
entry             496 drivers/block/ps3vram.c 		entry = ps3vram_cache_match(dev, to);
entry             497 drivers/block/ps3vram.c 		cached = CACHE_OFFSET + entry * priv->cache.page_size + offset;
entry             507 drivers/block/ps3vram.c 		priv->cache.tags[entry].flags |= CACHE_PAGE_DIRTY;
entry             590 drivers/block/zram/zram_drv.c 			unsigned long entry, struct bio *parent)
entry             598 drivers/block/zram/zram_drv.c 	bio->bi_iter.bi_sector = entry * (PAGE_SIZE >> 9);
entry             775 drivers/block/zram/zram_drv.c 	unsigned long entry;
entry             785 drivers/block/zram/zram_drv.c 	unsigned long entry = zw->entry;
entry             788 drivers/block/zram/zram_drv.c 	read_from_bdev_async(zram, &zw->bvec, entry, bio);
entry             797 drivers/block/zram/zram_drv.c 				unsigned long entry, struct bio *bio)
entry             803 drivers/block/zram/zram_drv.c 	work.entry = entry;
entry             815 drivers/block/zram/zram_drv.c 				unsigned long entry, struct bio *bio)
entry             823 drivers/block/zram/zram_drv.c 			unsigned long entry, struct bio *parent, bool sync)
entry             827 drivers/block/zram/zram_drv.c 		return read_from_bdev_sync(zram, bvec, entry, parent);
entry             829 drivers/block/zram/zram_drv.c 		return read_from_bdev_async(zram, bvec, entry, parent);
entry             834 drivers/block/zram/zram_drv.c 			unsigned long entry, struct bio *parent, bool sync)
entry            1409 drivers/bluetooth/btmrvl_sdio.c 		struct memory_type_mapping *entry = &mem_type_mapping_tbl[idx];
entry            1411 drivers/bluetooth/btmrvl_sdio.c 		if (entry->mem_ptr) {
entry            1412 drivers/bluetooth/btmrvl_sdio.c 			vfree(entry->mem_ptr);
entry            1413 drivers/bluetooth/btmrvl_sdio.c 			entry->mem_ptr = NULL;
entry            1415 drivers/bluetooth/btmrvl_sdio.c 		entry->mem_size = 0;
entry            1438 drivers/bluetooth/btmrvl_sdio.c 		struct memory_type_mapping *entry = &mem_type_mapping_tbl[idx];
entry            1467 drivers/bluetooth/btmrvl_sdio.c 		BT_INFO("%s_SIZE=0x%x", entry->mem_name, memory_size);
entry            1468 drivers/bluetooth/btmrvl_sdio.c 		entry->mem_ptr = vzalloc(memory_size + 1);
entry            1469 drivers/bluetooth/btmrvl_sdio.c 		entry->mem_size = memory_size;
entry            1470 drivers/bluetooth/btmrvl_sdio.c 		if (!entry->mem_ptr) {
entry            1471 drivers/bluetooth/btmrvl_sdio.c 			BT_ERR("Vzalloc %s failed", entry->mem_name);
entry            1476 drivers/bluetooth/btmrvl_sdio.c 				strlen(entry->mem_name) +
entry            1481 drivers/bluetooth/btmrvl_sdio.c 		dbg_ptr = entry->mem_ptr;
entry            1484 drivers/bluetooth/btmrvl_sdio.c 		doneflag = entry->done_flag;
entry            1486 drivers/bluetooth/btmrvl_sdio.c 			entry->mem_name);
entry            1511 drivers/bluetooth/btmrvl_sdio.c 					entry->mem_name,
entry            1512 drivers/bluetooth/btmrvl_sdio.c 					dbg_ptr - entry->mem_ptr);
entry            1538 drivers/bluetooth/btmrvl_sdio.c 		struct memory_type_mapping *entry = &mem_type_mapping_tbl[idx];
entry            1540 drivers/bluetooth/btmrvl_sdio.c 		if (entry->mem_ptr) {
entry            1544 drivers/bluetooth/btmrvl_sdio.c 			strcpy(fw_dump_ptr, entry->mem_name);
entry            1545 drivers/bluetooth/btmrvl_sdio.c 			fw_dump_ptr += strlen(entry->mem_name);
entry            1550 drivers/bluetooth/btmrvl_sdio.c 			memcpy(fw_dump_ptr, entry->mem_ptr, entry->mem_size);
entry            1551 drivers/bluetooth/btmrvl_sdio.c 			fw_dump_ptr += entry->mem_size;
entry             180 drivers/bluetooth/btqca.c 		       le32_to_cpu(tlv_patch->entry));
entry             107 drivers/bluetooth/btqca.h 	__le32 entry;
entry             734 drivers/bluetooth/btrtl.c 	struct rtl_vendor_config_entry *entry;
entry             756 drivers/bluetooth/btrtl.c 		entry = ((void *)config->entry) + i;
entry             758 drivers/bluetooth/btrtl.c 		switch (le16_to_cpu(entry->offset)) {
entry             760 drivers/bluetooth/btrtl.c 			if (entry->len < sizeof(*device_baudrate)) {
entry             765 drivers/bluetooth/btrtl.c 			*device_baudrate = get_unaligned_le32(entry->data);
entry             769 drivers/bluetooth/btrtl.c 			if (entry->len >= 13)
entry             770 drivers/bluetooth/btrtl.c 				*flow_control = !!(entry->data[12] & BIT(2));
entry             779 drivers/bluetooth/btrtl.c 				   le16_to_cpu(entry->offset), entry->len);
entry             783 drivers/bluetooth/btrtl.c 		i += sizeof(*entry) + entry->len;
entry              47 drivers/bluetooth/btrtl.h 	struct rtl_vendor_config_entry entry[0];
entry            1055 drivers/bluetooth/hci_bcm.c 	struct resource_entry *entry;
entry            1065 drivers/bluetooth/hci_bcm.c 	resource_list_for_each_entry(entry, &resources) {
entry            1066 drivers/bluetooth/hci_bcm.c 		if (resource_type(entry->res) == IORESOURCE_IRQ) {
entry            1067 drivers/bluetooth/hci_bcm.c 			dev->irq = entry->res->start;
entry             558 drivers/bus/moxtet.c 	struct dentry *root, *entry;
entry             565 drivers/bus/moxtet.c 	entry = debugfs_create_file_unsafe("input", 0444, root, moxtet,
entry             567 drivers/bus/moxtet.c 	if (IS_ERR(entry))
entry             570 drivers/bus/moxtet.c 	entry = debugfs_create_file_unsafe("output", 0644, root, moxtet,
entry             572 drivers/bus/moxtet.c 	if (IS_ERR(entry))
entry             580 drivers/bus/moxtet.c 	return PTR_ERR(entry);
entry             991 drivers/cdrom/cdrom.c 	struct cdrom_tocentry entry;
entry            1015 drivers/cdrom/cdrom.c 	entry.cdte_format = CDROM_MSF;
entry            1017 drivers/cdrom/cdrom.c 		entry.cdte_track = i;
entry            1018 drivers/cdrom/cdrom.c 		if (cdi->ops->audio_ioctl(cdi, CDROMREADTOCENTRY, &entry)) {
entry            1022 drivers/cdrom/cdrom.c 		if (entry.cdte_ctrl & CDROM_DATA_TRACK) {
entry            1023 drivers/cdrom/cdrom.c 			if (entry.cdte_format == 0x10)
entry            1025 drivers/cdrom/cdrom.c 			else if (entry.cdte_format == 0x20)
entry            1033 drivers/cdrom/cdrom.c 		       i, entry.cdte_format, entry.cdte_ctrl);
entry            2669 drivers/cdrom/cdrom.c 	struct cdrom_tocentry entry;
entry            2675 drivers/cdrom/cdrom.c 	if (copy_from_user(&entry, argp, sizeof(entry)))
entry            2678 drivers/cdrom/cdrom.c 	requested_format = entry.cdte_format;
entry            2682 drivers/cdrom/cdrom.c 	entry.cdte_format = CDROM_MSF;
entry            2683 drivers/cdrom/cdrom.c 	ret = cdi->ops->audio_ioctl(cdi, CDROMREADTOCENTRY, &entry);
entry            2686 drivers/cdrom/cdrom.c 	sanitize_format(&entry.cdte_addr, &entry.cdte_format, requested_format);
entry            2688 drivers/cdrom/cdrom.c 	if (copy_to_user(argp, &entry, sizeof(entry)))
entry              91 drivers/cdrom/gdrom.c 	unsigned int entry[99];
entry             337 drivers/cdrom/gdrom.c 		data = gd.toc->entry[track - 1];
entry              66 drivers/char/agp/amd-k7-agp.c 	struct amd_page_map *entry;
entry              70 drivers/char/agp/amd-k7-agp.c 		entry = tables[i];
entry              71 drivers/char/agp/amd-k7-agp.c 		if (entry != NULL) {
entry              72 drivers/char/agp/amd-k7-agp.c 			if (entry->real != NULL)
entry              73 drivers/char/agp/amd-k7-agp.c 				amd_free_page_map(entry);
entry              74 drivers/char/agp/amd-k7-agp.c 			kfree(entry);
entry              84 drivers/char/agp/amd-k7-agp.c 	struct amd_page_map *entry;
entry              94 drivers/char/agp/amd-k7-agp.c 		entry = kzalloc(sizeof(struct amd_page_map), GFP_KERNEL);
entry              95 drivers/char/agp/amd-k7-agp.c 		tables[i] = entry;
entry              96 drivers/char/agp/amd-k7-agp.c 		if (entry == NULL) {
entry             100 drivers/char/agp/amd-k7-agp.c 		retval = amd_create_page_map(entry);
entry              89 drivers/char/agp/ati-agp.c 	struct ati_page_map *entry;
entry              93 drivers/char/agp/ati-agp.c 		entry = tables[i];
entry              94 drivers/char/agp/ati-agp.c 		if (entry != NULL) {
entry              95 drivers/char/agp/ati-agp.c 			if (entry->real != NULL)
entry              96 drivers/char/agp/ati-agp.c 				ati_free_page_map(entry);
entry              97 drivers/char/agp/ati-agp.c 			kfree(entry);
entry             107 drivers/char/agp/ati-agp.c 	struct ati_page_map *entry;
entry             117 drivers/char/agp/ati-agp.c 		entry = kzalloc(sizeof(struct ati_page_map), GFP_KERNEL);
entry             118 drivers/char/agp/ati-agp.c 		tables[i] = entry;
entry             119 drivers/char/agp/ati-agp.c 		if (entry == NULL) {
entry             123 drivers/char/agp/ati-agp.c 		retval = ati_create_page_map(entry);
entry              54 drivers/char/agp/intel-gtt.c 	void (*write_entry)(dma_addr_t addr, unsigned int entry, unsigned int flags);
entry             319 drivers/char/agp/intel-gtt.c static void i810_write_entry(dma_addr_t addr, unsigned int entry,
entry             333 drivers/char/agp/intel-gtt.c 	writel_relaxed(addr | pte_flags, intel_private.gtt + entry);
entry             733 drivers/char/agp/intel-gtt.c static void i830_write_entry(dma_addr_t addr, unsigned int entry,
entry             741 drivers/char/agp/intel-gtt.c 	writel_relaxed(addr | pte_flags, intel_private.gtt + entry);
entry            1115 drivers/char/agp/intel-gtt.c 			     unsigned int entry,
entry            1126 drivers/char/agp/intel-gtt.c 	writel_relaxed(addr | pte_flags, intel_private.gtt + entry);
entry              77 drivers/char/agp/sworks-agp.c 	struct serverworks_page_map *entry;
entry              81 drivers/char/agp/sworks-agp.c 		entry = tables[i];
entry              82 drivers/char/agp/sworks-agp.c 		if (entry != NULL) {
entry              83 drivers/char/agp/sworks-agp.c 			if (entry->real != NULL) {
entry              84 drivers/char/agp/sworks-agp.c 				serverworks_free_page_map(entry);
entry              86 drivers/char/agp/sworks-agp.c 			kfree(entry);
entry              95 drivers/char/agp/sworks-agp.c 	struct serverworks_page_map *entry;
entry             105 drivers/char/agp/sworks-agp.c 		entry = kzalloc(sizeof(struct serverworks_page_map), GFP_KERNEL);
entry             106 drivers/char/agp/sworks-agp.c 		if (entry == NULL) {
entry             110 drivers/char/agp/sworks-agp.c 		tables[i] = entry;
entry             111 drivers/char/agp/sworks-agp.c 		retval = serverworks_create_page_map(entry);
entry             204 drivers/char/ipmi/ipmi_devintf.c 	struct list_head *entry;
entry             226 drivers/char/ipmi/ipmi_devintf.c 	entry = priv->recv_msgs.next;
entry             227 drivers/char/ipmi/ipmi_devintf.c 	msg = list_entry(entry, struct ipmi_recv_msg, link);
entry             228 drivers/char/ipmi/ipmi_devintf.c 	list_del(entry);
entry             280 drivers/char/ipmi/ipmi_devintf.c 	list_add(entry, &priv->recv_msgs);
entry             814 drivers/char/ipmi/ipmi_devintf.c 	struct ipmi_reg_list *entry;
entry             816 drivers/char/ipmi/ipmi_devintf.c 	entry = kmalloc(sizeof(*entry), GFP_KERNEL);
entry             817 drivers/char/ipmi/ipmi_devintf.c 	if (!entry) {
entry             821 drivers/char/ipmi/ipmi_devintf.c 	entry->dev = dev;
entry             825 drivers/char/ipmi/ipmi_devintf.c 	list_add(&entry->link, &reg_list);
entry             832 drivers/char/ipmi/ipmi_devintf.c 	struct ipmi_reg_list *entry;
entry             835 drivers/char/ipmi/ipmi_devintf.c 	list_for_each_entry(entry, &reg_list, link) {
entry             836 drivers/char/ipmi/ipmi_devintf.c 		if (entry->dev == dev) {
entry             837 drivers/char/ipmi/ipmi_devintf.c 			list_del(&entry->link);
entry             838 drivers/char/ipmi/ipmi_devintf.c 			kfree(entry);
entry             893 drivers/char/ipmi/ipmi_devintf.c 	struct ipmi_reg_list *entry, *entry2;
entry             895 drivers/char/ipmi/ipmi_devintf.c 	list_for_each_entry_safe(entry, entry2, &reg_list, link) {
entry             896 drivers/char/ipmi/ipmi_devintf.c 		list_del(&entry->link);
entry             897 drivers/char/ipmi/ipmi_devintf.c 		device_destroy(ipmi_class, entry->dev);
entry             898 drivers/char/ipmi/ipmi_devintf.c 		kfree(entry);
entry            3548 drivers/char/ipmi/ipmi_msghandler.c 	struct list_head *entry;
entry            3569 drivers/char/ipmi/ipmi_msghandler.c 		entry = tmplist.next;
entry            3570 drivers/char/ipmi/ipmi_msghandler.c 		list_del(entry);
entry            3571 drivers/char/ipmi/ipmi_msghandler.c 		msg = list_entry(entry, struct ipmi_smi_msg, link);
entry            4504 drivers/char/ipmi/ipmi_msghandler.c 		struct list_head *entry = NULL;
entry            4508 drivers/char/ipmi/ipmi_msghandler.c 			entry = intf->hp_xmit_msgs.next;
entry            4510 drivers/char/ipmi/ipmi_msghandler.c 			entry = intf->xmit_msgs.next;
entry            4512 drivers/char/ipmi/ipmi_msghandler.c 		if (entry) {
entry            4513 drivers/char/ipmi/ipmi_msghandler.c 			list_del(entry);
entry            4514 drivers/char/ipmi/ipmi_msghandler.c 			newmsg = list_entry(entry, struct ipmi_smi_msg, link);
entry             407 drivers/char/xillybus/xillybus_core.c 	int i, entry, rc;
entry             482 drivers/char/xillybus/xillybus_core.c 	for (entry = 0; entry < entries; entry++, chandesc += 4) {
entry             147 drivers/clk/bcm/clk-bcm63xx-gate.c 	const struct clk_bcm63xx_table_entry *entry, *table;
entry             156 drivers/clk/bcm/clk-bcm63xx-gate.c 	for (entry = table; entry->name; entry++)
entry             157 drivers/clk/bcm/clk-bcm63xx-gate.c 		maxbit = max_t(u8, maxbit, entry->bit);
entry             176 drivers/clk/bcm/clk-bcm63xx-gate.c 	for (entry = table; entry->name; entry++) {
entry             179 drivers/clk/bcm/clk-bcm63xx-gate.c 		clk = clk_hw_register_gate(&pdev->dev, entry->name, NULL,
entry             180 drivers/clk/bcm/clk-bcm63xx-gate.c 					   entry->flags, hw->regs, entry->bit,
entry             187 drivers/clk/bcm/clk-bcm63xx-gate.c 		hw->data.hws[entry->bit] = clk;
entry             133 drivers/clk/clk-max9485.c 	const struct max9485_rate *entry;
entry             135 drivers/clk/clk-max9485.c 	for (entry = max9485_rates; entry->out != 0; entry++)
entry             136 drivers/clk/clk-max9485.c 		if (entry->out == rate)
entry             139 drivers/clk/clk-max9485.c 	if (entry->out == 0)
entry             144 drivers/clk/clk-max9485.c 				   entry->reg_value);
entry             153 drivers/clk/clk-max9485.c 	const struct max9485_rate *entry;
entry             155 drivers/clk/clk-max9485.c 	for (entry = max9485_rates; entry->out != 0; entry++)
entry             156 drivers/clk/clk-max9485.c 		if (val == entry->reg_value)
entry             157 drivers/clk/clk-max9485.c 			return entry->out;
entry             422 drivers/clk/clk.c 	struct clk_parent_map *entry = &core->parents[index];
entry             425 drivers/clk/clk.c 	if (entry->hw) {
entry             426 drivers/clk/clk.c 		parent = entry->hw->core;
entry             436 drivers/clk/clk.c 		if (IS_ERR(parent) && PTR_ERR(parent) == -ENOENT && entry->name)
entry             437 drivers/clk/clk.c 			parent = clk_core_lookup(entry->name);
entry             442 drivers/clk/clk.c 		entry->core = parent;
entry              65 drivers/clk/tegra/cvb.c 		const struct cvb_table_freq_entry *entry = &table->entries[i];
entry              67 drivers/clk/tegra/cvb.c 		if (!entry->freq || (entry->freq > max_freq))
entry              71 drivers/clk/tegra/cvb.c 					  &entry->coefficients);
entry              76 drivers/clk/tegra/cvb.c 		ret = dev_pm_opp_add(dev, entry->freq, dfll_mv * 1000);
entry             134 drivers/clk/tegra/cvb.c 		const struct cvb_table_freq_entry *entry = &table->entries[i];
entry             136 drivers/clk/tegra/cvb.c 		if (!entry->freq || (entry->freq > max_freq))
entry             139 drivers/clk/tegra/cvb.c 		dev_pm_opp_remove(dev, entry->freq);
entry             232 drivers/clk/ti/clkctrl.c 	struct omap_clkctrl_clk *entry;
entry             241 drivers/clk/ti/clkctrl.c 	list_for_each_entry(entry, &provider->clocks, node) {
entry             242 drivers/clk/ti/clkctrl.c 		if (entry->reg_offset == clkspec->args[0] &&
entry             243 drivers/clk/ti/clkctrl.c 		    entry->bit_offset == clkspec->args[1]) {
entry             252 drivers/clk/ti/clkctrl.c 	return entry->clk;
entry             460 drivers/cpufreq/acpi-cpufreq.c 	struct cpufreq_frequency_table *entry;
entry             471 drivers/cpufreq/acpi-cpufreq.c 	entry = &policy->freq_table[index];
entry             472 drivers/cpufreq/acpi-cpufreq.c 	next_freq = entry->frequency;
entry             473 drivers/cpufreq/acpi-cpufreq.c 	next_perf_state = entry->driver_data;
entry             165 drivers/cpufreq/powernow-k7.c static void invalidate_entry(unsigned int entry)
entry             167 drivers/cpufreq/powernow-k7.c 	powernow_table[entry].frequency = CPUFREQ_ENTRY_INVALID;
entry             538 drivers/cpufreq/powernow-k8.c 		unsigned int entry)
entry             540 drivers/cpufreq/powernow-k8.c 	powernow_table[entry].frequency = CPUFREQ_ENTRY_INVALID;
entry            1042 drivers/crypto/amcc/crypto4xx_core.c 			list_add_tail(&alg->entry, &sec_dev->alg_list);
entry            1052 drivers/crypto/amcc/crypto4xx_core.c 	list_for_each_entry_safe(alg, tmp, &sec_dev->alg_list, entry) {
entry            1053 drivers/crypto/amcc/crypto4xx_core.c 		list_del(&alg->entry);
entry             144 drivers/crypto/amcc/crypto4xx_core.h 	struct list_head  entry;
entry            4503 drivers/crypto/caam/caamalg_qi2.c 	struct list_head entry;
entry            5303 drivers/crypto/caam/caamalg_qi2.c 			list_add_tail(&t_alg->entry, &hash_list);
entry            5322 drivers/crypto/caam/caamalg_qi2.c 			list_add_tail(&t_alg->entry, &hash_list);
entry            5372 drivers/crypto/caam/caamalg_qi2.c 		list_for_each_entry_safe(t_hash_alg, p, &hash_list, entry) {
entry            5374 drivers/crypto/caam/caamalg_qi2.c 			list_del(&t_hash_alg->entry);
entry            1801 drivers/crypto/caam/caamhash.c 	struct list_head entry;
entry            1931 drivers/crypto/caam/caamhash.c 	list_for_each_entry_safe(t_alg, n, &hash_list, entry) {
entry            1933 drivers/crypto/caam/caamhash.c 		list_del(&t_alg->entry);
entry            2044 drivers/crypto/caam/caamhash.c 			list_add_tail(&t_alg->entry, &hash_list);
entry            2064 drivers/crypto/caam/caamhash.c 			list_add_tail(&t_alg->entry, &hash_list);
entry             414 drivers/crypto/cavium/nitrox/nitrox_isr.c 	ndev->iov.msix.entry = NON_RING_MSIX_BASE;
entry             157 drivers/crypto/ccp/ccp-crypto-aes-cmac.c 	INIT_LIST_HEAD(&rctx->cmd.entry);
entry             361 drivers/crypto/ccp/ccp-crypto-aes-cmac.c 	INIT_LIST_HEAD(&ccp_alg->entry);
entry             398 drivers/crypto/ccp/ccp-crypto-aes-cmac.c 	list_add(&ccp_alg->entry, head);
entry             118 drivers/crypto/ccp/ccp-crypto-aes-galois.c 	INIT_LIST_HEAD(&rctx->cmd.entry);
entry             221 drivers/crypto/ccp/ccp-crypto-aes-galois.c 	INIT_LIST_HEAD(&ccp_aead->entry);
entry             242 drivers/crypto/ccp/ccp-crypto-aes-galois.c 	list_add(&ccp_aead->entry, head);
entry             169 drivers/crypto/ccp/ccp-crypto-aes-xts.c 	INIT_LIST_HEAD(&rctx->cmd.entry);
entry             238 drivers/crypto/ccp/ccp-crypto-aes-xts.c 	INIT_LIST_HEAD(&ccp_alg->entry);
entry             270 drivers/crypto/ccp/ccp-crypto-aes-xts.c 	list_add(&ccp_alg->entry, head);
entry              94 drivers/crypto/ccp/ccp-crypto-aes.c 	INIT_LIST_HEAD(&rctx->cmd.entry);
entry             334 drivers/crypto/ccp/ccp-crypto-aes.c 	INIT_LIST_HEAD(&ccp_alg->entry);
entry             355 drivers/crypto/ccp/ccp-crypto-aes.c 	list_add(&ccp_alg->entry, head);
entry              88 drivers/crypto/ccp/ccp-crypto-des3.c 	INIT_LIST_HEAD(&rctx->cmd.entry);
entry             197 drivers/crypto/ccp/ccp-crypto-des3.c 	INIT_LIST_HEAD(&ccp_alg->entry);
entry             218 drivers/crypto/ccp/ccp-crypto-des3.c 	list_add(&ccp_alg->entry, head);
entry              65 drivers/crypto/ccp/ccp-crypto-main.c 	struct list_head entry;
entry             110 drivers/crypto/ccp/ccp-crypto-main.c 	list_for_each_entry_continue(tmp, &req_queue.cmds, entry) {
entry             123 drivers/crypto/ccp/ccp-crypto-main.c 		if (req_queue.backlog == &crypto_cmd->entry)
entry             124 drivers/crypto/ccp/ccp-crypto-main.c 			req_queue.backlog = crypto_cmd->entry.next;
entry             127 drivers/crypto/ccp/ccp-crypto-main.c 					struct ccp_crypto_cmd, entry);
entry             131 drivers/crypto/ccp/ccp-crypto-main.c 		if (req_queue.backlog == &crypto_cmd->entry)
entry             132 drivers/crypto/ccp/ccp-crypto-main.c 			req_queue.backlog = crypto_cmd->entry.next;
entry             137 drivers/crypto/ccp/ccp-crypto-main.c 	list_del(&crypto_cmd->entry);
entry             232 drivers/crypto/ccp/ccp-crypto-main.c 	list_for_each_entry(tmp, &req_queue.cmds, entry) {
entry             249 drivers/crypto/ccp/ccp-crypto-main.c 			req_queue.backlog = &crypto_cmd->entry;
entry             254 drivers/crypto/ccp/ccp-crypto-main.c 	list_add_tail(&crypto_cmd->entry, &req_queue.cmds);
entry             378 drivers/crypto/ccp/ccp-crypto-main.c 	list_for_each_entry_safe(ahash_alg, ahash_tmp, &hash_algs, entry) {
entry             380 drivers/crypto/ccp/ccp-crypto-main.c 		list_del(&ahash_alg->entry);
entry             384 drivers/crypto/ccp/ccp-crypto-main.c 	list_for_each_entry_safe(ablk_alg, ablk_tmp, &cipher_algs, entry) {
entry             386 drivers/crypto/ccp/ccp-crypto-main.c 		list_del(&ablk_alg->entry);
entry             390 drivers/crypto/ccp/ccp-crypto-main.c 	list_for_each_entry_safe(aead_alg, aead_tmp, &aead_algs, entry) {
entry             392 drivers/crypto/ccp/ccp-crypto-main.c 		list_del(&aead_alg->entry);
entry             396 drivers/crypto/ccp/ccp-crypto-main.c 	list_for_each_entry_safe(akc_alg, akc_tmp, &akcipher_algs, entry) {
entry             398 drivers/crypto/ccp/ccp-crypto-main.c 		list_del(&akc_alg->entry);
entry              72 drivers/crypto/ccp/ccp-crypto-rsa.c 	INIT_LIST_HEAD(&rctx->cmd.entry);
entry             256 drivers/crypto/ccp/ccp-crypto-rsa.c 	INIT_LIST_HEAD(&ccp_alg->entry);
entry             271 drivers/crypto/ccp/ccp-crypto-rsa.c 	list_add(&ccp_alg->entry, head);
entry             132 drivers/crypto/ccp/ccp-crypto-sha.c 	INIT_LIST_HEAD(&rctx->cmd.entry);
entry             432 drivers/crypto/ccp/ccp-crypto-sha.c 	INIT_LIST_HEAD(&ccp_alg->entry);
entry             456 drivers/crypto/ccp/ccp-crypto-sha.c 	list_add(&ccp_alg->entry, head);
entry             474 drivers/crypto/ccp/ccp-crypto-sha.c 	INIT_LIST_HEAD(&ccp_alg->entry);
entry             513 drivers/crypto/ccp/ccp-crypto-sha.c 	list_add(&ccp_alg->entry, head);
entry              35 drivers/crypto/ccp/ccp-crypto.h 	struct list_head entry;
entry              43 drivers/crypto/ccp/ccp-crypto.h 	struct list_head entry;
entry              51 drivers/crypto/ccp/ccp-crypto.h 	struct list_head entry;
entry              64 drivers/crypto/ccp/ccp-crypto.h 	struct list_head entry;
entry             556 drivers/crypto/ccp/ccp-dev-v3.c 		cmd = list_first_entry(&ccp->cmd, struct ccp_cmd, entry);
entry             557 drivers/crypto/ccp/ccp-dev-v3.c 		list_del(&cmd->entry);
entry             562 drivers/crypto/ccp/ccp-dev-v3.c 		cmd = list_first_entry(&ccp->backlog, struct ccp_cmd, entry);
entry             563 drivers/crypto/ccp/ccp-dev-v3.c 		list_del(&cmd->entry);
entry            1042 drivers/crypto/ccp/ccp-dev-v5.c 		cmd = list_first_entry(&ccp->cmd, struct ccp_cmd, entry);
entry            1043 drivers/crypto/ccp/ccp-dev-v5.c 		list_del(&cmd->entry);
entry            1048 drivers/crypto/ccp/ccp-dev-v5.c 		cmd = list_first_entry(&ccp->backlog, struct ccp_cmd, entry);
entry            1049 drivers/crypto/ccp/ccp-dev-v5.c 		list_del(&cmd->entry);
entry             140 drivers/crypto/ccp/ccp-dev.c 	list_add_tail(&ccp->entry, &ccp_units);
entry             168 drivers/crypto/ccp/ccp-dev.c 		if (list_is_last(&ccp_rr->entry, &ccp_units))
entry             170 drivers/crypto/ccp/ccp-dev.c 						  entry);
entry             172 drivers/crypto/ccp/ccp-dev.c 			ccp_rr = list_next_entry(ccp_rr, entry);
entry             174 drivers/crypto/ccp/ccp-dev.c 	list_del(&ccp->entry);
entry             215 drivers/crypto/ccp/ccp-dev.c 		if (list_is_last(&ccp_rr->entry, &ccp_units))
entry             217 drivers/crypto/ccp/ccp-dev.c 						  entry);
entry             219 drivers/crypto/ccp/ccp-dev.c 			ccp_rr = list_next_entry(ccp_rr, entry);
entry             259 drivers/crypto/ccp/ccp-dev.c 		dp = list_first_entry(&ccp_units, struct ccp_device, entry);
entry             315 drivers/crypto/ccp/ccp-dev.c 			list_add_tail(&cmd->entry, &ccp->backlog);
entry             322 drivers/crypto/ccp/ccp-dev.c 		list_add_tail(&cmd->entry, &ccp->cmd);
entry             357 drivers/crypto/ccp/ccp-dev.c 	list_add_tail(&cmd->entry, &ccp->cmd);
entry             397 drivers/crypto/ccp/ccp-dev.c 		cmd = list_first_entry(&ccp->cmd, struct ccp_cmd, entry);
entry             398 drivers/crypto/ccp/ccp-dev.c 		list_del(&cmd->entry);
entry             405 drivers/crypto/ccp/ccp-dev.c 					   entry);
entry             406 drivers/crypto/ccp/ccp-dev.c 		list_del(&backlog->entry);
entry             227 drivers/crypto/ccp/ccp-dev.h 	struct list_head entry;
entry             233 drivers/crypto/ccp/ccp-dev.h 	struct list_head entry;
entry             340 drivers/crypto/ccp/ccp-dev.h 	struct list_head entry;
entry              67 drivers/crypto/ccp/ccp-dmaengine.c 	list_for_each_entry_safe(cmd, ctmp, list, entry) {
entry              68 drivers/crypto/ccp/ccp-dmaengine.c 		list_del(&cmd->entry);
entry              78 drivers/crypto/ccp/ccp-dmaengine.c 	list_for_each_entry_safe(desc, dtmp, list, entry) {
entry              82 drivers/crypto/ccp/ccp-dmaengine.c 		list_del(&desc->entry);
entry             110 drivers/crypto/ccp/ccp-dmaengine.c 	list_for_each_entry_safe_reverse(desc, dtmp, list, entry) {
entry             119 drivers/crypto/ccp/ccp-dmaengine.c 		list_del(&desc->entry);
entry             144 drivers/crypto/ccp/ccp-dmaengine.c 	cmd = list_first_entry(&desc->pending, struct ccp_dma_cmd, entry);
entry             145 drivers/crypto/ccp/ccp-dmaengine.c 	list_move(&cmd->entry, &desc->active);
entry             165 drivers/crypto/ccp/ccp-dmaengine.c 				       entry);
entry             172 drivers/crypto/ccp/ccp-dmaengine.c 	list_del(&cmd->entry);
entry             181 drivers/crypto/ccp/ccp-dmaengine.c 		list_move(&desc->entry, &chan->complete);
entry             185 drivers/crypto/ccp/ccp-dmaengine.c 					entry);
entry             253 drivers/crypto/ccp/ccp-dmaengine.c 		? list_first_entry(&chan->pending, struct ccp_dma_desc, entry)
entry             310 drivers/crypto/ccp/ccp-dmaengine.c 	list_del(&desc->entry);
entry             311 drivers/crypto/ccp/ccp-dmaengine.c 	list_add_tail(&desc->entry, &chan->pending);
entry             345 drivers/crypto/ccp/ccp-dmaengine.c 	INIT_LIST_HEAD(&desc->entry);
entry             441 drivers/crypto/ccp/ccp-dmaengine.c 		list_add_tail(&cmd->entry, &desc->pending);
entry             466 drivers/crypto/ccp/ccp-dmaengine.c 	list_add_tail(&desc->entry, &chan->created);
entry             561 drivers/crypto/ccp/ccp-dmaengine.c 		list_for_each_entry(desc, &chan->complete, entry) {
entry             600 drivers/crypto/ccp/ccp-dmaengine.c 					entry);
entry              44 drivers/crypto/ccp/psp-dev.h 	struct list_head entry;
entry              48 drivers/crypto/ccp/sp-dev.c 	list_add_tail(&sp->entry, &sp_units);
entry              59 drivers/crypto/ccp/sp-dev.c 	list_del(&sp->entry);
entry             251 drivers/crypto/ccp/sp-dev.c 	list_for_each_entry(i, &sp_units, entry) {
entry              60 drivers/crypto/ccp/sp-dev.h 	struct list_head entry;
entry              43 drivers/crypto/ccp/sp-pci.c 		sp_pci->msix_entry[v].entry = v;
entry            2660 drivers/crypto/ccree/cc_aead.c 					 entry) {
entry            2662 drivers/crypto/ccree/cc_aead.c 			list_del(&t_alg->entry);
entry            2718 drivers/crypto/ccree/cc_aead.c 			list_add_tail(&t_alg->entry, &aead_handle->aead_list);
entry              33 drivers/crypto/ccree/cc_buffer_mgr.c 	union buffer_array_entry entry[MAX_NUM_OF_BUFFERS_IN_MLLI];
entry             213 drivers/crypto/ccree/cc_buffer_mgr.c 		union buffer_array_entry *entry = &sg_data->entry[i];
entry             218 drivers/crypto/ccree/cc_buffer_mgr.c 			rc = cc_render_sg_to_mlli(dev, entry->sgl, tot_len,
entry             222 drivers/crypto/ccree/cc_buffer_mgr.c 			rc = cc_render_buff_to_mlli(dev, entry->buffer_dma,
entry             260 drivers/crypto/ccree/cc_buffer_mgr.c 	sgl_data->entry[index].buffer_dma = buffer_dma;
entry             281 drivers/crypto/ccree/cc_buffer_mgr.c 	sgl_data->entry[index].sgl = sgl;
entry            1677 drivers/crypto/ccree/cc_cipher.c 					 entry) {
entry            1679 drivers/crypto/ccree/cc_cipher.c 			list_del(&t_alg->entry);
entry            1733 drivers/crypto/ccree/cc_cipher.c 			list_add_tail(&t_alg->entry,
entry             168 drivers/crypto/ccree/cc_driver.h 	struct list_head entry;
entry              61 drivers/crypto/ccree/cc_hash.c 	struct list_head entry;
entry            2068 drivers/crypto/ccree/cc_hash.c 				list_add_tail(&t_alg->entry,
entry            2093 drivers/crypto/ccree/cc_hash.c 			list_add_tail(&t_alg->entry, &hash_handle->hash_list);
entry            2112 drivers/crypto/ccree/cc_hash.c 					 &hash_handle->hash_list, entry) {
entry            2114 drivers/crypto/ccree/cc_hash.c 			list_del(&t_hash_alg->entry);
entry              89 drivers/crypto/chelsio/chcr_core.c 		if (list_is_last(&drv_data.last_dev->entry, &drv_data.act_dev))
entry              91 drivers/crypto/chelsio/chcr_core.c 						  struct uld_ctx, entry);
entry              94 drivers/crypto/chelsio/chcr_core.c 				list_next_entry(drv_data.last_dev, entry);
entry             108 drivers/crypto/chelsio/chcr_core.c 	list_move(&u_ctx->entry, &drv_data.act_dev);
entry             127 drivers/crypto/chelsio/chcr_core.c 	list_add_tail(&u_ctx->entry, &drv_data.inact_dev);
entry             137 drivers/crypto/chelsio/chcr_core.c 		if (list_is_last(&drv_data.last_dev->entry, &drv_data.act_dev))
entry             139 drivers/crypto/chelsio/chcr_core.c 						  struct uld_ctx, entry);
entry             142 drivers/crypto/chelsio/chcr_core.c 				list_next_entry(drv_data.last_dev, entry);
entry             144 drivers/crypto/chelsio/chcr_core.c 	list_move(&u_ctx->entry, &drv_data.inact_dev);
entry             306 drivers/crypto/chelsio/chcr_core.c 	list_for_each_entry_safe(u_ctx, tmp, &drv_data.act_dev, entry) {
entry             307 drivers/crypto/chelsio/chcr_core.c 		list_del(&u_ctx->entry);
entry             310 drivers/crypto/chelsio/chcr_core.c 	list_for_each_entry_safe(u_ctx, tmp, &drv_data.inact_dev, entry) {
entry             311 drivers/crypto/chelsio/chcr_core.c 		list_del(&u_ctx->entry);
entry             154 drivers/crypto/chelsio/chcr_core.h 	struct list_head entry;
entry             598 drivers/crypto/hifn_795x.c 	struct list_head	entry;
entry            2406 drivers/crypto/hifn_795x.c 	list_add_tail(&alg->entry, &dev->alg_list);
entry            2410 drivers/crypto/hifn_795x.c 		list_del(&alg->entry);
entry            2421 drivers/crypto/hifn_795x.c 	list_for_each_entry_safe(a, n, &dev->alg_list, entry) {
entry            2422 drivers/crypto/hifn_795x.c 		list_del(&a->entry);
entry              92 drivers/crypto/n2_core.c 	struct list_head	entry;
entry             249 drivers/crypto/n2_core.c 	struct list_head	entry;
entry             674 drivers/crypto/n2_core.c 	struct list_head	entry;
entry             712 drivers/crypto/n2_core.c 	struct list_head	entry;
entry             898 drivers/crypto/n2_core.c 	INIT_LIST_HEAD(&chunk->entry);
entry             927 drivers/crypto/n2_core.c 				list_add_tail(&chunk->entry,
entry             934 drivers/crypto/n2_core.c 				INIT_LIST_HEAD(&chunk->entry);
entry             955 drivers/crypto/n2_core.c 		list_add_tail(&chunk->entry, &rctx->chunk_list);
entry             970 drivers/crypto/n2_core.c 	list_for_each_entry_safe(c, tmp, &rctx->chunk_list, entry) {
entry             971 drivers/crypto/n2_core.c 		list_del(&c->entry);
entry             997 drivers/crypto/n2_core.c 	list_for_each_entry_safe(c, tmp, &rctx->chunk_list, entry) {
entry            1001 drivers/crypto/n2_core.c 		list_del(&c->entry);
entry            1055 drivers/crypto/n2_core.c 					 entry) {
entry            1061 drivers/crypto/n2_core.c 			list_del(&c->entry);
entry            1068 drivers/crypto/n2_core.c 						 entry) {
entry            1090 drivers/crypto/n2_core.c 			list_del(&c->entry);
entry            1351 drivers/crypto/n2_core.c 	list_for_each_entry_safe(cipher, cipher_tmp, &cipher_algs, entry) {
entry            1353 drivers/crypto/n2_core.c 		list_del(&cipher->entry);
entry            1356 drivers/crypto/n2_core.c 	list_for_each_entry_safe(hmac, hmac_tmp, &hmac_algs, derived.entry) {
entry            1358 drivers/crypto/n2_core.c 		list_del(&hmac->derived.entry);
entry            1361 drivers/crypto/n2_core.c 	list_for_each_entry_safe(alg, alg_tmp, &ahash_algs, entry) {
entry            1363 drivers/crypto/n2_core.c 		list_del(&alg->entry);
entry            1398 drivers/crypto/n2_core.c 	list_add(&p->entry, &cipher_algs);
entry            1402 drivers/crypto/n2_core.c 		list_del(&p->entry);
entry            1422 drivers/crypto/n2_core.c 	INIT_LIST_HEAD(&p->derived.entry);
entry            1436 drivers/crypto/n2_core.c 	list_add(&p->derived.entry, &hmac_algs);
entry            1440 drivers/crypto/n2_core.c 		list_del(&p->derived.entry);
entry            1490 drivers/crypto/n2_core.c 	list_add(&p->entry, &ahash_algs);
entry            1494 drivers/crypto/n2_core.c 		list_del(&p->entry);
entry             173 drivers/crypto/nx/nx-842-pseries.c 	struct nx842_slentry *entry;
entry             177 drivers/crypto/nx/nx-842-pseries.c 	entry = sl->entries;
entry             179 drivers/crypto/nx/nx-842-pseries.c 		entry->ptr = cpu_to_be64(nx842_get_pa((void *)buf));
entry             182 drivers/crypto/nx/nx-842-pseries.c 		entry->len = cpu_to_be64(entrylen);
entry             188 drivers/crypto/nx/nx-842-pseries.c 		entry++;
entry              96 drivers/crypto/picoxcell_crypto.c 	struct list_head		entry;
entry             139 drivers/crypto/picoxcell_crypto.c 	struct list_head		entry;
entry            1737 drivers/crypto/picoxcell_crypto.c 			list_add_tail(&engine->algs[i].entry,
entry            1754 drivers/crypto/picoxcell_crypto.c 			list_add_tail(&engine->aeads[i].entry,
entry            1788 drivers/crypto/picoxcell_crypto.c 	list_for_each_entry_safe(aead, an, &engine->registered_aeads, entry) {
entry            1789 drivers/crypto/picoxcell_crypto.c 		list_del(&aead->entry);
entry            1793 drivers/crypto/picoxcell_crypto.c 	list_for_each_entry_safe(alg, next, &engine->registered_algs, entry) {
entry            1794 drivers/crypto/picoxcell_crypto.c 		list_del(&alg->entry);
entry              74 drivers/crypto/qat/qat_common/adf_isr.c 			pci_dev_info->msix_entries.entries[i].entry = i;
entry              76 drivers/crypto/qat/qat_common/adf_isr.c 		pci_dev_info->msix_entries.entries[0].entry =
entry             391 drivers/crypto/qce/ablkcipher.c 	INIT_LIST_HEAD(&tmpl->entry);
entry             403 drivers/crypto/qce/ablkcipher.c 	list_add_tail(&tmpl->entry, &ablkcipher_algs);
entry             412 drivers/crypto/qce/ablkcipher.c 	list_for_each_entry_safe(tmpl, n, &ablkcipher_algs, entry) {
entry             414 drivers/crypto/qce/ablkcipher.c 		list_del(&tmpl->entry);
entry              77 drivers/crypto/qce/common.h 	struct list_head entry;
entry             508 drivers/crypto/qce/sha.c 	INIT_LIST_HEAD(&tmpl->entry);
entry             520 drivers/crypto/qce/sha.c 	list_add_tail(&tmpl->entry, &ahash_algs);
entry             529 drivers/crypto/qce/sha.c 	list_for_each_entry_safe(tmpl, n, &ahash_algs, entry) {
entry             531 drivers/crypto/qce/sha.c 		list_del(&tmpl->entry);
entry            3008 drivers/crypto/talitos.c 	struct list_head entry;
entry            3113 drivers/crypto/talitos.c 	list_for_each_entry_safe(t_alg, n, &priv->alg_list, entry) {
entry            3124 drivers/crypto/talitos.c 		list_del(&t_alg->entry);
entry            3487 drivers/crypto/talitos.c 				list_add_tail(&t_alg->entry, &priv->alg_list);
entry             192 drivers/dma/ioat/dca.c 	u8 entry;
entry             198 drivers/dma/ioat/dca.c 		entry = ioatdca->tag_map[i];
entry             199 drivers/dma/ioat/dca.c 		if (entry & DCA3_TAG_MAP_BIT_TO_SEL) {
entry             200 drivers/dma/ioat/dca.c 			bit = entry &
entry             203 drivers/dma/ioat/dca.c 		} else if (entry & DCA3_TAG_MAP_BIT_TO_INV) {
entry             204 drivers/dma/ioat/dca.c 			bit = entry & ~DCA3_TAG_MAP_BIT_TO_INV;
entry             207 drivers/dma/ioat/dca.c 			value = (entry & DCA3_TAG_MAP_LITERAL_VAL) ? 1 : 0;
entry             426 drivers/dma/ioat/init.c 		ioat_dma->msix_entries[i].entry = i;
entry              44 drivers/dma/ioat/sysfs.c 	struct ioat_sysfs_entry *entry;
entry              47 drivers/dma/ioat/sysfs.c 	entry = container_of(attr, struct ioat_sysfs_entry, attr);
entry              50 drivers/dma/ioat/sysfs.c 	if (!entry->show)
entry              52 drivers/dma/ioat/sysfs.c 	return entry->show(&ioat_chan->dma_chan, page);
entry              59 drivers/dma/ioat/sysfs.c 	struct ioat_sysfs_entry *entry;
entry              62 drivers/dma/ioat/sysfs.c 	entry = container_of(attr, struct ioat_sysfs_entry, attr);
entry              65 drivers/dma/ioat/sysfs.c 	if (!entry->store)
entry              67 drivers/dma/ioat/sysfs.c 	return entry->store(&ioat_chan->dma_chan, page, count);
entry             108 drivers/edac/ghes_edac.c 		struct memdev_dmi_entry *entry = (struct memdev_dmi_entry *)dh;
entry             114 drivers/edac/ghes_edac.c 		if (entry->size == 0xffff) {
entry             118 drivers/edac/ghes_edac.c 		} else if (entry->size == 0x7fff) {
entry             119 drivers/edac/ghes_edac.c 			dimm->nr_pages = MiB_TO_PAGES(entry->extended_size);
entry             121 drivers/edac/ghes_edac.c 			if (entry->size & BIT(15))
entry             122 drivers/edac/ghes_edac.c 				dimm->nr_pages = MiB_TO_PAGES((entry->size & 0x7fff) << 10);
entry             124 drivers/edac/ghes_edac.c 				dimm->nr_pages = MiB_TO_PAGES(entry->size);
entry             127 drivers/edac/ghes_edac.c 		switch (entry->memory_type) {
entry             129 drivers/edac/ghes_edac.c 			if (entry->type_detail & BIT(13))
entry             135 drivers/edac/ghes_edac.c 			if (entry->type_detail & BIT(13))
entry             144 drivers/edac/ghes_edac.c 			if (entry->type_detail & BIT(12))
entry             146 drivers/edac/ghes_edac.c 			else if (entry->type_detail & BIT(13))
entry             152 drivers/edac/ghes_edac.c 			if (entry->type_detail & BIT(12))
entry             154 drivers/edac/ghes_edac.c 			else if (entry->type_detail & BIT(13))
entry             160 drivers/edac/ghes_edac.c 			if (entry->type_detail & BIT(6))
entry             162 drivers/edac/ghes_edac.c 			else if ((entry->type_detail & rdr_mask) == rdr_mask)
entry             164 drivers/edac/ghes_edac.c 			else if (entry->type_detail & BIT(7))
entry             166 drivers/edac/ghes_edac.c 			else if (entry->type_detail & BIT(9))
entry             176 drivers/edac/ghes_edac.c 		if (entry->total_width == entry->data_width)
entry             194 drivers/edac/ghes_edac.c 				entry->memory_type, entry->type_detail,
entry             195 drivers/edac/ghes_edac.c 				entry->total_width, entry->data_width);
entry             198 drivers/edac/ghes_edac.c 		dimm->smbios_handle = entry->handle;
entry            1148 drivers/edac/sb_edac.c 		const int entry,
entry            1171 drivers/edac/sb_edac.c 			knl_tad_dram_limit_lo[entry], &reg_limit_lo);
entry            1173 drivers/edac/sb_edac.c 			knl_tad_dram_offset_lo[entry], &reg_offset_lo);
entry            1175 drivers/edac/sb_edac.c 			knl_tad_dram_hi[entry], &reg_hi);
entry            1228 drivers/edac/sb_edac.c static u32 knl_get_edc_route(int entry, u32 reg)
entry            1230 drivers/edac/sb_edac.c 	WARN_ON(entry >= KNL_MAX_EDCS);
entry            1231 drivers/edac/sb_edac.c 	return GET_BITFIELD(reg, entry*3, (entry*3)+2);
entry            1251 drivers/edac/sb_edac.c static u32 knl_get_mc_route(int entry, u32 reg)
entry            1255 drivers/edac/sb_edac.c 	WARN_ON(entry >= KNL_MAX_CHANNELS);
entry            1257 drivers/edac/sb_edac.c 	mc = GET_BITFIELD(reg, entry*3, (entry*3)+2);
entry            1258 drivers/edac/sb_edac.c 	chan = GET_BITFIELD(reg, (entry*2) + 18, (entry*2) + 18 + 1);
entry             723 drivers/edac/thunderx_edac.c 	lmc->msix_ent.entry = 0;
entry            1079 drivers/edac/thunderx_edac.c 						msix_ent[msix->entry]);
entry            1108 drivers/edac/thunderx_edac.c 						msix_ent[msix->entry]);
entry            1172 drivers/edac/thunderx_edac.c 						msix_ent[msix->entry]);
entry            1177 drivers/edac/thunderx_edac.c 	ctx->link = msix->entry;
entry            1191 drivers/edac/thunderx_edac.c 						msix_ent[msix->entry]);
entry            1391 drivers/edac/thunderx_edac.c 		ocx->msix_ent[i].entry = i;
entry            2029 drivers/edac/thunderx_edac.c 	l2c->msix_ent.entry = 0;
entry             875 drivers/extcon/extcon.c 	list_for_each_entry(sd, &extcon_dev_list, entry) {
entry            1256 drivers/extcon/extcon.c 	list_add(&edev->entry, &extcon_dev_list);
entry            1297 drivers/extcon/extcon.c 	list_del(&edev->entry);
entry            1341 drivers/extcon/extcon.c 	list_for_each_entry(edev, &extcon_dev_list, entry)
entry              51 drivers/extcon/extcon.h 	struct list_head entry;
entry            1791 drivers/firewire/ohci.c 	u32 entry;
entry            1794 drivers/firewire/ohci.c 		entry = ohci->self_id_buffer[i];
entry            1795 drivers/firewire/ohci.c 		if ((self_id & 0xff000000) == (entry & 0xff000000))
entry            1797 drivers/firewire/ohci.c 		if ((self_id & 0xff000000) < (entry & 0xff000000))
entry              54 drivers/firmware/dmi-sysfs.c 	ssize_t (*show)(struct dmi_sysfs_entry *entry, char *buf);
entry              69 drivers/firmware/dmi-sysfs.c 	ssize_t (*show)(struct dmi_sysfs_entry *entry,
entry             101 drivers/firmware/dmi-sysfs.c 	struct dmi_sysfs_entry *entry = to_entry(kobj);
entry             108 drivers/firmware/dmi-sysfs.c 	return attr->show(entry, buf);
entry             119 drivers/firmware/dmi-sysfs.c 	struct dmi_sysfs_entry	*entry;
entry             130 drivers/firmware/dmi-sysfs.c 	struct dmi_sysfs_entry *entry = data->entry;
entry             133 drivers/firmware/dmi-sysfs.c 	if (dh->type != entry->dh.type)
entry             149 drivers/firmware/dmi-sysfs.c 	data->ret = data->callback(entry, dh, data->private);
entry             159 drivers/firmware/dmi-sysfs.c static ssize_t find_dmi_entry(struct dmi_sysfs_entry *entry,
entry             163 drivers/firmware/dmi-sysfs.c 		.entry = entry,
entry             166 drivers/firmware/dmi-sysfs.c 		.instance_countdown = entry->instance,
entry             203 drivers/firmware/dmi-sysfs.c static ssize_t dmi_entry_attr_show_helper(struct dmi_sysfs_entry *entry,
entry             212 drivers/firmware/dmi-sysfs.c 	return attr->show(entry, dh, data->buf);
entry             269 drivers/firmware/dmi-sysfs.c static ssize_t dmi_sysfs_sel_##_field(struct dmi_sysfs_entry *entry, \
entry             361 drivers/firmware/dmi-sysfs.c static ssize_t dmi_sel_raw_read_io(struct dmi_sysfs_entry *entry,
entry             378 drivers/firmware/dmi-sysfs.c static ssize_t dmi_sel_raw_read_phys32(struct dmi_sysfs_entry *entry,
entry             399 drivers/firmware/dmi-sysfs.c static ssize_t dmi_sel_raw_read_helper(struct dmi_sysfs_entry *entry,
entry             415 drivers/firmware/dmi-sysfs.c 		return dmi_sel_raw_read_io(entry, &sel, state->buf,
entry             418 drivers/firmware/dmi-sysfs.c 		return dmi_sel_raw_read_phys32(entry, &sel, state->buf,
entry             434 drivers/firmware/dmi-sysfs.c 	struct dmi_sysfs_entry *entry = to_entry(kobj->parent);
entry             441 drivers/firmware/dmi-sysfs.c 	return find_dmi_entry(entry, dmi_sel_raw_read_helper, &state);
entry             449 drivers/firmware/dmi-sysfs.c static int dmi_system_event_log(struct dmi_sysfs_entry *entry)
entry             453 drivers/firmware/dmi-sysfs.c 	entry->child = kzalloc(sizeof(*entry->child), GFP_KERNEL);
entry             454 drivers/firmware/dmi-sysfs.c 	if (!entry->child)
entry             456 drivers/firmware/dmi-sysfs.c 	ret = kobject_init_and_add(entry->child,
entry             458 drivers/firmware/dmi-sysfs.c 				   &entry->kobj,
entry             463 drivers/firmware/dmi-sysfs.c 	ret = sysfs_create_bin_file(entry->child, &dmi_sel_raw_attr);
entry             470 drivers/firmware/dmi-sysfs.c 	kobject_del(entry->child);
entry             472 drivers/firmware/dmi-sysfs.c 	kfree(entry->child);
entry             480 drivers/firmware/dmi-sysfs.c static ssize_t dmi_sysfs_entry_length(struct dmi_sysfs_entry *entry, char *buf)
entry             482 drivers/firmware/dmi-sysfs.c 	return sprintf(buf, "%d\n", entry->dh.length);
entry             485 drivers/firmware/dmi-sysfs.c static ssize_t dmi_sysfs_entry_handle(struct dmi_sysfs_entry *entry, char *buf)
entry             487 drivers/firmware/dmi-sysfs.c 	return sprintf(buf, "%d\n", entry->dh.handle);
entry             490 drivers/firmware/dmi-sysfs.c static ssize_t dmi_sysfs_entry_type(struct dmi_sysfs_entry *entry, char *buf)
entry             492 drivers/firmware/dmi-sysfs.c 	return sprintf(buf, "%d\n", entry->dh.type);
entry             495 drivers/firmware/dmi-sysfs.c static ssize_t dmi_sysfs_entry_instance(struct dmi_sysfs_entry *entry,
entry             498 drivers/firmware/dmi-sysfs.c 	return sprintf(buf, "%d\n", entry->instance);
entry             501 drivers/firmware/dmi-sysfs.c static ssize_t dmi_sysfs_entry_position(struct dmi_sysfs_entry *entry,
entry             504 drivers/firmware/dmi-sysfs.c 	return sprintf(buf, "%d\n", entry->position);
entry             507 drivers/firmware/dmi-sysfs.c static DMI_SYSFS_ATTR(entry, length);
entry             508 drivers/firmware/dmi-sysfs.c static DMI_SYSFS_ATTR(entry, handle);
entry             509 drivers/firmware/dmi-sysfs.c static DMI_SYSFS_ATTR(entry, type);
entry             510 drivers/firmware/dmi-sysfs.c static DMI_SYSFS_ATTR(entry, instance);
entry             511 drivers/firmware/dmi-sysfs.c static DMI_SYSFS_ATTR(entry, position);
entry             522 drivers/firmware/dmi-sysfs.c static ssize_t dmi_entry_raw_read_helper(struct dmi_sysfs_entry *entry,
entry             540 drivers/firmware/dmi-sysfs.c 	struct dmi_sysfs_entry *entry = to_entry(kobj);
entry             547 drivers/firmware/dmi-sysfs.c 	return find_dmi_entry(entry, dmi_entry_raw_read_helper, &state);
entry             557 drivers/firmware/dmi-sysfs.c 	struct dmi_sysfs_entry *entry = to_entry(kobj);
entry             560 drivers/firmware/dmi-sysfs.c 	list_del(&entry->list);
entry             562 drivers/firmware/dmi-sysfs.c 	kfree(entry);
entry             582 drivers/firmware/dmi-sysfs.c 	struct dmi_sysfs_entry *entry;
entry             590 drivers/firmware/dmi-sysfs.c 	entry = kzalloc(sizeof(*entry), GFP_KERNEL);
entry             591 drivers/firmware/dmi-sysfs.c 	if (!entry) {
entry             597 drivers/firmware/dmi-sysfs.c 	memcpy(&entry->dh, dh, sizeof(*dh));
entry             598 drivers/firmware/dmi-sysfs.c 	entry->instance = instance_counts[dh->type]++;
entry             599 drivers/firmware/dmi-sysfs.c 	entry->position = position_count++;
entry             601 drivers/firmware/dmi-sysfs.c 	entry->kobj.kset = dmi_kset;
entry             602 drivers/firmware/dmi-sysfs.c 	*ret = kobject_init_and_add(&entry->kobj, &dmi_sysfs_entry_ktype, NULL,
entry             603 drivers/firmware/dmi-sysfs.c 				    "%d-%d", dh->type, entry->instance);
entry             606 drivers/firmware/dmi-sysfs.c 		kfree(entry);
entry             612 drivers/firmware/dmi-sysfs.c 	list_add_tail(&entry->list, &entry_list);
entry             618 drivers/firmware/dmi-sysfs.c 		*ret = dmi_system_event_log(entry);
entry             628 drivers/firmware/dmi-sysfs.c 	*ret = sysfs_create_bin_file(&entry->kobj, &dmi_entry_raw_attr);
entry             634 drivers/firmware/dmi-sysfs.c 	kobject_put(entry->child);
entry             635 drivers/firmware/dmi-sysfs.c 	kobject_put(&entry->kobj);
entry             641 drivers/firmware/dmi-sysfs.c 	struct dmi_sysfs_entry *entry, *next;
entry             644 drivers/firmware/dmi-sysfs.c 	list_for_each_entry_safe(entry, next, &entry_list, list) {
entry             645 drivers/firmware/dmi-sysfs.c 		kobject_put(entry->child);
entry             646 drivers/firmware/dmi-sysfs.c 		kobject_put(&entry->kobj);
entry              50 drivers/firmware/efi/apple-properties.c 					     struct property_entry entry[])
entry              88 drivers/firmware/efi/apple-properties.c 		entry[i].name = key;
entry              89 drivers/firmware/efi/apple-properties.c 		entry[i].length = val_len - sizeof(val_len);
entry              90 drivers/firmware/efi/apple-properties.c 		entry[i].is_array = !!entry[i].length;
entry              91 drivers/firmware/efi/apple-properties.c 		entry[i].type = DEV_PROP_U8;
entry              92 drivers/firmware/efi/apple-properties.c 		entry[i].pointer.u8_data = ptr + key_len + sizeof(val_len);
entry              95 drivers/firmware/efi/apple-properties.c 			dev_info(dev, "property: %s\n", entry[i].name);
entry              97 drivers/firmware/efi/apple-properties.c 				16, 1, entry[i].pointer.u8_data,
entry              98 drivers/firmware/efi/apple-properties.c 				entry[i].length, true);
entry             121 drivers/firmware/efi/apple-properties.c 		struct property_entry *entry = NULL;
entry             146 drivers/firmware/efi/apple-properties.c 		entry = kcalloc(dev_header->prop_count + 1, sizeof(*entry),
entry             148 drivers/firmware/efi/apple-properties.c 		if (!entry) {
entry             153 drivers/firmware/efi/apple-properties.c 		unmarshal_key_value_pairs(dev_header, dev, ptr, entry);
entry             154 drivers/firmware/efi/apple-properties.c 		if (!entry[0].name)
entry             157 drivers/firmware/efi/apple-properties.c 		ret = device_add_properties(dev, entry); /* makes deep copy */
entry             161 drivers/firmware/efi/apple-properties.c 		for (i = 0; entry[i].name; i++)
entry             162 drivers/firmware/efi/apple-properties.c 			kfree(entry[i].name);
entry             165 drivers/firmware/efi/apple-properties.c 		kfree(entry);
entry              38 drivers/firmware/efi/efi-pstore.c static int efi_pstore_read_func(struct efivar_entry *entry,
entry              49 drivers/firmware/efi/efi-pstore.c 	if (efi_guidcmp(entry->var.VendorGuid, vendor))
entry              53 drivers/firmware/efi/efi-pstore.c 		name[i] = entry->var.VariableName[i];
entry              93 drivers/firmware/efi/efi-pstore.c 	entry->var.DataSize = 1024;
entry              94 drivers/firmware/efi/efi-pstore.c 	__efivar_entry_get(entry, &entry->var.Attributes,
entry              95 drivers/firmware/efi/efi-pstore.c 			   &entry->var.DataSize, entry->var.Data);
entry              96 drivers/firmware/efi/efi-pstore.c 	size = entry->var.DataSize;
entry              97 drivers/firmware/efi/efi-pstore.c 	memcpy(record->buf, entry->var.Data,
entry             123 drivers/firmware/efi/efi-pstore.c static inline int __efi_pstore_scan_sysfs_exit(struct efivar_entry *entry,
entry             126 drivers/firmware/efi/efi-pstore.c 	if (entry->deleting) {
entry             127 drivers/firmware/efi/efi-pstore.c 		list_del(&entry->list);
entry             129 drivers/firmware/efi/efi-pstore.c 		efivar_unregister(entry);
entry             133 drivers/firmware/efi/efi-pstore.c 		entry->scanning = false;
entry             171 drivers/firmware/efi/efi-pstore.c 	struct efivar_entry *entry, *n;
entry             177 drivers/firmware/efi/efi-pstore.c 		list_for_each_entry_safe(entry, n, head, list) {
entry             178 drivers/firmware/efi/efi-pstore.c 			efi_pstore_scan_sysfs_enter(entry, n, head);
entry             180 drivers/firmware/efi/efi-pstore.c 			size = efi_pstore_read_func(entry, record);
entry             181 drivers/firmware/efi/efi-pstore.c 			ret = efi_pstore_scan_sysfs_exit(entry, n, head,
entry             275 drivers/firmware/efi/efi-pstore.c static int efi_pstore_erase_func(struct efivar_entry *entry, void *data)
entry             281 drivers/firmware/efi/efi-pstore.c 	if (efi_guidcmp(entry->var.VendorGuid, vendor))
entry             284 drivers/firmware/efi/efi-pstore.c 	if (ucs2_strncmp(entry->var.VariableName, efi_name, (size_t)ucs2_len))
entry             287 drivers/firmware/efi/efi-pstore.c 	if (entry->scanning) {
entry             292 drivers/firmware/efi/efi-pstore.c 		entry->deleting = true;
entry             294 drivers/firmware/efi/efi-pstore.c 		list_del(&entry->list);
entry             297 drivers/firmware/efi/efi-pstore.c 	__efivar_entry_delete(entry);
entry             304 drivers/firmware/efi/efi-pstore.c 	struct efivar_entry *entry = NULL;
entry             318 drivers/firmware/efi/efi-pstore.c 				    efi_name, &entry);
entry             321 drivers/firmware/efi/efi-pstore.c 	if (found && !entry->scanning)
entry             322 drivers/firmware/efi/efi-pstore.c 		efivar_unregister(entry);
entry             241 drivers/firmware/efi/efi.c 	struct efivar_entry *entry;
entry             250 drivers/firmware/efi/efi.c 	entry = kmalloc(sizeof(*entry), GFP_KERNEL);
entry             251 drivers/firmware/efi/efi.c 	if (!entry)
entry             254 drivers/firmware/efi/efi.c 	memcpy(entry->var.VariableName, name, name_size);
entry             255 drivers/firmware/efi/efi.c 	memcpy(&entry->var.VendorGuid, &vendor, sizeof(efi_guid_t));
entry             257 drivers/firmware/efi/efi.c 	efivar_entry_add(entry, list);
entry             265 drivers/firmware/efi/efi.c 	struct efivar_entry *entry, *aux;
entry             275 drivers/firmware/efi/efi.c 	list_for_each_entry_safe(entry, aux, &entries, list) {
entry             277 drivers/firmware/efi/efi.c 			&entry->var.VendorGuid);
entry             279 drivers/firmware/efi/efi.c 		list_del(&entry->list);
entry             281 drivers/firmware/efi/efi.c 		ret = efivar_entry_size(entry, &size);
entry             293 drivers/firmware/efi/efi.c 		ret = efivar_entry_get(entry, NULL, &size, data);
entry             311 drivers/firmware/efi/efi.c 		kfree(entry);
entry             613 drivers/firmware/efi/efi.c 				memblock_reserve(rsv->entry[i].base,
entry             614 drivers/firmware/efi/efi.c 						 rsv->entry[i].size);
entry            1011 drivers/firmware/efi/efi.c 			rsv->entry[index].base = addr;
entry            1012 drivers/firmware/efi/efi.c 			rsv->entry[index].size = size;
entry            1039 drivers/firmware/efi/efi.c 	rsv->entry[0].base = addr;
entry            1040 drivers/firmware/efi/efi.c 	rsv->entry[0].size = size;
entry              28 drivers/firmware/efi/efibc.c 	struct efivar_entry *entry;
entry              31 drivers/firmware/efi/efibc.c 	if (size > sizeof(entry->var.Data)) {
entry              36 drivers/firmware/efi/efibc.c 	entry = kmalloc(sizeof(*entry), GFP_KERNEL);
entry              37 drivers/firmware/efi/efibc.c 	if (!entry) {
entry              42 drivers/firmware/efi/efibc.c 	efibc_str_to_str16(name, entry->var.VariableName);
entry              43 drivers/firmware/efi/efibc.c 	efibc_str_to_str16(value, (efi_char16_t *)entry->var.Data);
entry              44 drivers/firmware/efi/efibc.c 	memcpy(&entry->var.VendorGuid, &guid, sizeof(guid));
entry              46 drivers/firmware/efi/efibc.c 	ret = efivar_entry_set_safe(entry->var.VariableName,
entry              47 drivers/firmware/efi/efibc.c 				    entry->var.VendorGuid,
entry              51 drivers/firmware/efi/efibc.c 				    false, size, entry->var.Data);
entry              57 drivers/firmware/efi/efibc.c 	kfree(entry);
entry              46 drivers/firmware/efi/efivars.c 	ssize_t (*show) (struct efivar_entry *entry, char *buf);
entry              47 drivers/firmware/efi/efivars.c 	ssize_t (*store)(struct efivar_entry *entry, const char *buf, size_t count);
entry              67 drivers/firmware/efi/efivars.c efivar_guid_read(struct efivar_entry *entry, char *buf)
entry              69 drivers/firmware/efi/efivars.c 	struct efi_variable *var = &entry->var;
entry              72 drivers/firmware/efi/efivars.c 	if (!entry || !buf)
entry              83 drivers/firmware/efi/efivars.c efivar_attr_read(struct efivar_entry *entry, char *buf)
entry              85 drivers/firmware/efi/efivars.c 	struct efi_variable *var = &entry->var;
entry              90 drivers/firmware/efi/efivars.c 	if (!entry || !buf)
entry              93 drivers/firmware/efi/efivars.c 	ret = efivar_entry_get(entry, &var->Attributes, &size, var->Data);
entry             119 drivers/firmware/efi/efivars.c efivar_size_read(struct efivar_entry *entry, char *buf)
entry             121 drivers/firmware/efi/efivars.c 	struct efi_variable *var = &entry->var;
entry             126 drivers/firmware/efi/efivars.c 	if (!entry || !buf)
entry             129 drivers/firmware/efi/efivars.c 	ret = efivar_entry_get(entry, &var->Attributes, &size, var->Data);
entry             139 drivers/firmware/efi/efivars.c efivar_data_read(struct efivar_entry *entry, char *buf)
entry             141 drivers/firmware/efi/efivars.c 	struct efi_variable *var = &entry->var;
entry             145 drivers/firmware/efi/efivars.c 	if (!entry || !buf)
entry             148 drivers/firmware/efi/efivars.c 	ret = efivar_entry_get(entry, &var->Attributes, &size, var->Data);
entry             201 drivers/firmware/efi/efivars.c efivar_store_raw(struct efivar_entry *entry, const char *buf, size_t count)
entry             203 drivers/firmware/efi/efivars.c 	struct efi_variable *new_var, *var = &entry->var;
entry             211 drivers/firmware/efi/efivars.c 	if (!entry || !buf)
entry             231 drivers/firmware/efi/efivars.c 		copy_out_compat(&entry->var, compat);
entry             248 drivers/firmware/efi/efivars.c 		memcpy(&entry->var, new_var, count);
entry             251 drivers/firmware/efi/efivars.c 	err = efivar_entry_set(entry, attributes, size, data, NULL);
entry             261 drivers/firmware/efi/efivars.c efivar_show_raw(struct efivar_entry *entry, char *buf)
entry             263 drivers/firmware/efi/efivars.c 	struct efi_variable *var = &entry->var;
entry             269 drivers/firmware/efi/efivars.c 	if (!entry || !buf)
entry             272 drivers/firmware/efi/efivars.c 	ret = efivar_entry_get(entry, &var->Attributes, &datasize, var->Data);
entry             440 drivers/firmware/efi/efivars.c 	struct efivar_entry *entry;
entry             465 drivers/firmware/efi/efivars.c 	entry = efivar_entry_find(name, vendor, &efivar_sysfs_list, true);
entry             466 drivers/firmware/efi/efivars.c 	if (!entry)
entry             468 drivers/firmware/efi/efivars.c 	else if (__efivar_entry_delete(entry))
entry             476 drivers/firmware/efi/efivars.c 	if (!entry->scanning) {
entry             478 drivers/firmware/efi/efivars.c 		efivar_unregister(entry);
entry             597 drivers/firmware/efi/efivars.c 	struct efivar_entry *entry = data;
entry             602 drivers/firmware/efi/efivars.c 	memcpy(entry->var.VariableName, name, name_size);
entry             603 drivers/firmware/efi/efivars.c 	memcpy(&(entry->var.VendorGuid), &vendor, sizeof(efi_guid_t));
entry             610 drivers/firmware/efi/efivars.c 	struct efivar_entry *entry;
entry             615 drivers/firmware/efi/efivars.c 		entry = kzalloc(sizeof(*entry), GFP_KERNEL);
entry             616 drivers/firmware/efi/efivars.c 		if (!entry)
entry             619 drivers/firmware/efi/efivars.c 		err = efivar_init(efivar_update_sysfs_entry, entry,
entry             624 drivers/firmware/efi/efivars.c 		efivar_create_sysfs_entry(entry);
entry             627 drivers/firmware/efi/efivars.c 	kfree(entry);
entry             633 drivers/firmware/efi/efivars.c 	struct efivar_entry *entry;
entry             635 drivers/firmware/efi/efivars.c 	entry = kzalloc(sizeof(*entry), GFP_KERNEL);
entry             636 drivers/firmware/efi/efivars.c 	if (!entry)
entry             639 drivers/firmware/efi/efivars.c 	memcpy(entry->var.VariableName, name, name_size);
entry             640 drivers/firmware/efi/efivars.c 	memcpy(&(entry->var.VendorGuid), &vendor, sizeof(efi_guid_t));
entry             642 drivers/firmware/efi/efivars.c 	efivar_create_sysfs_entry(entry);
entry             647 drivers/firmware/efi/efivars.c static int efivar_sysfs_destroy(struct efivar_entry *entry, void *data)
entry             649 drivers/firmware/efi/efivars.c 	int err = efivar_entry_remove(entry);
entry             653 drivers/firmware/efi/efivars.c 	efivar_unregister(entry);
entry              77 drivers/firmware/efi/esrt.c 	ssize_t (*show)(struct esre_entry *entry, char *buf);
entry              78 drivers/firmware/efi/esrt.c 	ssize_t (*store)(struct esre_entry *entry,
entry              95 drivers/firmware/efi/esrt.c 	struct esre_entry *entry = to_entry(kobj);
entry             102 drivers/firmware/efi/esrt.c 	return attr->show(entry, buf);
entry             110 drivers/firmware/efi/esrt.c static ssize_t fw_class_show(struct esre_entry *entry, char *buf)
entry             114 drivers/firmware/efi/esrt.c 	efi_guid_to_str(&entry->esre.esre1->fw_class, str);
entry             124 drivers/firmware/efi/esrt.c static ssize_t name##_show(struct esre_entry *entry, char *buf) \
entry             127 drivers/firmware/efi/esrt.c 		       le##size##_to_cpu(entry->esre.esre1->name)); \
entry             151 drivers/firmware/efi/esrt.c 	struct esre_entry *entry = to_entry(kobj);
entry             153 drivers/firmware/efi/esrt.c 	list_del(&entry->list);
entry             154 drivers/firmware/efi/esrt.c 	kfree(entry);
entry             169 drivers/firmware/efi/esrt.c 	struct esre_entry *entry;
entry             171 drivers/firmware/efi/esrt.c 	entry = kzalloc(sizeof(*entry), GFP_KERNEL);
entry             172 drivers/firmware/efi/esrt.c 	if (!entry)
entry             175 drivers/firmware/efi/esrt.c 	entry->kobj.kset = esrt_kset;
entry             180 drivers/firmware/efi/esrt.c 		entry->esre.esre1 = esre;
entry             181 drivers/firmware/efi/esrt.c 		rc = kobject_init_and_add(&entry->kobj, &esre1_ktype, NULL,
entry             184 drivers/firmware/efi/esrt.c 			kfree(entry);
entry             189 drivers/firmware/efi/esrt.c 	list_add_tail(&entry->list, &entry_list);
entry             369 drivers/firmware/efi/esrt.c 	struct esre_entry *entry, *next;
entry             371 drivers/firmware/efi/esrt.c 	list_for_each_entry_safe(entry, next, &entry_list, list) {
entry             372 drivers/firmware/efi/esrt.c 		kobject_put(&entry->kobj);
entry             166 drivers/firmware/efi/memattr.c 		valid = entry_is_valid((void *)tbl->entry + i * tbl->desc_size,
entry              25 drivers/firmware/efi/runtime-map.c 	ssize_t (*show)(struct efi_runtime_map_entry *entry, char *buf);
entry              33 drivers/firmware/efi/runtime-map.c static ssize_t type_show(struct efi_runtime_map_entry *entry, char *buf)
entry              35 drivers/firmware/efi/runtime-map.c 	return snprintf(buf, PAGE_SIZE, "0x%x\n", entry->md.type);
entry              38 drivers/firmware/efi/runtime-map.c #define EFI_RUNTIME_FIELD(var) entry->md.var
entry              41 drivers/firmware/efi/runtime-map.c static ssize_t name##_show(struct efi_runtime_map_entry *entry, char *buf) \
entry              59 drivers/firmware/efi/runtime-map.c 	struct efi_runtime_map_entry *entry = to_map_entry(kobj);
entry              62 drivers/firmware/efi/runtime-map.c 	return map_attr->show(entry, buf);
entry              89 drivers/firmware/efi/runtime-map.c 	struct efi_runtime_map_entry *entry;
entry              91 drivers/firmware/efi/runtime-map.c 	entry = to_map_entry(kobj);
entry              92 drivers/firmware/efi/runtime-map.c 	kfree(entry);
entry             108 drivers/firmware/efi/runtime-map.c 	struct efi_runtime_map_entry *entry;
entry             116 drivers/firmware/efi/runtime-map.c 	entry = kzalloc(sizeof(*entry), GFP_KERNEL);
entry             117 drivers/firmware/efi/runtime-map.c 	if (!entry) {
entry             123 drivers/firmware/efi/runtime-map.c 	memcpy(&entry->md, md, sizeof(efi_memory_desc_t));
entry             125 drivers/firmware/efi/runtime-map.c 	kobject_init(&entry->kobj, &map_ktype);
entry             126 drivers/firmware/efi/runtime-map.c 	entry->kobj.kset = map_kset;
entry             127 drivers/firmware/efi/runtime-map.c 	ret = kobject_add(&entry->kobj, NULL, "%d", nr);
entry             129 drivers/firmware/efi/runtime-map.c 		kobject_put(&entry->kobj);
entry             135 drivers/firmware/efi/runtime-map.c 	return entry;
entry             162 drivers/firmware/efi/runtime-map.c 	struct efi_runtime_map_entry *entry;
entry             168 drivers/firmware/efi/runtime-map.c 	map_entries = kcalloc(efi.memmap.nr_map, sizeof(entry), GFP_KERNEL);
entry             176 drivers/firmware/efi/runtime-map.c 		entry = add_sysfs_runtime_map_entry(efi_kobj, i, md);
entry             177 drivers/firmware/efi/runtime-map.c 		if (IS_ERR(entry)) {
entry             178 drivers/firmware/efi/runtime-map.c 			ret = PTR_ERR(entry);
entry             181 drivers/firmware/efi/runtime-map.c 		*(map_entries + i++) = entry;
entry             187 drivers/firmware/efi/runtime-map.c 		entry = *(map_entries + j);
entry             188 drivers/firmware/efi/runtime-map.c 		kobject_put(&entry->kobj);
entry             340 drivers/firmware/efi/vars.c 	struct efivar_entry *entry, *n;
entry             345 drivers/firmware/efi/vars.c 	list_for_each_entry_safe(entry, n, head, list) {
entry             346 drivers/firmware/efi/vars.c 		strsize2 = ucs2_strsize(entry->var.VariableName, 1024);
entry             348 drivers/firmware/efi/vars.c 			!memcmp(variable_name, &(entry->var.VariableName),
entry             350 drivers/firmware/efi/vars.c 			!efi_guidcmp(entry->var.VendorGuid,
entry             526 drivers/firmware/efi/vars.c int efivar_entry_add(struct efivar_entry *entry, struct list_head *head)
entry             530 drivers/firmware/efi/vars.c 	list_add(&entry->list, head);
entry             543 drivers/firmware/efi/vars.c int efivar_entry_remove(struct efivar_entry *entry)
entry             547 drivers/firmware/efi/vars.c 	list_del(&entry->list);
entry             565 drivers/firmware/efi/vars.c static void efivar_entry_list_del_unlock(struct efivar_entry *entry)
entry             567 drivers/firmware/efi/vars.c 	list_del(&entry->list);
entry             586 drivers/firmware/efi/vars.c int __efivar_entry_delete(struct efivar_entry *entry)
entry             593 drivers/firmware/efi/vars.c 	status = __efivars->ops->set_variable(entry->var.VariableName,
entry             594 drivers/firmware/efi/vars.c 					      &entry->var.VendorGuid,
entry             612 drivers/firmware/efi/vars.c int efivar_entry_delete(struct efivar_entry *entry)
entry             625 drivers/firmware/efi/vars.c 	status = ops->set_variable(entry->var.VariableName,
entry             626 drivers/firmware/efi/vars.c 				   &entry->var.VendorGuid,
entry             633 drivers/firmware/efi/vars.c 	efivar_entry_list_del_unlock(entry);
entry             659 drivers/firmware/efi/vars.c int efivar_entry_set(struct efivar_entry *entry, u32 attributes,
entry             664 drivers/firmware/efi/vars.c 	efi_char16_t *name = entry->var.VariableName;
entry             665 drivers/firmware/efi/vars.c 	efi_guid_t vendor = entry->var.VendorGuid;
entry             817 drivers/firmware/efi/vars.c 	struct efivar_entry *entry, *n;
entry             821 drivers/firmware/efi/vars.c 	list_for_each_entry_safe(entry, n, head, list) {
entry             823 drivers/firmware/efi/vars.c 		strsize2 = ucs2_strsize(entry->var.VariableName, 1024);
entry             825 drivers/firmware/efi/vars.c 		    !memcmp(name, &(entry->var.VariableName), strsize1) &&
entry             826 drivers/firmware/efi/vars.c 		    !efi_guidcmp(guid, entry->var.VendorGuid)) {
entry             836 drivers/firmware/efi/vars.c 		if (entry->scanning) {
entry             841 drivers/firmware/efi/vars.c 			entry->deleting = true;
entry             843 drivers/firmware/efi/vars.c 			list_del(&entry->list);
entry             846 drivers/firmware/efi/vars.c 	return entry;
entry             855 drivers/firmware/efi/vars.c int efivar_entry_size(struct efivar_entry *entry, unsigned long *size)
entry             869 drivers/firmware/efi/vars.c 	status = ops->get_variable(entry->var.VariableName,
entry             870 drivers/firmware/efi/vars.c 				   &entry->var.VendorGuid, NULL, size, NULL);
entry             891 drivers/firmware/efi/vars.c int __efivar_entry_get(struct efivar_entry *entry, u32 *attributes,
entry             899 drivers/firmware/efi/vars.c 	status = __efivars->ops->get_variable(entry->var.VariableName,
entry             900 drivers/firmware/efi/vars.c 					      &entry->var.VendorGuid,
entry             914 drivers/firmware/efi/vars.c int efivar_entry_get(struct efivar_entry *entry, u32 *attributes,
entry             927 drivers/firmware/efi/vars.c 	status = __efivars->ops->get_variable(entry->var.VariableName,
entry             928 drivers/firmware/efi/vars.c 					      &entry->var.VendorGuid,
entry             958 drivers/firmware/efi/vars.c int efivar_entry_set_get_size(struct efivar_entry *entry, u32 attributes,
entry             962 drivers/firmware/efi/vars.c 	efi_char16_t *name = entry->var.VariableName;
entry             963 drivers/firmware/efi/vars.c 	efi_guid_t *vendor = &entry->var.VendorGuid;
entry            1018 drivers/firmware/efi/vars.c 	status = ops->get_variable(entry->var.VariableName,
entry            1019 drivers/firmware/efi/vars.c 				   &entry->var.VendorGuid,
entry            1023 drivers/firmware/efi/vars.c 		efivar_entry_list_del_unlock(entry);
entry            1089 drivers/firmware/efi/vars.c 	struct efivar_entry *entry, *n;
entry            1093 drivers/firmware/efi/vars.c 		list_for_each_entry_safe(entry, n, head, list) {
entry            1094 drivers/firmware/efi/vars.c 			err = func(entry, data);
entry            1100 drivers/firmware/efi/vars.c 			*prev = entry;
entry              32 drivers/firmware/google/coreboot_table.c 	return device->entry.tag == driver->tag;
entry              92 drivers/firmware/google/coreboot_table.c 	struct coreboot_table_entry *entry;
entry              97 drivers/firmware/google/coreboot_table.c 		entry = ptr_entry;
entry              99 drivers/firmware/google/coreboot_table.c 		device = kzalloc(sizeof(struct device) + entry->size, GFP_KERNEL);
entry             107 drivers/firmware/google/coreboot_table.c 		memcpy(&device->entry, ptr_entry, entry->size);
entry             115 drivers/firmware/google/coreboot_table.c 		ptr_entry += entry->size;
entry              66 drivers/firmware/google/coreboot_table.h 		struct coreboot_table_entry entry;
entry             614 drivers/firmware/google/gsmi.c 	struct gsmi_log_entry_type_1 entry = {
entry             619 drivers/firmware/google/gsmi.c 		.data_len = sizeof(entry),
entry             636 drivers/firmware/google/gsmi.c 	memcpy(gsmi_dev.data_buf->start, &entry, sizeof(entry));
entry              37 drivers/firmware/google/vpd_decode.c 			    u32 *_consumed, const u8 **entry, u32 *entry_len)
entry              49 drivers/firmware/google/vpd_decode.c 	*entry = input_buf + consumed;
entry             253 drivers/firmware/iscsi_ibft.c 	struct ibft_kobject *entry = data;
entry             254 drivers/firmware/iscsi_ibft.c 	struct ibft_initiator *initiator = entry->initiator;
entry             255 drivers/firmware/iscsi_ibft.c 	void *ibft_loc = entry->header;
entry             294 drivers/firmware/iscsi_ibft.c 	struct ibft_kobject *entry = data;
entry             295 drivers/firmware/iscsi_ibft.c 	struct ibft_nic *nic = entry->nic;
entry             296 drivers/firmware/iscsi_ibft.c 	void *ibft_loc = entry->header;
entry             354 drivers/firmware/iscsi_ibft.c 	struct ibft_kobject *entry = data;
entry             355 drivers/firmware/iscsi_ibft.c 	struct ibft_tgt *tgt = entry->tgt;
entry             356 drivers/firmware/iscsi_ibft.c 	void *ibft_loc = entry->header;
entry             418 drivers/firmware/iscsi_ibft.c 	struct ibft_kobject *entry = data;
entry             424 drivers/firmware/iscsi_ibft.c 				      entry->header->header.signature);
entry             428 drivers/firmware/iscsi_ibft.c 				      entry->header->header.oem_id);
entry             432 drivers/firmware/iscsi_ibft.c 				      entry->header->header.oem_table_id);
entry             473 drivers/firmware/iscsi_ibft.c 	struct ibft_kobject *entry = data;
entry             474 drivers/firmware/iscsi_ibft.c 	struct ibft_nic *nic = entry->nic;
entry             527 drivers/firmware/iscsi_ibft.c 	struct ibft_kobject *entry = data;
entry             528 drivers/firmware/iscsi_ibft.c 	struct ibft_tgt *tgt = entry->tgt;
entry             564 drivers/firmware/iscsi_ibft.c 	struct ibft_kobject *entry = data;
entry             565 drivers/firmware/iscsi_ibft.c 	struct ibft_initiator *init = entry->initiator;
entry              43 drivers/firmware/memmap.c static ssize_t start_show(struct firmware_map_entry *entry, char *buf);
entry              44 drivers/firmware/memmap.c static ssize_t end_show(struct firmware_map_entry *entry, char *buf);
entry              45 drivers/firmware/memmap.c static ssize_t type_show(struct firmware_map_entry *entry, char *buf);
entry              56 drivers/firmware/memmap.c 	ssize_t (*show)(struct firmware_map_entry *entry, char *buf);
entry              99 drivers/firmware/memmap.c 	struct firmware_map_entry *entry = to_memmap_entry(kobj);
entry             101 drivers/firmware/memmap.c 	if (PageReserved(virt_to_page(entry))) {
entry             109 drivers/firmware/memmap.c 		list_add(&entry->list, &map_entries_bootmem);
entry             115 drivers/firmware/memmap.c 	kfree(entry);
entry             143 drivers/firmware/memmap.c 				  struct firmware_map_entry *entry)
entry             147 drivers/firmware/memmap.c 	entry->start = start;
entry             148 drivers/firmware/memmap.c 	entry->end = end - 1;
entry             149 drivers/firmware/memmap.c 	entry->type = type;
entry             150 drivers/firmware/memmap.c 	INIT_LIST_HEAD(&entry->list);
entry             151 drivers/firmware/memmap.c 	kobject_init(&entry->kobj, &memmap_ktype);
entry             154 drivers/firmware/memmap.c 	list_add_tail(&entry->list, &map_entries);
entry             167 drivers/firmware/memmap.c static inline void firmware_map_remove_entry(struct firmware_map_entry *entry)
entry             169 drivers/firmware/memmap.c 	list_del(&entry->list);
entry             175 drivers/firmware/memmap.c static int add_sysfs_fw_map_entry(struct firmware_map_entry *entry)
entry             180 drivers/firmware/memmap.c 	if (entry->kobj.state_in_sysfs)
entry             189 drivers/firmware/memmap.c 	entry->kobj.kset = mmap_kset;
entry             190 drivers/firmware/memmap.c 	if (kobject_add(&entry->kobj, NULL, "%d", map_entries_nr++))
entry             191 drivers/firmware/memmap.c 		kobject_put(&entry->kobj);
entry             199 drivers/firmware/memmap.c static inline void remove_sysfs_fw_map_entry(struct firmware_map_entry *entry)
entry             201 drivers/firmware/memmap.c 	kobject_put(&entry->kobj);
entry             221 drivers/firmware/memmap.c 	struct firmware_map_entry *entry;
entry             223 drivers/firmware/memmap.c 	list_for_each_entry(entry, list, list)
entry             224 drivers/firmware/memmap.c 		if ((entry->start == start) && (entry->end == end) &&
entry             225 drivers/firmware/memmap.c 		    (!strcmp(entry->type, type))) {
entry             226 drivers/firmware/memmap.c 			return entry;
entry             283 drivers/firmware/memmap.c 	struct firmware_map_entry *entry;
entry             285 drivers/firmware/memmap.c 	entry = firmware_map_find_entry(start, end - 1, type);
entry             286 drivers/firmware/memmap.c 	if (entry)
entry             289 drivers/firmware/memmap.c 	entry = firmware_map_find_entry_bootmem(start, end - 1, type);
entry             290 drivers/firmware/memmap.c 	if (!entry) {
entry             291 drivers/firmware/memmap.c 		entry = kzalloc(sizeof(struct firmware_map_entry), GFP_ATOMIC);
entry             292 drivers/firmware/memmap.c 		if (!entry)
entry             297 drivers/firmware/memmap.c 		list_del(&entry->list);
entry             300 drivers/firmware/memmap.c 		memset(entry, 0, sizeof(*entry));
entry             303 drivers/firmware/memmap.c 	firmware_map_add_entry(start, end, type, entry);
entry             305 drivers/firmware/memmap.c 	add_sysfs_fw_map_entry(entry);
entry             325 drivers/firmware/memmap.c 	struct firmware_map_entry *entry;
entry             327 drivers/firmware/memmap.c 	entry = memblock_alloc(sizeof(struct firmware_map_entry),
entry             329 drivers/firmware/memmap.c 	if (WARN_ON(!entry))
entry             332 drivers/firmware/memmap.c 	return firmware_map_add_entry(start, end, type, entry);
entry             347 drivers/firmware/memmap.c 	struct firmware_map_entry *entry;
entry             350 drivers/firmware/memmap.c 	entry = firmware_map_find_entry(start, end - 1, type);
entry             351 drivers/firmware/memmap.c 	if (!entry) {
entry             356 drivers/firmware/memmap.c 	firmware_map_remove_entry(entry);
entry             360 drivers/firmware/memmap.c 	remove_sysfs_fw_map_entry(entry);
entry             369 drivers/firmware/memmap.c static ssize_t start_show(struct firmware_map_entry *entry, char *buf)
entry             372 drivers/firmware/memmap.c 		(unsigned long long)entry->start);
entry             375 drivers/firmware/memmap.c static ssize_t end_show(struct firmware_map_entry *entry, char *buf)
entry             378 drivers/firmware/memmap.c 		(unsigned long long)entry->end);
entry             381 drivers/firmware/memmap.c static ssize_t type_show(struct firmware_map_entry *entry, char *buf)
entry             383 drivers/firmware/memmap.c 	return snprintf(buf, PAGE_SIZE, "%s\n", entry->type);
entry             394 drivers/firmware/memmap.c 	struct firmware_map_entry *entry = to_memmap_entry(kobj);
entry             397 drivers/firmware/memmap.c 	return memmap_attr->show(entry, buf);
entry             410 drivers/firmware/memmap.c 	struct firmware_map_entry *entry;
entry             412 drivers/firmware/memmap.c 	list_for_each_entry(entry, &map_entries, list)
entry             413 drivers/firmware/memmap.c 		add_sysfs_fw_map_entry(entry);
entry              29 drivers/firmware/qcom_scm-32.c 	void *entry;
entry             336 drivers/firmware/qcom_scm-32.c int __qcom_scm_set_cold_boot_addr(void *entry, const cpumask_t *cpus)
entry             358 drivers/firmware/qcom_scm-32.c 				    flags, virt_to_phys(entry));
entry             369 drivers/firmware/qcom_scm-32.c int __qcom_scm_set_warm_boot_addr(struct device *dev, void *entry,
entry             385 drivers/firmware/qcom_scm-32.c 		if (entry == qcom_scm_wb[cpu].entry)
entry             394 drivers/firmware/qcom_scm-32.c 	cmd.addr = cpu_to_le32(virt_to_phys(entry));
entry             400 drivers/firmware/qcom_scm-32.c 			qcom_scm_wb[cpu].entry = entry;
entry             167 drivers/firmware/qcom_scm-64.c int __qcom_scm_set_cold_boot_addr(void *entry, const cpumask_t *cpus)
entry             181 drivers/firmware/qcom_scm-64.c int __qcom_scm_set_warm_boot_addr(struct device *dev, void *entry,
entry              98 drivers/firmware/qcom_scm.c int qcom_scm_set_cold_boot_addr(void *entry, const cpumask_t *cpus)
entry             100 drivers/firmware/qcom_scm.c 	return __qcom_scm_set_cold_boot_addr(entry, cpus);
entry             112 drivers/firmware/qcom_scm.c int qcom_scm_set_warm_boot_addr(void *entry, const cpumask_t *cpus)
entry             114 drivers/firmware/qcom_scm.c 	return __qcom_scm_set_warm_boot_addr(__scm->dev, entry, cpus);
entry              18 drivers/firmware/qcom_scm.h extern int __qcom_scm_set_warm_boot_addr(struct device *dev, void *entry,
entry              20 drivers/firmware/qcom_scm.h extern int __qcom_scm_set_cold_boot_addr(void *entry, const cpumask_t *cpus);
entry             356 drivers/firmware/qemu_fw_cfg.c 	ssize_t (*show)(struct fw_cfg_sysfs_entry *entry, char *buf);
entry             371 drivers/firmware/qemu_fw_cfg.c static inline void fw_cfg_sysfs_cache_enlist(struct fw_cfg_sysfs_entry *entry)
entry             374 drivers/firmware/qemu_fw_cfg.c 	list_add_tail(&entry->list, &fw_cfg_entry_cache);
entry             378 drivers/firmware/qemu_fw_cfg.c static inline void fw_cfg_sysfs_cache_delist(struct fw_cfg_sysfs_entry *entry)
entry             381 drivers/firmware/qemu_fw_cfg.c 	list_del(&entry->list);
entry             387 drivers/firmware/qemu_fw_cfg.c 	struct fw_cfg_sysfs_entry *entry, *next;
entry             389 drivers/firmware/qemu_fw_cfg.c 	list_for_each_entry_safe(entry, next, &fw_cfg_entry_cache, list) {
entry             393 drivers/firmware/qemu_fw_cfg.c 		kobject_put(&entry->kobj);
entry             435 drivers/firmware/qemu_fw_cfg.c 	struct fw_cfg_sysfs_entry *entry = to_entry(kobj);
entry             438 drivers/firmware/qemu_fw_cfg.c 	return attr->show(entry, buf);
entry             448 drivers/firmware/qemu_fw_cfg.c 	struct fw_cfg_sysfs_entry *entry = to_entry(kobj);
entry             450 drivers/firmware/qemu_fw_cfg.c 	fw_cfg_sysfs_cache_delist(entry);
entry             451 drivers/firmware/qemu_fw_cfg.c 	kfree(entry);
entry             466 drivers/firmware/qemu_fw_cfg.c 	struct fw_cfg_sysfs_entry *entry = to_entry(kobj);
entry             468 drivers/firmware/qemu_fw_cfg.c 	if (pos > entry->size)
entry             471 drivers/firmware/qemu_fw_cfg.c 	if (count > entry->size - pos)
entry             472 drivers/firmware/qemu_fw_cfg.c 		count = entry->size - pos;
entry             474 drivers/firmware/qemu_fw_cfg.c 	return fw_cfg_read_blob(entry->select, buf, pos, count);
entry             566 drivers/firmware/qemu_fw_cfg.c 	list_for_each_entry_safe(k, next, &kset->list, entry)
entry             584 drivers/firmware/qemu_fw_cfg.c 	struct fw_cfg_sysfs_entry *entry;
entry             596 drivers/firmware/qemu_fw_cfg.c 	entry = kzalloc(sizeof(*entry), GFP_KERNEL);
entry             597 drivers/firmware/qemu_fw_cfg.c 	if (!entry)
entry             601 drivers/firmware/qemu_fw_cfg.c 	entry->size = be32_to_cpu(f->size);
entry             602 drivers/firmware/qemu_fw_cfg.c 	entry->select = be16_to_cpu(f->select);
entry             603 drivers/firmware/qemu_fw_cfg.c 	memcpy(entry->name, f->name, FW_CFG_MAX_FILE_PATH);
entry             606 drivers/firmware/qemu_fw_cfg.c 	err = kobject_init_and_add(&entry->kobj, &fw_cfg_sysfs_entry_ktype,
entry             607 drivers/firmware/qemu_fw_cfg.c 				   fw_cfg_sel_ko, "%d", entry->select);
entry             612 drivers/firmware/qemu_fw_cfg.c 	err = sysfs_create_bin_file(&entry->kobj, &fw_cfg_sysfs_attr_raw);
entry             617 drivers/firmware/qemu_fw_cfg.c 	fw_cfg_build_symlink(fw_cfg_fname_kset, &entry->kobj, entry->name);
entry             620 drivers/firmware/qemu_fw_cfg.c 	fw_cfg_sysfs_cache_enlist(entry);
entry             624 drivers/firmware/qemu_fw_cfg.c 	kobject_del(&entry->kobj);
entry             626 drivers/firmware/qemu_fw_cfg.c 	kfree(entry);
entry             390 drivers/firmware/tegra/bpmp.c 	struct tegra_bpmp_mrq *entry;
entry             392 drivers/firmware/tegra/bpmp.c 	list_for_each_entry(entry, &bpmp->mrqs, list)
entry             393 drivers/firmware/tegra/bpmp.c 		if (entry->mrq == mrq)
entry             394 drivers/firmware/tegra/bpmp.c 			return entry;
entry             440 drivers/firmware/tegra/bpmp.c 	struct tegra_bpmp_mrq *entry;
entry             445 drivers/firmware/tegra/bpmp.c 	entry = tegra_bpmp_find_mrq(bpmp, mrq);
entry             446 drivers/firmware/tegra/bpmp.c 	if (!entry) {
entry             452 drivers/firmware/tegra/bpmp.c 	entry->handler(mrq, channel, entry->data);
entry             460 drivers/firmware/tegra/bpmp.c 	struct tegra_bpmp_mrq *entry;
entry             466 drivers/firmware/tegra/bpmp.c 	entry = devm_kzalloc(bpmp->dev, sizeof(*entry), GFP_KERNEL);
entry             467 drivers/firmware/tegra/bpmp.c 	if (!entry)
entry             472 drivers/firmware/tegra/bpmp.c 	entry->mrq = mrq;
entry             473 drivers/firmware/tegra/bpmp.c 	entry->handler = handler;
entry             474 drivers/firmware/tegra/bpmp.c 	entry->data = data;
entry             475 drivers/firmware/tegra/bpmp.c 	list_add(&entry->list, &bpmp->mrqs);
entry             485 drivers/firmware/tegra/bpmp.c 	struct tegra_bpmp_mrq *entry;
entry             490 drivers/firmware/tegra/bpmp.c 	entry = tegra_bpmp_find_mrq(bpmp, mrq);
entry             491 drivers/firmware/tegra/bpmp.c 	if (!entry)
entry             494 drivers/firmware/tegra/bpmp.c 	list_del(&entry->list);
entry             495 drivers/firmware/tegra/bpmp.c 	devm_kfree(bpmp->dev, entry);
entry             519 drivers/gpio/gpio-thunderx.c 		txgpio->msix_entries[i].entry = txgpio->base_msi + (2 * i);
entry             746 drivers/gpu/drm/amd/amdgpu/amdgpu_acpi.c 	struct acpi_bus_event *entry = (struct acpi_bus_event *)data;
entry             748 drivers/gpu/drm/amd/amdgpu/amdgpu_acpi.c 	if (strcmp(entry->device_class, ACPI_AC_CLASS) == 0) {
entry             758 drivers/gpu/drm/amd/amdgpu/amdgpu_acpi.c 	return amdgpu_atif_handler(adev, entry);
entry              86 drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd_gpuvm.c 	struct kfd_bo_va_list *entry;
entry              88 drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd_gpuvm.c 	list_for_each_entry(entry, &mem->bo_va_list, bo_list)
entry              89 drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd_gpuvm.c 		if (entry->bo_va->base.vm == avm)
entry             433 drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd_gpuvm.c 		struct kfd_bo_va_list *entry, unsigned long size)
entry             436 drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd_gpuvm.c 			entry->va,
entry             437 drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd_gpuvm.c 			entry->va + size, entry);
entry             438 drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd_gpuvm.c 	amdgpu_vm_bo_rmv(adev, entry->bo_va);
entry             439 drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd_gpuvm.c 	list_del(&entry->bo_list);
entry             440 drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd_gpuvm.c 	kfree(entry);
entry             447 drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd_gpuvm.c 	struct ttm_validate_buffer *entry = &mem->validate_list;
entry             450 drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd_gpuvm.c 	INIT_LIST_HEAD(&entry->head);
entry             451 drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd_gpuvm.c 	entry->num_shared = 1;
entry             452 drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd_gpuvm.c 	entry->bo = &bo->tbo;
entry             455 drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd_gpuvm.c 		list_add_tail(&entry->head, &process_info->userptr_valid_list);
entry             457 drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd_gpuvm.c 		list_add_tail(&entry->head, &process_info->kfd_bo_list);
entry             616 drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd_gpuvm.c 	struct kfd_bo_va_list *entry;
entry             628 drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd_gpuvm.c 	list_for_each_entry(entry, &mem->bo_va_list, bo_list) {
entry             629 drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd_gpuvm.c 		if ((vm && vm != entry->bo_va->base.vm) ||
entry             630 drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd_gpuvm.c 			(entry->is_mapped != map_type
entry             650 drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd_gpuvm.c 	list_for_each_entry(entry, &mem->bo_va_list, bo_list) {
entry             651 drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd_gpuvm.c 		if ((vm && vm != entry->bo_va->base.vm) ||
entry             652 drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd_gpuvm.c 			(entry->is_mapped != map_type
entry             656 drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd_gpuvm.c 		amdgpu_vm_get_pd_bo(entry->bo_va->base.vm, &ctx->list,
entry             707 drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd_gpuvm.c 				struct kfd_bo_va_list *entry,
entry             710 drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd_gpuvm.c 	struct amdgpu_bo_va *bo_va = entry->bo_va;
entry             713 drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd_gpuvm.c 	amdgpu_vm_bo_unmap(adev, bo_va, entry->va);
entry             723 drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd_gpuvm.c 		struct kfd_bo_va_list *entry,
entry             727 drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd_gpuvm.c 	struct amdgpu_bo_va *bo_va = entry->bo_va;
entry             740 drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd_gpuvm.c 		struct kfd_bo_va_list *entry, struct amdgpu_sync *sync,
entry             746 drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd_gpuvm.c 	ret = amdgpu_vm_bo_map(adev, entry->bo_va, entry->va, 0,
entry             747 drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd_gpuvm.c 			       amdgpu_bo_size(entry->bo_va->base.bo),
entry             748 drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd_gpuvm.c 			       entry->pte_flags);
entry             751 drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd_gpuvm.c 				entry->va, ret);
entry             758 drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd_gpuvm.c 	ret = update_gpuvm_pte(adev, entry, sync);
entry             767 drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd_gpuvm.c 	unmap_bo_from_gpuvm(adev, entry, sync);
entry            1231 drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd_gpuvm.c 	struct kfd_bo_va_list *entry, *tmp;
entry            1273 drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd_gpuvm.c 	list_for_each_entry_safe(entry, tmp, &mem->bo_va_list, bo_list)
entry            1274 drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd_gpuvm.c 		remove_bo_from_vm((struct amdgpu_device *)entry->kgd_dev,
entry            1275 drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd_gpuvm.c 				entry, bo_size);
entry            1306 drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd_gpuvm.c 	struct kfd_bo_va_list *entry;
entry            1388 drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd_gpuvm.c 	list_for_each_entry(entry, &mem->bo_va_list, bo_list) {
entry            1389 drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd_gpuvm.c 		if (entry->bo_va->base.vm == vm && !entry->is_mapped) {
entry            1391 drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd_gpuvm.c 					entry->va, entry->va + bo_size,
entry            1392 drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd_gpuvm.c 					entry);
entry            1394 drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd_gpuvm.c 			ret = map_bo_to_gpuvm(adev, entry, ctx.sync,
entry            1407 drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd_gpuvm.c 			entry->is_mapped = true;
entry            1443 drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd_gpuvm.c 	struct kfd_bo_va_list *entry;
entry            1467 drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd_gpuvm.c 	list_for_each_entry(entry, &mem->bo_va_list, bo_list) {
entry            1468 drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd_gpuvm.c 		if (entry->bo_va->base.vm == vm && entry->is_mapped) {
entry            1470 drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd_gpuvm.c 					entry->va,
entry            1471 drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd_gpuvm.c 					entry->va + bo_size,
entry            1472 drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd_gpuvm.c 					entry);
entry            1474 drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd_gpuvm.c 			ret = unmap_bo_from_gpuvm(adev, entry, ctx.sync);
entry            1476 drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd_gpuvm.c 				entry->is_mapped = false;
entry              93 drivers/gpu/drm/amd/amdgpu/amdgpu_bo_list.c 		struct amdgpu_bo_list_entry *entry;
entry             114 drivers/gpu/drm/amd/amdgpu/amdgpu_bo_list.c 			entry = &array[--first_userptr];
entry             116 drivers/gpu/drm/amd/amdgpu/amdgpu_bo_list.c 			entry = &array[last_entry++];
entry             119 drivers/gpu/drm/amd/amdgpu/amdgpu_bo_list.c 		entry->priority = min(info[i].bo_priority,
entry             121 drivers/gpu/drm/amd/amdgpu/amdgpu_bo_list.c 		entry->tv.bo = &bo->tbo;
entry             232 drivers/gpu/drm/amd/amdgpu/amdgpu_dpm.c 	ATOM_PPLIB_Clock_Voltage_Dependency_Record *entry;
entry             238 drivers/gpu/drm/amd/amdgpu/amdgpu_dpm.c 	entry = &atom_table->entries[0];
entry             240 drivers/gpu/drm/amd/amdgpu/amdgpu_dpm.c 		amdgpu_table->entries[i].clk = le16_to_cpu(entry->usClockLow) |
entry             241 drivers/gpu/drm/amd/amdgpu/amdgpu_dpm.c 			(entry->ucClockHigh << 16);
entry             242 drivers/gpu/drm/amd/amdgpu/amdgpu_dpm.c 		amdgpu_table->entries[i].v = le16_to_cpu(entry->usVoltage);
entry             243 drivers/gpu/drm/amd/amdgpu/amdgpu_dpm.c 		entry = (ATOM_PPLIB_Clock_Voltage_Dependency_Record *)
entry             244 drivers/gpu/drm/amd/amdgpu/amdgpu_dpm.c 			((u8 *)entry + sizeof(ATOM_PPLIB_Clock_Voltage_Dependency_Record));
entry             397 drivers/gpu/drm/amd/amdgpu/amdgpu_dpm.c 			ATOM_PPLIB_PhaseSheddingLimits_Record *entry;
entry             408 drivers/gpu/drm/amd/amdgpu/amdgpu_dpm.c 			entry = &psl->entries[0];
entry             411 drivers/gpu/drm/amd/amdgpu/amdgpu_dpm.c 					le16_to_cpu(entry->usSclkLow) | (entry->ucSclkHigh << 16);
entry             413 drivers/gpu/drm/amd/amdgpu/amdgpu_dpm.c 					le16_to_cpu(entry->usMclkLow) | (entry->ucMclkHigh << 16);
entry             415 drivers/gpu/drm/amd/amdgpu/amdgpu_dpm.c 					le16_to_cpu(entry->usVoltage);
entry             416 drivers/gpu/drm/amd/amdgpu/amdgpu_dpm.c 				entry = (ATOM_PPLIB_PhaseSheddingLimits_Record *)
entry             417 drivers/gpu/drm/amd/amdgpu/amdgpu_dpm.c 					((u8 *)entry + sizeof(ATOM_PPLIB_PhaseSheddingLimits_Record));
entry             444 drivers/gpu/drm/amd/amdgpu/amdgpu_dpm.c 			ATOM_PPLIB_CAC_Leakage_Record *entry;
entry             451 drivers/gpu/drm/amd/amdgpu/amdgpu_dpm.c 			entry = &cac_table->entries[0];
entry             455 drivers/gpu/drm/amd/amdgpu/amdgpu_dpm.c 						le16_to_cpu(entry->usVddc1);
entry             457 drivers/gpu/drm/amd/amdgpu/amdgpu_dpm.c 						le16_to_cpu(entry->usVddc2);
entry             459 drivers/gpu/drm/amd/amdgpu/amdgpu_dpm.c 						le16_to_cpu(entry->usVddc3);
entry             462 drivers/gpu/drm/amd/amdgpu/amdgpu_dpm.c 						le16_to_cpu(entry->usVddc);
entry             464 drivers/gpu/drm/amd/amdgpu/amdgpu_dpm.c 						le32_to_cpu(entry->ulLeakageValue);
entry             466 drivers/gpu/drm/amd/amdgpu/amdgpu_dpm.c 				entry = (ATOM_PPLIB_CAC_Leakage_Record *)
entry             467 drivers/gpu/drm/amd/amdgpu/amdgpu_dpm.c 					((u8 *)entry + sizeof(ATOM_PPLIB_CAC_Leakage_Record));
entry             495 drivers/gpu/drm/amd/amdgpu/amdgpu_dpm.c 			ATOM_PPLIB_VCE_Clock_Voltage_Limit_Record *entry;
entry             508 drivers/gpu/drm/amd/amdgpu/amdgpu_dpm.c 			entry = &limits->entries[0];
entry             513 drivers/gpu/drm/amd/amdgpu/amdgpu_dpm.c 					 (entry->ucVCEClockInfoIndex * sizeof(VCEClockInfo)));
entry             519 drivers/gpu/drm/amd/amdgpu/amdgpu_dpm.c 					le16_to_cpu(entry->usVoltage);
entry             520 drivers/gpu/drm/amd/amdgpu/amdgpu_dpm.c 				entry = (ATOM_PPLIB_VCE_Clock_Voltage_Limit_Record *)
entry             521 drivers/gpu/drm/amd/amdgpu/amdgpu_dpm.c 					((u8 *)entry + sizeof(ATOM_PPLIB_VCE_Clock_Voltage_Limit_Record));
entry             552 drivers/gpu/drm/amd/amdgpu/amdgpu_dpm.c 			ATOM_PPLIB_UVD_Clock_Voltage_Limit_Record *entry;
entry             563 drivers/gpu/drm/amd/amdgpu/amdgpu_dpm.c 			entry = &limits->entries[0];
entry             567 drivers/gpu/drm/amd/amdgpu/amdgpu_dpm.c 					 (entry->ucUVDClockInfoIndex * sizeof(UVDClockInfo)));
entry             573 drivers/gpu/drm/amd/amdgpu/amdgpu_dpm.c 					le16_to_cpu(entry->usVoltage);
entry             574 drivers/gpu/drm/amd/amdgpu/amdgpu_dpm.c 				entry = (ATOM_PPLIB_UVD_Clock_Voltage_Limit_Record *)
entry             575 drivers/gpu/drm/amd/amdgpu/amdgpu_dpm.c 					((u8 *)entry + sizeof(ATOM_PPLIB_UVD_Clock_Voltage_Limit_Record));
entry             584 drivers/gpu/drm/amd/amdgpu/amdgpu_dpm.c 			ATOM_PPLIB_SAMClk_Voltage_Limit_Record *entry;
entry             595 drivers/gpu/drm/amd/amdgpu/amdgpu_dpm.c 			entry = &limits->entries[0];
entry             598 drivers/gpu/drm/amd/amdgpu/amdgpu_dpm.c 					le16_to_cpu(entry->usSAMClockLow) | (entry->ucSAMClockHigh << 16);
entry             600 drivers/gpu/drm/amd/amdgpu/amdgpu_dpm.c 					le16_to_cpu(entry->usVoltage);
entry             601 drivers/gpu/drm/amd/amdgpu/amdgpu_dpm.c 				entry = (ATOM_PPLIB_SAMClk_Voltage_Limit_Record *)
entry             602 drivers/gpu/drm/amd/amdgpu/amdgpu_dpm.c 					((u8 *)entry + sizeof(ATOM_PPLIB_SAMClk_Voltage_Limit_Record));
entry             642 drivers/gpu/drm/amd/amdgpu/amdgpu_dpm.c 			ATOM_PPLIB_ACPClk_Voltage_Limit_Record *entry;
entry             653 drivers/gpu/drm/amd/amdgpu/amdgpu_dpm.c 			entry = &limits->entries[0];
entry             656 drivers/gpu/drm/amd/amdgpu/amdgpu_dpm.c 					le16_to_cpu(entry->usACPClockLow) | (entry->ucACPClockHigh << 16);
entry             658 drivers/gpu/drm/amd/amdgpu/amdgpu_dpm.c 					le16_to_cpu(entry->usVoltage);
entry             659 drivers/gpu/drm/amd/amdgpu/amdgpu_dpm.c 				entry = (ATOM_PPLIB_ACPClk_Voltage_Limit_Record *)
entry             660 drivers/gpu/drm/amd/amdgpu/amdgpu_dpm.c 					((u8 *)entry + sizeof(ATOM_PPLIB_ACPClk_Voltage_Limit_Record));
entry              63 drivers/gpu/drm/amd/amdgpu/amdgpu_ih.h 			  struct amdgpu_iv_entry *entry);
entry             380 drivers/gpu/drm/amd/amdgpu/amdgpu_irq.c 	struct amdgpu_iv_entry entry;
entry             386 drivers/gpu/drm/amd/amdgpu/amdgpu_irq.c 	entry.iv_entry = (const uint32_t *)&ih->ring[ring_index];
entry             387 drivers/gpu/drm/amd/amdgpu/amdgpu_irq.c 	amdgpu_ih_decode_iv(adev, &entry);
entry             389 drivers/gpu/drm/amd/amdgpu/amdgpu_irq.c 	trace_amdgpu_iv(ih - &adev->irq.ih, &entry);
entry             391 drivers/gpu/drm/amd/amdgpu/amdgpu_irq.c 	client_id = entry.client_id;
entry             392 drivers/gpu/drm/amd/amdgpu/amdgpu_irq.c 	src_id = entry.src_id;
entry             408 drivers/gpu/drm/amd/amdgpu/amdgpu_irq.c 		r = src->funcs->process(adev, src, &entry);
entry             420 drivers/gpu/drm/amd/amdgpu/amdgpu_irq.c 		amdgpu_amdkfd_interrupt(adev, entry.iv_entry);
entry              78 drivers/gpu/drm/amd/amdgpu/amdgpu_irq.h 		       struct amdgpu_iv_entry *entry);
entry              36 drivers/gpu/drm/amd/amdgpu/amdgpu_pmu.c 	struct list_head entry;
entry             241 drivers/gpu/drm/amd/amdgpu/amdgpu_pmu.c 	list_add_tail(&pmu_entry->entry, &amdgpu_pmu_list);
entry             273 drivers/gpu/drm/amd/amdgpu/amdgpu_pmu.c 	list_for_each_entry_safe(pe, temp, &amdgpu_pmu_list, entry) {
entry             275 drivers/gpu/drm/amd/amdgpu/amdgpu_pmu.c 			list_del(&pe->entry);
entry            1006 drivers/gpu/drm/amd/amdgpu/amdgpu_ras.c 	struct amdgpu_iv_entry entry;
entry            1012 drivers/gpu/drm/amd/amdgpu/amdgpu_ras.c 		memcpy(&entry, &data->ring[data->rptr],
entry            1023 drivers/gpu/drm/amd/amdgpu/amdgpu_ras.c 			ret = data->cb(obj->adev, &err_data, &entry);
entry            1063 drivers/gpu/drm/amd/amdgpu/amdgpu_ras.c 	memcpy(&data->ring[data->wptr], info->entry,
entry             369 drivers/gpu/drm/amd/amdgpu/amdgpu_ras.h 		struct amdgpu_iv_entry *entry);
entry             446 drivers/gpu/drm/amd/amdgpu/amdgpu_ras.h 	struct amdgpu_iv_entry *entry;
entry             350 drivers/gpu/drm/amd/amdgpu/amdgpu_vm.c 	struct amdgpu_vm_pt *entry;
entry             370 drivers/gpu/drm/amd/amdgpu/amdgpu_vm.c 	cursor->entry = &vm->root;
entry             389 drivers/gpu/drm/amd/amdgpu/amdgpu_vm.c 	if (!cursor->entry->entries)
entry             392 drivers/gpu/drm/amd/amdgpu/amdgpu_vm.c 	BUG_ON(!cursor->entry->base.bo);
entry             398 drivers/gpu/drm/amd/amdgpu/amdgpu_vm.c 	cursor->parent = cursor->entry;
entry             399 drivers/gpu/drm/amd/amdgpu/amdgpu_vm.c 	cursor->entry = &cursor->entry->entries[idx];
entry             426 drivers/gpu/drm/amd/amdgpu/amdgpu_vm.c 	if (cursor->entry == &cursor->parent->entries[num_entries - 1])
entry             431 drivers/gpu/drm/amd/amdgpu/amdgpu_vm.c 	++cursor->entry;
entry             450 drivers/gpu/drm/amd/amdgpu/amdgpu_vm.c 	cursor->entry = cursor->parent;
entry             511 drivers/gpu/drm/amd/amdgpu/amdgpu_vm.c 				      struct amdgpu_vm_pt *entry)
entry             513 drivers/gpu/drm/amd/amdgpu/amdgpu_vm.c 	return entry && (!start || entry != start->entry);
entry             527 drivers/gpu/drm/amd/amdgpu/amdgpu_vm.c 	if (!cursor->entry)
entry             531 drivers/gpu/drm/amd/amdgpu/amdgpu_vm.c 		cursor->entry = NULL;
entry             541 drivers/gpu/drm/amd/amdgpu/amdgpu_vm.c #define for_each_amdgpu_vm_pt_dfs_safe(adev, vm, start, cursor, entry)		\
entry             543 drivers/gpu/drm/amd/amdgpu/amdgpu_vm.c 	     (entry) = (cursor).entry, amdgpu_vm_pt_next_dfs((adev), &(cursor));\
entry             544 drivers/gpu/drm/amd/amdgpu/amdgpu_vm.c 	     amdgpu_vm_pt_continue_dfs((start), (entry));			\
entry             545 drivers/gpu/drm/amd/amdgpu/amdgpu_vm.c 	     (entry) = (cursor).entry, amdgpu_vm_pt_next_dfs((adev), &(cursor)))
entry             559 drivers/gpu/drm/amd/amdgpu/amdgpu_vm.c 			 struct amdgpu_bo_list_entry *entry)
entry             561 drivers/gpu/drm/amd/amdgpu/amdgpu_vm.c 	entry->priority = 0;
entry             562 drivers/gpu/drm/amd/amdgpu/amdgpu_vm.c 	entry->tv.bo = &vm->root.base.bo->tbo;
entry             564 drivers/gpu/drm/amd/amdgpu/amdgpu_vm.c 	entry->tv.num_shared = 3;
entry             565 drivers/gpu/drm/amd/amdgpu/amdgpu_vm.c 	entry->user_pages = NULL;
entry             566 drivers/gpu/drm/amd/amdgpu/amdgpu_vm.c 	list_add(&entry->tv.head, validated);
entry             855 drivers/gpu/drm/amd/amdgpu/amdgpu_vm.c 	struct amdgpu_vm_pt *entry = cursor->entry;
entry             860 drivers/gpu/drm/amd/amdgpu/amdgpu_vm.c 	if (cursor->level < AMDGPU_VM_PTB && !entry->entries) {
entry             864 drivers/gpu/drm/amd/amdgpu/amdgpu_vm.c 		entry->entries = kvmalloc_array(num_entries,
entry             865 drivers/gpu/drm/amd/amdgpu/amdgpu_vm.c 						sizeof(*entry->entries),
entry             867 drivers/gpu/drm/amd/amdgpu/amdgpu_vm.c 		if (!entry->entries)
entry             871 drivers/gpu/drm/amd/amdgpu/amdgpu_vm.c 	if (entry->base.bo)
entry             884 drivers/gpu/drm/amd/amdgpu/amdgpu_vm.c 	amdgpu_vm_bo_base_init(&entry->base, vm, pt);
entry             903 drivers/gpu/drm/amd/amdgpu/amdgpu_vm.c static void amdgpu_vm_free_table(struct amdgpu_vm_pt *entry)
entry             905 drivers/gpu/drm/amd/amdgpu/amdgpu_vm.c 	if (entry->base.bo) {
entry             906 drivers/gpu/drm/amd/amdgpu/amdgpu_vm.c 		entry->base.bo->vm_bo = NULL;
entry             907 drivers/gpu/drm/amd/amdgpu/amdgpu_vm.c 		list_del(&entry->base.vm_status);
entry             908 drivers/gpu/drm/amd/amdgpu/amdgpu_vm.c 		amdgpu_bo_unref(&entry->base.bo->shadow);
entry             909 drivers/gpu/drm/amd/amdgpu/amdgpu_vm.c 		amdgpu_bo_unref(&entry->base.bo);
entry             911 drivers/gpu/drm/amd/amdgpu/amdgpu_vm.c 	kvfree(entry->entries);
entry             912 drivers/gpu/drm/amd/amdgpu/amdgpu_vm.c 	entry->entries = NULL;
entry             929 drivers/gpu/drm/amd/amdgpu/amdgpu_vm.c 	struct amdgpu_vm_pt *entry;
entry             933 drivers/gpu/drm/amd/amdgpu/amdgpu_vm.c 	for_each_amdgpu_vm_pt_dfs_safe(adev, vm, start, cursor, entry)
entry             934 drivers/gpu/drm/amd/amdgpu/amdgpu_vm.c 		amdgpu_vm_free_table(entry);
entry             937 drivers/gpu/drm/amd/amdgpu/amdgpu_vm.c 		amdgpu_vm_free_table(start->entry);
entry            1192 drivers/gpu/drm/amd/amdgpu/amdgpu_vm.c 				struct amdgpu_vm_pt *entry)
entry            1194 drivers/gpu/drm/amd/amdgpu/amdgpu_vm.c 	struct amdgpu_vm_pt *parent = amdgpu_vm_pt_parent(entry);
entry            1203 drivers/gpu/drm/amd/amdgpu/amdgpu_vm.c 	amdgpu_gmc_get_pde_for_bo(entry->base.bo, level, &pt, &flags);
entry            1204 drivers/gpu/drm/amd/amdgpu/amdgpu_vm.c 	pde = (entry - parent->entries) * 8;
entry            1220 drivers/gpu/drm/amd/amdgpu/amdgpu_vm.c 	struct amdgpu_vm_pt *entry;
entry            1222 drivers/gpu/drm/amd/amdgpu/amdgpu_vm.c 	for_each_amdgpu_vm_pt_dfs_safe(adev, vm, NULL, cursor, entry)
entry            1223 drivers/gpu/drm/amd/amdgpu/amdgpu_vm.c 		if (entry->base.bo && !entry->base.moved)
entry            1224 drivers/gpu/drm/amd/amdgpu/amdgpu_vm.c 			amdgpu_vm_bo_relocated(&entry->base);
entry            1256 drivers/gpu/drm/amd/amdgpu/amdgpu_vm.c 		struct amdgpu_vm_pt *entry;
entry            1258 drivers/gpu/drm/amd/amdgpu/amdgpu_vm.c 		entry = list_first_entry(&vm->relocated, struct amdgpu_vm_pt,
entry            1260 drivers/gpu/drm/amd/amdgpu/amdgpu_vm.c 		amdgpu_vm_bo_idle(&entry->base);
entry            1262 drivers/gpu/drm/amd/amdgpu/amdgpu_vm.c 		r = amdgpu_vm_update_pde(&params, vm, entry);
entry            1404 drivers/gpu/drm/amd/amdgpu/amdgpu_vm.c 		pt = cursor.entry->base.bo;
entry             354 drivers/gpu/drm/amd/amdgpu/amdgpu_vm.h 			 struct amdgpu_bo_list_entry *entry);
entry             330 drivers/gpu/drm/amd/amdgpu/amdgpu_xgmi.c 	struct amdgpu_xgmi	*entry;
entry             369 drivers/gpu/drm/amd/amdgpu/amdgpu_xgmi.c 	list_for_each_entry(entry, &hive->device_list, head)
entry             370 drivers/gpu/drm/amd/amdgpu/amdgpu_xgmi.c 		top_info->nodes[count++].node_id = entry->node_id;
entry             243 drivers/gpu/drm/amd/amdgpu/cik_ih.c 			     struct amdgpu_iv_entry *entry)
entry             254 drivers/gpu/drm/amd/amdgpu/cik_ih.c 	entry->client_id = AMDGPU_IRQ_CLIENTID_LEGACY;
entry             255 drivers/gpu/drm/amd/amdgpu/cik_ih.c 	entry->src_id = dw[0] & 0xff;
entry             256 drivers/gpu/drm/amd/amdgpu/cik_ih.c 	entry->src_data[0] = dw[1] & 0xfffffff;
entry             257 drivers/gpu/drm/amd/amdgpu/cik_ih.c 	entry->ring_id = dw[2] & 0xff;
entry             258 drivers/gpu/drm/amd/amdgpu/cik_ih.c 	entry->vmid = (dw[2] >> 8) & 0xff;
entry             259 drivers/gpu/drm/amd/amdgpu/cik_ih.c 	entry->pasid = (dw[2] >> 16) & 0xffff;
entry            1159 drivers/gpu/drm/amd/amdgpu/cik_sdma.c 				     struct amdgpu_iv_entry *entry)
entry            1163 drivers/gpu/drm/amd/amdgpu/cik_sdma.c 	instance_id = (entry->ring_id & 0x3) >> 0;
entry            1164 drivers/gpu/drm/amd/amdgpu/cik_sdma.c 	queue_id = (entry->ring_id & 0xc) >> 2;
entry            1200 drivers/gpu/drm/amd/amdgpu/cik_sdma.c 					     struct amdgpu_iv_entry *entry)
entry            1205 drivers/gpu/drm/amd/amdgpu/cik_sdma.c 	instance_id = (entry->ring_id & 0x3) >> 0;
entry             222 drivers/gpu/drm/amd/amdgpu/cz_ih.c 			    struct amdgpu_iv_entry *entry)
entry             233 drivers/gpu/drm/amd/amdgpu/cz_ih.c 	entry->client_id = AMDGPU_IRQ_CLIENTID_LEGACY;
entry             234 drivers/gpu/drm/amd/amdgpu/cz_ih.c 	entry->src_id = dw[0] & 0xff;
entry             235 drivers/gpu/drm/amd/amdgpu/cz_ih.c 	entry->src_data[0] = dw[1] & 0xfffffff;
entry             236 drivers/gpu/drm/amd/amdgpu/cz_ih.c 	entry->ring_id = dw[2] & 0xff;
entry             237 drivers/gpu/drm/amd/amdgpu/cz_ih.c 	entry->vmid = (dw[2] >> 8) & 0xff;
entry             238 drivers/gpu/drm/amd/amdgpu/cz_ih.c 	entry->pasid = (dw[2] >> 16) & 0xffff;
entry            3113 drivers/gpu/drm/amd/amdgpu/dce_v10_0.c 				  struct amdgpu_iv_entry *entry)
entry            3120 drivers/gpu/drm/amd/amdgpu/dce_v10_0.c 	crtc_id = (entry->src_id - 8) >> 1;
entry            3211 drivers/gpu/drm/amd/amdgpu/dce_v10_0.c 			      struct amdgpu_iv_entry *entry)
entry            3213 drivers/gpu/drm/amd/amdgpu/dce_v10_0.c 	unsigned crtc = entry->src_id - 1;
entry            3217 drivers/gpu/drm/amd/amdgpu/dce_v10_0.c 	switch (entry->src_data[0]) {
entry            3240 drivers/gpu/drm/amd/amdgpu/dce_v10_0.c 		DRM_DEBUG("Unhandled interrupt: %d %d\n", entry->src_id, entry->src_data[0]);
entry            3249 drivers/gpu/drm/amd/amdgpu/dce_v10_0.c 			     struct amdgpu_iv_entry *entry)
entry            3254 drivers/gpu/drm/amd/amdgpu/dce_v10_0.c 	if (entry->src_data[0] >= adev->mode_info.num_hpd) {
entry            3255 drivers/gpu/drm/amd/amdgpu/dce_v10_0.c 		DRM_DEBUG("Unhandled interrupt: %d %d\n", entry->src_id, entry->src_data[0]);
entry            3259 drivers/gpu/drm/amd/amdgpu/dce_v10_0.c 	hpd = entry->src_data[0];
entry            3239 drivers/gpu/drm/amd/amdgpu/dce_v11_0.c 				  struct amdgpu_iv_entry *entry)
entry            3246 drivers/gpu/drm/amd/amdgpu/dce_v11_0.c 	crtc_id = (entry->src_id - 8) >> 1;
entry            3337 drivers/gpu/drm/amd/amdgpu/dce_v11_0.c 				struct amdgpu_iv_entry *entry)
entry            3339 drivers/gpu/drm/amd/amdgpu/dce_v11_0.c 	unsigned crtc = entry->src_id - 1;
entry            3344 drivers/gpu/drm/amd/amdgpu/dce_v11_0.c 	switch (entry->src_data[0]) {
entry            3367 drivers/gpu/drm/amd/amdgpu/dce_v11_0.c 		DRM_DEBUG("Unhandled interrupt: %d %d\n", entry->src_id, entry->src_data[0]);
entry            3376 drivers/gpu/drm/amd/amdgpu/dce_v11_0.c 			     struct amdgpu_iv_entry *entry)
entry            3381 drivers/gpu/drm/amd/amdgpu/dce_v11_0.c 	if (entry->src_data[0] >= adev->mode_info.num_hpd) {
entry            3382 drivers/gpu/drm/amd/amdgpu/dce_v11_0.c 		DRM_DEBUG("Unhandled interrupt: %d %d\n", entry->src_id, entry->src_data[0]);
entry            3386 drivers/gpu/drm/amd/amdgpu/dce_v11_0.c 	hpd = entry->src_data[0];
entry            2930 drivers/gpu/drm/amd/amdgpu/dce_v6_0.c 			     struct amdgpu_iv_entry *entry)
entry            2932 drivers/gpu/drm/amd/amdgpu/dce_v6_0.c 	unsigned crtc = entry->src_id - 1;
entry            2937 drivers/gpu/drm/amd/amdgpu/dce_v6_0.c 	switch (entry->src_data[0]) {
entry            2958 drivers/gpu/drm/amd/amdgpu/dce_v6_0.c 		DRM_DEBUG("Unhandled interrupt: %d %d\n", entry->src_id, entry->src_data[0]);
entry            2990 drivers/gpu/drm/amd/amdgpu/dce_v6_0.c 				 struct amdgpu_iv_entry *entry)
entry            2997 drivers/gpu/drm/amd/amdgpu/dce_v6_0.c 	crtc_id = (entry->src_id - 8) >> 1;
entry            3043 drivers/gpu/drm/amd/amdgpu/dce_v6_0.c 			    struct amdgpu_iv_entry *entry)
entry            3048 drivers/gpu/drm/amd/amdgpu/dce_v6_0.c 	if (entry->src_data[0] >= adev->mode_info.num_hpd) {
entry            3049 drivers/gpu/drm/amd/amdgpu/dce_v6_0.c 		DRM_DEBUG("Unhandled interrupt: %d %d\n", entry->src_id, entry->src_data[0]);
entry            3053 drivers/gpu/drm/amd/amdgpu/dce_v6_0.c 	hpd = entry->src_data[0];
entry            3022 drivers/gpu/drm/amd/amdgpu/dce_v8_0.c 			     struct amdgpu_iv_entry *entry)
entry            3024 drivers/gpu/drm/amd/amdgpu/dce_v8_0.c 	unsigned crtc = entry->src_id - 1;
entry            3029 drivers/gpu/drm/amd/amdgpu/dce_v8_0.c 	switch (entry->src_data[0]) {
entry            3050 drivers/gpu/drm/amd/amdgpu/dce_v8_0.c 		DRM_DEBUG("Unhandled interrupt: %d %d\n", entry->src_id, entry->src_data[0]);
entry            3082 drivers/gpu/drm/amd/amdgpu/dce_v8_0.c 				struct amdgpu_iv_entry *entry)
entry            3089 drivers/gpu/drm/amd/amdgpu/dce_v8_0.c 	crtc_id = (entry->src_id - 8) >> 1;
entry            3135 drivers/gpu/drm/amd/amdgpu/dce_v8_0.c 			    struct amdgpu_iv_entry *entry)
entry            3140 drivers/gpu/drm/amd/amdgpu/dce_v8_0.c 	if (entry->src_data[0] >= adev->mode_info.num_hpd) {
entry            3141 drivers/gpu/drm/amd/amdgpu/dce_v8_0.c 		DRM_DEBUG("Unhandled interrupt: %d %d\n", entry->src_id, entry->src_data[0]);
entry            3145 drivers/gpu/drm/amd/amdgpu/dce_v8_0.c 	hpd = entry->src_data[0];
entry            4992 drivers/gpu/drm/amd/amdgpu/gfx_v10_0.c 			     struct amdgpu_iv_entry *entry)
entry            4999 drivers/gpu/drm/amd/amdgpu/gfx_v10_0.c 	me_id = (entry->ring_id & 0x0c) >> 2;
entry            5000 drivers/gpu/drm/amd/amdgpu/gfx_v10_0.c 	pipe_id = (entry->ring_id & 0x03) >> 0;
entry            5001 drivers/gpu/drm/amd/amdgpu/gfx_v10_0.c 	queue_id = (entry->ring_id & 0x70) >> 4;
entry            5063 drivers/gpu/drm/amd/amdgpu/gfx_v10_0.c 					struct amdgpu_iv_entry *entry)
entry            5069 drivers/gpu/drm/amd/amdgpu/gfx_v10_0.c 	me_id = (entry->ring_id & 0x0c) >> 2;
entry            5070 drivers/gpu/drm/amd/amdgpu/gfx_v10_0.c 	pipe_id = (entry->ring_id & 0x03) >> 0;
entry            5071 drivers/gpu/drm/amd/amdgpu/gfx_v10_0.c 	queue_id = (entry->ring_id & 0x70) >> 4;
entry            5098 drivers/gpu/drm/amd/amdgpu/gfx_v10_0.c 				  struct amdgpu_iv_entry *entry)
entry            5101 drivers/gpu/drm/amd/amdgpu/gfx_v10_0.c 	gfx_v10_0_handle_priv_fault(adev, entry);
entry            5107 drivers/gpu/drm/amd/amdgpu/gfx_v10_0.c 				   struct amdgpu_iv_entry *entry)
entry            5110 drivers/gpu/drm/amd/amdgpu/gfx_v10_0.c 	gfx_v10_0_handle_priv_fault(adev, entry);
entry            5161 drivers/gpu/drm/amd/amdgpu/gfx_v10_0.c 			     struct amdgpu_iv_entry *entry)
entry            5166 drivers/gpu/drm/amd/amdgpu/gfx_v10_0.c 	me_id = (entry->ring_id & 0x0c) >> 2;
entry            5167 drivers/gpu/drm/amd/amdgpu/gfx_v10_0.c 	pipe_id = (entry->ring_id & 0x03) >> 0;
entry            5168 drivers/gpu/drm/amd/amdgpu/gfx_v10_0.c 	queue_id = (entry->ring_id & 0x70) >> 4;
entry            3370 drivers/gpu/drm/amd/amdgpu/gfx_v6_0.c 			    struct amdgpu_iv_entry *entry)
entry            3372 drivers/gpu/drm/amd/amdgpu/gfx_v6_0.c 	switch (entry->ring_id) {
entry            3378 drivers/gpu/drm/amd/amdgpu/gfx_v6_0.c 		amdgpu_fence_process(&adev->gfx.compute_ring[entry->ring_id - 1]);
entry            3387 drivers/gpu/drm/amd/amdgpu/gfx_v6_0.c 			   struct amdgpu_iv_entry *entry)
entry            3391 drivers/gpu/drm/amd/amdgpu/gfx_v6_0.c 	switch (entry->ring_id) {
entry            3397 drivers/gpu/drm/amd/amdgpu/gfx_v6_0.c 		ring = &adev->gfx.compute_ring[entry->ring_id - 1];
entry            3407 drivers/gpu/drm/amd/amdgpu/gfx_v6_0.c 				 struct amdgpu_iv_entry *entry)
entry            3410 drivers/gpu/drm/amd/amdgpu/gfx_v6_0.c 	gfx_v6_0_fault(adev, entry);
entry            3416 drivers/gpu/drm/amd/amdgpu/gfx_v6_0.c 				  struct amdgpu_iv_entry *entry)
entry            3419 drivers/gpu/drm/amd/amdgpu/gfx_v6_0.c 	gfx_v6_0_fault(adev, entry);
entry            4865 drivers/gpu/drm/amd/amdgpu/gfx_v7_0.c 			    struct amdgpu_iv_entry *entry)
entry            4872 drivers/gpu/drm/amd/amdgpu/gfx_v7_0.c 	me_id = (entry->ring_id & 0x0c) >> 2;
entry            4873 drivers/gpu/drm/amd/amdgpu/gfx_v7_0.c 	pipe_id = (entry->ring_id & 0x03) >> 0;
entry            4891 drivers/gpu/drm/amd/amdgpu/gfx_v7_0.c 			   struct amdgpu_iv_entry *entry)
entry            4897 drivers/gpu/drm/amd/amdgpu/gfx_v7_0.c 	me_id = (entry->ring_id & 0x0c) >> 2;
entry            4898 drivers/gpu/drm/amd/amdgpu/gfx_v7_0.c 	pipe_id = (entry->ring_id & 0x03) >> 0;
entry            4916 drivers/gpu/drm/amd/amdgpu/gfx_v7_0.c 				 struct amdgpu_iv_entry *entry)
entry            4919 drivers/gpu/drm/amd/amdgpu/gfx_v7_0.c 	gfx_v7_0_fault(adev, entry);
entry            4925 drivers/gpu/drm/amd/amdgpu/gfx_v7_0.c 				  struct amdgpu_iv_entry *entry)
entry            4929 drivers/gpu/drm/amd/amdgpu/gfx_v7_0.c 	gfx_v7_0_fault(adev, entry);
entry            6729 drivers/gpu/drm/amd/amdgpu/gfx_v8_0.c 			    struct amdgpu_iv_entry *entry)
entry            6736 drivers/gpu/drm/amd/amdgpu/gfx_v8_0.c 	me_id = (entry->ring_id & 0x0c) >> 2;
entry            6737 drivers/gpu/drm/amd/amdgpu/gfx_v8_0.c 	pipe_id = (entry->ring_id & 0x03) >> 0;
entry            6738 drivers/gpu/drm/amd/amdgpu/gfx_v8_0.c 	queue_id = (entry->ring_id & 0x70) >> 4;
entry            6760 drivers/gpu/drm/amd/amdgpu/gfx_v8_0.c 			   struct amdgpu_iv_entry *entry)
entry            6766 drivers/gpu/drm/amd/amdgpu/gfx_v8_0.c 	me_id = (entry->ring_id & 0x0c) >> 2;
entry            6767 drivers/gpu/drm/amd/amdgpu/gfx_v8_0.c 	pipe_id = (entry->ring_id & 0x03) >> 0;
entry            6768 drivers/gpu/drm/amd/amdgpu/gfx_v8_0.c 	queue_id = (entry->ring_id & 0x70) >> 4;
entry            6788 drivers/gpu/drm/amd/amdgpu/gfx_v8_0.c 				 struct amdgpu_iv_entry *entry)
entry            6791 drivers/gpu/drm/amd/amdgpu/gfx_v8_0.c 	gfx_v8_0_fault(adev, entry);
entry            6797 drivers/gpu/drm/amd/amdgpu/gfx_v8_0.c 				  struct amdgpu_iv_entry *entry)
entry            6800 drivers/gpu/drm/amd/amdgpu/gfx_v8_0.c 	gfx_v8_0_fault(adev, entry);
entry            6806 drivers/gpu/drm/amd/amdgpu/gfx_v8_0.c 				     struct amdgpu_iv_entry *entry)
entry            6893 drivers/gpu/drm/amd/amdgpu/gfx_v8_0.c 			   struct amdgpu_iv_entry *entry)
entry            6895 drivers/gpu/drm/amd/amdgpu/gfx_v8_0.c 	unsigned ih_data = entry->src_data[0];
entry            4417 drivers/gpu/drm/amd/amdgpu/gfx_v9_0.c 		struct amdgpu_iv_entry *entry);
entry            5664 drivers/gpu/drm/amd/amdgpu/gfx_v9_0.c 			    struct amdgpu_iv_entry *entry)
entry            5671 drivers/gpu/drm/amd/amdgpu/gfx_v9_0.c 	me_id = (entry->ring_id & 0x0c) >> 2;
entry            5672 drivers/gpu/drm/amd/amdgpu/gfx_v9_0.c 	pipe_id = (entry->ring_id & 0x03) >> 0;
entry            5673 drivers/gpu/drm/amd/amdgpu/gfx_v9_0.c 	queue_id = (entry->ring_id & 0x70) >> 4;
entry            5695 drivers/gpu/drm/amd/amdgpu/gfx_v9_0.c 			   struct amdgpu_iv_entry *entry)
entry            5701 drivers/gpu/drm/amd/amdgpu/gfx_v9_0.c 	me_id = (entry->ring_id & 0x0c) >> 2;
entry            5702 drivers/gpu/drm/amd/amdgpu/gfx_v9_0.c 	pipe_id = (entry->ring_id & 0x03) >> 0;
entry            5703 drivers/gpu/drm/amd/amdgpu/gfx_v9_0.c 	queue_id = (entry->ring_id & 0x70) >> 4;
entry            5723 drivers/gpu/drm/amd/amdgpu/gfx_v9_0.c 				 struct amdgpu_iv_entry *entry)
entry            5726 drivers/gpu/drm/amd/amdgpu/gfx_v9_0.c 	gfx_v9_0_fault(adev, entry);
entry            5732 drivers/gpu/drm/amd/amdgpu/gfx_v9_0.c 				  struct amdgpu_iv_entry *entry)
entry            5735 drivers/gpu/drm/amd/amdgpu/gfx_v9_0.c 	gfx_v9_0_fault(adev, entry);
entry            5741 drivers/gpu/drm/amd/amdgpu/gfx_v9_0.c 		struct amdgpu_iv_entry *entry)
entry            6161 drivers/gpu/drm/amd/amdgpu/gfx_v9_0.c 				  struct amdgpu_iv_entry *entry)
entry            6165 drivers/gpu/drm/amd/amdgpu/gfx_v9_0.c 		.entry = entry,
entry             129 drivers/gpu/drm/amd/amdgpu/gmc_v10_0.c 				       struct amdgpu_iv_entry *entry)
entry             131 drivers/gpu/drm/amd/amdgpu/gmc_v10_0.c 	struct amdgpu_vmhub *hub = &adev->vmhub[entry->vmid_src];
entry             135 drivers/gpu/drm/amd/amdgpu/gmc_v10_0.c 	addr = (u64)entry->src_data[0] << 12;
entry             136 drivers/gpu/drm/amd/amdgpu/gmc_v10_0.c 	addr |= ((u64)entry->src_data[1] & 0xf) << 44;
entry             144 drivers/gpu/drm/amd/amdgpu/gmc_v10_0.c 		if (entry->vmid_src == AMDGPU_GFXHUB_0)
entry             155 drivers/gpu/drm/amd/amdgpu/gmc_v10_0.c 		amdgpu_vm_get_task_info(adev, entry->pasid, &task_info);
entry             160 drivers/gpu/drm/amd/amdgpu/gmc_v10_0.c 			entry->vmid_src ? "mmhub" : "gfxhub",
entry             161 drivers/gpu/drm/amd/amdgpu/gmc_v10_0.c 			entry->src_id, entry->ring_id, entry->vmid,
entry             162 drivers/gpu/drm/amd/amdgpu/gmc_v10_0.c 			entry->pasid, task_info.process_name, task_info.tgid,
entry             165 drivers/gpu/drm/amd/amdgpu/gmc_v10_0.c 			addr, entry->client_id);
entry            1095 drivers/gpu/drm/amd/amdgpu/gmc_v6_0.c 				      struct amdgpu_iv_entry *entry)
entry            1111 drivers/gpu/drm/amd/amdgpu/gmc_v6_0.c 			entry->src_id, entry->src_data[0]);
entry            1245 drivers/gpu/drm/amd/amdgpu/gmc_v7_0.c 				      struct amdgpu_iv_entry *entry)
entry            1263 drivers/gpu/drm/amd/amdgpu/gmc_v7_0.c 			entry->src_id, entry->src_data[0]);
entry            1269 drivers/gpu/drm/amd/amdgpu/gmc_v7_0.c 					 entry->pasid);
entry            1417 drivers/gpu/drm/amd/amdgpu/gmc_v8_0.c 				      struct amdgpu_iv_entry *entry)
entry            1423 drivers/gpu/drm/amd/amdgpu/gmc_v8_0.c 			entry->src_id, entry->src_data[0]);
entry            1444 drivers/gpu/drm/amd/amdgpu/gmc_v8_0.c 		amdgpu_vm_get_task_info(adev, entry->pasid, &task_info);
entry            1447 drivers/gpu/drm/amd/amdgpu/gmc_v8_0.c 			entry->src_id, entry->src_data[0], task_info.process_name,
entry            1454 drivers/gpu/drm/amd/amdgpu/gmc_v8_0.c 					 entry->pasid);
entry             248 drivers/gpu/drm/amd/amdgpu/gmc_v9_0.c 		struct amdgpu_iv_entry *entry)
entry             268 drivers/gpu/drm/amd/amdgpu/gmc_v9_0.c 		struct amdgpu_iv_entry *entry)
entry             272 drivers/gpu/drm/amd/amdgpu/gmc_v9_0.c 		.entry = entry,
entry             331 drivers/gpu/drm/amd/amdgpu/gmc_v9_0.c 				struct amdgpu_iv_entry *entry)
entry             334 drivers/gpu/drm/amd/amdgpu/gmc_v9_0.c 	bool retry_fault = !!(entry->src_data[1] & 0x80);
entry             339 drivers/gpu/drm/amd/amdgpu/gmc_v9_0.c 	addr = (u64)entry->src_data[0] << 12;
entry             340 drivers/gpu/drm/amd/amdgpu/gmc_v9_0.c 	addr |= ((u64)entry->src_data[1] & 0xf) << 44;
entry             342 drivers/gpu/drm/amd/amdgpu/gmc_v9_0.c 	if (retry_fault && amdgpu_gmc_filter_faults(adev, addr, entry->pasid,
entry             343 drivers/gpu/drm/amd/amdgpu/gmc_v9_0.c 						    entry->timestamp))
entry             346 drivers/gpu/drm/amd/amdgpu/gmc_v9_0.c 	if (entry->client_id == SOC15_IH_CLIENTID_VMC) {
entry             349 drivers/gpu/drm/amd/amdgpu/gmc_v9_0.c 	} else if (entry->client_id == SOC15_IH_CLIENTID_VMC1) {
entry             364 drivers/gpu/drm/amd/amdgpu/gmc_v9_0.c 		if (entry->vmid_src == AMDGPU_GFXHUB_0)
entry             375 drivers/gpu/drm/amd/amdgpu/gmc_v9_0.c 		amdgpu_vm_get_task_info(adev, entry->pasid, &task_info);
entry             381 drivers/gpu/drm/amd/amdgpu/gmc_v9_0.c 			entry->src_id, entry->ring_id, entry->vmid,
entry             382 drivers/gpu/drm/amd/amdgpu/gmc_v9_0.c 			entry->pasid, task_info.process_name, task_info.tgid,
entry             385 drivers/gpu/drm/amd/amdgpu/gmc_v9_0.c 			addr, entry->client_id);
entry             222 drivers/gpu/drm/amd/amdgpu/iceland_ih.c 				 struct amdgpu_iv_entry *entry)
entry             233 drivers/gpu/drm/amd/amdgpu/iceland_ih.c 	entry->client_id = AMDGPU_IRQ_CLIENTID_LEGACY;
entry             234 drivers/gpu/drm/amd/amdgpu/iceland_ih.c 	entry->src_id = dw[0] & 0xff;
entry             235 drivers/gpu/drm/amd/amdgpu/iceland_ih.c 	entry->src_data[0] = dw[1] & 0xfffffff;
entry             236 drivers/gpu/drm/amd/amdgpu/iceland_ih.c 	entry->ring_id = dw[2] & 0xff;
entry             237 drivers/gpu/drm/amd/amdgpu/iceland_ih.c 	entry->vmid = (dw[2] >> 8) & 0xff;
entry             238 drivers/gpu/drm/amd/amdgpu/iceland_ih.c 	entry->pasid = (dw[2] >> 16) & 0xffff;
entry            3188 drivers/gpu/drm/amd/amdgpu/kv_dpm.c 				    struct amdgpu_iv_entry *entry)
entry            3192 drivers/gpu/drm/amd/amdgpu/kv_dpm.c 	if (entry == NULL)
entry            3195 drivers/gpu/drm/amd/amdgpu/kv_dpm.c 	switch (entry->src_id) {
entry             292 drivers/gpu/drm/amd/amdgpu/mxgpu_ai.c 					struct amdgpu_iv_entry *entry)
entry             367 drivers/gpu/drm/amd/amdgpu/mxgpu_ai.c 				   struct amdgpu_iv_entry *entry)
entry             491 drivers/gpu/drm/amd/amdgpu/mxgpu_vi.c 				   struct amdgpu_iv_entry *entry)
entry             543 drivers/gpu/drm/amd/amdgpu/mxgpu_vi.c 				   struct amdgpu_iv_entry *entry)
entry             254 drivers/gpu/drm/amd/amdgpu/navi10_ih.c 				struct amdgpu_iv_entry *entry)
entry             269 drivers/gpu/drm/amd/amdgpu/navi10_ih.c 	entry->client_id = dw[0] & 0xff;
entry             270 drivers/gpu/drm/amd/amdgpu/navi10_ih.c 	entry->src_id = (dw[0] >> 8) & 0xff;
entry             271 drivers/gpu/drm/amd/amdgpu/navi10_ih.c 	entry->ring_id = (dw[0] >> 16) & 0xff;
entry             272 drivers/gpu/drm/amd/amdgpu/navi10_ih.c 	entry->vmid = (dw[0] >> 24) & 0xf;
entry             273 drivers/gpu/drm/amd/amdgpu/navi10_ih.c 	entry->vmid_src = (dw[0] >> 31);
entry             274 drivers/gpu/drm/amd/amdgpu/navi10_ih.c 	entry->timestamp = dw[1] | ((u64)(dw[2] & 0xffff) << 32);
entry             275 drivers/gpu/drm/amd/amdgpu/navi10_ih.c 	entry->timestamp_src = dw[2] >> 31;
entry             276 drivers/gpu/drm/amd/amdgpu/navi10_ih.c 	entry->pasid = dw[3] & 0xffff;
entry             277 drivers/gpu/drm/amd/amdgpu/navi10_ih.c 	entry->pasid_src = dw[3] >> 31;
entry             278 drivers/gpu/drm/amd/amdgpu/navi10_ih.c 	entry->src_data[0] = dw[4];
entry             279 drivers/gpu/drm/amd/amdgpu/navi10_ih.c 	entry->src_data[1] = dw[5];
entry             280 drivers/gpu/drm/amd/amdgpu/navi10_ih.c 	entry->src_data[2] = dw[6];
entry             281 drivers/gpu/drm/amd/amdgpu/navi10_ih.c 	entry->src_data[3] = dw[7];
entry            1052 drivers/gpu/drm/amd/amdgpu/sdma_v2_4.c 				      struct amdgpu_iv_entry *entry)
entry            1056 drivers/gpu/drm/amd/amdgpu/sdma_v2_4.c 	instance_id = (entry->ring_id & 0x3) >> 0;
entry            1057 drivers/gpu/drm/amd/amdgpu/sdma_v2_4.c 	queue_id = (entry->ring_id & 0xc) >> 2;
entry            1092 drivers/gpu/drm/amd/amdgpu/sdma_v2_4.c 					      struct amdgpu_iv_entry *entry)
entry            1097 drivers/gpu/drm/amd/amdgpu/sdma_v2_4.c 	instance_id = (entry->ring_id & 0x3) >> 0;
entry            1098 drivers/gpu/drm/amd/amdgpu/sdma_v2_4.c 	queue_id = (entry->ring_id & 0xc) >> 2;
entry            1386 drivers/gpu/drm/amd/amdgpu/sdma_v3_0.c 				      struct amdgpu_iv_entry *entry)
entry            1390 drivers/gpu/drm/amd/amdgpu/sdma_v3_0.c 	instance_id = (entry->ring_id & 0x3) >> 0;
entry            1391 drivers/gpu/drm/amd/amdgpu/sdma_v3_0.c 	queue_id = (entry->ring_id & 0xc) >> 2;
entry            1426 drivers/gpu/drm/amd/amdgpu/sdma_v3_0.c 					      struct amdgpu_iv_entry *entry)
entry            1431 drivers/gpu/drm/amd/amdgpu/sdma_v3_0.c 	instance_id = (entry->ring_id & 0x3) >> 0;
entry            1432 drivers/gpu/drm/amd/amdgpu/sdma_v3_0.c 	queue_id = (entry->ring_id & 0xc) >> 2;
entry            1695 drivers/gpu/drm/amd/amdgpu/sdma_v4_0.c 		struct amdgpu_iv_entry *entry);
entry            2002 drivers/gpu/drm/amd/amdgpu/sdma_v4_0.c 				      struct amdgpu_iv_entry *entry)
entry            2007 drivers/gpu/drm/amd/amdgpu/sdma_v4_0.c 	instance = sdma_v4_0_irq_id_to_seq(entry->client_id);
entry            2008 drivers/gpu/drm/amd/amdgpu/sdma_v4_0.c 	switch (entry->ring_id) {
entry            2029 drivers/gpu/drm/amd/amdgpu/sdma_v4_0.c 		struct amdgpu_iv_entry *entry)
entry            2034 drivers/gpu/drm/amd/amdgpu/sdma_v4_0.c 	instance = sdma_v4_0_irq_id_to_seq(entry->client_id);
entry            2038 drivers/gpu/drm/amd/amdgpu/sdma_v4_0.c 	switch (entry->src_id) {
entry            2058 drivers/gpu/drm/amd/amdgpu/sdma_v4_0.c 				      struct amdgpu_iv_entry *entry)
entry            2062 drivers/gpu/drm/amd/amdgpu/sdma_v4_0.c 		.entry = entry,
entry            2076 drivers/gpu/drm/amd/amdgpu/sdma_v4_0.c 					      struct amdgpu_iv_entry *entry)
entry            2082 drivers/gpu/drm/amd/amdgpu/sdma_v4_0.c 	instance = sdma_v4_0_irq_id_to_seq(entry->client_id);
entry            2086 drivers/gpu/drm/amd/amdgpu/sdma_v4_0.c 	switch (entry->ring_id) {
entry            1428 drivers/gpu/drm/amd/amdgpu/sdma_v5_0.c 				      struct amdgpu_iv_entry *entry)
entry            1431 drivers/gpu/drm/amd/amdgpu/sdma_v5_0.c 	switch (entry->client_id) {
entry            1433 drivers/gpu/drm/amd/amdgpu/sdma_v5_0.c 		switch (entry->ring_id) {
entry            1449 drivers/gpu/drm/amd/amdgpu/sdma_v5_0.c 		switch (entry->ring_id) {
entry            1470 drivers/gpu/drm/amd/amdgpu/sdma_v5_0.c 					      struct amdgpu_iv_entry *entry)
entry             634 drivers/gpu/drm/amd/amdgpu/si_dma.c 				      struct amdgpu_iv_entry *entry)
entry             636 drivers/gpu/drm/amd/amdgpu/si_dma.c 	if (entry->src_id == 224)
entry            6057 drivers/gpu/drm/amd/amdgpu/si_dpm.c static void si_convert_mc_registers(const struct si_mc_reg_entry *entry,
entry            6065 drivers/gpu/drm/amd/amdgpu/si_dpm.c 			data->value[i] = cpu_to_be32(entry->mc_data[j]);
entry            7554 drivers/gpu/drm/amd/amdgpu/si_dpm.c 				    struct amdgpu_iv_entry *entry)
entry            7558 drivers/gpu/drm/amd/amdgpu/si_dpm.c 	if (entry == NULL)
entry            7561 drivers/gpu/drm/amd/amdgpu/si_dpm.c 	switch (entry->src_id) {
entry             126 drivers/gpu/drm/amd/amdgpu/si_ih.c 			    struct amdgpu_iv_entry *entry)
entry             136 drivers/gpu/drm/amd/amdgpu/si_ih.c 	entry->client_id = AMDGPU_IRQ_CLIENTID_LEGACY;
entry             137 drivers/gpu/drm/amd/amdgpu/si_ih.c 	entry->src_id = dw[0] & 0xff;
entry             138 drivers/gpu/drm/amd/amdgpu/si_ih.c 	entry->src_data[0] = dw[1] & 0xfffffff;
entry             139 drivers/gpu/drm/amd/amdgpu/si_ih.c 	entry->ring_id = dw[2] & 0xff;
entry             140 drivers/gpu/drm/amd/amdgpu/si_ih.c 	entry->vmid = (dw[2] >> 8) & 0xff;
entry             440 drivers/gpu/drm/amd/amdgpu/soc15.c 	const struct soc15_reg_golden *entry;
entry             445 drivers/gpu/drm/amd/amdgpu/soc15.c 		entry = &regs[i];
entry             446 drivers/gpu/drm/amd/amdgpu/soc15.c 		reg =  adev->reg_offset[entry->hwip][entry->instance][entry->segment] + entry->reg;
entry             448 drivers/gpu/drm/amd/amdgpu/soc15.c 		if (entry->and_mask == 0xffffffff) {
entry             449 drivers/gpu/drm/amd/amdgpu/soc15.c 			tmp = entry->or_mask;
entry             452 drivers/gpu/drm/amd/amdgpu/soc15.c 			tmp &= ~(entry->and_mask);
entry             453 drivers/gpu/drm/amd/amdgpu/soc15.c 			tmp |= (entry->or_mask & entry->and_mask);
entry              65 drivers/gpu/drm/amd/amdgpu/soc15.h #define SOC15_REG_ENTRY_OFFSET(entry)	(adev->reg_offset[entry.hwip][entry.inst][entry.seg] + entry.reg_offset)
entry             224 drivers/gpu/drm/amd/amdgpu/tonga_ih.c 			       struct amdgpu_iv_entry *entry)
entry             235 drivers/gpu/drm/amd/amdgpu/tonga_ih.c 	entry->client_id = AMDGPU_IRQ_CLIENTID_LEGACY;
entry             236 drivers/gpu/drm/amd/amdgpu/tonga_ih.c 	entry->src_id = dw[0] & 0xff;
entry             237 drivers/gpu/drm/amd/amdgpu/tonga_ih.c 	entry->src_data[0] = dw[1] & 0xfffffff;
entry             238 drivers/gpu/drm/amd/amdgpu/tonga_ih.c 	entry->ring_id = dw[2] & 0xff;
entry             239 drivers/gpu/drm/amd/amdgpu/tonga_ih.c 	entry->vmid = (dw[2] >> 8) & 0xff;
entry             240 drivers/gpu/drm/amd/amdgpu/tonga_ih.c 	entry->pasid = (dw[2] >> 16) & 0xffff;
entry             672 drivers/gpu/drm/amd/amdgpu/uvd_v4_2.c 				      struct amdgpu_iv_entry *entry)
entry             595 drivers/gpu/drm/amd/amdgpu/uvd_v5_0.c 				      struct amdgpu_iv_entry *entry)
entry            1218 drivers/gpu/drm/amd/amdgpu/uvd_v6_0.c 				      struct amdgpu_iv_entry *entry)
entry            1223 drivers/gpu/drm/amd/amdgpu/uvd_v6_0.c 	switch (entry->src_id) {
entry            1243 drivers/gpu/drm/amd/amdgpu/uvd_v6_0.c 			  entry->src_id, entry->src_data[0]);
entry            1542 drivers/gpu/drm/amd/amdgpu/uvd_v7_0.c 				      struct amdgpu_iv_entry *entry)
entry            1546 drivers/gpu/drm/amd/amdgpu/uvd_v7_0.c 	switch (entry->client_id) {
entry            1554 drivers/gpu/drm/amd/amdgpu/uvd_v7_0.c 		DRM_ERROR("Unhandled client id: %d\n", entry->client_id);
entry            1560 drivers/gpu/drm/amd/amdgpu/uvd_v7_0.c 	switch (entry->src_id) {
entry            1573 drivers/gpu/drm/amd/amdgpu/uvd_v7_0.c 			  entry->src_id, entry->src_data[0]);
entry             533 drivers/gpu/drm/amd/amdgpu/vce_v2_0.c 				      struct amdgpu_iv_entry *entry)
entry             536 drivers/gpu/drm/amd/amdgpu/vce_v2_0.c 	switch (entry->src_data[0]) {
entry             539 drivers/gpu/drm/amd/amdgpu/vce_v2_0.c 		amdgpu_fence_process(&adev->vce.ring[entry->src_data[0]]);
entry             543 drivers/gpu/drm/amd/amdgpu/vce_v2_0.c 			  entry->src_id, entry->src_data[0]);
entry             717 drivers/gpu/drm/amd/amdgpu/vce_v3_0.c 				      struct amdgpu_iv_entry *entry)
entry             723 drivers/gpu/drm/amd/amdgpu/vce_v3_0.c 	switch (entry->src_data[0]) {
entry             727 drivers/gpu/drm/amd/amdgpu/vce_v3_0.c 		amdgpu_fence_process(&adev->vce.ring[entry->src_data[0]]);
entry             731 drivers/gpu/drm/amd/amdgpu/vce_v3_0.c 			  entry->src_id, entry->src_data[0]);
entry            1028 drivers/gpu/drm/amd/amdgpu/vce_v4_0.c 				      struct amdgpu_iv_entry *entry)
entry            1032 drivers/gpu/drm/amd/amdgpu/vce_v4_0.c 	switch (entry->src_data[0]) {
entry            1036 drivers/gpu/drm/amd/amdgpu/vce_v4_0.c 		amdgpu_fence_process(&adev->vce.ring[entry->src_data[0]]);
entry            1040 drivers/gpu/drm/amd/amdgpu/vce_v4_0.c 			  entry->src_id, entry->src_data[0]);
entry            2113 drivers/gpu/drm/amd/amdgpu/vcn_v1_0.c 				      struct amdgpu_iv_entry *entry)
entry            2117 drivers/gpu/drm/amd/amdgpu/vcn_v1_0.c 	switch (entry->src_id) {
entry            2132 drivers/gpu/drm/amd/amdgpu/vcn_v1_0.c 			  entry->src_id, entry->src_data[0]);
entry            2069 drivers/gpu/drm/amd/amdgpu/vcn_v2_0.c 				      struct amdgpu_iv_entry *entry)
entry            2073 drivers/gpu/drm/amd/amdgpu/vcn_v2_0.c 	switch (entry->src_id) {
entry            2088 drivers/gpu/drm/amd/amdgpu/vcn_v2_0.c 			  entry->src_id, entry->src_data[0]);
entry            1328 drivers/gpu/drm/amd/amdgpu/vcn_v2_5.c 				      struct amdgpu_iv_entry *entry)
entry            1332 drivers/gpu/drm/amd/amdgpu/vcn_v2_5.c 	switch (entry->client_id) {
entry            1340 drivers/gpu/drm/amd/amdgpu/vcn_v2_5.c 		DRM_ERROR("Unhandled client id: %d\n", entry->client_id);
entry            1346 drivers/gpu/drm/amd/amdgpu/vcn_v2_5.c 	switch (entry->src_id) {
entry            1361 drivers/gpu/drm/amd/amdgpu/vcn_v2_5.c 			  entry->src_id, entry->src_data[0]);
entry             437 drivers/gpu/drm/amd/amdgpu/vega10_ih.c 				struct amdgpu_iv_entry *entry)
entry             452 drivers/gpu/drm/amd/amdgpu/vega10_ih.c 	entry->client_id = dw[0] & 0xff;
entry             453 drivers/gpu/drm/amd/amdgpu/vega10_ih.c 	entry->src_id = (dw[0] >> 8) & 0xff;
entry             454 drivers/gpu/drm/amd/amdgpu/vega10_ih.c 	entry->ring_id = (dw[0] >> 16) & 0xff;
entry             455 drivers/gpu/drm/amd/amdgpu/vega10_ih.c 	entry->vmid = (dw[0] >> 24) & 0xf;
entry             456 drivers/gpu/drm/amd/amdgpu/vega10_ih.c 	entry->vmid_src = (dw[0] >> 31);
entry             457 drivers/gpu/drm/amd/amdgpu/vega10_ih.c 	entry->timestamp = dw[1] | ((u64)(dw[2] & 0xffff) << 32);
entry             458 drivers/gpu/drm/amd/amdgpu/vega10_ih.c 	entry->timestamp_src = dw[2] >> 31;
entry             459 drivers/gpu/drm/amd/amdgpu/vega10_ih.c 	entry->pasid = dw[3] & 0xffff;
entry             460 drivers/gpu/drm/amd/amdgpu/vega10_ih.c 	entry->pasid_src = dw[3] >> 31;
entry             461 drivers/gpu/drm/amd/amdgpu/vega10_ih.c 	entry->src_data[0] = dw[4];
entry             462 drivers/gpu/drm/amd/amdgpu/vega10_ih.c 	entry->src_data[1] = dw[5];
entry             463 drivers/gpu/drm/amd/amdgpu/vega10_ih.c 	entry->src_data[2] = dw[6];
entry             464 drivers/gpu/drm/amd/amdgpu/vega10_ih.c 	entry->src_data[3] = dw[7];
entry             539 drivers/gpu/drm/amd/amdgpu/vega10_ih.c 			      struct amdgpu_iv_entry *entry)
entry             541 drivers/gpu/drm/amd/amdgpu/vega10_ih.c 	uint32_t wptr = cpu_to_le32(entry->src_data[0]);
entry             543 drivers/gpu/drm/amd/amdgpu/vega10_ih.c 	switch (entry->ring_id) {
entry             312 drivers/gpu/drm/amd/amdkfd/kfd_crat.h 	uint8_t		entry[1];
entry             243 drivers/gpu/drm/amd/amdkfd/kfd_events.c 	list_for_each_entry(waiter, &ev->wq.head, wait.entry)
entry             400 drivers/gpu/drm/amd/amdkfd/kfd_events.c 	list_for_each_entry(waiter, &ev->wq.head, wait.entry)
entry              35 drivers/gpu/drm/amd/amdkfd/soc15_int.h #define SOC15_CLIENT_ID_FROM_IH_ENTRY(entry) (le32_to_cpu(entry[0]) & 0xff)
entry              36 drivers/gpu/drm/amd/amdkfd/soc15_int.h #define SOC15_SOURCE_ID_FROM_IH_ENTRY(entry) (le32_to_cpu(entry[0]) >> 8 & 0xff)
entry              37 drivers/gpu/drm/amd/amdkfd/soc15_int.h #define SOC15_RING_ID_FROM_IH_ENTRY(entry) (le32_to_cpu(entry[0]) >> 16 & 0xff)
entry              38 drivers/gpu/drm/amd/amdkfd/soc15_int.h #define SOC15_VMID_FROM_IH_ENTRY(entry) (le32_to_cpu(entry[0]) >> 24 & 0xf)
entry              39 drivers/gpu/drm/amd/amdkfd/soc15_int.h #define SOC15_VMID_TYPE_FROM_IH_ENTRY(entry) (le32_to_cpu(entry[0]) >> 31 & 0x1)
entry              40 drivers/gpu/drm/amd/amdkfd/soc15_int.h #define SOC15_PASID_FROM_IH_ENTRY(entry) (le32_to_cpu(entry[3]) & 0xffff)
entry              41 drivers/gpu/drm/amd/amdkfd/soc15_int.h #define SOC15_CONTEXT_ID0_FROM_IH_ENTRY(entry) (le32_to_cpu(entry[4]))
entry              42 drivers/gpu/drm/amd/amdkfd/soc15_int.h #define SOC15_CONTEXT_ID1_FROM_IH_ENTRY(entry) (le32_to_cpu(entry[5]))
entry              43 drivers/gpu/drm/amd/amdkfd/soc15_int.h #define SOC15_CONTEXT_ID2_FROM_IH_ENTRY(entry) (le32_to_cpu(entry[6]))
entry              44 drivers/gpu/drm/amd/amdkfd/soc15_int.h #define SOC15_CONTEXT_ID3_FROM_IH_ENTRY(entry) (le32_to_cpu(entry[7]))
entry             114 drivers/gpu/drm/amd/display/amdgpu_dm/amdgpu_dm_irq.c 	struct list_head *entry;
entry             120 drivers/gpu/drm/amd/display/amdgpu_dm/amdgpu_dm_irq.c 	list_for_each(entry, handler_list) {
entry             121 drivers/gpu/drm/amd/display/amdgpu_dm/amdgpu_dm_irq.c 		handler_data = list_entry(entry,
entry             148 drivers/gpu/drm/amd/display/amdgpu_dm/amdgpu_dm_irq.c 	struct list_head *entry, *tmp;
entry             168 drivers/gpu/drm/amd/display/amdgpu_dm/amdgpu_dm_irq.c 	list_for_each_safe(entry, tmp, hnd_list) {
entry             170 drivers/gpu/drm/amd/display/amdgpu_dm/amdgpu_dm_irq.c 		handler = list_entry(entry, struct amdgpu_dm_irq_handler_data,
entry             531 drivers/gpu/drm/amd/display/amdgpu_dm/amdgpu_dm_irq.c 	struct list_head *entry;
entry             537 drivers/gpu/drm/amd/display/amdgpu_dm/amdgpu_dm_irq.c 		entry,
entry             540 drivers/gpu/drm/amd/display/amdgpu_dm/amdgpu_dm_irq.c 		handler_data = list_entry(entry,
entry             563 drivers/gpu/drm/amd/display/amdgpu_dm/amdgpu_dm_irq.c 				 struct amdgpu_iv_entry *entry)
entry             569 drivers/gpu/drm/amd/display/amdgpu_dm/amdgpu_dm_irq.c 			entry->src_id,
entry             570 drivers/gpu/drm/amd/display/amdgpu_dm/amdgpu_dm_irq.c 			entry->src_data[0]);
entry            2989 drivers/gpu/drm/amd/display/dc/core/dc_link_dp.c 	uint32_t entry;
entry            3004 drivers/gpu/drm/amd/display/dc/core/dc_link_dp.c 		for (entry = 0; entry < 16; entry += 2) {
entry            3007 drivers/gpu/drm/amd/display/dc/core/dc_link_dp.c 			link_rate_in_khz = (supported_link_rates[entry+1] * 0x100 +
entry            3008 drivers/gpu/drm/amd/display/dc/core/dc_link_dp.c 										supported_link_rates[entry]) * 200;
entry             152 drivers/gpu/drm/amd/display/include/logger_interface.h #define DISPLAY_STATS_BEGIN(entry) (void)(entry)
entry             156 drivers/gpu/drm/amd/display/include/logger_interface.h #define DISPLAY_STATS_END(entry) (void)(entry)
entry              83 drivers/gpu/drm/amd/powerplay/hwmgr/common_baco.c 				 const struct soc15_baco_cmd_entry *entry,
entry              90 drivers/gpu/drm/amd/powerplay/hwmgr/common_baco.c 		if ((entry[i].cmd == CMD_WRITE) ||
entry              91 drivers/gpu/drm/amd/powerplay/hwmgr/common_baco.c 		    (entry[i].cmd == CMD_READMODIFYWRITE) ||
entry              92 drivers/gpu/drm/amd/powerplay/hwmgr/common_baco.c 		    (entry[i].cmd == CMD_WAITFOR))
entry              93 drivers/gpu/drm/amd/powerplay/hwmgr/common_baco.c 			reg = adev->reg_offset[entry[i].hwip][entry[i].inst][entry[i].seg]
entry              94 drivers/gpu/drm/amd/powerplay/hwmgr/common_baco.c 				+ entry[i].reg_offset;
entry              95 drivers/gpu/drm/amd/powerplay/hwmgr/common_baco.c 		if (!baco_cmd_handler(hwmgr, entry[i].cmd, reg, entry[i].mask,
entry              96 drivers/gpu/drm/amd/powerplay/hwmgr/common_baco.c 				     entry[i].shift, entry[i].val, entry[i].timeout))
entry              48 drivers/gpu/drm/amd/powerplay/hwmgr/common_baco.h 					const struct soc15_baco_cmd_entry *entry,
entry            1374 drivers/gpu/drm/amd/powerplay/hwmgr/ppatomctrl.c 		table->entry[i].ucVco_setting = psmu_info->asSclkFcwRangeEntry[i].ucVco_setting;
entry            1375 drivers/gpu/drm/amd/powerplay/hwmgr/ppatomctrl.c 		table->entry[i].ucPostdiv = psmu_info->asSclkFcwRangeEntry[i].ucPostdiv;
entry            1376 drivers/gpu/drm/amd/powerplay/hwmgr/ppatomctrl.c 		table->entry[i].usFcw_pcc =
entry            1378 drivers/gpu/drm/amd/powerplay/hwmgr/ppatomctrl.c 		table->entry[i].usFcw_trans_upper =
entry            1380 drivers/gpu/drm/amd/powerplay/hwmgr/ppatomctrl.c 		table->entry[i].usRcw_trans_lower =
entry             244 drivers/gpu/drm/amd/powerplay/hwmgr/ppatomctrl.h 	struct pp_atom_ctrl_sclk_range_table_entry entry[MAX_SCLK_RANGE];
entry            1126 drivers/gpu/drm/amd/powerplay/hwmgr/processpptables.c 		const UVDClockInfo *entry =
entry            1129 drivers/gpu/drm/amd/powerplay/hwmgr/processpptables.c 		uvd_table->entries[i].vclk = ((unsigned long)entry->ucVClkHigh << 16)
entry            1130 drivers/gpu/drm/amd/powerplay/hwmgr/processpptables.c 					 | le16_to_cpu(entry->usVClkLow);
entry            1131 drivers/gpu/drm/amd/powerplay/hwmgr/processpptables.c 		uvd_table->entries[i].dclk = ((unsigned long)entry->ucDClkHigh << 16)
entry            1132 drivers/gpu/drm/amd/powerplay/hwmgr/processpptables.c 					 | le16_to_cpu(entry->usDClkLow);
entry            1158 drivers/gpu/drm/amd/powerplay/hwmgr/processpptables.c 		const VCEClockInfo *entry = &array->entries[table->entries[i].ucVCEClockInfoIndex];
entry            1161 drivers/gpu/drm/amd/powerplay/hwmgr/processpptables.c 		vce_table->entries[i].evclk = ((unsigned long)entry->ucEVClkHigh << 16)
entry            1162 drivers/gpu/drm/amd/powerplay/hwmgr/processpptables.c 					| le16_to_cpu(entry->usEVClkLow);
entry            1163 drivers/gpu/drm/amd/powerplay/hwmgr/processpptables.c 		vce_table->entries[i].ecclk = ((unsigned long)entry->ucECClkHigh << 16)
entry            1164 drivers/gpu/drm/amd/powerplay/hwmgr/processpptables.c 					| le16_to_cpu(entry->usECClkLow);
entry             779 drivers/gpu/drm/amd/powerplay/hwmgr/smu10_hwmgr.c 		    unsigned long entry, struct pp_power_state *ps)
entry             788 drivers/gpu/drm/amd/powerplay/hwmgr/smu10_hwmgr.c 	result = pp_tables_get_entry(hwmgr, entry, ps,
entry             108 drivers/gpu/drm/amd/powerplay/hwmgr/smu10_hwmgr.h 			uint32_t entry : 1;
entry            1374 drivers/gpu/drm/amd/powerplay/hwmgr/smu8_hwmgr.c 		    unsigned long entry, struct pp_power_state *ps)
entry            1383 drivers/gpu/drm/amd/powerplay/hwmgr/smu8_hwmgr.c 	result = pp_tables_get_entry(hwmgr, entry, ps,
entry             131 drivers/gpu/drm/amd/powerplay/hwmgr/smu8_hwmgr.h 			uint32_t entry : 1;
entry             590 drivers/gpu/drm/amd/powerplay/hwmgr/smu_helper.c 			   struct amdgpu_iv_entry *entry)
entry             592 drivers/gpu/drm/amd/powerplay/hwmgr/smu_helper.c 	uint32_t client_id = entry->client_id;
entry             593 drivers/gpu/drm/amd/powerplay/hwmgr/smu_helper.c 	uint32_t src_id = entry->src_id;
entry             110 drivers/gpu/drm/amd/powerplay/hwmgr/smu_helper.h 			   struct amdgpu_iv_entry *entry);
entry              98 drivers/gpu/drm/amd/powerplay/inc/smu_ucode_xfer_vi.h 	struct SMU_Entry entry[SMU_MAX_ENTRIES];
entry            1528 drivers/gpu/drm/amd/powerplay/smu_v11_0.c 				 struct amdgpu_iv_entry *entry)
entry            1530 drivers/gpu/drm/amd/powerplay/smu_v11_0.c 	uint32_t client_id = entry->client_id;
entry            1531 drivers/gpu/drm/amd/powerplay/smu_v11_0.c 	uint32_t src_id = entry->src_id;
entry            1746 drivers/gpu/drm/amd/powerplay/smumgr/ci_smumgr.c 	const struct ci_mc_reg_entry *entry,
entry            1754 drivers/gpu/drm/amd/powerplay/smumgr/ci_smumgr.c 			data->value[i] = PP_HOST_TO_SMC_UL(entry->mc_data[j]);
entry            1714 drivers/gpu/drm/amd/powerplay/smumgr/iceland_smumgr.c 	const struct iceland_mc_reg_entry *entry,
entry            1722 drivers/gpu/drm/amd/powerplay/smumgr/iceland_smumgr.c 			data->value[i] = PP_HOST_TO_SMC_UL(entry->mc_data[j]);
entry             810 drivers/gpu/drm/amd/powerplay/smumgr/polaris10_smumgr.c 			table->SclkFcwRangeTable[i].vco_setting = range_table_from_vbios.entry[i].ucVco_setting;
entry             811 drivers/gpu/drm/amd/powerplay/smumgr/polaris10_smumgr.c 			table->SclkFcwRangeTable[i].postdiv = range_table_from_vbios.entry[i].ucPostdiv;
entry             812 drivers/gpu/drm/amd/powerplay/smumgr/polaris10_smumgr.c 			table->SclkFcwRangeTable[i].fcw_pcc = range_table_from_vbios.entry[i].usFcw_pcc;
entry             814 drivers/gpu/drm/amd/powerplay/smumgr/polaris10_smumgr.c 			table->SclkFcwRangeTable[i].fcw_trans_upper = range_table_from_vbios.entry[i].usFcw_trans_upper;
entry             815 drivers/gpu/drm/amd/powerplay/smumgr/polaris10_smumgr.c 			table->SclkFcwRangeTable[i].fcw_trans_lower = range_table_from_vbios.entry[i].usRcw_trans_lower;
entry             125 drivers/gpu/drm/amd/powerplay/smumgr/smu10_smumgr.c 	PP_ASSERT_WITH_CODE(priv->smu_tables.entry[table_id].version != 0,
entry             127 drivers/gpu/drm/amd/powerplay/smumgr/smu10_smumgr.c 	PP_ASSERT_WITH_CODE(priv->smu_tables.entry[table_id].size != 0,
entry             131 drivers/gpu/drm/amd/powerplay/smumgr/smu10_smumgr.c 			upper_32_bits(priv->smu_tables.entry[table_id].mc_addr));
entry             134 drivers/gpu/drm/amd/powerplay/smumgr/smu10_smumgr.c 			lower_32_bits(priv->smu_tables.entry[table_id].mc_addr));
entry             137 drivers/gpu/drm/amd/powerplay/smumgr/smu10_smumgr.c 			priv->smu_tables.entry[table_id].table_id);
entry             142 drivers/gpu/drm/amd/powerplay/smumgr/smu10_smumgr.c 	memcpy(table, (uint8_t *)priv->smu_tables.entry[table_id].table,
entry             143 drivers/gpu/drm/amd/powerplay/smumgr/smu10_smumgr.c 			priv->smu_tables.entry[table_id].size);
entry             156 drivers/gpu/drm/amd/powerplay/smumgr/smu10_smumgr.c 	PP_ASSERT_WITH_CODE(priv->smu_tables.entry[table_id].version != 0,
entry             158 drivers/gpu/drm/amd/powerplay/smumgr/smu10_smumgr.c 	PP_ASSERT_WITH_CODE(priv->smu_tables.entry[table_id].size != 0,
entry             161 drivers/gpu/drm/amd/powerplay/smumgr/smu10_smumgr.c 	memcpy(priv->smu_tables.entry[table_id].table, table,
entry             162 drivers/gpu/drm/amd/powerplay/smumgr/smu10_smumgr.c 			priv->smu_tables.entry[table_id].size);
entry             166 drivers/gpu/drm/amd/powerplay/smumgr/smu10_smumgr.c 			upper_32_bits(priv->smu_tables.entry[table_id].mc_addr));
entry             169 drivers/gpu/drm/amd/powerplay/smumgr/smu10_smumgr.c 			lower_32_bits(priv->smu_tables.entry[table_id].mc_addr));
entry             172 drivers/gpu/drm/amd/powerplay/smumgr/smu10_smumgr.c 			priv->smu_tables.entry[table_id].table_id);
entry             200 drivers/gpu/drm/amd/powerplay/smumgr/smu10_smumgr.c 		amdgpu_bo_free_kernel(&priv->smu_tables.entry[SMU10_WMTABLE].handle,
entry             201 drivers/gpu/drm/amd/powerplay/smumgr/smu10_smumgr.c 					&priv->smu_tables.entry[SMU10_WMTABLE].mc_addr,
entry             202 drivers/gpu/drm/amd/powerplay/smumgr/smu10_smumgr.c 					&priv->smu_tables.entry[SMU10_WMTABLE].table);
entry             203 drivers/gpu/drm/amd/powerplay/smumgr/smu10_smumgr.c 		amdgpu_bo_free_kernel(&priv->smu_tables.entry[SMU10_CLOCKTABLE].handle,
entry             204 drivers/gpu/drm/amd/powerplay/smumgr/smu10_smumgr.c 					&priv->smu_tables.entry[SMU10_CLOCKTABLE].mc_addr,
entry             205 drivers/gpu/drm/amd/powerplay/smumgr/smu10_smumgr.c 					&priv->smu_tables.entry[SMU10_CLOCKTABLE].table);
entry             248 drivers/gpu/drm/amd/powerplay/smumgr/smu10_smumgr.c 			&priv->smu_tables.entry[SMU10_WMTABLE].handle,
entry             249 drivers/gpu/drm/amd/powerplay/smumgr/smu10_smumgr.c 			&priv->smu_tables.entry[SMU10_WMTABLE].mc_addr,
entry             250 drivers/gpu/drm/amd/powerplay/smumgr/smu10_smumgr.c 			&priv->smu_tables.entry[SMU10_WMTABLE].table);
entry             255 drivers/gpu/drm/amd/powerplay/smumgr/smu10_smumgr.c 	priv->smu_tables.entry[SMU10_WMTABLE].version = 0x01;
entry             256 drivers/gpu/drm/amd/powerplay/smumgr/smu10_smumgr.c 	priv->smu_tables.entry[SMU10_WMTABLE].size = sizeof(Watermarks_t);
entry             257 drivers/gpu/drm/amd/powerplay/smumgr/smu10_smumgr.c 	priv->smu_tables.entry[SMU10_WMTABLE].table_id = TABLE_WATERMARKS;
entry             264 drivers/gpu/drm/amd/powerplay/smumgr/smu10_smumgr.c 			&priv->smu_tables.entry[SMU10_CLOCKTABLE].handle,
entry             265 drivers/gpu/drm/amd/powerplay/smumgr/smu10_smumgr.c 			&priv->smu_tables.entry[SMU10_CLOCKTABLE].mc_addr,
entry             266 drivers/gpu/drm/amd/powerplay/smumgr/smu10_smumgr.c 			&priv->smu_tables.entry[SMU10_CLOCKTABLE].table);
entry             271 drivers/gpu/drm/amd/powerplay/smumgr/smu10_smumgr.c 	priv->smu_tables.entry[SMU10_CLOCKTABLE].version = 0x01;
entry             272 drivers/gpu/drm/amd/powerplay/smumgr/smu10_smumgr.c 	priv->smu_tables.entry[SMU10_CLOCKTABLE].size = sizeof(DpmClocks_t);
entry             273 drivers/gpu/drm/amd/powerplay/smumgr/smu10_smumgr.c 	priv->smu_tables.entry[SMU10_CLOCKTABLE].table_id = TABLE_DPMCLOCKS;
entry             278 drivers/gpu/drm/amd/powerplay/smumgr/smu10_smumgr.c 	amdgpu_bo_free_kernel(&priv->smu_tables.entry[SMU10_WMTABLE].handle,
entry             279 drivers/gpu/drm/amd/powerplay/smumgr/smu10_smumgr.c 				&priv->smu_tables.entry[SMU10_WMTABLE].mc_addr,
entry             280 drivers/gpu/drm/amd/powerplay/smumgr/smu10_smumgr.c 				&priv->smu_tables.entry[SMU10_WMTABLE].table);
entry              42 drivers/gpu/drm/amd/powerplay/smumgr/smu10_smumgr.h 	struct smu_table_entry entry[MAX_SMU_TABLE];
entry             307 drivers/gpu/drm/amd/powerplay/smumgr/smu7_smumgr.c 						struct SMU_Entry *entry)
entry             317 drivers/gpu/drm/amd/powerplay/smumgr/smu7_smumgr.c 		entry->version = info.fw_version;
entry             318 drivers/gpu/drm/amd/powerplay/smumgr/smu7_smumgr.c 		entry->id = (uint16_t)fw_type;
entry             319 drivers/gpu/drm/amd/powerplay/smumgr/smu7_smumgr.c 		entry->image_addr_high = upper_32_bits(info.mc_addr);
entry             320 drivers/gpu/drm/amd/powerplay/smumgr/smu7_smumgr.c 		entry->image_addr_low = lower_32_bits(info.mc_addr);
entry             321 drivers/gpu/drm/amd/powerplay/smumgr/smu7_smumgr.c 		entry->meta_data_addr_high = 0;
entry             322 drivers/gpu/drm/amd/powerplay/smumgr/smu7_smumgr.c 		entry->meta_data_addr_low = 0;
entry             327 drivers/gpu/drm/amd/powerplay/smumgr/smu7_smumgr.c 		entry->data_size_byte = info.image_size;
entry             328 drivers/gpu/drm/amd/powerplay/smumgr/smu7_smumgr.c 		entry->num_register_entries = 0;
entry             333 drivers/gpu/drm/amd/powerplay/smumgr/smu7_smumgr.c 		entry->flags = 1;
entry             335 drivers/gpu/drm/amd/powerplay/smumgr/smu7_smumgr.c 		entry->flags = 0;
entry             393 drivers/gpu/drm/amd/powerplay/smumgr/smu7_smumgr.c 				UCODE_ID_RLC_G, &toc->entry[toc->num_entries++]),
entry             396 drivers/gpu/drm/amd/powerplay/smumgr/smu7_smumgr.c 				UCODE_ID_CP_CE, &toc->entry[toc->num_entries++]),
entry             399 drivers/gpu/drm/amd/powerplay/smumgr/smu7_smumgr.c 				UCODE_ID_CP_PFP, &toc->entry[toc->num_entries++]),
entry             402 drivers/gpu/drm/amd/powerplay/smumgr/smu7_smumgr.c 				UCODE_ID_CP_ME, &toc->entry[toc->num_entries++]),
entry             405 drivers/gpu/drm/amd/powerplay/smumgr/smu7_smumgr.c 				UCODE_ID_CP_MEC, &toc->entry[toc->num_entries++]),
entry             408 drivers/gpu/drm/amd/powerplay/smumgr/smu7_smumgr.c 				UCODE_ID_CP_MEC_JT1, &toc->entry[toc->num_entries++]),
entry             411 drivers/gpu/drm/amd/powerplay/smumgr/smu7_smumgr.c 				UCODE_ID_CP_MEC_JT2, &toc->entry[toc->num_entries++]),
entry             414 drivers/gpu/drm/amd/powerplay/smumgr/smu7_smumgr.c 				UCODE_ID_SDMA0, &toc->entry[toc->num_entries++]),
entry             417 drivers/gpu/drm/amd/powerplay/smumgr/smu7_smumgr.c 				UCODE_ID_SDMA1, &toc->entry[toc->num_entries++]),
entry             421 drivers/gpu/drm/amd/powerplay/smumgr/smu7_smumgr.c 				UCODE_ID_MEC_STORAGE, &toc->entry[toc->num_entries++]),
entry             586 drivers/gpu/drm/amd/powerplay/smumgr/smu8_smumgr.c 				struct smu8_buffer_entry *entry)
entry             591 drivers/gpu/drm/amd/powerplay/smumgr/smu8_smumgr.c 	entry->data_size = ulsize_byte;
entry             592 drivers/gpu/drm/amd/powerplay/smumgr/smu8_smumgr.c 	entry->kaddr = (char *) smu8_smu->smu_buffer.kaddr +
entry             594 drivers/gpu/drm/amd/powerplay/smumgr/smu8_smumgr.c 	entry->mc_addr = smu8_smu->smu_buffer.mc_addr + smu8_smu->smu_buffer_used_bytes;
entry             595 drivers/gpu/drm/amd/powerplay/smumgr/smu8_smumgr.c 	entry->firmware_ID = scratch_type;
entry            2092 drivers/gpu/drm/amd/powerplay/smumgr/tonga_smumgr.c 	const struct tonga_mc_reg_entry *entry,
entry            2100 drivers/gpu/drm/amd/powerplay/smumgr/tonga_smumgr.c 			data->value[i] = PP_HOST_TO_SMC_UL(entry->mc_data[j]);
entry              46 drivers/gpu/drm/amd/powerplay/smumgr/vega10_smumgr.c 	PP_ASSERT_WITH_CODE(priv->smu_tables.entry[table_id].version != 0,
entry              48 drivers/gpu/drm/amd/powerplay/smumgr/vega10_smumgr.c 	PP_ASSERT_WITH_CODE(priv->smu_tables.entry[table_id].size != 0,
entry              52 drivers/gpu/drm/amd/powerplay/smumgr/vega10_smumgr.c 			upper_32_bits(priv->smu_tables.entry[table_id].mc_addr));
entry              55 drivers/gpu/drm/amd/powerplay/smumgr/vega10_smumgr.c 			lower_32_bits(priv->smu_tables.entry[table_id].mc_addr));
entry              58 drivers/gpu/drm/amd/powerplay/smumgr/vega10_smumgr.c 			priv->smu_tables.entry[table_id].table_id);
entry              63 drivers/gpu/drm/amd/powerplay/smumgr/vega10_smumgr.c 	memcpy(table, priv->smu_tables.entry[table_id].table,
entry              64 drivers/gpu/drm/amd/powerplay/smumgr/vega10_smumgr.c 			priv->smu_tables.entry[table_id].size);
entry              76 drivers/gpu/drm/amd/powerplay/smumgr/vega10_smumgr.c 	PP_ASSERT_WITH_CODE(priv->smu_tables.entry[table_id].version != 0,
entry              78 drivers/gpu/drm/amd/powerplay/smumgr/vega10_smumgr.c 	PP_ASSERT_WITH_CODE(priv->smu_tables.entry[table_id].size != 0,
entry              81 drivers/gpu/drm/amd/powerplay/smumgr/vega10_smumgr.c 	memcpy(priv->smu_tables.entry[table_id].table, table,
entry              82 drivers/gpu/drm/amd/powerplay/smumgr/vega10_smumgr.c 			priv->smu_tables.entry[table_id].size);
entry              86 drivers/gpu/drm/amd/powerplay/smumgr/vega10_smumgr.c 			upper_32_bits(priv->smu_tables.entry[table_id].mc_addr));
entry              89 drivers/gpu/drm/amd/powerplay/smumgr/vega10_smumgr.c 			lower_32_bits(priv->smu_tables.entry[table_id].mc_addr));
entry              92 drivers/gpu/drm/amd/powerplay/smumgr/vega10_smumgr.c 			priv->smu_tables.entry[table_id].table_id);
entry             135 drivers/gpu/drm/amd/powerplay/smumgr/vega10_smumgr.c 	if (priv->smu_tables.entry[TOOLSTABLE].mc_addr) {
entry             138 drivers/gpu/drm/amd/powerplay/smumgr/vega10_smumgr.c 				upper_32_bits(priv->smu_tables.entry[TOOLSTABLE].mc_addr));
entry             141 drivers/gpu/drm/amd/powerplay/smumgr/vega10_smumgr.c 				lower_32_bits(priv->smu_tables.entry[TOOLSTABLE].mc_addr));
entry             201 drivers/gpu/drm/amd/powerplay/smumgr/vega10_smumgr.c 			&priv->smu_tables.entry[PPTABLE].handle,
entry             202 drivers/gpu/drm/amd/powerplay/smumgr/vega10_smumgr.c 			&priv->smu_tables.entry[PPTABLE].mc_addr,
entry             203 drivers/gpu/drm/amd/powerplay/smumgr/vega10_smumgr.c 			&priv->smu_tables.entry[PPTABLE].table);
entry             207 drivers/gpu/drm/amd/powerplay/smumgr/vega10_smumgr.c 	priv->smu_tables.entry[PPTABLE].version = 0x01;
entry             208 drivers/gpu/drm/amd/powerplay/smumgr/vega10_smumgr.c 	priv->smu_tables.entry[PPTABLE].size = sizeof(PPTable_t);
entry             209 drivers/gpu/drm/amd/powerplay/smumgr/vega10_smumgr.c 	priv->smu_tables.entry[PPTABLE].table_id = TABLE_PPTABLE;
entry             216 drivers/gpu/drm/amd/powerplay/smumgr/vega10_smumgr.c 			&priv->smu_tables.entry[WMTABLE].handle,
entry             217 drivers/gpu/drm/amd/powerplay/smumgr/vega10_smumgr.c 			&priv->smu_tables.entry[WMTABLE].mc_addr,
entry             218 drivers/gpu/drm/amd/powerplay/smumgr/vega10_smumgr.c 			&priv->smu_tables.entry[WMTABLE].table);
entry             223 drivers/gpu/drm/amd/powerplay/smumgr/vega10_smumgr.c 	priv->smu_tables.entry[WMTABLE].version = 0x01;
entry             224 drivers/gpu/drm/amd/powerplay/smumgr/vega10_smumgr.c 	priv->smu_tables.entry[WMTABLE].size = sizeof(Watermarks_t);
entry             225 drivers/gpu/drm/amd/powerplay/smumgr/vega10_smumgr.c 	priv->smu_tables.entry[WMTABLE].table_id = TABLE_WATERMARKS;
entry             232 drivers/gpu/drm/amd/powerplay/smumgr/vega10_smumgr.c 			&priv->smu_tables.entry[AVFSTABLE].handle,
entry             233 drivers/gpu/drm/amd/powerplay/smumgr/vega10_smumgr.c 			&priv->smu_tables.entry[AVFSTABLE].mc_addr,
entry             234 drivers/gpu/drm/amd/powerplay/smumgr/vega10_smumgr.c 			&priv->smu_tables.entry[AVFSTABLE].table);
entry             239 drivers/gpu/drm/amd/powerplay/smumgr/vega10_smumgr.c 	priv->smu_tables.entry[AVFSTABLE].version = 0x01;
entry             240 drivers/gpu/drm/amd/powerplay/smumgr/vega10_smumgr.c 	priv->smu_tables.entry[AVFSTABLE].size = sizeof(AvfsTable_t);
entry             241 drivers/gpu/drm/amd/powerplay/smumgr/vega10_smumgr.c 	priv->smu_tables.entry[AVFSTABLE].table_id = TABLE_AVFS;
entry             249 drivers/gpu/drm/amd/powerplay/smumgr/vega10_smumgr.c 				&priv->smu_tables.entry[TOOLSTABLE].handle,
entry             250 drivers/gpu/drm/amd/powerplay/smumgr/vega10_smumgr.c 				&priv->smu_tables.entry[TOOLSTABLE].mc_addr,
entry             251 drivers/gpu/drm/amd/powerplay/smumgr/vega10_smumgr.c 				&priv->smu_tables.entry[TOOLSTABLE].table);
entry             254 drivers/gpu/drm/amd/powerplay/smumgr/vega10_smumgr.c 		priv->smu_tables.entry[TOOLSTABLE].version = 0x01;
entry             255 drivers/gpu/drm/amd/powerplay/smumgr/vega10_smumgr.c 		priv->smu_tables.entry[TOOLSTABLE].size = tools_size;
entry             256 drivers/gpu/drm/amd/powerplay/smumgr/vega10_smumgr.c 		priv->smu_tables.entry[TOOLSTABLE].table_id = TABLE_PMSTATUSLOG;
entry             264 drivers/gpu/drm/amd/powerplay/smumgr/vega10_smumgr.c 			&priv->smu_tables.entry[AVFSFUSETABLE].handle,
entry             265 drivers/gpu/drm/amd/powerplay/smumgr/vega10_smumgr.c 			&priv->smu_tables.entry[AVFSFUSETABLE].mc_addr,
entry             266 drivers/gpu/drm/amd/powerplay/smumgr/vega10_smumgr.c 			&priv->smu_tables.entry[AVFSFUSETABLE].table);
entry             270 drivers/gpu/drm/amd/powerplay/smumgr/vega10_smumgr.c 	priv->smu_tables.entry[AVFSFUSETABLE].version = 0x01;
entry             271 drivers/gpu/drm/amd/powerplay/smumgr/vega10_smumgr.c 	priv->smu_tables.entry[AVFSFUSETABLE].size = sizeof(AvfsFuseOverride_t);
entry             272 drivers/gpu/drm/amd/powerplay/smumgr/vega10_smumgr.c 	priv->smu_tables.entry[AVFSFUSETABLE].table_id = TABLE_AVFS_FUSE_OVERRIDE;
entry             278 drivers/gpu/drm/amd/powerplay/smumgr/vega10_smumgr.c 	if (priv->smu_tables.entry[TOOLSTABLE].table)
entry             279 drivers/gpu/drm/amd/powerplay/smumgr/vega10_smumgr.c 		amdgpu_bo_free_kernel(&priv->smu_tables.entry[TOOLSTABLE].handle,
entry             280 drivers/gpu/drm/amd/powerplay/smumgr/vega10_smumgr.c 				&priv->smu_tables.entry[TOOLSTABLE].mc_addr,
entry             281 drivers/gpu/drm/amd/powerplay/smumgr/vega10_smumgr.c 				&priv->smu_tables.entry[TOOLSTABLE].table);
entry             283 drivers/gpu/drm/amd/powerplay/smumgr/vega10_smumgr.c 	amdgpu_bo_free_kernel(&priv->smu_tables.entry[AVFSTABLE].handle,
entry             284 drivers/gpu/drm/amd/powerplay/smumgr/vega10_smumgr.c 				&priv->smu_tables.entry[AVFSTABLE].mc_addr,
entry             285 drivers/gpu/drm/amd/powerplay/smumgr/vega10_smumgr.c 				&priv->smu_tables.entry[AVFSTABLE].table);
entry             287 drivers/gpu/drm/amd/powerplay/smumgr/vega10_smumgr.c 	amdgpu_bo_free_kernel(&priv->smu_tables.entry[WMTABLE].handle,
entry             288 drivers/gpu/drm/amd/powerplay/smumgr/vega10_smumgr.c 				&priv->smu_tables.entry[WMTABLE].mc_addr,
entry             289 drivers/gpu/drm/amd/powerplay/smumgr/vega10_smumgr.c 				&priv->smu_tables.entry[WMTABLE].table);
entry             291 drivers/gpu/drm/amd/powerplay/smumgr/vega10_smumgr.c 	amdgpu_bo_free_kernel(&priv->smu_tables.entry[PPTABLE].handle,
entry             292 drivers/gpu/drm/amd/powerplay/smumgr/vega10_smumgr.c 			&priv->smu_tables.entry[PPTABLE].mc_addr,
entry             293 drivers/gpu/drm/amd/powerplay/smumgr/vega10_smumgr.c 			&priv->smu_tables.entry[PPTABLE].table);
entry             305 drivers/gpu/drm/amd/powerplay/smumgr/vega10_smumgr.c 		amdgpu_bo_free_kernel(&priv->smu_tables.entry[PPTABLE].handle,
entry             306 drivers/gpu/drm/amd/powerplay/smumgr/vega10_smumgr.c 				&priv->smu_tables.entry[PPTABLE].mc_addr,
entry             307 drivers/gpu/drm/amd/powerplay/smumgr/vega10_smumgr.c 				&priv->smu_tables.entry[PPTABLE].table);
entry             308 drivers/gpu/drm/amd/powerplay/smumgr/vega10_smumgr.c 		amdgpu_bo_free_kernel(&priv->smu_tables.entry[WMTABLE].handle,
entry             309 drivers/gpu/drm/amd/powerplay/smumgr/vega10_smumgr.c 					&priv->smu_tables.entry[WMTABLE].mc_addr,
entry             310 drivers/gpu/drm/amd/powerplay/smumgr/vega10_smumgr.c 					&priv->smu_tables.entry[WMTABLE].table);
entry             311 drivers/gpu/drm/amd/powerplay/smumgr/vega10_smumgr.c 		amdgpu_bo_free_kernel(&priv->smu_tables.entry[AVFSTABLE].handle,
entry             312 drivers/gpu/drm/amd/powerplay/smumgr/vega10_smumgr.c 					&priv->smu_tables.entry[AVFSTABLE].mc_addr,
entry             313 drivers/gpu/drm/amd/powerplay/smumgr/vega10_smumgr.c 					&priv->smu_tables.entry[AVFSTABLE].table);
entry             314 drivers/gpu/drm/amd/powerplay/smumgr/vega10_smumgr.c 		if (priv->smu_tables.entry[TOOLSTABLE].table)
entry             315 drivers/gpu/drm/amd/powerplay/smumgr/vega10_smumgr.c 			amdgpu_bo_free_kernel(&priv->smu_tables.entry[TOOLSTABLE].handle,
entry             316 drivers/gpu/drm/amd/powerplay/smumgr/vega10_smumgr.c 					&priv->smu_tables.entry[TOOLSTABLE].mc_addr,
entry             317 drivers/gpu/drm/amd/powerplay/smumgr/vega10_smumgr.c 					&priv->smu_tables.entry[TOOLSTABLE].table);
entry             318 drivers/gpu/drm/amd/powerplay/smumgr/vega10_smumgr.c 		amdgpu_bo_free_kernel(&priv->smu_tables.entry[AVFSFUSETABLE].handle,
entry             319 drivers/gpu/drm/amd/powerplay/smumgr/vega10_smumgr.c 					&priv->smu_tables.entry[AVFSFUSETABLE].mc_addr,
entry             320 drivers/gpu/drm/amd/powerplay/smumgr/vega10_smumgr.c 					&priv->smu_tables.entry[AVFSFUSETABLE].table);
entry              38 drivers/gpu/drm/amd/powerplay/smumgr/vega10_smumgr.h 	struct smu_table_entry entry[MAX_SMU_TABLE];
entry              49 drivers/gpu/drm/amd/powerplay/smumgr/vega12_smumgr.c 	PP_ASSERT_WITH_CODE(priv->smu_tables.entry[table_id].version != 0,
entry              51 drivers/gpu/drm/amd/powerplay/smumgr/vega12_smumgr.c 	PP_ASSERT_WITH_CODE(priv->smu_tables.entry[table_id].size != 0,
entry              55 drivers/gpu/drm/amd/powerplay/smumgr/vega12_smumgr.c 			upper_32_bits(priv->smu_tables.entry[table_id].mc_addr)) == 0,
entry              59 drivers/gpu/drm/amd/powerplay/smumgr/vega12_smumgr.c 			lower_32_bits(priv->smu_tables.entry[table_id].mc_addr)) == 0,
entry              71 drivers/gpu/drm/amd/powerplay/smumgr/vega12_smumgr.c 	memcpy(table, priv->smu_tables.entry[table_id].table,
entry              72 drivers/gpu/drm/amd/powerplay/smumgr/vega12_smumgr.c 			priv->smu_tables.entry[table_id].size);
entry              90 drivers/gpu/drm/amd/powerplay/smumgr/vega12_smumgr.c 	PP_ASSERT_WITH_CODE(priv->smu_tables.entry[table_id].version != 0,
entry              92 drivers/gpu/drm/amd/powerplay/smumgr/vega12_smumgr.c 	PP_ASSERT_WITH_CODE(priv->smu_tables.entry[table_id].size != 0,
entry              95 drivers/gpu/drm/amd/powerplay/smumgr/vega12_smumgr.c 	memcpy(priv->smu_tables.entry[table_id].table, table,
entry              96 drivers/gpu/drm/amd/powerplay/smumgr/vega12_smumgr.c 			priv->smu_tables.entry[table_id].size);
entry             100 drivers/gpu/drm/amd/powerplay/smumgr/vega12_smumgr.c 			upper_32_bits(priv->smu_tables.entry[table_id].mc_addr)) == 0,
entry             105 drivers/gpu/drm/amd/powerplay/smumgr/vega12_smumgr.c 			lower_32_bits(priv->smu_tables.entry[table_id].mc_addr)) == 0,
entry             191 drivers/gpu/drm/amd/powerplay/smumgr/vega12_smumgr.c 	if (priv->smu_tables.entry[TABLE_PMSTATUSLOG].mc_addr) {
entry             194 drivers/gpu/drm/amd/powerplay/smumgr/vega12_smumgr.c 				upper_32_bits(priv->smu_tables.entry[TABLE_PMSTATUSLOG].mc_addr)))
entry             197 drivers/gpu/drm/amd/powerplay/smumgr/vega12_smumgr.c 					lower_32_bits(priv->smu_tables.entry[TABLE_PMSTATUSLOG].mc_addr));
entry             225 drivers/gpu/drm/amd/powerplay/smumgr/vega12_smumgr.c 			&priv->smu_tables.entry[TABLE_PPTABLE].handle,
entry             226 drivers/gpu/drm/amd/powerplay/smumgr/vega12_smumgr.c 			&priv->smu_tables.entry[TABLE_PPTABLE].mc_addr,
entry             227 drivers/gpu/drm/amd/powerplay/smumgr/vega12_smumgr.c 			&priv->smu_tables.entry[TABLE_PPTABLE].table);
entry             231 drivers/gpu/drm/amd/powerplay/smumgr/vega12_smumgr.c 	priv->smu_tables.entry[TABLE_PPTABLE].version = 0x01;
entry             232 drivers/gpu/drm/amd/powerplay/smumgr/vega12_smumgr.c 	priv->smu_tables.entry[TABLE_PPTABLE].size = sizeof(PPTable_t);
entry             239 drivers/gpu/drm/amd/powerplay/smumgr/vega12_smumgr.c 				      &priv->smu_tables.entry[TABLE_WATERMARKS].handle,
entry             240 drivers/gpu/drm/amd/powerplay/smumgr/vega12_smumgr.c 				      &priv->smu_tables.entry[TABLE_WATERMARKS].mc_addr,
entry             241 drivers/gpu/drm/amd/powerplay/smumgr/vega12_smumgr.c 				      &priv->smu_tables.entry[TABLE_WATERMARKS].table);
entry             246 drivers/gpu/drm/amd/powerplay/smumgr/vega12_smumgr.c 	priv->smu_tables.entry[TABLE_WATERMARKS].version = 0x01;
entry             247 drivers/gpu/drm/amd/powerplay/smumgr/vega12_smumgr.c 	priv->smu_tables.entry[TABLE_WATERMARKS].size = sizeof(Watermarks_t);
entry             255 drivers/gpu/drm/amd/powerplay/smumgr/vega12_smumgr.c 					      &priv->smu_tables.entry[TABLE_PMSTATUSLOG].handle,
entry             256 drivers/gpu/drm/amd/powerplay/smumgr/vega12_smumgr.c 					      &priv->smu_tables.entry[TABLE_PMSTATUSLOG].mc_addr,
entry             257 drivers/gpu/drm/amd/powerplay/smumgr/vega12_smumgr.c 					      &priv->smu_tables.entry[TABLE_PMSTATUSLOG].table);
entry             261 drivers/gpu/drm/amd/powerplay/smumgr/vega12_smumgr.c 		priv->smu_tables.entry[TABLE_PMSTATUSLOG].version = 0x01;
entry             262 drivers/gpu/drm/amd/powerplay/smumgr/vega12_smumgr.c 		priv->smu_tables.entry[TABLE_PMSTATUSLOG].size = tools_size;
entry             270 drivers/gpu/drm/amd/powerplay/smumgr/vega12_smumgr.c 				      &priv->smu_tables.entry[TABLE_AVFS_FUSE_OVERRIDE].handle,
entry             271 drivers/gpu/drm/amd/powerplay/smumgr/vega12_smumgr.c 				      &priv->smu_tables.entry[TABLE_AVFS_FUSE_OVERRIDE].mc_addr,
entry             272 drivers/gpu/drm/amd/powerplay/smumgr/vega12_smumgr.c 				      &priv->smu_tables.entry[TABLE_AVFS_FUSE_OVERRIDE].table);
entry             277 drivers/gpu/drm/amd/powerplay/smumgr/vega12_smumgr.c 	priv->smu_tables.entry[TABLE_AVFS_FUSE_OVERRIDE].version = 0x01;
entry             278 drivers/gpu/drm/amd/powerplay/smumgr/vega12_smumgr.c 	priv->smu_tables.entry[TABLE_AVFS_FUSE_OVERRIDE].size = sizeof(AvfsFuseOverride_t);
entry             285 drivers/gpu/drm/amd/powerplay/smumgr/vega12_smumgr.c 				      &priv->smu_tables.entry[TABLE_OVERDRIVE].handle,
entry             286 drivers/gpu/drm/amd/powerplay/smumgr/vega12_smumgr.c 				      &priv->smu_tables.entry[TABLE_OVERDRIVE].mc_addr,
entry             287 drivers/gpu/drm/amd/powerplay/smumgr/vega12_smumgr.c 				      &priv->smu_tables.entry[TABLE_OVERDRIVE].table);
entry             291 drivers/gpu/drm/amd/powerplay/smumgr/vega12_smumgr.c 	priv->smu_tables.entry[TABLE_OVERDRIVE].version = 0x01;
entry             292 drivers/gpu/drm/amd/powerplay/smumgr/vega12_smumgr.c 	priv->smu_tables.entry[TABLE_OVERDRIVE].size = sizeof(OverDriveTable_t);
entry             299 drivers/gpu/drm/amd/powerplay/smumgr/vega12_smumgr.c 				      &priv->smu_tables.entry[TABLE_SMU_METRICS].handle,
entry             300 drivers/gpu/drm/amd/powerplay/smumgr/vega12_smumgr.c 				      &priv->smu_tables.entry[TABLE_SMU_METRICS].mc_addr,
entry             301 drivers/gpu/drm/amd/powerplay/smumgr/vega12_smumgr.c 				      &priv->smu_tables.entry[TABLE_SMU_METRICS].table);
entry             305 drivers/gpu/drm/amd/powerplay/smumgr/vega12_smumgr.c 	priv->smu_tables.entry[TABLE_SMU_METRICS].version = 0x01;
entry             306 drivers/gpu/drm/amd/powerplay/smumgr/vega12_smumgr.c 	priv->smu_tables.entry[TABLE_SMU_METRICS].size = sizeof(SmuMetrics_t);
entry             311 drivers/gpu/drm/amd/powerplay/smumgr/vega12_smumgr.c 	amdgpu_bo_free_kernel(&priv->smu_tables.entry[TABLE_OVERDRIVE].handle,
entry             312 drivers/gpu/drm/amd/powerplay/smumgr/vega12_smumgr.c 				&priv->smu_tables.entry[TABLE_OVERDRIVE].mc_addr,
entry             313 drivers/gpu/drm/amd/powerplay/smumgr/vega12_smumgr.c 				&priv->smu_tables.entry[TABLE_OVERDRIVE].table);
entry             315 drivers/gpu/drm/amd/powerplay/smumgr/vega12_smumgr.c 	amdgpu_bo_free_kernel(&priv->smu_tables.entry[TABLE_AVFS_FUSE_OVERRIDE].handle,
entry             316 drivers/gpu/drm/amd/powerplay/smumgr/vega12_smumgr.c 				&priv->smu_tables.entry[TABLE_AVFS_FUSE_OVERRIDE].mc_addr,
entry             317 drivers/gpu/drm/amd/powerplay/smumgr/vega12_smumgr.c 				&priv->smu_tables.entry[TABLE_AVFS_FUSE_OVERRIDE].table);
entry             319 drivers/gpu/drm/amd/powerplay/smumgr/vega12_smumgr.c 	if (priv->smu_tables.entry[TABLE_PMSTATUSLOG].table)
entry             320 drivers/gpu/drm/amd/powerplay/smumgr/vega12_smumgr.c 		amdgpu_bo_free_kernel(&priv->smu_tables.entry[TABLE_PMSTATUSLOG].handle,
entry             321 drivers/gpu/drm/amd/powerplay/smumgr/vega12_smumgr.c 				&priv->smu_tables.entry[TABLE_PMSTATUSLOG].mc_addr,
entry             322 drivers/gpu/drm/amd/powerplay/smumgr/vega12_smumgr.c 				&priv->smu_tables.entry[TABLE_PMSTATUSLOG].table);
entry             324 drivers/gpu/drm/amd/powerplay/smumgr/vega12_smumgr.c 	amdgpu_bo_free_kernel(&priv->smu_tables.entry[TABLE_WATERMARKS].handle,
entry             325 drivers/gpu/drm/amd/powerplay/smumgr/vega12_smumgr.c 				&priv->smu_tables.entry[TABLE_WATERMARKS].mc_addr,
entry             326 drivers/gpu/drm/amd/powerplay/smumgr/vega12_smumgr.c 				&priv->smu_tables.entry[TABLE_WATERMARKS].table);
entry             328 drivers/gpu/drm/amd/powerplay/smumgr/vega12_smumgr.c 	amdgpu_bo_free_kernel(&priv->smu_tables.entry[TABLE_PPTABLE].handle,
entry             329 drivers/gpu/drm/amd/powerplay/smumgr/vega12_smumgr.c 			&priv->smu_tables.entry[TABLE_PPTABLE].mc_addr,
entry             330 drivers/gpu/drm/amd/powerplay/smumgr/vega12_smumgr.c 			&priv->smu_tables.entry[TABLE_PPTABLE].table);
entry             343 drivers/gpu/drm/amd/powerplay/smumgr/vega12_smumgr.c 		amdgpu_bo_free_kernel(&priv->smu_tables.entry[TABLE_PPTABLE].handle,
entry             344 drivers/gpu/drm/amd/powerplay/smumgr/vega12_smumgr.c 				      &priv->smu_tables.entry[TABLE_PPTABLE].mc_addr,
entry             345 drivers/gpu/drm/amd/powerplay/smumgr/vega12_smumgr.c 				      &priv->smu_tables.entry[TABLE_PPTABLE].table);
entry             346 drivers/gpu/drm/amd/powerplay/smumgr/vega12_smumgr.c 		amdgpu_bo_free_kernel(&priv->smu_tables.entry[TABLE_WATERMARKS].handle,
entry             347 drivers/gpu/drm/amd/powerplay/smumgr/vega12_smumgr.c 				      &priv->smu_tables.entry[TABLE_WATERMARKS].mc_addr,
entry             348 drivers/gpu/drm/amd/powerplay/smumgr/vega12_smumgr.c 				      &priv->smu_tables.entry[TABLE_WATERMARKS].table);
entry             349 drivers/gpu/drm/amd/powerplay/smumgr/vega12_smumgr.c 		if (priv->smu_tables.entry[TABLE_PMSTATUSLOG].table)
entry             350 drivers/gpu/drm/amd/powerplay/smumgr/vega12_smumgr.c 			amdgpu_bo_free_kernel(&priv->smu_tables.entry[TABLE_PMSTATUSLOG].handle,
entry             351 drivers/gpu/drm/amd/powerplay/smumgr/vega12_smumgr.c 					      &priv->smu_tables.entry[TABLE_PMSTATUSLOG].mc_addr,
entry             352 drivers/gpu/drm/amd/powerplay/smumgr/vega12_smumgr.c 					      &priv->smu_tables.entry[TABLE_PMSTATUSLOG].table);
entry             353 drivers/gpu/drm/amd/powerplay/smumgr/vega12_smumgr.c 		amdgpu_bo_free_kernel(&priv->smu_tables.entry[TABLE_AVFS_FUSE_OVERRIDE].handle,
entry             354 drivers/gpu/drm/amd/powerplay/smumgr/vega12_smumgr.c 				      &priv->smu_tables.entry[TABLE_AVFS_FUSE_OVERRIDE].mc_addr,
entry             355 drivers/gpu/drm/amd/powerplay/smumgr/vega12_smumgr.c 				      &priv->smu_tables.entry[TABLE_AVFS_FUSE_OVERRIDE].table);
entry             356 drivers/gpu/drm/amd/powerplay/smumgr/vega12_smumgr.c 		amdgpu_bo_free_kernel(&priv->smu_tables.entry[TABLE_OVERDRIVE].handle,
entry             357 drivers/gpu/drm/amd/powerplay/smumgr/vega12_smumgr.c 				      &priv->smu_tables.entry[TABLE_OVERDRIVE].mc_addr,
entry             358 drivers/gpu/drm/amd/powerplay/smumgr/vega12_smumgr.c 				      &priv->smu_tables.entry[TABLE_OVERDRIVE].table);
entry             359 drivers/gpu/drm/amd/powerplay/smumgr/vega12_smumgr.c 		amdgpu_bo_free_kernel(&priv->smu_tables.entry[TABLE_SMU_METRICS].handle,
entry             360 drivers/gpu/drm/amd/powerplay/smumgr/vega12_smumgr.c 				      &priv->smu_tables.entry[TABLE_SMU_METRICS].mc_addr,
entry             361 drivers/gpu/drm/amd/powerplay/smumgr/vega12_smumgr.c 				      &priv->smu_tables.entry[TABLE_SMU_METRICS].table);
entry              39 drivers/gpu/drm/amd/powerplay/smumgr/vega12_smumgr.h 	struct smu_table_entry entry[TABLE_COUNT];
entry             171 drivers/gpu/drm/amd/powerplay/smumgr/vega20_smumgr.c 	PP_ASSERT_WITH_CODE(priv->smu_tables.entry[table_id].version != 0,
entry             173 drivers/gpu/drm/amd/powerplay/smumgr/vega20_smumgr.c 	PP_ASSERT_WITH_CODE(priv->smu_tables.entry[table_id].size != 0,
entry             178 drivers/gpu/drm/amd/powerplay/smumgr/vega20_smumgr.c 			upper_32_bits(priv->smu_tables.entry[table_id].mc_addr))) == 0,
entry             183 drivers/gpu/drm/amd/powerplay/smumgr/vega20_smumgr.c 			lower_32_bits(priv->smu_tables.entry[table_id].mc_addr))) == 0,
entry             194 drivers/gpu/drm/amd/powerplay/smumgr/vega20_smumgr.c 	memcpy(table, priv->smu_tables.entry[table_id].table,
entry             195 drivers/gpu/drm/amd/powerplay/smumgr/vega20_smumgr.c 			priv->smu_tables.entry[table_id].size);
entry             214 drivers/gpu/drm/amd/powerplay/smumgr/vega20_smumgr.c 	PP_ASSERT_WITH_CODE(priv->smu_tables.entry[table_id].version != 0,
entry             216 drivers/gpu/drm/amd/powerplay/smumgr/vega20_smumgr.c 	PP_ASSERT_WITH_CODE(priv->smu_tables.entry[table_id].size != 0,
entry             219 drivers/gpu/drm/amd/powerplay/smumgr/vega20_smumgr.c 	memcpy(priv->smu_tables.entry[table_id].table, table,
entry             220 drivers/gpu/drm/amd/powerplay/smumgr/vega20_smumgr.c 			priv->smu_tables.entry[table_id].size);
entry             224 drivers/gpu/drm/amd/powerplay/smumgr/vega20_smumgr.c 			upper_32_bits(priv->smu_tables.entry[table_id].mc_addr))) == 0,
entry             229 drivers/gpu/drm/amd/powerplay/smumgr/vega20_smumgr.c 			lower_32_bits(priv->smu_tables.entry[table_id].mc_addr))) == 0,
entry             247 drivers/gpu/drm/amd/powerplay/smumgr/vega20_smumgr.c 	memcpy(priv->smu_tables.entry[TABLE_ACTIVITY_MONITOR_COEFF].table, table,
entry             248 drivers/gpu/drm/amd/powerplay/smumgr/vega20_smumgr.c 			priv->smu_tables.entry[TABLE_ACTIVITY_MONITOR_COEFF].size);
entry             252 drivers/gpu/drm/amd/powerplay/smumgr/vega20_smumgr.c 			upper_32_bits(priv->smu_tables.entry[TABLE_ACTIVITY_MONITOR_COEFF].mc_addr))) == 0,
entry             257 drivers/gpu/drm/amd/powerplay/smumgr/vega20_smumgr.c 			lower_32_bits(priv->smu_tables.entry[TABLE_ACTIVITY_MONITOR_COEFF].mc_addr))) == 0,
entry             278 drivers/gpu/drm/amd/powerplay/smumgr/vega20_smumgr.c 			upper_32_bits(priv->smu_tables.entry[TABLE_ACTIVITY_MONITOR_COEFF].mc_addr))) == 0,
entry             283 drivers/gpu/drm/amd/powerplay/smumgr/vega20_smumgr.c 			lower_32_bits(priv->smu_tables.entry[TABLE_ACTIVITY_MONITOR_COEFF].mc_addr))) == 0,
entry             295 drivers/gpu/drm/amd/powerplay/smumgr/vega20_smumgr.c 	memcpy(table, priv->smu_tables.entry[TABLE_ACTIVITY_MONITOR_COEFF].table,
entry             296 drivers/gpu/drm/amd/powerplay/smumgr/vega20_smumgr.c 			priv->smu_tables.entry[TABLE_ACTIVITY_MONITOR_COEFF].size);
entry             365 drivers/gpu/drm/amd/powerplay/smumgr/vega20_smumgr.c 	if (priv->smu_tables.entry[TABLE_PMSTATUSLOG].mc_addr) {
entry             368 drivers/gpu/drm/amd/powerplay/smumgr/vega20_smumgr.c 				upper_32_bits(priv->smu_tables.entry[TABLE_PMSTATUSLOG].mc_addr));
entry             372 drivers/gpu/drm/amd/powerplay/smumgr/vega20_smumgr.c 					lower_32_bits(priv->smu_tables.entry[TABLE_PMSTATUSLOG].mc_addr));
entry             386 drivers/gpu/drm/amd/powerplay/smumgr/vega20_smumgr.c 			upper_32_bits(priv->smu_tables.entry[TABLE_PPTABLE].mc_addr))) == 0,
entry             391 drivers/gpu/drm/amd/powerplay/smumgr/vega20_smumgr.c 			lower_32_bits(priv->smu_tables.entry[TABLE_PPTABLE].mc_addr))) == 0,
entry             423 drivers/gpu/drm/amd/powerplay/smumgr/vega20_smumgr.c 			&priv->smu_tables.entry[TABLE_PPTABLE].handle,
entry             424 drivers/gpu/drm/amd/powerplay/smumgr/vega20_smumgr.c 			&priv->smu_tables.entry[TABLE_PPTABLE].mc_addr,
entry             425 drivers/gpu/drm/amd/powerplay/smumgr/vega20_smumgr.c 			&priv->smu_tables.entry[TABLE_PPTABLE].table);
entry             429 drivers/gpu/drm/amd/powerplay/smumgr/vega20_smumgr.c 	priv->smu_tables.entry[TABLE_PPTABLE].version = 0x01;
entry             430 drivers/gpu/drm/amd/powerplay/smumgr/vega20_smumgr.c 	priv->smu_tables.entry[TABLE_PPTABLE].size = sizeof(PPTable_t);
entry             437 drivers/gpu/drm/amd/powerplay/smumgr/vega20_smumgr.c 			&priv->smu_tables.entry[TABLE_WATERMARKS].handle,
entry             438 drivers/gpu/drm/amd/powerplay/smumgr/vega20_smumgr.c 			&priv->smu_tables.entry[TABLE_WATERMARKS].mc_addr,
entry             439 drivers/gpu/drm/amd/powerplay/smumgr/vega20_smumgr.c 			&priv->smu_tables.entry[TABLE_WATERMARKS].table);
entry             443 drivers/gpu/drm/amd/powerplay/smumgr/vega20_smumgr.c 	priv->smu_tables.entry[TABLE_WATERMARKS].version = 0x01;
entry             444 drivers/gpu/drm/amd/powerplay/smumgr/vega20_smumgr.c 	priv->smu_tables.entry[TABLE_WATERMARKS].size = sizeof(Watermarks_t);
entry             451 drivers/gpu/drm/amd/powerplay/smumgr/vega20_smumgr.c 			&priv->smu_tables.entry[TABLE_PMSTATUSLOG].handle,
entry             452 drivers/gpu/drm/amd/powerplay/smumgr/vega20_smumgr.c 			&priv->smu_tables.entry[TABLE_PMSTATUSLOG].mc_addr,
entry             453 drivers/gpu/drm/amd/powerplay/smumgr/vega20_smumgr.c 			&priv->smu_tables.entry[TABLE_PMSTATUSLOG].table);
entry             457 drivers/gpu/drm/amd/powerplay/smumgr/vega20_smumgr.c 	priv->smu_tables.entry[TABLE_PMSTATUSLOG].version = 0x01;
entry             458 drivers/gpu/drm/amd/powerplay/smumgr/vega20_smumgr.c 	priv->smu_tables.entry[TABLE_PMSTATUSLOG].size = tools_size;
entry             465 drivers/gpu/drm/amd/powerplay/smumgr/vega20_smumgr.c 			&priv->smu_tables.entry[TABLE_OVERDRIVE].handle,
entry             466 drivers/gpu/drm/amd/powerplay/smumgr/vega20_smumgr.c 			&priv->smu_tables.entry[TABLE_OVERDRIVE].mc_addr,
entry             467 drivers/gpu/drm/amd/powerplay/smumgr/vega20_smumgr.c 			&priv->smu_tables.entry[TABLE_OVERDRIVE].table);
entry             471 drivers/gpu/drm/amd/powerplay/smumgr/vega20_smumgr.c 	priv->smu_tables.entry[TABLE_OVERDRIVE].version = 0x01;
entry             472 drivers/gpu/drm/amd/powerplay/smumgr/vega20_smumgr.c 	priv->smu_tables.entry[TABLE_OVERDRIVE].size = sizeof(OverDriveTable_t);
entry             479 drivers/gpu/drm/amd/powerplay/smumgr/vega20_smumgr.c 			&priv->smu_tables.entry[TABLE_SMU_METRICS].handle,
entry             480 drivers/gpu/drm/amd/powerplay/smumgr/vega20_smumgr.c 			&priv->smu_tables.entry[TABLE_SMU_METRICS].mc_addr,
entry             481 drivers/gpu/drm/amd/powerplay/smumgr/vega20_smumgr.c 			&priv->smu_tables.entry[TABLE_SMU_METRICS].table);
entry             485 drivers/gpu/drm/amd/powerplay/smumgr/vega20_smumgr.c 	priv->smu_tables.entry[TABLE_SMU_METRICS].version = 0x01;
entry             486 drivers/gpu/drm/amd/powerplay/smumgr/vega20_smumgr.c 	priv->smu_tables.entry[TABLE_SMU_METRICS].size = sizeof(SmuMetrics_t);
entry             493 drivers/gpu/drm/amd/powerplay/smumgr/vega20_smumgr.c 			&priv->smu_tables.entry[TABLE_ACTIVITY_MONITOR_COEFF].handle,
entry             494 drivers/gpu/drm/amd/powerplay/smumgr/vega20_smumgr.c 			&priv->smu_tables.entry[TABLE_ACTIVITY_MONITOR_COEFF].mc_addr,
entry             495 drivers/gpu/drm/amd/powerplay/smumgr/vega20_smumgr.c 			&priv->smu_tables.entry[TABLE_ACTIVITY_MONITOR_COEFF].table);
entry             499 drivers/gpu/drm/amd/powerplay/smumgr/vega20_smumgr.c 	priv->smu_tables.entry[TABLE_ACTIVITY_MONITOR_COEFF].version = 0x01;
entry             500 drivers/gpu/drm/amd/powerplay/smumgr/vega20_smumgr.c 	priv->smu_tables.entry[TABLE_ACTIVITY_MONITOR_COEFF].size = sizeof(DpmActivityMonitorCoeffInt_t);
entry             505 drivers/gpu/drm/amd/powerplay/smumgr/vega20_smumgr.c 	amdgpu_bo_free_kernel(&priv->smu_tables.entry[TABLE_SMU_METRICS].handle,
entry             506 drivers/gpu/drm/amd/powerplay/smumgr/vega20_smumgr.c 			&priv->smu_tables.entry[TABLE_SMU_METRICS].mc_addr,
entry             507 drivers/gpu/drm/amd/powerplay/smumgr/vega20_smumgr.c 			&priv->smu_tables.entry[TABLE_SMU_METRICS].table);
entry             509 drivers/gpu/drm/amd/powerplay/smumgr/vega20_smumgr.c 	amdgpu_bo_free_kernel(&priv->smu_tables.entry[TABLE_OVERDRIVE].handle,
entry             510 drivers/gpu/drm/amd/powerplay/smumgr/vega20_smumgr.c 			&priv->smu_tables.entry[TABLE_OVERDRIVE].mc_addr,
entry             511 drivers/gpu/drm/amd/powerplay/smumgr/vega20_smumgr.c 			&priv->smu_tables.entry[TABLE_OVERDRIVE].table);
entry             513 drivers/gpu/drm/amd/powerplay/smumgr/vega20_smumgr.c 	amdgpu_bo_free_kernel(&priv->smu_tables.entry[TABLE_PMSTATUSLOG].handle,
entry             514 drivers/gpu/drm/amd/powerplay/smumgr/vega20_smumgr.c 			&priv->smu_tables.entry[TABLE_PMSTATUSLOG].mc_addr,
entry             515 drivers/gpu/drm/amd/powerplay/smumgr/vega20_smumgr.c 			&priv->smu_tables.entry[TABLE_PMSTATUSLOG].table);
entry             517 drivers/gpu/drm/amd/powerplay/smumgr/vega20_smumgr.c 	amdgpu_bo_free_kernel(&priv->smu_tables.entry[TABLE_WATERMARKS].handle,
entry             518 drivers/gpu/drm/amd/powerplay/smumgr/vega20_smumgr.c 			&priv->smu_tables.entry[TABLE_WATERMARKS].mc_addr,
entry             519 drivers/gpu/drm/amd/powerplay/smumgr/vega20_smumgr.c 			&priv->smu_tables.entry[TABLE_WATERMARKS].table);
entry             521 drivers/gpu/drm/amd/powerplay/smumgr/vega20_smumgr.c 	amdgpu_bo_free_kernel(&priv->smu_tables.entry[TABLE_PPTABLE].handle,
entry             522 drivers/gpu/drm/amd/powerplay/smumgr/vega20_smumgr.c 			&priv->smu_tables.entry[TABLE_PPTABLE].mc_addr,
entry             523 drivers/gpu/drm/amd/powerplay/smumgr/vega20_smumgr.c 			&priv->smu_tables.entry[TABLE_PPTABLE].table);
entry             536 drivers/gpu/drm/amd/powerplay/smumgr/vega20_smumgr.c 		amdgpu_bo_free_kernel(&priv->smu_tables.entry[TABLE_PPTABLE].handle,
entry             537 drivers/gpu/drm/amd/powerplay/smumgr/vega20_smumgr.c 				&priv->smu_tables.entry[TABLE_PPTABLE].mc_addr,
entry             538 drivers/gpu/drm/amd/powerplay/smumgr/vega20_smumgr.c 				&priv->smu_tables.entry[TABLE_PPTABLE].table);
entry             539 drivers/gpu/drm/amd/powerplay/smumgr/vega20_smumgr.c 		amdgpu_bo_free_kernel(&priv->smu_tables.entry[TABLE_WATERMARKS].handle,
entry             540 drivers/gpu/drm/amd/powerplay/smumgr/vega20_smumgr.c 				&priv->smu_tables.entry[TABLE_WATERMARKS].mc_addr,
entry             541 drivers/gpu/drm/amd/powerplay/smumgr/vega20_smumgr.c 				&priv->smu_tables.entry[TABLE_WATERMARKS].table);
entry             542 drivers/gpu/drm/amd/powerplay/smumgr/vega20_smumgr.c 		amdgpu_bo_free_kernel(&priv->smu_tables.entry[TABLE_PMSTATUSLOG].handle,
entry             543 drivers/gpu/drm/amd/powerplay/smumgr/vega20_smumgr.c 				&priv->smu_tables.entry[TABLE_PMSTATUSLOG].mc_addr,
entry             544 drivers/gpu/drm/amd/powerplay/smumgr/vega20_smumgr.c 				&priv->smu_tables.entry[TABLE_PMSTATUSLOG].table);
entry             545 drivers/gpu/drm/amd/powerplay/smumgr/vega20_smumgr.c 		amdgpu_bo_free_kernel(&priv->smu_tables.entry[TABLE_OVERDRIVE].handle,
entry             546 drivers/gpu/drm/amd/powerplay/smumgr/vega20_smumgr.c 				&priv->smu_tables.entry[TABLE_OVERDRIVE].mc_addr,
entry             547 drivers/gpu/drm/amd/powerplay/smumgr/vega20_smumgr.c 				&priv->smu_tables.entry[TABLE_OVERDRIVE].table);
entry             548 drivers/gpu/drm/amd/powerplay/smumgr/vega20_smumgr.c 		amdgpu_bo_free_kernel(&priv->smu_tables.entry[TABLE_SMU_METRICS].handle,
entry             549 drivers/gpu/drm/amd/powerplay/smumgr/vega20_smumgr.c 				&priv->smu_tables.entry[TABLE_SMU_METRICS].mc_addr,
entry             550 drivers/gpu/drm/amd/powerplay/smumgr/vega20_smumgr.c 				&priv->smu_tables.entry[TABLE_SMU_METRICS].table);
entry             551 drivers/gpu/drm/amd/powerplay/smumgr/vega20_smumgr.c 		amdgpu_bo_free_kernel(&priv->smu_tables.entry[TABLE_ACTIVITY_MONITOR_COEFF].handle,
entry             552 drivers/gpu/drm/amd/powerplay/smumgr/vega20_smumgr.c 				&priv->smu_tables.entry[TABLE_ACTIVITY_MONITOR_COEFF].mc_addr,
entry             553 drivers/gpu/drm/amd/powerplay/smumgr/vega20_smumgr.c 				&priv->smu_tables.entry[TABLE_ACTIVITY_MONITOR_COEFF].table);
entry              38 drivers/gpu/drm/amd/powerplay/smumgr/vega20_smumgr.h 	struct smu_table_entry entry[TABLE_COUNT];
entry             679 drivers/gpu/drm/amd/powerplay/smumgr/vegam_smumgr.c 					range_table_from_vbios.entry[i].ucVco_setting;
entry             681 drivers/gpu/drm/amd/powerplay/smumgr/vegam_smumgr.c 					range_table_from_vbios.entry[i].ucPostdiv;
entry             683 drivers/gpu/drm/amd/powerplay/smumgr/vegam_smumgr.c 					range_table_from_vbios.entry[i].usFcw_pcc;
entry             686 drivers/gpu/drm/amd/powerplay/smumgr/vegam_smumgr.c 					range_table_from_vbios.entry[i].usFcw_trans_upper;
entry             688 drivers/gpu/drm/amd/powerplay/smumgr/vegam_smumgr.c 					range_table_from_vbios.entry[i].usRcw_trans_lower;
entry              63 drivers/gpu/drm/ati_pcigart.c 	struct drm_sg_mem *entry = dev->sg;
entry              69 drivers/gpu/drm/ati_pcigart.c 	if (!entry) {
entry              77 drivers/gpu/drm/ati_pcigart.c 		pages = (entry->pages <= max_pages)
entry              78 drivers/gpu/drm/ati_pcigart.c 		  ? entry->pages : max_pages;
entry              81 drivers/gpu/drm/ati_pcigart.c 			if (!entry->busaddr[i])
entry              83 drivers/gpu/drm/ati_pcigart.c 			pci_unmap_page(dev->pdev, entry->busaddr[i],
entry             103 drivers/gpu/drm/ati_pcigart.c 	struct drm_sg_mem *entry = dev->sg;
entry             111 drivers/gpu/drm/ati_pcigart.c 	if (!entry) {
entry             146 drivers/gpu/drm/ati_pcigart.c 	pages = (entry->pages <= max_real_pages)
entry             147 drivers/gpu/drm/ati_pcigart.c 	    ? entry->pages : max_real_pages;
entry             158 drivers/gpu/drm/ati_pcigart.c 		entry->busaddr[i] = pci_map_page(dev->pdev, entry->pagelist[i],
entry             160 drivers/gpu/drm/ati_pcigart.c 		if (pci_dma_mapping_error(dev->pdev, entry->busaddr[i])) {
entry             168 drivers/gpu/drm/ati_pcigart.c 		page_base = (u32) entry->busaddr[i];
entry             204 drivers/gpu/drm/drm_agpsupport.c 	struct drm_agp_mem *entry;
entry             211 drivers/gpu/drm/drm_agpsupport.c 	entry = kzalloc(sizeof(*entry), GFP_KERNEL);
entry             212 drivers/gpu/drm/drm_agpsupport.c 	if (!entry)
entry             219 drivers/gpu/drm/drm_agpsupport.c 		kfree(entry);
entry             223 drivers/gpu/drm/drm_agpsupport.c 	entry->handle = (unsigned long)memory->key + 1;
entry             224 drivers/gpu/drm/drm_agpsupport.c 	entry->memory = memory;
entry             225 drivers/gpu/drm/drm_agpsupport.c 	entry->bound = 0;
entry             226 drivers/gpu/drm/drm_agpsupport.c 	entry->pages = pages;
entry             227 drivers/gpu/drm/drm_agpsupport.c 	list_add(&entry->head, &dev->agp->memory);
entry             229 drivers/gpu/drm/drm_agpsupport.c 	request->handle = entry->handle;
entry             257 drivers/gpu/drm/drm_agpsupport.c 	struct drm_agp_mem *entry;
entry             259 drivers/gpu/drm/drm_agpsupport.c 	list_for_each_entry(entry, &dev->agp->memory, head) {
entry             260 drivers/gpu/drm/drm_agpsupport.c 		if (entry->handle == handle)
entry             261 drivers/gpu/drm/drm_agpsupport.c 			return entry;
entry             280 drivers/gpu/drm/drm_agpsupport.c 	struct drm_agp_mem *entry;
entry             285 drivers/gpu/drm/drm_agpsupport.c 	entry = drm_agp_lookup_entry(dev, request->handle);
entry             286 drivers/gpu/drm/drm_agpsupport.c 	if (!entry || !entry->bound)
entry             288 drivers/gpu/drm/drm_agpsupport.c 	ret = drm_unbind_agp(entry->memory);
entry             290 drivers/gpu/drm/drm_agpsupport.c 		entry->bound = 0;
entry             319 drivers/gpu/drm/drm_agpsupport.c 	struct drm_agp_mem *entry;
entry             325 drivers/gpu/drm/drm_agpsupport.c 	entry = drm_agp_lookup_entry(dev, request->handle);
entry             326 drivers/gpu/drm/drm_agpsupport.c 	if (!entry || entry->bound)
entry             329 drivers/gpu/drm/drm_agpsupport.c 	retcode = drm_bind_agp(entry->memory, page);
entry             332 drivers/gpu/drm/drm_agpsupport.c 	entry->bound = dev->agp->base + (page << PAGE_SHIFT);
entry             334 drivers/gpu/drm/drm_agpsupport.c 		  dev->agp->base, entry->bound);
entry             364 drivers/gpu/drm/drm_agpsupport.c 	struct drm_agp_mem *entry;
entry             368 drivers/gpu/drm/drm_agpsupport.c 	entry = drm_agp_lookup_entry(dev, request->handle);
entry             369 drivers/gpu/drm/drm_agpsupport.c 	if (!entry)
entry             371 drivers/gpu/drm/drm_agpsupport.c 	if (entry->bound)
entry             372 drivers/gpu/drm/drm_agpsupport.c 		drm_unbind_agp(entry->memory);
entry             374 drivers/gpu/drm/drm_agpsupport.c 	list_del(&entry->head);
entry             376 drivers/gpu/drm/drm_agpsupport.c 	drm_free_agp(entry->memory, entry->pages);
entry             377 drivers/gpu/drm/drm_agpsupport.c 	kfree(entry);
entry             447 drivers/gpu/drm/drm_agpsupport.c 	struct drm_agp_mem *entry, *tempe;
entry             454 drivers/gpu/drm/drm_agpsupport.c 	list_for_each_entry_safe(entry, tempe, &dev->agp->memory, head) {
entry             455 drivers/gpu/drm/drm_agpsupport.c 		if (entry->bound)
entry             456 drivers/gpu/drm/drm_agpsupport.c 			drm_unbind_agp(entry->memory);
entry             457 drivers/gpu/drm/drm_agpsupport.c 		drm_free_agp(entry->memory, entry->pages);
entry             458 drivers/gpu/drm/drm_agpsupport.c 		kfree(entry);
entry              55 drivers/gpu/drm/drm_bufs.c 	struct drm_map_list *entry;
entry              56 drivers/gpu/drm/drm_bufs.c 	list_for_each_entry(entry, &dev->maplist, head) {
entry              65 drivers/gpu/drm/drm_bufs.c 		if (!entry->map ||
entry              66 drivers/gpu/drm/drm_bufs.c 		    map->type != entry->map->type ||
entry              67 drivers/gpu/drm/drm_bufs.c 		    entry->master != dev->master)
entry              73 drivers/gpu/drm/drm_bufs.c 			return entry;
entry              76 drivers/gpu/drm/drm_bufs.c 			if ((entry->map->offset & 0xffffffff) ==
entry              78 drivers/gpu/drm/drm_bufs.c 				return entry;
entry              82 drivers/gpu/drm/drm_bufs.c 		if (entry->map->offset == map->offset)
entry              83 drivers/gpu/drm/drm_bufs.c 			return entry;
entry             270 drivers/gpu/drm/drm_bufs.c 		struct drm_agp_mem *entry;
entry             299 drivers/gpu/drm/drm_bufs.c 		list_for_each_entry(entry, &dev->agp->memory, head) {
entry             300 drivers/gpu/drm/drm_bufs.c 			if ((map->offset >= entry->bound) &&
entry             301 drivers/gpu/drm/drm_bufs.c 			    (map->offset + map->size <= entry->bound + entry->pages * PAGE_SIZE)) {
entry             673 drivers/gpu/drm/drm_bufs.c 				  struct drm_buf_entry *entry)
entry             677 drivers/gpu/drm/drm_bufs.c 	if (entry->seg_count) {
entry             678 drivers/gpu/drm/drm_bufs.c 		for (i = 0; i < entry->seg_count; i++) {
entry             679 drivers/gpu/drm/drm_bufs.c 			if (entry->seglist[i]) {
entry             680 drivers/gpu/drm/drm_bufs.c 				drm_pci_free(dev, entry->seglist[i]);
entry             683 drivers/gpu/drm/drm_bufs.c 		kfree(entry->seglist);
entry             685 drivers/gpu/drm/drm_bufs.c 		entry->seg_count = 0;
entry             688 drivers/gpu/drm/drm_bufs.c 	if (entry->buf_count) {
entry             689 drivers/gpu/drm/drm_bufs.c 		for (i = 0; i < entry->buf_count; i++) {
entry             690 drivers/gpu/drm/drm_bufs.c 			kfree(entry->buflist[i].dev_private);
entry             692 drivers/gpu/drm/drm_bufs.c 		kfree(entry->buflist);
entry             694 drivers/gpu/drm/drm_bufs.c 		entry->buf_count = 0;
entry             714 drivers/gpu/drm/drm_bufs.c 	struct drm_buf_entry *entry;
entry             777 drivers/gpu/drm/drm_bufs.c 	entry = &dma->bufs[order];
entry             778 drivers/gpu/drm/drm_bufs.c 	if (entry->buf_count) {
entry             790 drivers/gpu/drm/drm_bufs.c 	entry->buflist = kcalloc(count, sizeof(*entry->buflist), GFP_KERNEL);
entry             791 drivers/gpu/drm/drm_bufs.c 	if (!entry->buflist) {
entry             797 drivers/gpu/drm/drm_bufs.c 	entry->buf_size = size;
entry             798 drivers/gpu/drm/drm_bufs.c 	entry->page_order = page_order;
entry             802 drivers/gpu/drm/drm_bufs.c 	while (entry->buf_count < count) {
entry             803 drivers/gpu/drm/drm_bufs.c 		buf = &entry->buflist[entry->buf_count];
entry             804 drivers/gpu/drm/drm_bufs.c 		buf->idx = dma->buf_count + entry->buf_count;
entry             821 drivers/gpu/drm/drm_bufs.c 			entry->buf_count = count;
entry             822 drivers/gpu/drm/drm_bufs.c 			drm_cleanup_buf_error(dev, entry);
entry             828 drivers/gpu/drm/drm_bufs.c 		DRM_DEBUG("buffer %d @ %p\n", entry->buf_count, buf->address);
entry             831 drivers/gpu/drm/drm_bufs.c 		entry->buf_count++;
entry             838 drivers/gpu/drm/drm_bufs.c 				(dma->buf_count + entry->buf_count) *
entry             842 drivers/gpu/drm/drm_bufs.c 		drm_cleanup_buf_error(dev, entry);
entry             849 drivers/gpu/drm/drm_bufs.c 	for (i = 0; i < entry->buf_count; i++) {
entry             850 drivers/gpu/drm/drm_bufs.c 		dma->buflist[i + dma->buf_count] = &entry->buflist[i];
entry             853 drivers/gpu/drm/drm_bufs.c 	dma->buf_count += entry->buf_count;
entry             854 drivers/gpu/drm/drm_bufs.c 	dma->seg_count += entry->seg_count;
entry             859 drivers/gpu/drm/drm_bufs.c 	DRM_DEBUG("entry->buf_count : %d\n", entry->buf_count);
entry             863 drivers/gpu/drm/drm_bufs.c 	request->count = entry->buf_count;
entry             883 drivers/gpu/drm/drm_bufs.c 	struct drm_buf_entry *entry;
entry             927 drivers/gpu/drm/drm_bufs.c 	entry = &dma->bufs[order];
entry             928 drivers/gpu/drm/drm_bufs.c 	if (entry->buf_count) {
entry             940 drivers/gpu/drm/drm_bufs.c 	entry->buflist = kcalloc(count, sizeof(*entry->buflist), GFP_KERNEL);
entry             941 drivers/gpu/drm/drm_bufs.c 	if (!entry->buflist) {
entry             947 drivers/gpu/drm/drm_bufs.c 	entry->seglist = kcalloc(count, sizeof(*entry->seglist), GFP_KERNEL);
entry             948 drivers/gpu/drm/drm_bufs.c 	if (!entry->seglist) {
entry             949 drivers/gpu/drm/drm_bufs.c 		kfree(entry->buflist);
entry             962 drivers/gpu/drm/drm_bufs.c 		kfree(entry->buflist);
entry             963 drivers/gpu/drm/drm_bufs.c 		kfree(entry->seglist);
entry             973 drivers/gpu/drm/drm_bufs.c 	entry->buf_size = size;
entry             974 drivers/gpu/drm/drm_bufs.c 	entry->page_order = page_order;
entry             978 drivers/gpu/drm/drm_bufs.c 	while (entry->buf_count < count) {
entry             984 drivers/gpu/drm/drm_bufs.c 			entry->buf_count = count;
entry             985 drivers/gpu/drm/drm_bufs.c 			entry->seg_count = count;
entry             986 drivers/gpu/drm/drm_bufs.c 			drm_cleanup_buf_error(dev, entry);
entry             992 drivers/gpu/drm/drm_bufs.c 		entry->seglist[entry->seg_count++] = dmah;
entry            1001 drivers/gpu/drm/drm_bufs.c 		     offset + size <= total && entry->buf_count < count;
entry            1002 drivers/gpu/drm/drm_bufs.c 		     offset += alignment, ++entry->buf_count) {
entry            1003 drivers/gpu/drm/drm_bufs.c 			buf = &entry->buflist[entry->buf_count];
entry            1004 drivers/gpu/drm/drm_bufs.c 			buf->idx = dma->buf_count + entry->buf_count;
entry            1021 drivers/gpu/drm/drm_bufs.c 				entry->buf_count = count;
entry            1022 drivers/gpu/drm/drm_bufs.c 				entry->seg_count = count;
entry            1023 drivers/gpu/drm/drm_bufs.c 				drm_cleanup_buf_error(dev, entry);
entry            1031 drivers/gpu/drm/drm_bufs.c 				  entry->buf_count, buf->address);
entry            1037 drivers/gpu/drm/drm_bufs.c 				(dma->buf_count + entry->buf_count) *
entry            1041 drivers/gpu/drm/drm_bufs.c 		drm_cleanup_buf_error(dev, entry);
entry            1049 drivers/gpu/drm/drm_bufs.c 	for (i = 0; i < entry->buf_count; i++) {
entry            1050 drivers/gpu/drm/drm_bufs.c 		dma->buflist[i + dma->buf_count] = &entry->buflist[i];
entry            1061 drivers/gpu/drm/drm_bufs.c 	dma->buf_count += entry->buf_count;
entry            1062 drivers/gpu/drm/drm_bufs.c 	dma->seg_count += entry->seg_count;
entry            1063 drivers/gpu/drm/drm_bufs.c 	dma->page_count += entry->seg_count << page_order;
entry            1064 drivers/gpu/drm/drm_bufs.c 	dma->byte_count += PAGE_SIZE * (entry->seg_count << page_order);
entry            1068 drivers/gpu/drm/drm_bufs.c 	request->count = entry->buf_count;
entry            1084 drivers/gpu/drm/drm_bufs.c 	struct drm_buf_entry *entry;
entry            1139 drivers/gpu/drm/drm_bufs.c 	entry = &dma->bufs[order];
entry            1140 drivers/gpu/drm/drm_bufs.c 	if (entry->buf_count) {
entry            1152 drivers/gpu/drm/drm_bufs.c 	entry->buflist = kcalloc(count, sizeof(*entry->buflist), GFP_KERNEL);
entry            1153 drivers/gpu/drm/drm_bufs.c 	if (!entry->buflist) {
entry            1159 drivers/gpu/drm/drm_bufs.c 	entry->buf_size = size;
entry            1160 drivers/gpu/drm/drm_bufs.c 	entry->page_order = page_order;
entry            1164 drivers/gpu/drm/drm_bufs.c 	while (entry->buf_count < count) {
entry            1165 drivers/gpu/drm/drm_bufs.c 		buf = &entry->buflist[entry->buf_count];
entry            1166 drivers/gpu/drm/drm_bufs.c 		buf->idx = dma->buf_count + entry->buf_count;
entry            1184 drivers/gpu/drm/drm_bufs.c 			entry->buf_count = count;
entry            1185 drivers/gpu/drm/drm_bufs.c 			drm_cleanup_buf_error(dev, entry);
entry            1191 drivers/gpu/drm/drm_bufs.c 		DRM_DEBUG("buffer %d @ %p\n", entry->buf_count, buf->address);
entry            1194 drivers/gpu/drm/drm_bufs.c 		entry->buf_count++;
entry            1201 drivers/gpu/drm/drm_bufs.c 				(dma->buf_count + entry->buf_count) *
entry            1205 drivers/gpu/drm/drm_bufs.c 		drm_cleanup_buf_error(dev, entry);
entry            1212 drivers/gpu/drm/drm_bufs.c 	for (i = 0; i < entry->buf_count; i++) {
entry            1213 drivers/gpu/drm/drm_bufs.c 		dma->buflist[i + dma->buf_count] = &entry->buflist[i];
entry            1216 drivers/gpu/drm/drm_bufs.c 	dma->buf_count += entry->buf_count;
entry            1217 drivers/gpu/drm/drm_bufs.c 	dma->seg_count += entry->seg_count;
entry            1222 drivers/gpu/drm/drm_bufs.c 	DRM_DEBUG("entry->buf_count : %d\n", entry->buf_count);
entry            1226 drivers/gpu/drm/drm_bufs.c 	request->count = entry->buf_count;
entry            1387 drivers/gpu/drm/drm_bufs.c 	struct drm_buf_entry *entry;
entry            1403 drivers/gpu/drm/drm_bufs.c 	entry = &dma->bufs[order];
entry            1405 drivers/gpu/drm/drm_bufs.c 	if (request->low_mark < 0 || request->low_mark > entry->buf_count)
entry            1407 drivers/gpu/drm/drm_bufs.c 	if (request->high_mark < 0 || request->high_mark > entry->buf_count)
entry            1410 drivers/gpu/drm/drm_bufs.c 	entry->low_mark = request->low_mark;
entry            1411 drivers/gpu/drm/drm_bufs.c 	entry->high_mark = request->high_mark;
entry            1594 drivers/gpu/drm/drm_bufs.c 	struct drm_map_list *entry;
entry            1596 drivers/gpu/drm/drm_bufs.c 	list_for_each_entry(entry, &dev->maplist, head) {
entry            1597 drivers/gpu/drm/drm_bufs.c 		if (entry->map && entry->map->type == _DRM_SHM &&
entry            1598 drivers/gpu/drm/drm_bufs.c 		    (entry->map->flags & _DRM_CONTAINS_LOCK)) {
entry            1599 drivers/gpu/drm/drm_bufs.c 			return entry->map;
entry             483 drivers/gpu/drm/drm_color_mgmt.c 	const struct drm_color_lut *entry;
entry             489 drivers/gpu/drm/drm_color_mgmt.c 	entry = lut->data;
entry             492 drivers/gpu/drm/drm_color_mgmt.c 			if (entry[i].red != entry[i].blue ||
entry             493 drivers/gpu/drm/drm_color_mgmt.c 			    entry[i].red != entry[i].green) {
entry             500 drivers/gpu/drm/drm_color_mgmt.c 			if (entry[i].red < entry[i - 1].red ||
entry             501 drivers/gpu/drm/drm_color_mgmt.c 			    entry[i].green < entry[i - 1].green ||
entry             502 drivers/gpu/drm/drm_color_mgmt.c 			    entry[i].blue < entry[i - 1].blue) {
entry             282 drivers/gpu/drm/drm_debugfs_crc.c 	struct drm_crtc_crc_entry *entry;
entry             310 drivers/gpu/drm/drm_debugfs_crc.c 	entry = &crc->entries[crc->tail];
entry             322 drivers/gpu/drm/drm_debugfs_crc.c 	if (entry->has_frame_counter)
entry             323 drivers/gpu/drm/drm_debugfs_crc.c 		sprintf(buf, "0x%08x", entry->frame);
entry             328 drivers/gpu/drm/drm_debugfs_crc.c 		sprintf(buf + 10 + i * 11, " 0x%08x", entry->crcs[i]);
entry             392 drivers/gpu/drm/drm_debugfs_crc.c 	struct drm_crtc_crc_entry *entry;
entry             419 drivers/gpu/drm/drm_debugfs_crc.c 	entry = &crc->entries[head];
entry             420 drivers/gpu/drm/drm_debugfs_crc.c 	entry->frame = frame;
entry             421 drivers/gpu/drm/drm_debugfs_crc.c 	entry->has_frame_counter = has_frame;
entry             422 drivers/gpu/drm/drm_debugfs_crc.c 	memcpy(&entry->crcs, crcs, sizeof(*crcs) * crc->values_cnt);
entry            1359 drivers/gpu/drm/drm_gem.c 	struct dma_fence *entry;
entry            1371 drivers/gpu/drm/drm_gem.c 	xa_for_each(fence_array, index, entry) {
entry            1372 drivers/gpu/drm/drm_gem.c 		if (entry->context != fence->context)
entry            1375 drivers/gpu/drm/drm_gem.c 		if (dma_fence_is_later(fence, entry)) {
entry            1376 drivers/gpu/drm/drm_gem.c 			dma_fence_put(entry);
entry              65 drivers/gpu/drm/drm_hashtab.c 	struct drm_hash_item *entry;
entry              73 drivers/gpu/drm/drm_hashtab.c 	hlist_for_each_entry(entry, h_list, head)
entry              74 drivers/gpu/drm/drm_hashtab.c 		DRM_DEBUG("count %d, key: 0x%08lx\n", count++, entry->key);
entry              80 drivers/gpu/drm/drm_hashtab.c 	struct drm_hash_item *entry;
entry              86 drivers/gpu/drm/drm_hashtab.c 	hlist_for_each_entry(entry, h_list, head) {
entry              87 drivers/gpu/drm/drm_hashtab.c 		if (entry->key == key)
entry              88 drivers/gpu/drm/drm_hashtab.c 			return &entry->head;
entry              89 drivers/gpu/drm/drm_hashtab.c 		if (entry->key > key)
entry              98 drivers/gpu/drm/drm_hashtab.c 	struct drm_hash_item *entry;
entry             104 drivers/gpu/drm/drm_hashtab.c 	hlist_for_each_entry_rcu(entry, h_list, head) {
entry             105 drivers/gpu/drm/drm_hashtab.c 		if (entry->key == key)
entry             106 drivers/gpu/drm/drm_hashtab.c 			return &entry->head;
entry             107 drivers/gpu/drm/drm_hashtab.c 		if (entry->key > key)
entry             115 drivers/gpu/drm/drm_hashtab.c 	struct drm_hash_item *entry;
entry             124 drivers/gpu/drm/drm_hashtab.c 	hlist_for_each_entry(entry, h_list, head) {
entry             125 drivers/gpu/drm/drm_hashtab.c 		if (entry->key == key)
entry             127 drivers/gpu/drm/drm_hashtab.c 		if (entry->key > key)
entry             129 drivers/gpu/drm/drm_hashtab.c 		parent = &entry->head;
entry             203 drivers/gpu/drm/drm_lease.c 	void *entry;
entry             215 drivers/gpu/drm/drm_lease.c 	idr_for_each_entry(leases, entry, object) {
entry             303 drivers/gpu/drm/drm_lease.c 	void *entry;
entry             316 drivers/gpu/drm/drm_lease.c 		idr_for_each_entry(&master->leases, entry, object)
entry             664 drivers/gpu/drm/drm_lease.c 	void *entry;
entry             687 drivers/gpu/drm/drm_lease.c 	idr_for_each_entry(object_idr, entry, object) {
entry             167 drivers/gpu/drm/drm_lock.c 	DECLARE_WAITQUEUE(entry, current);
entry             188 drivers/gpu/drm/drm_lock.c 	add_wait_queue(&master->lock.lock_queue, &entry);
entry             220 drivers/gpu/drm/drm_lock.c 	remove_wait_queue(&master->lock.lock_queue, &entry);
entry             945 drivers/gpu/drm/drm_mm.c static u64 drm_mm_dump_hole(struct drm_printer *p, const struct drm_mm_node *entry)
entry             949 drivers/gpu/drm/drm_mm.c 	size = entry->hole_size;
entry             951 drivers/gpu/drm/drm_mm.c 		start = drm_mm_hole_node_start(entry);
entry             965 drivers/gpu/drm/drm_mm.c 	const struct drm_mm_node *entry;
entry             970 drivers/gpu/drm/drm_mm.c 	drm_mm_for_each_node(entry, mm) {
entry             971 drivers/gpu/drm/drm_mm.c 		drm_printf(p, "%#018llx-%#018llx: %llu: used\n", entry->start,
entry             972 drivers/gpu/drm/drm_mm.c 			   entry->start + entry->size, entry->size);
entry             973 drivers/gpu/drm/drm_mm.c 		total_used += entry->size;
entry             974 drivers/gpu/drm/drm_mm.c 		total_free += drm_mm_dump_hole(p, entry);
entry              55 drivers/gpu/drm/drm_scatter.c static void drm_sg_cleanup(struct drm_sg_mem * entry)
entry              60 drivers/gpu/drm/drm_scatter.c 	for (i = 0; i < entry->pages; i++) {
entry              61 drivers/gpu/drm/drm_scatter.c 		page = entry->pagelist[i];
entry              66 drivers/gpu/drm/drm_scatter.c 	vfree(entry->virtual);
entry              68 drivers/gpu/drm/drm_scatter.c 	kfree(entry->busaddr);
entry              69 drivers/gpu/drm/drm_scatter.c 	kfree(entry->pagelist);
entry              70 drivers/gpu/drm/drm_scatter.c 	kfree(entry);
entry              91 drivers/gpu/drm/drm_scatter.c 	struct drm_sg_mem *entry;
entry             105 drivers/gpu/drm/drm_scatter.c 	entry = kzalloc(sizeof(*entry), GFP_KERNEL);
entry             106 drivers/gpu/drm/drm_scatter.c 	if (!entry)
entry             112 drivers/gpu/drm/drm_scatter.c 	entry->pages = pages;
entry             113 drivers/gpu/drm/drm_scatter.c 	entry->pagelist = kcalloc(pages, sizeof(*entry->pagelist), GFP_KERNEL);
entry             114 drivers/gpu/drm/drm_scatter.c 	if (!entry->pagelist) {
entry             115 drivers/gpu/drm/drm_scatter.c 		kfree(entry);
entry             119 drivers/gpu/drm/drm_scatter.c 	entry->busaddr = kcalloc(pages, sizeof(*entry->busaddr), GFP_KERNEL);
entry             120 drivers/gpu/drm/drm_scatter.c 	if (!entry->busaddr) {
entry             121 drivers/gpu/drm/drm_scatter.c 		kfree(entry->pagelist);
entry             122 drivers/gpu/drm/drm_scatter.c 		kfree(entry);
entry             126 drivers/gpu/drm/drm_scatter.c 	entry->virtual = drm_vmalloc_dma(pages << PAGE_SHIFT);
entry             127 drivers/gpu/drm/drm_scatter.c 	if (!entry->virtual) {
entry             128 drivers/gpu/drm/drm_scatter.c 		kfree(entry->busaddr);
entry             129 drivers/gpu/drm/drm_scatter.c 		kfree(entry->pagelist);
entry             130 drivers/gpu/drm/drm_scatter.c 		kfree(entry);
entry             137 drivers/gpu/drm/drm_scatter.c 	memset(entry->virtual, 0, pages << PAGE_SHIFT);
entry             139 drivers/gpu/drm/drm_scatter.c 	entry->handle = ScatterHandle((unsigned long)entry->virtual);
entry             141 drivers/gpu/drm/drm_scatter.c 	DRM_DEBUG("handle  = %08lx\n", entry->handle);
entry             142 drivers/gpu/drm/drm_scatter.c 	DRM_DEBUG("virtual = %p\n", entry->virtual);
entry             144 drivers/gpu/drm/drm_scatter.c 	for (i = (unsigned long)entry->virtual, j = 0; j < pages;
entry             146 drivers/gpu/drm/drm_scatter.c 		entry->pagelist[j] = vmalloc_to_page((void *)i);
entry             147 drivers/gpu/drm/drm_scatter.c 		if (!entry->pagelist[j])
entry             149 drivers/gpu/drm/drm_scatter.c 		SetPageReserved(entry->pagelist[j]);
entry             152 drivers/gpu/drm/drm_scatter.c 	request->handle = entry->handle;
entry             154 drivers/gpu/drm/drm_scatter.c 	dev->sg = entry;
entry             166 drivers/gpu/drm/drm_scatter.c 			tmp = page_address(entry->pagelist[i]);
entry             172 drivers/gpu/drm/drm_scatter.c 			tmp = (unsigned long *)((u8 *) entry->virtual +
entry             184 drivers/gpu/drm/drm_scatter.c 			tmp = page_address(entry->pagelist[i]);
entry             199 drivers/gpu/drm/drm_scatter.c 	drm_sg_cleanup(entry);
entry             207 drivers/gpu/drm/drm_scatter.c 	struct drm_sg_mem *entry;
entry             215 drivers/gpu/drm/drm_scatter.c 	entry = dev->sg;
entry             218 drivers/gpu/drm/drm_scatter.c 	if (!entry || entry->handle != request->handle)
entry             221 drivers/gpu/drm/drm_scatter.c 	DRM_DEBUG("virtual  = %p\n", entry->virtual);
entry             223 drivers/gpu/drm/drm_scatter.c 	drm_sg_cleanup(entry);
entry             348 drivers/gpu/drm/drm_vm.c 	struct drm_sg_mem *entry = dev->sg;
entry             354 drivers/gpu/drm/drm_vm.c 	if (!entry)
entry             356 drivers/gpu/drm/drm_vm.c 	if (!entry->pagelist)
entry             362 drivers/gpu/drm/drm_vm.c 	page = entry->pagelist[page_offset];
entry             267 drivers/gpu/drm/drm_vma_manager.c 	struct drm_vma_offset_file *new, *entry;
entry             274 drivers/gpu/drm/drm_vma_manager.c 	new = kmalloc(sizeof(*entry), GFP_KERNEL);
entry             282 drivers/gpu/drm/drm_vma_manager.c 		entry = rb_entry(*iter, struct drm_vma_offset_file, vm_rb);
entry             284 drivers/gpu/drm/drm_vma_manager.c 		if (tag == entry->vm_tag) {
entry             285 drivers/gpu/drm/drm_vma_manager.c 			entry->vm_count++;
entry             287 drivers/gpu/drm/drm_vma_manager.c 		} else if (tag > entry->vm_tag) {
entry             328 drivers/gpu/drm/drm_vma_manager.c 	struct drm_vma_offset_file *entry;
entry             335 drivers/gpu/drm/drm_vma_manager.c 		entry = rb_entry(iter, struct drm_vma_offset_file, vm_rb);
entry             336 drivers/gpu/drm/drm_vma_manager.c 		if (tag == entry->vm_tag) {
entry             337 drivers/gpu/drm/drm_vma_manager.c 			if (!--entry->vm_count) {
entry             338 drivers/gpu/drm/drm_vma_manager.c 				rb_erase(&entry->vm_rb, &node->vm_files);
entry             339 drivers/gpu/drm/drm_vma_manager.c 				kfree(entry);
entry             342 drivers/gpu/drm/drm_vma_manager.c 		} else if (tag > entry->vm_tag) {
entry             369 drivers/gpu/drm/drm_vma_manager.c 	struct drm_vma_offset_file *entry;
entry             376 drivers/gpu/drm/drm_vma_manager.c 		entry = rb_entry(iter, struct drm_vma_offset_file, vm_rb);
entry             377 drivers/gpu/drm/drm_vma_manager.c 		if (tag == entry->vm_tag)
entry             379 drivers/gpu/drm/drm_vma_manager.c 		else if (tag > entry->vm_tag)
entry              98 drivers/gpu/drm/etnaviv/etnaviv_iommu_v2.c 	u32 entry = lower_32_bits(paddr) | MMUv2_PTE_PRESENT;
entry             104 drivers/gpu/drm/etnaviv/etnaviv_iommu_v2.c 		entry |= (upper_32_bits(paddr) & 0xff) << 4;
entry             107 drivers/gpu/drm/etnaviv/etnaviv_iommu_v2.c 		entry |= MMUv2_PTE_WRITEABLE;
entry             116 drivers/gpu/drm/etnaviv/etnaviv_iommu_v2.c 	v2_context->stlb_cpu[mtlb_entry][stlb_entry] = entry;
entry             222 drivers/gpu/drm/gma500/intel_bios.c 	struct bdb_lvds_lfp_data_entry *entry;
entry             245 drivers/gpu/drm/gma500/intel_bios.c 	entry = &lvds_lfp_data->data[lvds_options->panel_type];
entry             246 drivers/gpu/drm/gma500/intel_bios.c 	dvo_timing = &entry->dvo_timing;
entry             549 drivers/gpu/drm/gma500/psb_intel_display.c 	int entry = 0;
entry             555 drivers/gpu/drm/gma500/psb_intel_display.c 			index_mask |= (1 << entry);
entry             556 drivers/gpu/drm/gma500/psb_intel_display.c 		entry++;
entry             178 drivers/gpu/drm/i915/display/intel_bios.c 	char *entry = (char *)lvds_lfp_data->data + lfp_data_size * index;
entry             180 drivers/gpu/drm/i915/display/intel_bios.c 	return (struct lvds_dvo_timing *)(entry + dvo_timing_offset);
entry             307 drivers/gpu/drm/i915/display/intel_bios.c 	const struct lfp_backlight_data_entry *entry;
entry             320 drivers/gpu/drm/i915/display/intel_bios.c 	entry = &backlight_data->data[panel_type];
entry             322 drivers/gpu/drm/i915/display/intel_bios.c 	dev_priv->vbt.backlight.present = entry->type == BDB_BACKLIGHT_TYPE_PWM;
entry             325 drivers/gpu/drm/i915/display/intel_bios.c 			      entry->type);
entry             339 drivers/gpu/drm/i915/display/intel_bios.c 	dev_priv->vbt.backlight.pwm_freq_hz = entry->pwm_freq_hz;
entry             340 drivers/gpu/drm/i915/display/intel_bios.c 	dev_priv->vbt.backlight.active_low_pwm = entry->active_low_pwm;
entry             341 drivers/gpu/drm/i915/display/intel_bios.c 	dev_priv->vbt.backlight.min_brightness = entry->min_brightness;
entry             568 drivers/gpu/drm/i915/display/intel_color.c 		const struct drm_color_lut *entry =
entry             572 drivers/gpu/drm/i915/display/intel_color.c 		I915_WRITE(PREC_PAL_DATA(pipe), ilk_lut_10(entry));
entry             598 drivers/gpu/drm/i915/display/intel_color.c 		const struct drm_color_lut *entry =
entry             601 drivers/gpu/drm/i915/display/intel_color.c 		I915_WRITE(PREC_PAL_DATA(pipe), ilk_lut_10(entry));
entry             819 drivers/gpu/drm/i915/display/intel_color.c 		const struct drm_color_lut *entry = &lut[i];
entry             822 drivers/gpu/drm/i915/display/intel_color.c 			   ilk_lut_12p4_ldw(entry));
entry             824 drivers/gpu/drm/i915/display/intel_color.c 			   ilk_lut_12p4_udw(entry));
entry             835 drivers/gpu/drm/i915/display/intel_color.c 	const struct drm_color_lut *entry;
entry             852 drivers/gpu/drm/i915/display/intel_color.c 		entry = &lut[i * 8];
entry             853 drivers/gpu/drm/i915/display/intel_color.c 		I915_WRITE(PREC_PAL_DATA(pipe), ilk_lut_12p4_ldw(entry));
entry             854 drivers/gpu/drm/i915/display/intel_color.c 		I915_WRITE(PREC_PAL_DATA(pipe), ilk_lut_12p4_udw(entry));
entry             870 drivers/gpu/drm/i915/display/intel_color.c 		entry = &lut[i * 8 * 128];
entry             871 drivers/gpu/drm/i915/display/intel_color.c 		I915_WRITE(PREC_PAL_DATA(pipe), ilk_lut_12p4_ldw(entry));
entry             872 drivers/gpu/drm/i915/display/intel_color.c 		I915_WRITE(PREC_PAL_DATA(pipe), ilk_lut_12p4_udw(entry));
entry             876 drivers/gpu/drm/i915/display/intel_color.c 	entry = &lut[256 * 8 * 128];
entry             877 drivers/gpu/drm/i915/display/intel_color.c 	icl_load_gcmax(crtc_state, entry);
entry            14268 drivers/gpu/drm/i915/display/intel_display.c 	list_del(&wait->wait.entry);
entry            15231 drivers/gpu/drm/i915/display/intel_display.c 	int entry = 0;
entry            15235 drivers/gpu/drm/i915/display/intel_display.c 			index_mask |= (1 << entry);
entry            15237 drivers/gpu/drm/i915/display/intel_display.c 		entry++;
entry             351 drivers/gpu/drm/i915/gem/i915_gem_execbuffer.c eb_vma_misplaced(const struct drm_i915_gem_exec_object2 *entry,
entry             355 drivers/gpu/drm/i915/gem/i915_gem_execbuffer.c 	if (vma->node.size < entry->pad_to_size)
entry             358 drivers/gpu/drm/i915/gem/i915_gem_execbuffer.c 	if (entry->alignment && !IS_ALIGNED(vma->node.start, entry->alignment))
entry             362 drivers/gpu/drm/i915/gem/i915_gem_execbuffer.c 	    vma->node.start != entry->offset)
entry             382 drivers/gpu/drm/i915/gem/i915_gem_execbuffer.c 	   const struct drm_i915_gem_exec_object2 *entry,
entry             391 drivers/gpu/drm/i915/gem/i915_gem_execbuffer.c 		pin_flags = entry->offset & PIN_OFFSET_MASK;
entry             411 drivers/gpu/drm/i915/gem/i915_gem_execbuffer.c 	return !eb_vma_misplaced(entry, vma, exec_flags);
entry             436 drivers/gpu/drm/i915/gem/i915_gem_execbuffer.c 		struct drm_i915_gem_exec_object2 *entry,
entry             439 drivers/gpu/drm/i915/gem/i915_gem_execbuffer.c 	if (unlikely(entry->flags & eb->invalid_flags))
entry             442 drivers/gpu/drm/i915/gem/i915_gem_execbuffer.c 	if (unlikely(entry->alignment &&
entry             443 drivers/gpu/drm/i915/gem/i915_gem_execbuffer.c 		     !is_power_of_2_u64(entry->alignment)))
entry             450 drivers/gpu/drm/i915/gem/i915_gem_execbuffer.c 	if (unlikely(entry->flags & EXEC_OBJECT_PINNED &&
entry             451 drivers/gpu/drm/i915/gem/i915_gem_execbuffer.c 		     entry->offset != gen8_canonical_addr(entry->offset & I915_GTT_PAGE_MASK)))
entry             455 drivers/gpu/drm/i915/gem/i915_gem_execbuffer.c 	if (entry->flags & EXEC_OBJECT_PAD_TO_SIZE) {
entry             456 drivers/gpu/drm/i915/gem/i915_gem_execbuffer.c 		if (unlikely(offset_in_page(entry->pad_to_size)))
entry             459 drivers/gpu/drm/i915/gem/i915_gem_execbuffer.c 		entry->pad_to_size = 0;
entry             464 drivers/gpu/drm/i915/gem/i915_gem_execbuffer.c 			  entry->handle, (int)(entry - eb->exec));
entry             473 drivers/gpu/drm/i915/gem/i915_gem_execbuffer.c 	entry->offset = gen8_noncanonical_addr(entry->offset);
entry             476 drivers/gpu/drm/i915/gem/i915_gem_execbuffer.c 		entry->flags &= ~EXEC_OBJECT_NEEDS_FENCE;
entry             478 drivers/gpu/drm/i915/gem/i915_gem_execbuffer.c 		if ((entry->flags & EXEC_OBJECT_NEEDS_FENCE ||
entry             481 drivers/gpu/drm/i915/gem/i915_gem_execbuffer.c 			entry->flags |= EXEC_OBJECT_NEEDS_GTT | __EXEC_OBJECT_NEEDS_MAP;
entry             484 drivers/gpu/drm/i915/gem/i915_gem_execbuffer.c 	if (!(entry->flags & EXEC_OBJECT_PINNED))
entry             485 drivers/gpu/drm/i915/gem/i915_gem_execbuffer.c 		entry->flags |= eb->context_flags;
entry             495 drivers/gpu/drm/i915/gem/i915_gem_execbuffer.c 	struct drm_i915_gem_exec_object2 *entry = &eb->exec[i];
entry             501 drivers/gpu/drm/i915/gem/i915_gem_execbuffer.c 		err = eb_validate_vma(eb, entry, vma);
entry             507 drivers/gpu/drm/i915/gem/i915_gem_execbuffer.c 		vma->exec_handle = entry->handle;
entry             509 drivers/gpu/drm/i915/gem/i915_gem_execbuffer.c 			       &eb->buckets[hash_32(entry->handle,
entry             513 drivers/gpu/drm/i915/gem/i915_gem_execbuffer.c 	if (entry->relocation_count)
entry             523 drivers/gpu/drm/i915/gem/i915_gem_execbuffer.c 	eb->flags[i] = entry->flags;
entry             536 drivers/gpu/drm/i915/gem/i915_gem_execbuffer.c 		if (entry->relocation_count &&
entry             546 drivers/gpu/drm/i915/gem/i915_gem_execbuffer.c 	if (eb_pin_vma(eb, entry, vma)) {
entry             547 drivers/gpu/drm/i915/gem/i915_gem_execbuffer.c 		if (entry->offset != vma->node.start) {
entry             548 drivers/gpu/drm/i915/gem/i915_gem_execbuffer.c 			entry->offset = vma->node.start | UPDATE;
entry             583 drivers/gpu/drm/i915/gem/i915_gem_execbuffer.c 	struct drm_i915_gem_exec_object2 *entry = exec_entry(eb, vma);
entry             603 drivers/gpu/drm/i915/gem/i915_gem_execbuffer.c 		pin_flags |= entry->offset | PIN_OFFSET_FIXED;
entry             610 drivers/gpu/drm/i915/gem/i915_gem_execbuffer.c 			   entry->pad_to_size, entry->alignment,
entry             615 drivers/gpu/drm/i915/gem/i915_gem_execbuffer.c 	if (entry->offset != vma->node.start) {
entry             616 drivers/gpu/drm/i915/gem/i915_gem_execbuffer.c 		entry->offset = vma->node.start | UPDATE;
entry             632 drivers/gpu/drm/i915/gem/i915_gem_execbuffer.c 	GEM_BUG_ON(eb_vma_misplaced(entry, vma, exec_flags));
entry            1448 drivers/gpu/drm/i915/gem/i915_gem_execbuffer.c 	const struct drm_i915_gem_exec_object2 *entry = exec_entry(eb, vma);
entry            1451 drivers/gpu/drm/i915/gem/i915_gem_execbuffer.c 	urelocs = u64_to_user_ptr(entry->relocs_ptr);
entry            1452 drivers/gpu/drm/i915/gem/i915_gem_execbuffer.c 	remain = entry->relocation_count;
entry            1533 drivers/gpu/drm/i915/gem/i915_gem_execbuffer.c 	const struct drm_i915_gem_exec_object2 *entry = exec_entry(eb, vma);
entry            1535 drivers/gpu/drm/i915/gem/i915_gem_execbuffer.c 		u64_to_ptr(typeof(*relocs), entry->relocs_ptr);
entry            1539 drivers/gpu/drm/i915/gem/i915_gem_execbuffer.c 	for (i = 0; i < entry->relocation_count; i++) {
entry            1553 drivers/gpu/drm/i915/gem/i915_gem_execbuffer.c static int check_relocations(const struct drm_i915_gem_exec_object2 *entry)
entry            1559 drivers/gpu/drm/i915/gem/i915_gem_execbuffer.c 	size = entry->relocation_count;
entry            1566 drivers/gpu/drm/i915/gem/i915_gem_execbuffer.c 	addr = u64_to_user_ptr(entry->relocs_ptr);
entry            1771 drivers/gpu/drm/i915/gem/i915_gem_execbuffer.c 			const struct drm_i915_gem_exec_object2 *entry =
entry            1775 drivers/gpu/drm/i915/gem/i915_gem_execbuffer.c 			if (!entry->relocation_count)
entry            1778 drivers/gpu/drm/i915/gem/i915_gem_execbuffer.c 			relocs = u64_to_ptr(typeof(*relocs), entry->relocs_ptr);
entry             410 drivers/gpu/drm/i915/gem/i915_gem_pages.c 		void *entry;
entry             426 drivers/gpu/drm/i915/gem/i915_gem_pages.c 		entry = xa_mk_value(idx);
entry             428 drivers/gpu/drm/i915/gem/i915_gem_pages.c 			ret = radix_tree_insert(&iter->radix, idx + i, entry);
entry             527 drivers/gpu/drm/i915/gvt/gtt.c 	struct intel_gvt_gtt_entry *entry, bool ips)
entry             529 drivers/gpu/drm/i915/gvt/gtt.c 	switch (entry->type) {
entry             532 drivers/gpu/drm/i915/gvt/gtt.c 		if (pte_ops->test_pse(entry))
entry             533 drivers/gpu/drm/i915/gvt/gtt.c 			entry->type = get_pse_type(entry->type);
entry             537 drivers/gpu/drm/i915/gvt/gtt.c 			entry->type = get_pse_type(entry->type);
entry             540 drivers/gpu/drm/i915/gvt/gtt.c 		GEM_BUG_ON(!gtt_type_is_entry(entry->type));
entry             543 drivers/gpu/drm/i915/gvt/gtt.c 	GEM_BUG_ON(entry->type == GTT_TYPE_INVALID);
entry             550 drivers/gpu/drm/i915/gvt/gtt.c 		struct intel_gvt_gtt_entry *entry, unsigned long index,
entry             557 drivers/gpu/drm/i915/gvt/gtt.c 	entry->type = mm->ppgtt_mm.root_entry_type;
entry             560 drivers/gpu/drm/i915/gvt/gtt.c 			   entry, index, false, 0, mm->vgpu);
entry             561 drivers/gpu/drm/i915/gvt/gtt.c 	update_entry_type_for_real(pte_ops, entry, false);
entry             565 drivers/gpu/drm/i915/gvt/gtt.c 		struct intel_gvt_gtt_entry *entry, unsigned long index)
entry             567 drivers/gpu/drm/i915/gvt/gtt.c 	_ppgtt_get_root_entry(mm, entry, index, true);
entry             571 drivers/gpu/drm/i915/gvt/gtt.c 		struct intel_gvt_gtt_entry *entry, unsigned long index)
entry             573 drivers/gpu/drm/i915/gvt/gtt.c 	_ppgtt_get_root_entry(mm, entry, index, false);
entry             577 drivers/gpu/drm/i915/gvt/gtt.c 		struct intel_gvt_gtt_entry *entry, unsigned long index,
entry             584 drivers/gpu/drm/i915/gvt/gtt.c 			   entry, index, false, 0, mm->vgpu);
entry             588 drivers/gpu/drm/i915/gvt/gtt.c 		struct intel_gvt_gtt_entry *entry, unsigned long index)
entry             590 drivers/gpu/drm/i915/gvt/gtt.c 	_ppgtt_set_root_entry(mm, entry, index, true);
entry             594 drivers/gpu/drm/i915/gvt/gtt.c 		struct intel_gvt_gtt_entry *entry, unsigned long index)
entry             596 drivers/gpu/drm/i915/gvt/gtt.c 	_ppgtt_set_root_entry(mm, entry, index, false);
entry             600 drivers/gpu/drm/i915/gvt/gtt.c 		struct intel_gvt_gtt_entry *entry, unsigned long index)
entry             606 drivers/gpu/drm/i915/gvt/gtt.c 	entry->type = GTT_TYPE_GGTT_PTE;
entry             607 drivers/gpu/drm/i915/gvt/gtt.c 	pte_ops->get_entry(mm->ggtt_mm.virtual_ggtt, entry, index,
entry             612 drivers/gpu/drm/i915/gvt/gtt.c 		struct intel_gvt_gtt_entry *entry, unsigned long index)
entry             618 drivers/gpu/drm/i915/gvt/gtt.c 	pte_ops->set_entry(mm->ggtt_mm.virtual_ggtt, entry, index,
entry             623 drivers/gpu/drm/i915/gvt/gtt.c 		struct intel_gvt_gtt_entry *entry, unsigned long index)
entry             629 drivers/gpu/drm/i915/gvt/gtt.c 	pte_ops->get_entry(NULL, entry, index, false, 0, mm->vgpu);
entry             633 drivers/gpu/drm/i915/gvt/gtt.c 		struct intel_gvt_gtt_entry *entry, unsigned long index)
entry             639 drivers/gpu/drm/i915/gvt/gtt.c 	pte_ops->set_entry(NULL, entry, index, false, 0, mm->vgpu);
entry             975 drivers/gpu/drm/i915/gvt/gtt.c 		struct intel_gvt_gtt_entry *entry)
entry             982 drivers/gpu/drm/i915/gvt/gtt.c 	pfn = ops->get_pfn(entry);
entry            1151 drivers/gpu/drm/i915/gvt/gtt.c 	struct intel_gvt_gtt_entry *entry)
entry            1159 drivers/gpu/drm/i915/gvt/gtt.c 	pfn = intel_gvt_hypervisor_gfn_to_mfn(vgpu, ops->get_pfn(entry));
entry            1218 drivers/gpu/drm/i915/gvt/gtt.c 	struct intel_gvt_gtt_entry entry = *se;
entry            1229 drivers/gpu/drm/i915/gvt/gtt.c 	entry.type = GTT_TYPE_PPGTT_PTE_4K_ENTRY;
entry            1230 drivers/gpu/drm/i915/gvt/gtt.c 	ops->set_64k_splited(&entry);
entry            1238 drivers/gpu/drm/i915/gvt/gtt.c 		ops->set_pfn(&entry, dma_addr >> PAGE_SHIFT);
entry            1239 drivers/gpu/drm/i915/gvt/gtt.c 		ppgtt_set_shadow_entry(spt, &entry, index + i);
entry            2195 drivers/gpu/drm/i915/gvt/gtt.c 		struct intel_gvt_gtt_entry *entry)
entry            2200 drivers/gpu/drm/i915/gvt/gtt.c 	pfn = pte_ops->get_pfn(entry);
entry            2784 drivers/gpu/drm/i915/gvt/gtt.c 	struct intel_gvt_gtt_entry entry = {.type = GTT_TYPE_GGTT_PTE};
entry            2789 drivers/gpu/drm/i915/gvt/gtt.c 	pte_ops->set_pfn(&entry, gvt->gtt.scratch_mfn);
entry            2790 drivers/gpu/drm/i915/gvt/gtt.c 	pte_ops->set_present(&entry);
entry            2799 drivers/gpu/drm/i915/gvt/gtt.c 		ggtt_set_host_entry(vgpu->gtt.ggtt_mm, &entry, index++);
entry            2809 drivers/gpu/drm/i915/gvt/gtt.c 		ggtt_set_host_entry(vgpu->gtt.ggtt_mm, &entry, index++);
entry             306 drivers/gpu/drm/i915/gvt/kvmgt.c 				struct gvt_dma *entry)
entry             308 drivers/gpu/drm/i915/gvt/kvmgt.c 	rb_erase(&entry->gfn_node, &vgpu->vdev.gfn_cache);
entry             309 drivers/gpu/drm/i915/gvt/kvmgt.c 	rb_erase(&entry->dma_addr_node, &vgpu->vdev.dma_addr_cache);
entry             310 drivers/gpu/drm/i915/gvt/kvmgt.c 	kfree(entry);
entry             705 drivers/gpu/drm/i915/gvt/kvmgt.c 		struct gvt_dma *entry;
entry             713 drivers/gpu/drm/i915/gvt/kvmgt.c 			entry = __gvt_cache_find_gfn(vgpu, iov_pfn);
entry             714 drivers/gpu/drm/i915/gvt/kvmgt.c 			if (!entry)
entry             717 drivers/gpu/drm/i915/gvt/kvmgt.c 			gvt_dma_unmap_page(vgpu, entry->gfn, entry->dma_addr,
entry             718 drivers/gpu/drm/i915/gvt/kvmgt.c 					   entry->size);
entry             719 drivers/gpu/drm/i915/gvt/kvmgt.c 			__gvt_cache_remove_entry(vgpu, entry);
entry            1889 drivers/gpu/drm/i915/gvt/kvmgt.c 	struct gvt_dma *entry;
entry            1900 drivers/gpu/drm/i915/gvt/kvmgt.c 	entry = __gvt_cache_find_gfn(info->vgpu, gfn);
entry            1901 drivers/gpu/drm/i915/gvt/kvmgt.c 	if (!entry) {
entry            1909 drivers/gpu/drm/i915/gvt/kvmgt.c 	} else if (entry->size != size) {
entry            1911 drivers/gpu/drm/i915/gvt/kvmgt.c 		gvt_dma_unmap_page(vgpu, gfn, entry->dma_addr, entry->size);
entry            1912 drivers/gpu/drm/i915/gvt/kvmgt.c 		__gvt_cache_remove_entry(vgpu, entry);
entry            1922 drivers/gpu/drm/i915/gvt/kvmgt.c 		kref_get(&entry->ref);
entry            1923 drivers/gpu/drm/i915/gvt/kvmgt.c 		*dma_addr = entry->dma_addr;
entry            1938 drivers/gpu/drm/i915/gvt/kvmgt.c 	struct gvt_dma *entry = container_of(ref, typeof(*entry), ref);
entry            1940 drivers/gpu/drm/i915/gvt/kvmgt.c 	gvt_dma_unmap_page(entry->vgpu, entry->gfn, entry->dma_addr,
entry            1941 drivers/gpu/drm/i915/gvt/kvmgt.c 			   entry->size);
entry            1942 drivers/gpu/drm/i915/gvt/kvmgt.c 	__gvt_cache_remove_entry(entry->vgpu, entry);
entry            1948 drivers/gpu/drm/i915/gvt/kvmgt.c 	struct gvt_dma *entry;
entry            1956 drivers/gpu/drm/i915/gvt/kvmgt.c 	entry = __gvt_cache_find_dma_addr(info->vgpu, dma_addr);
entry            1957 drivers/gpu/drm/i915/gvt/kvmgt.c 	if (entry)
entry            1958 drivers/gpu/drm/i915/gvt/kvmgt.c 		kref_put(&entry->ref, __gvt_dma_release);
entry            2957 drivers/gpu/drm/i915/i915_debugfs.c 	struct skl_ddb_entry *entry;
entry            2976 drivers/gpu/drm/i915/i915_debugfs.c 			entry = &crtc_state->wm.skl.plane_ddb_y[plane_id];
entry            2978 drivers/gpu/drm/i915/i915_debugfs.c 				   entry->start, entry->end,
entry            2979 drivers/gpu/drm/i915/i915_debugfs.c 				   skl_ddb_entry_size(entry));
entry            2982 drivers/gpu/drm/i915/i915_debugfs.c 		entry = &crtc_state->wm.skl.plane_ddb_y[PLANE_CURSOR];
entry            2983 drivers/gpu/drm/i915/i915_debugfs.c 		seq_printf(m, "  %-13s%8u%8u%8u\n", "Cursor", entry->start,
entry            2984 drivers/gpu/drm/i915/i915_debugfs.c 			   entry->end, skl_ddb_entry_size(entry));
entry             881 drivers/gpu/drm/i915/i915_drv.h static inline u16 skl_ddb_entry_size(const struct skl_ddb_entry *entry)
entry             883 drivers/gpu/drm/i915/i915_drv.h 	return entry->end - entry->start;
entry             776 drivers/gpu/drm/i915/i915_gem_gtt.c 	GEM_BUG_ON(atomic_read(px_used(pd)) > 2 * ARRAY_SIZE(pd->entry));
entry             779 drivers/gpu/drm/i915/i915_gem_gtt.c 	pd->entry[idx] = to;
entry             794 drivers/gpu/drm/i915/i915_gem_gtt.c 	pd->entry[idx] = NULL;
entry             939 drivers/gpu/drm/i915/i915_gem_gtt.c 		void **pde = pd->entry;
entry             979 drivers/gpu/drm/i915/i915_gem_gtt.c 		struct i915_page_table *pt = pd->entry[idx];
entry            1057 drivers/gpu/drm/i915/i915_gem_gtt.c 		struct i915_page_table *pt = pd->entry[idx];
entry            1091 drivers/gpu/drm/i915/i915_gem_gtt.c 			if (likely(!pd->entry[idx]))
entry            1094 drivers/gpu/drm/i915/i915_gem_gtt.c 				alloc = pt, pt = pd->entry[idx];
entry            1202 drivers/gpu/drm/i915/i915_gem_gtt.c 				pd = pdp->entry[gen8_pd_index(idx, 2)];
entry            1453 drivers/gpu/drm/i915/i915_gem_gtt.c 	GEM_BUG_ON(count > ARRAY_SIZE(pd->entry));
entry            1455 drivers/gpu/drm/i915/i915_gem_gtt.c 	pd = __alloc_pd(offsetof(typeof(*pd), entry[count]));
entry            1657 drivers/gpu/drm/i915/i915_gem_gtt.c 	GEM_BUG_ON(pd->entry[act_pt] == &vm->scratch[1]);
entry            1717 drivers/gpu/drm/i915/i915_gem_gtt.c 			if (pd->entry[pde] == &vm->scratch[1]) {
entry            1718 drivers/gpu/drm/i915/i915_gem_gtt.c 				pd->entry[pde] = pt;
entry            1726 drivers/gpu/drm/i915/i915_gem_gtt.c 				pt = pd->entry[pde];
entry            1770 drivers/gpu/drm/i915/i915_gem_gtt.c 	memset_p(pd->entry, &vm->scratch[1], I915_PDES);
entry            1856 drivers/gpu/drm/i915/i915_gem_gtt.c 		pd->entry[pde] = scratch;
entry            2681 drivers/gpu/drm/i915/i915_gem_gtt.c 	struct drm_mm_node *entry;
entry            2715 drivers/gpu/drm/i915/i915_gem_gtt.c 	drm_mm_for_each_hole(entry, &ggtt->vm.mm, hole_start, hole_end) {
entry             244 drivers/gpu/drm/i915/i915_gem_gtt.h 	void *entry[512];
entry             520 drivers/gpu/drm/i915/i915_gem_gtt.h 	return pd->entry[n];
entry             527 drivers/gpu/drm/i915/i915_gem_gtt.h 	return pdp->entry[n];
entry             533 drivers/gpu/drm/i915/i915_gem_gtt.h 	struct i915_page_dma *pt = ppgtt->pd->entry[n];
entry             150 drivers/gpu/drm/i915/i915_sw_fence.c 		list_for_each_entry_safe(pos, next, &x->head, entry) {
entry             154 drivers/gpu/drm/i915/i915_sw_fence.c 				list_move_tail(&pos->entry, continuation);
entry             160 drivers/gpu/drm/i915/i915_sw_fence.c 			list_for_each_entry_safe(pos, next, &x->head, entry) {
entry             240 drivers/gpu/drm/i915/i915_sw_fence.c 	list_del(&wq->entry);
entry             259 drivers/gpu/drm/i915/i915_sw_fence.c 	list_for_each_entry(wq, &fence->wait.head, entry) {
entry             277 drivers/gpu/drm/i915/i915_sw_fence.c 	list_for_each_entry(wq, &fence->wait.head, entry) {
entry             338 drivers/gpu/drm/i915/i915_sw_fence.c 	INIT_LIST_HEAD(&wq->entry);
entry            3974 drivers/gpu/drm/i915/intel_pm.c 				       struct skl_ddb_entry *entry, u32 reg)
entry            3977 drivers/gpu/drm/i915/intel_pm.c 	entry->start = reg & DDB_ENTRY_MASK;
entry            3978 drivers/gpu/drm/i915/intel_pm.c 	entry->end = (reg >> DDB_ENTRY_END_SHIFT) & DDB_ENTRY_MASK;
entry            3980 drivers/gpu/drm/i915/intel_pm.c 	if (entry->end)
entry            3981 drivers/gpu/drm/i915/intel_pm.c 		entry->end += 1;
entry            5123 drivers/gpu/drm/i915/intel_pm.c 				const struct skl_ddb_entry *entry)
entry            5125 drivers/gpu/drm/i915/intel_pm.c 	if (entry->end)
entry            5126 drivers/gpu/drm/i915/intel_pm.c 		I915_WRITE_FW(reg, (entry->end - 1) << 16 | entry->start);
entry             821 drivers/gpu/drm/i915/intel_uncore.c static int fw_range_cmp(u32 offset, const struct intel_forcewake_range *entry)
entry             823 drivers/gpu/drm/i915/intel_uncore.c 	if (offset < entry->start)
entry             825 drivers/gpu/drm/i915/intel_uncore.c 	else if (offset > entry->end)
entry             853 drivers/gpu/drm/i915/intel_uncore.c 	const struct intel_forcewake_range *entry;
entry             855 drivers/gpu/drm/i915/intel_uncore.c 	entry = BSEARCH(offset,
entry             860 drivers/gpu/drm/i915/intel_uncore.c 	if (!entry)
entry             868 drivers/gpu/drm/i915/intel_uncore.c 	if (entry->domains == FORCEWAKE_ALL)
entry             871 drivers/gpu/drm/i915/intel_uncore.c 	WARN(entry->domains & ~uncore->fw_domains,
entry             873 drivers/gpu/drm/i915/intel_uncore.c 	     entry->domains & ~uncore->fw_domains, offset);
entry             875 drivers/gpu/drm/i915/intel_uncore.c 	return entry->domains;
entry            1820 drivers/gpu/drm/i915/intel_uncore.c 	struct reg_whitelist const *entry;
entry            1826 drivers/gpu/drm/i915/intel_uncore.c 	entry = reg_read_whitelist;
entry            1829 drivers/gpu/drm/i915/intel_uncore.c 		u32 entry_offset = i915_mmio_reg_offset(entry->offset_ldw);
entry            1831 drivers/gpu/drm/i915/intel_uncore.c 		GEM_BUG_ON(!is_power_of_2(entry->size));
entry            1832 drivers/gpu/drm/i915/intel_uncore.c 		GEM_BUG_ON(entry->size > 8);
entry            1833 drivers/gpu/drm/i915/intel_uncore.c 		GEM_BUG_ON(entry_offset & (entry->size - 1));
entry            1835 drivers/gpu/drm/i915/intel_uncore.c 		if (INTEL_INFO(i915)->gen_mask & entry->gen_mask &&
entry            1836 drivers/gpu/drm/i915/intel_uncore.c 		    entry_offset == (reg->offset & -entry->size))
entry            1838 drivers/gpu/drm/i915/intel_uncore.c 		entry++;
entry            1845 drivers/gpu/drm/i915/intel_uncore.c 	flags = reg->offset & (entry->size - 1);
entry            1848 drivers/gpu/drm/i915/intel_uncore.c 		if (entry->size == 8 && flags == I915_REG_READ_8B_WA)
entry            1850 drivers/gpu/drm/i915/intel_uncore.c 							    entry->offset_ldw,
entry            1851 drivers/gpu/drm/i915/intel_uncore.c 							    entry->offset_udw);
entry            1852 drivers/gpu/drm/i915/intel_uncore.c 		else if (entry->size == 8 && flags == 0)
entry            1854 drivers/gpu/drm/i915/intel_uncore.c 						       entry->offset_ldw);
entry            1855 drivers/gpu/drm/i915/intel_uncore.c 		else if (entry->size == 4 && flags == 0)
entry            1856 drivers/gpu/drm/i915/intel_uncore.c 			reg->val = intel_uncore_read(uncore, entry->offset_ldw);
entry            1857 drivers/gpu/drm/i915/intel_uncore.c 		else if (entry->size == 2 && flags == 0)
entry            1859 drivers/gpu/drm/i915/intel_uncore.c 						       entry->offset_ldw);
entry            1860 drivers/gpu/drm/i915/intel_uncore.c 		else if (entry->size == 1 && flags == 0)
entry            1862 drivers/gpu/drm/i915/intel_uncore.c 						      entry->offset_ldw);
entry             226 drivers/gpu/drm/mga/mga_dma.c 	drm_mga_freelist_t *entry;
entry             235 drivers/gpu/drm/mga/mga_dma.c 	for (entry = dev_priv->head->next; entry; entry = entry->next) {
entry             237 drivers/gpu/drm/mga/mga_dma.c 			 entry, entry->buf->idx, entry->age.head,
entry             238 drivers/gpu/drm/mga/mga_dma.c 			 (unsigned long)(entry->age.head - dev_priv->primary->offset));
entry             249 drivers/gpu/drm/mga/mga_dma.c 	drm_mga_freelist_t *entry;
entry             263 drivers/gpu/drm/mga/mga_dma.c 		entry = kzalloc(sizeof(drm_mga_freelist_t), GFP_KERNEL);
entry             264 drivers/gpu/drm/mga/mga_dma.c 		if (entry == NULL)
entry             267 drivers/gpu/drm/mga/mga_dma.c 		entry->next = dev_priv->head->next;
entry             268 drivers/gpu/drm/mga/mga_dma.c 		entry->prev = dev_priv->head;
entry             269 drivers/gpu/drm/mga/mga_dma.c 		SET_AGE(&entry->age, MGA_BUFFER_FREE, 0);
entry             270 drivers/gpu/drm/mga/mga_dma.c 		entry->buf = buf;
entry             273 drivers/gpu/drm/mga/mga_dma.c 			dev_priv->head->next->prev = entry;
entry             274 drivers/gpu/drm/mga/mga_dma.c 		if (entry->next == NULL)
entry             275 drivers/gpu/drm/mga/mga_dma.c 			dev_priv->tail = entry;
entry             277 drivers/gpu/drm/mga/mga_dma.c 		buf_priv->list_entry = entry;
entry             281 drivers/gpu/drm/mga/mga_dma.c 		dev_priv->head->next = entry;
entry             290 drivers/gpu/drm/mga/mga_dma.c 	drm_mga_freelist_t *entry;
entry             294 drivers/gpu/drm/mga/mga_dma.c 	entry = dev_priv->head;
entry             295 drivers/gpu/drm/mga/mga_dma.c 	while (entry) {
entry             296 drivers/gpu/drm/mga/mga_dma.c 		next = entry->next;
entry             297 drivers/gpu/drm/mga/mga_dma.c 		kfree(entry);
entry             298 drivers/gpu/drm/mga/mga_dma.c 		entry = next;
entry             359 drivers/gpu/drm/mga/mga_dma.c 	drm_mga_freelist_t *head, *entry, *prev;
entry             366 drivers/gpu/drm/mga/mga_dma.c 	entry = buf_priv->list_entry;
entry             370 drivers/gpu/drm/mga/mga_dma.c 		SET_AGE(&entry->age, MGA_BUFFER_FREE, 0);
entry             372 drivers/gpu/drm/mga/mga_dma.c 		prev->next = entry;
entry             373 drivers/gpu/drm/mga/mga_dma.c 		entry->prev = prev;
entry             374 drivers/gpu/drm/mga/mga_dma.c 		entry->next = NULL;
entry             377 drivers/gpu/drm/mga/mga_dma.c 		head->next = entry;
entry             378 drivers/gpu/drm/mga/mga_dma.c 		prev->prev = entry;
entry             379 drivers/gpu/drm/mga/mga_dma.c 		entry->prev = head;
entry             380 drivers/gpu/drm/mga/mga_dma.c 		entry->next = prev;
entry             368 drivers/gpu/drm/mga/mga_drv.h 	drm_mga_freelist_t *entry = (buf_priv)->list_entry;		\
entry             370 drivers/gpu/drm/mga/mga_drv.h 		entry->age.head = (dev_priv->prim.tail +		\
entry             372 drivers/gpu/drm/mga/mga_drv.h 		entry->age.wrap = dev_priv->sarea_priv->last_wrap;	\
entry             374 drivers/gpu/drm/mga/mga_drv.h 		entry->age.head = 0;					\
entry             375 drivers/gpu/drm/mga/mga_drv.h 		entry->age.wrap = 0;					\
entry             186 drivers/gpu/drm/msm/adreno/adreno_device.c static inline bool _rev_match(uint8_t entry, uint8_t id)
entry             188 drivers/gpu/drm/msm/adreno/adreno_device.c 	return (entry == ANY_ID) || (entry == id);
entry             446 drivers/gpu/drm/msm/disp/dpu1/dpu_core_perf.c 	struct dentry *entry;
entry             448 drivers/gpu/drm/msm/disp/dpu1/dpu_core_perf.c 	entry = debugfs_create_dir("core_perf", parent);
entry             450 drivers/gpu/drm/msm/disp/dpu1/dpu_core_perf.c 	debugfs_create_u64("max_core_clk_rate", 0600, entry,
entry             452 drivers/gpu/drm/msm/disp/dpu1/dpu_core_perf.c 	debugfs_create_u64("core_clk_rate", 0600, entry,
entry             454 drivers/gpu/drm/msm/disp/dpu1/dpu_core_perf.c 	debugfs_create_u32("enable_bw_release", 0600, entry,
entry             456 drivers/gpu/drm/msm/disp/dpu1/dpu_core_perf.c 	debugfs_create_u32("threshold_low", 0600, entry,
entry             458 drivers/gpu/drm/msm/disp/dpu1/dpu_core_perf.c 	debugfs_create_u32("threshold_high", 0600, entry,
entry             460 drivers/gpu/drm/msm/disp/dpu1/dpu_core_perf.c 	debugfs_create_u32("min_core_ib", 0600, entry,
entry             462 drivers/gpu/drm/msm/disp/dpu1/dpu_core_perf.c 	debugfs_create_u32("min_llcc_ib", 0600, entry,
entry             464 drivers/gpu/drm/msm/disp/dpu1/dpu_core_perf.c 	debugfs_create_u32("min_dram_ib", 0600, entry,
entry             466 drivers/gpu/drm/msm/disp/dpu1/dpu_core_perf.c 	debugfs_create_file("perf_mode", 0600, entry,
entry             468 drivers/gpu/drm/msm/disp/dpu1/dpu_core_perf.c 	debugfs_create_u64("fix_core_clk_rate", 0600, entry,
entry             470 drivers/gpu/drm/msm/disp/dpu1/dpu_core_perf.c 	debugfs_create_u64("fix_core_ib_vote", 0600, entry,
entry             472 drivers/gpu/drm/msm/disp/dpu1/dpu_core_perf.c 	debugfs_create_u64("fix_core_ab_vote", 0600, entry,
entry             135 drivers/gpu/drm/msm/disp/dpu1/dpu_kms.c 	struct dentry *entry = debugfs_create_dir("danger", parent);
entry             137 drivers/gpu/drm/msm/disp/dpu1/dpu_kms.c 	debugfs_create_file("danger_status", 0600, entry,
entry             139 drivers/gpu/drm/msm/disp/dpu1/dpu_kms.c 	debugfs_create_file("safe_status", 0600, entry,
entry             226 drivers/gpu/drm/msm/disp/dpu1/dpu_kms.c 	struct dentry *entry;
entry             231 drivers/gpu/drm/msm/disp/dpu1/dpu_kms.c 	entry = debugfs_create_dir("debug", minor->debugfs_root);
entry             233 drivers/gpu/drm/msm/disp/dpu1/dpu_kms.c 	debugfs_create_x32(DPU_DEBUGFS_HWMASKNAME, 0600, entry, p);
entry             235 drivers/gpu/drm/msm/disp/dpu1/dpu_kms.c 	dpu_debugfs_danger_init(dpu_kms, entry);
entry             236 drivers/gpu/drm/msm/disp/dpu1/dpu_kms.c 	dpu_debugfs_vbif_init(dpu_kms, entry);
entry             237 drivers/gpu/drm/msm/disp/dpu1/dpu_kms.c 	dpu_debugfs_core_irq_init(dpu_kms, entry);
entry             239 drivers/gpu/drm/msm/disp/dpu1/dpu_kms.c 	return dpu_core_perf_debugfs_init(dpu_kms, entry);
entry             300 drivers/gpu/drm/msm/disp/dpu1/dpu_vbif.c 	struct dentry *entry, *debugfs_vbif;
entry             303 drivers/gpu/drm/msm/disp/dpu1/dpu_vbif.c 	entry = debugfs_create_dir("vbif", debugfs_root);
entry             310 drivers/gpu/drm/msm/disp/dpu1/dpu_vbif.c 		debugfs_vbif = debugfs_create_dir(vbif_name, entry);
entry             357 drivers/gpu/drm/msm/hdmi/hdmi.c #define HDMI_CFG(item, entry) \
entry             358 drivers/gpu/drm/msm/hdmi/hdmi.c 	.item ## _names = item ##_names_ ## entry, \
entry             359 drivers/gpu/drm/msm/hdmi/hdmi.c 	.item ## _cnt   = ARRAY_SIZE(item ## _names_ ## entry)
entry              21 drivers/gpu/drm/msm/msm_submitqueue.c 	struct msm_gpu_submitqueue *entry;
entry              28 drivers/gpu/drm/msm/msm_submitqueue.c 	list_for_each_entry(entry, &ctx->submitqueues, node) {
entry              29 drivers/gpu/drm/msm/msm_submitqueue.c 		if (entry->id == id) {
entry              30 drivers/gpu/drm/msm/msm_submitqueue.c 			kref_get(&entry->ref);
entry              33 drivers/gpu/drm/msm/msm_submitqueue.c 			return entry;
entry              43 drivers/gpu/drm/msm/msm_submitqueue.c 	struct msm_gpu_submitqueue *entry, *tmp;
entry              52 drivers/gpu/drm/msm/msm_submitqueue.c 	list_for_each_entry_safe(entry, tmp, &ctx->submitqueues, node)
entry              53 drivers/gpu/drm/msm/msm_submitqueue.c 		msm_submitqueue_put(entry);
entry             159 drivers/gpu/drm/msm/msm_submitqueue.c 	struct msm_gpu_submitqueue *entry;
entry             173 drivers/gpu/drm/msm/msm_submitqueue.c 	list_for_each_entry(entry, &ctx->submitqueues, node) {
entry             174 drivers/gpu/drm/msm/msm_submitqueue.c 		if (entry->id == id) {
entry             175 drivers/gpu/drm/msm/msm_submitqueue.c 			list_del(&entry->node);
entry             178 drivers/gpu/drm/msm/msm_submitqueue.c 			msm_submitqueue_put(entry);
entry             527 drivers/gpu/drm/nouveau/dispnv04/dac.c nv04_dac_create(struct drm_connector *connector, struct dcb_output *entry)
entry             540 drivers/gpu/drm/nouveau/dispnv04/dac.c 	nv_encoder->dcb = entry;
entry             541 drivers/gpu/drm/nouveau/dispnv04/dac.c 	nv_encoder->or = ffs(entry->or) - 1;
entry             555 drivers/gpu/drm/nouveau/dispnv04/dac.c 	encoder->possible_crtcs = entry->heads;
entry             677 drivers/gpu/drm/nouveau/dispnv04/dfp.c nv04_dfp_create(struct drm_connector *connector, struct dcb_output *entry)
entry             684 drivers/gpu/drm/nouveau/dispnv04/dfp.c 	switch (entry->type) {
entry             706 drivers/gpu/drm/nouveau/dispnv04/dfp.c 	nv_encoder->dcb = entry;
entry             707 drivers/gpu/drm/nouveau/dispnv04/dfp.c 	nv_encoder->or = ffs(entry->or) - 1;
entry             712 drivers/gpu/drm/nouveau/dispnv04/dfp.c 	encoder->possible_crtcs = entry->heads;
entry             715 drivers/gpu/drm/nouveau/dispnv04/dfp.c 	if (entry->type == DCB_OUTPUT_TMDS &&
entry             716 drivers/gpu/drm/nouveau/dispnv04/dfp.c 	    entry->location != DCB_LOC_ON_CHIP)
entry             228 drivers/gpu/drm/nouveau/dispnv04/disp.c 		struct dcb_output *dcbent = &dcb->entry[i];
entry             202 drivers/gpu/drm/nouveau/dispnv04/tvnv04.c nv04_tv_create(struct drm_connector *connector, struct dcb_output *entry)
entry             209 drivers/gpu/drm/nouveau/dispnv04/tvnv04.c 	struct nvkm_i2c_bus *bus = nvkm_i2c_bus_find(i2c, entry->i2c_index);
entry             213 drivers/gpu/drm/nouveau/dispnv04/tvnv04.c 	type = nv04_tv_identify(dev, entry->i2c_index);
entry             232 drivers/gpu/drm/nouveau/dispnv04/tvnv04.c 	encoder->possible_crtcs = entry->heads;
entry             234 drivers/gpu/drm/nouveau/dispnv04/tvnv04.c 	nv_encoder->dcb = entry;
entry             235 drivers/gpu/drm/nouveau/dispnv04/tvnv04.c 	nv_encoder->or = ffs(entry->or) - 1;
entry             790 drivers/gpu/drm/nouveau/dispnv04/tvnv17.c nv17_tv_create(struct drm_connector *connector, struct dcb_output *entry)
entry             811 drivers/gpu/drm/nouveau/dispnv04/tvnv17.c 	tv_enc->base.dcb = entry;
entry             812 drivers/gpu/drm/nouveau/dispnv04/tvnv17.c 	tv_enc->base.or = ffs(entry->or) - 1;
entry             822 drivers/gpu/drm/nouveau/dispnv04/tvnv17.c 	encoder->possible_crtcs = entry->heads;
entry            2382 drivers/gpu/drm/nouveau/dispnv50/disp.c 	for (i = 0, dcbe = &dcb->entry[0]; i < dcb->entries; i++, dcbe++) {
entry             134 drivers/gpu/drm/nouveau/include/nvif/list.h __list_add(struct list_head *entry,
entry             137 drivers/gpu/drm/nouveau/include/nvif/list.h     next->prev = entry;
entry             138 drivers/gpu/drm/nouveau/include/nvif/list.h     entry->next = next;
entry             139 drivers/gpu/drm/nouveau/include/nvif/list.h     entry->prev = prev;
entry             140 drivers/gpu/drm/nouveau/include/nvif/list.h     prev->next = entry;
entry             159 drivers/gpu/drm/nouveau/include/nvif/list.h list_add(struct list_head *entry, struct list_head *head)
entry             161 drivers/gpu/drm/nouveau/include/nvif/list.h     __list_add(entry, head, head->next);
entry             180 drivers/gpu/drm/nouveau/include/nvif/list.h list_add_tail(struct list_head *entry, struct list_head *head)
entry             182 drivers/gpu/drm/nouveau/include/nvif/list.h     __list_add(entry, head->prev, head);
entry             207 drivers/gpu/drm/nouveau/include/nvif/list.h list_del(struct list_head *entry)
entry             209 drivers/gpu/drm/nouveau/include/nvif/list.h     __list_del(entry->prev, entry->next);
entry             213 drivers/gpu/drm/nouveau/include/nvif/list.h list_del_init(struct list_head *entry)
entry             215 drivers/gpu/drm/nouveau/include/nvif/list.h     __list_del(entry->prev, entry->next);
entry             216 drivers/gpu/drm/nouveau/include/nvif/list.h     INIT_LIST_HEAD(entry);
entry              66 drivers/gpu/drm/nouveau/include/nvkm/subdev/bios/dcb.h 		     (struct nvkm_bios *, void *, int index, u16 entry));
entry             974 drivers/gpu/drm/nouveau/nouveau_bios.c 	u8 entries, *entry;
entry             980 drivers/gpu/drm/nouveau/nouveau_bios.c 	entry   = &bios->data[bios->offset + 12];
entry             982 drivers/gpu/drm/nouveau/nouveau_bios.c 		if (entry[0] == id) {
entry             983 drivers/gpu/drm/nouveau/nouveau_bios.c 			bit->id = entry[0];
entry             984 drivers/gpu/drm/nouveau/nouveau_bios.c 			bit->version = entry[1];
entry             985 drivers/gpu/drm/nouveau/nouveau_bios.c 			bit->length = ROM16(entry[2]);
entry             986 drivers/gpu/drm/nouveau/nouveau_bios.c 			bit->offset = ROM16(entry[4]);
entry             987 drivers/gpu/drm/nouveau/nouveau_bios.c 			bit->data = ROMPTR(dev, entry[4]);
entry             991 drivers/gpu/drm/nouveau/nouveau_bios.c 		entry += bios->data[bios->offset + 9];
entry            1373 drivers/gpu/drm/nouveau/nouveau_bios.c 	struct dcb_output *entry = &dcb->entry[dcb->entries];
entry            1375 drivers/gpu/drm/nouveau/nouveau_bios.c 	memset(entry, 0, sizeof(struct dcb_output));
entry            1376 drivers/gpu/drm/nouveau/nouveau_bios.c 	entry->index = dcb->entries++;
entry            1378 drivers/gpu/drm/nouveau/nouveau_bios.c 	return entry;
entry            1384 drivers/gpu/drm/nouveau/nouveau_bios.c 	struct dcb_output *entry = new_dcb_entry(dcb);
entry            1386 drivers/gpu/drm/nouveau/nouveau_bios.c 	entry->type = type;
entry            1387 drivers/gpu/drm/nouveau/nouveau_bios.c 	entry->i2c_index = i2c;
entry            1388 drivers/gpu/drm/nouveau/nouveau_bios.c 	entry->heads = heads;
entry            1390 drivers/gpu/drm/nouveau/nouveau_bios.c 		entry->location = !DCB_LOC_ON_CHIP; /* ie OFF CHIP */
entry            1391 drivers/gpu/drm/nouveau/nouveau_bios.c 	entry->or = or;
entry            1396 drivers/gpu/drm/nouveau/nouveau_bios.c 		  uint32_t conn, uint32_t conf, struct dcb_output *entry)
entry            1401 drivers/gpu/drm/nouveau/nouveau_bios.c 	entry->type = conn & 0xf;
entry            1402 drivers/gpu/drm/nouveau/nouveau_bios.c 	entry->i2c_index = (conn >> 4) & 0xf;
entry            1403 drivers/gpu/drm/nouveau/nouveau_bios.c 	entry->heads = (conn >> 8) & 0xf;
entry            1404 drivers/gpu/drm/nouveau/nouveau_bios.c 	entry->connector = (conn >> 12) & 0xf;
entry            1405 drivers/gpu/drm/nouveau/nouveau_bios.c 	entry->bus = (conn >> 16) & 0xf;
entry            1406 drivers/gpu/drm/nouveau/nouveau_bios.c 	entry->location = (conn >> 20) & 0x3;
entry            1407 drivers/gpu/drm/nouveau/nouveau_bios.c 	entry->or = (conn >> 24) & 0xf;
entry            1409 drivers/gpu/drm/nouveau/nouveau_bios.c 	switch (entry->type) {
entry            1415 drivers/gpu/drm/nouveau/nouveau_bios.c 		entry->crtconf.maxfreq = (dcb->version < 0x30) ?
entry            1423 drivers/gpu/drm/nouveau/nouveau_bios.c 			entry->lvdsconf.use_straps_for_mode = true;
entry            1431 drivers/gpu/drm/nouveau/nouveau_bios.c 			entry->lvdsconf.use_straps_for_mode = true;
entry            1437 drivers/gpu/drm/nouveau/nouveau_bios.c 				entry->lvdsconf.use_power_scripts = true;
entry            1441 drivers/gpu/drm/nouveau/nouveau_bios.c 				entry->lvdsconf.use_acpi_for_edid = true;
entry            1443 drivers/gpu/drm/nouveau/nouveau_bios.c 				entry->lvdsconf.use_power_scripts = true;
entry            1444 drivers/gpu/drm/nouveau/nouveau_bios.c 			entry->lvdsconf.sor.link = (conf & 0x00000030) >> 4;
entry            1445 drivers/gpu/drm/nouveau/nouveau_bios.c 			link = entry->lvdsconf.sor.link;
entry            1463 drivers/gpu/drm/nouveau/nouveau_bios.c 			entry->tvconf.has_component_output = conf & (0x8 << 4);
entry            1465 drivers/gpu/drm/nouveau/nouveau_bios.c 			entry->tvconf.has_component_output = false;
entry            1470 drivers/gpu/drm/nouveau/nouveau_bios.c 		entry->dpconf.sor.link = (conf & 0x00000030) >> 4;
entry            1471 drivers/gpu/drm/nouveau/nouveau_bios.c 		entry->extdev = (conf & 0x0000ff00) >> 8;
entry            1474 drivers/gpu/drm/nouveau/nouveau_bios.c 			entry->dpconf.link_bw = 162000;
entry            1477 drivers/gpu/drm/nouveau/nouveau_bios.c 			entry->dpconf.link_bw = 270000;
entry            1480 drivers/gpu/drm/nouveau/nouveau_bios.c 			entry->dpconf.link_bw = 540000;
entry            1484 drivers/gpu/drm/nouveau/nouveau_bios.c 			entry->dpconf.link_bw = 810000;
entry            1490 drivers/gpu/drm/nouveau/nouveau_bios.c 			entry->dpconf.link_nr = 4;
entry            1494 drivers/gpu/drm/nouveau/nouveau_bios.c 			entry->dpconf.link_nr = 2;
entry            1497 drivers/gpu/drm/nouveau/nouveau_bios.c 			entry->dpconf.link_nr = 1;
entry            1500 drivers/gpu/drm/nouveau/nouveau_bios.c 		link = entry->dpconf.sor.link;
entry            1504 drivers/gpu/drm/nouveau/nouveau_bios.c 			entry->tmdsconf.sor.link = (conf & 0x00000030) >> 4;
entry            1505 drivers/gpu/drm/nouveau/nouveau_bios.c 			entry->extdev = (conf & 0x0000ff00) >> 8;
entry            1506 drivers/gpu/drm/nouveau/nouveau_bios.c 			link = entry->tmdsconf.sor.link;
entry            1509 drivers/gpu/drm/nouveau/nouveau_bios.c 			entry->tmdsconf.slave_addr = (conf & 0x00000700) >> 8;
entry            1511 drivers/gpu/drm/nouveau/nouveau_bios.c 			entry->tmdsconf.slave_addr = (conf & 0x00000070) >> 4;
entry            1525 drivers/gpu/drm/nouveau/nouveau_bios.c 		entry->duallink_possible =
entry            1526 drivers/gpu/drm/nouveau/nouveau_bios.c 			((1 << (ffs(entry->or) - 1)) * 3 == entry->or);
entry            1528 drivers/gpu/drm/nouveau/nouveau_bios.c 		entry->duallink_possible = (entry->sorconf.link == 3);
entry            1533 drivers/gpu/drm/nouveau/nouveau_bios.c 		entry->i2c_upper_default = true;
entry            1535 drivers/gpu/drm/nouveau/nouveau_bios.c 	entry->hasht = (entry->extdev << 8) | (entry->location << 4) |
entry            1536 drivers/gpu/drm/nouveau/nouveau_bios.c 			entry->type;
entry            1537 drivers/gpu/drm/nouveau/nouveau_bios.c 	entry->hashm = (entry->heads << 8) | (link << 6) | entry->or;
entry            1543 drivers/gpu/drm/nouveau/nouveau_bios.c 		  uint32_t conn, uint32_t conf, struct dcb_output *entry)
entry            1549 drivers/gpu/drm/nouveau/nouveau_bios.c 		entry->type = DCB_OUTPUT_ANALOG;
entry            1552 drivers/gpu/drm/nouveau/nouveau_bios.c 		entry->type = DCB_OUTPUT_TV;
entry            1557 drivers/gpu/drm/nouveau/nouveau_bios.c 			entry->type = DCB_OUTPUT_LVDS;
entry            1559 drivers/gpu/drm/nouveau/nouveau_bios.c 			entry->type = DCB_OUTPUT_TMDS;
entry            1562 drivers/gpu/drm/nouveau/nouveau_bios.c 		entry->type = DCB_OUTPUT_LVDS;
entry            1569 drivers/gpu/drm/nouveau/nouveau_bios.c 	entry->i2c_index = (conn & 0x0003c000) >> 14;
entry            1570 drivers/gpu/drm/nouveau/nouveau_bios.c 	entry->heads = ((conn & 0x001c0000) >> 18) + 1;
entry            1571 drivers/gpu/drm/nouveau/nouveau_bios.c 	entry->or = entry->heads; /* same as heads, hopefully safe enough */
entry            1572 drivers/gpu/drm/nouveau/nouveau_bios.c 	entry->location = (conn & 0x01e00000) >> 21;
entry            1573 drivers/gpu/drm/nouveau/nouveau_bios.c 	entry->bus = (conn & 0x0e000000) >> 25;
entry            1574 drivers/gpu/drm/nouveau/nouveau_bios.c 	entry->duallink_possible = false;
entry            1576 drivers/gpu/drm/nouveau/nouveau_bios.c 	switch (entry->type) {
entry            1578 drivers/gpu/drm/nouveau/nouveau_bios.c 		entry->crtconf.maxfreq = (conf & 0xffff) * 10;
entry            1581 drivers/gpu/drm/nouveau/nouveau_bios.c 		entry->tvconf.has_component_output = false;
entry            1585 drivers/gpu/drm/nouveau/nouveau_bios.c 			entry->lvdsconf.use_straps_for_mode = true;
entry            1586 drivers/gpu/drm/nouveau/nouveau_bios.c 		entry->lvdsconf.use_power_scripts = true;
entry            1608 drivers/gpu/drm/nouveau/nouveau_bios.c 		struct dcb_output *ient = &dcb->entry[i];
entry            1612 drivers/gpu/drm/nouveau/nouveau_bios.c 			struct dcb_output *jent = &dcb->entry[j];
entry            1632 drivers/gpu/drm/nouveau/nouveau_bios.c 		if (dcb->entry[i].type == 100)
entry            1636 drivers/gpu/drm/nouveau/nouveau_bios.c 			dcb->entry[newentries] = dcb->entry[i];
entry            1637 drivers/gpu/drm/nouveau/nouveau_bios.c 			dcb->entry[newentries].index = newentries;
entry            1793 drivers/gpu/drm/nouveau/nouveau_bios.c 		struct dcb_output *entry = new_dcb_entry(dcb);
entry            1798 drivers/gpu/drm/nouveau/nouveau_bios.c 			ret = parse_dcb20_entry(dev, dcb, conn, conf, entry);
entry            1800 drivers/gpu/drm/nouveau/nouveau_bios.c 			ret = parse_dcb15_entry(dev, dcb, conn, conf, entry);
entry            1808 drivers/gpu/drm/nouveau/nouveau_bios.c 		if (entry->type == DCB_OUTPUT_TV &&
entry            1809 drivers/gpu/drm/nouveau/nouveau_bios.c 		    entry->location == DCB_LOC_ON_CHIP)
entry            1810 drivers/gpu/drm/nouveau/nouveau_bios.c 			entry->i2c_index = 0x0f;
entry            1830 drivers/gpu/drm/nouveau/nouveau_bios.c 			if (dcbt->entry[i].connector)
entry            1842 drivers/gpu/drm/nouveau/nouveau_bios.c 		u8 i2c = dcbt->entry[i].i2c_index;
entry            1844 drivers/gpu/drm/nouveau/nouveau_bios.c 			dcbt->entry[i].connector = idx++;
entry            1848 drivers/gpu/drm/nouveau/nouveau_bios.c 			dcbt->entry[i].connector = map[i2c] - 1;
entry            1909 drivers/gpu/drm/nouveau/nouveau_bios.c static int load_nv17_hwsq_ucode_entry(struct drm_device *dev, struct nvbios *bios, uint16_t hwsq_offset, int entry)
entry            1926 drivers/gpu/drm/nouveau/nouveau_bios.c 	if (bios->data[hwsq_offset] <= entry) {
entry            1941 drivers/gpu/drm/nouveau/nouveau_bios.c 	hwsq_entry_offset = hwsq_offset + 2 + entry * bytes_to_write;
entry              57 drivers/gpu/drm/nouveau/nouveau_bios.h 	struct dcb_output entry[DCB_MAX_NUM_ENTRIES];
entry             213 drivers/gpu/drm/nouveau/nouveau_bo.c 	INIT_LIST_HEAD(&nvbo->entry);
entry              23 drivers/gpu/drm/nouveau/nouveau_bo.h 	struct list_head entry;
entry            1289 drivers/gpu/drm/nouveau/nouveau_connector.c 		u32 entry = ROM16(nv_connector->dcb[0]);
entry            1291 drivers/gpu/drm/nouveau/nouveau_connector.c 			entry |= (u32)ROM16(nv_connector->dcb[2]) << 16;
entry            1326 drivers/gpu/drm/nouveau/nouveau_connector.c 			if (dcbt->entry[i].connector == nv_connector->index)
entry            1327 drivers/gpu/drm/nouveau/nouveau_connector.c 				encoders |= (1 << dcbt->entry[i].type);
entry             338 drivers/gpu/drm/nouveau/nouveau_gem.c 		nvbo = list_entry(op->list.next, struct nouveau_bo, entry);
entry             358 drivers/gpu/drm/nouveau/nouveau_gem.c 		list_del(&nvbo->entry);
entry             457 drivers/gpu/drm/nouveau/nouveau_gem.c 			list_add_tail(&nvbo->entry, &both_list);
entry             460 drivers/gpu/drm/nouveau/nouveau_gem.c 			list_add_tail(&nvbo->entry, &vram_list);
entry             463 drivers/gpu/drm/nouveau/nouveau_gem.c 			list_add_tail(&nvbo->entry, &gart_list);
entry             467 drivers/gpu/drm/nouveau/nouveau_gem.c 			list_add_tail(&nvbo->entry, &both_list);
entry             496 drivers/gpu/drm/nouveau/nouveau_gem.c 	list_for_each_entry(nvbo, list, entry) {
entry              32 drivers/gpu/drm/nouveau/nvkm/subdev/bios/bit.c 		u32 entry   = bios->bit_offset + 12;
entry              34 drivers/gpu/drm/nouveau/nvkm/subdev/bios/bit.c 			if (nvbios_rd08(bios, entry + 0) == id) {
entry              35 drivers/gpu/drm/nouveau/nvkm/subdev/bios/bit.c 				bit->id      = nvbios_rd08(bios, entry + 0);
entry              36 drivers/gpu/drm/nouveau/nvkm/subdev/bios/bit.c 				bit->version = nvbios_rd08(bios, entry + 1);
entry              37 drivers/gpu/drm/nouveau/nvkm/subdev/bios/bit.c 				bit->length  = nvbios_rd16(bios, entry + 2);
entry              38 drivers/gpu/drm/nouveau/nvkm/subdev/bios/bit.c 				bit->offset  = nvbios_rd16(bios, entry + 4);
entry              42 drivers/gpu/drm/nouveau/nvkm/subdev/bios/bit.c 			entry += nvbios_rd08(bios, bios->bit_offset + 9);
entry              74 drivers/gpu/drm/nouveau/nvkm/subdev/bios/extdev.c 		   struct nvbios_extdev_func *entry)
entry              76 drivers/gpu/drm/nouveau/nvkm/subdev/bios/extdev.c 	entry->type = nvbios_rd08(bios, offset + 0);
entry              77 drivers/gpu/drm/nouveau/nvkm/subdev/bios/extdev.c 	entry->addr = nvbios_rd08(bios, offset + 1);
entry              78 drivers/gpu/drm/nouveau/nvkm/subdev/bios/extdev.c 	entry->bus = (nvbios_rd08(bios, offset + 2) >> 4) & 1;
entry              86 drivers/gpu/drm/nouveau/nvkm/subdev/bios/extdev.c 	u16 entry;
entry              88 drivers/gpu/drm/nouveau/nvkm/subdev/bios/extdev.c 	if (!(entry = nvbios_extdev_entry(bios, idx, &ver, &len)))
entry              91 drivers/gpu/drm/nouveau/nvkm/subdev/bios/extdev.c 	extdev_parse_entry(bios, entry, func);
entry             100 drivers/gpu/drm/nouveau/nvkm/subdev/bios/extdev.c 	u16 entry;
entry             103 drivers/gpu/drm/nouveau/nvkm/subdev/bios/extdev.c 	while ((entry = nvbios_extdev_entry(bios, i++, &ver, &len))) {
entry             104 drivers/gpu/drm/nouveau/nvkm/subdev/bios/extdev.c 		extdev_parse_entry(bios, entry, func);
entry              64 drivers/gpu/drm/nouveau/nvkm/subdev/bios/iccsense.c 	u32 table, entry;
entry              87 drivers/gpu/drm/nouveau/nvkm/subdev/bios/iccsense.c 		entry = table + hdr + i * len;
entry              91 drivers/gpu/drm/nouveau/nvkm/subdev/bios/iccsense.c 			if ((nvbios_rd08(bios, entry + 0x1) & 0xf8) == 0xf8)
entry              95 drivers/gpu/drm/nouveau/nvkm/subdev/bios/iccsense.c 			rail->extdev_id = nvbios_rd08(bios, entry + 0x2);
entry              99 drivers/gpu/drm/nouveau/nvkm/subdev/bios/iccsense.c 			rail->mode = nvbios_rd08(bios, entry);
entry             100 drivers/gpu/drm/nouveau/nvkm/subdev/bios/iccsense.c 			rail->extdev_id = nvbios_rd08(bios, entry + 0x1);
entry             122 drivers/gpu/drm/nouveau/nvkm/subdev/bios/iccsense.c 			rail->resistors[r].mohm = nvbios_rd08(bios, entry + res_start + r * 2);
entry             123 drivers/gpu/drm/nouveau/nvkm/subdev/bios/iccsense.c 			rail->resistors[r].enabled = !(nvbios_rd08(bios, entry + res_start + r * 2 + 1) & 0x40);
entry             125 drivers/gpu/drm/nouveau/nvkm/subdev/bios/iccsense.c 		rail->config = nvbios_rd16(bios, entry + res_start + rail->resistor_count * 2);
entry             105 drivers/gpu/drm/nouveau/nvkm/subdev/bios/power_budget.c                           u8 idx, struct nvbios_power_budget_entry *entry)
entry             110 drivers/gpu/drm/nouveau/nvkm/subdev/bios/power_budget.c 		|| !entry)
entry             116 drivers/gpu/drm/nouveau/nvkm/subdev/bios/power_budget.c 		entry->min_w = nvbios_rd32(bios, entry_offset + 0x2);
entry             117 drivers/gpu/drm/nouveau/nvkm/subdev/bios/power_budget.c 		entry->avg_w = nvbios_rd32(bios, entry_offset + 0x6);
entry             118 drivers/gpu/drm/nouveau/nvkm/subdev/bios/power_budget.c 		entry->max_w = nvbios_rd32(bios, entry_offset + 0xa);
entry             120 drivers/gpu/drm/nouveau/nvkm/subdev/bios/power_budget.c 		entry->min_w = 0;
entry             121 drivers/gpu/drm/nouveau/nvkm/subdev/bios/power_budget.c 		entry->max_w = nvbios_rd32(bios, entry_offset + 0x2);
entry             122 drivers/gpu/drm/nouveau/nvkm/subdev/bios/power_budget.c 		entry->avg_w = entry->max_w;
entry              73 drivers/gpu/drm/nouveau/nvkm/subdev/bios/therm.c 	u32 entry;
entry              83 drivers/gpu/drm/nouveau/nvkm/subdev/bios/therm.c 	while ((entry = nvbios_therm_entry(bios, i++, &ver, &len))) {
entry              84 drivers/gpu/drm/nouveau/nvkm/subdev/bios/therm.c 		s16 value = nvbios_rd16(bios, entry + 1);
entry              86 drivers/gpu/drm/nouveau/nvkm/subdev/bios/therm.c 		switch (nvbios_rd08(bios, entry + 0)) {
entry              95 drivers/gpu/drm/nouveau/nvkm/subdev/bios/therm.c 				offset = ((s8) nvbios_rd08(bios, entry + 2)) / 2;
entry             157 drivers/gpu/drm/nouveau/nvkm/subdev/bios/therm.c 	u32 entry;
entry             165 drivers/gpu/drm/nouveau/nvkm/subdev/bios/therm.c 	while ((entry = nvbios_therm_entry(bios, i++, &ver, &len))) {
entry             166 drivers/gpu/drm/nouveau/nvkm/subdev/bios/therm.c 		s16 value = nvbios_rd16(bios, entry + 1);
entry             168 drivers/gpu/drm/nouveau/nvkm/subdev/bios/therm.c 		switch (nvbios_rd08(bios, entry + 0)) {
entry             199 drivers/gpu/drm/nouveau/nvkm/subdev/bios/therm.c 			fan->linear_min_temp = nvbios_rd08(bios, entry + 1);
entry             200 drivers/gpu/drm/nouveau/nvkm/subdev/bios/therm.c 			fan->linear_max_temp = nvbios_rd08(bios, entry + 2);
entry              32 drivers/gpu/drm/nouveau/nvkm/subdev/gpio/gf119.c 	u16 entry;
entry              35 drivers/gpu/drm/nouveau/nvkm/subdev/gpio/gf119.c 	while ((entry = dcb_gpio_entry(bios, 0, ++ent, &ver, &len))) {
entry              36 drivers/gpu/drm/nouveau/nvkm/subdev/gpio/gf119.c 		u32 data = nvbios_rd32(bios, entry);
entry              32 drivers/gpu/drm/nouveau/nvkm/subdev/gpio/nv50.c 	u16 entry;
entry              35 drivers/gpu/drm/nouveau/nvkm/subdev/gpio/nv50.c 	while ((entry = dcb_gpio_entry(bios, 0, ++ent, &ver, &len))) {
entry              37 drivers/gpu/drm/nouveau/nvkm/subdev/gpio/nv50.c 		u32 data = nvbios_rd32(bios, entry);
entry             229 drivers/gpu/drm/nouveau/nvkm/subdev/iccsense/base.c 		struct nvbios_power_budget_entry entry;
entry             231 drivers/gpu/drm/nouveau/nvkm/subdev/iccsense/base.c 		                                budget.cap_entry, &entry);
entry             233 drivers/gpu/drm/nouveau/nvkm/subdev/iccsense/base.c 			iccsense->power_w_max  = entry.avg_w;
entry             234 drivers/gpu/drm/nouveau/nvkm/subdev/iccsense/base.c 			iccsense->power_w_crit = entry.max_w;
entry             908 drivers/gpu/drm/omapdrm/dss/dss.c 	struct dss_debugfs_entry *entry = inode->i_private;
entry             910 drivers/gpu/drm/omapdrm/dss/dss.c 	return single_open(file, entry->show_fn, entry->data);
entry             925 drivers/gpu/drm/omapdrm/dss/dss.c 	struct dss_debugfs_entry *entry;
entry             927 drivers/gpu/drm/omapdrm/dss/dss.c 	entry = kzalloc(sizeof(*entry), GFP_KERNEL);
entry             928 drivers/gpu/drm/omapdrm/dss/dss.c 	if (!entry)
entry             931 drivers/gpu/drm/omapdrm/dss/dss.c 	entry->show_fn = show_fn;
entry             932 drivers/gpu/drm/omapdrm/dss/dss.c 	entry->data = data;
entry             933 drivers/gpu/drm/omapdrm/dss/dss.c 	entry->dentry = debugfs_create_file(name, 0444, dss->debugfs.root,
entry             934 drivers/gpu/drm/omapdrm/dss/dss.c 					    entry, &dss_debug_fops);
entry             936 drivers/gpu/drm/omapdrm/dss/dss.c 	return entry;
entry             939 drivers/gpu/drm/omapdrm/dss/dss.c void dss_debugfs_remove_file(struct dss_debugfs_entry *entry)
entry             941 drivers/gpu/drm/omapdrm/dss/dss.c 	if (IS_ERR_OR_NULL(entry))
entry             944 drivers/gpu/drm/omapdrm/dss/dss.c 	debugfs_remove(entry->dentry);
entry             945 drivers/gpu/drm/omapdrm/dss/dss.c 	kfree(entry);
entry             287 drivers/gpu/drm/omapdrm/dss/dss.h void dss_debugfs_remove_file(struct dss_debugfs_entry *entry);
entry             297 drivers/gpu/drm/omapdrm/dss/dss.h static inline void dss_debugfs_remove_file(struct dss_debugfs_entry *entry)
entry             121 drivers/gpu/drm/omapdrm/omap_gem.c 	struct omap_drm_usergart_entry entry[NUM_USERGART_ENTRIES];
entry             167 drivers/gpu/drm/omapdrm/omap_gem.c 		enum tiler_fmt fmt, struct omap_drm_usergart_entry *entry)
entry             174 drivers/gpu/drm/omapdrm/omap_gem.c 			(entry->obj_pgoff << PAGE_SHIFT);
entry             190 drivers/gpu/drm/omapdrm/omap_gem.c 	entry->obj = NULL;
entry             204 drivers/gpu/drm/omapdrm/omap_gem.c 			struct omap_drm_usergart_entry *entry =
entry             205 drivers/gpu/drm/omapdrm/omap_gem.c 				&priv->usergart[fmt].entry[i];
entry             207 drivers/gpu/drm/omapdrm/omap_gem.c 			if (entry->obj == obj)
entry             208 drivers/gpu/drm/omapdrm/omap_gem.c 				omap_gem_evict_entry(obj, fmt, entry);
entry             377 drivers/gpu/drm/omapdrm/omap_gem.c 	struct omap_drm_usergart_entry *entry;
entry             416 drivers/gpu/drm/omapdrm/omap_gem.c 	entry = &priv->usergart[fmt].entry[priv->usergart[fmt].last];
entry             419 drivers/gpu/drm/omapdrm/omap_gem.c 	if (entry->obj)
entry             420 drivers/gpu/drm/omapdrm/omap_gem.c 		omap_gem_evict_entry(entry->obj, fmt, entry);
entry             422 drivers/gpu/drm/omapdrm/omap_gem.c 	entry->obj = obj;
entry             423 drivers/gpu/drm/omapdrm/omap_gem.c 	entry->obj_pgoff = base_pgoff;
entry             431 drivers/gpu/drm/omapdrm/omap_gem.c 		entry->obj_pgoff += off;
entry             450 drivers/gpu/drm/omapdrm/omap_gem.c 	err = tiler_pin(entry->block, pages, ARRAY_SIZE(pages), 0, true);
entry             457 drivers/gpu/drm/omapdrm/omap_gem.c 	pfn = entry->dma_addr >> PAGE_SHIFT;
entry            1349 drivers/gpu/drm/omapdrm/omap_gem.c 			struct omap_drm_usergart_entry *entry;
entry            1352 drivers/gpu/drm/omapdrm/omap_gem.c 			entry = &usergart[i].entry[j];
entry            1360 drivers/gpu/drm/omapdrm/omap_gem.c 			entry->dma_addr = tiler_ssptr(block);
entry            1361 drivers/gpu/drm/omapdrm/omap_gem.c 			entry->block = block;
entry            1364 drivers/gpu/drm/omapdrm/omap_gem.c 					&entry->dma_addr,
entry             240 drivers/gpu/drm/panel/panel-raydium-rm67191.c 		const struct cmd_set_entry *entry = &manufacturer_cmd_set[i];
entry             241 drivers/gpu/drm/panel/panel-raydium-rm67191.c 		u8 buffer[2] = { entry->cmd, entry->param };
entry             163 drivers/gpu/drm/qxl/qxl_release.c 		struct qxl_bo_list *entry;
entry             166 drivers/gpu/drm/qxl/qxl_release.c 		entry = container_of(release->bos.next,
entry             168 drivers/gpu/drm/qxl/qxl_release.c 		bo = to_qxl_bo(entry->tv.bo);
entry             170 drivers/gpu/drm/qxl/qxl_release.c 		list_del(&entry->tv.head);
entry             171 drivers/gpu/drm/qxl/qxl_release.c 		kfree(entry);
entry             211 drivers/gpu/drm/qxl/qxl_release.c 	struct qxl_bo_list *entry;
entry             213 drivers/gpu/drm/qxl/qxl_release.c 	list_for_each_entry(entry, &release->bos, tv.head) {
entry             214 drivers/gpu/drm/qxl/qxl_release.c 		if (entry->tv.bo == &bo->tbo)
entry             218 drivers/gpu/drm/qxl/qxl_release.c 	entry = kmalloc(sizeof(struct qxl_bo_list), GFP_KERNEL);
entry             219 drivers/gpu/drm/qxl/qxl_release.c 	if (!entry)
entry             223 drivers/gpu/drm/qxl/qxl_release.c 	entry->tv.bo = &bo->tbo;
entry             224 drivers/gpu/drm/qxl/qxl_release.c 	entry->tv.num_shared = 0;
entry             225 drivers/gpu/drm/qxl/qxl_release.c 	list_add_tail(&entry->tv.head, &release->bos);
entry             255 drivers/gpu/drm/qxl/qxl_release.c 	struct qxl_bo_list *entry;
entry             267 drivers/gpu/drm/qxl/qxl_release.c 	list_for_each_entry(entry, &release->bos, tv.head) {
entry             268 drivers/gpu/drm/qxl/qxl_release.c 		struct qxl_bo *bo = to_qxl_bo(entry->tv.bo);
entry             434 drivers/gpu/drm/qxl/qxl_release.c 	struct ttm_validate_buffer *entry;
entry             458 drivers/gpu/drm/qxl/qxl_release.c 	list_for_each_entry(entry, &release->bos, head) {
entry             459 drivers/gpu/drm/qxl/qxl_release.c 		bo = entry->bo;
entry             779 drivers/gpu/drm/r128/r128_cce.c 	drm_r128_freelist_t *entry;
entry             792 drivers/gpu/drm/r128/r128_cce.c 		entry = kmalloc(sizeof(drm_r128_freelist_t), GFP_KERNEL);
entry             793 drivers/gpu/drm/r128/r128_cce.c 		if (!entry)
entry             796 drivers/gpu/drm/r128/r128_cce.c 		entry->age = R128_BUFFER_FREE;
entry             797 drivers/gpu/drm/r128/r128_cce.c 		entry->buf = buf;
entry             798 drivers/gpu/drm/r128/r128_cce.c 		entry->prev = dev_priv->head;
entry             799 drivers/gpu/drm/r128/r128_cce.c 		entry->next = dev_priv->head->next;
entry             800 drivers/gpu/drm/r128/r128_cce.c 		if (!entry->next)
entry             801 drivers/gpu/drm/r128/r128_cce.c 			dev_priv->tail = entry;
entry             805 drivers/gpu/drm/r128/r128_cce.c 		buf_priv->list_entry = entry;
entry             807 drivers/gpu/drm/r128/r128_cce.c 		dev_priv->head->next = entry;
entry             810 drivers/gpu/drm/r128/r128_cce.c 			dev_priv->head->next->prev = entry;
entry            4704 drivers/gpu/drm/radeon/ci_dpm.c static void ci_convert_mc_registers(const struct ci_mc_reg_entry *entry,
entry            4712 drivers/gpu/drm/radeon/ci_dpm.c 			data->value[i] = cpu_to_be32(entry->mc_data[j]);
entry             814 drivers/gpu/drm/radeon/cypress_dpm.c static void cypress_convert_mc_registers(struct evergreen_mc_reg_entry *entry,
entry             822 drivers/gpu/drm/radeon/cypress_dpm.c 			data->value[i] = cpu_to_be32(entry->mc_data[j]);
entry            1032 drivers/gpu/drm/radeon/cypress_dpm.c 						     struct evergreen_mc_reg_entry *entry)
entry            1038 drivers/gpu/drm/radeon/cypress_dpm.c 		entry->mc_data[i] =
entry            2943 drivers/gpu/drm/radeon/ni_dpm.c static void ni_convert_mc_registers(struct ni_mc_reg_entry *entry,
entry            2951 drivers/gpu/drm/radeon/ni_dpm.c 			data->value[i] = cpu_to_be32(entry->mc_data[j]);
entry             699 drivers/gpu/drm/radeon/r100.c 			    uint64_t entry)
entry             702 drivers/gpu/drm/radeon/r100.c 	gtt[i] = cpu_to_le32(lower_32_bits(entry));
entry             121 drivers/gpu/drm/radeon/r300.c 			      uint64_t entry)
entry             128 drivers/gpu/drm/radeon/r300.c 	writel(entry, ((void __iomem *)ptr) + (i * 4));
entry             826 drivers/gpu/drm/radeon/r600_dpm.c 	ATOM_PPLIB_Clock_Voltage_Dependency_Record *entry;
entry             832 drivers/gpu/drm/radeon/r600_dpm.c 	entry = &atom_table->entries[0];
entry             834 drivers/gpu/drm/radeon/r600_dpm.c 		radeon_table->entries[i].clk = le16_to_cpu(entry->usClockLow) |
entry             835 drivers/gpu/drm/radeon/r600_dpm.c 			(entry->ucClockHigh << 16);
entry             836 drivers/gpu/drm/radeon/r600_dpm.c 		radeon_table->entries[i].v = le16_to_cpu(entry->usVoltage);
entry             837 drivers/gpu/drm/radeon/r600_dpm.c 		entry = (ATOM_PPLIB_Clock_Voltage_Dependency_Record *)
entry             838 drivers/gpu/drm/radeon/r600_dpm.c 			((u8 *)entry + sizeof(ATOM_PPLIB_Clock_Voltage_Dependency_Record));
entry             990 drivers/gpu/drm/radeon/r600_dpm.c 			ATOM_PPLIB_PhaseSheddingLimits_Record *entry;
entry            1001 drivers/gpu/drm/radeon/r600_dpm.c 			entry = &psl->entries[0];
entry            1004 drivers/gpu/drm/radeon/r600_dpm.c 					le16_to_cpu(entry->usSclkLow) | (entry->ucSclkHigh << 16);
entry            1006 drivers/gpu/drm/radeon/r600_dpm.c 					le16_to_cpu(entry->usMclkLow) | (entry->ucMclkHigh << 16);
entry            1008 drivers/gpu/drm/radeon/r600_dpm.c 					le16_to_cpu(entry->usVoltage);
entry            1009 drivers/gpu/drm/radeon/r600_dpm.c 				entry = (ATOM_PPLIB_PhaseSheddingLimits_Record *)
entry            1010 drivers/gpu/drm/radeon/r600_dpm.c 					((u8 *)entry + sizeof(ATOM_PPLIB_PhaseSheddingLimits_Record));
entry            1037 drivers/gpu/drm/radeon/r600_dpm.c 			ATOM_PPLIB_CAC_Leakage_Record *entry;
entry            1044 drivers/gpu/drm/radeon/r600_dpm.c 			entry = &cac_table->entries[0];
entry            1048 drivers/gpu/drm/radeon/r600_dpm.c 						le16_to_cpu(entry->usVddc1);
entry            1050 drivers/gpu/drm/radeon/r600_dpm.c 						le16_to_cpu(entry->usVddc2);
entry            1052 drivers/gpu/drm/radeon/r600_dpm.c 						le16_to_cpu(entry->usVddc3);
entry            1055 drivers/gpu/drm/radeon/r600_dpm.c 						le16_to_cpu(entry->usVddc);
entry            1057 drivers/gpu/drm/radeon/r600_dpm.c 						le32_to_cpu(entry->ulLeakageValue);
entry            1059 drivers/gpu/drm/radeon/r600_dpm.c 				entry = (ATOM_PPLIB_CAC_Leakage_Record *)
entry            1060 drivers/gpu/drm/radeon/r600_dpm.c 					((u8 *)entry + sizeof(ATOM_PPLIB_CAC_Leakage_Record));
entry            1088 drivers/gpu/drm/radeon/r600_dpm.c 			ATOM_PPLIB_VCE_Clock_Voltage_Limit_Record *entry;
entry            1101 drivers/gpu/drm/radeon/r600_dpm.c 			entry = &limits->entries[0];
entry            1106 drivers/gpu/drm/radeon/r600_dpm.c 					 (entry->ucVCEClockInfoIndex * sizeof(VCEClockInfo)));
entry            1112 drivers/gpu/drm/radeon/r600_dpm.c 					le16_to_cpu(entry->usVoltage);
entry            1113 drivers/gpu/drm/radeon/r600_dpm.c 				entry = (ATOM_PPLIB_VCE_Clock_Voltage_Limit_Record *)
entry            1114 drivers/gpu/drm/radeon/r600_dpm.c 					((u8 *)entry + sizeof(ATOM_PPLIB_VCE_Clock_Voltage_Limit_Record));
entry            1144 drivers/gpu/drm/radeon/r600_dpm.c 			ATOM_PPLIB_UVD_Clock_Voltage_Limit_Record *entry;
entry            1155 drivers/gpu/drm/radeon/r600_dpm.c 			entry = &limits->entries[0];
entry            1159 drivers/gpu/drm/radeon/r600_dpm.c 					 (entry->ucUVDClockInfoIndex * sizeof(UVDClockInfo)));
entry            1165 drivers/gpu/drm/radeon/r600_dpm.c 					le16_to_cpu(entry->usVoltage);
entry            1166 drivers/gpu/drm/radeon/r600_dpm.c 				entry = (ATOM_PPLIB_UVD_Clock_Voltage_Limit_Record *)
entry            1167 drivers/gpu/drm/radeon/r600_dpm.c 					((u8 *)entry + sizeof(ATOM_PPLIB_UVD_Clock_Voltage_Limit_Record));
entry            1176 drivers/gpu/drm/radeon/r600_dpm.c 			ATOM_PPLIB_SAMClk_Voltage_Limit_Record *entry;
entry            1187 drivers/gpu/drm/radeon/r600_dpm.c 			entry = &limits->entries[0];
entry            1190 drivers/gpu/drm/radeon/r600_dpm.c 					le16_to_cpu(entry->usSAMClockLow) | (entry->ucSAMClockHigh << 16);
entry            1192 drivers/gpu/drm/radeon/r600_dpm.c 					le16_to_cpu(entry->usVoltage);
entry            1193 drivers/gpu/drm/radeon/r600_dpm.c 				entry = (ATOM_PPLIB_SAMClk_Voltage_Limit_Record *)
entry            1194 drivers/gpu/drm/radeon/r600_dpm.c 					((u8 *)entry + sizeof(ATOM_PPLIB_SAMClk_Voltage_Limit_Record));
entry            1234 drivers/gpu/drm/radeon/r600_dpm.c 			ATOM_PPLIB_ACPClk_Voltage_Limit_Record *entry;
entry            1245 drivers/gpu/drm/radeon/r600_dpm.c 			entry = &limits->entries[0];
entry            1248 drivers/gpu/drm/radeon/r600_dpm.c 					le16_to_cpu(entry->usACPClockLow) | (entry->ucACPClockHigh << 16);
entry            1250 drivers/gpu/drm/radeon/r600_dpm.c 					le16_to_cpu(entry->usVoltage);
entry            1251 drivers/gpu/drm/radeon/r600_dpm.c 				entry = (ATOM_PPLIB_ACPClk_Voltage_Limit_Record *)
entry            1252 drivers/gpu/drm/radeon/r600_dpm.c 					((u8 *)entry + sizeof(ATOM_PPLIB_ACPClk_Voltage_Limit_Record));
entry             252 drivers/gpu/drm/radeon/radeon.h 	uint64_t	entry;
entry            1865 drivers/gpu/drm/radeon/radeon.h 				 uint64_t entry);
entry             685 drivers/gpu/drm/radeon/radeon_acpi.c 	struct acpi_bus_event *entry = (struct acpi_bus_event *)data;
entry             687 drivers/gpu/drm/radeon/radeon_acpi.c 	if (strcmp(entry->device_class, ACPI_AC_CLASS) == 0) {
entry             697 drivers/gpu/drm/radeon/radeon_acpi.c 	return radeon_atif_handler(rdev, entry);
entry              72 drivers/gpu/drm/radeon/radeon_asic.h 			    uint64_t entry);
entry             178 drivers/gpu/drm/radeon/radeon_asic.h 				     uint64_t entry);
entry             215 drivers/gpu/drm/radeon/radeon_asic.h 			 uint64_t entry);
entry             240 drivers/gpu/drm/radeon/radeon_asic.h 			 uint64_t entry);
entry            2309 drivers/gpu/drm/radeon/radeon_combios.c 	uint32_t conn_info, entry, devices;
entry            2320 drivers/gpu/drm/radeon/radeon_combios.c 			entry = conn_info + 2 + i * 2;
entry            2322 drivers/gpu/drm/radeon/radeon_combios.c 			if (!RBIOS16(entry))
entry            2325 drivers/gpu/drm/radeon/radeon_combios.c 			tmp = RBIOS16(entry);
entry             795 drivers/gpu/drm/radeon/radeon_device.c 	rdev->dummy_page.entry = radeon_gart_get_page_entry(rdev->dummy_page.addr,
entry             261 drivers/gpu/drm/radeon/radeon_gart.c 				rdev->gart.pages_entry[t] = rdev->dummy_page.entry;
entry             264 drivers/gpu/drm/radeon/radeon_gart.c 							     rdev->dummy_page.entry);
entry             367 drivers/gpu/drm/radeon/radeon_gart.c 		rdev->gart.pages_entry[i] = rdev->dummy_page.entry;
entry             552 drivers/gpu/drm/radeon/radeon_gem.c 	struct ttm_validate_buffer tv, *entry;
entry             573 drivers/gpu/drm/radeon/radeon_gem.c 	list_for_each_entry(entry, &list, head) {
entry             574 drivers/gpu/drm/radeon/radeon_gem.c 		domain = radeon_mem_type_to_domain(entry->bo->mem.mem_type);
entry             222 drivers/gpu/drm/radeon/rs400.c 	uint32_t entry;
entry             224 drivers/gpu/drm/radeon/rs400.c 	entry = (lower_32_bits(addr) & PAGE_MASK) |
entry             227 drivers/gpu/drm/radeon/rs400.c 		entry |= RS400_PTE_READABLE;
entry             229 drivers/gpu/drm/radeon/rs400.c 		entry |= RS400_PTE_WRITEABLE;
entry             231 drivers/gpu/drm/radeon/rs400.c 		entry |= RS400_PTE_UNSNOOPED;
entry             232 drivers/gpu/drm/radeon/rs400.c 	return entry;
entry             236 drivers/gpu/drm/radeon/rs400.c 			 uint64_t entry)
entry             239 drivers/gpu/drm/radeon/rs400.c 	gtt[i] = cpu_to_le32(lower_32_bits(entry));
entry             654 drivers/gpu/drm/radeon/rs600.c 			 uint64_t entry)
entry             657 drivers/gpu/drm/radeon/rs600.c 	writeq(entry, ptr + (i * 8));
entry             598 drivers/gpu/drm/radeon/rv6xx_dpm.c 					     u32 entry, u32 clock)
entry             606 drivers/gpu/drm/radeon/rv6xx_dpm.c 	rv6xx_memory_clock_entry_set_reference_divider(rdev, entry, dividers.ref_div);
entry             607 drivers/gpu/drm/radeon/rv6xx_dpm.c 	rv6xx_memory_clock_entry_set_feedback_divider(rdev, entry, dividers.fb_div);
entry             608 drivers/gpu/drm/radeon/rv6xx_dpm.c 	rv6xx_memory_clock_entry_set_post_divider(rdev, entry, dividers.post_div);
entry             611 drivers/gpu/drm/radeon/rv6xx_dpm.c 		rv6xx_memory_clock_entry_enable_post_divider(rdev, entry, true);
entry             613 drivers/gpu/drm/radeon/rv6xx_dpm.c 		rv6xx_memory_clock_entry_enable_post_divider(rdev, entry, false);
entry             702 drivers/gpu/drm/radeon/rv6xx_dpm.c 						u32 entry, u16 voltage)
entry             713 drivers/gpu/drm/radeon/rv6xx_dpm.c 	r600_voltage_control_program_voltages(rdev, entry, set_pins);
entry            5603 drivers/gpu/drm/radeon/si_dpm.c static void si_convert_mc_registers(const struct si_mc_reg_entry *entry,
entry            5611 drivers/gpu/drm/radeon/si_dpm.c 			data->value[i] = cpu_to_be32(entry->mc_data[j]);
entry             220 drivers/gpu/drm/savage/savage_bci.c 	drm_savage_buf_priv_t *entry;
entry             234 drivers/gpu/drm/savage/savage_bci.c 		entry = buf->dev_private;
entry             236 drivers/gpu/drm/savage/savage_bci.c 		SET_AGE(&entry->age, 0, 0);
entry             237 drivers/gpu/drm/savage/savage_bci.c 		entry->buf = buf;
entry             239 drivers/gpu/drm/savage/savage_bci.c 		entry->next = dev_priv->head.next;
entry             240 drivers/gpu/drm/savage/savage_bci.c 		entry->prev = &dev_priv->head;
entry             241 drivers/gpu/drm/savage/savage_bci.c 		dev_priv->head.next->prev = entry;
entry             242 drivers/gpu/drm/savage/savage_bci.c 		dev_priv->head.next = entry;
entry             284 drivers/gpu/drm/savage/savage_bci.c 	drm_savage_buf_priv_t *entry = buf->dev_private, *prev, *next;
entry             286 drivers/gpu/drm/savage/savage_bci.c 	DRM_DEBUG("age=0x%04x wrap=%d\n", entry->age.event, entry->age.wrap);
entry             288 drivers/gpu/drm/savage/savage_bci.c 	if (entry->next != NULL || entry->prev != NULL) {
entry             295 drivers/gpu/drm/savage/savage_bci.c 	prev->next = entry;
entry             296 drivers/gpu/drm/savage/savage_bci.c 	next->prev = entry;
entry             297 drivers/gpu/drm/savage/savage_bci.c 	entry->prev = prev;
entry             298 drivers/gpu/drm/savage/savage_bci.c 	entry->next = next;
entry             227 drivers/gpu/drm/sis/sis_mm.c 	struct drm_map_list *entry;
entry             230 drivers/gpu/drm/sis/sis_mm.c 	list_for_each_entry(entry, &dev->maplist, head) {
entry             231 drivers/gpu/drm/sis/sis_mm.c 		map = entry->map;
entry             318 drivers/gpu/drm/sis/sis_mm.c 	struct sis_memblock *entry, *next;
entry             336 drivers/gpu/drm/sis/sis_mm.c 	list_for_each_entry_safe(entry, next, &file_priv->obj_list,
entry             338 drivers/gpu/drm/sis/sis_mm.c 		list_del(&entry->owner_list);
entry             339 drivers/gpu/drm/sis/sis_mm.c 		if (drm_mm_node_allocated(&entry->mm_node))
entry             340 drivers/gpu/drm/sis/sis_mm.c 			drm_mm_remove_node(&entry->mm_node);
entry             343 drivers/gpu/drm/sis/sis_mm.c 			sis_free(entry->req.offset);
entry             345 drivers/gpu/drm/sis/sis_mm.c 		kfree(entry);
entry              37 drivers/gpu/drm/ttm/ttm_execbuf_util.c 					      struct ttm_validate_buffer *entry)
entry              39 drivers/gpu/drm/ttm/ttm_execbuf_util.c 	list_for_each_entry_continue_reverse(entry, list, head) {
entry              40 drivers/gpu/drm/ttm/ttm_execbuf_util.c 		struct ttm_buffer_object *bo = entry->bo;
entry              48 drivers/gpu/drm/ttm/ttm_execbuf_util.c 	struct ttm_validate_buffer *entry;
entry              50 drivers/gpu/drm/ttm/ttm_execbuf_util.c 	list_for_each_entry(entry, list, head) {
entry              51 drivers/gpu/drm/ttm/ttm_execbuf_util.c 		struct ttm_buffer_object *bo = entry->bo;
entry              59 drivers/gpu/drm/ttm/ttm_execbuf_util.c 	struct ttm_validate_buffer *entry;
entry              65 drivers/gpu/drm/ttm/ttm_execbuf_util.c 	entry = list_first_entry(list, struct ttm_validate_buffer, head);
entry              66 drivers/gpu/drm/ttm/ttm_execbuf_util.c 	glob = entry->bo->bdev->glob;
entry              69 drivers/gpu/drm/ttm/ttm_execbuf_util.c 	list_for_each_entry(entry, list, head) {
entry              70 drivers/gpu/drm/ttm/ttm_execbuf_util.c 		struct ttm_buffer_object *bo = entry->bo;
entry             100 drivers/gpu/drm/ttm/ttm_execbuf_util.c 	struct ttm_validate_buffer *entry;
entry             106 drivers/gpu/drm/ttm/ttm_execbuf_util.c 	entry = list_first_entry(list, struct ttm_validate_buffer, head);
entry             107 drivers/gpu/drm/ttm/ttm_execbuf_util.c 	glob = entry->bo->bdev->glob;
entry             112 drivers/gpu/drm/ttm/ttm_execbuf_util.c 	list_for_each_entry(entry, list, head) {
entry             113 drivers/gpu/drm/ttm/ttm_execbuf_util.c 		struct ttm_buffer_object *bo = entry->bo;
entry             122 drivers/gpu/drm/ttm/ttm_execbuf_util.c 			struct ttm_validate_buffer *safe = entry;
entry             123 drivers/gpu/drm/ttm/ttm_execbuf_util.c 			entry = list_prev_entry(entry, head);
entry             130 drivers/gpu/drm/ttm/ttm_execbuf_util.c 			if (!entry->num_shared)
entry             134 drivers/gpu/drm/ttm/ttm_execbuf_util.c 								entry->num_shared);
entry             143 drivers/gpu/drm/ttm/ttm_execbuf_util.c 		ttm_eu_backoff_reservation_reverse(list, entry);
entry             155 drivers/gpu/drm/ttm/ttm_execbuf_util.c 		if (!ret && entry->num_shared)
entry             157 drivers/gpu/drm/ttm/ttm_execbuf_util.c 								entry->num_shared);
entry             172 drivers/gpu/drm/ttm/ttm_execbuf_util.c 		list_del(&entry->head);
entry             173 drivers/gpu/drm/ttm/ttm_execbuf_util.c 		list_add(&entry->head, list);
entry             189 drivers/gpu/drm/ttm/ttm_execbuf_util.c 	struct ttm_validate_buffer *entry;
entry             201 drivers/gpu/drm/ttm/ttm_execbuf_util.c 	list_for_each_entry(entry, list, head) {
entry             202 drivers/gpu/drm/ttm/ttm_execbuf_util.c 		bo = entry->bo;
entry             203 drivers/gpu/drm/ttm/ttm_execbuf_util.c 		if (entry->num_shared)
entry              38 drivers/gpu/drm/udl/udl_drv.h 	struct list_head entry;
entry             150 drivers/gpu/drm/udl/udl_main.c 	list_add_tail(&unode->entry, &udl->urbs.list);
entry             187 drivers/gpu/drm/udl/udl_main.c 		unode = list_entry(node, struct urb_node, entry);
entry             251 drivers/gpu/drm/udl/udl_main.c 		list_add_tail(&unode->entry, &udl->urbs.list);
entry             267 drivers/gpu/drm/udl/udl_main.c 	struct list_head *entry;
entry             283 drivers/gpu/drm/udl/udl_main.c 	entry = udl->urbs.list.next;
entry             284 drivers/gpu/drm/udl/udl_main.c 	list_del_init(entry);
entry             289 drivers/gpu/drm/udl/udl_main.c 	unode = list_entry(entry, struct urb_node, entry);
entry              27 drivers/gpu/drm/vc4/vc4_debugfs.c 	struct vc4_debugfs_info_entry *entry;
entry              32 drivers/gpu/drm/vc4/vc4_debugfs.c 	list_for_each_entry(entry, &vc4->debugfs_list, link) {
entry              33 drivers/gpu/drm/vc4/vc4_debugfs.c 		int ret = drm_debugfs_create_files(&entry->info, 1,
entry              70 drivers/gpu/drm/vc4/vc4_debugfs.c 	struct vc4_debugfs_info_entry *entry =
entry              71 drivers/gpu/drm/vc4/vc4_debugfs.c 		devm_kzalloc(dev->dev, sizeof(*entry), GFP_KERNEL);
entry              73 drivers/gpu/drm/vc4/vc4_debugfs.c 	if (!entry)
entry              76 drivers/gpu/drm/vc4/vc4_debugfs.c 	entry->info.name = name;
entry              77 drivers/gpu/drm/vc4/vc4_debugfs.c 	entry->info.show = show;
entry              78 drivers/gpu/drm/vc4/vc4_debugfs.c 	entry->info.data = data;
entry              80 drivers/gpu/drm/vc4/vc4_debugfs.c 	list_add(&entry->link, &vc4->debugfs_list);
entry            1008 drivers/gpu/drm/vc4/vc4_gem.c 	list_for_each_entry_safe(cb, cb_temp, &vc4->seqno_cb_list, work.entry) {
entry            1010 drivers/gpu/drm/vc4/vc4_gem.c 			list_del_init(&cb->work.entry);
entry            1039 drivers/gpu/drm/vc4/vc4_gem.c 		list_add_tail(&cb->work.entry, &vc4->seqno_cb_list);
entry             164 drivers/gpu/drm/via/via_drv.h 	DECLARE_WAITQUEUE(entry, current);			\
entry             166 drivers/gpu/drm/via/via_drv.h 	add_wait_queue(&(queue), &entry);			\
entry             183 drivers/gpu/drm/via/via_drv.h 	remove_wait_queue(&(queue), &entry);			\
entry             214 drivers/gpu/drm/via/via_mm.c 	struct via_memblock *entry, *next;
entry             231 drivers/gpu/drm/via/via_mm.c 	list_for_each_entry_safe(entry, next, &file_priv->obj_list,
entry             233 drivers/gpu/drm/via/via_mm.c 		list_del(&entry->owner_list);
entry             234 drivers/gpu/drm/via/via_mm.c 		drm_mm_remove_node(&entry->mm_node);
entry             235 drivers/gpu/drm/via/via_mm.c 		kfree(entry);
entry             182 drivers/gpu/drm/virtio/virtgpu_vq.c 	struct virtio_gpu_vbuffer *entry, *tmp;
entry             195 drivers/gpu/drm/virtio/virtgpu_vq.c 	list_for_each_entry_safe(entry, tmp, &reclaim_list, list) {
entry             196 drivers/gpu/drm/virtio/virtgpu_vq.c 		resp = (struct virtio_gpu_ctrl_hdr *)entry->resp_buf;
entry             203 drivers/gpu/drm/virtio/virtgpu_vq.c 				cmd = (struct virtio_gpu_ctrl_hdr *)entry->buf;
entry             220 drivers/gpu/drm/virtio/virtgpu_vq.c 		if (entry->resp_cb)
entry             221 drivers/gpu/drm/virtio/virtgpu_vq.c 			entry->resp_cb(vgdev, entry);
entry             223 drivers/gpu/drm/virtio/virtgpu_vq.c 		list_del(&entry->list);
entry             224 drivers/gpu/drm/virtio/virtgpu_vq.c 		free_vbuf(vgdev, entry);
entry             238 drivers/gpu/drm/virtio/virtgpu_vq.c 	struct virtio_gpu_vbuffer *entry, *tmp;
entry             248 drivers/gpu/drm/virtio/virtgpu_vq.c 	list_for_each_entry_safe(entry, tmp, &reclaim_list, list) {
entry             249 drivers/gpu/drm/virtio/virtgpu_vq.c 		list_del(&entry->list);
entry             250 drivers/gpu/drm/virtio/virtgpu_vq.c 		free_vbuf(vgdev, entry);
entry             353 drivers/gpu/drm/vmwgfx/vmwgfx_binding.c 	struct vmw_ctx_bindinfo *entry, *next;
entry             356 drivers/gpu/drm/vmwgfx/vmwgfx_binding.c 	list_for_each_entry_safe(entry, next, &cbs->list, ctx_list)
entry             357 drivers/gpu/drm/vmwgfx/vmwgfx_binding.c 		vmw_binding_drop(entry);
entry             371 drivers/gpu/drm/vmwgfx/vmwgfx_binding.c 	struct vmw_ctx_bindinfo *entry;
entry             373 drivers/gpu/drm/vmwgfx/vmwgfx_binding.c 	list_for_each_entry(entry, &cbs->list, ctx_list) {
entry             374 drivers/gpu/drm/vmwgfx/vmwgfx_binding.c 		if (!entry->scrubbed) {
entry             375 drivers/gpu/drm/vmwgfx/vmwgfx_binding.c 			(void) vmw_binding_infos[entry->bt].scrub_func
entry             376 drivers/gpu/drm/vmwgfx/vmwgfx_binding.c 				(entry, false);
entry             377 drivers/gpu/drm/vmwgfx/vmwgfx_binding.c 			entry->scrubbed = true;
entry             395 drivers/gpu/drm/vmwgfx/vmwgfx_binding.c 	struct vmw_ctx_bindinfo *entry, *next;
entry             398 drivers/gpu/drm/vmwgfx/vmwgfx_binding.c 	list_for_each_entry_safe(entry, next, head, res_list)
entry             399 drivers/gpu/drm/vmwgfx/vmwgfx_binding.c 		vmw_binding_drop(entry);
entry             413 drivers/gpu/drm/vmwgfx/vmwgfx_binding.c 	struct vmw_ctx_bindinfo *entry;
entry             415 drivers/gpu/drm/vmwgfx/vmwgfx_binding.c 	list_for_each_entry(entry, head, res_list) {
entry             416 drivers/gpu/drm/vmwgfx/vmwgfx_binding.c 		if (!entry->scrubbed) {
entry             417 drivers/gpu/drm/vmwgfx/vmwgfx_binding.c 			(void) vmw_binding_infos[entry->bt].scrub_func
entry             418 drivers/gpu/drm/vmwgfx/vmwgfx_binding.c 				(entry, false);
entry             419 drivers/gpu/drm/vmwgfx/vmwgfx_binding.c 			entry->scrubbed = true;
entry             423 drivers/gpu/drm/vmwgfx/vmwgfx_binding.c 	list_for_each_entry(entry, head, res_list) {
entry             425 drivers/gpu/drm/vmwgfx/vmwgfx_binding.c 			vmw_context_binding_state(entry->ctx);
entry             447 drivers/gpu/drm/vmwgfx/vmwgfx_binding.c 	struct vmw_ctx_bindinfo *entry, *next;
entry             449 drivers/gpu/drm/vmwgfx/vmwgfx_binding.c 	list_for_each_entry_safe(entry, next, &from->list, ctx_list) {
entry             450 drivers/gpu/drm/vmwgfx/vmwgfx_binding.c 		vmw_binding_transfer(to, from, entry);
entry             451 drivers/gpu/drm/vmwgfx/vmwgfx_binding.c 		vmw_binding_drop(entry);
entry             465 drivers/gpu/drm/vmwgfx/vmwgfx_binding.c 	struct vmw_ctx_bindinfo *entry;
entry             468 drivers/gpu/drm/vmwgfx/vmwgfx_binding.c 	list_for_each_entry(entry, &cbs->list, ctx_list) {
entry             469 drivers/gpu/drm/vmwgfx/vmwgfx_binding.c 		if (likely(!entry->scrubbed))
entry             472 drivers/gpu/drm/vmwgfx/vmwgfx_binding.c 		if ((entry->res == NULL || entry->res->id ==
entry             476 drivers/gpu/drm/vmwgfx/vmwgfx_binding.c 		ret = vmw_binding_infos[entry->bt].scrub_func(entry, true);
entry             480 drivers/gpu/drm/vmwgfx/vmwgfx_binding.c 		entry->scrubbed = false;
entry            1238 drivers/gpu/drm/vmwgfx/vmwgfx_binding.c 	struct vmw_ctx_bindinfo *entry, *next;
entry            1240 drivers/gpu/drm/vmwgfx/vmwgfx_binding.c 	list_for_each_entry_safe(entry, next, &cbs->list, ctx_list)
entry            1241 drivers/gpu/drm/vmwgfx/vmwgfx_binding.c 		vmw_binding_drop(entry);
entry             345 drivers/gpu/drm/vmwgfx/vmwgfx_cmdbuf.c 		struct vmw_cmdbuf_header *entry;
entry             348 drivers/gpu/drm/vmwgfx/vmwgfx_cmdbuf.c 		entry = list_first_entry(&ctx->submitted,
entry             352 drivers/gpu/drm/vmwgfx/vmwgfx_cmdbuf.c 		status = vmw_cmdbuf_header_submit(entry);
entry             356 drivers/gpu/drm/vmwgfx/vmwgfx_cmdbuf.c 			entry->cb_header->status = SVGA_CB_STATUS_NONE;
entry             360 drivers/gpu/drm/vmwgfx/vmwgfx_cmdbuf.c 		list_del(&entry->list);
entry             361 drivers/gpu/drm/vmwgfx/vmwgfx_cmdbuf.c 		list_add_tail(&entry->list, &ctx->hw_submitted);
entry             381 drivers/gpu/drm/vmwgfx/vmwgfx_cmdbuf.c 	struct vmw_cmdbuf_header *entry, *next;
entry             385 drivers/gpu/drm/vmwgfx/vmwgfx_cmdbuf.c 	list_for_each_entry_safe(entry, next, &ctx->hw_submitted, list) {
entry             386 drivers/gpu/drm/vmwgfx/vmwgfx_cmdbuf.c 		SVGACBStatus status = entry->cb_header->status;
entry             391 drivers/gpu/drm/vmwgfx/vmwgfx_cmdbuf.c 		list_del(&entry->list);
entry             396 drivers/gpu/drm/vmwgfx/vmwgfx_cmdbuf.c 			__vmw_cmdbuf_header_free(entry);
entry             400 drivers/gpu/drm/vmwgfx/vmwgfx_cmdbuf.c 			entry->cb_header->status = SVGA_CB_STATUS_NONE;
entry             401 drivers/gpu/drm/vmwgfx/vmwgfx_cmdbuf.c 			list_add_tail(&entry->list, &man->error);
entry             405 drivers/gpu/drm/vmwgfx/vmwgfx_cmdbuf.c 			entry->cb_header->status = SVGA_CB_STATUS_NONE;
entry             406 drivers/gpu/drm/vmwgfx/vmwgfx_cmdbuf.c 			list_add_tail(&entry->list, &ctx->preempted);
entry             410 drivers/gpu/drm/vmwgfx/vmwgfx_cmdbuf.c 			__vmw_cmdbuf_header_free(entry);
entry             414 drivers/gpu/drm/vmwgfx/vmwgfx_cmdbuf.c 			__vmw_cmdbuf_header_free(entry);
entry             516 drivers/gpu/drm/vmwgfx/vmwgfx_cmdbuf.c 	struct vmw_cmdbuf_header *entry, *next;
entry             529 drivers/gpu/drm/vmwgfx/vmwgfx_cmdbuf.c 	list_for_each_entry_safe(entry, next, &man->error, list) {
entry             530 drivers/gpu/drm/vmwgfx/vmwgfx_cmdbuf.c 		SVGACBHeader *cb_hdr = entry->cb_header;
entry             532 drivers/gpu/drm/vmwgfx/vmwgfx_cmdbuf.c 			(entry->cmd + cb_hdr->errorOffset);
entry             536 drivers/gpu/drm/vmwgfx/vmwgfx_cmdbuf.c 		list_del_init(&entry->list);
entry             543 drivers/gpu/drm/vmwgfx/vmwgfx_cmdbuf.c 			__vmw_cmdbuf_header_free(entry);
entry             558 drivers/gpu/drm/vmwgfx/vmwgfx_cmdbuf.c 			__vmw_cmdbuf_header_free(entry);
entry             568 drivers/gpu/drm/vmwgfx/vmwgfx_cmdbuf.c 		entry->cmd += new_start_offset;
entry             573 drivers/gpu/drm/vmwgfx/vmwgfx_cmdbuf.c 		list_add_tail(&entry->list, &restart_head[entry->cb_context]);
entry             105 drivers/gpu/drm/vmwgfx/vmwgfx_cmdbuf_res.c 				struct vmw_cmdbuf_res *entry)
entry             107 drivers/gpu/drm/vmwgfx/vmwgfx_cmdbuf_res.c 	list_del(&entry->head);
entry             108 drivers/gpu/drm/vmwgfx/vmwgfx_cmdbuf_res.c 	WARN_ON(drm_ht_remove_item(&man->resources, &entry->hash));
entry             109 drivers/gpu/drm/vmwgfx/vmwgfx_cmdbuf_res.c 	vmw_resource_unreference(&entry->res);
entry             110 drivers/gpu/drm/vmwgfx/vmwgfx_cmdbuf_res.c 	kfree(entry);
entry             125 drivers/gpu/drm/vmwgfx/vmwgfx_cmdbuf_res.c 	struct vmw_cmdbuf_res *entry, *next;
entry             127 drivers/gpu/drm/vmwgfx/vmwgfx_cmdbuf_res.c 	list_for_each_entry_safe(entry, next, list, head) {
entry             128 drivers/gpu/drm/vmwgfx/vmwgfx_cmdbuf_res.c 		list_del(&entry->head);
entry             129 drivers/gpu/drm/vmwgfx/vmwgfx_cmdbuf_res.c 		if (entry->res->func->commit_notify)
entry             130 drivers/gpu/drm/vmwgfx/vmwgfx_cmdbuf_res.c 			entry->res->func->commit_notify(entry->res,
entry             131 drivers/gpu/drm/vmwgfx/vmwgfx_cmdbuf_res.c 							entry->state);
entry             132 drivers/gpu/drm/vmwgfx/vmwgfx_cmdbuf_res.c 		switch (entry->state) {
entry             134 drivers/gpu/drm/vmwgfx/vmwgfx_cmdbuf_res.c 			entry->state = VMW_CMDBUF_RES_COMMITTED;
entry             135 drivers/gpu/drm/vmwgfx/vmwgfx_cmdbuf_res.c 			list_add_tail(&entry->head, &entry->man->list);
entry             138 drivers/gpu/drm/vmwgfx/vmwgfx_cmdbuf_res.c 			vmw_resource_unreference(&entry->res);
entry             139 drivers/gpu/drm/vmwgfx/vmwgfx_cmdbuf_res.c 			kfree(entry);
entry             162 drivers/gpu/drm/vmwgfx/vmwgfx_cmdbuf_res.c 	struct vmw_cmdbuf_res *entry, *next;
entry             165 drivers/gpu/drm/vmwgfx/vmwgfx_cmdbuf_res.c 	list_for_each_entry_safe(entry, next, list, head) {
entry             166 drivers/gpu/drm/vmwgfx/vmwgfx_cmdbuf_res.c 		switch (entry->state) {
entry             168 drivers/gpu/drm/vmwgfx/vmwgfx_cmdbuf_res.c 			vmw_cmdbuf_res_free(entry->man, entry);
entry             171 drivers/gpu/drm/vmwgfx/vmwgfx_cmdbuf_res.c 			ret = drm_ht_insert_item(&entry->man->resources,
entry             172 drivers/gpu/drm/vmwgfx/vmwgfx_cmdbuf_res.c 						 &entry->hash);
entry             173 drivers/gpu/drm/vmwgfx/vmwgfx_cmdbuf_res.c 			list_del(&entry->head);
entry             174 drivers/gpu/drm/vmwgfx/vmwgfx_cmdbuf_res.c 			list_add_tail(&entry->head, &entry->man->list);
entry             175 drivers/gpu/drm/vmwgfx/vmwgfx_cmdbuf_res.c 			entry->state = VMW_CMDBUF_RES_COMMITTED;
entry             248 drivers/gpu/drm/vmwgfx/vmwgfx_cmdbuf_res.c 	struct vmw_cmdbuf_res *entry;
entry             257 drivers/gpu/drm/vmwgfx/vmwgfx_cmdbuf_res.c 	entry = drm_hash_entry(hash, struct vmw_cmdbuf_res, hash);
entry             259 drivers/gpu/drm/vmwgfx/vmwgfx_cmdbuf_res.c 	switch (entry->state) {
entry             261 drivers/gpu/drm/vmwgfx/vmwgfx_cmdbuf_res.c 		vmw_cmdbuf_res_free(man, entry);
entry             265 drivers/gpu/drm/vmwgfx/vmwgfx_cmdbuf_res.c 		(void) drm_ht_remove_item(&man->resources, &entry->hash);
entry             266 drivers/gpu/drm/vmwgfx/vmwgfx_cmdbuf_res.c 		list_del(&entry->head);
entry             267 drivers/gpu/drm/vmwgfx/vmwgfx_cmdbuf_res.c 		entry->state = VMW_CMDBUF_RES_DEL;
entry             268 drivers/gpu/drm/vmwgfx/vmwgfx_cmdbuf_res.c 		list_add_tail(&entry->head, list);
entry             269 drivers/gpu/drm/vmwgfx/vmwgfx_cmdbuf_res.c 		*res_p = entry->res;
entry             320 drivers/gpu/drm/vmwgfx/vmwgfx_cmdbuf_res.c 	struct vmw_cmdbuf_res *entry, *next;
entry             322 drivers/gpu/drm/vmwgfx/vmwgfx_cmdbuf_res.c 	list_for_each_entry_safe(entry, next, &man->list, head)
entry             323 drivers/gpu/drm/vmwgfx/vmwgfx_cmdbuf_res.c 		vmw_cmdbuf_res_free(man, entry);
entry             172 drivers/gpu/drm/vmwgfx/vmwgfx_execbuf.c 	struct vmw_ctx_validation_info *entry;
entry             174 drivers/gpu/drm/vmwgfx/vmwgfx_execbuf.c 	list_for_each_entry(entry, &sw_context->ctx_list, head) {
entry             176 drivers/gpu/drm/vmwgfx/vmwgfx_execbuf.c 			vmw_binding_state_commit(entry->cur, entry->staged);
entry             178 drivers/gpu/drm/vmwgfx/vmwgfx_execbuf.c 		if (entry->staged != sw_context->staged_bindings)
entry             179 drivers/gpu/drm/vmwgfx/vmwgfx_execbuf.c 			vmw_binding_state_free(entry->staged);
entry             458 drivers/gpu/drm/vmwgfx/vmwgfx_execbuf.c 	struct vmw_ctx_bindinfo *entry;
entry             481 drivers/gpu/drm/vmwgfx/vmwgfx_execbuf.c 	list_for_each_entry(entry, binding_list, ctx_list) {
entry             482 drivers/gpu/drm/vmwgfx/vmwgfx_execbuf.c 		if (vmw_res_type(entry->res) == vmw_res_view)
entry             483 drivers/gpu/drm/vmwgfx/vmwgfx_execbuf.c 			ret = vmw_view_res_val_add(sw_context, entry->res);
entry             486 drivers/gpu/drm/vmwgfx/vmwgfx_execbuf.c 				(sw_context, entry->res,
entry             487 drivers/gpu/drm/vmwgfx/vmwgfx_execbuf.c 				 vmw_binding_dirtying(entry->bt));
entry            3162 drivers/gpu/drm/vmwgfx/vmwgfx_execbuf.c 		const struct vmw_cmd_entry *entry;
entry            3170 drivers/gpu/drm/vmwgfx/vmwgfx_execbuf.c 		entry = &vmw_cmd_entries[cmd_id];
entry            3171 drivers/gpu/drm/vmwgfx/vmwgfx_execbuf.c 		*cmd = entry->cmd_name;
entry            3209 drivers/gpu/drm/vmwgfx/vmwgfx_execbuf.c 	const struct vmw_cmd_entry *entry;
entry            3228 drivers/gpu/drm/vmwgfx/vmwgfx_execbuf.c 	entry = &vmw_cmd_entries[cmd_id];
entry            3229 drivers/gpu/drm/vmwgfx/vmwgfx_execbuf.c 	if (unlikely(!entry->func))
entry            3232 drivers/gpu/drm/vmwgfx/vmwgfx_execbuf.c 	if (unlikely(!entry->user_allow && !sw_context->kernel))
entry            3235 drivers/gpu/drm/vmwgfx/vmwgfx_execbuf.c 	if (unlikely(entry->gb_disable && gb))
entry            3238 drivers/gpu/drm/vmwgfx/vmwgfx_execbuf.c 	if (unlikely(entry->gb_enable && !gb))
entry            3241 drivers/gpu/drm/vmwgfx/vmwgfx_execbuf.c 	ret = entry->func(dev_priv, sw_context, header);
entry              81 drivers/gpu/drm/vmwgfx/vmwgfx_ldu.c 	struct vmw_legacy_display_unit *entry;
entry              91 drivers/gpu/drm/vmwgfx/vmwgfx_ldu.c 		list_for_each_entry(entry, &lds->active, active) {
entry              92 drivers/gpu/drm/vmwgfx/vmwgfx_ldu.c 			crtc = &entry->base.crtc;
entry             100 drivers/gpu/drm/vmwgfx/vmwgfx_ldu.c 		fb = entry->base.crtc.primary->state->fb;
entry             108 drivers/gpu/drm/vmwgfx/vmwgfx_ldu.c 		entry = list_entry(lds->active.next, typeof(*entry), active);
entry             109 drivers/gpu/drm/vmwgfx/vmwgfx_ldu.c 		fb = entry->base.crtc.primary->state->fb;
entry             120 drivers/gpu/drm/vmwgfx/vmwgfx_ldu.c 	list_for_each_entry(entry, &lds->active, active) {
entry             121 drivers/gpu/drm/vmwgfx/vmwgfx_ldu.c 		crtc = &entry->base.crtc;
entry             165 drivers/gpu/drm/vmwgfx/vmwgfx_ldu.c 	struct vmw_legacy_display_unit *entry;
entry             182 drivers/gpu/drm/vmwgfx/vmwgfx_ldu.c 	list_for_each_entry(entry, &ld->active, active) {
entry             183 drivers/gpu/drm/vmwgfx/vmwgfx_ldu.c 		if (entry->base.unit > ldu->base.unit)
entry             186 drivers/gpu/drm/vmwgfx/vmwgfx_ldu.c 		at = &entry->active;
entry             550 drivers/gpu/drm/vmwgfx/vmwgfx_shader.c 	struct vmw_dx_shader *entry, *next;
entry             554 drivers/gpu/drm/vmwgfx/vmwgfx_shader.c 	list_for_each_entry_safe(entry, next, list, cotable_head) {
entry             555 drivers/gpu/drm/vmwgfx/vmwgfx_shader.c 		WARN_ON(vmw_dx_shader_scrub(&entry->res));
entry             557 drivers/gpu/drm/vmwgfx/vmwgfx_shader.c 			entry->committed = false;
entry             436 drivers/gpu/drm/vmwgfx/vmwgfx_so.c 	struct vmw_view *entry, *next;
entry             440 drivers/gpu/drm/vmwgfx/vmwgfx_so.c 	list_for_each_entry_safe(entry, next, list, cotable_head)
entry             441 drivers/gpu/drm/vmwgfx/vmwgfx_so.c 		WARN_ON(vmw_view_destroy(&entry->res));
entry             456 drivers/gpu/drm/vmwgfx/vmwgfx_so.c 	struct vmw_view *entry, *next;
entry             460 drivers/gpu/drm/vmwgfx/vmwgfx_so.c 	list_for_each_entry_safe(entry, next, list, srf_head)
entry             461 drivers/gpu/drm/vmwgfx/vmwgfx_so.c 		WARN_ON(vmw_view_destroy(&entry->res));
entry             149 drivers/gpu/drm/vmwgfx/vmwgfx_validation.c 	struct page *entry, *next;
entry             151 drivers/gpu/drm/vmwgfx/vmwgfx_validation.c 	list_for_each_entry_safe(entry, next, &ctx->page_list, lru) {
entry             152 drivers/gpu/drm/vmwgfx/vmwgfx_validation.c 		list_del_init(&entry->lru);
entry             153 drivers/gpu/drm/vmwgfx/vmwgfx_validation.c 		__free_page(entry);
entry             188 drivers/gpu/drm/vmwgfx/vmwgfx_validation.c 		struct  vmw_validation_bo_node *entry;
entry             190 drivers/gpu/drm/vmwgfx/vmwgfx_validation.c 		list_for_each_entry(entry, &ctx->bo_list, base.head) {
entry             191 drivers/gpu/drm/vmwgfx/vmwgfx_validation.c 			if (entry->base.bo == &vbo->base) {
entry             192 drivers/gpu/drm/vmwgfx/vmwgfx_validation.c 				bo_node = entry;
entry             225 drivers/gpu/drm/vmwgfx/vmwgfx_validation.c 		struct  vmw_validation_res_node *entry;
entry             227 drivers/gpu/drm/vmwgfx/vmwgfx_validation.c 		list_for_each_entry(entry, &ctx->resource_ctx_list, head) {
entry             228 drivers/gpu/drm/vmwgfx/vmwgfx_validation.c 			if (entry->res == res) {
entry             229 drivers/gpu/drm/vmwgfx/vmwgfx_validation.c 				res_node = entry;
entry             234 drivers/gpu/drm/vmwgfx/vmwgfx_validation.c 		list_for_each_entry(entry, &ctx->resource_list, head) {
entry             235 drivers/gpu/drm/vmwgfx/vmwgfx_validation.c 			if (entry->res == res) {
entry             236 drivers/gpu/drm/vmwgfx/vmwgfx_validation.c 				res_node = entry;
entry             561 drivers/gpu/drm/vmwgfx/vmwgfx_validation.c 	struct vmw_validation_bo_node *entry;
entry             564 drivers/gpu/drm/vmwgfx/vmwgfx_validation.c 	list_for_each_entry(entry, &ctx->bo_list, base.head) {
entry             565 drivers/gpu/drm/vmwgfx/vmwgfx_validation.c 		if (entry->cpu_blit) {
entry             571 drivers/gpu/drm/vmwgfx/vmwgfx_validation.c 			ret = ttm_bo_validate(entry->base.bo,
entry             575 drivers/gpu/drm/vmwgfx/vmwgfx_validation.c 			(entry->base.bo, intr, entry->as_mob);
entry             639 drivers/gpu/drm/vmwgfx/vmwgfx_validation.c 	struct vmw_validation_bo_node *entry;
entry             645 drivers/gpu/drm/vmwgfx/vmwgfx_validation.c 	list_for_each_entry(entry, &ctx->bo_list, base.head)
entry             646 drivers/gpu/drm/vmwgfx/vmwgfx_validation.c 		(void) drm_ht_remove_item(ctx->ht, &entry->hash);
entry             667 drivers/gpu/drm/vmwgfx/vmwgfx_validation.c 	struct vmw_validation_bo_node *entry;
entry             670 drivers/gpu/drm/vmwgfx/vmwgfx_validation.c 	list_for_each_entry(entry, &ctx->bo_list, base.head) {
entry             671 drivers/gpu/drm/vmwgfx/vmwgfx_validation.c 		ttm_bo_put(entry->base.bo);
entry             672 drivers/gpu/drm/vmwgfx/vmwgfx_validation.c 		entry->base.bo = NULL;
entry             174 drivers/gpu/host1x/dev.c 		const struct host1x_sid_entry *entry = &info->sid_table[i];
entry             176 drivers/gpu/host1x/dev.c 		host1x_hypervisor_writel(host, entry->offset, entry->base);
entry             177 drivers/gpu/host1x/dev.c 		host1x_hypervisor_writel(host, entry->limit, entry->base + 4);
entry             303 drivers/hid/hid-lg4ff.c 	struct lg4ff_device_entry *entry = drv_data->device_props;
entry             306 drivers/hid/hid-lg4ff.c 	if (!entry) {
entry             311 drivers/hid/hid-lg4ff.c 	switch (entry->wdata.product_id) {
entry             315 drivers/hid/hid-lg4ff.c 			new_value = lg4ff_adjust_dfp_x_axis(value, entry->wdata.range);
entry             330 drivers/hid/hid-lg4ff.c 	struct lg4ff_device_entry *entry = drv_data->device_props;
entry             332 drivers/hid/hid-lg4ff.c 	if (!entry)
entry             336 drivers/hid/hid-lg4ff.c 	if (entry->wdata.combine) {
entry             337 drivers/hid/hid-lg4ff.c 		switch (entry->wdata.product_id) {
entry             409 drivers/hid/hid-lg4ff.c 	struct lg4ff_device_entry *entry;
entry             421 drivers/hid/hid-lg4ff.c 	entry = drv_data->device_props;
entry             422 drivers/hid/hid-lg4ff.c 	if (!entry) {
entry             426 drivers/hid/hid-lg4ff.c 	value = entry->report->field[0]->value;
entry             435 drivers/hid/hid-lg4ff.c 		spin_lock_irqsave(&entry->report_lock, flags);
entry             446 drivers/hid/hid-lg4ff.c 			hid_hw_request(hid, entry->report, HID_REQ_SET_REPORT);
entry             447 drivers/hid/hid-lg4ff.c 			spin_unlock_irqrestore(&entry->report_lock, flags);
entry             459 drivers/hid/hid-lg4ff.c 		hid_hw_request(hid, entry->report, HID_REQ_SET_REPORT);
entry             460 drivers/hid/hid-lg4ff.c 		spin_unlock_irqrestore(&entry->report_lock, flags);
entry             473 drivers/hid/hid-lg4ff.c 	struct lg4ff_device_entry *entry;
entry             483 drivers/hid/hid-lg4ff.c 	entry = drv_data->device_props;
entry             484 drivers/hid/hid-lg4ff.c 	if (!entry) {
entry             488 drivers/hid/hid-lg4ff.c 	value = entry->report->field[0]->value;
entry             491 drivers/hid/hid-lg4ff.c 	spin_lock_irqsave(&entry->report_lock, flags);
entry             501 drivers/hid/hid-lg4ff.c 		hid_hw_request(hid, entry->report, HID_REQ_SET_REPORT);
entry             502 drivers/hid/hid-lg4ff.c 		spin_unlock_irqrestore(&entry->report_lock, flags);
entry             515 drivers/hid/hid-lg4ff.c 	switch (entry->wdata.product_id) {
entry             532 drivers/hid/hid-lg4ff.c 	hid_hw_request(hid, entry->report, HID_REQ_SET_REPORT);
entry             543 drivers/hid/hid-lg4ff.c 	hid_hw_request(hid, entry->report, HID_REQ_SET_REPORT);
entry             544 drivers/hid/hid-lg4ff.c 	spin_unlock_irqrestore(&entry->report_lock, flags);
entry             551 drivers/hid/hid-lg4ff.c 	struct lg4ff_device_entry *entry;
entry             563 drivers/hid/hid-lg4ff.c 	entry = drv_data->device_props;
entry             564 drivers/hid/hid-lg4ff.c 	if (!entry) {
entry             568 drivers/hid/hid-lg4ff.c 	value = entry->report->field[0]->value;
entry             570 drivers/hid/hid-lg4ff.c 	spin_lock_irqsave(&entry->report_lock, flags);
entry             579 drivers/hid/hid-lg4ff.c 	hid_hw_request(hid, entry->report, HID_REQ_SET_REPORT);
entry             580 drivers/hid/hid-lg4ff.c 	spin_unlock_irqrestore(&entry->report_lock, flags);
entry             586 drivers/hid/hid-lg4ff.c 	struct lg4ff_device_entry *entry;
entry             597 drivers/hid/hid-lg4ff.c 	entry = drv_data->device_props;
entry             598 drivers/hid/hid-lg4ff.c 	if (!entry) {
entry             602 drivers/hid/hid-lg4ff.c 	value = entry->report->field[0]->value;
entry             605 drivers/hid/hid-lg4ff.c 	spin_lock_irqsave(&entry->report_lock, flags);
entry             614 drivers/hid/hid-lg4ff.c 	hid_hw_request(hid, entry->report, HID_REQ_SET_REPORT);
entry             615 drivers/hid/hid-lg4ff.c 	spin_unlock_irqrestore(&entry->report_lock, flags);
entry             621 drivers/hid/hid-lg4ff.c 	struct lg4ff_device_entry *entry;
entry             633 drivers/hid/hid-lg4ff.c 	entry = drv_data->device_props;
entry             634 drivers/hid/hid-lg4ff.c 	if (!entry) {
entry             638 drivers/hid/hid-lg4ff.c 	value = entry->report->field[0]->value;
entry             642 drivers/hid/hid-lg4ff.c 	spin_lock_irqsave(&entry->report_lock, flags);
entry             658 drivers/hid/hid-lg4ff.c 	hid_hw_request(hid, entry->report, HID_REQ_SET_REPORT);
entry             670 drivers/hid/hid-lg4ff.c 		hid_hw_request(hid, entry->report, HID_REQ_SET_REPORT);
entry             671 drivers/hid/hid-lg4ff.c 		spin_unlock_irqrestore(&entry->report_lock, flags);
entry             685 drivers/hid/hid-lg4ff.c 	hid_hw_request(hid, entry->report, HID_REQ_SET_REPORT);
entry             686 drivers/hid/hid-lg4ff.c 	spin_unlock_irqrestore(&entry->report_lock, flags);
entry             765 drivers/hid/hid-lg4ff.c 	struct lg4ff_device_entry *entry;
entry             777 drivers/hid/hid-lg4ff.c 	entry = drv_data->device_props;
entry             778 drivers/hid/hid-lg4ff.c 	if (!entry) {
entry             782 drivers/hid/hid-lg4ff.c 	value = entry->report->field[0]->value;
entry             784 drivers/hid/hid-lg4ff.c 	spin_lock_irqsave(&entry->report_lock, flags);
entry             791 drivers/hid/hid-lg4ff.c 		hid_hw_request(hid, entry->report, HID_REQ_SET_REPORT);
entry             793 drivers/hid/hid-lg4ff.c 	spin_unlock_irqrestore(&entry->report_lock, flags);
entry             801 drivers/hid/hid-lg4ff.c 	struct lg4ff_device_entry *entry;
entry             812 drivers/hid/hid-lg4ff.c 	entry = drv_data->device_props;
entry             813 drivers/hid/hid-lg4ff.c 	if (!entry) {
entry             818 drivers/hid/hid-lg4ff.c 	if (!entry->wdata.real_name) {
entry             824 drivers/hid/hid-lg4ff.c 		if (entry->wdata.alternate_modes & BIT(i)) {
entry             828 drivers/hid/hid-lg4ff.c 					   !lg4ff_alternate_modes[i].product_id ? entry->wdata.real_name : lg4ff_alternate_modes[i].name);
entry             833 drivers/hid/hid-lg4ff.c 			if (lg4ff_alternate_modes[i].product_id == entry->wdata.product_id ||
entry             834 drivers/hid/hid-lg4ff.c 			    (lg4ff_alternate_modes[i].product_id == 0 && entry->wdata.product_id == entry->wdata.real_product_id))
entry             850 drivers/hid/hid-lg4ff.c 	struct lg4ff_device_entry *entry;
entry             863 drivers/hid/hid-lg4ff.c 	entry = drv_data->device_props;
entry             864 drivers/hid/hid-lg4ff.c 	if (!entry) {
entry             887 drivers/hid/hid-lg4ff.c 		if (entry->wdata.alternate_modes & BIT(i)) {
entry             890 drivers/hid/hid-lg4ff.c 					target_product_id = entry->wdata.real_product_id;
entry             905 drivers/hid/hid-lg4ff.c 	if (target_product_id == entry->wdata.product_id) /* Nothing to do */
entry             911 drivers/hid/hid-lg4ff.c 			 entry->wdata.real_name);
entry             916 drivers/hid/hid-lg4ff.c 	if ((entry->wdata.real_product_id == USB_DEVICE_ID_LOGITECH_DFP_WHEEL || entry->wdata.real_product_id == USB_DEVICE_ID_LOGITECH_G25_WHEEL) &&
entry             917 drivers/hid/hid-lg4ff.c 	    entry->wdata.product_id > target_product_id) {
entry             918 drivers/hid/hid-lg4ff.c 		hid_info(hid, "\"%s\" cannot be switched back into \"%s\" mode\n", entry->wdata.real_name, lg4ff_alternate_modes[i].name);
entry             922 drivers/hid/hid-lg4ff.c 	s = lg4ff_get_mode_switch_command(entry->wdata.real_product_id, target_product_id);
entry             937 drivers/hid/hid-lg4ff.c 	struct lg4ff_device_entry *entry;
entry             947 drivers/hid/hid-lg4ff.c 	entry = drv_data->device_props;
entry             948 drivers/hid/hid-lg4ff.c 	if (!entry) {
entry             953 drivers/hid/hid-lg4ff.c 	count = scnprintf(buf, PAGE_SIZE, "%u\n", entry->wdata.combine);
entry             961 drivers/hid/hid-lg4ff.c 	struct lg4ff_device_entry *entry;
entry             971 drivers/hid/hid-lg4ff.c 	entry = drv_data->device_props;
entry             972 drivers/hid/hid-lg4ff.c 	if (!entry) {
entry             980 drivers/hid/hid-lg4ff.c 	entry->wdata.combine = combine;
entry             990 drivers/hid/hid-lg4ff.c 	struct lg4ff_device_entry *entry;
entry            1000 drivers/hid/hid-lg4ff.c 	entry = drv_data->device_props;
entry            1001 drivers/hid/hid-lg4ff.c 	if (!entry) {
entry            1006 drivers/hid/hid-lg4ff.c 	count = scnprintf(buf, PAGE_SIZE, "%u\n", entry->wdata.range);
entry            1016 drivers/hid/hid-lg4ff.c 	struct lg4ff_device_entry *entry;
entry            1026 drivers/hid/hid-lg4ff.c 	entry = drv_data->device_props;
entry            1027 drivers/hid/hid-lg4ff.c 	if (!entry) {
entry            1033 drivers/hid/hid-lg4ff.c 		range = entry->wdata.max_range;
entry            1037 drivers/hid/hid-lg4ff.c 	if (entry->wdata.set_range && range >= entry->wdata.min_range && range <= entry->wdata.max_range) {
entry            1038 drivers/hid/hid-lg4ff.c 		entry->wdata.set_range(hid, range);
entry            1039 drivers/hid/hid-lg4ff.c 		entry->wdata.range = range;
entry            1049 drivers/hid/hid-lg4ff.c 	struct lg4ff_device_entry *entry;
entry            1059 drivers/hid/hid-lg4ff.c 	entry = drv_data->device_props;
entry            1060 drivers/hid/hid-lg4ff.c 	if (!entry) {
entry            1065 drivers/hid/hid-lg4ff.c 	if (!entry->wdata.real_tag || !entry->wdata.real_name) {
entry            1070 drivers/hid/hid-lg4ff.c 	count = scnprintf(buf, PAGE_SIZE, "%s: %s\n", entry->wdata.real_tag, entry->wdata.real_name);
entry            1085 drivers/hid/hid-lg4ff.c 	struct lg4ff_device_entry *entry;
entry            1095 drivers/hid/hid-lg4ff.c 	entry = drv_data->device_props;
entry            1096 drivers/hid/hid-lg4ff.c 	if (!entry) {
entry            1100 drivers/hid/hid-lg4ff.c 	value = entry->report->field[0]->value;
entry            1102 drivers/hid/hid-lg4ff.c 	spin_lock_irqsave(&entry->report_lock, flags);
entry            1110 drivers/hid/hid-lg4ff.c 	hid_hw_request(hid, entry->report, HID_REQ_SET_REPORT);
entry            1111 drivers/hid/hid-lg4ff.c 	spin_unlock_irqrestore(&entry->report_lock, flags);
entry            1120 drivers/hid/hid-lg4ff.c 	struct lg4ff_device_entry *entry;
entry            1128 drivers/hid/hid-lg4ff.c 	entry = drv_data->device_props;
entry            1130 drivers/hid/hid-lg4ff.c 	if (!entry) {
entry            1136 drivers/hid/hid-lg4ff.c 		if (led_cdev != entry->wdata.led[i])
entry            1138 drivers/hid/hid-lg4ff.c 		state = (entry->wdata.led_state >> i) & 1;
entry            1140 drivers/hid/hid-lg4ff.c 			entry->wdata.led_state &= ~(1 << i);
entry            1141 drivers/hid/hid-lg4ff.c 			lg4ff_set_leds(hid, entry->wdata.led_state);
entry            1143 drivers/hid/hid-lg4ff.c 			entry->wdata.led_state |= 1 << i;
entry            1144 drivers/hid/hid-lg4ff.c 			lg4ff_set_leds(hid, entry->wdata.led_state);
entry            1155 drivers/hid/hid-lg4ff.c 	struct lg4ff_device_entry *entry;
entry            1163 drivers/hid/hid-lg4ff.c 	entry = drv_data->device_props;
entry            1165 drivers/hid/hid-lg4ff.c 	if (!entry) {
entry            1171 drivers/hid/hid-lg4ff.c 		if (led_cdev == entry->wdata.led[i]) {
entry            1172 drivers/hid/hid-lg4ff.c 			value = (entry->wdata.led_state >> i) & 1;
entry            1263 drivers/hid/hid-lg4ff.c 	struct lg4ff_device_entry *entry;
entry            1285 drivers/hid/hid-lg4ff.c 	entry = kzalloc(sizeof(*entry), GFP_KERNEL);
entry            1286 drivers/hid/hid-lg4ff.c 	if (!entry)
entry            1288 drivers/hid/hid-lg4ff.c 	spin_lock_init(&entry->report_lock);
entry            1289 drivers/hid/hid-lg4ff.c 	entry->report = report;
entry            1290 drivers/hid/hid-lg4ff.c 	drv_data->device_props = entry;
entry            1350 drivers/hid/hid-lg4ff.c 	lg4ff_init_wheel_data(&entry->wdata, &lg4ff_devices[i], mmode_wheel, real_product_id);
entry            1383 drivers/hid/hid-lg4ff.c 	entry->wdata.range = entry->wdata.max_range;
entry            1384 drivers/hid/hid-lg4ff.c 	if (entry->wdata.set_range)
entry            1385 drivers/hid/hid-lg4ff.c 		entry->wdata.set_range(hid, entry->wdata.range);
entry            1389 drivers/hid/hid-lg4ff.c 	entry->wdata.led_state = 0;
entry            1391 drivers/hid/hid-lg4ff.c 		entry->wdata.led[j] = NULL;
entry            1418 drivers/hid/hid-lg4ff.c 			entry->wdata.led[j] = led;
entry            1426 drivers/hid/hid-lg4ff.c 					led = entry->wdata.led[j];
entry            1427 drivers/hid/hid-lg4ff.c 					entry->wdata.led[j] = NULL;
entry            1444 drivers/hid/hid-lg4ff.c 	kfree(entry);
entry            1450 drivers/hid/hid-lg4ff.c 	struct lg4ff_device_entry *entry;
entry            1458 drivers/hid/hid-lg4ff.c 	entry = drv_data->device_props;
entry            1459 drivers/hid/hid-lg4ff.c 	if (!entry)
entry            1463 drivers/hid/hid-lg4ff.c 	if (entry->wdata.alternate_modes) {
entry            1478 drivers/hid/hid-lg4ff.c 			led = entry->wdata.led[j];
entry            1479 drivers/hid/hid-lg4ff.c 			entry->wdata.led[j] = NULL;
entry            1489 drivers/hid/hid-lg4ff.c 	kfree(entry);
entry            2396 drivers/hid/hid-sony.c 	struct sony_sc *entry;
entry            2402 drivers/hid/hid-sony.c 	list_for_each_entry(entry, &sony_device_list, list_node) {
entry            2403 drivers/hid/hid-sony.c 		ret = memcmp(sc->mac_address, entry->mac_address,
entry            2406 drivers/hid/hid-sony.c 			if (sony_compare_connection_type(sc, entry)) {
entry             133 drivers/hsi/clients/cmt_speech.c 	struct char_queue *entry;
entry             142 drivers/hsi/clients/cmt_speech.c 	entry = kmalloc(sizeof(*entry), GFP_ATOMIC);
entry             143 drivers/hsi/clients/cmt_speech.c 	if (!entry) {
entry             150 drivers/hsi/clients/cmt_speech.c 	entry->msg = message;
entry             151 drivers/hsi/clients/cmt_speech.c 	list_add_tail(&entry->list, head);
entry             164 drivers/hsi/clients/cmt_speech.c 	struct char_queue *entry;
entry             167 drivers/hsi/clients/cmt_speech.c 	entry = list_entry(head->next, struct char_queue, list);
entry             168 drivers/hsi/clients/cmt_speech.c 	data = entry->msg;
entry             169 drivers/hsi/clients/cmt_speech.c 	list_del(&entry->list);
entry             170 drivers/hsi/clients/cmt_speech.c 	kfree(entry);
entry            1322 drivers/hsi/clients/cmt_speech.c 	struct char_queue *entry;
entry            1327 drivers/hsi/clients/cmt_speech.c 			entry = list_entry(cursor, struct char_queue, list);
entry            1328 drivers/hsi/clients/cmt_speech.c 			list_del(&entry->list);
entry            1329 drivers/hsi/clients/cmt_speech.c 			kfree(entry);
entry            1058 drivers/hv/vmbus_drv.c 	const struct vmbus_channel_message_table_entry *entry;
entry            1075 drivers/hv/vmbus_drv.c 	entry = &channel_message_table[hdr->msgtype];
entry            1076 drivers/hv/vmbus_drv.c 	if (entry->handler_type	== VMHT_BLOCKING) {
entry            1111 drivers/hv/vmbus_drv.c 		entry->message_handler(hdr);
entry             301 drivers/hwmon/applesmc.c static int applesmc_read_entry(const struct applesmc_entry *entry,
entry             306 drivers/hwmon/applesmc.c 	if (entry->len != len)
entry             309 drivers/hwmon/applesmc.c 	ret = read_smc(APPLESMC_READ_CMD, entry->key, buf, len);
entry             315 drivers/hwmon/applesmc.c static int applesmc_write_entry(const struct applesmc_entry *entry,
entry             320 drivers/hwmon/applesmc.c 	if (entry->len != len)
entry             323 drivers/hwmon/applesmc.c 	ret = write_smc(APPLESMC_WRITE_CMD, entry->key, buf, len);
entry             366 drivers/hwmon/applesmc.c 	const struct applesmc_entry *entry;
entry             370 drivers/hwmon/applesmc.c 		entry = applesmc_get_entry_by_index(middle);
entry             371 drivers/hwmon/applesmc.c 		if (IS_ERR(entry)) {
entry             373 drivers/hwmon/applesmc.c 			return PTR_ERR(entry);
entry             375 drivers/hwmon/applesmc.c 		if (strcmp(entry->key, key) < 0)
entry             388 drivers/hwmon/applesmc.c 	const struct applesmc_entry *entry;
entry             392 drivers/hwmon/applesmc.c 		entry = applesmc_get_entry_by_index(middle);
entry             393 drivers/hwmon/applesmc.c 		if (IS_ERR(entry)) {
entry             395 drivers/hwmon/applesmc.c 			return PTR_ERR(entry);
entry             397 drivers/hwmon/applesmc.c 		if (strcmp(key, entry->key) < 0)
entry             426 drivers/hwmon/applesmc.c 	const struct applesmc_entry *entry;
entry             428 drivers/hwmon/applesmc.c 	entry = applesmc_get_entry_by_key(key);
entry             429 drivers/hwmon/applesmc.c 	if (IS_ERR(entry))
entry             430 drivers/hwmon/applesmc.c 		return PTR_ERR(entry);
entry             432 drivers/hwmon/applesmc.c 	return applesmc_read_entry(entry, buffer, len);
entry             437 drivers/hwmon/applesmc.c 	const struct applesmc_entry *entry;
entry             439 drivers/hwmon/applesmc.c 	entry = applesmc_get_entry_by_key(key);
entry             440 drivers/hwmon/applesmc.c 	if (IS_ERR(entry))
entry             441 drivers/hwmon/applesmc.c 		return PTR_ERR(entry);
entry             443 drivers/hwmon/applesmc.c 	return applesmc_write_entry(entry, buffer, len);
entry             448 drivers/hwmon/applesmc.c 	const struct applesmc_entry *entry;
entry             450 drivers/hwmon/applesmc.c 	entry = applesmc_get_entry_by_key(key);
entry             451 drivers/hwmon/applesmc.c 	if (IS_ERR(entry) && PTR_ERR(entry) != -EINVAL)
entry             452 drivers/hwmon/applesmc.c 		return PTR_ERR(entry);
entry             454 drivers/hwmon/applesmc.c 	*value = !IS_ERR(entry);
entry             500 drivers/hwmon/applesmc.c 	const struct applesmc_entry *entry;
entry             511 drivers/hwmon/applesmc.c 		entry = applesmc_get_entry_by_index(i);
entry             512 drivers/hwmon/applesmc.c 		if (IS_ERR(entry))
entry             514 drivers/hwmon/applesmc.c 		if (strcmp(entry->type, TEMP_SENSOR_TYPE))
entry             516 drivers/hwmon/applesmc.c 		s->index[s->index_count++] = entry->key;
entry             734 drivers/hwmon/applesmc.c 	const struct applesmc_entry *entry;
entry             741 drivers/hwmon/applesmc.c 		entry = applesmc_get_entry_by_key(LIGHT_SENSOR_LEFT_KEY);
entry             742 drivers/hwmon/applesmc.c 		if (IS_ERR(entry))
entry             743 drivers/hwmon/applesmc.c 			return PTR_ERR(entry);
entry             744 drivers/hwmon/applesmc.c 		if (entry->len > 10)
entry             746 drivers/hwmon/applesmc.c 		data_length = entry->len;
entry             961 drivers/hwmon/applesmc.c 	const struct applesmc_entry *entry;
entry             964 drivers/hwmon/applesmc.c 	entry = applesmc_get_entry_by_index(key_at_index);
entry             965 drivers/hwmon/applesmc.c 	if (IS_ERR(entry))
entry             966 drivers/hwmon/applesmc.c 		return PTR_ERR(entry);
entry             967 drivers/hwmon/applesmc.c 	ret = applesmc_read_entry(entry, sysfsbuf, entry->len);
entry             971 drivers/hwmon/applesmc.c 	return entry->len;
entry             977 drivers/hwmon/applesmc.c 	const struct applesmc_entry *entry;
entry             979 drivers/hwmon/applesmc.c 	entry = applesmc_get_entry_by_index(key_at_index);
entry             980 drivers/hwmon/applesmc.c 	if (IS_ERR(entry))
entry             981 drivers/hwmon/applesmc.c 		return PTR_ERR(entry);
entry             983 drivers/hwmon/applesmc.c 	return snprintf(sysfsbuf, PAGE_SIZE, "%d\n", entry->len);
entry             989 drivers/hwmon/applesmc.c 	const struct applesmc_entry *entry;
entry             991 drivers/hwmon/applesmc.c 	entry = applesmc_get_entry_by_index(key_at_index);
entry             992 drivers/hwmon/applesmc.c 	if (IS_ERR(entry))
entry             993 drivers/hwmon/applesmc.c 		return PTR_ERR(entry);
entry             995 drivers/hwmon/applesmc.c 	return snprintf(sysfsbuf, PAGE_SIZE, "%s\n", entry->type);
entry            1001 drivers/hwmon/applesmc.c 	const struct applesmc_entry *entry;
entry            1003 drivers/hwmon/applesmc.c 	entry = applesmc_get_entry_by_index(key_at_index);
entry            1004 drivers/hwmon/applesmc.c 	if (IS_ERR(entry))
entry            1005 drivers/hwmon/applesmc.c 		return PTR_ERR(entry);
entry            1007 drivers/hwmon/applesmc.c 	return snprintf(sysfsbuf, PAGE_SIZE, "%s\n", entry->key);
entry             324 drivers/hwmon/k10temp.c 		const struct tctl_offset *entry = &tctl_offset_table[i];
entry             326 drivers/hwmon/k10temp.c 		if (boot_cpu_data.x86 == entry->model &&
entry             327 drivers/hwmon/k10temp.c 		    strstr(boot_cpu_data.x86_model_id, entry->id)) {
entry             328 drivers/hwmon/k10temp.c 			data->temp_offset = entry->offset;
entry            2256 drivers/hwmon/pmbus/pmbus_core.c 	struct pmbus_debugfs_entry *entry = data;
entry            2258 drivers/hwmon/pmbus/pmbus_core.c 	rc = _pmbus_read_byte_data(entry->client, entry->page, entry->reg);
entry            2272 drivers/hwmon/pmbus/pmbus_core.c 	struct pmbus_debugfs_entry *entry = data;
entry            2273 drivers/hwmon/pmbus/pmbus_core.c 	struct pmbus_data *pdata = i2c_get_clientdata(entry->client);
entry            2275 drivers/hwmon/pmbus/pmbus_core.c 	rc = pdata->read_status(entry->client, entry->page);
entry             372 drivers/hwmon/pmbus/ucd9000.c 	struct ucd9000_debugfs_entry *entry = data;
entry             373 drivers/hwmon/pmbus/ucd9000.c 	struct i2c_client *client = entry->client;
entry             385 drivers/hwmon/pmbus/ucd9000.c 	*val = !!(buffer[1] & BIT(entry->index));
entry             101 drivers/hwtracing/coresight/coresight-catu.c #define CATU_ENTRY_ADDR(entry)	((cate_t)(entry) & ~((cate_t)CATU_ENTRY_VALID))
entry             416 drivers/hwtracing/coresight/coresight-etm-perf.c 	list_for_each_entry(filter, filters, entry) {
entry             459 drivers/hwtracing/coresight/coresight-etm-perf.c 	list_for_each_entry(filter, &head->list, entry) {
entry             487 drivers/hwtracing/coresight/coresight-etm-perf.c 	char entry[sizeof("cpu9999999")];
entry             492 drivers/hwtracing/coresight/coresight-etm-perf.c 	sprintf(entry, "cpu%d", cpu);
entry             498 drivers/hwtracing/coresight/coresight-etm-perf.c 		ret = sysfs_create_link(&pmu_dev->kobj, &cs_dev->kobj, entry);
entry             503 drivers/hwtracing/coresight/coresight-etm-perf.c 		sysfs_remove_link(&pmu_dev->kobj, entry);
entry              96 drivers/hwtracing/coresight/coresight-tmc-etr.c #define ETR_SG_ADDR(entry) \
entry              97 drivers/hwtracing/coresight/coresight-tmc-etr.c 	(((dma_addr_t)(entry) >> ETR_SG_ADDR_SHIFT) << ETR_SG_PAGE_SHIFT)
entry              98 drivers/hwtracing/coresight/coresight-tmc-etr.c #define ETR_SG_ET(entry)		((entry) & ETR_SG_ET_MASK)
entry              68 drivers/hwtracing/intel_th/msu.c 	struct list_head	entry;
entry              93 drivers/hwtracing/intel_th/msu.c 	struct list_head	entry;
entry             176 drivers/hwtracing/intel_th/msu.c 	struct list_head	entry;
entry             187 drivers/hwtracing/intel_th/msu.c 	list_for_each_entry(mbe, &msu_buffer_list, entry) {
entry             239 drivers/hwtracing/intel_th/msu.c 	list_add_tail(&mbe->entry, &msu_buffer_list);
entry             254 drivers/hwtracing/intel_th/msu.c 		list_del(&mbe->entry);
entry             302 drivers/hwtracing/intel_th/msu.c 	return win->entry.next == &win->msc->win_list;
entry             315 drivers/hwtracing/intel_th/msu.c 					entry);
entry             317 drivers/hwtracing/intel_th/msu.c 	return list_next_entry(win, entry);
entry             363 drivers/hwtracing/intel_th/msu.c 	list_for_each_entry(win, &msc->win_list, entry) {
entry             398 drivers/hwtracing/intel_th/msu.c 	return list_first_entry(&msc->win_list, struct msc_window, entry);
entry             460 drivers/hwtracing/intel_th/msu.c 	list_add_tail(&iter->entry, &msc->iter_list);
entry             470 drivers/hwtracing/intel_th/msu.c 	list_del(&iter->entry);
entry             655 drivers/hwtracing/intel_th/msu.c 	list_for_each_entry(win, &msc->win_list, entry) {
entry            1082 drivers/hwtracing/intel_th/msu.c 							  entry);
entry            1107 drivers/hwtracing/intel_th/msu.c 	list_add_tail(&win->entry, &msc->win_list);
entry            1145 drivers/hwtracing/intel_th/msu.c 	list_del(&win->entry);
entry            1173 drivers/hwtracing/intel_th/msu.c 	list_for_each_entry(win, &msc->win_list, entry) {
entry            1185 drivers/hwtracing/intel_th/msu.c 						    struct msc_window, entry);
entry            1187 drivers/hwtracing/intel_th/msu.c 			next_win = list_next_entry(win, entry);
entry            1226 drivers/hwtracing/intel_th/msu.c 	list_for_each_entry_safe(win, iter, &msc->win_list, entry)
entry            1376 drivers/hwtracing/intel_th/msu.c 	list_for_each_entry(win, &msc->win_list, entry)
entry            1687 drivers/hwtracing/intel_th/msu.c 	first = list_first_entry(&msc->win_list, struct msc_window, entry);
entry            1692 drivers/hwtracing/intel_th/msu.c 		msc->cur_win = list_next_entry(msc->cur_win, entry);
entry            1933 drivers/hwtracing/intel_th/msu.c 		list_for_each_entry(win, &msc->win_list, entry) {
entry             373 drivers/hwtracing/stm/core.c 	struct list_head			entry;
entry             389 drivers/hwtracing/stm/core.c 	list_for_each_entry(pe, &stm_pdrv_head, entry) {
entry             419 drivers/hwtracing/stm/core.c 	list_add_tail(&pe->entry, &stm_pdrv_head);
entry             439 drivers/hwtracing/stm/core.c 	list_for_each_entry_safe(pe, iter, &stm_pdrv_head, entry) {
entry             441 drivers/hwtracing/stm/core.c 			list_del(&pe->entry);
entry              59 drivers/infiniband/core/agent.c 	struct ib_agent_port_private *entry;
entry              61 drivers/infiniband/core/agent.c 	list_for_each_entry(entry, &ib_agent_port_list, port_list) {
entry              62 drivers/infiniband/core/agent.c 		if (entry->agent[1]->device == device &&
entry              63 drivers/infiniband/core/agent.c 		    entry->agent[1]->port_num == port_num)
entry              64 drivers/infiniband/core/agent.c 			return entry;
entry              72 drivers/infiniband/core/agent.c 	struct ib_agent_port_private *entry;
entry              76 drivers/infiniband/core/agent.c 	entry = __ib_get_agent_port(device, port_num);
entry              78 drivers/infiniband/core/agent.c 	return entry;
entry             201 drivers/infiniband/core/cache.c static bool is_gid_entry_free(const struct ib_gid_table_entry *entry)
entry             203 drivers/infiniband/core/cache.c 	return !entry;
entry             206 drivers/infiniband/core/cache.c static bool is_gid_entry_valid(const struct ib_gid_table_entry *entry)
entry             208 drivers/infiniband/core/cache.c 	return entry && entry->state == GID_TABLE_ENTRY_VALID;
entry             213 drivers/infiniband/core/cache.c 	struct ib_gid_table_entry *entry =
entry             216 drivers/infiniband/core/cache.c 	queue_work(ib_wq, &entry->del_work);
entry             233 drivers/infiniband/core/cache.c static void free_gid_entry_locked(struct ib_gid_table_entry *entry)
entry             235 drivers/infiniband/core/cache.c 	struct ib_device *device = entry->attr.device;
entry             236 drivers/infiniband/core/cache.c 	u8 port_num = entry->attr.port_num;
entry             240 drivers/infiniband/core/cache.c 		port_num, entry->attr.index, entry->attr.gid.raw);
entry             250 drivers/infiniband/core/cache.c 	if (entry == table->data_vec[entry->attr.index])
entry             251 drivers/infiniband/core/cache.c 		table->data_vec[entry->attr.index] = NULL;
entry             255 drivers/infiniband/core/cache.c 	if (entry->ndev_storage)
entry             256 drivers/infiniband/core/cache.c 		call_rcu(&entry->ndev_storage->rcu_head, put_gid_ndev);
entry             257 drivers/infiniband/core/cache.c 	kfree(entry);
entry             262 drivers/infiniband/core/cache.c 	struct ib_gid_table_entry *entry =
entry             265 drivers/infiniband/core/cache.c 	free_gid_entry_locked(entry);
entry             278 drivers/infiniband/core/cache.c 	struct ib_gid_table_entry *entry =
entry             280 drivers/infiniband/core/cache.c 	struct ib_device *device = entry->attr.device;
entry             281 drivers/infiniband/core/cache.c 	u8 port_num = entry->attr.port_num;
entry             285 drivers/infiniband/core/cache.c 	free_gid_entry_locked(entry);
entry             292 drivers/infiniband/core/cache.c 	struct ib_gid_table_entry *entry;
entry             295 drivers/infiniband/core/cache.c 	entry = kzalloc(sizeof(*entry), GFP_KERNEL);
entry             296 drivers/infiniband/core/cache.c 	if (!entry)
entry             301 drivers/infiniband/core/cache.c 		entry->ndev_storage = kzalloc(sizeof(*entry->ndev_storage),
entry             303 drivers/infiniband/core/cache.c 		if (!entry->ndev_storage) {
entry             304 drivers/infiniband/core/cache.c 			kfree(entry);
entry             308 drivers/infiniband/core/cache.c 		entry->ndev_storage->ndev = ndev;
entry             310 drivers/infiniband/core/cache.c 	kref_init(&entry->kref);
entry             311 drivers/infiniband/core/cache.c 	memcpy(&entry->attr, attr, sizeof(*attr));
entry             312 drivers/infiniband/core/cache.c 	INIT_WORK(&entry->del_work, free_gid_work);
entry             313 drivers/infiniband/core/cache.c 	entry->state = GID_TABLE_ENTRY_INVALID;
entry             314 drivers/infiniband/core/cache.c 	return entry;
entry             318 drivers/infiniband/core/cache.c 			    struct ib_gid_table_entry *entry)
entry             320 drivers/infiniband/core/cache.c 	entry->state = GID_TABLE_ENTRY_VALID;
entry             322 drivers/infiniband/core/cache.c 	dev_dbg(&entry->attr.device->dev, "%s port=%d index=%d gid %pI6\n",
entry             323 drivers/infiniband/core/cache.c 		__func__, entry->attr.port_num, entry->attr.index,
entry             324 drivers/infiniband/core/cache.c 		entry->attr.gid.raw);
entry             328 drivers/infiniband/core/cache.c 	table->data_vec[entry->attr.index] = entry;
entry             332 drivers/infiniband/core/cache.c static void get_gid_entry(struct ib_gid_table_entry *entry)
entry             334 drivers/infiniband/core/cache.c 	kref_get(&entry->kref);
entry             337 drivers/infiniband/core/cache.c static void put_gid_entry(struct ib_gid_table_entry *entry)
entry             339 drivers/infiniband/core/cache.c 	kref_put(&entry->kref, schedule_free_gid);
entry             342 drivers/infiniband/core/cache.c static void put_gid_entry_locked(struct ib_gid_table_entry *entry)
entry             344 drivers/infiniband/core/cache.c 	kref_put(&entry->kref, free_gid_entry);
entry             347 drivers/infiniband/core/cache.c static int add_roce_gid(struct ib_gid_table_entry *entry)
entry             349 drivers/infiniband/core/cache.c 	const struct ib_gid_attr *attr = &entry->attr;
entry             358 drivers/infiniband/core/cache.c 		ret = attr->device->ops.add_gid(attr, &entry->context);
entry             382 drivers/infiniband/core/cache.c 	struct ib_gid_table_entry *entry;
entry             390 drivers/infiniband/core/cache.c 	entry = table->data_vec[ix];
entry             391 drivers/infiniband/core/cache.c 	entry->state = GID_TABLE_ENTRY_PENDING_DEL;
entry             399 drivers/infiniband/core/cache.c 	ndev_storage = entry->ndev_storage;
entry             401 drivers/infiniband/core/cache.c 		entry->ndev_storage = NULL;
entry             402 drivers/infiniband/core/cache.c 		rcu_assign_pointer(entry->attr.ndev, NULL);
entry             407 drivers/infiniband/core/cache.c 		ib_dev->ops.del_gid(&entry->attr, &entry->context);
entry             409 drivers/infiniband/core/cache.c 	put_gid_entry_locked(entry);
entry             425 drivers/infiniband/core/cache.c 	struct ib_gid_table_entry *entry;
entry             443 drivers/infiniband/core/cache.c 	entry = alloc_gid_entry(attr);
entry             444 drivers/infiniband/core/cache.c 	if (!entry)
entry             448 drivers/infiniband/core/cache.c 		ret = add_roce_gid(entry);
entry             453 drivers/infiniband/core/cache.c 	store_gid_entry(table, entry);
entry             457 drivers/infiniband/core/cache.c 	put_gid_entry(entry);
entry             750 drivers/infiniband/core/cache.c 		struct ib_gid_table_entry *entry = table->data_vec[i];
entry             752 drivers/infiniband/core/cache.c 		if (!is_gid_entry_valid(entry))
entry             755 drivers/infiniband/core/cache.c 		if (memcmp(gid, &entry->attr.gid, sizeof(*gid)))
entry             758 drivers/infiniband/core/cache.c 		if (filter(gid, &entry->attr, context)) {
entry             759 drivers/infiniband/core/cache.c 			get_gid_entry(entry);
entry             760 drivers/infiniband/core/cache.c 			res = &entry->attr;
entry            1238 drivers/infiniband/core/cache.c 	struct ib_gid_table_entry *entry =
entry            1241 drivers/infiniband/core/cache.c 	put_gid_entry(entry);
entry            1257 drivers/infiniband/core/cache.c 	struct ib_gid_table_entry *entry =
entry            1260 drivers/infiniband/core/cache.c 	get_gid_entry(entry);
entry            1278 drivers/infiniband/core/cache.c 	struct ib_gid_table_entry *entry =
entry            1280 drivers/infiniband/core/cache.c 	struct ib_device *device = entry->attr.device;
entry            1282 drivers/infiniband/core/cache.c 	u8 port_num = entry->attr.port_num;
entry             161 drivers/infiniband/core/device.c 	void *entry;
entry             165 drivers/infiniband/core/device.c 		entry = xas_find_marked(&xas, ULONG_MAX, filter);
entry             166 drivers/infiniband/core/device.c 		if (xa_is_zero(entry))
entry             168 drivers/infiniband/core/device.c 	} while (xas_retry(&xas, entry));
entry             171 drivers/infiniband/core/device.c 	if (entry) {
entry             173 drivers/infiniband/core/device.c 		if (xa_is_zero(entry))
entry             175 drivers/infiniband/core/device.c 		return entry;
entry             179 drivers/infiniband/core/device.c #define xan_for_each_marked(xa, index, entry, filter)                          \
entry             180 drivers/infiniband/core/device.c 	for (index = 0, entry = xan_find_marked(xa, &(index), filter);         \
entry             181 drivers/infiniband/core/device.c 	     !xa_is_err(entry);                                                \
entry             182 drivers/infiniband/core/device.c 	     (index)++, entry = xan_find_marked(xa, &(index), filter))
entry              60 drivers/infiniband/core/mad.c 			  struct trace_event_raw_ib_mad_send_template *entry)
entry              71 drivers/infiniband/core/mad.c 	entry->sl = attr.sl;
entry              73 drivers/infiniband/core/mad.c 	entry->pkey = pkey;
entry              74 drivers/infiniband/core/mad.c 	entry->rqpn = wr->remote_qpn;
entry              75 drivers/infiniband/core/mad.c 	entry->rqkey = wr->remote_qkey;
entry              76 drivers/infiniband/core/mad.c 	entry->dlid = rdma_ah_get_dlid(&attr);
entry             124 drivers/infiniband/core/mad.c 	struct ib_mad_port_private *entry;
entry             126 drivers/infiniband/core/mad.c 	list_for_each_entry(entry, &ib_mad_port_list, port_list) {
entry             127 drivers/infiniband/core/mad.c 		if (entry->device == device && entry->port_num == port_num)
entry             128 drivers/infiniband/core/mad.c 			return entry;
entry             140 drivers/infiniband/core/mad.c 	struct ib_mad_port_private *entry;
entry             144 drivers/infiniband/core/mad.c 	entry = __ib_get_mad_port(device, port_num);
entry             147 drivers/infiniband/core/mad.c 	return entry;
entry            1129 drivers/infiniband/core/nldev.c 	u32 entry;
entry            1142 drivers/infiniband/core/nldev.c 		.entry = RDMA_NLDEV_ATTR_RES_QP_ENTRY,
entry            1149 drivers/infiniband/core/nldev.c 		.entry = RDMA_NLDEV_ATTR_RES_CM_ID_ENTRY,
entry            1157 drivers/infiniband/core/nldev.c 		.entry = RDMA_NLDEV_ATTR_RES_CQ_ENTRY,
entry            1165 drivers/infiniband/core/nldev.c 		.entry = RDMA_NLDEV_ATTR_RES_MR_ENTRY,
entry            1173 drivers/infiniband/core/nldev.c 		.entry = RDMA_NLDEV_ATTR_RES_PD_ENTRY,
entry            1180 drivers/infiniband/core/nldev.c 		.entry = RDMA_NLDEV_ATTR_STAT_COUNTER_ENTRY,
entry            1351 drivers/infiniband/core/nldev.c 		entry_attr = nla_nest_start_noflag(skb, fe->entry);
entry             726 drivers/infiniband/core/rdma_core.c 	struct ib_uobject *entry;
entry             737 drivers/infiniband/core/rdma_core.c 	xa_for_each(&ufile->idr, id, entry) {
entry             738 drivers/infiniband/core/rdma_core.c 		WARN_ON(entry->object);
entry             739 drivers/infiniband/core/rdma_core.c 		uverbs_uobject_put(entry);
entry             354 drivers/infiniband/core/roce_gid_mgmt.c 		struct sin_list *entry = kzalloc(sizeof(*entry), GFP_ATOMIC);
entry             356 drivers/infiniband/core/roce_gid_mgmt.c 		if (!entry)
entry             359 drivers/infiniband/core/roce_gid_mgmt.c 		entry->ip.sin_family = AF_INET;
entry             360 drivers/infiniband/core/roce_gid_mgmt.c 		entry->ip.sin_addr.s_addr = ifa->ifa_address;
entry             361 drivers/infiniband/core/roce_gid_mgmt.c 		list_add_tail(&entry->list, &sin_list);
entry             397 drivers/infiniband/core/roce_gid_mgmt.c 		struct sin6_list *entry = kzalloc(sizeof(*entry), GFP_ATOMIC);
entry             399 drivers/infiniband/core/roce_gid_mgmt.c 		if (!entry)
entry             402 drivers/infiniband/core/roce_gid_mgmt.c 		entry->sin6.sin6_family = AF_INET6;
entry             403 drivers/infiniband/core/roce_gid_mgmt.c 		entry->sin6.sin6_addr = ifp->addr;
entry             404 drivers/infiniband/core/roce_gid_mgmt.c 		list_add_tail(&entry->list, &sin6_list);
entry             536 drivers/infiniband/core/roce_gid_mgmt.c 	struct upper_list *entry = kmalloc(sizeof(*entry), GFP_ATOMIC);
entry             539 drivers/infiniband/core/roce_gid_mgmt.c 	if (!entry)
entry             542 drivers/infiniband/core/roce_gid_mgmt.c 	list_add_tail(&entry->list, upper_list);
entry             544 drivers/infiniband/core/roce_gid_mgmt.c 	entry->upper = upper;
entry            1151 drivers/infiniband/core/sysfs.c 	list_add_tail(&p->kobj.entry, &coredev->port_list);
entry            1312 drivers/infiniband/core/sysfs.c 	list_for_each_entry_safe(p, t, &coredev->port_list, entry) {
entry            1315 drivers/infiniband/core/sysfs.c 		list_del(&p->entry);
entry            1401 drivers/infiniband/core/sysfs.c 	list_for_each_entry_safe(p, t, &device->coredev.port_list, entry) {
entry             481 drivers/infiniband/core/uverbs_cmd.c 	struct xrcd_table_entry *entry, *scan;
entry             485 drivers/infiniband/core/uverbs_cmd.c 	entry = kmalloc(sizeof *entry, GFP_KERNEL);
entry             486 drivers/infiniband/core/uverbs_cmd.c 	if (!entry)
entry             489 drivers/infiniband/core/uverbs_cmd.c 	entry->xrcd  = xrcd;
entry             490 drivers/infiniband/core/uverbs_cmd.c 	entry->inode = inode;
entry             501 drivers/infiniband/core/uverbs_cmd.c 			kfree(entry);
entry             506 drivers/infiniband/core/uverbs_cmd.c 	rb_link_node(&entry->node, parent, p);
entry             507 drivers/infiniband/core/uverbs_cmd.c 	rb_insert_color(&entry->node, &dev->xrcd_tree);
entry             515 drivers/infiniband/core/uverbs_cmd.c 	struct xrcd_table_entry *entry;
entry             519 drivers/infiniband/core/uverbs_cmd.c 		entry = rb_entry(p, struct xrcd_table_entry, node);
entry             521 drivers/infiniband/core/uverbs_cmd.c 		if (inode < entry->inode)
entry             523 drivers/infiniband/core/uverbs_cmd.c 		else if (inode > entry->inode)
entry             526 drivers/infiniband/core/uverbs_cmd.c 			return entry;
entry             534 drivers/infiniband/core/uverbs_cmd.c 	struct xrcd_table_entry *entry;
entry             536 drivers/infiniband/core/uverbs_cmd.c 	entry = xrcd_table_search(dev, inode);
entry             537 drivers/infiniband/core/uverbs_cmd.c 	if (!entry)
entry             540 drivers/infiniband/core/uverbs_cmd.c 	return entry->xrcd;
entry             546 drivers/infiniband/core/uverbs_cmd.c 	struct xrcd_table_entry *entry;
entry             548 drivers/infiniband/core/uverbs_cmd.c 	entry = xrcd_table_search(dev, inode);
entry             549 drivers/infiniband/core/uverbs_cmd.c 	if (entry) {
entry             551 drivers/infiniband/core/uverbs_cmd.c 		rb_erase(&entry->node, &dev->xrcd_tree);
entry             552 drivers/infiniband/core/uverbs_cmd.c 		kfree(entry);
entry             372 drivers/infiniband/core/uverbs_ioctl.c 		void *entry;
entry             375 drivers/infiniband/core/uverbs_ioctl.c 		entry = rcu_dereference_raw(*slot);
entry             376 drivers/infiniband/core/uverbs_ioctl.c 		if (likely(!radix_tree_is_internal_node(entry) && entry))
entry             353 drivers/infiniband/core/uverbs_main.c 	struct ib_uverbs_event *entry, *tmp;
entry             360 drivers/infiniband/core/uverbs_main.c 	list_for_each_entry_safe(entry, tmp, &file->ev_queue.event_list, list) {
entry             361 drivers/infiniband/core/uverbs_main.c 		if (entry->counter)
entry             362 drivers/infiniband/core/uverbs_main.c 			list_del(&entry->obj_list);
entry             363 drivers/infiniband/core/uverbs_main.c 		kfree(entry);
entry             383 drivers/infiniband/core/uverbs_main.c 	struct ib_uverbs_event *entry, *tmp;
entry             386 drivers/infiniband/core/uverbs_main.c 	list_for_each_entry_safe(entry, tmp, &file->ev_queue.event_list, list) {
entry             387 drivers/infiniband/core/uverbs_main.c 		if (entry->counter)
entry             388 drivers/infiniband/core/uverbs_main.c 			list_del(&entry->obj_list);
entry             389 drivers/infiniband/core/uverbs_main.c 		kfree(entry);
entry             421 drivers/infiniband/core/uverbs_main.c 	struct ib_uverbs_event	       *entry;
entry             433 drivers/infiniband/core/uverbs_main.c 	entry = kmalloc(sizeof(*entry), GFP_ATOMIC);
entry             434 drivers/infiniband/core/uverbs_main.c 	if (!entry) {
entry             441 drivers/infiniband/core/uverbs_main.c 	entry->desc.comp.cq_handle = cq->uobject->user_handle;
entry             442 drivers/infiniband/core/uverbs_main.c 	entry->counter		   = &uobj->comp_events_reported;
entry             444 drivers/infiniband/core/uverbs_main.c 	list_add_tail(&entry->list, &ev_queue->event_list);
entry             445 drivers/infiniband/core/uverbs_main.c 	list_add_tail(&entry->obj_list, &uobj->comp_list);
entry             457 drivers/infiniband/core/uverbs_main.c 	struct ib_uverbs_event *entry;
entry             466 drivers/infiniband/core/uverbs_main.c 	entry = kmalloc(sizeof(*entry), GFP_ATOMIC);
entry             467 drivers/infiniband/core/uverbs_main.c 	if (!entry) {
entry             472 drivers/infiniband/core/uverbs_main.c 	entry->desc.async.element    = element;
entry             473 drivers/infiniband/core/uverbs_main.c 	entry->desc.async.event_type = event;
entry             474 drivers/infiniband/core/uverbs_main.c 	entry->desc.async.reserved   = 0;
entry             475 drivers/infiniband/core/uverbs_main.c 	entry->counter               = counter;
entry             477 drivers/infiniband/core/uverbs_main.c 	list_add_tail(&entry->list, &file->async_file->ev_queue.event_list);
entry             479 drivers/infiniband/core/uverbs_main.c 		list_add_tail(&entry->obj_list, obj_list);
entry              55 drivers/infiniband/hw/cxgb3/cxio_hal.c 	list_for_each_entry(rdev, &rdev_list, entry)
entry              65 drivers/infiniband/hw/cxgb3/cxio_hal.c 	list_for_each_entry(rdev, &rdev_list, entry)
entry             191 drivers/infiniband/hw/cxgb3/cxio_hal.c 	struct cxio_qpid_list *entry;
entry             197 drivers/infiniband/hw/cxgb3/cxio_hal.c 		entry = list_entry(uctx->qpids.next, struct cxio_qpid_list,
entry             198 drivers/infiniband/hw/cxgb3/cxio_hal.c 				   entry);
entry             199 drivers/infiniband/hw/cxgb3/cxio_hal.c 		list_del(&entry->entry);
entry             200 drivers/infiniband/hw/cxgb3/cxio_hal.c 		qpid = entry->qpid;
entry             201 drivers/infiniband/hw/cxgb3/cxio_hal.c 		kfree(entry);
entry             207 drivers/infiniband/hw/cxgb3/cxio_hal.c 			entry = kmalloc(sizeof(*entry), GFP_KERNEL);
entry             208 drivers/infiniband/hw/cxgb3/cxio_hal.c 			if (!entry)
entry             210 drivers/infiniband/hw/cxgb3/cxio_hal.c 			entry->qpid = i;
entry             211 drivers/infiniband/hw/cxgb3/cxio_hal.c 			list_add_tail(&entry->entry, &uctx->qpids);
entry             223 drivers/infiniband/hw/cxgb3/cxio_hal.c 	struct cxio_qpid_list *entry;
entry             225 drivers/infiniband/hw/cxgb3/cxio_hal.c 	entry = kmalloc(sizeof(*entry), GFP_KERNEL);
entry             226 drivers/infiniband/hw/cxgb3/cxio_hal.c 	if (!entry)
entry             229 drivers/infiniband/hw/cxgb3/cxio_hal.c 	entry->qpid = qpid;
entry             231 drivers/infiniband/hw/cxgb3/cxio_hal.c 	list_add_tail(&entry->entry, &uctx->qpids);
entry             238 drivers/infiniband/hw/cxgb3/cxio_hal.c 	struct cxio_qpid_list *entry;
entry             242 drivers/infiniband/hw/cxgb3/cxio_hal.c 		entry = list_entry(pos, struct cxio_qpid_list, entry);
entry             243 drivers/infiniband/hw/cxgb3/cxio_hal.c 		list_del_init(&entry->entry);
entry             244 drivers/infiniband/hw/cxgb3/cxio_hal.c 		if (!(entry->qpid & rdev_p->qpmask))
entry             245 drivers/infiniband/hw/cxgb3/cxio_hal.c 			cxio_hal_put_qpid(rdev_p->rscp, entry->qpid);
entry             246 drivers/infiniband/hw/cxgb3/cxio_hal.c 		kfree(entry);
entry             914 drivers/infiniband/hw/cxgb3/cxio_hal.c 	list_add_tail(&rdev_p->entry, &rdev_list);
entry            1008 drivers/infiniband/hw/cxgb3/cxio_hal.c 	list_del(&rdev_p->entry);
entry            1017 drivers/infiniband/hw/cxgb3/cxio_hal.c 		list_del(&rdev_p->entry);
entry            1037 drivers/infiniband/hw/cxgb3/cxio_hal.c 	list_for_each_entry_safe(rdev, tmp, &rdev_list, entry)
entry              90 drivers/infiniband/hw/cxgb3/cxio_hal.h 	struct list_head entry;
entry             113 drivers/infiniband/hw/cxgb3/cxio_hal.h 	struct list_head entry;
entry              53 drivers/infiniband/hw/cxgb3/cxio_resource.c 	u32 i, j, entry = 0, idx;
entry              62 drivers/infiniband/hw/cxgb3/cxio_resource.c 		kfifo_in(fifo, (unsigned char *) &entry, sizeof(u32));
entry              89 drivers/infiniband/hw/cxgb3/cxio_resource.c 		if (kfifo_out_locked(fifo, (unsigned char *) &entry,
entry             183 drivers/infiniband/hw/cxgb3/cxio_resource.c 	u32 entry;
entry             184 drivers/infiniband/hw/cxgb3/cxio_resource.c 	if (kfifo_out_locked(fifo, (unsigned char *) &entry, sizeof(u32), lock))
entry             185 drivers/infiniband/hw/cxgb3/cxio_resource.c 		return entry;
entry             191 drivers/infiniband/hw/cxgb3/cxio_resource.c 		u32 entry)
entry             194 drivers/infiniband/hw/cxgb3/cxio_resource.c 	kfifo_in_locked(fifo, (unsigned char *) &entry, sizeof(u32), lock)
entry             160 drivers/infiniband/hw/cxgb3/iwch.c 	list_add_tail(&rnicp->entry, &dev_list);
entry             177 drivers/infiniband/hw/cxgb3/iwch.c 	list_for_each_entry_safe(dev, tmp, &dev_list, entry) {
entry             182 drivers/infiniband/hw/cxgb3/iwch.c 			list_del(&dev->entry);
entry             112 drivers/infiniband/hw/cxgb3/iwch.h 	struct list_head entry;
entry              72 drivers/infiniband/hw/cxgb3/iwch_provider.c 	list_for_each_entry_safe(mm, tmp, &ucontext->mmaps, entry)
entry             200 drivers/infiniband/hw/cxgb3/iwch_provider.h 	struct list_head entry;
entry             215 drivers/infiniband/hw/cxgb3/iwch_provider.h 		mm = list_entry(pos, struct iwch_mm_entry, entry);
entry             217 drivers/infiniband/hw/cxgb3/iwch_provider.h 			list_del_init(&mm->entry);
entry             235 drivers/infiniband/hw/cxgb3/iwch_provider.h 	list_add_tail(&mm->entry, &ucontext->mmaps);
entry            4283 drivers/infiniband/hw/cxgb4/cm.c 		ep = list_entry(tmp, struct c4iw_ep, entry);
entry            4329 drivers/infiniband/hw/cxgb4/cm.c 		if (!ep->entry.next) {
entry            4330 drivers/infiniband/hw/cxgb4/cm.c 			list_add_tail(&ep->entry, &timeout_list);
entry             751 drivers/infiniband/hw/cxgb4/device.c 	struct c4iw_qid_list *entry;
entry             755 drivers/infiniband/hw/cxgb4/device.c 		entry = list_entry(pos, struct c4iw_qid_list, entry);
entry             756 drivers/infiniband/hw/cxgb4/device.c 		list_del_init(&entry->entry);
entry             757 drivers/infiniband/hw/cxgb4/device.c 		if (!(entry->qid & rdev->qpmask)) {
entry             759 drivers/infiniband/hw/cxgb4/device.c 					  entry->qid);
entry             764 drivers/infiniband/hw/cxgb4/device.c 		kfree(entry);
entry             768 drivers/infiniband/hw/cxgb4/device.c 		entry = list_entry(pos, struct c4iw_qid_list, entry);
entry             769 drivers/infiniband/hw/cxgb4/device.c 		list_del_init(&entry->entry);
entry             770 drivers/infiniband/hw/cxgb4/device.c 		kfree(entry);
entry            1094 drivers/infiniband/hw/cxgb4/device.c 	list_add_tail(&ctx->entry, &uld_ctx_list);
entry            1558 drivers/infiniband/hw/cxgb4/device.c 	list_for_each_entry_safe(ctx, tmp, &uld_ctx_list, entry) {
entry             104 drivers/infiniband/hw/cxgb4/iw_cxgb4.h 	struct list_head entry;
entry             333 drivers/infiniband/hw/cxgb4/iw_cxgb4.h 	struct list_head entry;
entry             542 drivers/infiniband/hw/cxgb4/iw_cxgb4.h 	struct list_head entry;
entry             557 drivers/infiniband/hw/cxgb4/iw_cxgb4.h 		mm = list_entry(pos, struct c4iw_mm_entry, entry);
entry             559 drivers/infiniband/hw/cxgb4/iw_cxgb4.h 			list_del_init(&mm->entry);
entry             576 drivers/infiniband/hw/cxgb4/iw_cxgb4.h 	list_add_tail(&mm->entry, &ucontext->mmaps);
entry             883 drivers/infiniband/hw/cxgb4/iw_cxgb4.h 	struct list_head entry;
entry             950 drivers/infiniband/hw/cxgb4/iw_cxgb4.h void c4iw_put_resource(struct c4iw_id_table *id_table, u32 entry);
entry              70 drivers/infiniband/hw/cxgb4/provider.c 	list_for_each_entry_safe(mm, tmp, &ucontext->mmaps, entry)
entry             905 drivers/infiniband/hw/cxgb4/qp.c static void add_to_fc_list(struct list_head *head, struct list_head *entry)
entry             907 drivers/infiniband/hw/cxgb4/qp.c 	if (list_empty(entry))
entry             908 drivers/infiniband/hw/cxgb4/qp.c 		list_add_tail(entry, head);
entry              95 drivers/infiniband/hw/cxgb4/resource.c 	u32 entry;
entry              96 drivers/infiniband/hw/cxgb4/resource.c 	entry = c4iw_id_alloc(id_table);
entry              97 drivers/infiniband/hw/cxgb4/resource.c 	if (entry == (u32)(-1))
entry              99 drivers/infiniband/hw/cxgb4/resource.c 	return entry;
entry             102 drivers/infiniband/hw/cxgb4/resource.c void c4iw_put_resource(struct c4iw_id_table *id_table, u32 entry)
entry             104 drivers/infiniband/hw/cxgb4/resource.c 	pr_debug("entry 0x%x\n", entry);
entry             105 drivers/infiniband/hw/cxgb4/resource.c 	c4iw_id_free(id_table, entry);
entry             110 drivers/infiniband/hw/cxgb4/resource.c 	struct c4iw_qid_list *entry;
entry             116 drivers/infiniband/hw/cxgb4/resource.c 		entry = list_entry(uctx->cqids.next, struct c4iw_qid_list,
entry             117 drivers/infiniband/hw/cxgb4/resource.c 				   entry);
entry             118 drivers/infiniband/hw/cxgb4/resource.c 		list_del(&entry->entry);
entry             119 drivers/infiniband/hw/cxgb4/resource.c 		qid = entry->qid;
entry             120 drivers/infiniband/hw/cxgb4/resource.c 		kfree(entry);
entry             129 drivers/infiniband/hw/cxgb4/resource.c 			entry = kmalloc(sizeof(*entry), GFP_KERNEL);
entry             130 drivers/infiniband/hw/cxgb4/resource.c 			if (!entry)
entry             132 drivers/infiniband/hw/cxgb4/resource.c 			entry->qid = i;
entry             133 drivers/infiniband/hw/cxgb4/resource.c 			list_add_tail(&entry->entry, &uctx->cqids);
entry             140 drivers/infiniband/hw/cxgb4/resource.c 		entry = kmalloc(sizeof(*entry), GFP_KERNEL);
entry             141 drivers/infiniband/hw/cxgb4/resource.c 		if (!entry)
entry             143 drivers/infiniband/hw/cxgb4/resource.c 		entry->qid = qid;
entry             144 drivers/infiniband/hw/cxgb4/resource.c 		list_add_tail(&entry->entry, &uctx->qpids);
entry             146 drivers/infiniband/hw/cxgb4/resource.c 			entry = kmalloc(sizeof(*entry), GFP_KERNEL);
entry             147 drivers/infiniband/hw/cxgb4/resource.c 			if (!entry)
entry             149 drivers/infiniband/hw/cxgb4/resource.c 			entry->qid = i;
entry             150 drivers/infiniband/hw/cxgb4/resource.c 			list_add_tail(&entry->entry, &uctx->qpids);
entry             166 drivers/infiniband/hw/cxgb4/resource.c 	struct c4iw_qid_list *entry;
entry             168 drivers/infiniband/hw/cxgb4/resource.c 	entry = kmalloc(sizeof(*entry), GFP_KERNEL);
entry             169 drivers/infiniband/hw/cxgb4/resource.c 	if (!entry)
entry             172 drivers/infiniband/hw/cxgb4/resource.c 	entry->qid = qid;
entry             174 drivers/infiniband/hw/cxgb4/resource.c 	list_add_tail(&entry->entry, &uctx->cqids);
entry             180 drivers/infiniband/hw/cxgb4/resource.c 	struct c4iw_qid_list *entry;
entry             186 drivers/infiniband/hw/cxgb4/resource.c 		entry = list_entry(uctx->qpids.next, struct c4iw_qid_list,
entry             187 drivers/infiniband/hw/cxgb4/resource.c 				   entry);
entry             188 drivers/infiniband/hw/cxgb4/resource.c 		list_del(&entry->entry);
entry             189 drivers/infiniband/hw/cxgb4/resource.c 		qid = entry->qid;
entry             190 drivers/infiniband/hw/cxgb4/resource.c 		kfree(entry);
entry             203 drivers/infiniband/hw/cxgb4/resource.c 			entry = kmalloc(sizeof(*entry), GFP_KERNEL);
entry             204 drivers/infiniband/hw/cxgb4/resource.c 			if (!entry)
entry             206 drivers/infiniband/hw/cxgb4/resource.c 			entry->qid = i;
entry             207 drivers/infiniband/hw/cxgb4/resource.c 			list_add_tail(&entry->entry, &uctx->qpids);
entry             214 drivers/infiniband/hw/cxgb4/resource.c 		entry = kmalloc(sizeof(*entry), GFP_KERNEL);
entry             215 drivers/infiniband/hw/cxgb4/resource.c 		if (!entry)
entry             217 drivers/infiniband/hw/cxgb4/resource.c 		entry->qid = qid;
entry             218 drivers/infiniband/hw/cxgb4/resource.c 		list_add_tail(&entry->entry, &uctx->cqids);
entry             220 drivers/infiniband/hw/cxgb4/resource.c 			entry = kmalloc(sizeof(*entry), GFP_KERNEL);
entry             221 drivers/infiniband/hw/cxgb4/resource.c 			if (!entry)
entry             223 drivers/infiniband/hw/cxgb4/resource.c 			entry->qid = i;
entry             224 drivers/infiniband/hw/cxgb4/resource.c 			list_add_tail(&entry->entry, &uctx->cqids);
entry             240 drivers/infiniband/hw/cxgb4/resource.c 	struct c4iw_qid_list *entry;
entry             242 drivers/infiniband/hw/cxgb4/resource.c 	entry = kmalloc(sizeof(*entry), GFP_KERNEL);
entry             243 drivers/infiniband/hw/cxgb4/resource.c 	if (!entry)
entry             246 drivers/infiniband/hw/cxgb4/resource.c 	entry->qid = qid;
entry             248 drivers/infiniband/hw/cxgb4/resource.c 	list_add_tail(&entry->entry, &uctx->qpids);
entry             185 drivers/infiniband/hw/efa/efa_verbs.c 	struct efa_mmap_entry *entry;
entry             188 drivers/infiniband/hw/efa/efa_verbs.c 	xa_for_each(&ucontext->mmap_xa, mmap_page, entry) {
entry             194 drivers/infiniband/hw/efa/efa_verbs.c 			entry->obj, get_mmap_key(entry), entry->address,
entry             195 drivers/infiniband/hw/efa/efa_verbs.c 			entry->length);
entry             196 drivers/infiniband/hw/efa/efa_verbs.c 		if (entry->mmap_flag == EFA_MMAP_DMA_PAGE)
entry             198 drivers/infiniband/hw/efa/efa_verbs.c 			free_pages_exact(phys_to_virt(entry->address),
entry             199 drivers/infiniband/hw/efa/efa_verbs.c 					 entry->length);
entry             200 drivers/infiniband/hw/efa/efa_verbs.c 		kfree(entry);
entry             208 drivers/infiniband/hw/efa/efa_verbs.c 	struct efa_mmap_entry *entry;
entry             215 drivers/infiniband/hw/efa/efa_verbs.c 	entry = xa_load(&ucontext->mmap_xa, mmap_page);
entry             216 drivers/infiniband/hw/efa/efa_verbs.c 	if (!entry || get_mmap_key(entry) != key || entry->length != len)
entry             221 drivers/infiniband/hw/efa/efa_verbs.c 		  entry->obj, key, entry->address, entry->length);
entry             223 drivers/infiniband/hw/efa/efa_verbs.c 	return entry;
entry             233 drivers/infiniband/hw/efa/efa_verbs.c 	struct efa_mmap_entry *entry;
entry             237 drivers/infiniband/hw/efa/efa_verbs.c 	entry = kmalloc(sizeof(*entry), GFP_KERNEL);
entry             238 drivers/infiniband/hw/efa/efa_verbs.c 	if (!entry)
entry             241 drivers/infiniband/hw/efa/efa_verbs.c 	entry->obj = obj;
entry             242 drivers/infiniband/hw/efa/efa_verbs.c 	entry->address = address;
entry             243 drivers/infiniband/hw/efa/efa_verbs.c 	entry->length = length;
entry             244 drivers/infiniband/hw/efa/efa_verbs.c 	entry->mmap_flag = mmap_flag;
entry             252 drivers/infiniband/hw/efa/efa_verbs.c 	entry->mmap_page = ucontext->mmap_xa_page;
entry             254 drivers/infiniband/hw/efa/efa_verbs.c 	err = __xa_insert(&ucontext->mmap_xa, entry->mmap_page, entry,
entry             264 drivers/infiniband/hw/efa/efa_verbs.c 		entry->obj, entry->address, entry->length, get_mmap_key(entry));
entry             266 drivers/infiniband/hw/efa/efa_verbs.c 	return get_mmap_key(entry);
entry             270 drivers/infiniband/hw/efa/efa_verbs.c 	kfree(entry);
entry            1595 drivers/infiniband/hw/efa/efa_verbs.c 	struct efa_mmap_entry *entry;
entry            1600 drivers/infiniband/hw/efa/efa_verbs.c 	entry = mmap_entry_get(dev, ucontext, key, length);
entry            1601 drivers/infiniband/hw/efa/efa_verbs.c 	if (!entry) {
entry            1609 drivers/infiniband/hw/efa/efa_verbs.c 		  entry->address, length, entry->mmap_flag);
entry            1611 drivers/infiniband/hw/efa/efa_verbs.c 	pfn = entry->address >> PAGE_SHIFT;
entry            1612 drivers/infiniband/hw/efa/efa_verbs.c 	switch (entry->mmap_flag) {
entry            1637 drivers/infiniband/hw/efa/efa_verbs.c 			entry->address, length, entry->mmap_flag, err);
entry             224 drivers/infiniband/hw/hfi1/affinity.c static void node_affinity_destroy(struct hfi1_affinity_node *entry)
entry             226 drivers/infiniband/hw/hfi1/affinity.c 	free_percpu(entry->comp_vect_affinity);
entry             227 drivers/infiniband/hw/hfi1/affinity.c 	kfree(entry);
entry             233 drivers/infiniband/hw/hfi1/affinity.c 	struct hfi1_affinity_node *entry;
entry             237 drivers/infiniband/hw/hfi1/affinity.c 		entry = list_entry(pos, struct hfi1_affinity_node,
entry             240 drivers/infiniband/hw/hfi1/affinity.c 		node_affinity_destroy(entry);
entry             248 drivers/infiniband/hw/hfi1/affinity.c 	struct hfi1_affinity_node *entry;
entry             250 drivers/infiniband/hw/hfi1/affinity.c 	entry = kzalloc(sizeof(*entry), GFP_KERNEL);
entry             251 drivers/infiniband/hw/hfi1/affinity.c 	if (!entry)
entry             253 drivers/infiniband/hw/hfi1/affinity.c 	entry->node = node;
entry             254 drivers/infiniband/hw/hfi1/affinity.c 	entry->comp_vect_affinity = alloc_percpu(u16);
entry             255 drivers/infiniband/hw/hfi1/affinity.c 	INIT_LIST_HEAD(&entry->list);
entry             257 drivers/infiniband/hw/hfi1/affinity.c 	return entry;
entry             264 drivers/infiniband/hw/hfi1/affinity.c static void node_affinity_add_tail(struct hfi1_affinity_node *entry)
entry             266 drivers/infiniband/hw/hfi1/affinity.c 	list_add_tail(&entry->list, &node_affinity.list);
entry             273 drivers/infiniband/hw/hfi1/affinity.c 	struct hfi1_affinity_node *entry;
entry             276 drivers/infiniband/hw/hfi1/affinity.c 		entry = list_entry(pos, struct hfi1_affinity_node, list);
entry             277 drivers/infiniband/hw/hfi1/affinity.c 		if (entry->node == node)
entry             278 drivers/infiniband/hw/hfi1/affinity.c 			return entry;
entry             362 drivers/infiniband/hw/hfi1/affinity.c 				  struct hfi1_affinity_node *entry,
entry             387 drivers/infiniband/hw/hfi1/affinity.c 		       &entry->def_intr.used);
entry             441 drivers/infiniband/hw/hfi1/affinity.c 					  struct hfi1_affinity_node *entry)
entry             469 drivers/infiniband/hw/hfi1/affinity.c 		cpu = _dev_comp_vect_cpu_get(dd, entry, non_intr_cpus,
entry             497 drivers/infiniband/hw/hfi1/affinity.c 	struct hfi1_affinity_node *entry;
entry             500 drivers/infiniband/hw/hfi1/affinity.c 	entry = node_affinity_lookup(dd->node);
entry             501 drivers/infiniband/hw/hfi1/affinity.c 	if (!entry) {
entry             505 drivers/infiniband/hw/hfi1/affinity.c 	ret = _dev_comp_vect_mappings_create(dd, entry);
entry             534 drivers/infiniband/hw/hfi1/affinity.c 					struct hfi1_affinity_node *entry,
entry             550 drivers/infiniband/hw/hfi1/affinity.c 	if (cpumask_weight(&entry->comp_vect_mask) == 1) {
entry             556 drivers/infiniband/hw/hfi1/affinity.c 			cpumask_weight(&entry->comp_vect_mask) /
entry             565 drivers/infiniband/hw/hfi1/affinity.c 		    cpumask_weight(&entry->comp_vect_mask) %
entry             574 drivers/infiniband/hw/hfi1/affinity.c 		curr_cpu = per_cpu_affinity_get(&entry->comp_vect_mask,
entry             575 drivers/infiniband/hw/hfi1/affinity.c 						entry->comp_vect_affinity);
entry             591 drivers/infiniband/hw/hfi1/affinity.c 		per_cpu_affinity_put_max(&entry->comp_vect_mask,
entry             592 drivers/infiniband/hw/hfi1/affinity.c 					 entry->comp_vect_affinity);
entry             601 drivers/infiniband/hw/hfi1/affinity.c 					     struct hfi1_affinity_node *entry)
entry             612 drivers/infiniband/hw/hfi1/affinity.c 					       entry->comp_vect_affinity);
entry             635 drivers/infiniband/hw/hfi1/affinity.c 	struct hfi1_affinity_node *entry;
entry             655 drivers/infiniband/hw/hfi1/affinity.c 	entry = node_affinity_lookup(dd->node);
entry             661 drivers/infiniband/hw/hfi1/affinity.c 	if (!entry) {
entry             662 drivers/infiniband/hw/hfi1/affinity.c 		entry = node_affinity_allocate(node);
entry             663 drivers/infiniband/hw/hfi1/affinity.c 		if (!entry) {
entry             671 drivers/infiniband/hw/hfi1/affinity.c 		init_cpu_mask_set(&entry->def_intr);
entry             672 drivers/infiniband/hw/hfi1/affinity.c 		init_cpu_mask_set(&entry->rcv_intr);
entry             673 drivers/infiniband/hw/hfi1/affinity.c 		cpumask_clear(&entry->comp_vect_mask);
entry             674 drivers/infiniband/hw/hfi1/affinity.c 		cpumask_clear(&entry->general_intr_mask);
entry             676 drivers/infiniband/hw/hfi1/affinity.c 		cpumask_and(&entry->def_intr.mask, &node_affinity.real_cpu_mask,
entry             680 drivers/infiniband/hw/hfi1/affinity.c 		possible = cpumask_weight(&entry->def_intr.mask);
entry             681 drivers/infiniband/hw/hfi1/affinity.c 		curr_cpu = cpumask_first(&entry->def_intr.mask);
entry             685 drivers/infiniband/hw/hfi1/affinity.c 			cpumask_set_cpu(curr_cpu, &entry->rcv_intr.mask);
entry             686 drivers/infiniband/hw/hfi1/affinity.c 			cpumask_set_cpu(curr_cpu, &entry->general_intr_mask);
entry             693 drivers/infiniband/hw/hfi1/affinity.c 			cpumask_clear_cpu(curr_cpu, &entry->def_intr.mask);
entry             694 drivers/infiniband/hw/hfi1/affinity.c 			cpumask_set_cpu(curr_cpu, &entry->general_intr_mask);
entry             696 drivers/infiniband/hw/hfi1/affinity.c 						&entry->def_intr.mask);
entry             707 drivers/infiniband/hw/hfi1/affinity.c 						  &entry->def_intr.mask);
entry             709 drivers/infiniband/hw/hfi1/affinity.c 						&entry->rcv_intr.mask);
entry             711 drivers/infiniband/hw/hfi1/affinity.c 							&entry->def_intr.mask);
entry             721 drivers/infiniband/hw/hfi1/affinity.c 			if (cpumask_weight(&entry->def_intr.mask) == 0)
entry             722 drivers/infiniband/hw/hfi1/affinity.c 				cpumask_copy(&entry->def_intr.mask,
entry             723 drivers/infiniband/hw/hfi1/affinity.c 					     &entry->general_intr_mask);
entry             727 drivers/infiniband/hw/hfi1/affinity.c 		cpumask_and(&entry->comp_vect_mask,
entry             729 drivers/infiniband/hw/hfi1/affinity.c 		cpumask_andnot(&entry->comp_vect_mask,
entry             730 drivers/infiniband/hw/hfi1/affinity.c 			       &entry->comp_vect_mask,
entry             731 drivers/infiniband/hw/hfi1/affinity.c 			       &entry->rcv_intr.mask);
entry             732 drivers/infiniband/hw/hfi1/affinity.c 		cpumask_andnot(&entry->comp_vect_mask,
entry             733 drivers/infiniband/hw/hfi1/affinity.c 			       &entry->comp_vect_mask,
entry             734 drivers/infiniband/hw/hfi1/affinity.c 			       &entry->general_intr_mask);
entry             741 drivers/infiniband/hw/hfi1/affinity.c 		if (cpumask_weight(&entry->comp_vect_mask) == 0)
entry             742 drivers/infiniband/hw/hfi1/affinity.c 			cpumask_copy(&entry->comp_vect_mask,
entry             743 drivers/infiniband/hw/hfi1/affinity.c 				     &entry->general_intr_mask);
entry             746 drivers/infiniband/hw/hfi1/affinity.c 	ret = _dev_comp_vect_cpu_mask_init(dd, entry, new_entry);
entry             751 drivers/infiniband/hw/hfi1/affinity.c 		node_affinity_add_tail(entry);
entry             759 drivers/infiniband/hw/hfi1/affinity.c 		node_affinity_destroy(entry);
entry             766 drivers/infiniband/hw/hfi1/affinity.c 	struct hfi1_affinity_node *entry;
entry             772 drivers/infiniband/hw/hfi1/affinity.c 	entry = node_affinity_lookup(dd->node);
entry             773 drivers/infiniband/hw/hfi1/affinity.c 	if (!entry)
entry             780 drivers/infiniband/hw/hfi1/affinity.c 	_dev_comp_vect_cpu_mask_clean_up(dd, entry);
entry             795 drivers/infiniband/hw/hfi1/affinity.c 	struct hfi1_affinity_node *entry;
entry             803 drivers/infiniband/hw/hfi1/affinity.c 	entry = node_affinity_lookup(dd->node);
entry             804 drivers/infiniband/hw/hfi1/affinity.c 	if (!entry)
entry             820 drivers/infiniband/hw/hfi1/affinity.c 	set = &entry->def_intr;
entry             889 drivers/infiniband/hw/hfi1/affinity.c 	struct hfi1_affinity_node *entry;
entry             899 drivers/infiniband/hw/hfi1/affinity.c 	entry = node_affinity_lookup(dd->node);
entry             905 drivers/infiniband/hw/hfi1/affinity.c 		set = &entry->def_intr;
entry             908 drivers/infiniband/hw/hfi1/affinity.c 		cpu = cpumask_first(&entry->general_intr_mask);
entry             913 drivers/infiniband/hw/hfi1/affinity.c 			cpu = cpumask_first(&entry->general_intr_mask);
entry             915 drivers/infiniband/hw/hfi1/affinity.c 			set = &entry->rcv_intr;
entry             971 drivers/infiniband/hw/hfi1/affinity.c 	struct hfi1_affinity_node *entry;
entry             974 drivers/infiniband/hw/hfi1/affinity.c 	entry = node_affinity_lookup(dd->node);
entry             978 drivers/infiniband/hw/hfi1/affinity.c 		set = &entry->def_intr;
entry             988 drivers/infiniband/hw/hfi1/affinity.c 			set = &entry->rcv_intr;
entry            1040 drivers/infiniband/hw/hfi1/affinity.c 	struct hfi1_affinity_node *entry;
entry            1114 drivers/infiniband/hw/hfi1/affinity.c 	entry = node_affinity_lookup(node);
entry            1115 drivers/infiniband/hw/hfi1/affinity.c 	if (entry) {
entry            1116 drivers/infiniband/hw/hfi1/affinity.c 		cpumask_copy(intrs_mask, (entry->def_intr.gen ?
entry            1117 drivers/infiniband/hw/hfi1/affinity.c 					  &entry->def_intr.mask :
entry            1118 drivers/infiniband/hw/hfi1/affinity.c 					  &entry->def_intr.used));
entry            1119 drivers/infiniband/hw/hfi1/affinity.c 		cpumask_or(intrs_mask, intrs_mask, (entry->rcv_intr.gen ?
entry            1120 drivers/infiniband/hw/hfi1/affinity.c 						    &entry->rcv_intr.mask :
entry            1121 drivers/infiniband/hw/hfi1/affinity.c 						    &entry->rcv_intr.used));
entry            1122 drivers/infiniband/hw/hfi1/affinity.c 		cpumask_or(intrs_mask, intrs_mask, &entry->general_intr_mask);
entry            1393 drivers/infiniband/hw/hfi1/chip.c static u64 dev_access_u32_csr(const struct cntr_entry *entry,
entry            1397 drivers/infiniband/hw/hfi1/chip.c 	u64 csr = entry->csr;
entry            1399 drivers/infiniband/hw/hfi1/chip.c 	if (entry->flags & CNTR_SDMA) {
entry            1410 drivers/infiniband/hw/hfi1/chip.c static u64 access_sde_err_cnt(const struct cntr_entry *entry,
entry            1420 drivers/infiniband/hw/hfi1/chip.c static u64 access_sde_int_cnt(const struct cntr_entry *entry,
entry            1430 drivers/infiniband/hw/hfi1/chip.c static u64 access_sde_idle_int_cnt(const struct cntr_entry *entry,
entry            1440 drivers/infiniband/hw/hfi1/chip.c static u64 access_sde_progress_int_cnt(const struct cntr_entry *entry,
entry            1451 drivers/infiniband/hw/hfi1/chip.c static u64 dev_access_u64_csr(const struct cntr_entry *entry, void *context,
entry            1457 drivers/infiniband/hw/hfi1/chip.c 	u64 csr = entry->csr;
entry            1459 drivers/infiniband/hw/hfi1/chip.c 	if (entry->flags & CNTR_VL) {
entry            1472 drivers/infiniband/hw/hfi1/chip.c static u64 dc_access_lcb_cntr(const struct cntr_entry *entry, void *context,
entry            1476 drivers/infiniband/hw/hfi1/chip.c 	u32 csr = entry->csr;
entry            1496 drivers/infiniband/hw/hfi1/chip.c static u64 port_access_u32_csr(const struct cntr_entry *entry, void *context,
entry            1503 drivers/infiniband/hw/hfi1/chip.c 	return read_write_csr(ppd->dd, entry->csr, mode, data);
entry            1506 drivers/infiniband/hw/hfi1/chip.c static u64 port_access_u64_csr(const struct cntr_entry *entry,
entry            1511 drivers/infiniband/hw/hfi1/chip.c 	u64 csr = entry->csr;
entry            1513 drivers/infiniband/hw/hfi1/chip.c 	if (entry->flags & CNTR_VL) {
entry            1546 drivers/infiniband/hw/hfi1/chip.c static u64 access_sw_link_dn_cnt(const struct cntr_entry *entry, void *context,
entry            1556 drivers/infiniband/hw/hfi1/chip.c static u64 access_sw_link_up_cnt(const struct cntr_entry *entry, void *context,
entry            1566 drivers/infiniband/hw/hfi1/chip.c static u64 access_sw_unknown_frame_cnt(const struct cntr_entry *entry,
entry            1577 drivers/infiniband/hw/hfi1/chip.c static u64 access_sw_xmit_discards(const struct cntr_entry *entry,
entry            1594 drivers/infiniband/hw/hfi1/chip.c static u64 access_xmit_constraint_errs(const struct cntr_entry *entry,
entry            1607 drivers/infiniband/hw/hfi1/chip.c static u64 access_rcv_constraint_errs(const struct cntr_entry *entry,
entry            1654 drivers/infiniband/hw/hfi1/chip.c static u64 access_sw_cpu_intr(const struct cntr_entry *entry,
entry            1663 drivers/infiniband/hw/hfi1/chip.c static u64 access_sw_cpu_rcv_limit(const struct cntr_entry *entry,
entry            1672 drivers/infiniband/hw/hfi1/chip.c static u64 access_sw_pio_wait(const struct cntr_entry *entry,
entry            1680 drivers/infiniband/hw/hfi1/chip.c static u64 access_sw_pio_drain(const struct cntr_entry *entry,
entry            1688 drivers/infiniband/hw/hfi1/chip.c static u64 access_sw_ctx0_seq_drop(const struct cntr_entry *entry,
entry            1696 drivers/infiniband/hw/hfi1/chip.c static u64 access_sw_vtx_wait(const struct cntr_entry *entry,
entry            1704 drivers/infiniband/hw/hfi1/chip.c static u64 access_sw_kmem_wait(const struct cntr_entry *entry,
entry            1712 drivers/infiniband/hw/hfi1/chip.c static u64 access_sw_send_schedule(const struct cntr_entry *entry,
entry            1722 drivers/infiniband/hw/hfi1/chip.c static u64 access_misc_pll_lock_fail_err_cnt(const struct cntr_entry *entry,
entry            1731 drivers/infiniband/hw/hfi1/chip.c static u64 access_misc_mbist_fail_err_cnt(const struct cntr_entry *entry,
entry            1740 drivers/infiniband/hw/hfi1/chip.c static u64 access_misc_invalid_eep_cmd_err_cnt(const struct cntr_entry *entry,
entry            1749 drivers/infiniband/hw/hfi1/chip.c static u64 access_misc_efuse_done_parity_err_cnt(const struct cntr_entry *entry,
entry            1758 drivers/infiniband/hw/hfi1/chip.c static u64 access_misc_efuse_write_err_cnt(const struct cntr_entry *entry,
entry            1768 drivers/infiniband/hw/hfi1/chip.c 				const struct cntr_entry *entry,
entry            1776 drivers/infiniband/hw/hfi1/chip.c static u64 access_misc_efuse_csr_parity_err_cnt(const struct cntr_entry *entry,
entry            1785 drivers/infiniband/hw/hfi1/chip.c static u64 access_misc_fw_auth_failed_err_cnt(const struct cntr_entry *entry,
entry            1794 drivers/infiniband/hw/hfi1/chip.c static u64 access_misc_key_mismatch_err_cnt(const struct cntr_entry *entry,
entry            1803 drivers/infiniband/hw/hfi1/chip.c static u64 access_misc_sbus_write_failed_err_cnt(const struct cntr_entry *entry,
entry            1813 drivers/infiniband/hw/hfi1/chip.c 				const struct cntr_entry *entry,
entry            1821 drivers/infiniband/hw/hfi1/chip.c static u64 access_misc_csr_read_bad_addr_err_cnt(const struct cntr_entry *entry,
entry            1830 drivers/infiniband/hw/hfi1/chip.c static u64 access_misc_csr_parity_err_cnt(const struct cntr_entry *entry,
entry            1844 drivers/infiniband/hw/hfi1/chip.c 				const struct cntr_entry *entry,
entry            1856 drivers/infiniband/hw/hfi1/chip.c static u64 access_cce_msix_csr_parity_err_cnt(const struct cntr_entry *entry,
entry            1865 drivers/infiniband/hw/hfi1/chip.c static u64 access_cce_int_map_unc_err_cnt(const struct cntr_entry *entry,
entry            1874 drivers/infiniband/hw/hfi1/chip.c static u64 access_cce_int_map_cor_err_cnt(const struct cntr_entry *entry,
entry            1883 drivers/infiniband/hw/hfi1/chip.c static u64 access_cce_msix_table_unc_err_cnt(const struct cntr_entry *entry,
entry            1892 drivers/infiniband/hw/hfi1/chip.c static u64 access_cce_msix_table_cor_err_cnt(const struct cntr_entry *entry,
entry            1902 drivers/infiniband/hw/hfi1/chip.c 				const struct cntr_entry *entry,
entry            1911 drivers/infiniband/hw/hfi1/chip.c 				const struct cntr_entry *entry,
entry            1919 drivers/infiniband/hw/hfi1/chip.c static u64 access_cce_seg_write_bad_addr_err_cnt(const struct cntr_entry *entry,
entry            1928 drivers/infiniband/hw/hfi1/chip.c static u64 access_cce_seg_read_bad_addr_err_cnt(const struct cntr_entry *entry,
entry            1937 drivers/infiniband/hw/hfi1/chip.c static u64 access_la_triggered_cnt(const struct cntr_entry *entry,
entry            1945 drivers/infiniband/hw/hfi1/chip.c static u64 access_cce_trgt_cpl_timeout_err_cnt(const struct cntr_entry *entry,
entry            1954 drivers/infiniband/hw/hfi1/chip.c static u64 access_pcic_receive_parity_err_cnt(const struct cntr_entry *entry,
entry            1964 drivers/infiniband/hw/hfi1/chip.c 				const struct cntr_entry *entry,
entry            1973 drivers/infiniband/hw/hfi1/chip.c 				const struct cntr_entry *entry,
entry            1981 drivers/infiniband/hw/hfi1/chip.c static u64 access_pcic_cpl_dat_q_unc_err_cnt(const struct cntr_entry *entry,
entry            1990 drivers/infiniband/hw/hfi1/chip.c static u64 access_pcic_cpl_hd_q_unc_err_cnt(const struct cntr_entry *entry,
entry            1999 drivers/infiniband/hw/hfi1/chip.c static u64 access_pcic_post_dat_q_unc_err_cnt(const struct cntr_entry *entry,
entry            2008 drivers/infiniband/hw/hfi1/chip.c static u64 access_pcic_post_hd_q_unc_err_cnt(const struct cntr_entry *entry,
entry            2017 drivers/infiniband/hw/hfi1/chip.c static u64 access_pcic_retry_sot_mem_unc_err_cnt(const struct cntr_entry *entry,
entry            2026 drivers/infiniband/hw/hfi1/chip.c static u64 access_pcic_retry_mem_unc_err(const struct cntr_entry *entry,
entry            2036 drivers/infiniband/hw/hfi1/chip.c 				const struct cntr_entry *entry,
entry            2044 drivers/infiniband/hw/hfi1/chip.c static u64 access_pcic_n_post_h_q_parity_err_cnt(const struct cntr_entry *entry,
entry            2053 drivers/infiniband/hw/hfi1/chip.c static u64 access_pcic_cpl_dat_q_cor_err_cnt(const struct cntr_entry *entry,
entry            2062 drivers/infiniband/hw/hfi1/chip.c static u64 access_pcic_cpl_hd_q_cor_err_cnt(const struct cntr_entry *entry,
entry            2071 drivers/infiniband/hw/hfi1/chip.c static u64 access_pcic_post_dat_q_cor_err_cnt(const struct cntr_entry *entry,
entry            2080 drivers/infiniband/hw/hfi1/chip.c static u64 access_pcic_post_hd_q_cor_err_cnt(const struct cntr_entry *entry,
entry            2089 drivers/infiniband/hw/hfi1/chip.c static u64 access_pcic_retry_sot_mem_cor_err_cnt(const struct cntr_entry *entry,
entry            2098 drivers/infiniband/hw/hfi1/chip.c static u64 access_pcic_retry_mem_cor_err_cnt(const struct cntr_entry *entry,
entry            2108 drivers/infiniband/hw/hfi1/chip.c 				const struct cntr_entry *entry,
entry            2117 drivers/infiniband/hw/hfi1/chip.c 				const struct cntr_entry *entry,
entry            2126 drivers/infiniband/hw/hfi1/chip.c 				const struct cntr_entry *entry,
entry            2135 drivers/infiniband/hw/hfi1/chip.c 				const struct cntr_entry *entry,
entry            2144 drivers/infiniband/hw/hfi1/chip.c 				const struct cntr_entry *entry,
entry            2152 drivers/infiniband/hw/hfi1/chip.c static u64 access_cce_csr_cfg_bus_parity_err_cnt(const struct cntr_entry *entry,
entry            2162 drivers/infiniband/hw/hfi1/chip.c 				const struct cntr_entry *entry,
entry            2170 drivers/infiniband/hw/hfi1/chip.c static u64 access_cce_rspd_data_parity_err_cnt(const struct cntr_entry *entry,
entry            2179 drivers/infiniband/hw/hfi1/chip.c static u64 access_cce_trgt_access_err_cnt(const struct cntr_entry *entry,
entry            2189 drivers/infiniband/hw/hfi1/chip.c 				const struct cntr_entry *entry,
entry            2197 drivers/infiniband/hw/hfi1/chip.c static u64 access_cce_csr_write_bad_addr_err_cnt(const struct cntr_entry *entry,
entry            2206 drivers/infiniband/hw/hfi1/chip.c static u64 access_cce_csr_read_bad_addr_err_cnt(const struct cntr_entry *entry,
entry            2215 drivers/infiniband/hw/hfi1/chip.c static u64 access_ccs_csr_parity_err_cnt(const struct cntr_entry *entry,
entry            2228 drivers/infiniband/hw/hfi1/chip.c static u64 access_rx_csr_parity_err_cnt(const struct cntr_entry *entry,
entry            2237 drivers/infiniband/hw/hfi1/chip.c static u64 access_rx_csr_write_bad_addr_err_cnt(const struct cntr_entry *entry,
entry            2246 drivers/infiniband/hw/hfi1/chip.c static u64 access_rx_csr_read_bad_addr_err_cnt(const struct cntr_entry *entry,
entry            2255 drivers/infiniband/hw/hfi1/chip.c static u64 access_rx_dma_csr_unc_err_cnt(const struct cntr_entry *entry,
entry            2264 drivers/infiniband/hw/hfi1/chip.c static u64 access_rx_dma_dq_fsm_encoding_err_cnt(const struct cntr_entry *entry,
entry            2273 drivers/infiniband/hw/hfi1/chip.c static u64 access_rx_dma_eq_fsm_encoding_err_cnt(const struct cntr_entry *entry,
entry            2282 drivers/infiniband/hw/hfi1/chip.c static u64 access_rx_dma_csr_parity_err_cnt(const struct cntr_entry *entry,
entry            2291 drivers/infiniband/hw/hfi1/chip.c static u64 access_rx_rbuf_data_cor_err_cnt(const struct cntr_entry *entry,
entry            2300 drivers/infiniband/hw/hfi1/chip.c static u64 access_rx_rbuf_data_unc_err_cnt(const struct cntr_entry *entry,
entry            2310 drivers/infiniband/hw/hfi1/chip.c 				const struct cntr_entry *entry,
entry            2319 drivers/infiniband/hw/hfi1/chip.c 				const struct cntr_entry *entry,
entry            2327 drivers/infiniband/hw/hfi1/chip.c static u64 access_rx_dma_hdr_fifo_rd_cor_err_cnt(const struct cntr_entry *entry,
entry            2336 drivers/infiniband/hw/hfi1/chip.c static u64 access_rx_dma_hdr_fifo_rd_unc_err_cnt(const struct cntr_entry *entry,
entry            2345 drivers/infiniband/hw/hfi1/chip.c static u64 access_rx_rbuf_desc_part2_cor_err_cnt(const struct cntr_entry *entry,
entry            2354 drivers/infiniband/hw/hfi1/chip.c static u64 access_rx_rbuf_desc_part2_unc_err_cnt(const struct cntr_entry *entry,
entry            2363 drivers/infiniband/hw/hfi1/chip.c static u64 access_rx_rbuf_desc_part1_cor_err_cnt(const struct cntr_entry *entry,
entry            2372 drivers/infiniband/hw/hfi1/chip.c static u64 access_rx_rbuf_desc_part1_unc_err_cnt(const struct cntr_entry *entry,
entry            2381 drivers/infiniband/hw/hfi1/chip.c static u64 access_rx_hq_intr_fsm_err_cnt(const struct cntr_entry *entry,
entry            2391 drivers/infiniband/hw/hfi1/chip.c 				const struct cntr_entry *entry,
entry            2400 drivers/infiniband/hw/hfi1/chip.c 				const struct cntr_entry *entry,
entry            2409 drivers/infiniband/hw/hfi1/chip.c 				const struct cntr_entry *entry,
entry            2418 drivers/infiniband/hw/hfi1/chip.c 				const struct cntr_entry *entry,
entry            2427 drivers/infiniband/hw/hfi1/chip.c 				const struct cntr_entry *entry,
entry            2436 drivers/infiniband/hw/hfi1/chip.c 				const struct cntr_entry *entry,
entry            2445 drivers/infiniband/hw/hfi1/chip.c 				const struct cntr_entry *entry,
entry            2454 drivers/infiniband/hw/hfi1/chip.c 				const struct cntr_entry *entry,
entry            2463 drivers/infiniband/hw/hfi1/chip.c 				const struct cntr_entry *entry,
entry            2472 drivers/infiniband/hw/hfi1/chip.c 				const struct cntr_entry *entry,
entry            2481 drivers/infiniband/hw/hfi1/chip.c 				const struct cntr_entry *entry,
entry            2490 drivers/infiniband/hw/hfi1/chip.c 				const struct cntr_entry *entry,
entry            2499 drivers/infiniband/hw/hfi1/chip.c 				const struct cntr_entry *entry,
entry            2507 drivers/infiniband/hw/hfi1/chip.c static u64 access_rx_rbuf_empty_err_cnt(const struct cntr_entry *entry,
entry            2516 drivers/infiniband/hw/hfi1/chip.c static u64 access_rx_rbuf_full_err_cnt(const struct cntr_entry *entry,
entry            2525 drivers/infiniband/hw/hfi1/chip.c static u64 access_rbuf_bad_lookup_err_cnt(const struct cntr_entry *entry,
entry            2534 drivers/infiniband/hw/hfi1/chip.c static u64 access_rbuf_ctx_id_parity_err_cnt(const struct cntr_entry *entry,
entry            2543 drivers/infiniband/hw/hfi1/chip.c static u64 access_rbuf_csr_qeopdw_parity_err_cnt(const struct cntr_entry *entry,
entry            2553 drivers/infiniband/hw/hfi1/chip.c 				const struct cntr_entry *entry,
entry            2562 drivers/infiniband/hw/hfi1/chip.c 				const struct cntr_entry *entry,
entry            2571 drivers/infiniband/hw/hfi1/chip.c 				const struct cntr_entry *entry,
entry            2580 drivers/infiniband/hw/hfi1/chip.c 				const struct cntr_entry *entry,
entry            2589 drivers/infiniband/hw/hfi1/chip.c 				const struct cntr_entry *entry,
entry            2598 drivers/infiniband/hw/hfi1/chip.c 				const struct cntr_entry *entry,
entry            2607 drivers/infiniband/hw/hfi1/chip.c 				const struct cntr_entry *entry,
entry            2616 drivers/infiniband/hw/hfi1/chip.c 				const struct cntr_entry *entry,
entry            2625 drivers/infiniband/hw/hfi1/chip.c 				const struct cntr_entry *entry,
entry            2633 drivers/infiniband/hw/hfi1/chip.c static u64 access_rx_rbuf_lookup_des_cor_err_cnt(const struct cntr_entry *entry,
entry            2642 drivers/infiniband/hw/hfi1/chip.c static u64 access_rx_rbuf_lookup_des_unc_err_cnt(const struct cntr_entry *entry,
entry            2652 drivers/infiniband/hw/hfi1/chip.c 				const struct cntr_entry *entry,
entry            2661 drivers/infiniband/hw/hfi1/chip.c 				const struct cntr_entry *entry,
entry            2669 drivers/infiniband/hw/hfi1/chip.c static u64 access_rx_rbuf_free_list_cor_err_cnt(const struct cntr_entry *entry,
entry            2678 drivers/infiniband/hw/hfi1/chip.c static u64 access_rx_rbuf_free_list_unc_err_cnt(const struct cntr_entry *entry,
entry            2687 drivers/infiniband/hw/hfi1/chip.c static u64 access_rx_rcv_fsm_encoding_err_cnt(const struct cntr_entry *entry,
entry            2696 drivers/infiniband/hw/hfi1/chip.c static u64 access_rx_dma_flag_cor_err_cnt(const struct cntr_entry *entry,
entry            2705 drivers/infiniband/hw/hfi1/chip.c static u64 access_rx_dma_flag_unc_err_cnt(const struct cntr_entry *entry,
entry            2714 drivers/infiniband/hw/hfi1/chip.c static u64 access_rx_dc_sop_eop_parity_err_cnt(const struct cntr_entry *entry,
entry            2723 drivers/infiniband/hw/hfi1/chip.c static u64 access_rx_rcv_csr_parity_err_cnt(const struct cntr_entry *entry,
entry            2733 drivers/infiniband/hw/hfi1/chip.c 				const struct cntr_entry *entry,
entry            2742 drivers/infiniband/hw/hfi1/chip.c 				const struct cntr_entry *entry,
entry            2750 drivers/infiniband/hw/hfi1/chip.c static u64 access_rx_rcv_data_cor_err_cnt(const struct cntr_entry *entry,
entry            2759 drivers/infiniband/hw/hfi1/chip.c static u64 access_rx_rcv_data_unc_err_cnt(const struct cntr_entry *entry,
entry            2768 drivers/infiniband/hw/hfi1/chip.c static u64 access_rx_rcv_hdr_cor_err_cnt(const struct cntr_entry *entry,
entry            2777 drivers/infiniband/hw/hfi1/chip.c static u64 access_rx_rcv_hdr_unc_err_cnt(const struct cntr_entry *entry,
entry            2786 drivers/infiniband/hw/hfi1/chip.c static u64 access_rx_dc_intf_parity_err_cnt(const struct cntr_entry *entry,
entry            2795 drivers/infiniband/hw/hfi1/chip.c static u64 access_rx_dma_csr_cor_err_cnt(const struct cntr_entry *entry,
entry            2809 drivers/infiniband/hw/hfi1/chip.c 				const struct cntr_entry *entry,
entry            2818 drivers/infiniband/hw/hfi1/chip.c 				const struct cntr_entry *entry,
entry            2827 drivers/infiniband/hw/hfi1/chip.c 				const struct cntr_entry *entry,
entry            2836 drivers/infiniband/hw/hfi1/chip.c 				const struct cntr_entry *entry,
entry            2844 drivers/infiniband/hw/hfi1/chip.c static u64 access_pio_reserved_31_err_cnt(const struct cntr_entry *entry,
entry            2853 drivers/infiniband/hw/hfi1/chip.c static u64 access_pio_reserved_30_err_cnt(const struct cntr_entry *entry,
entry            2862 drivers/infiniband/hw/hfi1/chip.c static u64 access_pio_ppmc_sop_len_err_cnt(const struct cntr_entry *entry,
entry            2872 drivers/infiniband/hw/hfi1/chip.c 				const struct cntr_entry *entry,
entry            2880 drivers/infiniband/hw/hfi1/chip.c static u64 access_pio_vl_fifo_parity_err_cnt(const struct cntr_entry *entry,
entry            2889 drivers/infiniband/hw/hfi1/chip.c static u64 access_pio_vlf_sop_parity_err_cnt(const struct cntr_entry *entry,
entry            2898 drivers/infiniband/hw/hfi1/chip.c static u64 access_pio_vlf_v1_len_parity_err_cnt(const struct cntr_entry *entry,
entry            2908 drivers/infiniband/hw/hfi1/chip.c 				const struct cntr_entry *entry,
entry            2917 drivers/infiniband/hw/hfi1/chip.c 				const struct cntr_entry *entry,
entry            2925 drivers/infiniband/hw/hfi1/chip.c static u64 access_pio_state_machine_err_cnt(const struct cntr_entry *entry,
entry            2934 drivers/infiniband/hw/hfi1/chip.c static u64 access_pio_write_data_parity_err_cnt(const struct cntr_entry *entry,
entry            2943 drivers/infiniband/hw/hfi1/chip.c static u64 access_pio_host_addr_mem_cor_err_cnt(const struct cntr_entry *entry,
entry            2952 drivers/infiniband/hw/hfi1/chip.c static u64 access_pio_host_addr_mem_unc_err_cnt(const struct cntr_entry *entry,
entry            2962 drivers/infiniband/hw/hfi1/chip.c 				const struct cntr_entry *entry,
entry            2970 drivers/infiniband/hw/hfi1/chip.c static u64 access_pio_init_sm_in_err_cnt(const struct cntr_entry *entry,
entry            2979 drivers/infiniband/hw/hfi1/chip.c static u64 access_pio_ppmc_pbl_fifo_err_cnt(const struct cntr_entry *entry,
entry            2989 drivers/infiniband/hw/hfi1/chip.c 				const struct cntr_entry *entry,
entry            2998 drivers/infiniband/hw/hfi1/chip.c 				const struct cntr_entry *entry,
entry            3007 drivers/infiniband/hw/hfi1/chip.c 				const struct cntr_entry *entry,
entry            3016 drivers/infiniband/hw/hfi1/chip.c 				const struct cntr_entry *entry,
entry            3025 drivers/infiniband/hw/hfi1/chip.c 				const struct cntr_entry *entry,
entry            3034 drivers/infiniband/hw/hfi1/chip.c 				const struct cntr_entry *entry,
entry            3043 drivers/infiniband/hw/hfi1/chip.c 				const struct cntr_entry *entry,
entry            3052 drivers/infiniband/hw/hfi1/chip.c 				const struct cntr_entry *entry,
entry            3061 drivers/infiniband/hw/hfi1/chip.c 				const struct cntr_entry *entry,
entry            3069 drivers/infiniband/hw/hfi1/chip.c static u64 access_pio_pec_fifo_parity_err_cnt(const struct cntr_entry *entry,
entry            3078 drivers/infiniband/hw/hfi1/chip.c static u64 access_pio_pcc_fifo_parity_err_cnt(const struct cntr_entry *entry,
entry            3087 drivers/infiniband/hw/hfi1/chip.c static u64 access_pio_sb_mem_fifo1_err_cnt(const struct cntr_entry *entry,
entry            3096 drivers/infiniband/hw/hfi1/chip.c static u64 access_pio_sb_mem_fifo0_err_cnt(const struct cntr_entry *entry,
entry            3105 drivers/infiniband/hw/hfi1/chip.c static u64 access_pio_csr_parity_err_cnt(const struct cntr_entry *entry,
entry            3114 drivers/infiniband/hw/hfi1/chip.c static u64 access_pio_write_addr_parity_err_cnt(const struct cntr_entry *entry,
entry            3123 drivers/infiniband/hw/hfi1/chip.c static u64 access_pio_write_bad_ctxt_err_cnt(const struct cntr_entry *entry,
entry            3137 drivers/infiniband/hw/hfi1/chip.c 				const struct cntr_entry *entry,
entry            3146 drivers/infiniband/hw/hfi1/chip.c 				const struct cntr_entry *entry,
entry            3154 drivers/infiniband/hw/hfi1/chip.c static u64 access_sdma_csr_parity_err_cnt(const struct cntr_entry *entry,
entry            3163 drivers/infiniband/hw/hfi1/chip.c static u64 access_sdma_rpy_tag_err_cnt(const struct cntr_entry *entry,
entry            3177 drivers/infiniband/hw/hfi1/chip.c 				const struct cntr_entry *entry,
entry            3186 drivers/infiniband/hw/hfi1/chip.c 				const struct cntr_entry *entry,
entry            3194 drivers/infiniband/hw/hfi1/chip.c static u64 access_tx_egress_fifo_cor_err_cnt(const struct cntr_entry *entry,
entry            3203 drivers/infiniband/hw/hfi1/chip.c static u64 access_tx_read_pio_memory_cor_err_cnt(const struct cntr_entry *entry,
entry            3213 drivers/infiniband/hw/hfi1/chip.c 				const struct cntr_entry *entry,
entry            3221 drivers/infiniband/hw/hfi1/chip.c static u64 access_tx_sb_hdr_cor_err_cnt(const struct cntr_entry *entry,
entry            3230 drivers/infiniband/hw/hfi1/chip.c static u64 access_tx_credit_overrun_err_cnt(const struct cntr_entry *entry,
entry            3239 drivers/infiniband/hw/hfi1/chip.c static u64 access_tx_launch_fifo8_cor_err_cnt(const struct cntr_entry *entry,
entry            3248 drivers/infiniband/hw/hfi1/chip.c static u64 access_tx_launch_fifo7_cor_err_cnt(const struct cntr_entry *entry,
entry            3257 drivers/infiniband/hw/hfi1/chip.c static u64 access_tx_launch_fifo6_cor_err_cnt(const struct cntr_entry *entry,
entry            3266 drivers/infiniband/hw/hfi1/chip.c static u64 access_tx_launch_fifo5_cor_err_cnt(const struct cntr_entry *entry,
entry            3275 drivers/infiniband/hw/hfi1/chip.c static u64 access_tx_launch_fifo4_cor_err_cnt(const struct cntr_entry *entry,
entry            3284 drivers/infiniband/hw/hfi1/chip.c static u64 access_tx_launch_fifo3_cor_err_cnt(const struct cntr_entry *entry,
entry            3293 drivers/infiniband/hw/hfi1/chip.c static u64 access_tx_launch_fifo2_cor_err_cnt(const struct cntr_entry *entry,
entry            3302 drivers/infiniband/hw/hfi1/chip.c static u64 access_tx_launch_fifo1_cor_err_cnt(const struct cntr_entry *entry,
entry            3311 drivers/infiniband/hw/hfi1/chip.c static u64 access_tx_launch_fifo0_cor_err_cnt(const struct cntr_entry *entry,
entry            3320 drivers/infiniband/hw/hfi1/chip.c static u64 access_tx_credit_return_vl_err_cnt(const struct cntr_entry *entry,
entry            3329 drivers/infiniband/hw/hfi1/chip.c static u64 access_tx_hcrc_insertion_err_cnt(const struct cntr_entry *entry,
entry            3338 drivers/infiniband/hw/hfi1/chip.c static u64 access_tx_egress_fifo_unc_err_cnt(const struct cntr_entry *entry,
entry            3347 drivers/infiniband/hw/hfi1/chip.c static u64 access_tx_read_pio_memory_unc_err_cnt(const struct cntr_entry *entry,
entry            3357 drivers/infiniband/hw/hfi1/chip.c 				const struct cntr_entry *entry,
entry            3365 drivers/infiniband/hw/hfi1/chip.c static u64 access_tx_sb_hdr_unc_err_cnt(const struct cntr_entry *entry,
entry            3375 drivers/infiniband/hw/hfi1/chip.c 				const struct cntr_entry *entry,
entry            3384 drivers/infiniband/hw/hfi1/chip.c 				const struct cntr_entry *entry,
entry            3393 drivers/infiniband/hw/hfi1/chip.c 				const struct cntr_entry *entry,
entry            3402 drivers/infiniband/hw/hfi1/chip.c 				const struct cntr_entry *entry,
entry            3411 drivers/infiniband/hw/hfi1/chip.c 				const struct cntr_entry *entry,
entry            3420 drivers/infiniband/hw/hfi1/chip.c 				const struct cntr_entry *entry,
entry            3429 drivers/infiniband/hw/hfi1/chip.c 				const struct cntr_entry *entry,
entry            3438 drivers/infiniband/hw/hfi1/chip.c 				const struct cntr_entry *entry,
entry            3447 drivers/infiniband/hw/hfi1/chip.c 				const struct cntr_entry *entry,
entry            3456 drivers/infiniband/hw/hfi1/chip.c 				const struct cntr_entry *entry,
entry            3465 drivers/infiniband/hw/hfi1/chip.c 				const struct cntr_entry *entry,
entry            3474 drivers/infiniband/hw/hfi1/chip.c 				const struct cntr_entry *entry,
entry            3483 drivers/infiniband/hw/hfi1/chip.c 				const struct cntr_entry *entry,
entry            3492 drivers/infiniband/hw/hfi1/chip.c 				const struct cntr_entry *entry,
entry            3501 drivers/infiniband/hw/hfi1/chip.c 				const struct cntr_entry *entry,
entry            3510 drivers/infiniband/hw/hfi1/chip.c 				const struct cntr_entry *entry,
entry            3519 drivers/infiniband/hw/hfi1/chip.c 				const struct cntr_entry *entry,
entry            3528 drivers/infiniband/hw/hfi1/chip.c 				const struct cntr_entry *entry,
entry            3537 drivers/infiniband/hw/hfi1/chip.c 				const struct cntr_entry *entry,
entry            3546 drivers/infiniband/hw/hfi1/chip.c 				const struct cntr_entry *entry,
entry            3555 drivers/infiniband/hw/hfi1/chip.c 				const struct cntr_entry *entry,
entry            3564 drivers/infiniband/hw/hfi1/chip.c 				const struct cntr_entry *entry,
entry            3573 drivers/infiniband/hw/hfi1/chip.c 				const struct cntr_entry *entry,
entry            3582 drivers/infiniband/hw/hfi1/chip.c 				const struct cntr_entry *entry,
entry            3591 drivers/infiniband/hw/hfi1/chip.c 				const struct cntr_entry *entry,
entry            3600 drivers/infiniband/hw/hfi1/chip.c 				const struct cntr_entry *entry,
entry            3608 drivers/infiniband/hw/hfi1/chip.c static u64 access_tx_config_parity_err_cnt(const struct cntr_entry *entry,
entry            3617 drivers/infiniband/hw/hfi1/chip.c static u64 access_tx_sbrd_ctl_csr_parity_err_cnt(const struct cntr_entry *entry,
entry            3626 drivers/infiniband/hw/hfi1/chip.c static u64 access_tx_launch_csr_parity_err_cnt(const struct cntr_entry *entry,
entry            3635 drivers/infiniband/hw/hfi1/chip.c static u64 access_tx_illegal_vl_err_cnt(const struct cntr_entry *entry,
entry            3645 drivers/infiniband/hw/hfi1/chip.c 				const struct cntr_entry *entry,
entry            3653 drivers/infiniband/hw/hfi1/chip.c static u64 access_egress_reserved_10_err_cnt(const struct cntr_entry *entry,
entry            3662 drivers/infiniband/hw/hfi1/chip.c static u64 access_egress_reserved_9_err_cnt(const struct cntr_entry *entry,
entry            3672 drivers/infiniband/hw/hfi1/chip.c 				const struct cntr_entry *entry,
entry            3681 drivers/infiniband/hw/hfi1/chip.c 				const struct cntr_entry *entry,
entry            3689 drivers/infiniband/hw/hfi1/chip.c static u64 access_egress_reserved_6_err_cnt(const struct cntr_entry *entry,
entry            3699 drivers/infiniband/hw/hfi1/chip.c 				const struct cntr_entry *entry,
entry            3707 drivers/infiniband/hw/hfi1/chip.c static u64 access_tx_linkdown_err_cnt(const struct cntr_entry *entry,
entry            3717 drivers/infiniband/hw/hfi1/chip.c 				const struct cntr_entry *entry,
entry            3725 drivers/infiniband/hw/hfi1/chip.c static u64 access_egress_reserved_2_err_cnt(const struct cntr_entry *entry,
entry            3735 drivers/infiniband/hw/hfi1/chip.c 				const struct cntr_entry *entry,
entry            3744 drivers/infiniband/hw/hfi1/chip.c 				const struct cntr_entry *entry,
entry            3757 drivers/infiniband/hw/hfi1/chip.c 				const struct cntr_entry *entry,
entry            3765 drivers/infiniband/hw/hfi1/chip.c static u64 access_send_csr_read_bad_addr_err_cnt(const struct cntr_entry *entry,
entry            3774 drivers/infiniband/hw/hfi1/chip.c static u64 access_send_csr_parity_cnt(const struct cntr_entry *entry,
entry            3788 drivers/infiniband/hw/hfi1/chip.c 				const struct cntr_entry *entry,
entry            3796 drivers/infiniband/hw/hfi1/chip.c static u64 access_pio_write_overflow_err_cnt(const struct cntr_entry *entry,
entry            3806 drivers/infiniband/hw/hfi1/chip.c 				const struct cntr_entry *entry,
entry            3814 drivers/infiniband/hw/hfi1/chip.c static u64 access_pio_disallowed_packet_err_cnt(const struct cntr_entry *entry,
entry            3823 drivers/infiniband/hw/hfi1/chip.c static u64 access_pio_inconsistent_sop_err_cnt(const struct cntr_entry *entry,
entry            3837 drivers/infiniband/hw/hfi1/chip.c 				const struct cntr_entry *entry,
entry            3846 drivers/infiniband/hw/hfi1/chip.c 				const struct cntr_entry *entry,
entry            3855 drivers/infiniband/hw/hfi1/chip.c 				const struct cntr_entry *entry,
entry            3863 drivers/infiniband/hw/hfi1/chip.c static u64 access_sdma_assembly_cor_err_cnt(const struct cntr_entry *entry,
entry            3872 drivers/infiniband/hw/hfi1/chip.c static u64 access_sdma_desc_table_cor_err_cnt(const struct cntr_entry *entry,
entry            3882 drivers/infiniband/hw/hfi1/chip.c 				const struct cntr_entry *entry,
entry            3891 drivers/infiniband/hw/hfi1/chip.c 				const struct cntr_entry *entry,
entry            3900 drivers/infiniband/hw/hfi1/chip.c 				const struct cntr_entry *entry,
entry            3908 drivers/infiniband/hw/hfi1/chip.c static u64 access_sdma_assembly_unc_err_cnt(const struct cntr_entry *entry,
entry            3917 drivers/infiniband/hw/hfi1/chip.c static u64 access_sdma_desc_table_unc_err_cnt(const struct cntr_entry *entry,
entry            3926 drivers/infiniband/hw/hfi1/chip.c static u64 access_sdma_timeout_err_cnt(const struct cntr_entry *entry,
entry            3935 drivers/infiniband/hw/hfi1/chip.c static u64 access_sdma_header_length_err_cnt(const struct cntr_entry *entry,
entry            3944 drivers/infiniband/hw/hfi1/chip.c static u64 access_sdma_header_address_err_cnt(const struct cntr_entry *entry,
entry            3953 drivers/infiniband/hw/hfi1/chip.c static u64 access_sdma_header_select_err_cnt(const struct cntr_entry *entry,
entry            3962 drivers/infiniband/hw/hfi1/chip.c static u64 access_sdma_reserved_9_err_cnt(const struct cntr_entry *entry,
entry            3972 drivers/infiniband/hw/hfi1/chip.c 				const struct cntr_entry *entry,
entry            3980 drivers/infiniband/hw/hfi1/chip.c static u64 access_sdma_length_mismatch_err_cnt(const struct cntr_entry *entry,
entry            3989 drivers/infiniband/hw/hfi1/chip.c static u64 access_sdma_halt_err_cnt(const struct cntr_entry *entry,
entry            3997 drivers/infiniband/hw/hfi1/chip.c static u64 access_sdma_mem_read_err_cnt(const struct cntr_entry *entry,
entry            4006 drivers/infiniband/hw/hfi1/chip.c static u64 access_sdma_first_desc_err_cnt(const struct cntr_entry *entry,
entry            4016 drivers/infiniband/hw/hfi1/chip.c 				const struct cntr_entry *entry,
entry            4024 drivers/infiniband/hw/hfi1/chip.c static u64 access_sdma_too_long_err_cnt(const struct cntr_entry *entry,
entry            4033 drivers/infiniband/hw/hfi1/chip.c static u64 access_sdma_gen_mismatch_err_cnt(const struct cntr_entry *entry,
entry            4042 drivers/infiniband/hw/hfi1/chip.c static u64 access_sdma_wrong_dw_err_cnt(const struct cntr_entry *entry,
entry            4051 drivers/infiniband/hw/hfi1/chip.c static u64 access_dc_rcv_err_cnt(const struct cntr_entry *entry,
entry            4058 drivers/infiniband/hw/hfi1/chip.c 	u64 csr = entry->csr;
entry            4074 drivers/infiniband/hw/hfi1/chip.c static u64 access_sw_cpu_##cntr(const struct cntr_entry *entry,		      \
entry            4088 drivers/infiniband/hw/hfi1/chip.c static u64 access_ibp_##cntr(const struct cntr_entry *entry,		      \
entry            8285 drivers/infiniband/hw/hfi1/chip.c 	const struct is_table *entry;
entry            8288 drivers/infiniband/hw/hfi1/chip.c 	for (entry = &is_table[0]; entry->is_name; entry++) {
entry            8289 drivers/infiniband/hw/hfi1/chip.c 		if (source <= entry->end) {
entry            8290 drivers/infiniband/hw/hfi1/chip.c 			trace_hfi1_interrupt(dd, entry, source);
entry            8291 drivers/infiniband/hw/hfi1/chip.c 			entry->is_int(dd, source - entry->start);
entry            12076 drivers/infiniband/hw/hfi1/chip.c 		const struct cntr_entry *entry;
entry            12088 drivers/infiniband/hw/hfi1/chip.c 			entry = &dev_cntrs[i];
entry            12089 drivers/infiniband/hw/hfi1/chip.c 			hfi1_cdbg(CNTR, "reading %s", entry->name);
entry            12090 drivers/infiniband/hw/hfi1/chip.c 			if (entry->flags & CNTR_DISABLED) {
entry            12094 drivers/infiniband/hw/hfi1/chip.c 				if (entry->flags & CNTR_VL) {
entry            12097 drivers/infiniband/hw/hfi1/chip.c 						val = entry->rw_cntr(entry,
entry            12105 drivers/infiniband/hw/hfi1/chip.c 						dd->cntrs[entry->offset + j] =
entry            12108 drivers/infiniband/hw/hfi1/chip.c 				} else if (entry->flags & CNTR_SDMA) {
entry            12114 drivers/infiniband/hw/hfi1/chip.c 						entry->rw_cntr(entry, dd, j,
entry            12119 drivers/infiniband/hw/hfi1/chip.c 						dd->cntrs[entry->offset + j] =
entry            12123 drivers/infiniband/hw/hfi1/chip.c 					val = entry->rw_cntr(entry, dd,
entry            12126 drivers/infiniband/hw/hfi1/chip.c 					dd->cntrs[entry->offset] = val;
entry            12147 drivers/infiniband/hw/hfi1/chip.c 		const struct cntr_entry *entry;
entry            12154 drivers/infiniband/hw/hfi1/chip.c 			entry = &port_cntrs[i];
entry            12155 drivers/infiniband/hw/hfi1/chip.c 			hfi1_cdbg(CNTR, "reading %s", entry->name);
entry            12156 drivers/infiniband/hw/hfi1/chip.c 			if (entry->flags & CNTR_DISABLED) {
entry            12162 drivers/infiniband/hw/hfi1/chip.c 			if (entry->flags & CNTR_VL) {
entry            12165 drivers/infiniband/hw/hfi1/chip.c 					val = entry->rw_cntr(entry, ppd, j,
entry            12172 drivers/infiniband/hw/hfi1/chip.c 					ppd->cntrs[entry->offset + j] = val;
entry            12175 drivers/infiniband/hw/hfi1/chip.c 				val = entry->rw_cntr(entry, ppd,
entry            12179 drivers/infiniband/hw/hfi1/chip.c 				ppd->cntrs[entry->offset] = val;
entry            12221 drivers/infiniband/hw/hfi1/chip.c static u64 read_dev_port_cntr(struct hfi1_devdata *dd, struct cntr_entry *entry,
entry            12227 drivers/infiniband/hw/hfi1/chip.c 	if (entry->flags & CNTR_DISABLED) {
entry            12228 drivers/infiniband/hw/hfi1/chip.c 		dd_dev_err(dd, "Counter %s not enabled", entry->name);
entry            12232 drivers/infiniband/hw/hfi1/chip.c 	hfi1_cdbg(CNTR, "cntr: %s vl %d psval 0x%llx", entry->name, vl, *psval);
entry            12234 drivers/infiniband/hw/hfi1/chip.c 	val = entry->rw_cntr(entry, context, vl, CNTR_MODE_R, 0);
entry            12237 drivers/infiniband/hw/hfi1/chip.c 	if (entry->flags & CNTR_SYNTH) {
entry            12243 drivers/infiniband/hw/hfi1/chip.c 		if (entry->flags & CNTR_32BIT) {
entry            12273 drivers/infiniband/hw/hfi1/chip.c 			       struct cntr_entry *entry,
entry            12278 drivers/infiniband/hw/hfi1/chip.c 	if (entry->flags & CNTR_DISABLED) {
entry            12279 drivers/infiniband/hw/hfi1/chip.c 		dd_dev_err(dd, "Counter %s not enabled", entry->name);
entry            12283 drivers/infiniband/hw/hfi1/chip.c 	hfi1_cdbg(CNTR, "cntr: %s vl %d psval 0x%llx", entry->name, vl, *psval);
entry            12285 drivers/infiniband/hw/hfi1/chip.c 	if (entry->flags & CNTR_SYNTH) {
entry            12287 drivers/infiniband/hw/hfi1/chip.c 		if (entry->flags & CNTR_32BIT) {
entry            12288 drivers/infiniband/hw/hfi1/chip.c 			val = entry->rw_cntr(entry, context, vl, CNTR_MODE_W,
entry            12292 drivers/infiniband/hw/hfi1/chip.c 			val = entry->rw_cntr(entry, context, vl, CNTR_MODE_W,
entry            12296 drivers/infiniband/hw/hfi1/chip.c 		val = entry->rw_cntr(entry, context, vl, CNTR_MODE_W, data);
entry            12308 drivers/infiniband/hw/hfi1/chip.c 	struct cntr_entry *entry;
entry            12311 drivers/infiniband/hw/hfi1/chip.c 	entry = &dev_cntrs[index];
entry            12312 drivers/infiniband/hw/hfi1/chip.c 	sval = dd->scntrs + entry->offset;
entry            12317 drivers/infiniband/hw/hfi1/chip.c 	return read_dev_port_cntr(dd, entry, sval, dd, vl);
entry            12322 drivers/infiniband/hw/hfi1/chip.c 	struct cntr_entry *entry;
entry            12325 drivers/infiniband/hw/hfi1/chip.c 	entry = &dev_cntrs[index];
entry            12326 drivers/infiniband/hw/hfi1/chip.c 	sval = dd->scntrs + entry->offset;
entry            12331 drivers/infiniband/hw/hfi1/chip.c 	return write_dev_port_cntr(dd, entry, sval, dd, vl, data);
entry            12336 drivers/infiniband/hw/hfi1/chip.c 	struct cntr_entry *entry;
entry            12339 drivers/infiniband/hw/hfi1/chip.c 	entry = &port_cntrs[index];
entry            12340 drivers/infiniband/hw/hfi1/chip.c 	sval = ppd->scntrs + entry->offset;
entry            12351 drivers/infiniband/hw/hfi1/chip.c 	return read_dev_port_cntr(ppd->dd, entry, sval, ppd, vl);
entry            12356 drivers/infiniband/hw/hfi1/chip.c 	struct cntr_entry *entry;
entry            12359 drivers/infiniband/hw/hfi1/chip.c 	entry = &port_cntrs[index];
entry            12360 drivers/infiniband/hw/hfi1/chip.c 	sval = ppd->scntrs + entry->offset;
entry            12371 drivers/infiniband/hw/hfi1/chip.c 	return write_dev_port_cntr(ppd->dd, entry, sval, ppd, vl, data);
entry            12382 drivers/infiniband/hw/hfi1/chip.c 	struct cntr_entry *entry;
entry            12392 drivers/infiniband/hw/hfi1/chip.c 	entry = &dev_cntrs[C_DC_RCV_FLITS];
entry            12393 drivers/infiniband/hw/hfi1/chip.c 	cur_rx = entry->rw_cntr(entry, dd, CNTR_INVALID_VL, CNTR_MODE_R, 0);
entry            12395 drivers/infiniband/hw/hfi1/chip.c 	entry = &dev_cntrs[C_DC_XMIT_FLITS];
entry            12396 drivers/infiniband/hw/hfi1/chip.c 	cur_tx = entry->rw_cntr(entry, dd, CNTR_INVALID_VL, CNTR_MODE_R, 0);
entry            12426 drivers/infiniband/hw/hfi1/chip.c 			entry = &dev_cntrs[i];
entry            12427 drivers/infiniband/hw/hfi1/chip.c 			if (entry->flags & CNTR_VL) {
entry            12437 drivers/infiniband/hw/hfi1/chip.c 				entry = &port_cntrs[j];
entry            12438 drivers/infiniband/hw/hfi1/chip.c 				if (entry->flags & CNTR_VL) {
entry            12453 drivers/infiniband/hw/hfi1/chip.c 		entry = &dev_cntrs[C_DC_XMIT_FLITS];
entry            12454 drivers/infiniband/hw/hfi1/chip.c 		dd->last_tx = entry->rw_cntr(entry, dd, CNTR_INVALID_VL,
entry            12457 drivers/infiniband/hw/hfi1/chip.c 		entry = &dev_cntrs[C_DC_RCV_FLITS];
entry            12458 drivers/infiniband/hw/hfi1/chip.c 		dd->last_rx = entry->rw_cntr(entry, dd, CNTR_INVALID_VL,
entry             298 drivers/infiniband/hw/hfi1/eprom.c 	struct hfi1_eprom_table_entry *entry;
entry             342 drivers/infiniband/hw/hfi1/eprom.c 	for (entry = NULL, i = 0; i < footer->num_table_entries; i++) {
entry             344 drivers/infiniband/hw/hfi1/eprom.c 			entry = &table[i];
entry             348 drivers/infiniband/hw/hfi1/eprom.c 	if (!entry) {
entry             357 drivers/infiniband/hw/hfi1/eprom.c 	if (entry->size > (4 * 1024)) {
entry             359 drivers/infiniband/hw/hfi1/eprom.c 			   entry->size);
entry             365 drivers/infiniband/hw/hfi1/eprom.c 	if (entry->offset + entry->size < entry->offset) {
entry             368 drivers/infiniband/hw/hfi1/eprom.c 			   entry->offset, entry->size);
entry             374 drivers/infiniband/hw/hfi1/eprom.c 	buffer = kmalloc(entry->size, GFP_KERNEL);
entry             383 drivers/infiniband/hw/hfi1/eprom.c 	seg_offset = entry->offset % SEG_SIZE;
entry             384 drivers/infiniband/hw/hfi1/eprom.c 	seg_base = entry->offset - seg_offset;
entry             386 drivers/infiniband/hw/hfi1/eprom.c 	while (ncopied < entry->size) {
entry             400 drivers/infiniband/hw/hfi1/eprom.c 					   entry->offset);
entry             408 drivers/infiniband/hw/hfi1/eprom.c 		to_copy = entry->size - ncopied;
entry             439 drivers/infiniband/hw/hfi1/eprom.c 	*size = entry->size;
entry             573 drivers/infiniband/hw/hfi1/init.c 	cce = cc_state->cct.entries[max_ccti].entry;
entry            3969 drivers/infiniband/hw/hfi1/mad.c 		cc_table_attr->ccti_entries[j].entry =
entry            3970 drivers/infiniband/hw/hfi1/mad.c 			cpu_to_be16(entries[i].entry);
entry            4021 drivers/infiniband/hw/hfi1/mad.c 		entries[i].entry = be16_to_cpu(p->ccti_entries[j].entry);
entry             315 drivers/infiniband/hw/hfi1/mad.h 	__be16 entry; /* shift:2, multiplier:14 */
entry             319 drivers/infiniband/hw/hfi1/mad.h 	u16 entry; /* shift:2, multiplier:14 */
entry            1666 drivers/infiniband/hw/hfi1/tid_rdma.c u64 hfi1_access_sw_tid_wait(const struct cntr_entry *entry,
entry             241 drivers/infiniband/hw/hfi1/tid_rdma.h u64 hfi1_access_sw_tid_wait(const struct cntr_entry *entry,
entry             714 drivers/infiniband/hw/hfi1/trace_tid.h 	TP_PROTO(struct rvt_qp *qp, int index, u32 entry),
entry             715 drivers/infiniband/hw/hfi1/trace_tid.h 	TP_ARGS(qp, index, entry)
entry             732 drivers/infiniband/hw/hfi1/trace_tid.h 	TP_PROTO(struct rvt_qp *qp, int index, u32 entry),
entry             733 drivers/infiniband/hw/hfi1/trace_tid.h 	TP_ARGS(qp, index, entry)
entry             738 drivers/infiniband/hw/hfi1/trace_tid.h 	TP_PROTO(struct rvt_qp *qp, int index, u32 entry),
entry             739 drivers/infiniband/hw/hfi1/trace_tid.h 	TP_ARGS(qp, index, entry)
entry            3862 drivers/infiniband/hw/hns/hns_roce_hw_v1.c static struct hns_roce_aeqe *get_aeqe_v1(struct hns_roce_eq *eq, u32 entry)
entry            3864 drivers/infiniband/hw/hns/hns_roce_hw_v1.c 	unsigned long off = (entry & (eq->entries - 1)) *
entry            3970 drivers/infiniband/hw/hns/hns_roce_hw_v1.c static struct hns_roce_ceqe *get_ceqe_v1(struct hns_roce_eq *eq, u32 entry)
entry            3972 drivers/infiniband/hw/hns/hns_roce_hw_v1.c 	unsigned long off = (entry & (eq->entries - 1)) *
entry            1787 drivers/infiniband/hw/hns/hns_roce_hw_v2.c 	struct hns_roce_link_table_entry *entry;
entry            1806 drivers/infiniband/hw/hns/hns_roce_hw_v2.c 	entry = link_tbl->table.buf;
entry            1831 drivers/infiniband/hw/hns/hns_roce_hw_v2.c 			req_a->head_ba_l = cpu_to_le32(entry[0].blk_ba0);
entry            1833 drivers/infiniband/hw/hns/hns_roce_hw_v2.c 				cpu_to_le32(entry[0].blk_ba1_nxt_ptr);
entry            1839 drivers/infiniband/hw/hns/hns_roce_hw_v2.c 				cpu_to_le32(entry[page_num - 1].blk_ba0);
entry            1843 drivers/infiniband/hw/hns/hns_roce_hw_v2.c 				       entry[page_num - 1].blk_ba1_nxt_ptr &
entry            1848 drivers/infiniband/hw/hns/hns_roce_hw_v2.c 				       (entry[page_num - 2].blk_ba1_nxt_ptr &
entry            1864 drivers/infiniband/hw/hns/hns_roce_hw_v2.c 	struct hns_roce_link_table_entry *entry;
entry            1906 drivers/infiniband/hw/hns/hns_roce_hw_v2.c 	entry = link_tbl->table.buf;
entry            1915 drivers/infiniband/hw/hns/hns_roce_hw_v2.c 		entry[i].blk_ba0 = (u32)(t >> 12);
entry            1916 drivers/infiniband/hw/hns/hns_roce_hw_v2.c 		entry[i].blk_ba1_nxt_ptr = (u32)(t >> 44);
entry            1919 drivers/infiniband/hw/hns/hns_roce_hw_v2.c 			entry[i].blk_ba1_nxt_ptr |=
entry            4981 drivers/infiniband/hw/hns/hns_roce_hw_v2.c static struct hns_roce_aeqe *get_aeqe_v2(struct hns_roce_eq *eq, u32 entry)
entry            4987 drivers/infiniband/hw/hns/hns_roce_hw_v2.c 	off = (entry & (eq->entries - 1)) * HNS_ROCE_AEQ_ENTRY_SIZE;
entry            4993 drivers/infiniband/hw/hns/hns_roce_hw_v2.c static struct hns_roce_aeqe *mhop_get_aeqe(struct hns_roce_eq *eq, u32 entry)
entry            5000 drivers/infiniband/hw/hns/hns_roce_hw_v2.c 	off = (entry & (eq->entries - 1)) * HNS_ROCE_AEQ_ENTRY_SIZE;
entry            5111 drivers/infiniband/hw/hns/hns_roce_hw_v2.c static struct hns_roce_ceqe *get_ceqe_v2(struct hns_roce_eq *eq, u32 entry)
entry            5117 drivers/infiniband/hw/hns/hns_roce_hw_v2.c 	off = (entry & (eq->entries - 1)) * HNS_ROCE_CEQ_ENTRY_SIZE;
entry            5123 drivers/infiniband/hw/hns/hns_roce_hw_v2.c static struct hns_roce_ceqe *mhop_get_ceqe(struct hns_roce_eq *eq, u32 entry)
entry            5130 drivers/infiniband/hw/hns/hns_roce_hw_v2.c 	off = (entry & (eq->entries - 1)) * HNS_ROCE_CEQ_ENTRY_SIZE;
entry             306 drivers/infiniband/hw/i40iw/i40iw_ctrl.c static struct i40iw_sc_qp *i40iw_qp_from_entry(struct list_head *entry)
entry             308 drivers/infiniband/hw/i40iw/i40iw_ctrl.c 	if (!entry)
entry             311 drivers/infiniband/hw/i40iw/i40iw_ctrl.c 	return (struct i40iw_sc_qp *)((char *)entry - offsetof(struct i40iw_sc_qp, list));
entry             321 drivers/infiniband/hw/i40iw/i40iw_ctrl.c 	struct list_head *entry = NULL;
entry             328 drivers/infiniband/hw/i40iw/i40iw_ctrl.c 		entry = head->next;
entry             331 drivers/infiniband/hw/i40iw/i40iw_ctrl.c 		entry = (lastentry != head) ? lastentry->next : NULL;
entry             334 drivers/infiniband/hw/i40iw/i40iw_ctrl.c 	return i40iw_qp_from_entry(entry);
entry            3659 drivers/infiniband/hw/i40iw/i40iw_ctrl.c 		memcpy((char *)sdbuf->va + offset, &info->entry[3],
entry            3672 drivers/infiniband/hw/i40iw/i40iw_ctrl.c 			      (LS_64(info->entry[2].cmd, I40IW_CQPSQ_UPESD_SDCMD) |
entry            3675 drivers/infiniband/hw/i40iw/i40iw_ctrl.c 		set_64bit_val(wqe, 56, info->entry[2].data);
entry            3679 drivers/infiniband/hw/i40iw/i40iw_ctrl.c 			      (LS_64(info->entry[1].cmd, I40IW_CQPSQ_UPESD_SDCMD) |
entry            3682 drivers/infiniband/hw/i40iw/i40iw_ctrl.c 		set_64bit_val(wqe, 40, info->entry[1].data);
entry            3686 drivers/infiniband/hw/i40iw/i40iw_ctrl.c 			      LS_64(info->entry[0].cmd, I40IW_CQPSQ_UPESD_SDCMD));
entry            3688 drivers/infiniband/hw/i40iw/i40iw_ctrl.c 		set_64bit_val(wqe, 8, info->entry[0].data);
entry             115 drivers/infiniband/hw/i40iw/i40iw_hmc.c 				      struct update_sd_entry *entry)
entry             117 drivers/infiniband/hw/i40iw/i40iw_hmc.c 	entry->data = pa | (I40IW_HMC_MAX_BP_COUNT << I40E_PFHMC_SDDATALOW_PMSDBPCOUNT_SHIFT) |
entry             121 drivers/infiniband/hw/i40iw/i40iw_hmc.c 	entry->cmd = (idx | (1 << I40E_PFHMC_SDCMD_PMSDWR_SHIFT) | (1 << 15));
entry             131 drivers/infiniband/hw/i40iw/i40iw_hmc.c 				      struct update_sd_entry *entry)
entry             133 drivers/infiniband/hw/i40iw/i40iw_hmc.c 	entry->data = (I40IW_HMC_MAX_BP_COUNT <<
entry             137 drivers/infiniband/hw/i40iw/i40iw_hmc.c 	entry->cmd = (idx | (1 << I40E_PFHMC_SDCMD_PMSDWR_SHIFT) | (1 << 15));
entry             160 drivers/infiniband/hw/i40iw/i40iw_hmc.c 		i40iw_set_sd_entry(pa, sd_idx, type, sdinfo.entry);
entry             162 drivers/infiniband/hw/i40iw/i40iw_hmc.c 		i40iw_clr_sd_entry(sd_idx, type, sdinfo.entry);
entry             200 drivers/infiniband/hw/i40iw/i40iw_hmc.c 					   &sdinfo.entry[sdinfo.cnt]);
entry             203 drivers/infiniband/hw/i40iw/i40iw_hmc.c 					   &sdinfo.entry[sdinfo.cnt]);
entry             159 drivers/infiniband/hw/i40iw/i40iw_hmc.h 	struct update_sd_entry entry[I40IW_MAX_SD_ENTRIES];
entry            1457 drivers/infiniband/hw/i40iw/i40iw_main.c 		iwdev->iw_msixtbl[i].idx = ldev->msix_entries[i].entry;
entry             734 drivers/infiniband/hw/i40iw/i40iw_utils.c 	struct list_head *entry;
entry             739 drivers/infiniband/hw/i40iw/i40iw_utils.c 	entry = (void *)list->next;
entry             740 drivers/infiniband/hw/i40iw/i40iw_utils.c 	list_del(entry);
entry             741 drivers/infiniband/hw/i40iw/i40iw_utils.c 	return (void *)entry;
entry            2344 drivers/infiniband/hw/i40iw/i40iw_verbs.c 			 struct ib_wc *entry)
entry            2370 drivers/infiniband/hw/i40iw/i40iw_verbs.c 		entry->wc_flags = 0;
entry            2371 drivers/infiniband/hw/i40iw/i40iw_verbs.c 		entry->wr_id = cq_poll_info.wr_id;
entry            2373 drivers/infiniband/hw/i40iw/i40iw_verbs.c 			entry->status = IB_WC_WR_FLUSH_ERR;
entry            2374 drivers/infiniband/hw/i40iw/i40iw_verbs.c 			entry->vendor_err = cq_poll_info.major_err << 16 | cq_poll_info.minor_err;
entry            2376 drivers/infiniband/hw/i40iw/i40iw_verbs.c 			entry->status = IB_WC_SUCCESS;
entry            2381 drivers/infiniband/hw/i40iw/i40iw_verbs.c 			entry->opcode = IB_WC_RDMA_WRITE;
entry            2385 drivers/infiniband/hw/i40iw/i40iw_verbs.c 			entry->opcode = IB_WC_RDMA_READ;
entry            2391 drivers/infiniband/hw/i40iw/i40iw_verbs.c 			entry->opcode = IB_WC_SEND;
entry            2394 drivers/infiniband/hw/i40iw/i40iw_verbs.c 			entry->opcode = IB_WC_RECV;
entry            2397 drivers/infiniband/hw/i40iw/i40iw_verbs.c 			entry->opcode = IB_WC_RECV;
entry            2401 drivers/infiniband/hw/i40iw/i40iw_verbs.c 		entry->ex.imm_data = 0;
entry            2403 drivers/infiniband/hw/i40iw/i40iw_verbs.c 		entry->qp = (struct ib_qp *)qp->back_qp;
entry            2404 drivers/infiniband/hw/i40iw/i40iw_verbs.c 		entry->src_qp = cq_poll_info.qp_id;
entry            2412 drivers/infiniband/hw/i40iw/i40iw_verbs.c 		entry->byte_len = cq_poll_info.bytes_xfered;
entry            2413 drivers/infiniband/hw/i40iw/i40iw_verbs.c 		entry++;
entry             584 drivers/infiniband/hw/mlx4/alias_GUID.c 	int j, k, entry;
entry             590 drivers/infiniband/hw/mlx4/alias_GUID.c 			entry = j * NUM_ALIAS_GUID_IN_REC + k;
entry             592 drivers/infiniband/hw/mlx4/alias_GUID.c 			if (!entry || entry > dev->dev->persist->num_vfs ||
entry             593 drivers/infiniband/hw/mlx4/alias_GUID.c 			    !mlx4_is_slave_active(dev->dev, entry))
entry             595 drivers/infiniband/hw/mlx4/alias_GUID.c 			guid = mlx4_get_admin_guid(dev->dev, entry, port);
entry             600 drivers/infiniband/hw/mlx4/alias_GUID.c 				 entry,
entry             249 drivers/infiniband/hw/mlx4/main.c static void free_gid_entry(struct gid_entry *entry)
entry             251 drivers/infiniband/hw/mlx4/main.c 	memset(&entry->gid, 0, sizeof(entry->gid));
entry             252 drivers/infiniband/hw/mlx4/main.c 	kfree(entry->ctx);
entry             253 drivers/infiniband/hw/mlx4/main.c 	entry->ctx = NULL;
entry             686 drivers/infiniband/hw/mlx4/sysfs.c 	list_add_tail(&p->kobj.entry, &dev->pkeys.pkey_port_list[slave]);
entry             748 drivers/infiniband/hw/mlx4/sysfs.c 				 entry) {
entry             749 drivers/infiniband/hw/mlx4/sysfs.c 		list_del(&p->entry);
entry             793 drivers/infiniband/hw/mlx4/sysfs.c 					 entry) {
entry             794 drivers/infiniband/hw/mlx4/sysfs.c 			list_del(&p->entry);
entry            2429 drivers/infiniband/hw/mlx5/devx.c 	void *entry;
entry            2434 drivers/infiniband/hw/mlx5/devx.c 	xa_for_each(&table->event_xa, id, entry) {
entry            2435 drivers/infiniband/hw/mlx5/devx.c 		event = entry;
entry            2439 drivers/infiniband/hw/mlx5/devx.c 		kfree(entry);
entry            2504 drivers/infiniband/hw/mlx5/devx.c 	struct devx_async_data *entry, *tmp;
entry            2507 drivers/infiniband/hw/mlx5/devx.c 	list_for_each_entry_safe(entry, tmp,
entry            2509 drivers/infiniband/hw/mlx5/devx.c 		kvfree(entry);
entry            2647 drivers/infiniband/hw/mlx5/devx.c 	struct devx_async_event_data *entry, *tmp;
entry            2668 drivers/infiniband/hw/mlx5/devx.c 		list_for_each_entry_safe(entry, tmp,
entry            2670 drivers/infiniband/hw/mlx5/devx.c 			kfree(entry); /* read can't come any more */
entry              57 drivers/infiniband/hw/mlx5/mem.c 	int entry;
entry              65 drivers/infiniband/hw/mlx5/mem.c 	for_each_sg(umem->sg_head.sgl, sg, umem->nmap, entry) {
entry             139 drivers/infiniband/hw/mlx5/mem.c 	int entry;
entry             155 drivers/infiniband/hw/mlx5/mem.c 	for_each_sg(umem->sg_head.sgl, sg, umem->nmap, entry) {
entry            1225 drivers/infiniband/hw/mlx5/mlx5_ib.h struct mlx5_ib_mr *mlx5_mr_cache_alloc(struct mlx5_ib_dev *dev, int entry);
entry             422 drivers/infiniband/hw/mlx5/mr.c struct mlx5_ib_mr *mlx5_mr_cache_alloc(struct mlx5_ib_dev *dev, int entry)
entry             429 drivers/infiniband/hw/mlx5/mr.c 	if (entry < 0 || entry >= MAX_MR_CACHE_ENTRIES) {
entry             430 drivers/infiniband/hw/mlx5/mr.c 		mlx5_ib_err(dev, "cache entry %d is out of range\n", entry);
entry             434 drivers/infiniband/hw/mlx5/mr.c 	ent = &cache->ent[entry];
entry             440 drivers/infiniband/hw/mlx5/mr.c 			err = add_keys(dev, entry, 1);
entry             160 drivers/infiniband/hw/mthca/mthca_cq.c 						 int entry)
entry             163 drivers/infiniband/hw/mthca/mthca_cq.c 		return buf->queue.direct.buf + (entry * MTHCA_CQ_ENTRY_SIZE);
entry             165 drivers/infiniband/hw/mthca/mthca_cq.c 		return buf->queue.page_list[entry * MTHCA_CQ_ENTRY_SIZE / PAGE_SIZE].buf
entry             166 drivers/infiniband/hw/mthca/mthca_cq.c 			+ (entry * MTHCA_CQ_ENTRY_SIZE) % PAGE_SIZE;
entry             169 drivers/infiniband/hw/mthca/mthca_cq.c static inline struct mthca_cqe *get_cqe(struct mthca_cq *cq, int entry)
entry             171 drivers/infiniband/hw/mthca/mthca_cq.c 	return get_cqe_from_buf(&cq->buf, entry);
entry             375 drivers/infiniband/hw/mthca/mthca_cq.c 			     struct ib_wc *entry, int *free_cqe)
entry             394 drivers/infiniband/hw/mthca/mthca_cq.c 		entry->status = IB_WC_LOC_LEN_ERR;
entry             397 drivers/infiniband/hw/mthca/mthca_cq.c 		entry->status = IB_WC_LOC_QP_OP_ERR;
entry             400 drivers/infiniband/hw/mthca/mthca_cq.c 		entry->status = IB_WC_LOC_EEC_OP_ERR;
entry             403 drivers/infiniband/hw/mthca/mthca_cq.c 		entry->status = IB_WC_LOC_PROT_ERR;
entry             406 drivers/infiniband/hw/mthca/mthca_cq.c 		entry->status = IB_WC_WR_FLUSH_ERR;
entry             409 drivers/infiniband/hw/mthca/mthca_cq.c 		entry->status = IB_WC_MW_BIND_ERR;
entry             412 drivers/infiniband/hw/mthca/mthca_cq.c 		entry->status = IB_WC_BAD_RESP_ERR;
entry             415 drivers/infiniband/hw/mthca/mthca_cq.c 		entry->status = IB_WC_LOC_ACCESS_ERR;
entry             418 drivers/infiniband/hw/mthca/mthca_cq.c 		entry->status = IB_WC_REM_INV_REQ_ERR;
entry             421 drivers/infiniband/hw/mthca/mthca_cq.c 		entry->status = IB_WC_REM_ACCESS_ERR;
entry             424 drivers/infiniband/hw/mthca/mthca_cq.c 		entry->status = IB_WC_REM_OP_ERR;
entry             427 drivers/infiniband/hw/mthca/mthca_cq.c 		entry->status = IB_WC_RETRY_EXC_ERR;
entry             430 drivers/infiniband/hw/mthca/mthca_cq.c 		entry->status = IB_WC_RNR_RETRY_EXC_ERR;
entry             433 drivers/infiniband/hw/mthca/mthca_cq.c 		entry->status = IB_WC_LOC_RDD_VIOL_ERR;
entry             436 drivers/infiniband/hw/mthca/mthca_cq.c 		entry->status = IB_WC_REM_INV_RD_REQ_ERR;
entry             439 drivers/infiniband/hw/mthca/mthca_cq.c 		entry->status = IB_WC_REM_ABORT_ERR;
entry             442 drivers/infiniband/hw/mthca/mthca_cq.c 		entry->status = IB_WC_INV_EECN_ERR;
entry             445 drivers/infiniband/hw/mthca/mthca_cq.c 		entry->status = IB_WC_INV_EEC_STATE_ERR;
entry             448 drivers/infiniband/hw/mthca/mthca_cq.c 		entry->status = IB_WC_GENERAL_ERR;
entry             452 drivers/infiniband/hw/mthca/mthca_cq.c 	entry->vendor_err = cqe->vendor_err;
entry             482 drivers/infiniband/hw/mthca/mthca_cq.c 				 struct ib_wc *entry)
entry             531 drivers/infiniband/hw/mthca/mthca_cq.c 	entry->qp = &(*cur_qp)->ibqp;
entry             537 drivers/infiniband/hw/mthca/mthca_cq.c 		entry->wr_id = (*cur_qp)->wrid[wqe_index +
entry             544 drivers/infiniband/hw/mthca/mthca_cq.c 		entry->wr_id = srq->wrid[wqe_index];
entry             558 drivers/infiniband/hw/mthca/mthca_cq.c 		entry->wr_id = (*cur_qp)->wrid[wqe_index];
entry             573 drivers/infiniband/hw/mthca/mthca_cq.c 				 entry, &free_cqe);
entry             578 drivers/infiniband/hw/mthca/mthca_cq.c 		entry->wc_flags = 0;
entry             581 drivers/infiniband/hw/mthca/mthca_cq.c 			entry->opcode    = IB_WC_RDMA_WRITE;
entry             584 drivers/infiniband/hw/mthca/mthca_cq.c 			entry->opcode    = IB_WC_RDMA_WRITE;
entry             585 drivers/infiniband/hw/mthca/mthca_cq.c 			entry->wc_flags |= IB_WC_WITH_IMM;
entry             588 drivers/infiniband/hw/mthca/mthca_cq.c 			entry->opcode    = IB_WC_SEND;
entry             591 drivers/infiniband/hw/mthca/mthca_cq.c 			entry->opcode    = IB_WC_SEND;
entry             592 drivers/infiniband/hw/mthca/mthca_cq.c 			entry->wc_flags |= IB_WC_WITH_IMM;
entry             595 drivers/infiniband/hw/mthca/mthca_cq.c 			entry->opcode    = IB_WC_RDMA_READ;
entry             596 drivers/infiniband/hw/mthca/mthca_cq.c 			entry->byte_len  = be32_to_cpu(cqe->byte_cnt);
entry             599 drivers/infiniband/hw/mthca/mthca_cq.c 			entry->opcode    = IB_WC_COMP_SWAP;
entry             600 drivers/infiniband/hw/mthca/mthca_cq.c 			entry->byte_len  = MTHCA_ATOMIC_BYTE_LEN;
entry             603 drivers/infiniband/hw/mthca/mthca_cq.c 			entry->opcode    = IB_WC_FETCH_ADD;
entry             604 drivers/infiniband/hw/mthca/mthca_cq.c 			entry->byte_len  = MTHCA_ATOMIC_BYTE_LEN;
entry             607 drivers/infiniband/hw/mthca/mthca_cq.c 			entry->opcode    = MTHCA_OPCODE_INVALID;
entry             611 drivers/infiniband/hw/mthca/mthca_cq.c 		entry->byte_len = be32_to_cpu(cqe->byte_cnt);
entry             615 drivers/infiniband/hw/mthca/mthca_cq.c 			entry->wc_flags = IB_WC_WITH_IMM;
entry             616 drivers/infiniband/hw/mthca/mthca_cq.c 			entry->ex.imm_data = cqe->imm_etype_pkey_eec;
entry             617 drivers/infiniband/hw/mthca/mthca_cq.c 			entry->opcode = IB_WC_RECV;
entry             621 drivers/infiniband/hw/mthca/mthca_cq.c 			entry->wc_flags = IB_WC_WITH_IMM;
entry             622 drivers/infiniband/hw/mthca/mthca_cq.c 			entry->ex.imm_data = cqe->imm_etype_pkey_eec;
entry             623 drivers/infiniband/hw/mthca/mthca_cq.c 			entry->opcode = IB_WC_RECV_RDMA_WITH_IMM;
entry             626 drivers/infiniband/hw/mthca/mthca_cq.c 			entry->wc_flags = 0;
entry             627 drivers/infiniband/hw/mthca/mthca_cq.c 			entry->opcode = IB_WC_RECV;
entry             630 drivers/infiniband/hw/mthca/mthca_cq.c 		entry->slid 	   = be16_to_cpu(cqe->rlid);
entry             631 drivers/infiniband/hw/mthca/mthca_cq.c 		entry->sl   	   = cqe->sl_ipok >> 4;
entry             632 drivers/infiniband/hw/mthca/mthca_cq.c 		entry->src_qp 	   = be32_to_cpu(cqe->rqpn) & 0xffffff;
entry             633 drivers/infiniband/hw/mthca/mthca_cq.c 		entry->dlid_path_bits = cqe->g_mlpath & 0x7f;
entry             634 drivers/infiniband/hw/mthca/mthca_cq.c 		entry->pkey_index  = be32_to_cpu(cqe->imm_etype_pkey_eec) >> 16;
entry             635 drivers/infiniband/hw/mthca/mthca_cq.c 		entry->wc_flags   |= cqe->g_mlpath & 0x80 ? IB_WC_GRH : 0;
entry             638 drivers/infiniband/hw/mthca/mthca_cq.c 		entry->wc_flags	  |=  (cqe->sl_ipok & 1 && checksum == 0xffff) ?
entry             642 drivers/infiniband/hw/mthca/mthca_cq.c 	entry->status = IB_WC_SUCCESS;
entry             655 drivers/infiniband/hw/mthca/mthca_cq.c 		  struct ib_wc *entry)
entry             671 drivers/infiniband/hw/mthca/mthca_cq.c 				     &freed, entry + npolled);
entry             495 drivers/infiniband/hw/mthca/mthca_dev.h 		  struct ib_wc *entry);
entry             228 drivers/infiniband/hw/mthca/mthca_eq.c static inline struct mthca_eqe *get_eqe(struct mthca_eq *eq, u32 entry)
entry             230 drivers/infiniband/hw/mthca/mthca_eq.c 	unsigned long off = (entry & (eq->nent - 1)) * MTHCA_EQ_ENTRY_SIZE;
entry             287 drivers/infiniband/hw/ocrdma/ocrdma.h 	struct list_head entry;
entry             454 drivers/infiniband/hw/ocrdma/ocrdma.h 	struct list_head entry;
entry             209 drivers/infiniband/hw/ocrdma/ocrdma_verbs.c 	INIT_LIST_HEAD(&mm->entry);
entry             212 drivers/infiniband/hw/ocrdma/ocrdma_verbs.c 	list_add_tail(&mm->entry, &uctx->mm_head);
entry             223 drivers/infiniband/hw/ocrdma/ocrdma_verbs.c 	list_for_each_entry_safe(mm, tmp, &uctx->mm_head, entry) {
entry             227 drivers/infiniband/hw/ocrdma/ocrdma_verbs.c 		list_del(&mm->entry);
entry             241 drivers/infiniband/hw/ocrdma/ocrdma_verbs.c 	list_for_each_entry(mm, &uctx->mm_head, entry) {
entry             539 drivers/infiniband/hw/ocrdma/ocrdma_verbs.c 	list_for_each_entry_safe(mm, tmp, &uctx->mm_head, entry) {
entry             540 drivers/infiniband/hw/ocrdma/ocrdma_verbs.c 		list_del(&mm->entry);
entry             309 drivers/infiniband/hw/qedr/qedr.h 	struct list_head entry;
entry             278 drivers/infiniband/hw/qedr/verbs.c 	INIT_LIST_HEAD(&mm->entry);
entry             281 drivers/infiniband/hw/qedr/verbs.c 	list_add(&mm->entry, &uctx->mm_head);
entry             299 drivers/infiniband/hw/qedr/verbs.c 	list_for_each_entry(mm, &uctx->mm_head, entry) {
entry             378 drivers/infiniband/hw/qedr/verbs.c 	list_for_each_entry_safe(mm, tmp, &uctx->mm_head, entry) {
entry             382 drivers/infiniband/hw/qedr/verbs.c 		list_del(&mm->entry);
entry            2205 drivers/infiniband/hw/qib/qib_mad.c 		p->ccti_entries[i].entry = cpu_to_be16(entries[i].entry);
entry            2278 drivers/infiniband/hw/qib/qib_mad.c 		entries[i].entry = be16_to_cpu(p->ccti_entries[i].entry);
entry             262 drivers/infiniband/hw/qib/qib_mad.h 	__be16 entry; /* shift:2, multiplier:14 */
entry             266 drivers/infiniband/hw/qib/qib_mad.h 	u16 entry; /* shift:2, multiplier:14 */
entry              67 drivers/infiniband/sw/rdmavt/cq.c bool rvt_cq_enter(struct rvt_cq *cq, struct ib_wc *entry, bool solicited)
entry             120 drivers/infiniband/sw/rdmavt/cq.c 	trace_rvt_cq_enter(cq, entry, head);
entry             122 drivers/infiniband/sw/rdmavt/cq.c 		uqueue[head].wr_id = entry->wr_id;
entry             123 drivers/infiniband/sw/rdmavt/cq.c 		uqueue[head].status = entry->status;
entry             124 drivers/infiniband/sw/rdmavt/cq.c 		uqueue[head].opcode = entry->opcode;
entry             125 drivers/infiniband/sw/rdmavt/cq.c 		uqueue[head].vendor_err = entry->vendor_err;
entry             126 drivers/infiniband/sw/rdmavt/cq.c 		uqueue[head].byte_len = entry->byte_len;
entry             127 drivers/infiniband/sw/rdmavt/cq.c 		uqueue[head].ex.imm_data = entry->ex.imm_data;
entry             128 drivers/infiniband/sw/rdmavt/cq.c 		uqueue[head].qp_num = entry->qp->qp_num;
entry             129 drivers/infiniband/sw/rdmavt/cq.c 		uqueue[head].src_qp = entry->src_qp;
entry             130 drivers/infiniband/sw/rdmavt/cq.c 		uqueue[head].wc_flags = entry->wc_flags;
entry             131 drivers/infiniband/sw/rdmavt/cq.c 		uqueue[head].pkey_index = entry->pkey_index;
entry             132 drivers/infiniband/sw/rdmavt/cq.c 		uqueue[head].slid = ib_lid_cpu16(entry->slid);
entry             133 drivers/infiniband/sw/rdmavt/cq.c 		uqueue[head].sl = entry->sl;
entry             134 drivers/infiniband/sw/rdmavt/cq.c 		uqueue[head].dlid_path_bits = entry->dlid_path_bits;
entry             135 drivers/infiniband/sw/rdmavt/cq.c 		uqueue[head].port_num = entry->port_num;
entry             139 drivers/infiniband/sw/rdmavt/cq.c 		kqueue[head] = *entry;
entry             145 drivers/infiniband/sw/rdmavt/cq.c 	     (solicited || entry->status != IB_WC_SUCCESS))) {
entry             517 drivers/infiniband/sw/rdmavt/cq.c int rvt_poll_cq(struct ib_cq *ibcq, int num_entries, struct ib_wc *entry)
entry             535 drivers/infiniband/sw/rdmavt/cq.c 	for (npolled = 0; npolled < num_entries; ++npolled, ++entry) {
entry             540 drivers/infiniband/sw/rdmavt/cq.c 		*entry = wc->kqueue[tail];
entry              59 drivers/infiniband/sw/rdmavt/cq.h int rvt_poll_cq(struct ib_cq *ibcq, int num_entries, struct ib_wc *entry);
entry             245 drivers/infiniband/sw/rdmavt/qp.c 	int entry;
entry             272 drivers/infiniband/sw/rdmavt/qp.c 		entry = (atomic_inc_return(&wss->clean_entry) - 1)
entry             276 drivers/infiniband/sw/rdmavt/qp.c 		bits = xchg(&wss->entries[entry], 0);
entry             290 drivers/infiniband/sw/rdmavt/qp.c 	u32 entry = page / BITS_PER_LONG; /* assumes this ends up a shift */
entry             293 drivers/infiniband/sw/rdmavt/qp.c 	if (!test_and_set_bit(nr, &wss->entries[entry]))
entry             138 drivers/infiniband/sw/siw/siw_verbs.c 	void *entry;
entry             146 drivers/infiniband/sw/siw/siw_verbs.c 	xa_for_each(&uctx->xa, index, entry) {
entry             169 drivers/infiniband/ulp/opa_vnic/opa_vnic_encap.c 		struct __opa_vnic_mactable_entry *nentry = &node->entry;
entry             170 drivers/infiniband/ulp/opa_vnic/opa_vnic_encap.c 		struct opa_veswport_mactable_entry *entry;
entry             177 drivers/infiniband/ulp/opa_vnic/opa_vnic_encap.c 		entry = &tbl->tbl_entries[node->index - loffset];
entry             178 drivers/infiniband/ulp/opa_vnic/opa_vnic_encap.c 		memcpy(entry->mac_addr, nentry->mac_addr,
entry             179 drivers/infiniband/ulp/opa_vnic/opa_vnic_encap.c 		       ARRAY_SIZE(entry->mac_addr));
entry             180 drivers/infiniband/ulp/opa_vnic/opa_vnic_encap.c 		memcpy(entry->mac_addr_mask, nentry->mac_addr_mask,
entry             181 drivers/infiniband/ulp/opa_vnic/opa_vnic_encap.c 		       ARRAY_SIZE(entry->mac_addr_mask));
entry             182 drivers/infiniband/ulp/opa_vnic/opa_vnic_encap.c 		entry->dlid_sd = cpu_to_be32(nentry->dlid_sd);
entry             227 drivers/infiniband/ulp/opa_vnic/opa_vnic_encap.c 		struct opa_veswport_mactable_entry *entry =
entry             229 drivers/infiniband/ulp/opa_vnic/opa_vnic_encap.c 		u8 *mac_addr = entry->mac_addr;
entry             235 drivers/infiniband/ulp/opa_vnic/opa_vnic_encap.c 		      entry->dlid_sd);
entry             248 drivers/infiniband/ulp/opa_vnic/opa_vnic_encap.c 		nentry = &node->entry;
entry             249 drivers/infiniband/ulp/opa_vnic/opa_vnic_encap.c 		memcpy(nentry->mac_addr, entry->mac_addr,
entry             251 drivers/infiniband/ulp/opa_vnic/opa_vnic_encap.c 		memcpy(nentry->mac_addr_mask, entry->mac_addr_mask,
entry             253 drivers/infiniband/ulp/opa_vnic/opa_vnic_encap.c 		nentry->dlid_sd = be32_to_cpu(entry->dlid_sd);
entry             254 drivers/infiniband/ulp/opa_vnic/opa_vnic_encap.c 		key = node->entry.mac_addr[OPA_VNIC_MAC_HASH_IDX];
entry             275 drivers/infiniband/ulp/opa_vnic/opa_vnic_encap.c 		memcpy(&new_node->entry, &node->entry, sizeof(node->entry));
entry             276 drivers/infiniband/ulp/opa_vnic/opa_vnic_encap.c 		key = new_node->entry.mac_addr[OPA_VNIC_MAC_HASH_IDX];
entry             313 drivers/infiniband/ulp/opa_vnic/opa_vnic_encap.c 		struct __opa_vnic_mactable_entry *entry = &node->entry;
entry             316 drivers/infiniband/ulp/opa_vnic/opa_vnic_encap.c 		if (unlikely(OPA_VNIC_DLID_SD_IS_SRC_MAC(entry->dlid_sd)))
entry             319 drivers/infiniband/ulp/opa_vnic/opa_vnic_encap.c 		if (!memcmp(node->entry.mac_addr, mac_hdr->h_dest,
entry             320 drivers/infiniband/ulp/opa_vnic/opa_vnic_encap.c 			    ARRAY_SIZE(node->entry.mac_addr))) {
entry             322 drivers/infiniband/ulp/opa_vnic/opa_vnic_encap.c 			dlid = OPA_VNIC_DLID_SD_GET_DLID(node->entry.dlid_sd);
entry             443 drivers/infiniband/ulp/opa_vnic/opa_vnic_encap.h 	struct opa_vnic_iface_mac_entry entry[0];
entry             242 drivers/infiniband/ulp/opa_vnic/opa_vnic_internal.h 	struct __opa_vnic_mactable_entry     entry;
entry             323 drivers/infiniband/ulp/opa_vnic/opa_vnic_vema_iface.c 		struct opa_vnic_iface_mac_entry *entry = &macs->entry[count];
entry             329 drivers/infiniband/ulp/opa_vnic/opa_vnic_vema_iface.c 		memcpy(entry, ha->addr, sizeof(*entry));
entry             356 drivers/infiniband/ulp/opa_vnic/opa_vnic_vema_iface.c 		struct opa_vnic_iface_mac_entry *entry = &macs->entry[count];
entry             369 drivers/infiniband/ulp/opa_vnic/opa_vnic_vema_iface.c 		memcpy(entry, ha->addr, sizeof(*entry));
entry             375 drivers/infiniband/ulp/opa_vnic/opa_vnic_vema_iface.c 		struct opa_vnic_iface_mac_entry *entry = &macs->entry[count];
entry             381 drivers/infiniband/ulp/opa_vnic/opa_vnic_vema_iface.c 		memcpy(entry, ha->addr, sizeof(*entry));
entry             471 drivers/infiniband/ulp/srp/ib_srp.c 		list_add_tail(&d->entry, &pool->free_list);
entry             496 drivers/infiniband/ulp/srp/ib_srp.c 		d = list_first_entry(&pool->free_list, typeof(*d), entry);
entry             497 drivers/infiniband/ulp/srp/ib_srp.c 		list_del(&d->entry);
entry             521 drivers/infiniband/ulp/srp/ib_srp.c 		list_add(&desc[i]->entry, &pool->free_list);
entry             288 drivers/infiniband/ulp/srp/ib_srp.h 	struct list_head		entry;
entry            1965 drivers/infiniband/ulp/srpt/ib_srpt.c 	list_for_each_entry(nexus, &sport->nexus_list, entry) {
entry            2005 drivers/infiniband/ulp/srpt/ib_srpt.c 	list_for_each_entry(nexus, &sport->nexus_list, entry) {
entry            2029 drivers/infiniband/ulp/srpt/ib_srpt.c 		list_for_each_entry(n, &sport->nexus_list, entry) {
entry            2037 drivers/infiniband/ulp/srpt/ib_srpt.c 			list_add_tail_rcu(&tmp_nexus->entry,
entry            2939 drivers/infiniband/ulp/srpt/ib_srpt.c 	list_for_each_entry(nexus, &sport->nexus_list, entry)
entry            2967 drivers/infiniband/ulp/srpt/ib_srpt.c 		list_for_each_entry(nexus, &sport->nexus_list, entry) {
entry            2978 drivers/infiniband/ulp/srpt/ib_srpt.c 	list_for_each_entry_safe(nexus, next_n, &sport->nexus_list, entry) {
entry            2979 drivers/infiniband/ulp/srpt/ib_srpt.c 		list_del(&nexus->entry);
entry             343 drivers/infiniband/ulp/srpt/ib_srpt.h 	struct list_head	entry;
entry            1294 drivers/input/input.c 	struct proc_dir_entry *entry;
entry            1300 drivers/input/input.c 	entry = proc_create("devices", 0, proc_bus_input_dir,
entry            1302 drivers/input/input.c 	if (!entry)
entry            1305 drivers/input/input.c 	entry = proc_create("handlers", 0, proc_bus_input_dir,
entry            1307 drivers/input/input.c 	if (!entry)
entry            1179 drivers/input/misc/wistron_btns.c 					  struct key_entry *entry)
entry            1181 drivers/input/misc/wistron_btns.c 	switch (entry->type) {
entry            1186 drivers/input/misc/wistron_btns.c 			entry->type = KE_KEY;
entry            1187 drivers/input/misc/wistron_btns.c 			entry->keycode = KEY_WLAN;
entry            1193 drivers/input/misc/wistron_btns.c 			entry->type = KE_KEY;
entry            1194 drivers/input/misc/wistron_btns.c 			entry->keycode = KEY_BLUETOOTH;
entry            1199 drivers/input/misc/wistron_btns.c 		if (entry->code & FE_UNTESTED)
entry             101 drivers/input/rmi4/rmi_driver.c 	struct rmi_function *entry;
entry             104 drivers/input/rmi4/rmi_driver.c 	list_for_each_entry(entry, &data->function_list, node) {
entry             105 drivers/input/rmi4/rmi_driver.c 		retval = reset_one_function(entry);
entry             116 drivers/input/rmi4/rmi_driver.c 	struct rmi_function *entry;
entry             119 drivers/input/rmi4/rmi_driver.c 	list_for_each_entry(entry, &data->function_list, node) {
entry             120 drivers/input/rmi4/rmi_driver.c 		retval = configure_one_function(entry);
entry             246 drivers/input/rmi4/rmi_driver.c 	struct rmi_function *entry;
entry             248 drivers/input/rmi4/rmi_driver.c 	list_for_each_entry(entry, &data->function_list, node) {
entry             249 drivers/input/rmi4/rmi_driver.c 		if (entry->fd.function_number == number)
entry             250 drivers/input/rmi4/rmi_driver.c 			return entry;
entry             278 drivers/input/rmi4/rmi_driver.c 	struct rmi_function *entry;
entry             281 drivers/input/rmi4/rmi_driver.c 	list_for_each_entry(entry, &data->function_list, node) {
entry             282 drivers/input/rmi4/rmi_driver.c 		retval = suspend_one_function(entry);
entry             312 drivers/input/rmi4/rmi_driver.c 	struct rmi_function *entry;
entry             315 drivers/input/rmi4/rmi_driver.c 	list_for_each_entry(entry, &data->function_list, node) {
entry             316 drivers/input/rmi4/rmi_driver.c 		retval = resume_one_function(entry);
entry             460 drivers/input/rmi4/rmi_driver.c 			      struct pdt_entry *entry, u16 pdt_address)
entry             472 drivers/input/rmi4/rmi_driver.c 	entry->page_start = pdt_address & RMI4_PAGE_MASK;
entry             473 drivers/input/rmi4/rmi_driver.c 	entry->query_base_addr = buf[0];
entry             474 drivers/input/rmi4/rmi_driver.c 	entry->command_base_addr = buf[1];
entry             475 drivers/input/rmi4/rmi_driver.c 	entry->control_base_addr = buf[2];
entry             476 drivers/input/rmi4/rmi_driver.c 	entry->data_base_addr = buf[3];
entry             477 drivers/input/rmi4/rmi_driver.c 	entry->interrupt_source_count = buf[4] & RMI_PDT_INT_SOURCE_COUNT_MASK;
entry             478 drivers/input/rmi4/rmi_driver.c 	entry->function_version = (buf[4] & RMI_PDT_FUNCTION_VERSION_MASK) >> 5;
entry             479 drivers/input/rmi4/rmi_driver.c 	entry->function_number = buf[5];
entry             505 drivers/input/rmi4/rmi_driver.c 					     const struct pdt_entry *entry))
entry             544 drivers/input/rmi4/rmi_driver.c 		 void *ctx, const struct pdt_entry *entry))
entry              95 drivers/input/rmi4/rmi_driver.h 		 const struct pdt_entry *entry));
entry              91 drivers/input/rmi4/rmi_smbus.c 		struct mapping_table_entry *entry = &rmi_smb->mapping_table[i];
entry              93 drivers/input/rmi4/rmi_smbus.c 		if (le16_to_cpu(entry->rmiaddr) == rmiaddr) {
entry              95 drivers/input/rmi4/rmi_smbus.c 				if (entry->readcount == bytecount)
entry              98 drivers/input/rmi4/rmi_smbus.c 				if (entry->flags & RMI_SMB2_MAP_FLAGS_WE) {
entry             172 drivers/input/sparse-keymap.c 	struct key_entry *map, *entry;
entry             185 drivers/input/sparse-keymap.c 		entry = &map[i];
entry             188 drivers/input/sparse-keymap.c 			error = setup(dev, entry);
entry             193 drivers/input/sparse-keymap.c 		switch (entry->type) {
entry             196 drivers/input/sparse-keymap.c 			__set_bit(entry->keycode, dev->keybit);
entry             202 drivers/input/sparse-keymap.c 			__set_bit(entry->sw.code, dev->swbit);
entry             128 drivers/iommu/amd_iommu.c 				struct acpihid_map_entry *entry)
entry             143 drivers/iommu/amd_iommu.c 		return strcmp(hid, entry->hid);
entry             145 drivers/iommu/amd_iommu.c 	if (!(*entry->uid))
entry             146 drivers/iommu/amd_iommu.c 		return strcmp(hid, entry->hid);
entry             148 drivers/iommu/amd_iommu.c 	return (strcmp(hid, entry->hid) || strcmp(uid, entry->uid));
entry             159 drivers/iommu/amd_iommu.c 					struct acpihid_map_entry **entry)
entry             165 drivers/iommu/amd_iommu.c 			if (entry)
entry             166 drivers/iommu/amd_iommu.c 				*entry = p;
entry             313 drivers/iommu/amd_iommu.c 	struct acpihid_map_entry *p, *entry = NULL;
entry             316 drivers/iommu/amd_iommu.c 	devid = get_acpihid_device_id(dev, &entry);
entry             322 drivers/iommu/amd_iommu.c 			entry->group = p->group;
entry             325 drivers/iommu/amd_iommu.c 	if (!entry->group)
entry             326 drivers/iommu/amd_iommu.c 		entry->group = generic_device_group(dev);
entry             328 drivers/iommu/amd_iommu.c 		iommu_group_ref_get(entry->group);
entry             330 drivers/iommu/amd_iommu.c 	return entry->group;
entry             686 drivers/iommu/amd_iommu.c 		u64 entry[2];
entry             703 drivers/iommu/amd_iommu.c 		entry[0] = raw[0];
entry             704 drivers/iommu/amd_iommu.c 		entry[1] = raw[1];
entry             717 drivers/iommu/amd_iommu.c 		iommu_handle_ppr_entry(iommu, entry);
entry            2897 drivers/iommu/amd_iommu.c 	struct iommu_dev_data *entry;
entry            2903 drivers/iommu/amd_iommu.c 		entry = list_first_entry(&domain->dev_list,
entry            2905 drivers/iommu/amd_iommu.c 		BUG_ON(!entry->domain);
entry            2906 drivers/iommu/amd_iommu.c 		do_detach(entry);
entry            3180 drivers/iommu/amd_iommu.c 	struct unity_map_entry *entry;
entry            3187 drivers/iommu/amd_iommu.c 	list_for_each_entry(entry, &amd_iommu_unity_map, list) {
entry            3191 drivers/iommu/amd_iommu.c 		if (devid < entry->devid_start || devid > entry->devid_end)
entry            3195 drivers/iommu/amd_iommu.c 		length = entry->address_end - entry->address_start;
entry            3196 drivers/iommu/amd_iommu.c 		if (entry->prot & IOMMU_PROT_IR)
entry            3198 drivers/iommu/amd_iommu.c 		if (entry->prot & IOMMU_PROT_IW)
entry            3200 drivers/iommu/amd_iommu.c 		if (entry->prot & IOMMU_UNITY_MAP_FLAG_EXCL_RANGE)
entry            3204 drivers/iommu/amd_iommu.c 		region = iommu_alloc_resv_region(entry->address_start,
entry            3231 drivers/iommu/amd_iommu.c 	struct iommu_resv_region *entry, *next;
entry            3233 drivers/iommu/amd_iommu.c 	list_for_each_entry_safe(entry, next, head, list)
entry            3234 drivers/iommu/amd_iommu.c 		kfree(entry);
entry            3879 drivers/iommu/amd_iommu.c 	struct irte_ga *entry;
entry            3891 drivers/iommu/amd_iommu.c 	entry = (struct irte_ga *)table->table;
entry            3892 drivers/iommu/amd_iommu.c 	entry = &entry[index];
entry            3893 drivers/iommu/amd_iommu.c 	entry->lo.fields_remap.valid = 0;
entry            3894 drivers/iommu/amd_iommu.c 	entry->hi.val = irte->hi.val;
entry            3895 drivers/iommu/amd_iommu.c 	entry->lo.val = irte->lo.val;
entry            3896 drivers/iommu/amd_iommu.c 	entry->lo.fields_remap.valid = 1;
entry            3898 drivers/iommu/amd_iommu.c 		data->ref = entry;
entry            3954 drivers/iommu/amd_iommu.c static void irte_prepare(void *entry,
entry            3958 drivers/iommu/amd_iommu.c 	union irte *irte = (union irte *) entry;
entry            3968 drivers/iommu/amd_iommu.c static void irte_ga_prepare(void *entry,
entry            3972 drivers/iommu/amd_iommu.c 	struct irte_ga *irte = (struct irte_ga *) entry;
entry            3984 drivers/iommu/amd_iommu.c static void irte_activate(void *entry, u16 devid, u16 index)
entry            3986 drivers/iommu/amd_iommu.c 	union irte *irte = (union irte *) entry;
entry            3992 drivers/iommu/amd_iommu.c static void irte_ga_activate(void *entry, u16 devid, u16 index)
entry            3994 drivers/iommu/amd_iommu.c 	struct irte_ga *irte = (struct irte_ga *) entry;
entry            4000 drivers/iommu/amd_iommu.c static void irte_deactivate(void *entry, u16 devid, u16 index)
entry            4002 drivers/iommu/amd_iommu.c 	union irte *irte = (union irte *) entry;
entry            4008 drivers/iommu/amd_iommu.c static void irte_ga_deactivate(void *entry, u16 devid, u16 index)
entry            4010 drivers/iommu/amd_iommu.c 	struct irte_ga *irte = (struct irte_ga *) entry;
entry            4016 drivers/iommu/amd_iommu.c static void irte_set_affinity(void *entry, u16 devid, u16 index,
entry            4019 drivers/iommu/amd_iommu.c 	union irte *irte = (union irte *) entry;
entry            4026 drivers/iommu/amd_iommu.c static void irte_ga_set_affinity(void *entry, u16 devid, u16 index,
entry            4029 drivers/iommu/amd_iommu.c 	struct irte_ga *irte = (struct irte_ga *) entry;
entry            4171 drivers/iommu/amd_iommu.c 	struct IO_APIC_route_entry *entry;
entry            4179 drivers/iommu/amd_iommu.c 	iommu->irte_ops->prepare(data->entry, apic->irq_delivery_mode,
entry            4186 drivers/iommu/amd_iommu.c 		entry = info->ioapic_entry;
entry            4188 drivers/iommu/amd_iommu.c 		memset(entry, 0, sizeof(*entry));
entry            4189 drivers/iommu/amd_iommu.c 		entry->vector        = index;
entry            4190 drivers/iommu/amd_iommu.c 		entry->mask          = 0;
entry            4191 drivers/iommu/amd_iommu.c 		entry->trigger       = info->ioapic_trigger;
entry            4192 drivers/iommu/amd_iommu.c 		entry->polarity      = info->ioapic_polarity;
entry            4195 drivers/iommu/amd_iommu.c 			entry->mask = 1;
entry            4313 drivers/iommu/amd_iommu.c 			data->entry = kzalloc(sizeof(union irte), GFP_KERNEL);
entry            4315 drivers/iommu/amd_iommu.c 			data->entry = kzalloc(sizeof(struct irte_ga),
entry            4317 drivers/iommu/amd_iommu.c 		if (!data->entry) {
entry            4358 drivers/iommu/amd_iommu.c 			kfree(data->entry);
entry            4381 drivers/iommu/amd_iommu.c 	iommu->irte_ops->activate(data->entry, irte_info->devid,
entry            4395 drivers/iommu/amd_iommu.c 		iommu->irte_ops->deactivate(data->entry, irte_info->devid,
entry            4409 drivers/iommu/amd_iommu.c 	struct irte_ga *entry = (struct irte_ga *) ir_data->entry;
entry            4412 drivers/iommu/amd_iommu.c 	    !entry || entry->lo.fields_vapic.guest_mode)
entry            4415 drivers/iommu/amd_iommu.c 	entry->lo.val = 0;
entry            4416 drivers/iommu/amd_iommu.c 	entry->hi.val = 0;
entry            4418 drivers/iommu/amd_iommu.c 	entry->lo.fields_vapic.guest_mode  = 1;
entry            4419 drivers/iommu/amd_iommu.c 	entry->lo.fields_vapic.ga_log_intr = 1;
entry            4420 drivers/iommu/amd_iommu.c 	entry->hi.fields.ga_root_ptr       = ir_data->ga_root_ptr;
entry            4421 drivers/iommu/amd_iommu.c 	entry->hi.fields.vector            = ir_data->ga_vector;
entry            4422 drivers/iommu/amd_iommu.c 	entry->lo.fields_vapic.ga_tag      = ir_data->ga_tag;
entry            4425 drivers/iommu/amd_iommu.c 			      ir_data->irq_2_irte.index, entry, ir_data);
entry            4432 drivers/iommu/amd_iommu.c 	struct irte_ga *entry = (struct irte_ga *) ir_data->entry;
entry            4436 drivers/iommu/amd_iommu.c 	    !entry || !entry->lo.fields_vapic.guest_mode)
entry            4439 drivers/iommu/amd_iommu.c 	entry->lo.val = 0;
entry            4440 drivers/iommu/amd_iommu.c 	entry->hi.val = 0;
entry            4442 drivers/iommu/amd_iommu.c 	entry->lo.fields_remap.dm          = apic->irq_dest_mode;
entry            4443 drivers/iommu/amd_iommu.c 	entry->lo.fields_remap.int_type    = apic->irq_delivery_mode;
entry            4444 drivers/iommu/amd_iommu.c 	entry->hi.fields.vector            = cfg->vector;
entry            4445 drivers/iommu/amd_iommu.c 	entry->lo.fields_remap.destination =
entry            4447 drivers/iommu/amd_iommu.c 	entry->hi.fields.destination =
entry            4451 drivers/iommu/amd_iommu.c 			      ir_data->irq_2_irte.index, entry, ir_data);
entry            4522 drivers/iommu/amd_iommu.c 	iommu->irte_ops->set_affinity(ir_data->entry, irte_info->devid,
entry            4596 drivers/iommu/amd_iommu.c 	struct irte_ga *entry = (struct irte_ga *) ir_data->entry;
entry            4600 drivers/iommu/amd_iommu.c 	    !ref || !entry || !entry->lo.fields_vapic.guest_mode)
entry             348 drivers/iommu/amd_iommu_init.c 	u64 entry;
entry             353 drivers/iommu/amd_iommu_init.c 	entry = start | MMIO_EXCL_ENABLE_MASK;
entry             355 drivers/iommu/amd_iommu_init.c 			&entry, sizeof(entry));
entry             357 drivers/iommu/amd_iommu_init.c 	entry = limit;
entry             359 drivers/iommu/amd_iommu_init.c 			&entry, sizeof(entry));
entry             365 drivers/iommu/amd_iommu_init.c 	u64 entry;
entry             369 drivers/iommu/amd_iommu_init.c 	entry = iommu_virt_to_phys(amd_iommu_dev_table);
entry             370 drivers/iommu/amd_iommu_init.c 	entry |= (dev_table_size >> 12) - 1;
entry             372 drivers/iommu/amd_iommu_init.c 			&entry, sizeof(entry));
entry             626 drivers/iommu/amd_iommu_init.c 	u64 entry;
entry             630 drivers/iommu/amd_iommu_init.c 	entry = iommu_virt_to_phys(iommu->cmd_buf);
entry             631 drivers/iommu/amd_iommu_init.c 	entry |= MMIO_CMD_SIZE_512;
entry             634 drivers/iommu/amd_iommu_init.c 		    &entry, sizeof(entry));
entry             663 drivers/iommu/amd_iommu_init.c 	u64 entry;
entry             667 drivers/iommu/amd_iommu_init.c 	entry = iommu_virt_to_phys(iommu->evt_buf) | EVT_LEN_MASK;
entry             670 drivers/iommu/amd_iommu_init.c 		    &entry, sizeof(entry));
entry             703 drivers/iommu/amd_iommu_init.c 	u64 entry;
entry             708 drivers/iommu/amd_iommu_init.c 	entry = iommu_virt_to_phys(iommu->ppr_log) | PPR_LOG_SIZE_512;
entry             711 drivers/iommu/amd_iommu_init.c 		    &entry, sizeof(entry));
entry             773 drivers/iommu/amd_iommu_init.c 	u64 entry;
entry             788 drivers/iommu/amd_iommu_init.c 	entry = iommu_virt_to_phys(iommu->ga_log) | GA_LOG_SIZE_512;
entry             790 drivers/iommu/amd_iommu_init.c 		    &entry, sizeof(entry));
entry             791 drivers/iommu/amd_iommu_init.c 	entry = (iommu_virt_to_phys(iommu->ga_log_tail) &
entry             794 drivers/iommu/amd_iommu_init.c 		    &entry, sizeof(entry));
entry             864 drivers/iommu/amd_iommu_init.c 	u64 int_ctl, int_tab_len, entry = 0, last_entry = 0;
entry             881 drivers/iommu/amd_iommu_init.c 		entry = (((u64) hi) << 32) + lo;
entry             882 drivers/iommu/amd_iommu_init.c 		if (last_entry && last_entry != entry) {
entry             887 drivers/iommu/amd_iommu_init.c 		last_entry = entry;
entry             889 drivers/iommu/amd_iommu_init.c 		old_devtb_size = ((entry & ~PAGE_MASK) + 1) << 12;
entry             902 drivers/iommu/amd_iommu_init.c 	old_devtb_phys = __sme_clr(entry) & PAGE_MASK;
entry            1008 drivers/iommu/amd_iommu_init.c 	struct devid_map *entry;
entry            1018 drivers/iommu/amd_iommu_init.c 	list_for_each_entry(entry, list, list) {
entry            1019 drivers/iommu/amd_iommu_init.c 		if (!(entry->id == id && entry->cmd_line))
entry            1025 drivers/iommu/amd_iommu_init.c 		*devid = entry->devid;
entry            1030 drivers/iommu/amd_iommu_init.c 	entry = kzalloc(sizeof(*entry), GFP_KERNEL);
entry            1031 drivers/iommu/amd_iommu_init.c 	if (!entry)
entry            1034 drivers/iommu/amd_iommu_init.c 	entry->id	= id;
entry            1035 drivers/iommu/amd_iommu_init.c 	entry->devid	= *devid;
entry            1036 drivers/iommu/amd_iommu_init.c 	entry->cmd_line	= cmd_line;
entry            1038 drivers/iommu/amd_iommu_init.c 	list_add_tail(&entry->list, list);
entry            1046 drivers/iommu/amd_iommu_init.c 	struct acpihid_map_entry *entry;
entry            1049 drivers/iommu/amd_iommu_init.c 	list_for_each_entry(entry, list, list) {
entry            1050 drivers/iommu/amd_iommu_init.c 		if (strcmp(entry->hid, hid) ||
entry            1051 drivers/iommu/amd_iommu_init.c 		    (*uid && *entry->uid && strcmp(entry->uid, uid)) ||
entry            1052 drivers/iommu/amd_iommu_init.c 		    !entry->cmd_line)
entry            1057 drivers/iommu/amd_iommu_init.c 		*devid = entry->devid;
entry            1061 drivers/iommu/amd_iommu_init.c 	entry = kzalloc(sizeof(*entry), GFP_KERNEL);
entry            1062 drivers/iommu/amd_iommu_init.c 	if (!entry)
entry            1065 drivers/iommu/amd_iommu_init.c 	memcpy(entry->uid, uid, strlen(uid));
entry            1066 drivers/iommu/amd_iommu_init.c 	memcpy(entry->hid, hid, strlen(hid));
entry            1067 drivers/iommu/amd_iommu_init.c 	entry->devid = *devid;
entry            1068 drivers/iommu/amd_iommu_init.c 	entry->cmd_line	= cmd_line;
entry            1069 drivers/iommu/amd_iommu_init.c 	entry->root_devid = (entry->devid & (~0x7));
entry            1072 drivers/iommu/amd_iommu_init.c 		entry->cmd_line ? "cmd" : "ivrs",
entry            1073 drivers/iommu/amd_iommu_init.c 		entry->hid, entry->uid, entry->root_devid);
entry            1075 drivers/iommu/amd_iommu_init.c 	list_add_tail(&entry->list, list);
entry            2070 drivers/iommu/amd_iommu_init.c 	struct unity_map_entry *entry, *next;
entry            2072 drivers/iommu/amd_iommu_init.c 	list_for_each_entry_safe(entry, next, &amd_iommu_unity_map, list) {
entry            2073 drivers/iommu/amd_iommu_init.c 		list_del(&entry->list);
entry            2074 drivers/iommu/amd_iommu_init.c 		kfree(entry);
entry             763 drivers/iommu/amd_iommu_types.h 	struct devid_map *entry;
entry             765 drivers/iommu/amd_iommu_types.h 	list_for_each_entry(entry, &ioapic_map, list) {
entry             766 drivers/iommu/amd_iommu_types.h 		if (entry->id == id)
entry             767 drivers/iommu/amd_iommu_types.h 			return entry->devid;
entry             775 drivers/iommu/amd_iommu_types.h 	struct devid_map *entry;
entry             777 drivers/iommu/amd_iommu_types.h 	list_for_each_entry(entry, &hpet_map, list) {
entry             778 drivers/iommu/amd_iommu_types.h 		if (entry->id == id)
entry             779 drivers/iommu/amd_iommu_types.h 			return entry->devid;
entry             876 drivers/iommu/amd_iommu_types.h 	void *entry;    /* Pointer to union irte or struct irte_ga */
entry            2718 drivers/iommu/arm-smmu-v3.c 	struct iommu_resv_region *entry, *next;
entry            2720 drivers/iommu/arm-smmu-v3.c 	list_for_each_entry_safe(entry, next, head, list)
entry            2721 drivers/iommu/arm-smmu-v3.c 		kfree(entry);
entry            1549 drivers/iommu/arm-smmu.c 	struct iommu_resv_region *entry, *next;
entry            1551 drivers/iommu/arm-smmu.c 	list_for_each_entry_safe(entry, next, head, list)
entry            1552 drivers/iommu/arm-smmu.c 		kfree(entry);
entry             851 drivers/iommu/dmar.c dmar_validate_one_drhd(struct acpi_dmar_header *entry, void *arg)
entry             857 drivers/iommu/dmar.c 	drhd = (void *)entry;
entry            1079 drivers/iommu/exynos-iommu.c 	sysmmu_pte_t *entry;
entry            1089 drivers/iommu/exynos-iommu.c 	entry = section_entry(domain->pgtable, iova);
entry            1092 drivers/iommu/exynos-iommu.c 		ret = lv1set_section(domain, entry, iova, paddr, prot,
entry            1097 drivers/iommu/exynos-iommu.c 		pent = alloc_lv2entry(domain, entry, iova,
entry            1214 drivers/iommu/exynos-iommu.c 	sysmmu_pte_t *entry;
entry            1220 drivers/iommu/exynos-iommu.c 	entry = section_entry(domain->pgtable, iova);
entry            1222 drivers/iommu/exynos-iommu.c 	if (lv1ent_section(entry)) {
entry            1223 drivers/iommu/exynos-iommu.c 		phys = section_phys(entry) + section_offs(iova);
entry            1224 drivers/iommu/exynos-iommu.c 	} else if (lv1ent_page(entry)) {
entry            1225 drivers/iommu/exynos-iommu.c 		entry = page_entry(entry, iova);
entry            1227 drivers/iommu/exynos-iommu.c 		if (lv2ent_large(entry))
entry            1228 drivers/iommu/exynos-iommu.c 			phys = lpage_phys(entry) + lpage_offs(iova);
entry            1229 drivers/iommu/exynos-iommu.c 		else if (lv2ent_small(entry))
entry            1230 drivers/iommu/exynos-iommu.c 			phys = spage_phys(entry) + spage_offs(iova);
entry            1296 drivers/iommu/exynos-iommu.c 	struct sysmmu_drvdata *data, *entry;
entry            1315 drivers/iommu/exynos-iommu.c 	list_for_each_entry(entry, &owner->controllers, owner_node)
entry            1316 drivers/iommu/exynos-iommu.c 		if (entry == data)
entry              43 drivers/iommu/hyperv-iommu.c 	struct IO_APIC_route_entry *entry;
entry              54 drivers/iommu/hyperv-iommu.c 	entry = data->chip_data;
entry              55 drivers/iommu/hyperv-iommu.c 	entry->dest = cfg->dest_apicid;
entry              56 drivers/iommu/hyperv-iommu.c 	entry->vector = cfg->vector;
entry             126 drivers/iommu/hyperv-iommu.c 	struct IO_APIC_route_entry *entry = irq_data->chip_data;
entry             128 drivers/iommu/hyperv-iommu.c 	entry->dest = cfg->dest_apicid;
entry             129 drivers/iommu/hyperv-iommu.c 	entry->vector = cfg->vector;
entry             701 drivers/iommu/intel-iommu.c 	u64 *entry;
entry             703 drivers/iommu/intel-iommu.c 	entry = &root->lo;
entry             707 drivers/iommu/intel-iommu.c 			entry = &root->hi;
entry             711 drivers/iommu/intel-iommu.c 	if (*entry & 1)
entry             712 drivers/iommu/intel-iommu.c 		context = phys_to_virt(*entry & VTD_PAGE_MASK);
entry             724 drivers/iommu/intel-iommu.c 		*entry = phy_addr | 1;
entry             725 drivers/iommu/intel-iommu.c 		__iommu_flush_cache(iommu, entry, sizeof(*entry));
entry            5726 drivers/iommu/intel-iommu.c 	struct iommu_resv_region *entry, *next;
entry            5728 drivers/iommu/intel-iommu.c 	list_for_each_entry_safe(entry, next, head, list)
entry            5729 drivers/iommu/intel-iommu.c 		kfree(entry);
entry              23 drivers/iommu/intel-pasid.h #define is_pasid_enabled(entry)		(((entry)->lo >> 3) & 0x1)
entry              24 drivers/iommu/intel-pasid.h #define get_pasid_dir_size(entry)	(1 << ((((entry)->lo >> 9) & 0x7) + 7))
entry             239 drivers/iommu/intel_irq_remapping.c 	struct irte *start, *entry, *end;
entry             252 drivers/iommu/intel_irq_remapping.c 	for (entry = start; entry < end; entry++) {
entry             253 drivers/iommu/intel_irq_remapping.c 		set_64bit(&entry->low, 0);
entry             254 drivers/iommu/intel_irq_remapping.c 		set_64bit(&entry->high, 0);
entry            1264 drivers/iommu/intel_irq_remapping.c 	struct IR_IO_APIC_route_entry *entry;
entry            1280 drivers/iommu/intel_irq_remapping.c 		entry = (struct IR_IO_APIC_route_entry *)info->ioapic_entry;
entry            1282 drivers/iommu/intel_irq_remapping.c 		memset(entry, 0, sizeof(*entry));
entry            1283 drivers/iommu/intel_irq_remapping.c 		entry->index2	= (index >> 15) & 0x1;
entry            1284 drivers/iommu/intel_irq_remapping.c 		entry->zero	= 0;
entry            1285 drivers/iommu/intel_irq_remapping.c 		entry->format	= 1;
entry            1286 drivers/iommu/intel_irq_remapping.c 		entry->index	= (index & 0x7fff);
entry            1291 drivers/iommu/intel_irq_remapping.c 		entry->vector	= info->ioapic_pin;
entry            1292 drivers/iommu/intel_irq_remapping.c 		entry->mask	= 0;			/* enable IRQ */
entry            1293 drivers/iommu/intel_irq_remapping.c 		entry->trigger	= info->ioapic_trigger;
entry            1294 drivers/iommu/intel_irq_remapping.c 		entry->polarity	= info->ioapic_polarity;
entry            1296 drivers/iommu/intel_irq_remapping.c 			entry->mask = 1; /* Mask level triggered irqs. */
entry            1437 drivers/iommu/intel_irq_remapping.c 	struct irte entry;
entry            1439 drivers/iommu/intel_irq_remapping.c 	memset(&entry, 0, sizeof(entry));
entry            1440 drivers/iommu/intel_irq_remapping.c 	modify_irte(&data->irq_2_iommu, &entry);
entry             348 drivers/iommu/iommu.c 	struct iommu_resv_region *entry;
entry             351 drivers/iommu/iommu.c 	list_for_each_entry(entry, dev_resv_regions, list) {
entry             352 drivers/iommu/iommu.c 		ret = iommu_insert_resv_region(entry, group_resv_regions);
entry             627 drivers/iommu/iommu.c 	struct iommu_resv_region *entry;
entry             643 drivers/iommu/iommu.c 	list_for_each_entry(entry, &mappings, list) {
entry             647 drivers/iommu/iommu.c 			domain->ops->apply_resv_region(dev, domain, entry);
entry             649 drivers/iommu/iommu.c 		start = ALIGN(entry->start, pg_size);
entry             650 drivers/iommu/iommu.c 		end   = ALIGN(entry->start + entry->length, pg_size);
entry             652 drivers/iommu/iommu.c 		if (entry->type != IOMMU_RESV_DIRECT &&
entry             653 drivers/iommu/iommu.c 		    entry->type != IOMMU_RESV_DIRECT_RELAXABLE)
entry             663 drivers/iommu/iommu.c 			ret = iommu_map(domain, addr, addr, pg_size, entry->prot);
entry             811 drivers/iommu/iommu.c 	struct group_device *entry;
entry             814 drivers/iommu/iommu.c 	list_for_each_entry(entry, &group->devices, list)
entry             217 drivers/iommu/s390-iommu.c 	unsigned long *entry;
entry             230 drivers/iommu/s390-iommu.c 		entry = dma_walk_cpu_trans(s390_domain->dma_table, dma_addr);
entry             231 drivers/iommu/s390-iommu.c 		if (!entry) {
entry             235 drivers/iommu/s390-iommu.c 		dma_update_cpu_trans(entry, page_addr, flags);
entry             255 drivers/iommu/s390-iommu.c 			entry = dma_walk_cpu_trans(s390_domain->dma_table,
entry             257 drivers/iommu/s390-iommu.c 			if (!entry)
entry             259 drivers/iommu/s390-iommu.c 			dma_update_cpu_trans(entry, page_addr, flags);
entry             812 drivers/iommu/virtio-iommu.c 	struct iommu_resv_region *entry, *new_entry, *msi = NULL;
entry             817 drivers/iommu/virtio-iommu.c 	list_for_each_entry(entry, &vdev->resv_regions, list) {
entry             818 drivers/iommu/virtio-iommu.c 		if (entry->type == IOMMU_RESV_MSI)
entry             819 drivers/iommu/virtio-iommu.c 			msi = entry;
entry             821 drivers/iommu/virtio-iommu.c 		new_entry = kmemdup(entry, sizeof(*entry), GFP_KERNEL);
entry             845 drivers/iommu/virtio-iommu.c 	struct iommu_resv_region *entry, *next;
entry             847 drivers/iommu/virtio-iommu.c 	list_for_each_entry_safe(entry, next, head, list)
entry             848 drivers/iommu/virtio-iommu.c 		kfree(entry);
entry             197 drivers/irqchip/irq-crossbar.c 	u32 max = 0, entry, reg_size;
entry             242 drivers/irqchip/irq-crossbar.c 						   i, &entry);
entry             243 drivers/irqchip/irq-crossbar.c 			if (entry >= max) {
entry             248 drivers/irqchip/irq-crossbar.c 			cb->irq_map[entry] = IRQ_RESERVED;
entry             260 drivers/irqchip/irq-crossbar.c 						   i, &entry);
entry             261 drivers/irqchip/irq-crossbar.c 			if (entry >= max) {
entry             266 drivers/irqchip/irq-crossbar.c 			cb->irq_map[entry] = IRQ_SKIP;
entry              62 drivers/irqchip/irq-gic-v2m.c 	struct list_head entry;
entry             184 drivers/irqchip/irq-gic-v2m.c 	list_for_each_entry(tmp, &v2m_nodes, entry) {
entry             269 drivers/irqchip/irq-gic-v2m.c 	list_for_each_entry_safe(v2m, tmp, &v2m_nodes, entry) {
entry             270 drivers/irqchip/irq-gic-v2m.c 		list_del(&v2m->entry);
entry             285 drivers/irqchip/irq-gic-v2m.c 	v2m = list_first_entry_or_null(&v2m_nodes, struct v2m_data, entry);
entry             330 drivers/irqchip/irq-gic-v2m.c 	INIT_LIST_HEAD(&v2m->entry);
entry             397 drivers/irqchip/irq-gic-v2m.c 	list_add_tail(&v2m->entry, &v2m_nodes);
entry             468 drivers/irqchip/irq-gic-v2m.c 	data = list_first_entry_or_null(&v2m_nodes, struct v2m_data, entry);
entry              96 drivers/irqchip/irq-gic-v3-its.c 	struct list_head	entry;
entry             146 drivers/irqchip/irq-gic-v3-its.c 	struct list_head	entry;
entry             183 drivers/irqchip/irq-gic-v3-its.c 	list_for_each_entry(its, &its_nodes, entry) {
entry            1003 drivers/irqchip/irq-gic-v3-its.c 		its = list_first_entry(&its_nodes, struct its_node, entry);
entry            1023 drivers/irqchip/irq-gic-v3-its.c 	list_for_each_entry(its, &its_nodes, entry) {
entry            1475 drivers/irqchip/irq-gic-v3-its.c 	struct list_head	entry;
entry            1500 drivers/irqchip/irq-gic-v3-its.c 	list_for_each_entry_safe(range, tmp, &lpi_range_list, entry) {
entry            1507 drivers/irqchip/irq-gic-v3-its.c 				list_del(&range->entry);
entry            1524 drivers/irqchip/irq-gic-v3-its.c 	if (&a->entry == &lpi_range_list || &b->entry == &lpi_range_list)
entry            1530 drivers/irqchip/irq-gic-v3-its.c 	list_del(&a->entry);
entry            1544 drivers/irqchip/irq-gic-v3-its.c 	list_for_each_entry_reverse(old, &lpi_range_list, entry) {
entry            1555 drivers/irqchip/irq-gic-v3-its.c 	list_add(&new->entry, &old->entry);
entry            1561 drivers/irqchip/irq-gic-v3-its.c 	merge_lpi_ranges(new, list_next_entry(new, entry));
entry            2276 drivers/irqchip/irq-gic-v3-its.c 	list_for_each_entry(its, &its_nodes, entry)
entry            2289 drivers/irqchip/irq-gic-v3-its.c 	list_for_each_entry(tmp, &its->its_device_list, entry) {
entry            2380 drivers/irqchip/irq-gic-v3-its.c 	list_for_each_entry(its, &its_nodes, entry) {
entry            2455 drivers/irqchip/irq-gic-v3-its.c 	INIT_LIST_HEAD(&dev->entry);
entry            2458 drivers/irqchip/irq-gic-v3-its.c 	list_add(&dev->entry, &its->its_device_list);
entry            2472 drivers/irqchip/irq-gic-v3-its.c 	list_del(&its_dev->entry);
entry            2871 drivers/irqchip/irq-gic-v3-its.c 	list_for_each_entry(its, &its_nodes, entry) {
entry            3146 drivers/irqchip/irq-gic-v3-its.c 	list_for_each_entry(its, &its_nodes, entry) {
entry            3172 drivers/irqchip/irq-gic-v3-its.c 	list_for_each_entry(its, &its_nodes, entry) {
entry            3364 drivers/irqchip/irq-gic-v3-its.c 	list_for_each_entry(its, &its_nodes, entry) {
entry            3385 drivers/irqchip/irq-gic-v3-its.c 		list_for_each_entry_continue_reverse(its, &its_nodes, entry) {
entry            3406 drivers/irqchip/irq-gic-v3-its.c 	list_for_each_entry(its, &its_nodes, entry) {
entry            3502 drivers/irqchip/irq-gic-v3-its.c 	its = list_first_entry(&its_nodes, struct its_node, entry);
entry            3608 drivers/irqchip/irq-gic-v3-its.c 	INIT_LIST_HEAD(&its->entry);
entry            3695 drivers/irqchip/irq-gic-v3-its.c 	list_add(&its->entry, &its_nodes);
entry            4007 drivers/irqchip/irq-gic-v3-its.c 	list_for_each_entry(its, &its_nodes, entry)
entry              65 drivers/isdn/mISDN/dsp_pipeline.c 	struct dsp_element_entry *entry =
entry              67 drivers/isdn/mISDN/dsp_pipeline.c 	list_del(&entry->list);
entry              68 drivers/isdn/mISDN/dsp_pipeline.c 	kfree(entry);
entry              73 drivers/isdn/mISDN/dsp_pipeline.c 	struct dsp_element_entry *entry;
entry              79 drivers/isdn/mISDN/dsp_pipeline.c 	entry = kzalloc(sizeof(struct dsp_element_entry), GFP_ATOMIC);
entry              80 drivers/isdn/mISDN/dsp_pipeline.c 	if (!entry)
entry              83 drivers/isdn/mISDN/dsp_pipeline.c 	entry->elem = elem;
entry              85 drivers/isdn/mISDN/dsp_pipeline.c 	entry->dev.class = elements_class;
entry              86 drivers/isdn/mISDN/dsp_pipeline.c 	entry->dev.release = mISDN_dsp_dev_release;
entry              87 drivers/isdn/mISDN/dsp_pipeline.c 	dev_set_drvdata(&entry->dev, elem);
entry              88 drivers/isdn/mISDN/dsp_pipeline.c 	dev_set_name(&entry->dev, "%s", elem->name);
entry              89 drivers/isdn/mISDN/dsp_pipeline.c 	ret = device_register(&entry->dev);
entry              95 drivers/isdn/mISDN/dsp_pipeline.c 	list_add_tail(&entry->list, &dsp_elements);
entry              98 drivers/isdn/mISDN/dsp_pipeline.c 		ret = device_create_file(&entry->dev,
entry             114 drivers/isdn/mISDN/dsp_pipeline.c 	device_unregister(&entry->dev);
entry             117 drivers/isdn/mISDN/dsp_pipeline.c 	kfree(entry);
entry             124 drivers/isdn/mISDN/dsp_pipeline.c 	struct dsp_element_entry *entry, *n;
entry             129 drivers/isdn/mISDN/dsp_pipeline.c 	list_for_each_entry_safe(entry, n, &dsp_elements, list)
entry             130 drivers/isdn/mISDN/dsp_pipeline.c 		if (entry->elem == elem) {
entry             131 drivers/isdn/mISDN/dsp_pipeline.c 			device_unregister(&entry->dev);
entry             159 drivers/isdn/mISDN/dsp_pipeline.c 	struct dsp_element_entry *entry, *n;
entry             165 drivers/isdn/mISDN/dsp_pipeline.c 	list_for_each_entry_safe(entry, n, &dsp_elements, list) {
entry             166 drivers/isdn/mISDN/dsp_pipeline.c 		list_del(&entry->list);
entry             168 drivers/isdn/mISDN/dsp_pipeline.c 		       __func__, entry->elem->name);
entry             169 drivers/isdn/mISDN/dsp_pipeline.c 		kfree(entry);
entry             193 drivers/isdn/mISDN/dsp_pipeline.c 	struct dsp_pipeline_entry *entry, *n;
entry             195 drivers/isdn/mISDN/dsp_pipeline.c 	list_for_each_entry_safe(entry, n, &pipeline->list, list) {
entry             196 drivers/isdn/mISDN/dsp_pipeline.c 		list_del(&entry->list);
entry             197 drivers/isdn/mISDN/dsp_pipeline.c 		if (entry->elem == dsp_hwec)
entry             201 drivers/isdn/mISDN/dsp_pipeline.c 			entry->elem->free(entry->p);
entry             202 drivers/isdn/mISDN/dsp_pipeline.c 		kfree(entry);
entry             223 drivers/isdn/mISDN/dsp_pipeline.c 	struct dsp_element_entry *entry, *n;
entry             244 drivers/isdn/mISDN/dsp_pipeline.c 		list_for_each_entry_safe(entry, n, &dsp_elements, list)
entry             245 drivers/isdn/mISDN/dsp_pipeline.c 			if (!strcmp(entry->elem->name, name)) {
entry             246 drivers/isdn/mISDN/dsp_pipeline.c 				elem = entry->elem;
entry             316 drivers/isdn/mISDN/dsp_pipeline.c 	struct dsp_pipeline_entry *entry;
entry             321 drivers/isdn/mISDN/dsp_pipeline.c 	list_for_each_entry(entry, &pipeline->list, list)
entry             322 drivers/isdn/mISDN/dsp_pipeline.c 		if (entry->elem->process_tx)
entry             323 drivers/isdn/mISDN/dsp_pipeline.c 			entry->elem->process_tx(entry->p, data, len);
entry             329 drivers/isdn/mISDN/dsp_pipeline.c 	struct dsp_pipeline_entry *entry;
entry             334 drivers/isdn/mISDN/dsp_pipeline.c 	list_for_each_entry_reverse(entry, &pipeline->list, list)
entry             335 drivers/isdn/mISDN/dsp_pipeline.c 		if (entry->elem->process_rx)
entry             336 drivers/isdn/mISDN/dsp_pipeline.c 			entry->elem->process_rx(entry->p, data, len, txlen);
entry             114 drivers/lightnvm/pblk-rb.c 		struct pblk_rb_entry *entry;
entry             138 drivers/lightnvm/pblk-rb.c 		entry = &rb->entries[init_entry];
entry             139 drivers/lightnvm/pblk-rb.c 		entry->data = kaddr;
entry             140 drivers/lightnvm/pblk-rb.c 		entry->cacheline = pblk_cacheline_to_addr(init_entry++);
entry             141 drivers/lightnvm/pblk-rb.c 		entry->w_ctx.flags = PBLK_WRITABLE_ENTRY;
entry             145 drivers/lightnvm/pblk-rb.c 			entry = &rb->entries[init_entry];
entry             146 drivers/lightnvm/pblk-rb.c 			entry->cacheline = pblk_cacheline_to_addr(init_entry++);
entry             147 drivers/lightnvm/pblk-rb.c 			entry->data = kaddr + (i * rb->seg_size);
entry             148 drivers/lightnvm/pblk-rb.c 			entry->w_ctx.flags = PBLK_WRITABLE_ENTRY;
entry             149 drivers/lightnvm/pblk-rb.c 			bio_list_init(&entry->w_ctx.bios);
entry             241 drivers/lightnvm/pblk-rb.c 	struct pblk_rb_entry *entry;
entry             248 drivers/lightnvm/pblk-rb.c 		entry = &rb->entries[rb->l2p_update];
entry             249 drivers/lightnvm/pblk-rb.c 		w_ctx = &entry->w_ctx;
entry             251 drivers/lightnvm/pblk-rb.c 		flags = READ_ONCE(entry->w_ctx.flags);
entry             260 drivers/lightnvm/pblk-rb.c 							entry->cacheline);
entry             329 drivers/lightnvm/pblk-rb.c 				  struct pblk_rb_entry *entry)
entry             331 drivers/lightnvm/pblk-rb.c 	memcpy(entry->data, data, rb->seg_size);
entry             333 drivers/lightnvm/pblk-rb.c 	entry->w_ctx.lba = w_ctx.lba;
entry             334 drivers/lightnvm/pblk-rb.c 	entry->w_ctx.ppa = w_ctx.ppa;
entry             341 drivers/lightnvm/pblk-rb.c 	struct pblk_rb_entry *entry;
entry             344 drivers/lightnvm/pblk-rb.c 	entry = &rb->entries[ring_pos];
entry             345 drivers/lightnvm/pblk-rb.c 	flags = READ_ONCE(entry->w_ctx.flags);
entry             351 drivers/lightnvm/pblk-rb.c 	__pblk_rb_write_entry(rb, data, w_ctx, entry);
entry             353 drivers/lightnvm/pblk-rb.c 	pblk_update_map_cache(pblk, w_ctx.lba, entry->cacheline);
entry             357 drivers/lightnvm/pblk-rb.c 	smp_store_release(&entry->w_ctx.flags, flags);
entry             365 drivers/lightnvm/pblk-rb.c 	struct pblk_rb_entry *entry;
entry             368 drivers/lightnvm/pblk-rb.c 	entry = &rb->entries[ring_pos];
entry             369 drivers/lightnvm/pblk-rb.c 	flags = READ_ONCE(entry->w_ctx.flags);
entry             375 drivers/lightnvm/pblk-rb.c 	__pblk_rb_write_entry(rb, data, w_ctx, entry);
entry             377 drivers/lightnvm/pblk-rb.c 	if (!pblk_update_map_gc(pblk, w_ctx.lba, entry->cacheline, line, paddr))
entry             378 drivers/lightnvm/pblk-rb.c 		entry->w_ctx.lba = ADDR_EMPTY;
entry             383 drivers/lightnvm/pblk-rb.c 	smp_store_release(&entry->w_ctx.flags, flags);
entry             389 drivers/lightnvm/pblk-rb.c 	struct pblk_rb_entry *entry;
entry             405 drivers/lightnvm/pblk-rb.c 	entry = &rb->entries[flush_point];
entry             411 drivers/lightnvm/pblk-rb.c 		bio_list_add(&entry->w_ctx.bios, bio);
entry             558 drivers/lightnvm/pblk-rb.c 	struct pblk_rb_entry *entry;
entry             577 drivers/lightnvm/pblk-rb.c 		entry = &rb->entries[pos];
entry             583 drivers/lightnvm/pblk-rb.c 		flags = READ_ONCE(entry->w_ctx.flags);
entry             589 drivers/lightnvm/pblk-rb.c 		page = virt_to_page(entry->data);
entry             595 drivers/lightnvm/pblk-rb.c 			smp_store_release(&entry->w_ctx.flags, flags);
entry             605 drivers/lightnvm/pblk-rb.c 			smp_store_release(&entry->w_ctx.flags, flags);
entry             613 drivers/lightnvm/pblk-rb.c 		smp_store_release(&entry->w_ctx.flags, flags);
entry             648 drivers/lightnvm/pblk-rb.c 	struct pblk_rb_entry *entry;
entry             661 drivers/lightnvm/pblk-rb.c 	entry = &rb->entries[pos];
entry             662 drivers/lightnvm/pblk-rb.c 	w_ctx = &entry->w_ctx;
entry             677 drivers/lightnvm/pblk-rb.c 	memcpy(data, entry->data, rb->seg_size);
entry             686 drivers/lightnvm/pblk-rb.c 	unsigned int entry = pblk_rb_ptr_wrap(rb, pos, 0);
entry             688 drivers/lightnvm/pblk-rb.c 	return &rb->entries[entry].w_ctx;
entry             765 drivers/lightnvm/pblk-rb.c 	struct pblk_rb_entry *entry;
entry             784 drivers/lightnvm/pblk-rb.c 		entry = &rb->entries[i];
entry             786 drivers/lightnvm/pblk-rb.c 		if (!entry->data) {
entry             151 drivers/lightnvm/pblk-write.c 	struct pblk_rb_entry *entry;
entry             160 drivers/lightnvm/pblk-write.c 		entry = &rb->entries[pblk_rb_ptr_wrap(rb, sentry, i)];
entry             161 drivers/lightnvm/pblk-write.c 		w_ctx = &entry->w_ctx;
entry             166 drivers/lightnvm/pblk-write.c 			if (!pblk_ppa_comp(ppa_l2p, entry->cacheline))
entry             142 drivers/mailbox/tegra-hsp.c 	struct tegra_hsp_doorbell *entry;
entry             144 drivers/mailbox/tegra-hsp.c 	list_for_each_entry(entry, &hsp->doorbells, list)
entry             145 drivers/mailbox/tegra-hsp.c 		if (entry->master == master)
entry             146 drivers/mailbox/tegra-hsp.c 			return entry;
entry             336 drivers/md/dm-bio-prison-v1.c 	struct dm_deferred_entry *entry;
entry             339 drivers/md/dm-bio-prison-v1.c 	entry = ds->entries + ds->current_entry;
entry             340 drivers/md/dm-bio-prison-v1.c 	entry->count++;
entry             343 drivers/md/dm-bio-prison-v1.c 	return entry;
entry             364 drivers/md/dm-bio-prison-v1.c void dm_deferred_entry_dec(struct dm_deferred_entry *entry, struct list_head *head)
entry             368 drivers/md/dm-bio-prison-v1.c 	spin_lock_irqsave(&entry->ds->lock, flags);
entry             369 drivers/md/dm-bio-prison-v1.c 	BUG_ON(!entry->count);
entry             370 drivers/md/dm-bio-prison-v1.c 	--entry->count;
entry             371 drivers/md/dm-bio-prison-v1.c 	__sweep(entry->ds, head);
entry             372 drivers/md/dm-bio-prison-v1.c 	spin_unlock_irqrestore(&entry->ds->lock, flags);
entry             133 drivers/md/dm-bio-prison-v1.h void dm_deferred_entry_dec(struct dm_deferred_entry *entry, struct list_head *head);
entry              61 drivers/md/dm-cache-policy-smq.c 	struct entry *begin;
entry              62 drivers/md/dm-cache-policy-smq.c 	struct entry *end;
entry              72 drivers/md/dm-cache-policy-smq.c 	es->begin = vzalloc(array_size(nr_entries, sizeof(struct entry)));
entry              85 drivers/md/dm-cache-policy-smq.c static struct entry *__get_entry(struct entry_space *es, unsigned block)
entry              87 drivers/md/dm-cache-policy-smq.c 	struct entry *e;
entry              95 drivers/md/dm-cache-policy-smq.c static unsigned to_index(struct entry_space *es, struct entry *e)
entry             101 drivers/md/dm-cache-policy-smq.c static struct entry *to_entry(struct entry_space *es, unsigned block)
entry             122 drivers/md/dm-cache-policy-smq.c static struct entry *l_head(struct entry_space *es, struct ilist *l)
entry             127 drivers/md/dm-cache-policy-smq.c static struct entry *l_tail(struct entry_space *es, struct ilist *l)
entry             132 drivers/md/dm-cache-policy-smq.c static struct entry *l_next(struct entry_space *es, struct entry *e)
entry             137 drivers/md/dm-cache-policy-smq.c static struct entry *l_prev(struct entry_space *es, struct entry *e)
entry             147 drivers/md/dm-cache-policy-smq.c static void l_add_head(struct entry_space *es, struct ilist *l, struct entry *e)
entry             149 drivers/md/dm-cache-policy-smq.c 	struct entry *head = l_head(es, l);
entry             163 drivers/md/dm-cache-policy-smq.c static void l_add_tail(struct entry_space *es, struct ilist *l, struct entry *e)
entry             165 drivers/md/dm-cache-policy-smq.c 	struct entry *tail = l_tail(es, l);
entry             180 drivers/md/dm-cache-policy-smq.c 			 struct entry *old, struct entry *e)
entry             182 drivers/md/dm-cache-policy-smq.c 	struct entry *prev = l_prev(es, old);
entry             197 drivers/md/dm-cache-policy-smq.c static void l_del(struct entry_space *es, struct ilist *l, struct entry *e)
entry             199 drivers/md/dm-cache-policy-smq.c 	struct entry *prev = l_prev(es, e);
entry             200 drivers/md/dm-cache-policy-smq.c 	struct entry *next = l_next(es, e);
entry             216 drivers/md/dm-cache-policy-smq.c static struct entry *l_pop_head(struct entry_space *es, struct ilist *l)
entry             218 drivers/md/dm-cache-policy-smq.c 	struct entry *e;
entry             229 drivers/md/dm-cache-policy-smq.c static struct entry *l_pop_tail(struct entry_space *es, struct ilist *l)
entry             231 drivers/md/dm-cache-policy-smq.c 	struct entry *e;
entry             295 drivers/md/dm-cache-policy-smq.c static void q_push(struct queue *q, struct entry *e)
entry             305 drivers/md/dm-cache-policy-smq.c static void q_push_front(struct queue *q, struct entry *e)
entry             315 drivers/md/dm-cache-policy-smq.c static void q_push_before(struct queue *q, struct entry *old, struct entry *e)
entry             325 drivers/md/dm-cache-policy-smq.c static void q_del(struct queue *q, struct entry *e)
entry             335 drivers/md/dm-cache-policy-smq.c static struct entry *q_peek(struct queue *q, unsigned max_level, bool can_cross_sentinel)
entry             338 drivers/md/dm-cache-policy-smq.c 	struct entry *e;
entry             357 drivers/md/dm-cache-policy-smq.c static struct entry *q_pop(struct queue *q)
entry             359 drivers/md/dm-cache-policy-smq.c 	struct entry *e = q_peek(q, q->nr_levels, true);
entry             372 drivers/md/dm-cache-policy-smq.c static struct entry *__redist_pop_from(struct queue *q, unsigned level)
entry             374 drivers/md/dm-cache-policy-smq.c 	struct entry *e;
entry             431 drivers/md/dm-cache-policy-smq.c 	struct entry *e;
entry             470 drivers/md/dm-cache-policy-smq.c static void q_requeue(struct queue *q, struct entry *e, unsigned extra_levels,
entry             471 drivers/md/dm-cache-policy-smq.c 		      struct entry *s1, struct entry *s2)
entry             473 drivers/md/dm-cache-policy-smq.c 	struct entry *de;
entry             606 drivers/md/dm-cache-policy-smq.c static struct entry *h_head(struct smq_hash_table *ht, unsigned bucket)
entry             611 drivers/md/dm-cache-policy-smq.c static struct entry *h_next(struct smq_hash_table *ht, struct entry *e)
entry             616 drivers/md/dm-cache-policy-smq.c static void __h_insert(struct smq_hash_table *ht, unsigned bucket, struct entry *e)
entry             622 drivers/md/dm-cache-policy-smq.c static void h_insert(struct smq_hash_table *ht, struct entry *e)
entry             628 drivers/md/dm-cache-policy-smq.c static struct entry *__h_lookup(struct smq_hash_table *ht, unsigned h, dm_oblock_t oblock,
entry             629 drivers/md/dm-cache-policy-smq.c 				struct entry **prev)
entry             631 drivers/md/dm-cache-policy-smq.c 	struct entry *e;
entry             645 drivers/md/dm-cache-policy-smq.c 		       struct entry *e, struct entry *prev)
entry             656 drivers/md/dm-cache-policy-smq.c static struct entry *h_lookup(struct smq_hash_table *ht, dm_oblock_t oblock)
entry             658 drivers/md/dm-cache-policy-smq.c 	struct entry *e, *prev;
entry             674 drivers/md/dm-cache-policy-smq.c static void h_remove(struct smq_hash_table *ht, struct entry *e)
entry             677 drivers/md/dm-cache-policy-smq.c 	struct entry *prev;
entry             712 drivers/md/dm-cache-policy-smq.c static void init_entry(struct entry *e)
entry             728 drivers/md/dm-cache-policy-smq.c static struct entry *alloc_entry(struct entry_alloc *ea)
entry             730 drivers/md/dm-cache-policy-smq.c 	struct entry *e;
entry             745 drivers/md/dm-cache-policy-smq.c static struct entry *alloc_particular_entry(struct entry_alloc *ea, unsigned i)
entry             747 drivers/md/dm-cache-policy-smq.c 	struct entry *e = __get_entry(ea->es, ea->begin + i);
entry             758 drivers/md/dm-cache-policy-smq.c static void free_entry(struct entry_alloc *ea, struct entry *e)
entry             773 drivers/md/dm-cache-policy-smq.c static unsigned get_index(struct entry_alloc *ea, struct entry *e)
entry             778 drivers/md/dm-cache-policy-smq.c static struct entry *get_entry(struct entry_alloc *ea, unsigned index)
entry             862 drivers/md/dm-cache-policy-smq.c static struct entry *get_sentinel(struct entry_alloc *ea, unsigned level, bool which)
entry             867 drivers/md/dm-cache-policy-smq.c static struct entry *writeback_sentinel(struct smq_policy *mq, unsigned level)
entry             872 drivers/md/dm-cache-policy-smq.c static struct entry *demote_sentinel(struct smq_policy *mq, unsigned level)
entry             881 drivers/md/dm-cache-policy-smq.c 	struct entry *sentinel;
entry             894 drivers/md/dm-cache-policy-smq.c 	struct entry *sentinel;
entry             921 drivers/md/dm-cache-policy-smq.c 	struct entry *sentinel;
entry             950 drivers/md/dm-cache-policy-smq.c static void del_queue(struct smq_policy *mq, struct entry *e)
entry             955 drivers/md/dm-cache-policy-smq.c static void push_queue(struct smq_policy *mq, struct entry *e)
entry             964 drivers/md/dm-cache-policy-smq.c static void push(struct smq_policy *mq, struct entry *e)
entry             971 drivers/md/dm-cache-policy-smq.c static void push_queue_front(struct smq_policy *mq, struct entry *e)
entry             979 drivers/md/dm-cache-policy-smq.c static void push_front(struct smq_policy *mq, struct entry *e)
entry             986 drivers/md/dm-cache-policy-smq.c static dm_cblock_t infer_cblock(struct smq_policy *mq, struct entry *e)
entry             991 drivers/md/dm-cache-policy-smq.c static void requeue(struct smq_policy *mq, struct entry *e)
entry            1160 drivers/md/dm-cache-policy-smq.c static void mark_pending(struct smq_policy *mq, struct entry *e)
entry            1168 drivers/md/dm-cache-policy-smq.c static void clear_pending(struct smq_policy *mq, struct entry *e)
entry            1178 drivers/md/dm-cache-policy-smq.c 	struct entry *e;
entry            1201 drivers/md/dm-cache-policy-smq.c 	struct entry *e;
entry            1230 drivers/md/dm-cache-policy-smq.c 	struct entry *e;
entry            1280 drivers/md/dm-cache-policy-smq.c static enum promote_result should_promote(struct smq_policy *mq, struct entry *hs_e,
entry            1299 drivers/md/dm-cache-policy-smq.c static struct entry *update_hotspot_queue(struct smq_policy *mq, dm_oblock_t b)
entry            1303 drivers/md/dm-cache-policy-smq.c 	struct entry *e = h_lookup(&mq->hotspot_table, hb);
entry            1369 drivers/md/dm-cache-policy-smq.c 	struct entry *e, *hs_e;
entry            1462 drivers/md/dm-cache-policy-smq.c 	struct entry *e = get_entry(&mq->cache_alloc,
entry            1519 drivers/md/dm-cache-policy-smq.c 	struct entry *e = get_entry(&mq->cache_alloc, from_cblock(cblock));
entry            1560 drivers/md/dm-cache-policy-smq.c 	struct entry *e;
entry            1580 drivers/md/dm-cache-policy-smq.c 	struct entry *e = get_entry(&mq->cache_alloc, from_cblock(cblock));
entry            1595 drivers/md/dm-cache-policy-smq.c 	struct entry *e = get_entry(&mq->cache_alloc, from_cblock(cblock));
entry             198 drivers/md/dm-cache-target.c 	list_for_each_entry_safe(ws, tmp, &work_items, entry) {
entry             201 drivers/md/dm-cache-target.c 		INIT_LIST_HEAD(&ws->entry); /* to avoid a WARN_ON */
entry             246 drivers/md/dm-cache-target.c 	list_add_tail(&k->ws.entry, &b->work_items);
entry             211 drivers/md/dm-log-writes.c static int write_metadata(struct log_writes_c *lc, void *entry,
entry             241 drivers/md/dm-log-writes.c 	memcpy(ptr, entry, entrylen);
entry             263 drivers/md/dm-log-writes.c static int write_inline_data(struct log_writes_c *lc, void *entry,
entry             335 drivers/md/dm-log-writes.c 	struct log_write_entry entry;
entry             339 drivers/md/dm-log-writes.c 	entry.sector = cpu_to_le64(block->sector);
entry             340 drivers/md/dm-log-writes.c 	entry.nr_sectors = cpu_to_le64(block->nr_sectors);
entry             341 drivers/md/dm-log-writes.c 	entry.flags = cpu_to_le64(block->flags);
entry             342 drivers/md/dm-log-writes.c 	entry.data_len = cpu_to_le64(block->datalen);
entry             345 drivers/md/dm-log-writes.c 	if (write_metadata(lc, &entry, sizeof(entry), block->data,
entry             354 drivers/md/dm-log-writes.c 		if (write_inline_data(lc, &entry, sizeof(entry), block->data,
entry             514 drivers/md/dm-stats.c static void dm_stat_for_entry(struct dm_stat *s, size_t entry,
entry             519 drivers/md/dm-stats.c 	struct dm_stat_shared *shared = &s->stat_shared[entry];
entry             543 drivers/md/dm-stats.c 	p = &s->stat_percpu[smp_processor_id()][entry];
entry             590 drivers/md/dm-stats.c 	size_t entry;
entry             605 drivers/md/dm-stats.c 	entry = rel_sector;
entry             607 drivers/md/dm-stats.c 		if (WARN_ON_ONCE(entry >= s->n_entries)) {
entry             614 drivers/md/dm-stats.c 		dm_stat_for_entry(s, entry, bi_rw, fragment_len,
entry             617 drivers/md/dm-stats.c 		entry++;
entry            3485 drivers/md/md.c 	struct rdev_sysfs_entry *entry = container_of(attr, struct rdev_sysfs_entry, attr);
entry            3488 drivers/md/md.c 	if (!entry->show)
entry            3492 drivers/md/md.c 	return entry->show(rdev, page);
entry            3499 drivers/md/md.c 	struct rdev_sysfs_entry *entry = container_of(attr, struct rdev_sysfs_entry, attr);
entry            3504 drivers/md/md.c 	if (!entry->store)
entry            3513 drivers/md/md.c 			rv = entry->store(rdev, page, length);
entry            5325 drivers/md/md.c 	struct md_sysfs_entry *entry = container_of(attr, struct md_sysfs_entry, attr);
entry            5329 drivers/md/md.c 	if (!entry->show)
entry            5339 drivers/md/md.c 	rv = entry->show(mddev, page);
entry            5348 drivers/md/md.c 	struct md_sysfs_entry *entry = container_of(attr, struct md_sysfs_entry, attr);
entry            5352 drivers/md/md.c 	if (!entry->store)
entry            5363 drivers/md/md.c 	rv = entry->store(mddev, page, length);
entry              99 drivers/md/persistent-data/dm-array.c 	unsigned char *entry = (unsigned char *) (ab + 1);
entry             101 drivers/md/persistent-data/dm-array.c 	entry += index * info->value_type.size;
entry             103 drivers/md/persistent-data/dm-array.c 	return entry;
entry             775 drivers/md/persistent-data/dm-array.c 	unsigned entry, max_entries;
entry             784 drivers/md/persistent-data/dm-array.c 	entry = index % max_entries;
entry             785 drivers/md/persistent-data/dm-array.c 	if (entry >= le32_to_cpu(ab->nr_entries))
entry             788 drivers/md/persistent-data/dm-array.c 		memcpy(value_le, element_at(info, ab, entry),
entry             804 drivers/md/persistent-data/dm-array.c 	unsigned entry;
entry             816 drivers/md/persistent-data/dm-array.c 	entry = index % max_entries;
entry             817 drivers/md/persistent-data/dm-array.c 	if (entry >= le32_to_cpu(ab->nr_entries)) {
entry             822 drivers/md/persistent-data/dm-array.c 	old_value = element_at(info, ab, entry);
entry             105 drivers/media/cec/cec-adap.c 	struct cec_event_entry *entry;
entry             116 drivers/media/cec/cec-adap.c 		entry = &fh->core_events[ev_idx];
entry             118 drivers/media/cec/cec-adap.c 		entry = kmalloc(sizeof(*entry), GFP_KERNEL);
entry             119 drivers/media/cec/cec-adap.c 	if (entry) {
entry             122 drivers/media/cec/cec-adap.c 			entry->ev.lost_msgs.lost_msgs +=
entry             126 drivers/media/cec/cec-adap.c 		entry->ev = *new_ev;
entry             127 drivers/media/cec/cec-adap.c 		entry->ev.ts = ts;
entry             131 drivers/media/cec/cec-adap.c 			list_add_tail(&entry->list, &fh->events[ev_idx]);
entry             138 drivers/media/cec/cec-adap.c 			list_add_tail(&entry->list, &fh->events[ev_idx]);
entry             140 drivers/media/cec/cec-adap.c 			entry = list_first_entry(&fh->events[ev_idx],
entry             142 drivers/media/cec/cec-adap.c 			list_del(&entry->list);
entry             143 drivers/media/cec/cec-adap.c 			kfree(entry);
entry             147 drivers/media/cec/cec-adap.c 	entry = list_first_entry_or_null(&fh->events[ev_idx],
entry             149 drivers/media/cec/cec-adap.c 	if (entry)
entry             150 drivers/media/cec/cec-adap.c 		entry->ev.flags |= CEC_EVENT_FL_DROPPED_EVENTS;
entry             237 drivers/media/cec/cec-adap.c 	struct cec_msg_entry *entry;
entry             240 drivers/media/cec/cec-adap.c 	entry = kmalloc(sizeof(*entry), GFP_KERNEL);
entry             241 drivers/media/cec/cec-adap.c 	if (entry) {
entry             242 drivers/media/cec/cec-adap.c 		entry->msg = *msg;
entry             244 drivers/media/cec/cec-adap.c 		list_add_tail(&entry->list, &fh->msgs);
entry             258 drivers/media/cec/cec-adap.c 		entry = list_first_entry(&fh->msgs, struct cec_msg_entry, list);
entry             259 drivers/media/cec/cec-adap.c 		list_del(&entry->list);
entry             260 drivers/media/cec/cec-adap.c 		kfree(entry);
entry             229 drivers/media/cec/cec-api.c 			struct cec_msg_entry *entry =
entry             233 drivers/media/cec/cec-api.c 			list_del(&entry->list);
entry             234 drivers/media/cec/cec-api.c 			*msg = entry->msg;
entry             235 drivers/media/cec/cec-api.c 			kfree(entry);
entry             308 drivers/media/cec/cec-api.c 		struct cec_event_entry *entry =
entry             312 drivers/media/cec/cec-api.c 		if (entry && entry->ev.ts <= ts) {
entry             313 drivers/media/cec/cec-api.c 			ev = entry;
entry             650 drivers/media/cec/cec-api.c 		struct cec_msg_entry *entry =
entry             653 drivers/media/cec/cec-api.c 		list_del(&entry->list);
entry             654 drivers/media/cec/cec-api.c 		kfree(entry);
entry             658 drivers/media/cec/cec-api.c 			struct cec_event_entry *entry =
entry             662 drivers/media/cec/cec-api.c 			list_del(&entry->list);
entry             663 drivers/media/cec/cec-api.c 			kfree(entry);
entry              31 drivers/media/common/siano/smscoreapi.c 	struct list_head entry;
entry              36 drivers/media/common/siano/smscoreapi.c 	struct list_head entry;
entry              42 drivers/media/common/siano/smscoreapi.c 	struct list_head entry;
entry             409 drivers/media/common/siano/smscoreapi.c 	struct list_head entry;
entry             429 drivers/media/common/siano/smscoreapi.c 	struct smscore_registry_entry_t *entry;
entry             436 drivers/media/common/siano/smscoreapi.c 		entry = (struct smscore_registry_entry_t *) next;
entry             437 drivers/media/common/siano/smscoreapi.c 		if (!strncmp(entry->devpath, devpath, sizeof(entry->devpath))) {
entry             439 drivers/media/common/siano/smscoreapi.c 			return entry;
entry             442 drivers/media/common/siano/smscoreapi.c 	entry = kmalloc(sizeof(*entry), GFP_KERNEL);
entry             443 drivers/media/common/siano/smscoreapi.c 	if (entry) {
entry             444 drivers/media/common/siano/smscoreapi.c 		entry->mode = default_mode;
entry             445 drivers/media/common/siano/smscoreapi.c 		strscpy(entry->devpath, devpath, sizeof(entry->devpath));
entry             446 drivers/media/common/siano/smscoreapi.c 		list_add(&entry->entry, &g_smscore_registry);
entry             450 drivers/media/common/siano/smscoreapi.c 	return entry;
entry             455 drivers/media/common/siano/smscoreapi.c 	struct smscore_registry_entry_t *entry;
entry             457 drivers/media/common/siano/smscoreapi.c 	entry = smscore_find_registry(devpath);
entry             458 drivers/media/common/siano/smscoreapi.c 	if (entry)
entry             459 drivers/media/common/siano/smscoreapi.c 		return entry->mode;
entry             469 drivers/media/common/siano/smscoreapi.c 	struct smscore_registry_entry_t *entry;
entry             471 drivers/media/common/siano/smscoreapi.c 	entry = smscore_find_registry(devpath);
entry             472 drivers/media/common/siano/smscoreapi.c 	if (entry)
entry             473 drivers/media/common/siano/smscoreapi.c 		return entry->type;
entry             482 drivers/media/common/siano/smscoreapi.c 	struct smscore_registry_entry_t *entry;
entry             484 drivers/media/common/siano/smscoreapi.c 	entry = smscore_find_registry(devpath);
entry             485 drivers/media/common/siano/smscoreapi.c 	if (entry)
entry             486 drivers/media/common/siano/smscoreapi.c 		entry->mode = mode;
entry             494 drivers/media/common/siano/smscoreapi.c 	struct smscore_registry_entry_t *entry;
entry             496 drivers/media/common/siano/smscoreapi.c 	entry = smscore_find_registry(devpath);
entry             497 drivers/media/common/siano/smscoreapi.c 	if (entry)
entry             498 drivers/media/common/siano/smscoreapi.c 		entry->type = type;
entry             545 drivers/media/common/siano/smscoreapi.c 			list_add(&notifyee->entry, &g_smscore_notifyees);
entry             577 drivers/media/common/siano/smscoreapi.c 			list_del(&notifyee->entry);
entry             605 drivers/media/common/siano/smscoreapi.c 	list_for_each_entry(elem, &g_smscore_notifyees, entry) {
entry             660 drivers/media/common/siano/smscoreapi.c 	INIT_LIST_HEAD(&dev->entry);
entry             736 drivers/media/common/siano/smscoreapi.c 	list_add(&dev->entry, &g_smscore_devices);
entry            1214 drivers/media/common/siano/smscoreapi.c 			list_del(&cb->entry);
entry            1245 drivers/media/common/siano/smscoreapi.c 	list_del(&coredev->entry);
entry            1455 drivers/media/common/siano/smscoreapi.c 	list_for_each_entry(client, first, entry) {
entry            1457 drivers/media/common/siano/smscoreapi.c 		list_for_each_entry(client_id, firstid, entry) {
entry            1640 drivers/media/common/siano/smscoreapi.c 		list_del(&cb->entry);
entry            1667 drivers/media/common/siano/smscoreapi.c 	list_add_locked(&cb->entry, &coredev->buffers, &coredev->bufferslock);
entry            1696 drivers/media/common/siano/smscoreapi.c 	list_add_locked(&listentry->entry, &client->idlist,
entry            1736 drivers/media/common/siano/smscoreapi.c 	list_add_locked(&newclient->entry, &coredev->clients,
entry            1766 drivers/media/common/siano/smscoreapi.c 		list_del(&identry->entry);
entry            1772 drivers/media/common/siano/smscoreapi.c 	list_del(&client->entry);
entry            2142 drivers/media/common/siano/smscoreapi.c 		list_del(&notifyee->entry);
entry            2149 drivers/media/common/siano/smscoreapi.c 		struct smscore_registry_entry_t *entry =
entry            2153 drivers/media/common/siano/smscoreapi.c 		list_del(&entry->entry);
entry            2154 drivers/media/common/siano/smscoreapi.c 		kfree(entry);
entry             114 drivers/media/common/siano/smscoreapi.h 	struct list_head entry;
entry             153 drivers/media/common/siano/smscoreapi.h 	struct list_head entry;
entry             619 drivers/media/common/siano/smsdvb-main.c 	list_del(&client->entry);
entry            1156 drivers/media/common/siano/smsdvb-main.c 	list_add(&client->entry, &g_smsdvb_clients);
entry              21 drivers/media/common/siano/smsdvb.h 	struct list_head entry;
entry             660 drivers/media/dvb-core/dvb_demux.c 	struct dvb_demux_feed *entry;
entry             662 drivers/media/dvb-core/dvb_demux.c 	list_for_each_entry(entry, &feed->demux->feed_list, list_head)
entry             663 drivers/media/dvb-core/dvb_demux.c 		if (entry == feed)
entry             814 drivers/media/dvb-core/dvbdev.c 	struct list_head *entry;
entry             815 drivers/media/dvb-core/dvbdev.c 	list_for_each(entry, &dvb_adapter_list) {
entry             817 drivers/media/dvb-core/dvbdev.c 		adap = list_entry(entry, struct dvb_adapter, list_head);
entry             618 drivers/media/i2c/m5mols/m5mols_core.c 	fd->entry[0].length = info->cap.buf_size;
entry             619 drivers/media/i2c/m5mols/m5mols_core.c 	fd->entry[0].pixelcode = info->ffmt[M5MOLS_RESTYPE_CAPTURE].code;
entry             622 drivers/media/i2c/m5mols/m5mols_core.c 	fd->entry[0].flags = V4L2_MBUS_FRAME_DESC_FL_LEN_MAX;
entry             637 drivers/media/i2c/m5mols/m5mols_core.c 	fd->entry[0].flags = V4L2_MBUS_FRAME_DESC_FL_LEN_MAX;
entry             639 drivers/media/i2c/m5mols/m5mols_core.c 	fd->entry[0].length = clamp_t(u32, fd->entry[0].length,
entry             643 drivers/media/i2c/m5mols/m5mols_core.c 	info->cap.buf_size = fd->entry[0].length;
entry            1156 drivers/media/i2c/s5c73m3/s5c73m3-core.c 		fd->entry[i] = state->frame_desc.entry[i];
entry            1172 drivers/media/i2c/s5c73m3/s5c73m3-core.c 	fd->entry[0].length = 10 * SZ_1M;
entry            1173 drivers/media/i2c/s5c73m3/s5c73m3-core.c 	fd->entry[1].length = max_t(u32, fd->entry[1].length,
entry            1179 drivers/media/i2c/s5c73m3/s5c73m3-core.c 		frame_desc->entry[i] = fd->entry[i];
entry             679 drivers/media/pci/cx18/cx18-ioctl.c 	e_idx = &idx->entry[idx->entries];
entry             699 drivers/media/pci/cx18/cx18-ioctl.c 		e_idx = &idx->entry[idx->entries];
entry             137 drivers/media/pci/intel/ipu3/ipu3-cio2.c 				   struct cio2_fbpt_entry entry[CIO2_MAX_LOPS])
entry             151 drivers/media/pci/intel/ipu3/ipu3-cio2.c 	entry[0].first_entry.ctrl = CIO2_FBPT_CTRL_VALID |
entry             158 drivers/media/pci/intel/ipu3/ipu3-cio2.c 				       entry[CIO2_MAX_LOPS])
entry             162 drivers/media/pci/intel/ipu3/ipu3-cio2.c 	entry[0].first_entry.first_page_offset = 0;
entry             163 drivers/media/pci/intel/ipu3/ipu3-cio2.c 	entry[1].second_entry.num_of_pages =
entry             165 drivers/media/pci/intel/ipu3/ipu3-cio2.c 	entry[1].second_entry.last_page_available_bytes = CIO2_PAGE_SIZE - 1;
entry             168 drivers/media/pci/intel/ipu3/ipu3-cio2.c 		entry[i].lop_page_addr = cio2->dummy_lop_bus_addr >> PAGE_SHIFT;
entry             170 drivers/media/pci/intel/ipu3/ipu3-cio2.c 	cio2_fbpt_entry_enable(cio2, entry);
entry             177 drivers/media/pci/intel/ipu3/ipu3-cio2.c 				     entry[CIO2_MAX_LOPS])
entry             183 drivers/media/pci/intel/ipu3/ipu3-cio2.c 	entry[0].first_entry.first_page_offset = b->offset;
entry             184 drivers/media/pci/intel/ipu3/ipu3-cio2.c 	remaining = length + entry[0].first_entry.first_page_offset;
entry             185 drivers/media/pci/intel/ipu3/ipu3-cio2.c 	entry[1].second_entry.num_of_pages =
entry             195 drivers/media/pci/intel/ipu3/ipu3-cio2.c 	entry[1].second_entry.last_page_available_bytes =
entry             203 drivers/media/pci/intel/ipu3/ipu3-cio2.c 		entry->lop_page_addr = b->lop_bus_addr[i] >> PAGE_SHIFT;
entry             205 drivers/media/pci/intel/ipu3/ipu3-cio2.c 		entry++;
entry             212 drivers/media/pci/intel/ipu3/ipu3-cio2.c 	entry->lop_page_addr = cio2->dummy_lop_bus_addr >> PAGE_SHIFT;
entry             214 drivers/media/pci/intel/ipu3/ipu3-cio2.c 	cio2_fbpt_entry_enable(cio2, entry);
entry             558 drivers/media/pci/intel/ipu3/ipu3-cio2.c 		struct cio2_fbpt_entry *const entry =
entry             562 drivers/media/pci/intel/ipu3/ipu3-cio2.c 		if (entry->first_entry.ctrl & CIO2_FBPT_CTRL_VALID)
entry             567 drivers/media/pci/intel/ipu3/ipu3-cio2.c 			unsigned int bytes = entry[1].second_entry.num_of_bytes;
entry             584 drivers/media/pci/intel/ipu3/ipu3-cio2.c 		cio2_fbpt_entry_init_dummy(cio2, entry);
entry             904 drivers/media/pci/intel/ipu3/ipu3-cio2.c 	struct cio2_fbpt_entry *entry;
entry             951 drivers/media/pci/intel/ipu3/ipu3-cio2.c 			entry = &q->fbpt[next * CIO2_MAX_LOPS];
entry             952 drivers/media/pci/intel/ipu3/ipu3-cio2.c 			cio2_fbpt_entry_init_buf(cio2, b, entry);
entry            1257 drivers/media/pci/ivtv/ivtv-ioctl.c 	struct v4l2_enc_idx_entry *e = idx->entry;
entry             277 drivers/media/pci/saa7134/saa7134-core.c 			list_add_tail(&buf->entry, &q->queue);
entry             280 drivers/media/pci/saa7134/saa7134-core.c 					  entry);
entry             285 drivers/media/pci/saa7134/saa7134-core.c 		list_add_tail(&buf->entry, &q->queue);
entry             314 drivers/media/pci/saa7134/saa7134-core.c 		buf = list_entry(q->queue.next, struct saa7134_buf, entry);
entry             317 drivers/media/pci/saa7134/saa7134-core.c 		list_del(&buf->entry);
entry             319 drivers/media/pci/saa7134/saa7134-core.c 			next = list_entry(q->queue.next, struct saa7134_buf, entry);
entry             364 drivers/media/pci/saa7134/saa7134-core.c 			 tmp = list_entry(pos, struct saa7134_buf, entry);
entry            1394 drivers/media/pci/saa7134/saa7134-core.c 					  entry);
entry             137 drivers/media/pci/saa7134/saa7134-ts.c 		list_for_each_entry_safe(buf, tmp, &dmaq->queue, entry) {
entry             138 drivers/media/pci/saa7134/saa7134-ts.c 			list_del(&buf->entry);
entry             989 drivers/media/pci/saa7134/saa7134-video.c 		list_for_each_entry_safe(buf, tmp, &dmaq->queue, entry) {
entry             990 drivers/media/pci/saa7134/saa7134-video.c 			list_del(&buf->entry);
entry             473 drivers/media/pci/saa7134/saa7134.h 	struct list_head	entry;
entry             872 drivers/media/platform/exynos4-is/fimc-capture.c 		fd.entry[i].length = plane_fmt[i].sizeimage;
entry             887 drivers/media/platform/exynos4-is/fimc-capture.c 		plane_fmt[i].sizeimage = fd.entry[i].length;
entry             889 drivers/media/platform/exynos4-is/fimc-capture.c 	if (fd.entry[0].length > FIMC_MAX_JPEG_BUF_SIZE) {
entry             891 drivers/media/platform/exynos4-is/fimc-capture.c 			 fd.entry[0].length);
entry             516 drivers/media/rc/rc-main.c 	struct rc_map_table *entry;
entry             535 drivers/media/rc/rc-main.c 		entry = &rc_map->scan[index];
entry             538 drivers/media/rc/rc-main.c 		ke->keycode = entry->keycode;
entry             539 drivers/media/rc/rc-main.c 		ke->len = sizeof(entry->scancode);
entry             540 drivers/media/rc/rc-main.c 		memcpy(ke->scancode, &entry->scancode, sizeof(entry->scancode));
entry              39 drivers/media/usb/siano/smsusb.c 	struct list_head entry;
entry            1164 drivers/memstick/core/ms_block.c 	p->entry.disabled_block.start_addr
entry            1165 drivers/memstick/core/ms_block.c 		= be32_to_cpu(p->entry.disabled_block.start_addr);
entry            1166 drivers/memstick/core/ms_block.c 	p->entry.disabled_block.data_size
entry            1167 drivers/memstick/core/ms_block.c 		= be32_to_cpu(p->entry.disabled_block.data_size);
entry            1168 drivers/memstick/core/ms_block.c 	p->entry.cis_idi.start_addr
entry            1169 drivers/memstick/core/ms_block.c 		= be32_to_cpu(p->entry.cis_idi.start_addr);
entry            1170 drivers/memstick/core/ms_block.c 	p->entry.cis_idi.data_size
entry            1171 drivers/memstick/core/ms_block.c 		= be32_to_cpu(p->entry.cis_idi.data_size);
entry            1266 drivers/memstick/core/ms_block.c 	data_size = boot_block->entry.disabled_block.data_size;
entry            1268 drivers/memstick/core/ms_block.c 			boot_block->entry.disabled_block.start_addr;
entry             141 drivers/memstick/core/ms_block.h 	struct ms_system_entry   entry;
entry             129 drivers/mfd/dln2.c 	struct dln2_event_cb_entry *i, *entry;
entry             133 drivers/mfd/dln2.c 	entry = kzalloc(sizeof(*entry), GFP_KERNEL);
entry             134 drivers/mfd/dln2.c 	if (!entry)
entry             137 drivers/mfd/dln2.c 	entry->id = id;
entry             138 drivers/mfd/dln2.c 	entry->callback = event_cb;
entry             139 drivers/mfd/dln2.c 	entry->pdev = pdev;
entry             151 drivers/mfd/dln2.c 		list_add_rcu(&entry->list, &dln2->event_cb_list);
entry             156 drivers/mfd/dln2.c 		kfree(entry);
entry             155 drivers/mfd/syscon.c 	struct syscon *entry, *syscon = NULL;
entry             159 drivers/mfd/syscon.c 	list_for_each_entry(entry, &syscon_list, list)
entry             160 drivers/mfd/syscon.c 		if (entry->np == np) {
entry             161 drivers/mfd/syscon.c 			syscon = entry;
entry             703 drivers/mfd/timberdale.c 		msix_entries[i].entry = i;
entry              37 drivers/misc/cxl/fault.c 	unsigned int entry;
entry              47 drivers/misc/cxl/fault.c 	for (entry = 0, sste = primary; entry < 8; entry++, sste++) {
entry              98 drivers/misc/habanalabs/debugfs.c 	struct hl_debugfs_entry *entry = s->private;
entry              99 drivers/misc/habanalabs/debugfs.c 	struct hl_dbg_device_entry *dev_entry = entry->dev_entry;
entry             129 drivers/misc/habanalabs/debugfs.c 	struct hl_debugfs_entry *entry = s->private;
entry             130 drivers/misc/habanalabs/debugfs.c 	struct hl_dbg_device_entry *dev_entry = entry->dev_entry;
entry             160 drivers/misc/habanalabs/debugfs.c 	struct hl_debugfs_entry *entry = s->private;
entry             161 drivers/misc/habanalabs/debugfs.c 	struct hl_dbg_device_entry *dev_entry = entry->dev_entry;
entry             195 drivers/misc/habanalabs/debugfs.c 	struct hl_debugfs_entry *entry = s->private;
entry             196 drivers/misc/habanalabs/debugfs.c 	struct hl_dbg_device_entry *dev_entry = entry->dev_entry;
entry             226 drivers/misc/habanalabs/debugfs.c 	struct hl_debugfs_entry *entry = s->private;
entry             227 drivers/misc/habanalabs/debugfs.c 	struct hl_dbg_device_entry *dev_entry = entry->dev_entry;
entry             355 drivers/misc/habanalabs/debugfs.c 	struct hl_debugfs_entry *entry = s->private;
entry             356 drivers/misc/habanalabs/debugfs.c 	struct hl_dbg_device_entry *dev_entry = entry->dev_entry;
entry             464 drivers/misc/habanalabs/debugfs.c 	struct hl_debugfs_entry *entry = s->private;
entry             465 drivers/misc/habanalabs/debugfs.c 	struct hl_dbg_device_entry *dev_entry = entry->dev_entry;
entry             505 drivers/misc/habanalabs/debugfs.c 	struct hl_debugfs_entry *entry = s->private;
entry             506 drivers/misc/habanalabs/debugfs.c 	struct hl_dbg_device_entry *dev_entry = entry->dev_entry;
entry             604 drivers/misc/habanalabs/debugfs.c 	struct hl_dbg_device_entry *entry = file_inode(f)->i_private;
entry             605 drivers/misc/habanalabs/debugfs.c 	struct hl_device *hdev = entry->hdev;
entry             607 drivers/misc/habanalabs/debugfs.c 	u64 addr = entry->addr;
entry             634 drivers/misc/habanalabs/debugfs.c 	struct hl_dbg_device_entry *entry = file_inode(f)->i_private;
entry             635 drivers/misc/habanalabs/debugfs.c 	struct hl_device *hdev = entry->hdev;
entry             636 drivers/misc/habanalabs/debugfs.c 	u64 addr = entry->addr;
entry             663 drivers/misc/habanalabs/debugfs.c 	struct hl_dbg_device_entry *entry = file_inode(f)->i_private;
entry             664 drivers/misc/habanalabs/debugfs.c 	struct hl_device *hdev = entry->hdev;
entry             687 drivers/misc/habanalabs/debugfs.c 	struct hl_dbg_device_entry *entry = file_inode(f)->i_private;
entry             688 drivers/misc/habanalabs/debugfs.c 	struct hl_device *hdev = entry->hdev;
entry             715 drivers/misc/habanalabs/debugfs.c 	struct hl_dbg_device_entry *entry = file_inode(f)->i_private;
entry             716 drivers/misc/habanalabs/debugfs.c 	struct hl_device *hdev = entry->hdev;
entry             724 drivers/misc/habanalabs/debugfs.c 	rc = hl_debugfs_i2c_read(hdev, entry->i2c_bus, entry->i2c_addr,
entry             725 drivers/misc/habanalabs/debugfs.c 			entry->i2c_reg, &val);
entry             729 drivers/misc/habanalabs/debugfs.c 			entry->i2c_bus, entry->i2c_addr, entry->i2c_reg);
entry             743 drivers/misc/habanalabs/debugfs.c 	struct hl_dbg_device_entry *entry = file_inode(f)->i_private;
entry             744 drivers/misc/habanalabs/debugfs.c 	struct hl_device *hdev = entry->hdev;
entry             752 drivers/misc/habanalabs/debugfs.c 	rc = hl_debugfs_i2c_write(hdev, entry->i2c_bus, entry->i2c_addr,
entry             753 drivers/misc/habanalabs/debugfs.c 			entry->i2c_reg, value);
entry             757 drivers/misc/habanalabs/debugfs.c 			value, entry->i2c_bus, entry->i2c_addr, entry->i2c_reg);
entry             767 drivers/misc/habanalabs/debugfs.c 	struct hl_dbg_device_entry *entry = file_inode(f)->i_private;
entry             768 drivers/misc/habanalabs/debugfs.c 	struct hl_device *hdev = entry->hdev;
entry             786 drivers/misc/habanalabs/debugfs.c 	struct hl_dbg_device_entry *entry = file_inode(f)->i_private;
entry             787 drivers/misc/habanalabs/debugfs.c 	struct hl_device *hdev = entry->hdev;
entry             805 drivers/misc/habanalabs/debugfs.c 	struct hl_dbg_device_entry *entry = file_inode(f)->i_private;
entry             806 drivers/misc/habanalabs/debugfs.c 	struct hl_device *hdev = entry->hdev;
entry             832 drivers/misc/habanalabs/debugfs.c 	struct hl_dbg_device_entry *entry = file_inode(f)->i_private;
entry             833 drivers/misc/habanalabs/debugfs.c 	struct hl_device *hdev = entry->hdev;
entry             942 drivers/misc/habanalabs/debugfs.c 	struct hl_debugfs_entry *entry;
entry            1031 drivers/misc/habanalabs/debugfs.c 	for (i = 0, entry = dev_entry->entry_arr ; i < count ; i++, entry++) {
entry            1036 drivers/misc/habanalabs/debugfs.c 					entry,
entry            1038 drivers/misc/habanalabs/debugfs.c 		entry->dent = ent;
entry            1039 drivers/misc/habanalabs/debugfs.c 		entry->info_ent = &hl_debugfs_list[i];
entry            1040 drivers/misc/habanalabs/debugfs.c 		entry->dev_entry = dev_entry;
entry            1046 drivers/misc/habanalabs/debugfs.c 	struct hl_dbg_device_entry *entry = &hdev->hl_debugfs;
entry            1048 drivers/misc/habanalabs/debugfs.c 	debugfs_remove_recursive(entry->root);
entry            1050 drivers/misc/habanalabs/debugfs.c 	mutex_destroy(&entry->file_mutex);
entry            1051 drivers/misc/habanalabs/debugfs.c 	kfree(entry->entry_arr);
entry              40 drivers/misc/hpilo.c static inline int get_entry_id(int entry)
entry              42 drivers/misc/hpilo.c 	return (entry & ENTRY_MASK_DESCRIPTOR) >> ENTRY_BITPOS_DESCRIPTOR;
entry              45 drivers/misc/hpilo.c static inline int get_entry_len(int entry)
entry              47 drivers/misc/hpilo.c 	return ((entry & ENTRY_MASK_QWORDS) >> ENTRY_BITPOS_QWORDS) << 3;
entry              71 drivers/misc/hpilo.c static int fifo_enqueue(struct ilo_hwinfo *hw, char *fifobar, int entry)
entry              81 drivers/misc/hpilo.c 				(entry & ENTRY_MASK_NOSTATE) | fifo_q->merge;
entry              90 drivers/misc/hpilo.c static int fifo_dequeue(struct ilo_hwinfo *hw, char *fifobar, int *entry)
entry             100 drivers/misc/hpilo.c 		if (entry)
entry             101 drivers/misc/hpilo.c 			*entry = c & ENTRY_MASK_NOSTATE;
entry             133 drivers/misc/hpilo.c 	int entry;
entry             140 drivers/misc/hpilo.c 	entry = mk_entry(id, len);
entry             141 drivers/misc/hpilo.c 	return fifo_enqueue(hw, fifobar, entry);
entry             148 drivers/misc/hpilo.c 	int entry = 0, pkt_id = 0;
entry             159 drivers/misc/hpilo.c 	ret = fifo_dequeue(hw, fifobar, &entry);
entry             161 drivers/misc/hpilo.c 		pkt_id = get_entry_id(entry);
entry             165 drivers/misc/hpilo.c 			*len = get_entry_len(entry);
entry             580 drivers/misc/ibmasm/ibmasmfs.c 	struct list_head *entry;
entry             583 drivers/misc/ibmasm/ibmasmfs.c 	list_for_each(entry, &service_processors) {
entry             586 drivers/misc/ibmasm/ibmasmfs.c 		sp = list_entry(entry, struct service_processor, node);
entry              36 drivers/misc/mic/host/mic_debugfs.c 				   " ",  i, smpt_info->entry[i].dma_addr,
entry              37 drivers/misc/mic/host/mic_debugfs.c 				   smpt_info->entry[i].ref_count);
entry              63 drivers/misc/mic/host/mic_debugfs.c 	u16 entry;
entry              70 drivers/misc/mic/host/mic_debugfs.c 				entry = mdev->irq_info.msix_entries[i].entry;
entry              73 drivers/misc/mic/host/mic_debugfs.c 				entry = 0;
entry              77 drivers/misc/mic/host/mic_debugfs.c 			reg = mdev->intr_ops->read_msi_to_src_map(mdev, entry);
entry              80 drivers/misc/mic/host/mic_debugfs.c 				   "IRQ:", vector, "Entry:", entry, i, reg);
entry             200 drivers/misc/mic/host/mic_intr.c 		mdev->irq_info.msix_entries[i].entry = i;
entry             429 drivers/misc/mic/host/mic_intr.c 	u16 entry;
entry             458 drivers/misc/mic/host/mic_intr.c 		entry = msix->entry;
entry             459 drivers/misc/mic/host/mic_intr.c 		mdev->irq_info.mic_msi_map[entry] |= BIT(offset);
entry             461 drivers/misc/mic/host/mic_intr.c 				entry, offset, true);
entry             462 drivers/misc/mic/host/mic_intr.c 		cookie = MK_COOKIE(entry, offset);
entry             475 drivers/misc/mic/host/mic_intr.c 		entry = 0;
entry             477 drivers/misc/mic/host/mic_intr.c 			mdev->irq_info.mic_msi_map[entry] |= (1 << offset);
entry             479 drivers/misc/mic/host/mic_intr.c 				entry, offset, true);
entry             481 drivers/misc/mic/host/mic_intr.c 		cookie = MK_COOKIE(entry, intr_cb->cb_id);
entry             505 drivers/misc/mic/host/mic_intr.c 	u32 entry;
entry             510 drivers/misc/mic/host/mic_intr.c 	entry = GET_ENTRY((unsigned long)cookie);
entry             513 drivers/misc/mic/host/mic_intr.c 		if (entry >= mdev->irq_info.num_vectors) {
entry             516 drivers/misc/mic/host/mic_intr.c 				entry, mdev->irq_info.num_vectors);
entry             519 drivers/misc/mic/host/mic_intr.c 		irq = mdev->irq_info.msix_entries[entry].vector;
entry             521 drivers/misc/mic/host/mic_intr.c 		mdev->irq_info.mic_msi_map[entry] &= ~(BIT(offset));
entry             523 drivers/misc/mic/host/mic_intr.c 			entry, offset, false);
entry             534 drivers/misc/mic/host/mic_intr.c 			mdev->irq_info.mic_msi_map[entry] &= ~(BIT(src_id));
entry             536 drivers/misc/mic/host/mic_intr.c 				entry, src_id, false);
entry             620 drivers/misc/mic/host/mic_intr.c 	int entry, offset;
entry             626 drivers/misc/mic/host/mic_intr.c 	for (entry = 0; entry < mdev->irq_info.num_vectors; entry++) {
entry             628 drivers/misc/mic/host/mic_intr.c 			if (mdev->irq_info.mic_msi_map[entry] & BIT(offset))
entry             630 drivers/misc/mic/host/mic_intr.c 					entry, offset, true);
entry              74 drivers/misc/mic/host/mic_smpt.c 		if (!smpt_info->entry[i].ref_count &&
entry              75 drivers/misc/mic/host/mic_smpt.c 		    (smpt_info->entry[i].dma_addr != addr)) {
entry              77 drivers/misc/mic/host/mic_smpt.c 			smpt_info->entry[i].dma_addr = addr;
entry              79 drivers/misc/mic/host/mic_smpt.c 		smpt_info->entry[i].ref_count += ref[i - spt];
entry             102 drivers/misc/mic/host/mic_smpt.c 		if (smpt_info->entry[i].dma_addr == addr) {
entry             114 drivers/misc/mic/host/mic_smpt.c 		ae = (smpt_info->entry[i].ref_count == 0) ? ae + 1 : 0;
entry             177 drivers/misc/mic/host/mic_smpt.c 	dma_addr = smpt_info->entry[spt].dma_addr +
entry             273 drivers/misc/mic/host/mic_smpt.c 		smpt_info->entry[i].ref_count -= ref[i - spt];
entry             274 drivers/misc/mic/host/mic_smpt.c 		if (smpt_info->entry[i].ref_count < 0)
entry             354 drivers/misc/mic/host/mic_smpt.c 	smpt_info->entry = kmalloc_array(smpt_info->info.num_reg,
entry             355 drivers/misc/mic/host/mic_smpt.c 					 sizeof(*smpt_info->entry), GFP_KERNEL);
entry             356 drivers/misc/mic/host/mic_smpt.c 	if (!smpt_info->entry) {
entry             363 drivers/misc/mic/host/mic_smpt.c 		smpt_info->entry[i].dma_addr = dma_addr;
entry             364 drivers/misc/mic/host/mic_smpt.c 		smpt_info->entry[i].ref_count = 0;
entry             396 drivers/misc/mic/host/mic_smpt.c 			i, smpt_info->entry[i].dma_addr,
entry             397 drivers/misc/mic/host/mic_smpt.c 			smpt_info->entry[i].ref_count);
entry             398 drivers/misc/mic/host/mic_smpt.c 		if (smpt_info->entry[i].ref_count)
entry             402 drivers/misc/mic/host/mic_smpt.c 	kfree(smpt_info->entry);
entry             424 drivers/misc/mic/host/mic_smpt.c 		dma_addr = mdev->smpt->entry[i].dma_addr;
entry              55 drivers/misc/mic/host/mic_smpt.h 	struct mic_smpt *entry;
entry              64 drivers/misc/sgi-gru/grufile.c 	struct list_head *entry, *next;
entry              73 drivers/misc/sgi-gru/grufile.c 	list_for_each_safe(entry, next, &vdata->vd_head) {
entry              75 drivers/misc/sgi-gru/grufile.c 		    list_entry(entry, struct gru_thread_state, ts_next);
entry            1092 drivers/misc/sgi-xp/xpc_uv.c 	int entry;
entry            1101 drivers/misc/sgi-xp/xpc_uv.c 		for (entry = 0; entry < nentries; entry++) {
entry            1102 drivers/misc/sgi-xp/xpc_uv.c 			msg_slot = &ch_uv->send_msg_slots[entry];
entry            1104 drivers/misc/sgi-xp/xpc_uv.c 			msg_slot->msg_slot_number = entry;
entry            1126 drivers/misc/sgi-xp/xpc_uv.c 	int entry;
entry            1135 drivers/misc/sgi-xp/xpc_uv.c 		for (entry = 0; entry < nentries; entry++) {
entry            1137 drivers/misc/sgi-xp/xpc_uv.c 			    entry * ch->entry_size;
entry            1139 drivers/misc/sgi-xp/xpc_uv.c 			msg_slot->hdr.msg_slot_number = entry;
entry            1337 drivers/misc/sgi-xp/xpc_uv.c 	struct xpc_fifo_entry_uv *entry;
entry            1340 drivers/misc/sgi-xp/xpc_uv.c 		entry = xpc_get_fifo_entry_uv(&ch->sn.uv.msg_slot_free_list);
entry            1341 drivers/misc/sgi-xp/xpc_uv.c 		if (entry != NULL)
entry            1352 drivers/misc/sgi-xp/xpc_uv.c 	msg_slot = container_of(entry, struct xpc_send_msg_slot_uv, next);
entry            1396 drivers/misc/sgi-xp/xpc_uv.c 	int entry = msg->hdr.msg_slot_number % ch->local_nentries;
entry            1398 drivers/misc/sgi-xp/xpc_uv.c 	msg_slot = &ch->sn.uv.send_msg_slots[entry];
entry            1628 drivers/misc/sgi-xp/xpc_uv.c 	int entry;
entry            1632 drivers/misc/sgi-xp/xpc_uv.c 	for (entry = 0; entry < ch->local_nentries; entry++) {
entry            1637 drivers/misc/sgi-xp/xpc_uv.c 		msg_slot = &ch->sn.uv.send_msg_slots[entry];
entry            1649 drivers/misc/sgi-xp/xpc_uv.c 	struct xpc_fifo_entry_uv *entry;
entry            1654 drivers/misc/sgi-xp/xpc_uv.c 		entry = xpc_get_fifo_entry_uv(&ch->sn.uv.recv_msg_list);
entry            1655 drivers/misc/sgi-xp/xpc_uv.c 		if (entry != NULL) {
entry            1656 drivers/misc/sgi-xp/xpc_uv.c 			msg = container_of(entry, struct xpc_notify_mq_msg_uv,
entry             705 drivers/misc/vmw_vmci/vmci_context.c 	struct vmci_handle_list *entry;
entry             724 drivers/misc/vmw_vmci/vmci_context.c 	list_for_each_entry(entry, &context->notifier_list, node)
entry             725 drivers/misc/vmw_vmci/vmci_context.c 		notifiers[i++] = entry->handle.context;
entry              36 drivers/misc/vmw_vmci/vmci_datagram.c 	struct datagram_entry *entry;
entry              59 drivers/misc/vmw_vmci/vmci_datagram.c 	struct datagram_entry *entry;
entry              74 drivers/misc/vmw_vmci/vmci_datagram.c 	entry = kmalloc(sizeof(*entry), GFP_KERNEL);
entry              75 drivers/misc/vmw_vmci/vmci_datagram.c 	if (!entry) {
entry              80 drivers/misc/vmw_vmci/vmci_datagram.c 	entry->run_delayed = (flags & VMCI_FLAG_DG_DELAYED_CB) ? true : false;
entry              81 drivers/misc/vmw_vmci/vmci_datagram.c 	entry->flags = flags;
entry              82 drivers/misc/vmw_vmci/vmci_datagram.c 	entry->recv_cb = recv_cb;
entry              83 drivers/misc/vmw_vmci/vmci_datagram.c 	entry->client_data = client_data;
entry              84 drivers/misc/vmw_vmci/vmci_datagram.c 	entry->priv_flags = priv_flags;
entry              87 drivers/misc/vmw_vmci/vmci_datagram.c 	result = vmci_resource_add(&entry->resource,
entry              93 drivers/misc/vmw_vmci/vmci_datagram.c 		kfree(entry);
entry              97 drivers/misc/vmw_vmci/vmci_datagram.c 	*out_handle = vmci_resource_handle(&entry->resource);
entry             141 drivers/misc/vmw_vmci/vmci_datagram.c 	dg_info->entry->recv_cb(dg_info->entry->client_data, &dg_info->msg);
entry             143 drivers/misc/vmw_vmci/vmci_datagram.c 	vmci_resource_put(&dg_info->entry->resource);
entry             236 drivers/misc/vmw_vmci/vmci_datagram.c 			dg_info->entry = dst_entry;
entry             379 drivers/misc/vmw_vmci/vmci_datagram.c 		dg_info->entry = dst_entry;
entry             461 drivers/misc/vmw_vmci/vmci_datagram.c 	struct datagram_entry *entry;
entry             471 drivers/misc/vmw_vmci/vmci_datagram.c 	entry = container_of(resource, struct datagram_entry, resource);
entry             473 drivers/misc/vmw_vmci/vmci_datagram.c 	vmci_resource_put(&entry->resource);
entry             474 drivers/misc/vmw_vmci/vmci_datagram.c 	vmci_resource_remove(&entry->resource);
entry             475 drivers/misc/vmw_vmci/vmci_datagram.c 	kfree(entry);
entry              91 drivers/misc/vmw_vmci/vmci_doorbell.c 		struct dbell_entry *entry;
entry              99 drivers/misc/vmw_vmci/vmci_doorbell.c 		entry = container_of(resource, struct dbell_entry, resource);
entry             100 drivers/misc/vmw_vmci/vmci_doorbell.c 		*priv_flags = entry->priv_flags;
entry             137 drivers/misc/vmw_vmci/vmci_doorbell.c static void dbell_index_table_add(struct dbell_entry *entry)
entry             142 drivers/misc/vmw_vmci/vmci_doorbell.c 	vmci_resource_get(&entry->resource);
entry             186 drivers/misc/vmw_vmci/vmci_doorbell.c 	entry->idx = new_notify_idx;
entry             187 drivers/misc/vmw_vmci/vmci_doorbell.c 	bucket = VMCI_DOORBELL_HASH(entry->idx);
entry             188 drivers/misc/vmw_vmci/vmci_doorbell.c 	hlist_add_head(&entry->node, &vmci_doorbell_it.entries[bucket]);
entry             197 drivers/misc/vmw_vmci/vmci_doorbell.c static void dbell_index_table_remove(struct dbell_entry *entry)
entry             201 drivers/misc/vmw_vmci/vmci_doorbell.c 	hlist_del_init(&entry->node);
entry             204 drivers/misc/vmw_vmci/vmci_doorbell.c 	if (entry->idx == max_notify_idx - 1) {
entry             217 drivers/misc/vmw_vmci/vmci_doorbell.c 	last_notify_idx_released = entry->idx;
entry             221 drivers/misc/vmw_vmci/vmci_doorbell.c 	vmci_resource_put(&entry->resource);
entry             282 drivers/misc/vmw_vmci/vmci_doorbell.c 	struct dbell_entry *entry = container_of(work,
entry             285 drivers/misc/vmw_vmci/vmci_doorbell.c 	entry->notify_cb(entry->client_data);
entry             286 drivers/misc/vmw_vmci/vmci_doorbell.c 	vmci_resource_put(&entry->resource);
entry             294 drivers/misc/vmw_vmci/vmci_doorbell.c 	struct dbell_entry *entry;
entry             311 drivers/misc/vmw_vmci/vmci_doorbell.c 	entry = container_of(resource, struct dbell_entry, resource);
entry             312 drivers/misc/vmw_vmci/vmci_doorbell.c 	if (entry->run_delayed) {
entry             313 drivers/misc/vmw_vmci/vmci_doorbell.c 		if (!schedule_work(&entry->work))
entry             316 drivers/misc/vmw_vmci/vmci_doorbell.c 		entry->notify_cb(entry->client_data);
entry             414 drivers/misc/vmw_vmci/vmci_doorbell.c 	struct dbell_entry *entry;
entry             422 drivers/misc/vmw_vmci/vmci_doorbell.c 	entry = kmalloc(sizeof(*entry), GFP_KERNEL);
entry             423 drivers/misc/vmw_vmci/vmci_doorbell.c 	if (entry == NULL) {
entry             465 drivers/misc/vmw_vmci/vmci_doorbell.c 	entry->idx = 0;
entry             466 drivers/misc/vmw_vmci/vmci_doorbell.c 	INIT_HLIST_NODE(&entry->node);
entry             467 drivers/misc/vmw_vmci/vmci_doorbell.c 	entry->priv_flags = priv_flags;
entry             468 drivers/misc/vmw_vmci/vmci_doorbell.c 	INIT_WORK(&entry->work, dbell_delayed_dispatch);
entry             469 drivers/misc/vmw_vmci/vmci_doorbell.c 	entry->run_delayed = flags & VMCI_FLAG_DELAYED_CB;
entry             470 drivers/misc/vmw_vmci/vmci_doorbell.c 	entry->notify_cb = notify_cb;
entry             471 drivers/misc/vmw_vmci/vmci_doorbell.c 	entry->client_data = client_data;
entry             472 drivers/misc/vmw_vmci/vmci_doorbell.c 	atomic_set(&entry->active, 0);
entry             474 drivers/misc/vmw_vmci/vmci_doorbell.c 	result = vmci_resource_add(&entry->resource,
entry             483 drivers/misc/vmw_vmci/vmci_doorbell.c 	new_handle = vmci_resource_handle(&entry->resource);
entry             485 drivers/misc/vmw_vmci/vmci_doorbell.c 		dbell_index_table_add(entry);
entry             486 drivers/misc/vmw_vmci/vmci_doorbell.c 		result = dbell_link(new_handle, entry->idx);
entry             490 drivers/misc/vmw_vmci/vmci_doorbell.c 		atomic_set(&entry->active, 1);
entry             498 drivers/misc/vmw_vmci/vmci_doorbell.c 	dbell_index_table_remove(entry);
entry             499 drivers/misc/vmw_vmci/vmci_doorbell.c 	vmci_resource_remove(&entry->resource);
entry             501 drivers/misc/vmw_vmci/vmci_doorbell.c 	kfree(entry);
entry             515 drivers/misc/vmw_vmci/vmci_doorbell.c 	struct dbell_entry *entry;
entry             529 drivers/misc/vmw_vmci/vmci_doorbell.c 	entry = container_of(resource, struct dbell_entry, resource);
entry             531 drivers/misc/vmw_vmci/vmci_doorbell.c 	if (!hlist_unhashed(&entry->node)) {
entry             534 drivers/misc/vmw_vmci/vmci_doorbell.c 		dbell_index_table_remove(entry);
entry             561 drivers/misc/vmw_vmci/vmci_doorbell.c 	vmci_resource_put(&entry->resource);
entry             562 drivers/misc/vmw_vmci/vmci_doorbell.c 	vmci_resource_remove(&entry->resource);
entry             564 drivers/misc/vmw_vmci/vmci_doorbell.c 	kfree(entry);
entry             800 drivers/misc/vmw_vmci/vmci_queue_pair.c 	struct qp_entry *entry;
entry             805 drivers/misc/vmw_vmci/vmci_queue_pair.c 	list_for_each_entry(entry, &qp_list->head, list_item) {
entry             806 drivers/misc/vmw_vmci/vmci_queue_pair.c 		if (vmci_handle_is_equal(entry->handle, handle))
entry             807 drivers/misc/vmw_vmci/vmci_queue_pair.c 			return entry;
entry             819 drivers/misc/vmw_vmci/vmci_queue_pair.c 	struct qp_guest_endpoint *entry;
entry             822 drivers/misc/vmw_vmci/vmci_queue_pair.c 	entry = qp ? container_of(
entry             824 drivers/misc/vmw_vmci/vmci_queue_pair.c 	return entry;
entry             833 drivers/misc/vmw_vmci/vmci_queue_pair.c 	struct qp_broker_entry *entry;
entry             836 drivers/misc/vmw_vmci/vmci_queue_pair.c 	entry = qp ? container_of(
entry             838 drivers/misc/vmw_vmci/vmci_queue_pair.c 	return entry;
entry             879 drivers/misc/vmw_vmci/vmci_queue_pair.c 	struct qp_guest_endpoint *entry;
entry             890 drivers/misc/vmw_vmci/vmci_queue_pair.c 	entry = kzalloc(sizeof(*entry), GFP_KERNEL);
entry             891 drivers/misc/vmw_vmci/vmci_queue_pair.c 	if (entry) {
entry             892 drivers/misc/vmw_vmci/vmci_queue_pair.c 		entry->qp.peer = peer;
entry             893 drivers/misc/vmw_vmci/vmci_queue_pair.c 		entry->qp.flags = flags;
entry             894 drivers/misc/vmw_vmci/vmci_queue_pair.c 		entry->qp.produce_size = produce_size;
entry             895 drivers/misc/vmw_vmci/vmci_queue_pair.c 		entry->qp.consume_size = consume_size;
entry             896 drivers/misc/vmw_vmci/vmci_queue_pair.c 		entry->qp.ref_count = 0;
entry             897 drivers/misc/vmw_vmci/vmci_queue_pair.c 		entry->num_ppns = num_ppns;
entry             898 drivers/misc/vmw_vmci/vmci_queue_pair.c 		entry->produce_q = produce_q;
entry             899 drivers/misc/vmw_vmci/vmci_queue_pair.c 		entry->consume_q = consume_q;
entry             900 drivers/misc/vmw_vmci/vmci_queue_pair.c 		INIT_LIST_HEAD(&entry->qp.list_item);
entry             903 drivers/misc/vmw_vmci/vmci_queue_pair.c 		result = vmci_resource_add(&entry->resource,
entry             906 drivers/misc/vmw_vmci/vmci_queue_pair.c 		entry->qp.handle = vmci_resource_handle(&entry->resource);
entry             908 drivers/misc/vmw_vmci/vmci_queue_pair.c 		    qp_list_find(&qp_guest_endpoints, entry->qp.handle)) {
entry             911 drivers/misc/vmw_vmci/vmci_queue_pair.c 			kfree(entry);
entry             912 drivers/misc/vmw_vmci/vmci_queue_pair.c 			entry = NULL;
entry             915 drivers/misc/vmw_vmci/vmci_queue_pair.c 	return entry;
entry             921 drivers/misc/vmw_vmci/vmci_queue_pair.c static void qp_guest_endpoint_destroy(struct qp_guest_endpoint *entry)
entry             923 drivers/misc/vmw_vmci/vmci_queue_pair.c 	qp_free_ppn_set(&entry->ppn_set);
entry             924 drivers/misc/vmw_vmci/vmci_queue_pair.c 	qp_cleanup_queue_mutex(entry->produce_q, entry->consume_q);
entry             925 drivers/misc/vmw_vmci/vmci_queue_pair.c 	qp_free_queue(entry->produce_q, entry->qp.produce_size);
entry             926 drivers/misc/vmw_vmci/vmci_queue_pair.c 	qp_free_queue(entry->consume_q, entry->qp.consume_size);
entry             928 drivers/misc/vmw_vmci/vmci_queue_pair.c 	vmci_resource_remove(&entry->resource);
entry             930 drivers/misc/vmw_vmci/vmci_queue_pair.c 	kfree(entry);
entry             937 drivers/misc/vmw_vmci/vmci_queue_pair.c static int qp_alloc_hypercall(const struct qp_guest_endpoint *entry)
entry             944 drivers/misc/vmw_vmci/vmci_queue_pair.c 	if (!entry || entry->num_ppns <= 2)
entry             949 drivers/misc/vmw_vmci/vmci_queue_pair.c 	    (size_t) entry->num_ppns * ppn_size;
entry             958 drivers/misc/vmw_vmci/vmci_queue_pair.c 	alloc_msg->handle = entry->qp.handle;
entry             959 drivers/misc/vmw_vmci/vmci_queue_pair.c 	alloc_msg->peer = entry->qp.peer;
entry             960 drivers/misc/vmw_vmci/vmci_queue_pair.c 	alloc_msg->flags = entry->qp.flags;
entry             961 drivers/misc/vmw_vmci/vmci_queue_pair.c 	alloc_msg->produce_size = entry->qp.produce_size;
entry             962 drivers/misc/vmw_vmci/vmci_queue_pair.c 	alloc_msg->consume_size = entry->qp.consume_size;
entry             963 drivers/misc/vmw_vmci/vmci_queue_pair.c 	alloc_msg->num_ppns = entry->num_ppns;
entry             966 drivers/misc/vmw_vmci/vmci_queue_pair.c 				     &entry->ppn_set);
entry             995 drivers/misc/vmw_vmci/vmci_queue_pair.c static void qp_list_add_entry(struct qp_list *qp_list, struct qp_entry *entry)
entry             997 drivers/misc/vmw_vmci/vmci_queue_pair.c 	if (entry)
entry             998 drivers/misc/vmw_vmci/vmci_queue_pair.c 		list_add(&entry->list_item, &qp_list->head);
entry            1005 drivers/misc/vmw_vmci/vmci_queue_pair.c 				 struct qp_entry *entry)
entry            1007 drivers/misc/vmw_vmci/vmci_queue_pair.c 	if (entry)
entry            1008 drivers/misc/vmw_vmci/vmci_queue_pair.c 		list_del(&entry->list_item);
entry            1018 drivers/misc/vmw_vmci/vmci_queue_pair.c 	struct qp_guest_endpoint *entry;
entry            1023 drivers/misc/vmw_vmci/vmci_queue_pair.c 	entry = qp_guest_handle_to_entry(handle);
entry            1024 drivers/misc/vmw_vmci/vmci_queue_pair.c 	if (!entry) {
entry            1029 drivers/misc/vmw_vmci/vmci_queue_pair.c 	if (entry->qp.flags & VMCI_QPFLAG_LOCAL) {
entry            1032 drivers/misc/vmw_vmci/vmci_queue_pair.c 		if (entry->qp.ref_count > 1) {
entry            1064 drivers/misc/vmw_vmci/vmci_queue_pair.c 	entry->qp.ref_count--;
entry            1065 drivers/misc/vmw_vmci/vmci_queue_pair.c 	if (entry->qp.ref_count == 0)
entry            1066 drivers/misc/vmw_vmci/vmci_queue_pair.c 		qp_list_remove_entry(&qp_guest_endpoints, &entry->qp);
entry            1069 drivers/misc/vmw_vmci/vmci_queue_pair.c 	if (entry)
entry            1070 drivers/misc/vmw_vmci/vmci_queue_pair.c 		ref_count = entry->qp.ref_count;
entry            1075 drivers/misc/vmw_vmci/vmci_queue_pair.c 		qp_guest_endpoint_destroy(entry);
entry            1285 drivers/misc/vmw_vmci/vmci_queue_pair.c 	struct qp_broker_entry *entry = NULL;
entry            1313 drivers/misc/vmw_vmci/vmci_queue_pair.c 	entry = kzalloc(sizeof(*entry), GFP_ATOMIC);
entry            1314 drivers/misc/vmw_vmci/vmci_queue_pair.c 	if (!entry)
entry            1333 drivers/misc/vmw_vmci/vmci_queue_pair.c 	entry->qp.handle = handle;
entry            1334 drivers/misc/vmw_vmci/vmci_queue_pair.c 	entry->qp.peer = peer;
entry            1335 drivers/misc/vmw_vmci/vmci_queue_pair.c 	entry->qp.flags = flags;
entry            1336 drivers/misc/vmw_vmci/vmci_queue_pair.c 	entry->qp.produce_size = guest_produce_size;
entry            1337 drivers/misc/vmw_vmci/vmci_queue_pair.c 	entry->qp.consume_size = guest_consume_size;
entry            1338 drivers/misc/vmw_vmci/vmci_queue_pair.c 	entry->qp.ref_count = 1;
entry            1339 drivers/misc/vmw_vmci/vmci_queue_pair.c 	entry->create_id = context_id;
entry            1340 drivers/misc/vmw_vmci/vmci_queue_pair.c 	entry->attach_id = VMCI_INVALID_ID;
entry            1341 drivers/misc/vmw_vmci/vmci_queue_pair.c 	entry->state = VMCIQPB_NEW;
entry            1342 drivers/misc/vmw_vmci/vmci_queue_pair.c 	entry->require_trusted_attach =
entry            1344 drivers/misc/vmw_vmci/vmci_queue_pair.c 	entry->created_by_trusted =
entry            1346 drivers/misc/vmw_vmci/vmci_queue_pair.c 	entry->vmci_page_files = false;
entry            1347 drivers/misc/vmw_vmci/vmci_queue_pair.c 	entry->wakeup_cb = wakeup_cb;
entry            1348 drivers/misc/vmw_vmci/vmci_queue_pair.c 	entry->client_data = client_data;
entry            1349 drivers/misc/vmw_vmci/vmci_queue_pair.c 	entry->produce_q = qp_host_alloc_queue(guest_produce_size);
entry            1350 drivers/misc/vmw_vmci/vmci_queue_pair.c 	if (entry->produce_q == NULL) {
entry            1354 drivers/misc/vmw_vmci/vmci_queue_pair.c 	entry->consume_q = qp_host_alloc_queue(guest_consume_size);
entry            1355 drivers/misc/vmw_vmci/vmci_queue_pair.c 	if (entry->consume_q == NULL) {
entry            1360 drivers/misc/vmw_vmci/vmci_queue_pair.c 	qp_init_queue_mutex(entry->produce_q, entry->consume_q);
entry            1362 drivers/misc/vmw_vmci/vmci_queue_pair.c 	INIT_LIST_HEAD(&entry->qp.list_item);
entry            1367 drivers/misc/vmw_vmci/vmci_queue_pair.c 		entry->local_mem = kcalloc(QPE_NUM_PAGES(entry->qp),
entry            1369 drivers/misc/vmw_vmci/vmci_queue_pair.c 		if (entry->local_mem == NULL) {
entry            1373 drivers/misc/vmw_vmci/vmci_queue_pair.c 		entry->state = VMCIQPB_CREATED_MEM;
entry            1374 drivers/misc/vmw_vmci/vmci_queue_pair.c 		entry->produce_q->q_header = entry->local_mem;
entry            1375 drivers/misc/vmw_vmci/vmci_queue_pair.c 		tmp = (u8 *)entry->local_mem + PAGE_SIZE *
entry            1376 drivers/misc/vmw_vmci/vmci_queue_pair.c 		    (DIV_ROUND_UP(entry->qp.produce_size, PAGE_SIZE) + 1);
entry            1377 drivers/misc/vmw_vmci/vmci_queue_pair.c 		entry->consume_q->q_header = (struct vmci_queue_header *)tmp;
entry            1384 drivers/misc/vmw_vmci/vmci_queue_pair.c 						      entry->produce_q,
entry            1385 drivers/misc/vmw_vmci/vmci_queue_pair.c 						      entry->consume_q);
entry            1389 drivers/misc/vmw_vmci/vmci_queue_pair.c 		entry->state = VMCIQPB_CREATED_MEM;
entry            1398 drivers/misc/vmw_vmci/vmci_queue_pair.c 		entry->state = VMCIQPB_CREATED_NO_MEM;
entry            1401 drivers/misc/vmw_vmci/vmci_queue_pair.c 	qp_list_add_entry(&qp_broker_list, &entry->qp);
entry            1403 drivers/misc/vmw_vmci/vmci_queue_pair.c 		*ent = entry;
entry            1406 drivers/misc/vmw_vmci/vmci_queue_pair.c 	result = vmci_resource_add(&entry->resource,
entry            1415 drivers/misc/vmw_vmci/vmci_queue_pair.c 	entry->qp.handle = vmci_resource_handle(&entry->resource);
entry            1417 drivers/misc/vmw_vmci/vmci_queue_pair.c 		vmci_q_header_init(entry->produce_q->q_header,
entry            1418 drivers/misc/vmw_vmci/vmci_queue_pair.c 				   entry->qp.handle);
entry            1419 drivers/misc/vmw_vmci/vmci_queue_pair.c 		vmci_q_header_init(entry->consume_q->q_header,
entry            1420 drivers/misc/vmw_vmci/vmci_queue_pair.c 				   entry->qp.handle);
entry            1423 drivers/misc/vmw_vmci/vmci_queue_pair.c 	vmci_ctx_qp_create(context, entry->qp.handle);
entry            1428 drivers/misc/vmw_vmci/vmci_queue_pair.c 	if (entry != NULL) {
entry            1429 drivers/misc/vmw_vmci/vmci_queue_pair.c 		qp_host_free_queue(entry->produce_q, guest_produce_size);
entry            1430 drivers/misc/vmw_vmci/vmci_queue_pair.c 		qp_host_free_queue(entry->consume_q, guest_consume_size);
entry            1431 drivers/misc/vmw_vmci/vmci_queue_pair.c 		kfree(entry);
entry            1502 drivers/misc/vmw_vmci/vmci_queue_pair.c static int qp_broker_attach(struct qp_broker_entry *entry,
entry            1518 drivers/misc/vmw_vmci/vmci_queue_pair.c 	if (entry->state != VMCIQPB_CREATED_NO_MEM &&
entry            1519 drivers/misc/vmw_vmci/vmci_queue_pair.c 	    entry->state != VMCIQPB_CREATED_MEM)
entry            1523 drivers/misc/vmw_vmci/vmci_queue_pair.c 		if (!(entry->qp.flags & VMCI_QPFLAG_LOCAL) ||
entry            1524 drivers/misc/vmw_vmci/vmci_queue_pair.c 		    context_id != entry->create_id) {
entry            1527 drivers/misc/vmw_vmci/vmci_queue_pair.c 	} else if (context_id == entry->create_id ||
entry            1528 drivers/misc/vmw_vmci/vmci_queue_pair.c 		   context_id == entry->attach_id) {
entry            1533 drivers/misc/vmw_vmci/vmci_queue_pair.c 	    VMCI_CONTEXT_IS_VM(entry->create_id))
entry            1541 drivers/misc/vmw_vmci/vmci_queue_pair.c 	    !entry->created_by_trusted)
entry            1548 drivers/misc/vmw_vmci/vmci_queue_pair.c 	if (entry->require_trusted_attach &&
entry            1556 drivers/misc/vmw_vmci/vmci_queue_pair.c 	if (entry->qp.peer != VMCI_INVALID_ID && entry->qp.peer != context_id)
entry            1559 drivers/misc/vmw_vmci/vmci_queue_pair.c 	if (entry->create_id == VMCI_HOST_CONTEXT_ID) {
entry            1577 drivers/misc/vmw_vmci/vmci_queue_pair.c 		create_context = vmci_ctx_get(entry->create_id);
entry            1585 drivers/misc/vmw_vmci/vmci_queue_pair.c 	if ((entry->qp.flags & ~VMCI_QP_ASYMM) != (flags & ~VMCI_QP_ASYMM_PEER))
entry            1595 drivers/misc/vmw_vmci/vmci_queue_pair.c 		if (entry->qp.produce_size != produce_size ||
entry            1596 drivers/misc/vmw_vmci/vmci_queue_pair.c 		    entry->qp.consume_size != consume_size) {
entry            1599 drivers/misc/vmw_vmci/vmci_queue_pair.c 	} else if (entry->qp.produce_size != consume_size ||
entry            1600 drivers/misc/vmw_vmci/vmci_queue_pair.c 		   entry->qp.consume_size != produce_size) {
entry            1618 drivers/misc/vmw_vmci/vmci_queue_pair.c 		if (entry->state != VMCIQPB_CREATED_NO_MEM)
entry            1630 drivers/misc/vmw_vmci/vmci_queue_pair.c 							      entry->produce_q,
entry            1631 drivers/misc/vmw_vmci/vmci_queue_pair.c 							      entry->consume_q);
entry            1635 drivers/misc/vmw_vmci/vmci_queue_pair.c 			entry->state = VMCIQPB_ATTACHED_MEM;
entry            1637 drivers/misc/vmw_vmci/vmci_queue_pair.c 			entry->state = VMCIQPB_ATTACHED_NO_MEM;
entry            1639 drivers/misc/vmw_vmci/vmci_queue_pair.c 	} else if (entry->state == VMCIQPB_CREATED_NO_MEM) {
entry            1650 drivers/misc/vmw_vmci/vmci_queue_pair.c 		entry->state = VMCIQPB_ATTACHED_MEM;
entry            1653 drivers/misc/vmw_vmci/vmci_queue_pair.c 	if (entry->state == VMCIQPB_ATTACHED_MEM) {
entry            1655 drivers/misc/vmw_vmci/vmci_queue_pair.c 		    qp_notify_peer(true, entry->qp.handle, context_id,
entry            1656 drivers/misc/vmw_vmci/vmci_queue_pair.c 				   entry->create_id);
entry            1659 drivers/misc/vmw_vmci/vmci_queue_pair.c 				entry->create_id, entry->qp.handle.context,
entry            1660 drivers/misc/vmw_vmci/vmci_queue_pair.c 				entry->qp.handle.resource);
entry            1663 drivers/misc/vmw_vmci/vmci_queue_pair.c 	entry->attach_id = context_id;
entry            1664 drivers/misc/vmw_vmci/vmci_queue_pair.c 	entry->qp.ref_count++;
entry            1666 drivers/misc/vmw_vmci/vmci_queue_pair.c 		entry->wakeup_cb = wakeup_cb;
entry            1667 drivers/misc/vmw_vmci/vmci_queue_pair.c 		entry->client_data = client_data;
entry            1675 drivers/misc/vmw_vmci/vmci_queue_pair.c 		vmci_ctx_qp_create(context, entry->qp.handle);
entry            1678 drivers/misc/vmw_vmci/vmci_queue_pair.c 		*ent = entry;
entry            1702 drivers/misc/vmw_vmci/vmci_queue_pair.c 	struct qp_broker_entry *entry = NULL;
entry            1732 drivers/misc/vmw_vmci/vmci_queue_pair.c 		entry = qp_broker_handle_to_entry(handle);
entry            1734 drivers/misc/vmw_vmci/vmci_queue_pair.c 	if (!entry) {
entry            1743 drivers/misc/vmw_vmci/vmci_queue_pair.c 		    qp_broker_attach(entry, peer, flags, priv_flags,
entry            1774 drivers/misc/vmw_vmci/vmci_queue_pair.c 	struct qp_broker_entry *entry;
entry            1785 drivers/misc/vmw_vmci/vmci_queue_pair.c 	entry = NULL;
entry            1789 drivers/misc/vmw_vmci/vmci_queue_pair.c 			    wakeup_cb, client_data, &entry, &swap);
entry            1798 drivers/misc/vmw_vmci/vmci_queue_pair.c 			*produce_q = entry->consume_q;
entry            1799 drivers/misc/vmw_vmci/vmci_queue_pair.c 			*consume_q = entry->produce_q;
entry            1801 drivers/misc/vmw_vmci/vmci_queue_pair.c 			*produce_q = entry->produce_q;
entry            1802 drivers/misc/vmw_vmci/vmci_queue_pair.c 			*consume_q = entry->consume_q;
entry            1805 drivers/misc/vmw_vmci/vmci_queue_pair.c 		*handle = vmci_resource_handle(&entry->resource);
entry            1888 drivers/misc/vmw_vmci/vmci_queue_pair.c 		struct qp_entry *entry =
entry            1891 drivers/misc/vmw_vmci/vmci_queue_pair.c 		return entry;
entry            1899 drivers/misc/vmw_vmci/vmci_queue_pair.c 	struct qp_entry *entry;
entry            1904 drivers/misc/vmw_vmci/vmci_queue_pair.c 	while ((entry = qp_list_get_head(&qp_broker_list))) {
entry            1905 drivers/misc/vmw_vmci/vmci_queue_pair.c 		be = (struct qp_broker_entry *)entry;
entry            1907 drivers/misc/vmw_vmci/vmci_queue_pair.c 		qp_list_remove_entry(&qp_broker_list, entry);
entry            1956 drivers/misc/vmw_vmci/vmci_queue_pair.c 	struct qp_broker_entry *entry;
entry            1981 drivers/misc/vmw_vmci/vmci_queue_pair.c 	entry = qp_broker_handle_to_entry(handle);
entry            1982 drivers/misc/vmw_vmci/vmci_queue_pair.c 	if (!entry) {
entry            1993 drivers/misc/vmw_vmci/vmci_queue_pair.c 	if (entry->create_id != context_id &&
entry            1994 drivers/misc/vmw_vmci/vmci_queue_pair.c 	    (entry->create_id != VMCI_HOST_CONTEXT_ID ||
entry            1995 drivers/misc/vmw_vmci/vmci_queue_pair.c 	     entry->attach_id != context_id)) {
entry            2000 drivers/misc/vmw_vmci/vmci_queue_pair.c 	if (entry->state != VMCIQPB_CREATED_NO_MEM &&
entry            2001 drivers/misc/vmw_vmci/vmci_queue_pair.c 	    entry->state != VMCIQPB_ATTACHED_NO_MEM) {
entry            2007 drivers/misc/vmw_vmci/vmci_queue_pair.c 					 entry->produce_q, entry->consume_q);
entry            2011 drivers/misc/vmw_vmci/vmci_queue_pair.c 	result = qp_host_map_queues(entry->produce_q, entry->consume_q);
entry            2013 drivers/misc/vmw_vmci/vmci_queue_pair.c 		qp_host_unregister_user_memory(entry->produce_q,
entry            2014 drivers/misc/vmw_vmci/vmci_queue_pair.c 					       entry->consume_q);
entry            2018 drivers/misc/vmw_vmci/vmci_queue_pair.c 	if (entry->state == VMCIQPB_CREATED_NO_MEM)
entry            2019 drivers/misc/vmw_vmci/vmci_queue_pair.c 		entry->state = VMCIQPB_CREATED_MEM;
entry            2021 drivers/misc/vmw_vmci/vmci_queue_pair.c 		entry->state = VMCIQPB_ATTACHED_MEM;
entry            2023 drivers/misc/vmw_vmci/vmci_queue_pair.c 	entry->vmci_page_files = true;
entry            2025 drivers/misc/vmw_vmci/vmci_queue_pair.c 	if (entry->state == VMCIQPB_ATTACHED_MEM) {
entry            2027 drivers/misc/vmw_vmci/vmci_queue_pair.c 		    qp_notify_peer(true, handle, context_id, entry->create_id);
entry            2030 drivers/misc/vmw_vmci/vmci_queue_pair.c 				entry->create_id, entry->qp.handle.context,
entry            2031 drivers/misc/vmw_vmci/vmci_queue_pair.c 				entry->qp.handle.resource);
entry            2046 drivers/misc/vmw_vmci/vmci_queue_pair.c static void qp_reset_saved_headers(struct qp_broker_entry *entry)
entry            2048 drivers/misc/vmw_vmci/vmci_queue_pair.c 	entry->produce_q->saved_header = NULL;
entry            2049 drivers/misc/vmw_vmci/vmci_queue_pair.c 	entry->consume_q->saved_header = NULL;
entry            2072 drivers/misc/vmw_vmci/vmci_queue_pair.c 	struct qp_broker_entry *entry;
entry            2092 drivers/misc/vmw_vmci/vmci_queue_pair.c 	entry = qp_broker_handle_to_entry(handle);
entry            2093 drivers/misc/vmw_vmci/vmci_queue_pair.c 	if (!entry) {
entry            2100 drivers/misc/vmw_vmci/vmci_queue_pair.c 	if (context_id != entry->create_id && context_id != entry->attach_id) {
entry            2105 drivers/misc/vmw_vmci/vmci_queue_pair.c 	if (context_id == entry->create_id) {
entry            2106 drivers/misc/vmw_vmci/vmci_queue_pair.c 		peer_id = entry->attach_id;
entry            2107 drivers/misc/vmw_vmci/vmci_queue_pair.c 		entry->create_id = VMCI_INVALID_ID;
entry            2109 drivers/misc/vmw_vmci/vmci_queue_pair.c 		peer_id = entry->create_id;
entry            2110 drivers/misc/vmw_vmci/vmci_queue_pair.c 		entry->attach_id = VMCI_INVALID_ID;
entry            2112 drivers/misc/vmw_vmci/vmci_queue_pair.c 	entry->qp.ref_count--;
entry            2114 drivers/misc/vmw_vmci/vmci_queue_pair.c 	is_local = entry->qp.flags & VMCI_QPFLAG_LOCAL;
entry            2127 drivers/misc/vmw_vmci/vmci_queue_pair.c 		qp_acquire_queue_mutex(entry->produce_q);
entry            2128 drivers/misc/vmw_vmci/vmci_queue_pair.c 		headers_mapped = entry->produce_q->q_header ||
entry            2129 drivers/misc/vmw_vmci/vmci_queue_pair.c 		    entry->consume_q->q_header;
entry            2130 drivers/misc/vmw_vmci/vmci_queue_pair.c 		if (QPBROKERSTATE_HAS_MEM(entry)) {
entry            2133 drivers/misc/vmw_vmci/vmci_queue_pair.c 						 entry->produce_q,
entry            2134 drivers/misc/vmw_vmci/vmci_queue_pair.c 						 entry->consume_q);
entry            2140 drivers/misc/vmw_vmci/vmci_queue_pair.c 			qp_host_unregister_user_memory(entry->produce_q,
entry            2141 drivers/misc/vmw_vmci/vmci_queue_pair.c 						       entry->consume_q);
entry            2146 drivers/misc/vmw_vmci/vmci_queue_pair.c 			qp_reset_saved_headers(entry);
entry            2148 drivers/misc/vmw_vmci/vmci_queue_pair.c 		qp_release_queue_mutex(entry->produce_q);
entry            2150 drivers/misc/vmw_vmci/vmci_queue_pair.c 		if (!headers_mapped && entry->wakeup_cb)
entry            2151 drivers/misc/vmw_vmci/vmci_queue_pair.c 			entry->wakeup_cb(entry->client_data);
entry            2154 drivers/misc/vmw_vmci/vmci_queue_pair.c 		if (entry->wakeup_cb) {
entry            2155 drivers/misc/vmw_vmci/vmci_queue_pair.c 			entry->wakeup_cb = NULL;
entry            2156 drivers/misc/vmw_vmci/vmci_queue_pair.c 			entry->client_data = NULL;
entry            2160 drivers/misc/vmw_vmci/vmci_queue_pair.c 	if (entry->qp.ref_count == 0) {
entry            2161 drivers/misc/vmw_vmci/vmci_queue_pair.c 		qp_list_remove_entry(&qp_broker_list, &entry->qp);
entry            2164 drivers/misc/vmw_vmci/vmci_queue_pair.c 			kfree(entry->local_mem);
entry            2166 drivers/misc/vmw_vmci/vmci_queue_pair.c 		qp_cleanup_queue_mutex(entry->produce_q, entry->consume_q);
entry            2167 drivers/misc/vmw_vmci/vmci_queue_pair.c 		qp_host_free_queue(entry->produce_q, entry->qp.produce_size);
entry            2168 drivers/misc/vmw_vmci/vmci_queue_pair.c 		qp_host_free_queue(entry->consume_q, entry->qp.consume_size);
entry            2170 drivers/misc/vmw_vmci/vmci_queue_pair.c 		vmci_resource_remove(&entry->resource);
entry            2172 drivers/misc/vmw_vmci/vmci_queue_pair.c 		kfree(entry);
entry            2178 drivers/misc/vmw_vmci/vmci_queue_pair.c 		    QPBROKERSTATE_HAS_MEM(entry)) {
entry            2179 drivers/misc/vmw_vmci/vmci_queue_pair.c 			entry->state = VMCIQPB_SHUTDOWN_MEM;
entry            2181 drivers/misc/vmw_vmci/vmci_queue_pair.c 			entry->state = VMCIQPB_SHUTDOWN_NO_MEM;
entry            2204 drivers/misc/vmw_vmci/vmci_queue_pair.c 	struct qp_broker_entry *entry;
entry            2221 drivers/misc/vmw_vmci/vmci_queue_pair.c 	entry = qp_broker_handle_to_entry(handle);
entry            2222 drivers/misc/vmw_vmci/vmci_queue_pair.c 	if (!entry) {
entry            2229 drivers/misc/vmw_vmci/vmci_queue_pair.c 	if (context_id != entry->create_id && context_id != entry->attach_id) {
entry            2240 drivers/misc/vmw_vmci/vmci_queue_pair.c 		page_store.len = QPE_NUM_PAGES(entry->qp);
entry            2242 drivers/misc/vmw_vmci/vmci_queue_pair.c 		qp_acquire_queue_mutex(entry->produce_q);
entry            2243 drivers/misc/vmw_vmci/vmci_queue_pair.c 		qp_reset_saved_headers(entry);
entry            2246 drivers/misc/vmw_vmci/vmci_queue_pair.c 						 entry->produce_q,
entry            2247 drivers/misc/vmw_vmci/vmci_queue_pair.c 						 entry->consume_q);
entry            2248 drivers/misc/vmw_vmci/vmci_queue_pair.c 		qp_release_queue_mutex(entry->produce_q);
entry            2252 drivers/misc/vmw_vmci/vmci_queue_pair.c 			entry->state++;
entry            2254 drivers/misc/vmw_vmci/vmci_queue_pair.c 			if (entry->wakeup_cb)
entry            2255 drivers/misc/vmw_vmci/vmci_queue_pair.c 				entry->wakeup_cb(entry->client_data);
entry            2271 drivers/misc/vmw_vmci/vmci_queue_pair.c static int qp_save_headers(struct qp_broker_entry *entry)
entry            2275 drivers/misc/vmw_vmci/vmci_queue_pair.c 	if (entry->produce_q->saved_header != NULL &&
entry            2276 drivers/misc/vmw_vmci/vmci_queue_pair.c 	    entry->consume_q->saved_header != NULL) {
entry            2286 drivers/misc/vmw_vmci/vmci_queue_pair.c 	if (NULL == entry->produce_q->q_header ||
entry            2287 drivers/misc/vmw_vmci/vmci_queue_pair.c 	    NULL == entry->consume_q->q_header) {
entry            2288 drivers/misc/vmw_vmci/vmci_queue_pair.c 		result = qp_host_map_queues(entry->produce_q, entry->consume_q);
entry            2293 drivers/misc/vmw_vmci/vmci_queue_pair.c 	memcpy(&entry->saved_produce_q, entry->produce_q->q_header,
entry            2294 drivers/misc/vmw_vmci/vmci_queue_pair.c 	       sizeof(entry->saved_produce_q));
entry            2295 drivers/misc/vmw_vmci/vmci_queue_pair.c 	entry->produce_q->saved_header = &entry->saved_produce_q;
entry            2296 drivers/misc/vmw_vmci/vmci_queue_pair.c 	memcpy(&entry->saved_consume_q, entry->consume_q->q_header,
entry            2297 drivers/misc/vmw_vmci/vmci_queue_pair.c 	       sizeof(entry->saved_consume_q));
entry            2298 drivers/misc/vmw_vmci/vmci_queue_pair.c 	entry->consume_q->saved_header = &entry->saved_consume_q;
entry            2313 drivers/misc/vmw_vmci/vmci_queue_pair.c 	struct qp_broker_entry *entry;
entry            2330 drivers/misc/vmw_vmci/vmci_queue_pair.c 	entry = qp_broker_handle_to_entry(handle);
entry            2331 drivers/misc/vmw_vmci/vmci_queue_pair.c 	if (!entry) {
entry            2338 drivers/misc/vmw_vmci/vmci_queue_pair.c 	if (context_id != entry->create_id && context_id != entry->attach_id) {
entry            2344 drivers/misc/vmw_vmci/vmci_queue_pair.c 		qp_acquire_queue_mutex(entry->produce_q);
entry            2345 drivers/misc/vmw_vmci/vmci_queue_pair.c 		result = qp_save_headers(entry);
entry            2350 drivers/misc/vmw_vmci/vmci_queue_pair.c 		qp_host_unmap_queues(gid, entry->produce_q, entry->consume_q);
entry            2359 drivers/misc/vmw_vmci/vmci_queue_pair.c 		qp_host_unregister_user_memory(entry->produce_q,
entry            2360 drivers/misc/vmw_vmci/vmci_queue_pair.c 					       entry->consume_q);
entry            2365 drivers/misc/vmw_vmci/vmci_queue_pair.c 		entry->state--;
entry            2367 drivers/misc/vmw_vmci/vmci_queue_pair.c 		qp_release_queue_mutex(entry->produce_q);
entry            2385 drivers/misc/vmw_vmci/vmci_queue_pair.c 	struct qp_entry *entry;
entry            2390 drivers/misc/vmw_vmci/vmci_queue_pair.c 	while ((entry = qp_list_get_head(&qp_guest_endpoints))) {
entry            2391 drivers/misc/vmw_vmci/vmci_queue_pair.c 		ep = (struct qp_guest_endpoint *)entry;
entry            2394 drivers/misc/vmw_vmci/vmci_queue_pair.c 		if (!(entry->flags & VMCI_QPFLAG_LOCAL))
entry            2395 drivers/misc/vmw_vmci/vmci_queue_pair.c 			qp_detatch_hypercall(entry->handle);
entry            2398 drivers/misc/vmw_vmci/vmci_queue_pair.c 		entry->ref_count = 0;
entry            2399 drivers/misc/vmw_vmci/vmci_queue_pair.c 		qp_list_remove_entry(&qp_guest_endpoints, entry);
entry              75 drivers/mtd/nand/bbt.c 				 unsigned int entry)
entry              79 drivers/mtd/nand/bbt.c 			     ((entry * bits_per_block) / BITS_PER_LONG);
entry              80 drivers/mtd/nand/bbt.c 	unsigned int offs = (entry * bits_per_block) % BITS_PER_LONG;
entry              83 drivers/mtd/nand/bbt.c 	if (entry >= nanddev_neraseblocks(nand))
entry             107 drivers/mtd/nand/bbt.c int nanddev_bbt_set_block_status(struct nand_device *nand, unsigned int entry,
entry             112 drivers/mtd/nand/bbt.c 			     ((entry * bits_per_block) / BITS_PER_LONG);
entry             113 drivers/mtd/nand/bbt.c 	unsigned int offs = (entry * bits_per_block) % BITS_PER_LONG;
entry             116 drivers/mtd/nand/bbt.c 	if (entry >= nanddev_neraseblocks(nand))
entry              25 drivers/mtd/nand/core.c 		unsigned int entry;
entry              28 drivers/mtd/nand/core.c 		entry = nanddev_bbt_pos_to_entry(nand, pos);
entry              29 drivers/mtd/nand/core.c 		status = nanddev_bbt_get_block_status(nand, entry);
entry              37 drivers/mtd/nand/core.c 			nanddev_bbt_set_block_status(nand, entry, status);
entry              64 drivers/mtd/nand/core.c 	unsigned int entry;
entry              78 drivers/mtd/nand/core.c 	entry = nanddev_bbt_pos_to_entry(nand, pos);
entry              79 drivers/mtd/nand/core.c 	ret = nanddev_bbt_set_block_status(nand, entry, NAND_BBT_BLOCK_WORN);
entry             104 drivers/mtd/nand/core.c 	unsigned int entry;
entry             111 drivers/mtd/nand/core.c 	entry = nanddev_bbt_pos_to_entry(nand, pos);
entry             112 drivers/mtd/nand/core.c 	status = nanddev_bbt_get_block_status(nand, entry);
entry            2719 drivers/mtd/nand/raw/marvell_nand.c 	struct marvell_nand_chip *entry, *temp;
entry            2721 drivers/mtd/nand/raw/marvell_nand.c 	list_for_each_entry_safe(entry, temp, &nfc->chips, node) {
entry            2722 drivers/mtd/nand/raw/marvell_nand.c 		nand_release(&entry->chip);
entry            2723 drivers/mtd/nand/raw/marvell_nand.c 		list_del(&entry->node);
entry              78 drivers/mtd/nand/raw/nand_bbt.c 	uint8_t entry = chip->bbt[block >> BBT_ENTRY_SHIFT];
entry              79 drivers/mtd/nand/raw/nand_bbt.c 	entry >>= (block & BBT_ENTRY_MASK) * 2;
entry              80 drivers/mtd/nand/raw/nand_bbt.c 	return entry & BBT_ENTRY_MASK;
entry             108 drivers/mtd/rfd_ftl.c 		u16 entry;
entry             110 drivers/mtd/rfd_ftl.c 		entry = le16_to_cpu(part->header_cache[HEADER_MAP_OFFSET + i]);
entry             112 drivers/mtd/rfd_ftl.c 		if (entry == SECTOR_DELETED)
entry             115 drivers/mtd/rfd_ftl.c 		if (entry == SECTOR_FREE) {
entry             120 drivers/mtd/rfd_ftl.c 		if (entry == SECTOR_ZERO)
entry             121 drivers/mtd/rfd_ftl.c 			entry = 0;
entry             123 drivers/mtd/rfd_ftl.c 		if (entry >= part->sector_count) {
entry             127 drivers/mtd/rfd_ftl.c 				part->mbd.mtd->name, block_no, i, entry);
entry             131 drivers/mtd/rfd_ftl.c 		if (part->sector_map[entry] != -1) {
entry             134 drivers/mtd/rfd_ftl.c 				part->mbd.mtd->name, entry);
entry             139 drivers/mtd/rfd_ftl.c 		part->sector_map[entry] = block->offset +
entry             354 drivers/mtd/rfd_ftl.c 		u16 entry = le16_to_cpu(map[HEADER_MAP_OFFSET + i]);
entry             358 drivers/mtd/rfd_ftl.c 		if (entry == SECTOR_FREE || entry == SECTOR_DELETED)
entry             361 drivers/mtd/rfd_ftl.c 		if (entry == SECTOR_ZERO)
entry             362 drivers/mtd/rfd_ftl.c 			entry = 0;
entry             365 drivers/mtd/rfd_ftl.c 		if (entry >= part->sector_count)
entry             394 drivers/mtd/rfd_ftl.c 				entry, sector_data);
entry             612 drivers/mtd/rfd_ftl.c 	u16 entry;
entry             647 drivers/mtd/rfd_ftl.c 	entry = cpu_to_le16(sector == 0 ? SECTOR_ZERO : sector);
entry             649 drivers/mtd/rfd_ftl.c 	part->header_cache[i + HEADER_MAP_OFFSET] = entry;
entry             652 drivers/mtd/rfd_ftl.c 	rc = mtd_write(part->mbd.mtd, addr, sizeof(entry), &retlen,
entry             653 drivers/mtd/rfd_ftl.c 		       (u_char *)&entry);
entry             655 drivers/mtd/rfd_ftl.c 	if (!rc && retlen != sizeof(entry))
entry            1656 drivers/mtd/ubi/eba.c 				struct ubi_eba_entry *entry;
entry            1658 drivers/mtd/ubi/eba.c 				entry = &vol->eba_tbl->entries[aeb->lnum];
entry            1659 drivers/mtd/ubi/eba.c 				entry->pnum = aeb->pnum;
entry              74 drivers/net/bonding/bond_alb.c static inline void tlb_init_table_entry(struct tlb_client_info *entry, int save_load)
entry              77 drivers/net/bonding/bond_alb.c 		entry->load_history = 1 + entry->tx_bytes /
entry              79 drivers/net/bonding/bond_alb.c 		entry->tx_bytes = 0;
entry              82 drivers/net/bonding/bond_alb.c 	entry->tx_slave = NULL;
entry              83 drivers/net/bonding/bond_alb.c 	entry->next = TLB_NULL_INDEX;
entry              84 drivers/net/bonding/bond_alb.c 	entry->prev = TLB_NULL_INDEX;
entry             731 drivers/net/bonding/bond_alb.c static void rlb_init_table_entry_dst(struct rlb_client_info *entry)
entry             733 drivers/net/bonding/bond_alb.c 	entry->used_next = RLB_NULL_INDEX;
entry             734 drivers/net/bonding/bond_alb.c 	entry->used_prev = RLB_NULL_INDEX;
entry             735 drivers/net/bonding/bond_alb.c 	entry->assigned = 0;
entry             736 drivers/net/bonding/bond_alb.c 	entry->slave = NULL;
entry             737 drivers/net/bonding/bond_alb.c 	entry->vlan_id = 0;
entry             739 drivers/net/bonding/bond_alb.c static void rlb_init_table_entry_src(struct rlb_client_info *entry)
entry             741 drivers/net/bonding/bond_alb.c 	entry->src_first = RLB_NULL_INDEX;
entry             742 drivers/net/bonding/bond_alb.c 	entry->src_prev = RLB_NULL_INDEX;
entry             743 drivers/net/bonding/bond_alb.c 	entry->src_next = RLB_NULL_INDEX;
entry             746 drivers/net/bonding/bond_alb.c static void rlb_init_table_entry(struct rlb_client_info *entry)
entry             748 drivers/net/bonding/bond_alb.c 	memset(entry, 0, sizeof(struct rlb_client_info));
entry             749 drivers/net/bonding/bond_alb.c 	rlb_init_table_entry_dst(entry);
entry             750 drivers/net/bonding/bond_alb.c 	rlb_init_table_entry_src(entry);
entry             794 drivers/net/bonding/bond_alb.c 	struct rlb_client_info *entry = &(bond_info->rx_hashtbl[index]);
entry             797 drivers/net/bonding/bond_alb.c 	rlb_init_table_entry_dst(entry);
entry             831 drivers/net/bonding/bond_alb.c 		struct rlb_client_info *entry = &(bond_info->rx_hashtbl[index]);
entry             832 drivers/net/bonding/bond_alb.c 		u32 next_index = entry->src_next;
entry             833 drivers/net/bonding/bond_alb.c 		if (entry->ip_src == arp->ip_src &&
entry             834 drivers/net/bonding/bond_alb.c 		    !ether_addr_equal_64bits(arp->mac_src, entry->mac_src))
entry             440 drivers/net/can/mscan/mscan.c 			struct tx_queue_entry *entry =
entry             442 drivers/net/can/mscan/mscan.c 			u8 mask = entry->mask;
entry             450 drivers/net/can/mscan/mscan.c 			can_get_echo_skb(dev, entry->id);
entry             470 drivers/net/can/rcar/rcar_canfd.c #define RCANFD_GAFL_PAGENUM(entry)	((entry) / 16)
entry             252 drivers/net/dsa/b53/b53_common.c 		u32 entry = 0;
entry             255 drivers/net/dsa/b53/b53_common.c 			entry = ((vlan->untag & VA_UNTAG_MASK_25) <<
entry             258 drivers/net/dsa/b53/b53_common.c 				entry |= VA_VALID_25_R4 | vid << VA_VID_HIGH_S;
entry             260 drivers/net/dsa/b53/b53_common.c 				entry |= VA_VALID_25;
entry             263 drivers/net/dsa/b53/b53_common.c 		b53_write32(dev, B53_VLAN_PAGE, B53_VLAN_WRITE_25, entry);
entry             267 drivers/net/dsa/b53/b53_common.c 		u16 entry = 0;
entry             270 drivers/net/dsa/b53/b53_common.c 			entry = ((vlan->untag & VA_UNTAG_MASK_65) <<
entry             273 drivers/net/dsa/b53/b53_common.c 		b53_write16(dev, B53_VLAN_PAGE, B53_VLAN_WRITE_65, entry);
entry             292 drivers/net/dsa/b53/b53_common.c 		u32 entry = 0;
entry             296 drivers/net/dsa/b53/b53_common.c 		b53_read32(dev, B53_VLAN_PAGE, B53_VLAN_WRITE_25, &entry);
entry             299 drivers/net/dsa/b53/b53_common.c 			vlan->valid = !!(entry & VA_VALID_25_R4);
entry             301 drivers/net/dsa/b53/b53_common.c 			vlan->valid = !!(entry & VA_VALID_25);
entry             302 drivers/net/dsa/b53/b53_common.c 		vlan->members = entry & VA_MEMBER_MASK;
entry             303 drivers/net/dsa/b53/b53_common.c 		vlan->untag = (entry >> VA_UNTAG_S_25) & VA_UNTAG_MASK_25;
entry             306 drivers/net/dsa/b53/b53_common.c 		u16 entry = 0;
entry             310 drivers/net/dsa/b53/b53_common.c 		b53_read16(dev, B53_VLAN_PAGE, B53_VLAN_WRITE_65, &entry);
entry             312 drivers/net/dsa/b53/b53_common.c 		vlan->valid = !!(entry & VA_VALID_65);
entry             313 drivers/net/dsa/b53/b53_common.c 		vlan->members = entry & VA_MEMBER_MASK;
entry             314 drivers/net/dsa/b53/b53_common.c 		vlan->untag = (entry >> VA_UNTAG_S_65) & VA_UNTAG_MASK_65;
entry             316 drivers/net/dsa/b53/b53_common.c 		u32 entry = 0;
entry             320 drivers/net/dsa/b53/b53_common.c 		b53_read32(dev, B53_ARLIO_PAGE, dev->vta_regs[2], &entry);
entry             321 drivers/net/dsa/b53/b53_common.c 		vlan->members = entry & VTE_MEMBERS;
entry             322 drivers/net/dsa/b53/b53_common.c 		vlan->untag = (entry >> VTE_UNTAG_S) & VTE_MEMBERS;
entry            1001 drivers/net/dsa/mt7530.c 		   struct mt7530_hw_vlan_entry *entry)
entry            1006 drivers/net/dsa/mt7530.c 	new_members = entry->old_members | BIT(entry->port) |
entry            1018 drivers/net/dsa/mt7530.c 	val = entry->untagged ? MT7530_VLAN_EGRESS_UNTAG :
entry            1021 drivers/net/dsa/mt7530.c 		   ETAG_CTRL_P_MASK(entry->port),
entry            1022 drivers/net/dsa/mt7530.c 		   ETAG_CTRL_P(entry->port, val));
entry            1037 drivers/net/dsa/mt7530.c 		   struct mt7530_hw_vlan_entry *entry)
entry            1042 drivers/net/dsa/mt7530.c 	new_members = entry->old_members & ~BIT(entry->port);
entry            1067 drivers/net/dsa/mt7530.c 		      struct mt7530_hw_vlan_entry *entry,
entry            1077 drivers/net/dsa/mt7530.c 	entry->old_members = (val >> PORT_MEM_SHFT) & PORT_MEM_MASK;
entry            1080 drivers/net/dsa/mt7530.c 	vlan_op(priv, entry);
entry            1313 drivers/net/dsa/mv88e6xxx/chip.c 				 struct mv88e6xxx_vtu_entry *entry)
entry            1318 drivers/net/dsa/mv88e6xxx/chip.c 	return chip->info->ops->vtu_getnext(chip, entry);
entry            1322 drivers/net/dsa/mv88e6xxx/chip.c 				   struct mv88e6xxx_vtu_entry *entry)
entry            1327 drivers/net/dsa/mv88e6xxx/chip.c 	return chip->info->ops->vtu_loadpurge(chip, entry);
entry            1475 drivers/net/dsa/mv88e6xxx/chip.c 	struct mv88e6xxx_atu_entry entry;
entry            1500 drivers/net/dsa/mv88e6xxx/chip.c 	entry.state = 0;
entry            1501 drivers/net/dsa/mv88e6xxx/chip.c 	ether_addr_copy(entry.mac, addr);
entry            1502 drivers/net/dsa/mv88e6xxx/chip.c 	eth_addr_dec(entry.mac);
entry            1504 drivers/net/dsa/mv88e6xxx/chip.c 	err = mv88e6xxx_g1_atu_getnext(chip, fid, &entry);
entry            1509 drivers/net/dsa/mv88e6xxx/chip.c 	if (!entry.state || !ether_addr_equal(entry.mac, addr)) {
entry            1510 drivers/net/dsa/mv88e6xxx/chip.c 		memset(&entry, 0, sizeof(entry));
entry            1511 drivers/net/dsa/mv88e6xxx/chip.c 		ether_addr_copy(entry.mac, addr);
entry            1516 drivers/net/dsa/mv88e6xxx/chip.c 		entry.portvec &= ~BIT(port);
entry            1517 drivers/net/dsa/mv88e6xxx/chip.c 		if (!entry.portvec)
entry            1518 drivers/net/dsa/mv88e6xxx/chip.c 			entry.state = 0;
entry            1520 drivers/net/dsa/mv88e6xxx/chip.c 		entry.portvec |= BIT(port);
entry            1521 drivers/net/dsa/mv88e6xxx/chip.c 		entry.state = state;
entry            1524 drivers/net/dsa/mv88e6xxx/chip.c 	return mv88e6xxx_g1_atu_loadpurge(chip, fid, &entry);
entry             502 drivers/net/dsa/mv88e6xxx/chip.h 			   struct mv88e6xxx_vtu_entry *entry);
entry             504 drivers/net/dsa/mv88e6xxx/chip.h 			     struct mv88e6xxx_vtu_entry *entry);
entry             314 drivers/net/dsa/mv88e6xxx/global1.h 			     struct mv88e6xxx_atu_entry *entry);
entry             316 drivers/net/dsa/mv88e6xxx/global1.h 			       struct mv88e6xxx_atu_entry *entry);
entry             324 drivers/net/dsa/mv88e6xxx/global1.h 			     struct mv88e6xxx_vtu_entry *entry);
entry             326 drivers/net/dsa/mv88e6xxx/global1.h 			       struct mv88e6xxx_vtu_entry *entry);
entry             328 drivers/net/dsa/mv88e6xxx/global1.h 			     struct mv88e6xxx_vtu_entry *entry);
entry             330 drivers/net/dsa/mv88e6xxx/global1.h 			       struct mv88e6xxx_vtu_entry *entry);
entry             332 drivers/net/dsa/mv88e6xxx/global1.h 			     struct mv88e6xxx_vtu_entry *entry);
entry             334 drivers/net/dsa/mv88e6xxx/global1.h 			       struct mv88e6xxx_vtu_entry *entry);
entry             336 drivers/net/dsa/mv88e6xxx/global1.h 			     struct mv88e6xxx_vtu_entry *entry);
entry             338 drivers/net/dsa/mv88e6xxx/global1.h 			       struct mv88e6xxx_vtu_entry *entry);
entry             128 drivers/net/dsa/mv88e6xxx/global1_atu.c 				      struct mv88e6xxx_atu_entry *entry)
entry             137 drivers/net/dsa/mv88e6xxx/global1_atu.c 	entry->state = val & 0xf;
entry             138 drivers/net/dsa/mv88e6xxx/global1_atu.c 	if (entry->state) {
entry             139 drivers/net/dsa/mv88e6xxx/global1_atu.c 		entry->trunk = !!(val & MV88E6XXX_G1_ATU_DATA_TRUNK);
entry             140 drivers/net/dsa/mv88e6xxx/global1_atu.c 		entry->portvec = (val >> 4) & mv88e6xxx_port_mask(chip);
entry             147 drivers/net/dsa/mv88e6xxx/global1_atu.c 				       struct mv88e6xxx_atu_entry *entry)
entry             149 drivers/net/dsa/mv88e6xxx/global1_atu.c 	u16 data = entry->state & 0xf;
entry             151 drivers/net/dsa/mv88e6xxx/global1_atu.c 	if (entry->state) {
entry             152 drivers/net/dsa/mv88e6xxx/global1_atu.c 		if (entry->trunk)
entry             155 drivers/net/dsa/mv88e6xxx/global1_atu.c 		data |= (entry->portvec & mv88e6xxx_port_mask(chip)) << 4;
entry             167 drivers/net/dsa/mv88e6xxx/global1_atu.c 				     struct mv88e6xxx_atu_entry *entry)
entry             177 drivers/net/dsa/mv88e6xxx/global1_atu.c 		entry->mac[i * 2] = val >> 8;
entry             178 drivers/net/dsa/mv88e6xxx/global1_atu.c 		entry->mac[i * 2 + 1] = val & 0xff;
entry             185 drivers/net/dsa/mv88e6xxx/global1_atu.c 				      struct mv88e6xxx_atu_entry *entry)
entry             191 drivers/net/dsa/mv88e6xxx/global1_atu.c 		val = (entry->mac[i * 2] << 8) | entry->mac[i * 2 + 1];
entry             203 drivers/net/dsa/mv88e6xxx/global1_atu.c 			     struct mv88e6xxx_atu_entry *entry)
entry             212 drivers/net/dsa/mv88e6xxx/global1_atu.c 	if (!entry->state) {
entry             213 drivers/net/dsa/mv88e6xxx/global1_atu.c 		err = mv88e6xxx_g1_atu_mac_write(chip, entry);
entry             222 drivers/net/dsa/mv88e6xxx/global1_atu.c 	err = mv88e6xxx_g1_atu_data_read(chip, entry);
entry             226 drivers/net/dsa/mv88e6xxx/global1_atu.c 	return mv88e6xxx_g1_atu_mac_read(chip, entry);
entry             230 drivers/net/dsa/mv88e6xxx/global1_atu.c 			       struct mv88e6xxx_atu_entry *entry)
entry             238 drivers/net/dsa/mv88e6xxx/global1_atu.c 	err = mv88e6xxx_g1_atu_mac_write(chip, entry);
entry             242 drivers/net/dsa/mv88e6xxx/global1_atu.c 	err = mv88e6xxx_g1_atu_data_write(chip, entry);
entry             250 drivers/net/dsa/mv88e6xxx/global1_atu.c 				      struct mv88e6xxx_atu_entry *entry,
entry             260 drivers/net/dsa/mv88e6xxx/global1_atu.c 	err = mv88e6xxx_g1_atu_data_write(chip, entry);
entry             279 drivers/net/dsa/mv88e6xxx/global1_atu.c 	struct mv88e6xxx_atu_entry entry = {
entry             283 drivers/net/dsa/mv88e6xxx/global1_atu.c 	return mv88e6xxx_g1_atu_flushmove(chip, fid, &entry, all);
entry             289 drivers/net/dsa/mv88e6xxx/global1_atu.c 	struct mv88e6xxx_atu_entry entry = { 0 };
entry             299 drivers/net/dsa/mv88e6xxx/global1_atu.c 	entry.state = 0xf, /* Full EntryState means Move */
entry             300 drivers/net/dsa/mv88e6xxx/global1_atu.c 	entry.portvec = from_port & mask;
entry             301 drivers/net/dsa/mv88e6xxx/global1_atu.c 	entry.portvec |= (to_port & mask) << shift;
entry             303 drivers/net/dsa/mv88e6xxx/global1_atu.c 	return mv88e6xxx_g1_atu_flushmove(chip, fid, &entry, all);
entry             318 drivers/net/dsa/mv88e6xxx/global1_atu.c 	struct mv88e6xxx_atu_entry entry;
entry             334 drivers/net/dsa/mv88e6xxx/global1_atu.c 	err = mv88e6xxx_g1_atu_data_read(chip, &entry);
entry             338 drivers/net/dsa/mv88e6xxx/global1_atu.c 	err = mv88e6xxx_g1_atu_mac_read(chip, &entry);
entry             342 drivers/net/dsa/mv88e6xxx/global1_atu.c 	spid = entry.state;
entry             347 drivers/net/dsa/mv88e6xxx/global1_atu.c 				    entry.mac);
entry             353 drivers/net/dsa/mv88e6xxx/global1_atu.c 				    entry.mac, entry.portvec, spid);
entry             360 drivers/net/dsa/mv88e6xxx/global1_atu.c 				    entry.mac, entry.portvec, spid);
entry             367 drivers/net/dsa/mv88e6xxx/global1_atu.c 				    entry.mac, entry.portvec, spid);
entry              20 drivers/net/dsa/mv88e6xxx/global1_vtu.c 				     struct mv88e6xxx_vtu_entry *entry)
entry              29 drivers/net/dsa/mv88e6xxx/global1_vtu.c 	entry->fid = val & MV88E6352_G1_VTU_FID_MASK;
entry              35 drivers/net/dsa/mv88e6xxx/global1_vtu.c 				      struct mv88e6xxx_vtu_entry *entry)
entry              37 drivers/net/dsa/mv88e6xxx/global1_vtu.c 	u16 val = entry->fid & MV88E6352_G1_VTU_FID_MASK;
entry              45 drivers/net/dsa/mv88e6xxx/global1_vtu.c 				     struct mv88e6xxx_vtu_entry *entry)
entry              54 drivers/net/dsa/mv88e6xxx/global1_vtu.c 	entry->sid = val & MV88E6352_G1_VTU_SID_MASK;
entry              60 drivers/net/dsa/mv88e6xxx/global1_vtu.c 				      struct mv88e6xxx_vtu_entry *entry)
entry              62 drivers/net/dsa/mv88e6xxx/global1_vtu.c 	u16 val = entry->sid & MV88E6352_G1_VTU_SID_MASK;
entry              91 drivers/net/dsa/mv88e6xxx/global1_vtu.c 				     struct mv88e6xxx_vtu_entry *entry)
entry             100 drivers/net/dsa/mv88e6xxx/global1_vtu.c 	entry->vid = val & 0xfff;
entry             103 drivers/net/dsa/mv88e6xxx/global1_vtu.c 		entry->vid |= 0x1000;
entry             105 drivers/net/dsa/mv88e6xxx/global1_vtu.c 	entry->valid = !!(val & MV88E6XXX_G1_VTU_VID_VALID);
entry             111 drivers/net/dsa/mv88e6xxx/global1_vtu.c 				      struct mv88e6xxx_vtu_entry *entry)
entry             113 drivers/net/dsa/mv88e6xxx/global1_vtu.c 	u16 val = entry->vid & 0xfff;
entry             115 drivers/net/dsa/mv88e6xxx/global1_vtu.c 	if (entry->vid & 0x1000)
entry             118 drivers/net/dsa/mv88e6xxx/global1_vtu.c 	if (entry->valid)
entry             130 drivers/net/dsa/mv88e6xxx/global1_vtu.c 				      struct mv88e6xxx_vtu_entry *entry)
entry             150 drivers/net/dsa/mv88e6xxx/global1_vtu.c 		entry->member[i] = (regs[i / 4] >> member_offset) & 0x3;
entry             151 drivers/net/dsa/mv88e6xxx/global1_vtu.c 		entry->state[i] = (regs[i / 4] >> state_offset) & 0x3;
entry             158 drivers/net/dsa/mv88e6xxx/global1_vtu.c 				       struct mv88e6xxx_vtu_entry *entry)
entry             168 drivers/net/dsa/mv88e6xxx/global1_vtu.c 		regs[i / 4] |= (entry->member[i] & 0x3) << member_offset;
entry             169 drivers/net/dsa/mv88e6xxx/global1_vtu.c 		regs[i / 4] |= (entry->state[i] & 0x3) << state_offset;
entry             238 drivers/net/dsa/mv88e6xxx/global1_vtu.c 					struct mv88e6xxx_vtu_entry *entry)
entry             242 drivers/net/dsa/mv88e6xxx/global1_vtu.c 	err = mv88e6xxx_g1_vtu_sid_write(chip, entry);
entry             250 drivers/net/dsa/mv88e6xxx/global1_vtu.c 	err = mv88e6xxx_g1_vtu_sid_read(chip, entry);
entry             254 drivers/net/dsa/mv88e6xxx/global1_vtu.c 	return mv88e6xxx_g1_vtu_vid_read(chip, entry);
entry             280 drivers/net/dsa/mv88e6xxx/global1_vtu.c 				    struct mv88e6xxx_vtu_entry *entry)
entry             295 drivers/net/dsa/mv88e6xxx/global1_vtu.c 	if (!entry->valid) {
entry             296 drivers/net/dsa/mv88e6xxx/global1_vtu.c 		err = mv88e6xxx_g1_vtu_vid_write(chip, entry);
entry             305 drivers/net/dsa/mv88e6xxx/global1_vtu.c 	return mv88e6xxx_g1_vtu_vid_read(chip, entry);
entry             309 drivers/net/dsa/mv88e6xxx/global1_vtu.c 			     struct mv88e6xxx_vtu_entry *entry)
entry             314 drivers/net/dsa/mv88e6xxx/global1_vtu.c 	err = mv88e6xxx_g1_vtu_getnext(chip, entry);
entry             318 drivers/net/dsa/mv88e6xxx/global1_vtu.c 	if (entry->valid) {
entry             319 drivers/net/dsa/mv88e6xxx/global1_vtu.c 		err = mv88e6185_g1_vtu_data_read(chip, entry);
entry             330 drivers/net/dsa/mv88e6xxx/global1_vtu.c 		entry->fid = val & 0x000f;
entry             331 drivers/net/dsa/mv88e6xxx/global1_vtu.c 		entry->fid |= (val & 0x0300) >> 4;
entry             338 drivers/net/dsa/mv88e6xxx/global1_vtu.c 			     struct mv88e6xxx_vtu_entry *entry)
entry             343 drivers/net/dsa/mv88e6xxx/global1_vtu.c 	err = mv88e6xxx_g1_vtu_getnext(chip, entry);
entry             347 drivers/net/dsa/mv88e6xxx/global1_vtu.c 	if (entry->valid) {
entry             348 drivers/net/dsa/mv88e6xxx/global1_vtu.c 		err = mv88e6185_g1_vtu_data_read(chip, entry);
entry             359 drivers/net/dsa/mv88e6xxx/global1_vtu.c 		entry->fid = val & 0x000f;
entry             360 drivers/net/dsa/mv88e6xxx/global1_vtu.c 		entry->fid |= (val & 0x0f00) >> 4;
entry             367 drivers/net/dsa/mv88e6xxx/global1_vtu.c 			     struct mv88e6xxx_vtu_entry *entry)
entry             372 drivers/net/dsa/mv88e6xxx/global1_vtu.c 	err = mv88e6xxx_g1_vtu_getnext(chip, entry);
entry             376 drivers/net/dsa/mv88e6xxx/global1_vtu.c 	if (entry->valid) {
entry             378 drivers/net/dsa/mv88e6xxx/global1_vtu.c 		err = mv88e6xxx_g1_vtu_stu_get(chip, entry);
entry             382 drivers/net/dsa/mv88e6xxx/global1_vtu.c 		err = mv88e6185_g1_vtu_data_read(chip, entry);
entry             386 drivers/net/dsa/mv88e6xxx/global1_vtu.c 		err = mv88e6xxx_g1_vtu_fid_read(chip, entry);
entry             395 drivers/net/dsa/mv88e6xxx/global1_vtu.c 			     struct mv88e6xxx_vtu_entry *entry)
entry             400 drivers/net/dsa/mv88e6xxx/global1_vtu.c 	err = mv88e6xxx_g1_vtu_getnext(chip, entry);
entry             404 drivers/net/dsa/mv88e6xxx/global1_vtu.c 	if (entry->valid) {
entry             405 drivers/net/dsa/mv88e6xxx/global1_vtu.c 		err = mv88e6390_g1_vtu_data_read(chip, entry->member);
entry             410 drivers/net/dsa/mv88e6xxx/global1_vtu.c 		err = mv88e6xxx_g1_vtu_stu_get(chip, entry);
entry             414 drivers/net/dsa/mv88e6xxx/global1_vtu.c 		err = mv88e6390_g1_vtu_data_read(chip, entry->state);
entry             418 drivers/net/dsa/mv88e6xxx/global1_vtu.c 		err = mv88e6xxx_g1_vtu_fid_read(chip, entry);
entry             427 drivers/net/dsa/mv88e6xxx/global1_vtu.c 			       struct mv88e6xxx_vtu_entry *entry)
entry             436 drivers/net/dsa/mv88e6xxx/global1_vtu.c 	err = mv88e6xxx_g1_vtu_vid_write(chip, entry);
entry             440 drivers/net/dsa/mv88e6xxx/global1_vtu.c 	if (entry->valid) {
entry             441 drivers/net/dsa/mv88e6xxx/global1_vtu.c 		err = mv88e6185_g1_vtu_data_write(chip, entry);
entry             448 drivers/net/dsa/mv88e6xxx/global1_vtu.c 		op |= entry->fid & 0x000f;
entry             449 drivers/net/dsa/mv88e6xxx/global1_vtu.c 		op |= (entry->fid & 0x0030) << 4;
entry             456 drivers/net/dsa/mv88e6xxx/global1_vtu.c 			       struct mv88e6xxx_vtu_entry *entry)
entry             465 drivers/net/dsa/mv88e6xxx/global1_vtu.c 	err = mv88e6xxx_g1_vtu_vid_write(chip, entry);
entry             469 drivers/net/dsa/mv88e6xxx/global1_vtu.c 	if (entry->valid) {
entry             470 drivers/net/dsa/mv88e6xxx/global1_vtu.c 		err = mv88e6185_g1_vtu_data_write(chip, entry);
entry             477 drivers/net/dsa/mv88e6xxx/global1_vtu.c 		op |= entry->fid & 0x000f;
entry             478 drivers/net/dsa/mv88e6xxx/global1_vtu.c 		op |= (entry->fid & 0x00f0) << 4;
entry             485 drivers/net/dsa/mv88e6xxx/global1_vtu.c 			       struct mv88e6xxx_vtu_entry *entry)
entry             493 drivers/net/dsa/mv88e6xxx/global1_vtu.c 	err = mv88e6xxx_g1_vtu_vid_write(chip, entry);
entry             497 drivers/net/dsa/mv88e6xxx/global1_vtu.c 	if (entry->valid) {
entry             499 drivers/net/dsa/mv88e6xxx/global1_vtu.c 		err = mv88e6185_g1_vtu_data_write(chip, entry);
entry             503 drivers/net/dsa/mv88e6xxx/global1_vtu.c 		err = mv88e6xxx_g1_vtu_sid_write(chip, entry);
entry             513 drivers/net/dsa/mv88e6xxx/global1_vtu.c 		err = mv88e6xxx_g1_vtu_fid_write(chip, entry);
entry             523 drivers/net/dsa/mv88e6xxx/global1_vtu.c 			       struct mv88e6xxx_vtu_entry *entry)
entry             531 drivers/net/dsa/mv88e6xxx/global1_vtu.c 	err = mv88e6xxx_g1_vtu_vid_write(chip, entry);
entry             535 drivers/net/dsa/mv88e6xxx/global1_vtu.c 	if (entry->valid) {
entry             537 drivers/net/dsa/mv88e6xxx/global1_vtu.c 		err = mv88e6390_g1_vtu_data_write(chip, entry->state);
entry             541 drivers/net/dsa/mv88e6xxx/global1_vtu.c 		err = mv88e6xxx_g1_vtu_sid_write(chip, entry);
entry             552 drivers/net/dsa/mv88e6xxx/global1_vtu.c 		err = mv88e6390_g1_vtu_data_write(chip, entry->member);
entry             556 drivers/net/dsa/mv88e6xxx/global1_vtu.c 		err = mv88e6xxx_g1_vtu_fid_write(chip, entry);
entry             579 drivers/net/dsa/mv88e6xxx/global1_vtu.c 	struct mv88e6xxx_vtu_entry entry;
entry             594 drivers/net/dsa/mv88e6xxx/global1_vtu.c 	err = mv88e6xxx_g1_vtu_vid_read(chip, &entry);
entry             602 drivers/net/dsa/mv88e6xxx/global1_vtu.c 				    entry.vid, spid);
entry             608 drivers/net/dsa/mv88e6xxx/global1_vtu.c 				    entry.vid, spid);
entry             183 drivers/net/dsa/sja1105/sja1105.h 				int index, void *entry);
entry             186 drivers/net/dsa/sja1105/sja1105.h 				 int index, void *entry, bool keep);
entry             255 drivers/net/dsa/sja1105/sja1105_dynamic_config.c 	struct sja1105_l2_lookup_entry *entry = entry_ptr;
entry             259 drivers/net/dsa/sja1105/sja1105_dynamic_config.c 	sja1105_packing(cmd, &entry->lockeds, 28, 28, size, op);
entry             283 drivers/net/dsa/sja1105/sja1105_dynamic_config.c 	struct sja1105_l2_lookup_entry *entry = entry_ptr;
entry             287 drivers/net/dsa/sja1105/sja1105_dynamic_config.c 	sja1105_packing(cmd, &entry->lockeds, 28, 28, size, op);
entry             307 drivers/net/dsa/sja1105/sja1105_dynamic_config.c 	struct sja1105_mgmt_entry *entry = entry_ptr;
entry             316 drivers/net/dsa/sja1105/sja1105_dynamic_config.c 	sja1105_packing(buf, &entry->tsreg,     85, 85, size, op);
entry             317 drivers/net/dsa/sja1105/sja1105_dynamic_config.c 	sja1105_packing(buf, &entry->takets,    84, 84, size, op);
entry             318 drivers/net/dsa/sja1105/sja1105_dynamic_config.c 	sja1105_packing(buf, &entry->macaddr,   83, 36, size, op);
entry             319 drivers/net/dsa/sja1105/sja1105_dynamic_config.c 	sja1105_packing(buf, &entry->destports, 35, 31, size, op);
entry             320 drivers/net/dsa/sja1105/sja1105_dynamic_config.c 	sja1105_packing(buf, &entry->enfport,   30, 30, size, op);
entry             340 drivers/net/dsa/sja1105/sja1105_dynamic_config.c 	struct sja1105_mgmt_entry *entry = entry_ptr;
entry             346 drivers/net/dsa/sja1105/sja1105_dynamic_config.c 	sja1105_packing(buf, &entry->tsreg,     71, 71, size, op);
entry             347 drivers/net/dsa/sja1105/sja1105_dynamic_config.c 	sja1105_packing(buf, &entry->takets,    70, 70, size, op);
entry             348 drivers/net/dsa/sja1105/sja1105_dynamic_config.c 	sja1105_packing(buf, &entry->macaddr,   69, 22, size, op);
entry             349 drivers/net/dsa/sja1105/sja1105_dynamic_config.c 	sja1105_packing(buf, &entry->destports, 21, 17, size, op);
entry             350 drivers/net/dsa/sja1105/sja1105_dynamic_config.c 	sja1105_packing(buf, &entry->enfport,   16, 16, size, op);
entry             404 drivers/net/dsa/sja1105/sja1105_dynamic_config.c 	struct sja1105_mac_config_entry *entry = entry_ptr;
entry             409 drivers/net/dsa/sja1105/sja1105_dynamic_config.c 	sja1105_packing(reg1, &entry->speed,     30, 29, size, op);
entry             410 drivers/net/dsa/sja1105/sja1105_dynamic_config.c 	sja1105_packing(reg1, &entry->drpdtag,   23, 23, size, op);
entry             411 drivers/net/dsa/sja1105/sja1105_dynamic_config.c 	sja1105_packing(reg1, &entry->drpuntag,  22, 22, size, op);
entry             412 drivers/net/dsa/sja1105/sja1105_dynamic_config.c 	sja1105_packing(reg1, &entry->retag,     21, 21, size, op);
entry             413 drivers/net/dsa/sja1105/sja1105_dynamic_config.c 	sja1105_packing(reg1, &entry->dyn_learn, 20, 20, size, op);
entry             414 drivers/net/dsa/sja1105/sja1105_dynamic_config.c 	sja1105_packing(reg1, &entry->egress,    19, 19, size, op);
entry             415 drivers/net/dsa/sja1105/sja1105_dynamic_config.c 	sja1105_packing(reg1, &entry->ingress,   18, 18, size, op);
entry             416 drivers/net/dsa/sja1105/sja1105_dynamic_config.c 	sja1105_packing(reg1, &entry->ing_mirr,  17, 17, size, op);
entry             417 drivers/net/dsa/sja1105/sja1105_dynamic_config.c 	sja1105_packing(reg1, &entry->egr_mirr,  16, 16, size, op);
entry             418 drivers/net/dsa/sja1105/sja1105_dynamic_config.c 	sja1105_packing(reg1, &entry->vlanprio,  14, 12, size, op);
entry             419 drivers/net/dsa/sja1105/sja1105_dynamic_config.c 	sja1105_packing(reg1, &entry->vlanid,    11,  0, size, op);
entry             420 drivers/net/dsa/sja1105/sja1105_dynamic_config.c 	sja1105_packing(reg2, &entry->tp_delin,  31, 16, size, op);
entry             421 drivers/net/dsa/sja1105/sja1105_dynamic_config.c 	sja1105_packing(reg2, &entry->tp_delout, 15,  0, size, op);
entry             454 drivers/net/dsa/sja1105/sja1105_dynamic_config.c 	struct sja1105_l2_lookup_params_entry *entry = entry_ptr;
entry             456 drivers/net/dsa/sja1105/sja1105_dynamic_config.c 	sja1105_packing(buf, &entry->poly, 7, 0,
entry             476 drivers/net/dsa/sja1105/sja1105_dynamic_config.c 	struct sja1105_general_params_entry *entry = entry_ptr;
entry             479 drivers/net/dsa/sja1105/sja1105_dynamic_config.c 	sja1105_packing(buf, &entry->mirr_port, 2, 0, size, op);
entry             645 drivers/net/dsa/sja1105/sja1105_dynamic_config.c 				int index, void *entry)
entry             686 drivers/net/dsa/sja1105/sja1105_dynamic_config.c 		ops->entry_packing(packed_buf, entry, PACK);
entry             723 drivers/net/dsa/sja1105/sja1105_dynamic_config.c 	if (entry)
entry             724 drivers/net/dsa/sja1105/sja1105_dynamic_config.c 		ops->entry_packing(packed_buf, entry, UNPACK);
entry             730 drivers/net/dsa/sja1105/sja1105_dynamic_config.c 				 int index, void *entry, bool keep)
entry             771 drivers/net/dsa/sja1105/sja1105_dynamic_config.c 		ops->entry_packing(packed_buf, entry, PACK);
entry              98 drivers/net/dsa/sja1105/sja1105_static_config.c 	struct sja1105_avb_params_entry *entry = entry_ptr;
entry             100 drivers/net/dsa/sja1105/sja1105_static_config.c 	sja1105_packing(buf, &entry->destmeta, 95, 48, size, op);
entry             101 drivers/net/dsa/sja1105/sja1105_static_config.c 	sja1105_packing(buf, &entry->srcmeta,  47,  0, size, op);
entry             109 drivers/net/dsa/sja1105/sja1105_static_config.c 	struct sja1105_avb_params_entry *entry = entry_ptr;
entry             111 drivers/net/dsa/sja1105/sja1105_static_config.c 	sja1105_packing(buf, &entry->destmeta,   125,  78, size, op);
entry             112 drivers/net/dsa/sja1105/sja1105_static_config.c 	sja1105_packing(buf, &entry->srcmeta,     77,  30, size, op);
entry             120 drivers/net/dsa/sja1105/sja1105_static_config.c 	struct sja1105_general_params_entry *entry = entry_ptr;
entry             122 drivers/net/dsa/sja1105/sja1105_static_config.c 	sja1105_packing(buf, &entry->vllupformat, 319, 319, size, op);
entry             123 drivers/net/dsa/sja1105/sja1105_static_config.c 	sja1105_packing(buf, &entry->mirr_ptacu,  318, 318, size, op);
entry             124 drivers/net/dsa/sja1105/sja1105_static_config.c 	sja1105_packing(buf, &entry->switchid,    317, 315, size, op);
entry             125 drivers/net/dsa/sja1105/sja1105_static_config.c 	sja1105_packing(buf, &entry->hostprio,    314, 312, size, op);
entry             126 drivers/net/dsa/sja1105/sja1105_static_config.c 	sja1105_packing(buf, &entry->mac_fltres1, 311, 264, size, op);
entry             127 drivers/net/dsa/sja1105/sja1105_static_config.c 	sja1105_packing(buf, &entry->mac_fltres0, 263, 216, size, op);
entry             128 drivers/net/dsa/sja1105/sja1105_static_config.c 	sja1105_packing(buf, &entry->mac_flt1,    215, 168, size, op);
entry             129 drivers/net/dsa/sja1105/sja1105_static_config.c 	sja1105_packing(buf, &entry->mac_flt0,    167, 120, size, op);
entry             130 drivers/net/dsa/sja1105/sja1105_static_config.c 	sja1105_packing(buf, &entry->incl_srcpt1, 119, 119, size, op);
entry             131 drivers/net/dsa/sja1105/sja1105_static_config.c 	sja1105_packing(buf, &entry->incl_srcpt0, 118, 118, size, op);
entry             132 drivers/net/dsa/sja1105/sja1105_static_config.c 	sja1105_packing(buf, &entry->send_meta1,  117, 117, size, op);
entry             133 drivers/net/dsa/sja1105/sja1105_static_config.c 	sja1105_packing(buf, &entry->send_meta0,  116, 116, size, op);
entry             134 drivers/net/dsa/sja1105/sja1105_static_config.c 	sja1105_packing(buf, &entry->casc_port,   115, 113, size, op);
entry             135 drivers/net/dsa/sja1105/sja1105_static_config.c 	sja1105_packing(buf, &entry->host_port,   112, 110, size, op);
entry             136 drivers/net/dsa/sja1105/sja1105_static_config.c 	sja1105_packing(buf, &entry->mirr_port,   109, 107, size, op);
entry             137 drivers/net/dsa/sja1105/sja1105_static_config.c 	sja1105_packing(buf, &entry->vlmarker,    106,  75, size, op);
entry             138 drivers/net/dsa/sja1105/sja1105_static_config.c 	sja1105_packing(buf, &entry->vlmask,       74,  43, size, op);
entry             139 drivers/net/dsa/sja1105/sja1105_static_config.c 	sja1105_packing(buf, &entry->tpid,         42,  27, size, op);
entry             140 drivers/net/dsa/sja1105/sja1105_static_config.c 	sja1105_packing(buf, &entry->ignore2stf,   26,  26, size, op);
entry             141 drivers/net/dsa/sja1105/sja1105_static_config.c 	sja1105_packing(buf, &entry->tpid2,        25,  10, size, op);
entry             153 drivers/net/dsa/sja1105/sja1105_static_config.c 	struct sja1105_general_params_entry *entry = entry_ptr;
entry             155 drivers/net/dsa/sja1105/sja1105_static_config.c 	sja1105_packing(buf, &entry->vllupformat, 351, 351, size, op);
entry             156 drivers/net/dsa/sja1105/sja1105_static_config.c 	sja1105_packing(buf, &entry->mirr_ptacu,  350, 350, size, op);
entry             157 drivers/net/dsa/sja1105/sja1105_static_config.c 	sja1105_packing(buf, &entry->switchid,    349, 347, size, op);
entry             158 drivers/net/dsa/sja1105/sja1105_static_config.c 	sja1105_packing(buf, &entry->hostprio,    346, 344, size, op);
entry             159 drivers/net/dsa/sja1105/sja1105_static_config.c 	sja1105_packing(buf, &entry->mac_fltres1, 343, 296, size, op);
entry             160 drivers/net/dsa/sja1105/sja1105_static_config.c 	sja1105_packing(buf, &entry->mac_fltres0, 295, 248, size, op);
entry             161 drivers/net/dsa/sja1105/sja1105_static_config.c 	sja1105_packing(buf, &entry->mac_flt1,    247, 200, size, op);
entry             162 drivers/net/dsa/sja1105/sja1105_static_config.c 	sja1105_packing(buf, &entry->mac_flt0,    199, 152, size, op);
entry             163 drivers/net/dsa/sja1105/sja1105_static_config.c 	sja1105_packing(buf, &entry->incl_srcpt1, 151, 151, size, op);
entry             164 drivers/net/dsa/sja1105/sja1105_static_config.c 	sja1105_packing(buf, &entry->incl_srcpt0, 150, 150, size, op);
entry             165 drivers/net/dsa/sja1105/sja1105_static_config.c 	sja1105_packing(buf, &entry->send_meta1,  149, 149, size, op);
entry             166 drivers/net/dsa/sja1105/sja1105_static_config.c 	sja1105_packing(buf, &entry->send_meta0,  148, 148, size, op);
entry             167 drivers/net/dsa/sja1105/sja1105_static_config.c 	sja1105_packing(buf, &entry->casc_port,   147, 145, size, op);
entry             168 drivers/net/dsa/sja1105/sja1105_static_config.c 	sja1105_packing(buf, &entry->host_port,   144, 142, size, op);
entry             169 drivers/net/dsa/sja1105/sja1105_static_config.c 	sja1105_packing(buf, &entry->mirr_port,   141, 139, size, op);
entry             170 drivers/net/dsa/sja1105/sja1105_static_config.c 	sja1105_packing(buf, &entry->vlmarker,    138, 107, size, op);
entry             171 drivers/net/dsa/sja1105/sja1105_static_config.c 	sja1105_packing(buf, &entry->vlmask,      106,  75, size, op);
entry             172 drivers/net/dsa/sja1105/sja1105_static_config.c 	sja1105_packing(buf, &entry->tpid2,        74,  59, size, op);
entry             173 drivers/net/dsa/sja1105/sja1105_static_config.c 	sja1105_packing(buf, &entry->ignore2stf,   58,  58, size, op);
entry             174 drivers/net/dsa/sja1105/sja1105_static_config.c 	sja1105_packing(buf, &entry->tpid,         57,  42, size, op);
entry             175 drivers/net/dsa/sja1105/sja1105_static_config.c 	sja1105_packing(buf, &entry->queue_ts,     41,  41, size, op);
entry             176 drivers/net/dsa/sja1105/sja1105_static_config.c 	sja1105_packing(buf, &entry->egrmirrvid,   40,  29, size, op);
entry             177 drivers/net/dsa/sja1105/sja1105_static_config.c 	sja1105_packing(buf, &entry->egrmirrpcp,   28,  26, size, op);
entry             178 drivers/net/dsa/sja1105/sja1105_static_config.c 	sja1105_packing(buf, &entry->egrmirrdei,   25,  25, size, op);
entry             179 drivers/net/dsa/sja1105/sja1105_static_config.c 	sja1105_packing(buf, &entry->replay_port,  24,  22, size, op);
entry             188 drivers/net/dsa/sja1105/sja1105_static_config.c 	struct sja1105_l2_forwarding_params_entry *entry = entry_ptr;
entry             191 drivers/net/dsa/sja1105/sja1105_static_config.c 	sja1105_packing(buf, &entry->max_dynp, 95, 93, size, op);
entry             193 drivers/net/dsa/sja1105/sja1105_static_config.c 		sja1105_packing(buf, &entry->part_spc[i],
entry             202 drivers/net/dsa/sja1105/sja1105_static_config.c 	struct sja1105_l2_forwarding_entry *entry = entry_ptr;
entry             205 drivers/net/dsa/sja1105/sja1105_static_config.c 	sja1105_packing(buf, &entry->bc_domain,  63, 59, size, op);
entry             206 drivers/net/dsa/sja1105/sja1105_static_config.c 	sja1105_packing(buf, &entry->reach_port, 58, 54, size, op);
entry             207 drivers/net/dsa/sja1105/sja1105_static_config.c 	sja1105_packing(buf, &entry->fl_domain,  53, 49, size, op);
entry             209 drivers/net/dsa/sja1105/sja1105_static_config.c 		sja1105_packing(buf, &entry->vlan_pmap[i],
entry             219 drivers/net/dsa/sja1105/sja1105_static_config.c 	struct sja1105_l2_lookup_params_entry *entry = entry_ptr;
entry             221 drivers/net/dsa/sja1105/sja1105_static_config.c 	sja1105_packing(buf, &entry->maxage,         31, 17, size, op);
entry             222 drivers/net/dsa/sja1105/sja1105_static_config.c 	sja1105_packing(buf, &entry->dyn_tbsz,       16, 14, size, op);
entry             223 drivers/net/dsa/sja1105/sja1105_static_config.c 	sja1105_packing(buf, &entry->poly,           13,  6, size, op);
entry             224 drivers/net/dsa/sja1105/sja1105_static_config.c 	sja1105_packing(buf, &entry->shared_learn,    5,  5, size, op);
entry             225 drivers/net/dsa/sja1105/sja1105_static_config.c 	sja1105_packing(buf, &entry->no_enf_hostprt,  4,  4, size, op);
entry             226 drivers/net/dsa/sja1105/sja1105_static_config.c 	sja1105_packing(buf, &entry->no_mgmt_learn,   3,  3, size, op);
entry             235 drivers/net/dsa/sja1105/sja1105_static_config.c 	struct sja1105_l2_lookup_params_entry *entry = entry_ptr;
entry             239 drivers/net/dsa/sja1105/sja1105_static_config.c 		sja1105_packing(buf, &entry->maxaddrp[i],
entry             241 drivers/net/dsa/sja1105/sja1105_static_config.c 	sja1105_packing(buf, &entry->maxage,         57,  43, size, op);
entry             242 drivers/net/dsa/sja1105/sja1105_static_config.c 	sja1105_packing(buf, &entry->start_dynspc,   42,  33, size, op);
entry             243 drivers/net/dsa/sja1105/sja1105_static_config.c 	sja1105_packing(buf, &entry->drpnolearn,     32,  28, size, op);
entry             244 drivers/net/dsa/sja1105/sja1105_static_config.c 	sja1105_packing(buf, &entry->shared_learn,   27,  27, size, op);
entry             245 drivers/net/dsa/sja1105/sja1105_static_config.c 	sja1105_packing(buf, &entry->no_enf_hostprt, 26,  26, size, op);
entry             246 drivers/net/dsa/sja1105/sja1105_static_config.c 	sja1105_packing(buf, &entry->no_mgmt_learn,  25,  25, size, op);
entry             247 drivers/net/dsa/sja1105/sja1105_static_config.c 	sja1105_packing(buf, &entry->use_static,     24,  24, size, op);
entry             248 drivers/net/dsa/sja1105/sja1105_static_config.c 	sja1105_packing(buf, &entry->owr_dyn,        23,  23, size, op);
entry             249 drivers/net/dsa/sja1105/sja1105_static_config.c 	sja1105_packing(buf, &entry->learn_once,     22,  22, size, op);
entry             257 drivers/net/dsa/sja1105/sja1105_static_config.c 	struct sja1105_l2_lookup_entry *entry = entry_ptr;
entry             259 drivers/net/dsa/sja1105/sja1105_static_config.c 	sja1105_packing(buf, &entry->vlanid,    95, 84, size, op);
entry             260 drivers/net/dsa/sja1105/sja1105_static_config.c 	sja1105_packing(buf, &entry->macaddr,   83, 36, size, op);
entry             261 drivers/net/dsa/sja1105/sja1105_static_config.c 	sja1105_packing(buf, &entry->destports, 35, 31, size, op);
entry             262 drivers/net/dsa/sja1105/sja1105_static_config.c 	sja1105_packing(buf, &entry->enfport,   30, 30, size, op);
entry             263 drivers/net/dsa/sja1105/sja1105_static_config.c 	sja1105_packing(buf, &entry->index,     29, 20, size, op);
entry             271 drivers/net/dsa/sja1105/sja1105_static_config.c 	struct sja1105_l2_lookup_entry *entry = entry_ptr;
entry             273 drivers/net/dsa/sja1105/sja1105_static_config.c 	if (entry->lockeds) {
entry             274 drivers/net/dsa/sja1105/sja1105_static_config.c 		sja1105_packing(buf, &entry->tsreg,    159, 159, size, op);
entry             275 drivers/net/dsa/sja1105/sja1105_static_config.c 		sja1105_packing(buf, &entry->mirrvlan, 158, 147, size, op);
entry             276 drivers/net/dsa/sja1105/sja1105_static_config.c 		sja1105_packing(buf, &entry->takets,   146, 146, size, op);
entry             277 drivers/net/dsa/sja1105/sja1105_static_config.c 		sja1105_packing(buf, &entry->mirr,     145, 145, size, op);
entry             278 drivers/net/dsa/sja1105/sja1105_static_config.c 		sja1105_packing(buf, &entry->retag,    144, 144, size, op);
entry             280 drivers/net/dsa/sja1105/sja1105_static_config.c 		sja1105_packing(buf, &entry->touched,  159, 159, size, op);
entry             281 drivers/net/dsa/sja1105/sja1105_static_config.c 		sja1105_packing(buf, &entry->age,      158, 144, size, op);
entry             283 drivers/net/dsa/sja1105/sja1105_static_config.c 	sja1105_packing(buf, &entry->mask_iotag,   143, 143, size, op);
entry             284 drivers/net/dsa/sja1105/sja1105_static_config.c 	sja1105_packing(buf, &entry->mask_vlanid,  142, 131, size, op);
entry             285 drivers/net/dsa/sja1105/sja1105_static_config.c 	sja1105_packing(buf, &entry->mask_macaddr, 130,  83, size, op);
entry             286 drivers/net/dsa/sja1105/sja1105_static_config.c 	sja1105_packing(buf, &entry->iotag,         82,  82, size, op);
entry             287 drivers/net/dsa/sja1105/sja1105_static_config.c 	sja1105_packing(buf, &entry->vlanid,        81,  70, size, op);
entry             288 drivers/net/dsa/sja1105/sja1105_static_config.c 	sja1105_packing(buf, &entry->macaddr,       69,  22, size, op);
entry             289 drivers/net/dsa/sja1105/sja1105_static_config.c 	sja1105_packing(buf, &entry->destports,     21,  17, size, op);
entry             290 drivers/net/dsa/sja1105/sja1105_static_config.c 	sja1105_packing(buf, &entry->enfport,       16,  16, size, op);
entry             291 drivers/net/dsa/sja1105/sja1105_static_config.c 	sja1105_packing(buf, &entry->index,         15,   6, size, op);
entry             299 drivers/net/dsa/sja1105/sja1105_static_config.c 	struct sja1105_l2_policing_entry *entry = entry_ptr;
entry             301 drivers/net/dsa/sja1105/sja1105_static_config.c 	sja1105_packing(buf, &entry->sharindx,  63, 58, size, op);
entry             302 drivers/net/dsa/sja1105/sja1105_static_config.c 	sja1105_packing(buf, &entry->smax,      57, 42, size, op);
entry             303 drivers/net/dsa/sja1105/sja1105_static_config.c 	sja1105_packing(buf, &entry->rate,      41, 26, size, op);
entry             304 drivers/net/dsa/sja1105/sja1105_static_config.c 	sja1105_packing(buf, &entry->maxlen,    25, 15, size, op);
entry             305 drivers/net/dsa/sja1105/sja1105_static_config.c 	sja1105_packing(buf, &entry->partition, 14, 12, size, op);
entry             313 drivers/net/dsa/sja1105/sja1105_static_config.c 	struct sja1105_mac_config_entry *entry = entry_ptr;
entry             317 drivers/net/dsa/sja1105/sja1105_static_config.c 		sja1105_packing(buf, &entry->enabled[i],
entry             319 drivers/net/dsa/sja1105/sja1105_static_config.c 		sja1105_packing(buf, &entry->base[i],
entry             321 drivers/net/dsa/sja1105/sja1105_static_config.c 		sja1105_packing(buf, &entry->top[i],
entry             324 drivers/net/dsa/sja1105/sja1105_static_config.c 	sja1105_packing(buf, &entry->ifg,       71, 67, size, op);
entry             325 drivers/net/dsa/sja1105/sja1105_static_config.c 	sja1105_packing(buf, &entry->speed,     66, 65, size, op);
entry             326 drivers/net/dsa/sja1105/sja1105_static_config.c 	sja1105_packing(buf, &entry->tp_delin,  64, 49, size, op);
entry             327 drivers/net/dsa/sja1105/sja1105_static_config.c 	sja1105_packing(buf, &entry->tp_delout, 48, 33, size, op);
entry             328 drivers/net/dsa/sja1105/sja1105_static_config.c 	sja1105_packing(buf, &entry->maxage,    32, 25, size, op);
entry             329 drivers/net/dsa/sja1105/sja1105_static_config.c 	sja1105_packing(buf, &entry->vlanprio,  24, 22, size, op);
entry             330 drivers/net/dsa/sja1105/sja1105_static_config.c 	sja1105_packing(buf, &entry->vlanid,    21, 10, size, op);
entry             331 drivers/net/dsa/sja1105/sja1105_static_config.c 	sja1105_packing(buf, &entry->ing_mirr,   9,  9, size, op);
entry             332 drivers/net/dsa/sja1105/sja1105_static_config.c 	sja1105_packing(buf, &entry->egr_mirr,   8,  8, size, op);
entry             333 drivers/net/dsa/sja1105/sja1105_static_config.c 	sja1105_packing(buf, &entry->drpnona664, 7,  7, size, op);
entry             334 drivers/net/dsa/sja1105/sja1105_static_config.c 	sja1105_packing(buf, &entry->drpdtag,    6,  6, size, op);
entry             335 drivers/net/dsa/sja1105/sja1105_static_config.c 	sja1105_packing(buf, &entry->drpuntag,   5,  5, size, op);
entry             336 drivers/net/dsa/sja1105/sja1105_static_config.c 	sja1105_packing(buf, &entry->retag,      4,  4, size, op);
entry             337 drivers/net/dsa/sja1105/sja1105_static_config.c 	sja1105_packing(buf, &entry->dyn_learn,  3,  3, size, op);
entry             338 drivers/net/dsa/sja1105/sja1105_static_config.c 	sja1105_packing(buf, &entry->egress,     2,  2, size, op);
entry             339 drivers/net/dsa/sja1105/sja1105_static_config.c 	sja1105_packing(buf, &entry->ingress,    1,  1, size, op);
entry             347 drivers/net/dsa/sja1105/sja1105_static_config.c 	struct sja1105_mac_config_entry *entry = entry_ptr;
entry             351 drivers/net/dsa/sja1105/sja1105_static_config.c 		sja1105_packing(buf, &entry->enabled[i],
entry             353 drivers/net/dsa/sja1105/sja1105_static_config.c 		sja1105_packing(buf, &entry->base[i],
entry             355 drivers/net/dsa/sja1105/sja1105_static_config.c 		sja1105_packing(buf, &entry->top[i],
entry             358 drivers/net/dsa/sja1105/sja1105_static_config.c 	sja1105_packing(buf, &entry->ifg,       103, 99, size, op);
entry             359 drivers/net/dsa/sja1105/sja1105_static_config.c 	sja1105_packing(buf, &entry->speed,      98, 97, size, op);
entry             360 drivers/net/dsa/sja1105/sja1105_static_config.c 	sja1105_packing(buf, &entry->tp_delin,   96, 81, size, op);
entry             361 drivers/net/dsa/sja1105/sja1105_static_config.c 	sja1105_packing(buf, &entry->tp_delout,  80, 65, size, op);
entry             362 drivers/net/dsa/sja1105/sja1105_static_config.c 	sja1105_packing(buf, &entry->maxage,     64, 57, size, op);
entry             363 drivers/net/dsa/sja1105/sja1105_static_config.c 	sja1105_packing(buf, &entry->vlanprio,   56, 54, size, op);
entry             364 drivers/net/dsa/sja1105/sja1105_static_config.c 	sja1105_packing(buf, &entry->vlanid,     53, 42, size, op);
entry             365 drivers/net/dsa/sja1105/sja1105_static_config.c 	sja1105_packing(buf, &entry->ing_mirr,   41, 41, size, op);
entry             366 drivers/net/dsa/sja1105/sja1105_static_config.c 	sja1105_packing(buf, &entry->egr_mirr,   40, 40, size, op);
entry             367 drivers/net/dsa/sja1105/sja1105_static_config.c 	sja1105_packing(buf, &entry->drpnona664, 39, 39, size, op);
entry             368 drivers/net/dsa/sja1105/sja1105_static_config.c 	sja1105_packing(buf, &entry->drpdtag,    38, 38, size, op);
entry             369 drivers/net/dsa/sja1105/sja1105_static_config.c 	sja1105_packing(buf, &entry->drpuntag,   35, 35, size, op);
entry             370 drivers/net/dsa/sja1105/sja1105_static_config.c 	sja1105_packing(buf, &entry->retag,      34, 34, size, op);
entry             371 drivers/net/dsa/sja1105/sja1105_static_config.c 	sja1105_packing(buf, &entry->dyn_learn,  33, 33, size, op);
entry             372 drivers/net/dsa/sja1105/sja1105_static_config.c 	sja1105_packing(buf, &entry->egress,     32, 32, size, op);
entry             373 drivers/net/dsa/sja1105/sja1105_static_config.c 	sja1105_packing(buf, &entry->ingress,    31, 31, size, op);
entry             381 drivers/net/dsa/sja1105/sja1105_static_config.c 	struct sja1105_schedule_entry_points_params_entry *entry = entry_ptr;
entry             384 drivers/net/dsa/sja1105/sja1105_static_config.c 	sja1105_packing(buf, &entry->clksrc,    31, 30, size, op);
entry             385 drivers/net/dsa/sja1105/sja1105_static_config.c 	sja1105_packing(buf, &entry->actsubsch, 29, 27, size, op);
entry             393 drivers/net/dsa/sja1105/sja1105_static_config.c 	struct sja1105_schedule_entry_points_entry *entry = entry_ptr;
entry             396 drivers/net/dsa/sja1105/sja1105_static_config.c 	sja1105_packing(buf, &entry->subschindx, 31, 29, size, op);
entry             397 drivers/net/dsa/sja1105/sja1105_static_config.c 	sja1105_packing(buf, &entry->delta,      28, 11, size, op);
entry             398 drivers/net/dsa/sja1105/sja1105_static_config.c 	sja1105_packing(buf, &entry->address,    10, 1,  size, op);
entry             406 drivers/net/dsa/sja1105/sja1105_static_config.c 	struct sja1105_schedule_params_entry *entry = entry_ptr;
entry             410 drivers/net/dsa/sja1105/sja1105_static_config.c 		sja1105_packing(buf, &entry->subscheind[i],
entry             419 drivers/net/dsa/sja1105/sja1105_static_config.c 	struct sja1105_schedule_entry *entry = entry_ptr;
entry             421 drivers/net/dsa/sja1105/sja1105_static_config.c 	sja1105_packing(buf, &entry->winstindex,  63, 54, size, op);
entry             422 drivers/net/dsa/sja1105/sja1105_static_config.c 	sja1105_packing(buf, &entry->winend,      53, 53, size, op);
entry             423 drivers/net/dsa/sja1105/sja1105_static_config.c 	sja1105_packing(buf, &entry->winst,       52, 52, size, op);
entry             424 drivers/net/dsa/sja1105/sja1105_static_config.c 	sja1105_packing(buf, &entry->destports,   51, 47, size, op);
entry             425 drivers/net/dsa/sja1105/sja1105_static_config.c 	sja1105_packing(buf, &entry->setvalid,    46, 46, size, op);
entry             426 drivers/net/dsa/sja1105/sja1105_static_config.c 	sja1105_packing(buf, &entry->txen,        45, 45, size, op);
entry             427 drivers/net/dsa/sja1105/sja1105_static_config.c 	sja1105_packing(buf, &entry->resmedia_en, 44, 44, size, op);
entry             428 drivers/net/dsa/sja1105/sja1105_static_config.c 	sja1105_packing(buf, &entry->resmedia,    43, 36, size, op);
entry             429 drivers/net/dsa/sja1105/sja1105_static_config.c 	sja1105_packing(buf, &entry->vlindex,     35, 26, size, op);
entry             430 drivers/net/dsa/sja1105/sja1105_static_config.c 	sja1105_packing(buf, &entry->delta,       25, 8,  size, op);
entry             438 drivers/net/dsa/sja1105/sja1105_static_config.c 	struct sja1105_vlan_lookup_entry *entry = entry_ptr;
entry             440 drivers/net/dsa/sja1105/sja1105_static_config.c 	sja1105_packing(buf, &entry->ving_mirr,  63, 59, size, op);
entry             441 drivers/net/dsa/sja1105/sja1105_static_config.c 	sja1105_packing(buf, &entry->vegr_mirr,  58, 54, size, op);
entry             442 drivers/net/dsa/sja1105/sja1105_static_config.c 	sja1105_packing(buf, &entry->vmemb_port, 53, 49, size, op);
entry             443 drivers/net/dsa/sja1105/sja1105_static_config.c 	sja1105_packing(buf, &entry->vlan_bc,    48, 44, size, op);
entry             444 drivers/net/dsa/sja1105/sja1105_static_config.c 	sja1105_packing(buf, &entry->tag_port,   43, 39, size, op);
entry             445 drivers/net/dsa/sja1105/sja1105_static_config.c 	sja1105_packing(buf, &entry->vlanid,     38, 27, size, op);
entry             453 drivers/net/dsa/sja1105/sja1105_static_config.c 	struct sja1105_xmii_params_entry *entry = entry_ptr;
entry             457 drivers/net/dsa/sja1105/sja1105_static_config.c 		sja1105_packing(buf, &entry->xmii_mode[i],
entry             459 drivers/net/dsa/sja1105/sja1105_static_config.c 		sja1105_packing(buf, &entry->phy_mac[i],
entry             469 drivers/net/dsa/sja1105/sja1105_static_config.c 	struct sja1105_table_header *entry = entry_ptr;
entry             471 drivers/net/dsa/sja1105/sja1105_static_config.c 	sja1105_packing(buf, &entry->block_id, 31, 24, size, op);
entry             472 drivers/net/dsa/sja1105/sja1105_static_config.c 	sja1105_packing(buf, &entry->len,      55, 32, size, op);
entry             473 drivers/net/dsa/sja1105/sja1105_static_config.c 	sja1105_packing(buf, &entry->crc,      95, 64, size, op);
entry            1013 drivers/net/ethernet/3com/3c515.c 		int entry = vp->cur_tx % TX_RING_SIZE;
entry            1028 drivers/net/ethernet/3com/3c515.c 		vp->tx_skbuff[entry] = skb;
entry            1029 drivers/net/ethernet/3com/3c515.c 		vp->tx_ring[entry].next = 0;
entry            1030 drivers/net/ethernet/3com/3c515.c 		vp->tx_ring[entry].addr = isa_virt_to_bus(skb->data);
entry            1031 drivers/net/ethernet/3com/3c515.c 		vp->tx_ring[entry].length = skb->len | 0x80000000;
entry            1032 drivers/net/ethernet/3com/3c515.c 		vp->tx_ring[entry].status = skb->len | 0x80000000;
entry            1041 drivers/net/ethernet/3com/3c515.c 			prev_entry->next = isa_virt_to_bus(&vp->tx_ring[entry]);
entry            1043 drivers/net/ethernet/3com/3c515.c 			outl(isa_virt_to_bus(&vp->tx_ring[entry]),
entry            1176 drivers/net/ethernet/3com/3c515.c 				int entry = dirty_tx % TX_RING_SIZE;
entry            1177 drivers/net/ethernet/3com/3c515.c 				if (inl(ioaddr + DownListPtr) == isa_virt_to_bus(&lp->tx_ring[entry]))
entry            1179 drivers/net/ethernet/3com/3c515.c 				if (lp->tx_skbuff[entry]) {
entry            1180 drivers/net/ethernet/3com/3c515.c 					dev_consume_skb_irq(lp->tx_skbuff[entry]);
entry            1181 drivers/net/ethernet/3com/3c515.c 					lp->tx_skbuff[entry] = NULL;
entry            1333 drivers/net/ethernet/3com/3c515.c 	int entry = vp->cur_rx % RX_RING_SIZE;
entry            1340 drivers/net/ethernet/3com/3c515.c 	while ((rx_status = vp->rx_ring[entry].status) & RxDComplete) {
entry            1374 drivers/net/ethernet/3com/3c515.c 					     isa_bus_to_virt(vp->rx_ring[entry].addr),
entry            1380 drivers/net/ethernet/3com/3c515.c 				skb = vp->rx_skbuff[entry];
entry            1381 drivers/net/ethernet/3com/3c515.c 				vp->rx_skbuff[entry] = NULL;
entry            1384 drivers/net/ethernet/3com/3c515.c 				if (isa_bus_to_virt(vp->rx_ring[entry].addr) != temp)
entry            1387 drivers/net/ethernet/3com/3c515.c 						isa_bus_to_virt(vp->rx_ring[entry].addr),
entry            1395 drivers/net/ethernet/3com/3c515.c 		entry = (++vp->cur_rx) % RX_RING_SIZE;
entry            1400 drivers/net/ethernet/3com/3c515.c 		entry = vp->dirty_rx % RX_RING_SIZE;
entry            1401 drivers/net/ethernet/3com/3c515.c 		if (vp->rx_skbuff[entry] == NULL) {
entry            1406 drivers/net/ethernet/3com/3c515.c 			vp->rx_ring[entry].addr = isa_virt_to_bus(skb->data);
entry            1407 drivers/net/ethernet/3com/3c515.c 			vp->rx_skbuff[entry] = skb;
entry            1409 drivers/net/ethernet/3com/3c515.c 		vp->rx_ring[entry].status = 0;	/* Clear complete bit. */
entry            2114 drivers/net/ethernet/3com/3c59x.c 	int entry = vp->cur_tx % TX_RING_SIZE;
entry            2143 drivers/net/ethernet/3com/3c59x.c 	vp->tx_skbuff[entry] = skb;
entry            2145 drivers/net/ethernet/3com/3c59x.c 	vp->tx_ring[entry].next = 0;
entry            2148 drivers/net/ethernet/3com/3c59x.c 			vp->tx_ring[entry].status = cpu_to_le32(skb->len | TxIntrUploaded);
entry            2150 drivers/net/ethernet/3com/3c59x.c 			vp->tx_ring[entry].status = cpu_to_le32(skb->len | TxIntrUploaded | AddTCPChksum | AddUDPChksum);
entry            2158 drivers/net/ethernet/3com/3c59x.c 		vp->tx_ring[entry].frag[0].addr = cpu_to_le32(dma_addr);
entry            2159 drivers/net/ethernet/3com/3c59x.c 		vp->tx_ring[entry].frag[0].length = cpu_to_le32(skb->len | LAST_FRAG);
entry            2168 drivers/net/ethernet/3com/3c59x.c 		vp->tx_ring[entry].frag[0].addr = cpu_to_le32(dma_addr);
entry            2169 drivers/net/ethernet/3com/3c59x.c 		vp->tx_ring[entry].frag[0].length = cpu_to_le32(skb_headlen(skb));
entry            2181 drivers/net/ethernet/3com/3c59x.c 						       le32_to_cpu(vp->tx_ring[entry].frag[i+1].addr),
entry            2182 drivers/net/ethernet/3com/3c59x.c 						       le32_to_cpu(vp->tx_ring[entry].frag[i+1].length),
entry            2186 drivers/net/ethernet/3com/3c59x.c 						 le32_to_cpu(vp->tx_ring[entry].frag[0].addr),
entry            2187 drivers/net/ethernet/3com/3c59x.c 						 le32_to_cpu(vp->tx_ring[entry].frag[0].length),
entry            2193 drivers/net/ethernet/3com/3c59x.c 			vp->tx_ring[entry].frag[i+1].addr =
entry            2197 drivers/net/ethernet/3com/3c59x.c 					vp->tx_ring[entry].frag[i+1].length = cpu_to_le32(skb_frag_size(frag)|LAST_FRAG);
entry            2199 drivers/net/ethernet/3com/3c59x.c 					vp->tx_ring[entry].frag[i+1].length = cpu_to_le32(skb_frag_size(frag));
entry            2206 drivers/net/ethernet/3com/3c59x.c 	vp->tx_ring[entry].addr = cpu_to_le32(dma_addr);
entry            2207 drivers/net/ethernet/3com/3c59x.c 	vp->tx_ring[entry].length = cpu_to_le32(skb->len | LAST_FRAG);
entry            2208 drivers/net/ethernet/3com/3c59x.c 	vp->tx_ring[entry].status = cpu_to_le32(skb->len | TxIntrUploaded);
entry            2214 drivers/net/ethernet/3com/3c59x.c 	prev_entry->next = cpu_to_le32(vp->tx_ring_dma + entry * sizeof(struct boom_tx_desc));
entry            2216 drivers/net/ethernet/3com/3c59x.c 		iowrite32(vp->tx_ring_dma + entry * sizeof(struct boom_tx_desc), ioaddr + DownListPtr);
entry            2420 drivers/net/ethernet/3com/3c59x.c 				int entry = dirty_tx % TX_RING_SIZE;
entry            2423 drivers/net/ethernet/3com/3c59x.c 					vp->tx_ring_dma + entry * sizeof(struct boom_tx_desc))
entry            2426 drivers/net/ethernet/3com/3c59x.c 				if ((vp->tx_ring[entry].status & DN_COMPLETE) == 0)
entry            2430 drivers/net/ethernet/3com/3c59x.c 				if (vp->tx_skbuff[entry]) {
entry            2431 drivers/net/ethernet/3com/3c59x.c 					struct sk_buff *skb = vp->tx_skbuff[entry];
entry            2435 drivers/net/ethernet/3com/3c59x.c 							le32_to_cpu(vp->tx_ring[entry].frag[0].addr),
entry            2436 drivers/net/ethernet/3com/3c59x.c 							le32_to_cpu(vp->tx_ring[entry].frag[0].length)&0xFFF,
entry            2441 drivers/net/ethernet/3com/3c59x.c 											 le32_to_cpu(vp->tx_ring[entry].frag[i].addr),
entry            2442 drivers/net/ethernet/3com/3c59x.c 											 le32_to_cpu(vp->tx_ring[entry].frag[i].length)&0xFFF,
entry            2446 drivers/net/ethernet/3com/3c59x.c 						le32_to_cpu(vp->tx_ring[entry].addr), skb->len, DMA_TO_DEVICE);
entry            2451 drivers/net/ethernet/3com/3c59x.c 					vp->tx_skbuff[entry] = NULL;
entry            2592 drivers/net/ethernet/3com/3c59x.c 	int entry = vp->cur_rx % RX_RING_SIZE;
entry            2600 drivers/net/ethernet/3com/3c59x.c 	while ((rx_status = le32_to_cpu(vp->rx_ring[entry].status)) & RxDComplete){
entry            2618 drivers/net/ethernet/3com/3c59x.c 			dma_addr_t dma = le32_to_cpu(vp->rx_ring[entry].addr);
entry            2631 drivers/net/ethernet/3com/3c59x.c 				skb_put_data(skb, vp->rx_skbuff[entry]->data,
entry            2654 drivers/net/ethernet/3com/3c59x.c 				skb = vp->rx_skbuff[entry];
entry            2655 drivers/net/ethernet/3com/3c59x.c 				vp->rx_skbuff[entry] = newskb;
entry            2656 drivers/net/ethernet/3com/3c59x.c 				vp->rx_ring[entry].addr = cpu_to_le32(newdma);
entry            2676 drivers/net/ethernet/3com/3c59x.c 		vp->rx_ring[entry].status = 0;	/* Clear complete bit. */
entry            2678 drivers/net/ethernet/3com/3c59x.c 		entry = (++vp->cur_rx) % RX_RING_SIZE;
entry            1190 drivers/net/ethernet/adaptec/starfire.c 	unsigned int entry;
entry            1212 drivers/net/ethernet/adaptec/starfire.c 	entry = np->cur_tx % TX_RING_SIZE;
entry            1218 drivers/net/ethernet/adaptec/starfire.c 			np->tx_info[entry].skb = skb;
entry            1220 drivers/net/ethernet/adaptec/starfire.c 			if (entry >= TX_RING_SIZE - skb_num_frags(skb)) {
entry            1234 drivers/net/ethernet/adaptec/starfire.c 			np->tx_info[entry].mapping =
entry            1239 drivers/net/ethernet/adaptec/starfire.c 			np->tx_info[entry].mapping =
entry            1246 drivers/net/ethernet/adaptec/starfire.c 					  np->tx_info[entry].mapping)) {
entry            1251 drivers/net/ethernet/adaptec/starfire.c 		np->tx_ring[entry].addr = cpu_to_dma(np->tx_info[entry].mapping);
entry            1252 drivers/net/ethernet/adaptec/starfire.c 		np->tx_ring[entry].status = cpu_to_le32(status);
entry            1256 drivers/net/ethernet/adaptec/starfire.c 			       entry, status);
entry            1258 drivers/net/ethernet/adaptec/starfire.c 			np->tx_info[entry].used_slots = TX_RING_SIZE - entry;
entry            1259 drivers/net/ethernet/adaptec/starfire.c 			np->cur_tx += np->tx_info[entry].used_slots;
entry            1260 drivers/net/ethernet/adaptec/starfire.c 			entry = 0;
entry            1262 drivers/net/ethernet/adaptec/starfire.c 			np->tx_info[entry].used_slots = 1;
entry            1263 drivers/net/ethernet/adaptec/starfire.c 			np->cur_tx += np->tx_info[entry].used_slots;
entry            1264 drivers/net/ethernet/adaptec/starfire.c 			entry++;
entry            1277 drivers/net/ethernet/adaptec/starfire.c 	writel(entry * (sizeof(starfire_tx_desc) / 8), np->base + TxProducerIdx);
entry            1286 drivers/net/ethernet/adaptec/starfire.c 	entry = prev_tx % TX_RING_SIZE;
entry            1287 drivers/net/ethernet/adaptec/starfire.c 	np->tx_info[entry].skb = NULL;
entry            1290 drivers/net/ethernet/adaptec/starfire.c 				 np->tx_info[entry].mapping,
entry            1293 drivers/net/ethernet/adaptec/starfire.c 		np->tx_info[entry].mapping = 0;
entry            1294 drivers/net/ethernet/adaptec/starfire.c 		entry = (entry + np->tx_info[entry].used_slots) % TX_RING_SIZE;
entry            1297 drivers/net/ethernet/adaptec/starfire.c 					 np->tx_info[entry].mapping,
entry            1301 drivers/net/ethernet/adaptec/starfire.c 			entry++;
entry            1371 drivers/net/ethernet/adaptec/starfire.c 				u16 entry = (tx_status & 0x7fff) / sizeof(starfire_tx_desc);
entry            1372 drivers/net/ethernet/adaptec/starfire.c 				struct sk_buff *skb = np->tx_info[entry].skb;
entry            1373 drivers/net/ethernet/adaptec/starfire.c 				np->tx_info[entry].skb = NULL;
entry            1375 drivers/net/ethernet/adaptec/starfire.c 						 np->tx_info[entry].mapping,
entry            1378 drivers/net/ethernet/adaptec/starfire.c 				np->tx_info[entry].mapping = 0;
entry            1379 drivers/net/ethernet/adaptec/starfire.c 				np->dirty_tx += np->tx_info[entry].used_slots;
entry            1380 drivers/net/ethernet/adaptec/starfire.c 				entry = (entry + np->tx_info[entry].used_slots) % TX_RING_SIZE;
entry            1385 drivers/net/ethernet/adaptec/starfire.c 								 np->tx_info[entry].mapping,
entry            1389 drivers/net/ethernet/adaptec/starfire.c 						entry++;
entry            1448 drivers/net/ethernet/adaptec/starfire.c 		int entry;
entry            1470 drivers/net/ethernet/adaptec/starfire.c 		entry = (desc_status >> 16) & 0x7ff;
entry            1480 drivers/net/ethernet/adaptec/starfire.c 						    np->rx_info[entry].mapping,
entry            1482 drivers/net/ethernet/adaptec/starfire.c 			skb_copy_to_linear_data(skb, np->rx_info[entry].skb->data, pkt_len);
entry            1484 drivers/net/ethernet/adaptec/starfire.c 						       np->rx_info[entry].mapping,
entry            1488 drivers/net/ethernet/adaptec/starfire.c 			pci_unmap_single(np->pci_dev, np->rx_info[entry].mapping, np->rx_buf_sz, PCI_DMA_FROMDEVICE);
entry            1489 drivers/net/ethernet/adaptec/starfire.c 			skb = np->rx_info[entry].skb;
entry            1491 drivers/net/ethernet/adaptec/starfire.c 			np->rx_info[entry].skb = NULL;
entry            1492 drivers/net/ethernet/adaptec/starfire.c 			np->rx_info[entry].mapping = 0;
entry            1595 drivers/net/ethernet/adaptec/starfire.c 	int entry = -1;
entry            1599 drivers/net/ethernet/adaptec/starfire.c 		entry = np->dirty_rx % RX_RING_SIZE;
entry            1600 drivers/net/ethernet/adaptec/starfire.c 		if (np->rx_info[entry].skb == NULL) {
entry            1602 drivers/net/ethernet/adaptec/starfire.c 			np->rx_info[entry].skb = skb;
entry            1605 drivers/net/ethernet/adaptec/starfire.c 			np->rx_info[entry].mapping =
entry            1608 drivers/net/ethernet/adaptec/starfire.c 						np->rx_info[entry].mapping)) {
entry            1610 drivers/net/ethernet/adaptec/starfire.c 				np->rx_info[entry].skb = NULL;
entry            1613 drivers/net/ethernet/adaptec/starfire.c 			np->rx_ring[entry].rxaddr =
entry            1614 drivers/net/ethernet/adaptec/starfire.c 				cpu_to_dma(np->rx_info[entry].mapping | RxDescValid);
entry            1616 drivers/net/ethernet/adaptec/starfire.c 		if (entry == RX_RING_SIZE - 1)
entry            1617 drivers/net/ethernet/adaptec/starfire.c 			np->rx_ring[entry].rxaddr |= cpu_to_dma(RxDescEndRing);
entry            1619 drivers/net/ethernet/adaptec/starfire.c 	if (entry >= 0)
entry            1620 drivers/net/ethernet/adaptec/starfire.c 		writew(entry, np->base + RxDescQIdx);
entry            1545 drivers/net/ethernet/agere/et131x.c 	u32 entry;
entry            1597 drivers/net/ethernet/agere/et131x.c 		for (entry = 0; entry < fbr->num_entries; entry++) {
entry            1598 drivers/net/ethernet/agere/et131x.c 			fbr_entry->addr_hi = fbr->bus_high[entry];
entry            1599 drivers/net/ethernet/agere/et131x.c 			fbr_entry->addr_lo = fbr->bus_low[entry];
entry            1600 drivers/net/ethernet/agere/et131x.c 			fbr_entry->word2 = entry;
entry             423 drivers/net/ethernet/altera/altera_sgdma.c #define list_remove_head(list, entry, type, member)			\
entry             425 drivers/net/ethernet/altera/altera_sgdma.c 		entry = NULL;						\
entry             427 drivers/net/ethernet/altera/altera_sgdma.c 			entry = list_entry((list)->next, type, member);	\
entry             428 drivers/net/ethernet/altera/altera_sgdma.c 			list_del_init(&entry->member);			\
entry             432 drivers/net/ethernet/altera/altera_sgdma.c #define list_peek_head(list, entry, type, member)			\
entry             434 drivers/net/ethernet/altera/altera_sgdma.c 		entry = NULL;						\
entry             436 drivers/net/ethernet/altera/altera_sgdma.c 			entry = list_entry((list)->next, type, member);	\
entry             335 drivers/net/ethernet/altera/altera_tse_main.c 	unsigned int entry;
entry             340 drivers/net/ethernet/altera/altera_tse_main.c 		entry = priv->rx_prod % rxsize;
entry             341 drivers/net/ethernet/altera/altera_tse_main.c 		if (likely(priv->rx_ring[entry].skb == NULL)) {
entry             342 drivers/net/ethernet/altera/altera_tse_main.c 			ret = tse_init_rx_buffer(priv, &priv->rx_ring[entry],
entry             346 drivers/net/ethernet/altera/altera_tse_main.c 			priv->dmaops->add_rx_desc(priv, &priv->rx_ring[entry]);
entry             373 drivers/net/ethernet/altera/altera_tse_main.c 	unsigned int entry = priv->rx_cons % priv->rx_ring_size;
entry             402 drivers/net/ethernet/altera/altera_tse_main.c 		skb = priv->rx_ring[entry].skb;
entry             410 drivers/net/ethernet/altera/altera_tse_main.c 		priv->rx_ring[entry].skb = NULL;
entry             414 drivers/net/ethernet/altera/altera_tse_main.c 		dma_unmap_single(priv->device, priv->rx_ring[entry].dma_addr,
entry             415 drivers/net/ethernet/altera/altera_tse_main.c 				 priv->rx_ring[entry].len, DMA_FROM_DEVICE);
entry             434 drivers/net/ethernet/altera/altera_tse_main.c 		entry = next_entry;
entry             448 drivers/net/ethernet/altera/altera_tse_main.c 	unsigned int entry;
entry             458 drivers/net/ethernet/altera/altera_tse_main.c 		entry = priv->tx_cons % txsize;
entry             459 drivers/net/ethernet/altera/altera_tse_main.c 		tx_buff = &priv->tx_ring[entry];
entry             561 drivers/net/ethernet/altera/altera_tse_main.c 	unsigned int entry;
entry             583 drivers/net/ethernet/altera/altera_tse_main.c 	entry = priv->tx_prod % txsize;
entry             584 drivers/net/ethernet/altera/altera_tse_main.c 	buffer = &priv->tx_ring[entry];
entry             543 drivers/net/ethernet/amd/7990.c 	int entry, skblen, len;
entry             569 drivers/net/ethernet/amd/7990.c 	entry = lp->tx_new & lp->tx_ring_mod_mask;
entry             570 drivers/net/ethernet/amd/7990.c 	ib->btx_ring[entry].length = (-len) | 0xf000;
entry             571 drivers/net/ethernet/amd/7990.c 	ib->btx_ring[entry].misc = 0;
entry             574 drivers/net/ethernet/amd/7990.c 		memset((void *)&ib->tx_buf[entry][0], 0, ETH_ZLEN);
entry             575 drivers/net/ethernet/amd/7990.c 	skb_copy_from_linear_data(skb, (void *)&ib->tx_buf[entry][0], skblen);
entry             578 drivers/net/ethernet/amd/7990.c 	ib->btx_ring[entry].tmd1_bits = (LE_T1_POK|LE_T1_OWN);
entry             541 drivers/net/ethernet/amd/a2065.c 	int entry, skblen;
entry             559 drivers/net/ethernet/amd/a2065.c 	entry = lp->tx_new & lp->tx_ring_mod_mask;
entry             560 drivers/net/ethernet/amd/a2065.c 	ib->btx_ring[entry].length = (-skblen) | 0xf000;
entry             561 drivers/net/ethernet/amd/a2065.c 	ib->btx_ring[entry].misc = 0;
entry             563 drivers/net/ethernet/amd/a2065.c 	skb_copy_from_linear_data(skb, (void *)&ib->tx_buf[entry][0], skblen);
entry             566 drivers/net/ethernet/amd/a2065.c 	ib->btx_ring[entry].tmd1_bits = (LE_T1_POK | LE_T1_OWN);
entry             164 drivers/net/ethernet/amd/ariadne.c 	int entry = priv->cur_rx % RX_RING_SIZE;
entry             168 drivers/net/ethernet/amd/ariadne.c 	while (!(lowb(priv->rx_ring[entry]->RMD1) & RF_OWN)) {
entry             169 drivers/net/ethernet/amd/ariadne.c 		int status = lowb(priv->rx_ring[entry]->RMD1);
entry             189 drivers/net/ethernet/amd/ariadne.c 			priv->rx_ring[entry]->RMD1 &= 0xff00 | RF_STP | RF_ENP;
entry             192 drivers/net/ethernet/amd/ariadne.c 			short pkt_len = swapw(priv->rx_ring[entry]->RMD3);
entry             198 drivers/net/ethernet/amd/ariadne.c 					if (lowb(priv->rx_ring[(entry + i) % RX_RING_SIZE]->RMD1) & RF_OWN)
entry             203 drivers/net/ethernet/amd/ariadne.c 					priv->rx_ring[entry]->RMD1 |= RF_OWN;
entry             213 drivers/net/ethernet/amd/ariadne.c 						(const void *)priv->rx_buff[entry],
entry             226 drivers/net/ethernet/amd/ariadne.c 		priv->rx_ring[entry]->RMD1 |= RF_OWN;
entry             227 drivers/net/ethernet/amd/ariadne.c 		entry = (++priv->cur_rx) % RX_RING_SIZE;
entry             309 drivers/net/ethernet/amd/ariadne.c 				int entry = dirty_tx % TX_RING_SIZE;
entry             310 drivers/net/ethernet/amd/ariadne.c 				int status = lowb(priv->tx_ring[entry]->TMD1);
entry             315 drivers/net/ethernet/amd/ariadne.c 				priv->tx_ring[entry]->TMD1 &= 0xff00;
entry             319 drivers/net/ethernet/amd/ariadne.c 					int err_status = priv->tx_ring[entry]->TMD3;
entry             548 drivers/net/ethernet/amd/ariadne.c 	int entry;
entry             576 drivers/net/ethernet/amd/ariadne.c 	entry = priv->cur_tx % TX_RING_SIZE;
entry             581 drivers/net/ethernet/amd/ariadne.c 	priv->tx_ring[entry]->TMD2 = swapw((u_short)-skb->len);
entry             582 drivers/net/ethernet/amd/ariadne.c 	priv->tx_ring[entry]->TMD3 = 0x0000;
entry             583 drivers/net/ethernet/amd/ariadne.c 	memcpyw(priv->tx_buff[entry], (u_short *)skb->data, len);
entry             587 drivers/net/ethernet/amd/ariadne.c 		       (void *)priv->tx_buff[entry],
entry             591 drivers/net/ethernet/amd/ariadne.c 	priv->tx_ring[entry]->TMD1 = (priv->tx_ring[entry]->TMD1 & 0xff00)
entry             612 drivers/net/ethernet/amd/ariadne.c 	if (lowb(priv->tx_ring[(entry + 1) % TX_RING_SIZE]->TMD1) != 0) {
entry             778 drivers/net/ethernet/amd/atarilance.c 	int entry, len;
entry             815 drivers/net/ethernet/amd/atarilance.c 	entry = lp->cur_tx & TX_RING_MOD_MASK;
entry             816 drivers/net/ethernet/amd/atarilance.c 	head  = &(MEM->tx_head[entry]);
entry             838 drivers/net/ethernet/amd/atarilance.c 	if ((MEM->tx_head[(entry+1) & TX_RING_MOD_MASK].flag & TMD1_OWN) ==
entry             886 drivers/net/ethernet/amd/atarilance.c 				int entry = dirty_tx & TX_RING_MOD_MASK;
entry             887 drivers/net/ethernet/amd/atarilance.c 				int status = MEM->tx_head[entry].flag;
entry             892 drivers/net/ethernet/amd/atarilance.c 				MEM->tx_head[entry].flag = 0;
entry             896 drivers/net/ethernet/amd/atarilance.c 					int err_status = MEM->tx_head[entry].misc;
entry             965 drivers/net/ethernet/amd/atarilance.c 	int entry = lp->cur_rx & RX_RING_MOD_MASK;
entry             969 drivers/net/ethernet/amd/atarilance.c 				  MEM->rx_head[entry].flag ));
entry             972 drivers/net/ethernet/amd/atarilance.c 	while( (MEM->rx_head[entry].flag & RMD1_OWN) == RMD1_OWN_HOST ) {
entry             973 drivers/net/ethernet/amd/atarilance.c 		struct lance_rx_head *head = &(MEM->rx_head[entry]);
entry            1001 drivers/net/ethernet/amd/atarilance.c 						if (MEM->rx_head[(entry+i) & RX_RING_MOD_MASK].flag &
entry            1033 drivers/net/ethernet/amd/atarilance.c 		entry = (++lp->cur_rx) & RX_RING_MOD_MASK;
entry             562 drivers/net/ethernet/amd/declance.c 	int entry, len;
entry             587 drivers/net/ethernet/amd/declance.c 		entry = lp->rx_new;
entry             615 drivers/net/ethernet/amd/declance.c 					((lp->rx_buf_ptr_lnc[entry] >> 16) &
entry             617 drivers/net/ethernet/amd/declance.c 				lp->rx_new = (entry + 1) & RX_RING_MOD_MASK;
entry             626 drivers/net/ethernet/amd/declance.c 				    lp->rx_buf_ptr_cpu[entry], len);
entry             637 drivers/net/ethernet/amd/declance.c 			((lp->rx_buf_ptr_lnc[entry] >> 16) & 0xff) | LE_R1_OWN;
entry             638 drivers/net/ethernet/amd/declance.c 		lp->rx_new = (entry + 1) & RX_RING_MOD_MASK;
entry             904 drivers/net/ethernet/amd/declance.c 	int entry, len;
entry             918 drivers/net/ethernet/amd/declance.c 	entry = lp->tx_new;
entry             919 drivers/net/ethernet/amd/declance.c 	*lib_ptr(ib, btx_ring[entry].length, lp->type) = (-len);
entry             920 drivers/net/ethernet/amd/declance.c 	*lib_ptr(ib, btx_ring[entry].misc, lp->type) = 0;
entry             922 drivers/net/ethernet/amd/declance.c 	cp_to_buf(lp->type, lp->tx_buf_ptr_cpu[entry], skb->data, len);
entry             925 drivers/net/ethernet/amd/declance.c 	*lib_ptr(ib, btx_ring[entry].tmd1, lp->type) =
entry             926 drivers/net/ethernet/amd/declance.c 		((lp->tx_buf_ptr_lnc[entry] >> 16) & 0xff) |
entry             928 drivers/net/ethernet/amd/declance.c 	lp->tx_new = (entry + 1) & TX_RING_MOD_MASK;
entry             955 drivers/net/ethernet/amd/lance.c 	int entry;
entry             970 drivers/net/ethernet/amd/lance.c 	entry = lp->cur_tx & TX_RING_MOD_MASK;
entry             980 drivers/net/ethernet/amd/lance.c 			lp->tx_ring[entry].length = -ETH_ZLEN;
entry             983 drivers/net/ethernet/amd/lance.c 			lp->tx_ring[entry].length = -skb->len;
entry             985 drivers/net/ethernet/amd/lance.c 		lp->tx_ring[entry].length = -skb->len;
entry             987 drivers/net/ethernet/amd/lance.c 	lp->tx_ring[entry].misc = 0x0000;
entry             997 drivers/net/ethernet/amd/lance.c 		skb_copy_from_linear_data(skb, &lp->tx_bounce_buffs[entry], skb->len);
entry             998 drivers/net/ethernet/amd/lance.c 		lp->tx_ring[entry].base =
entry             999 drivers/net/ethernet/amd/lance.c 			((u32)isa_virt_to_bus((lp->tx_bounce_buffs + entry)) & 0xffffff) | 0x83000000;
entry            1002 drivers/net/ethernet/amd/lance.c 		lp->tx_skbuff[entry] = skb;
entry            1003 drivers/net/ethernet/amd/lance.c 		lp->tx_ring[entry].base = ((u32)isa_virt_to_bus(skb->data) & 0xffffff) | 0x83000000;
entry            1051 drivers/net/ethernet/amd/lance.c 				int entry = dirty_tx & TX_RING_MOD_MASK;
entry            1052 drivers/net/ethernet/amd/lance.c 				int status = lp->tx_ring[entry].base;
entry            1057 drivers/net/ethernet/amd/lance.c 				lp->tx_ring[entry].base = 0;
entry            1061 drivers/net/ethernet/amd/lance.c 					int err_status = lp->tx_ring[entry].misc;
entry            1086 drivers/net/ethernet/amd/lance.c 				if (lp->tx_skbuff[entry]) {
entry            1087 drivers/net/ethernet/amd/lance.c 					dev_consume_skb_irq(lp->tx_skbuff[entry]);
entry            1088 drivers/net/ethernet/amd/lance.c 					lp->tx_skbuff[entry] = NULL;
entry            1147 drivers/net/ethernet/amd/lance.c 	int entry = lp->cur_rx & RX_RING_MOD_MASK;
entry            1151 drivers/net/ethernet/amd/lance.c 	while (lp->rx_ring[entry].base >= 0) {
entry            1152 drivers/net/ethernet/amd/lance.c 		int status = lp->rx_ring[entry].base >> 24;
entry            1169 drivers/net/ethernet/amd/lance.c 			lp->rx_ring[entry].base &= 0x03ffffff;
entry            1174 drivers/net/ethernet/amd/lance.c 			short pkt_len = (lp->rx_ring[entry].msg_length & 0xfff)-4;
entry            1189 drivers/net/ethernet/amd/lance.c 						if (lp->rx_ring[(entry+i) & RX_RING_MOD_MASK].base < 0)
entry            1195 drivers/net/ethernet/amd/lance.c 						lp->rx_ring[entry].base |= 0x80000000;
entry            1203 drivers/net/ethernet/amd/lance.c 					(unsigned char *)isa_bus_to_virt((lp->rx_ring[entry].base & 0x00ffffff)),
entry            1213 drivers/net/ethernet/amd/lance.c 		lp->rx_ring[entry].buf_length = -PKT_BUF_SZ;
entry            1214 drivers/net/ethernet/amd/lance.c 		lp->rx_ring[entry].base |= 0x80000000;
entry            1215 drivers/net/ethernet/amd/lance.c 		entry = (++lp->cur_rx) & RX_RING_MOD_MASK;
entry            1181 drivers/net/ethernet/amd/pcnet32.c 			     int entry)
entry            1244 drivers/net/ethernet/amd/pcnet32.c 				skb = lp->rx_skbuff[entry];
entry            1246 drivers/net/ethernet/amd/pcnet32.c 						 lp->rx_dma_addr[entry],
entry            1250 drivers/net/ethernet/amd/pcnet32.c 				lp->rx_skbuff[entry] = newskb;
entry            1251 drivers/net/ethernet/amd/pcnet32.c 				lp->rx_dma_addr[entry] = new_dma_addr;
entry            1268 drivers/net/ethernet/amd/pcnet32.c 					    lp->rx_dma_addr[entry],
entry            1272 drivers/net/ethernet/amd/pcnet32.c 				 (unsigned char *)(lp->rx_skbuff[entry]->data),
entry            1275 drivers/net/ethernet/amd/pcnet32.c 					       lp->rx_dma_addr[entry],
entry            1288 drivers/net/ethernet/amd/pcnet32.c 	int entry = lp->cur_rx & lp->rx_mod_mask;
entry            1289 drivers/net/ethernet/amd/pcnet32.c 	struct pcnet32_rx_head *rxp = &lp->rx_ring[entry];
entry            1294 drivers/net/ethernet/amd/pcnet32.c 		pcnet32_rx_entry(dev, lp, rxp, entry);
entry            1303 drivers/net/ethernet/amd/pcnet32.c 		entry = (++lp->cur_rx) & lp->rx_mod_mask;
entry            1304 drivers/net/ethernet/amd/pcnet32.c 		rxp = &lp->rx_ring[entry];
entry            1318 drivers/net/ethernet/amd/pcnet32.c 		int entry = dirty_tx & lp->tx_mod_mask;
entry            1319 drivers/net/ethernet/amd/pcnet32.c 		int status = (short)le16_to_cpu(lp->tx_ring[entry].status);
entry            1324 drivers/net/ethernet/amd/pcnet32.c 		lp->tx_ring[entry].base = 0;
entry            1328 drivers/net/ethernet/amd/pcnet32.c 			int err_status = le32_to_cpu(lp->tx_ring[entry].misc);
entry            1365 drivers/net/ethernet/amd/pcnet32.c 		if (lp->tx_skbuff[entry]) {
entry            1367 drivers/net/ethernet/amd/pcnet32.c 					 lp->tx_dma_addr[entry],
entry            1368 drivers/net/ethernet/amd/pcnet32.c 					 lp->tx_skbuff[entry]->
entry            1370 drivers/net/ethernet/amd/pcnet32.c 			dev_kfree_skb_any(lp->tx_skbuff[entry]);
entry            1371 drivers/net/ethernet/amd/pcnet32.c 			lp->tx_skbuff[entry] = NULL;
entry            1372 drivers/net/ethernet/amd/pcnet32.c 			lp->tx_dma_addr[entry] = 0;
entry            2504 drivers/net/ethernet/amd/pcnet32.c 	int entry;
entry            2521 drivers/net/ethernet/amd/pcnet32.c 	entry = lp->cur_tx & lp->tx_mod_mask;
entry            2526 drivers/net/ethernet/amd/pcnet32.c 	lp->tx_ring[entry].length = cpu_to_le16(-skb->len);
entry            2528 drivers/net/ethernet/amd/pcnet32.c 	lp->tx_ring[entry].misc = 0x00000000;
entry            2530 drivers/net/ethernet/amd/pcnet32.c 	lp->tx_dma_addr[entry] =
entry            2532 drivers/net/ethernet/amd/pcnet32.c 	if (pci_dma_mapping_error(lp->pci_dev, lp->tx_dma_addr[entry])) {
entry            2537 drivers/net/ethernet/amd/pcnet32.c 	lp->tx_skbuff[entry] = skb;
entry            2538 drivers/net/ethernet/amd/pcnet32.c 	lp->tx_ring[entry].base = cpu_to_le32(lp->tx_dma_addr[entry]);
entry            2540 drivers/net/ethernet/amd/pcnet32.c 	lp->tx_ring[entry].status = cpu_to_le16(status);
entry            2548 drivers/net/ethernet/amd/pcnet32.c 	if (lp->tx_ring[(entry + 1) & lp->tx_mod_mask].base != 0) {
entry             519 drivers/net/ethernet/amd/sun3lance.c 	int entry, len;
entry             613 drivers/net/ethernet/amd/sun3lance.c 	entry = lp->new_tx;
entry             614 drivers/net/ethernet/amd/sun3lance.c 	head  = &(MEM->tx_head[entry]);
entry             644 drivers/net/ethernet/amd/sun3lance.c 	if ((MEM->tx_head[(entry+1) & TX_RING_MOD_MASK].flag & TMD1_OWN) ==
entry             785 drivers/net/ethernet/amd/sun3lance.c 	int entry = lp->new_rx;
entry             788 drivers/net/ethernet/amd/sun3lance.c 	while( (MEM->rx_head[entry].flag & RMD1_OWN) == RMD1_OWN_HOST ) {
entry             789 drivers/net/ethernet/amd/sun3lance.c 		struct lance_rx_head *head = &(MEM->rx_head[entry]);
entry             841 drivers/net/ethernet/amd/sun3lance.c 					printk( "%s: RX pkt %d type 0x%04x len %d\n ", dev->name, entry, ((u_short *)data)[6], pkt_len);
entry             862 drivers/net/ethernet/amd/sun3lance.c 		entry = lp->new_rx = (lp->new_rx +1) & RX_RING_MOD_MASK;
entry             514 drivers/net/ethernet/amd/sunlance.c 	int len, entry = lp->rx_new;
entry             517 drivers/net/ethernet/amd/sunlance.c 	for (rd = &ib->brx_ring [entry];
entry             519 drivers/net/ethernet/amd/sunlance.c 	     rd = &ib->brx_ring [entry]) {
entry             542 drivers/net/ethernet/amd/sunlance.c 				lp->rx_new = RX_NEXT(entry);
entry             551 drivers/net/ethernet/amd/sunlance.c 					 (unsigned char *)&(ib->rx_buf [entry][0]),
entry             561 drivers/net/ethernet/amd/sunlance.c 		entry = RX_NEXT(entry);
entry             564 drivers/net/ethernet/amd/sunlance.c 	lp->rx_new = entry;
entry             683 drivers/net/ethernet/amd/sunlance.c 	int len, entry;
entry             686 drivers/net/ethernet/amd/sunlance.c 	entry = lp->rx_new;
entry             687 drivers/net/ethernet/amd/sunlance.c 	for (rd = &ib->brx_ring [entry];
entry             689 drivers/net/ethernet/amd/sunlance.c 	     rd = &ib->brx_ring [entry]) {
entry             712 drivers/net/ethernet/amd/sunlance.c 				lp->rx_new = RX_NEXT(entry);
entry             720 drivers/net/ethernet/amd/sunlance.c 			lance_piocopy_to_skb(skb, &(ib->rx_buf[entry][0]), len);
entry             729 drivers/net/ethernet/amd/sunlance.c 		entry = RX_NEXT(entry);
entry             732 drivers/net/ethernet/amd/sunlance.c 	lp->rx_new = entry;
entry             882 drivers/net/ethernet/amd/sunlance.c 	int i, entry;
entry             884 drivers/net/ethernet/amd/sunlance.c 	entry = lp->tx_new & TX_RING_MOD_MASK;
entry             887 drivers/net/ethernet/amd/sunlance.c 		u16 __iomem *packet = (u16 __iomem *) &(ib->tx_buf[entry][0]);
entry             895 drivers/net/ethernet/amd/sunlance.c 		sbus_writew((-ETH_ZLEN) | 0xf000, &ib->btx_ring[entry].length);
entry             896 drivers/net/ethernet/amd/sunlance.c 		sbus_writew(0, &ib->btx_ring[entry].misc);
entry             897 drivers/net/ethernet/amd/sunlance.c 		sbus_writeb(LE_T1_POK|LE_T1_OWN, &ib->btx_ring[entry].tmd1_bits);
entry             900 drivers/net/ethernet/amd/sunlance.c 		u16 *packet = (u16 *) &(ib->tx_buf[entry][0]);
entry             907 drivers/net/ethernet/amd/sunlance.c 		ib->btx_ring[entry].length = (-ETH_ZLEN) | 0xf000;
entry             908 drivers/net/ethernet/amd/sunlance.c 		ib->btx_ring[entry].misc = 0;
entry             909 drivers/net/ethernet/amd/sunlance.c 		ib->btx_ring[entry].tmd1_bits = (LE_T1_POK|LE_T1_OWN);
entry             911 drivers/net/ethernet/amd/sunlance.c 	lp->tx_new = TX_NEXT(entry);
entry            1113 drivers/net/ethernet/amd/sunlance.c 	int entry, skblen, len;
entry            1123 drivers/net/ethernet/amd/sunlance.c 	entry = lp->tx_new & TX_RING_MOD_MASK;
entry            1126 drivers/net/ethernet/amd/sunlance.c 		sbus_writew((-len) | 0xf000, &ib->btx_ring[entry].length);
entry            1127 drivers/net/ethernet/amd/sunlance.c 		sbus_writew(0, &ib->btx_ring[entry].misc);
entry            1128 drivers/net/ethernet/amd/sunlance.c 		lance_piocopy_from_skb(&ib->tx_buf[entry][0], skb->data, skblen);
entry            1130 drivers/net/ethernet/amd/sunlance.c 			lance_piozero(&ib->tx_buf[entry][skblen], len - skblen);
entry            1131 drivers/net/ethernet/amd/sunlance.c 		sbus_writeb(LE_T1_POK | LE_T1_OWN, &ib->btx_ring[entry].tmd1_bits);
entry            1134 drivers/net/ethernet/amd/sunlance.c 		ib->btx_ring [entry].length = (-len) | 0xf000;
entry            1135 drivers/net/ethernet/amd/sunlance.c 		ib->btx_ring [entry].misc = 0;
entry            1136 drivers/net/ethernet/amd/sunlance.c 		skb_copy_from_linear_data(skb, &ib->tx_buf [entry][0], skblen);
entry            1138 drivers/net/ethernet/amd/sunlance.c 			memset((char *) &ib->tx_buf [entry][skblen], 0, len - skblen);
entry            1139 drivers/net/ethernet/amd/sunlance.c 		ib->btx_ring [entry].tmd1_bits = (LE_T1_POK | LE_T1_OWN);
entry            1142 drivers/net/ethernet/amd/sunlance.c 	lp->tx_new = TX_NEXT(entry);
entry              54 drivers/net/ethernet/atheros/alx/main.c static void alx_free_txbuf(struct alx_tx_queue *txq, int entry)
entry              56 drivers/net/ethernet/atheros/alx/main.c 	struct alx_buffer *txb = &txq->bufs[entry];
entry             978 drivers/net/ethernet/broadcom/b44.c 	u32 len, entry, ctrl;
entry            1019 drivers/net/ethernet/broadcom/b44.c 	entry = bp->tx_prod;
entry            1020 drivers/net/ethernet/broadcom/b44.c 	bp->tx_buffers[entry].skb = skb;
entry            1021 drivers/net/ethernet/broadcom/b44.c 	bp->tx_buffers[entry].mapping = mapping;
entry            1025 drivers/net/ethernet/broadcom/b44.c 	if (entry == (B44_TX_RING_SIZE - 1))
entry            1028 drivers/net/ethernet/broadcom/b44.c 	bp->tx_ring[entry].ctrl = cpu_to_le32(ctrl);
entry            1029 drivers/net/ethernet/broadcom/b44.c 	bp->tx_ring[entry].addr = cpu_to_le32((u32) mapping+bp->dma_offset);
entry            1033 drivers/net/ethernet/broadcom/b44.c 			                    entry * sizeof(bp->tx_ring[0]),
entry            1036 drivers/net/ethernet/broadcom/b44.c 	entry = NEXT_TX(entry);
entry            1038 drivers/net/ethernet/broadcom/b44.c 	bp->tx_prod = entry;
entry            1042 drivers/net/ethernet/broadcom/b44.c 	bw32(bp, B44_DMATX_PTR, entry * sizeof(struct dma_desc));
entry            1044 drivers/net/ethernet/broadcom/b44.c 		bw32(bp, B44_DMATX_PTR, entry * sizeof(struct dma_desc));
entry            3683 drivers/net/ethernet/broadcom/bnx2.c 		    const struct bnx2_mips_fw_file_entry *entry)
entry            3685 drivers/net/ethernet/broadcom/bnx2.c 	if (check_fw_section(fw, &entry->text, 4, true) ||
entry            3686 drivers/net/ethernet/broadcom/bnx2.c 	    check_fw_section(fw, &entry->data, 4, false) ||
entry            3687 drivers/net/ethernet/broadcom/bnx2.c 	    check_fw_section(fw, &entry->rodata, 4, false))
entry            6288 drivers/net/ethernet/broadcom/bnx2.c 		msix_ent[i].entry = i;
entry            1667 drivers/net/ethernet/broadcom/bnx2x/bnx2x_cmn.c 		bp->msix_table[msix_vec].entry = msix_vec;
entry            1669 drivers/net/ethernet/broadcom/bnx2x/bnx2x_cmn.c 			       bp->msix_table[0].entry);
entry            1675 drivers/net/ethernet/broadcom/bnx2x/bnx2x_cmn.c 		bp->msix_table[msix_vec].entry = msix_vec;
entry            1677 drivers/net/ethernet/broadcom/bnx2x/bnx2x_cmn.c 			       msix_vec, bp->msix_table[msix_vec].entry);
entry            1683 drivers/net/ethernet/broadcom/bnx2x/bnx2x_cmn.c 		bp->msix_table[msix_vec].entry = msix_vec;
entry             219 drivers/net/ethernet/broadcom/bnx2x/bnx2x_dcb.c 			struct dcbx_app_priority_entry *entry =
entry             223 drivers/net/ethernet/broadcom/bnx2x/bnx2x_dcb.c 			if (GET_FLAGS(entry[index].appBitfield,
entry             225 drivers/net/ethernet/broadcom/bnx2x/bnx2x_dcb.c 			    GET_FLAGS(entry[index].appBitfield,
entry             228 drivers/net/ethernet/broadcom/bnx2x/bnx2x_dcb.c 			} else if (GET_FLAGS(entry[index].appBitfield,
entry             230 drivers/net/ethernet/broadcom/bnx2x/bnx2x_dcb.c 				   TCP_PORT_ISCSI == entry[index].app_id) {
entry             233 drivers/net/ethernet/broadcom/bnx2x/bnx2x_dcb.c 			} else if (GET_FLAGS(entry[index].appBitfield,
entry             235 drivers/net/ethernet/broadcom/bnx2x/bnx2x_dcb.c 				   ETH_TYPE_FCOE == entry[index].app_id) {
entry             244 drivers/net/ethernet/broadcom/bnx2x/bnx2x_dcb.c 						   entry[index].pri_bitmap,
entry            1192 drivers/net/ethernet/broadcom/bnx2x/bnx2x_dcb.c 	u8	entry		= 0;
entry            1206 drivers/net/ethernet/broadcom/bnx2x/bnx2x_dcb.c 			entry = 1;
entry            1209 drivers/net/ethernet/broadcom/bnx2x/bnx2x_dcb.c 			entry = 0;
entry            1215 drivers/net/ethernet/broadcom/bnx2x/bnx2x_dcb.c 			bnx2x_dcbx_add_to_cos_bw(bp, &cos_data->data[entry],
entry            1221 drivers/net/ethernet/broadcom/bnx2x/bnx2x_dcb.c 			cos_data->data[entry].strict =
entry            1504 drivers/net/ethernet/broadcom/bnx2x/bnx2x_dcb.c 	u8 entry = 0;
entry            1545 drivers/net/ethernet/broadcom/bnx2x/bnx2x_dcb.c 				entry = 0;
entry            1553 drivers/net/ethernet/broadcom/bnx2x/bnx2x_dcb.c 					entry = 1;
entry            1554 drivers/net/ethernet/broadcom/bnx2x/bnx2x_dcb.c 				cos_data->data[entry].pri_join_mask |=
entry            1557 drivers/net/ethernet/broadcom/bnx2x/bnx2x_dcb.c 					&cos_data->data[entry],
entry            1623 drivers/net/ethernet/broadcom/bnx2x/bnx2x_dcb.c 					u8 entry,
entry            1635 drivers/net/ethernet/broadcom/bnx2x/bnx2x_dcb.c 								data[entry];
entry            1656 drivers/net/ethernet/broadcom/bnx2x/bnx2x_dcb.c 			entry++;
entry            1672 drivers/net/ethernet/broadcom/bnx2x/bnx2x_dcb.c 					 u8 entry,
entry            1676 drivers/net/ethernet/broadcom/bnx2x/bnx2x_dcb.c 	if (bnx2x_dcbx_spread_strict_pri(bp, cos_data, entry,
entry            1680 drivers/net/ethernet/broadcom/bnx2x/bnx2x_dcb.c 						    data[entry];
entry            1702 drivers/net/ethernet/broadcom/bnx2x/bnx2x_dcb.c 	u8 entry = 0;
entry            1723 drivers/net/ethernet/broadcom/bnx2x/bnx2x_dcb.c 							    data[entry];
entry            1731 drivers/net/ethernet/broadcom/bnx2x/bnx2x_dcb.c 			entry++;
entry            1742 drivers/net/ethernet/broadcom/bnx2x/bnx2x_dcb.c 			entry += bnx2x_dcbx_cee_fill_strict_pri(bp, cos_data,
entry            1743 drivers/net/ethernet/broadcom/bnx2x/bnx2x_dcb.c 				entry, need_num_of_entries, pg->pg_priority);
entry            1748 drivers/net/ethernet/broadcom/bnx2x/bnx2x_dcb.c 	cos_data->num_of_cos = entry;
entry            2811 drivers/net/ethernet/broadcom/bnx2x/bnx2x_ethtool.c 				struct code_entry *entry,
entry            2814 drivers/net/ethernet/broadcom/bnx2x/bnx2x_ethtool.c 	size_t size = entry->code_attribute & CODE_IMAGE_LENGTH_MASK;
entry            2815 drivers/net/ethernet/broadcom/bnx2x/bnx2x_ethtool.c 	u32 type = entry->code_attribute & CODE_IMAGE_TYPE_MASK;
entry            2822 drivers/net/ethernet/broadcom/bnx2x/bnx2x_ethtool.c 	rc = bnx2x_nvram_crc(bp, entry->nvm_start_addr, size, buff);
entry            2833 drivers/net/ethernet/broadcom/bnx2x/bnx2x_ethtool.c 	struct code_entry entry;
entry            2835 drivers/net/ethernet/broadcom/bnx2x/bnx2x_ethtool.c 	rc = bnx2x_nvram_read32(bp, addr, (u32 *)&entry, sizeof(entry));
entry            2839 drivers/net/ethernet/broadcom/bnx2x/bnx2x_ethtool.c 	return bnx2x_test_nvram_dir(bp, &entry, buff);
entry            2845 drivers/net/ethernet/broadcom/bnx2x/bnx2x_ethtool.c 	struct code_entry entry;
entry            2850 drivers/net/ethernet/broadcom/bnx2x/bnx2x_ethtool.c 				sizeof(entry) * CODE_ENTRY_EXTENDED_DIR_IDX,
entry            2851 drivers/net/ethernet/broadcom/bnx2x/bnx2x_ethtool.c 				(u32 *)&entry, sizeof(entry));
entry            2855 drivers/net/ethernet/broadcom/bnx2x/bnx2x_ethtool.c 	if (!EXTENDED_DIR_EXISTS(entry.code_attribute))
entry            2858 drivers/net/ethernet/broadcom/bnx2x/bnx2x_ethtool.c 	rc = bnx2x_nvram_read32(bp, entry.nvm_start_addr,
entry            2863 drivers/net/ethernet/broadcom/bnx2x/bnx2x_ethtool.c 	dir_offset = entry.nvm_start_addr + 8;
entry             672 drivers/net/ethernet/broadcom/bnx2x/bnx2x_main.c 					      int entry)
entry             676 drivers/net/ethernet/broadcom/bnx2x/bnx2x_main.c 		return XSTORM_ASSERT_LIST_OFFSET(entry);
entry             678 drivers/net/ethernet/broadcom/bnx2x/bnx2x_main.c 		return TSTORM_ASSERT_LIST_OFFSET(entry);
entry             680 drivers/net/ethernet/broadcom/bnx2x/bnx2x_main.c 		return CSTORM_ASSERT_LIST_OFFSET(entry);
entry             682 drivers/net/ethernet/broadcom/bnx2x/bnx2x_main.c 		return USTORM_ASSERT_LIST_OFFSET(entry);
entry            9147 drivers/net/ethernet/broadcom/bnx2x/bnx2x_main.c 		u8 entry = (BP_VN(bp) + 1)*8;
entry            9150 drivers/net/ethernet/broadcom/bnx2x/bnx2x_main.c 		EMAC_WR(bp, EMAC_REG_EMAC_MAC_MATCH + entry, val);
entry            9154 drivers/net/ethernet/broadcom/bnx2x/bnx2x_main.c 		EMAC_WR(bp, EMAC_REG_EMAC_MAC_MATCH + entry + 4, val);
entry            10686 drivers/net/ethernet/broadcom/bnx2x/bnx2x_main.c 	struct bnx2x_prev_path_list *entry;
entry            10691 drivers/net/ethernet/broadcom/bnx2x/bnx2x_main.c 	entry = bnx2x_prev_path_get_entry(bp);
entry            10692 drivers/net/ethernet/broadcom/bnx2x/bnx2x_main.c 	val = !!(entry && (entry->undi & (1 << BP_PORT(bp))));
entry             682 drivers/net/ethernet/broadcom/bnx2x/bnx2x_sp.h 	bool (*get_entry)(struct bnx2x_credit_pool_obj *o, int *entry);
entry             690 drivers/net/ethernet/broadcom/bnx2x/bnx2x_sp.h 	bool (*put_entry)(struct bnx2x_credit_pool_obj *o, int entry);
entry            8048 drivers/net/ethernet/broadcom/bnxt/bnxt.c 		msix_ent[i].entry = i;
entry            7705 drivers/net/ethernet/broadcom/tg3.c static bool tg3_tx_frag_set(struct tg3_napi *tnapi, u32 *entry, u32 *budget,
entry            7725 drivers/net/ethernet/broadcom/tg3.c 		u32 prvidx = *entry;
entry            7737 drivers/net/ethernet/broadcom/tg3.c 			tnapi->tx_buffers[*entry].fragmented = true;
entry            7739 drivers/net/ethernet/broadcom/tg3.c 			tg3_tx_set_bd(&tnapi->tx_ring[*entry], map,
entry            7742 drivers/net/ethernet/broadcom/tg3.c 			prvidx = *entry;
entry            7743 drivers/net/ethernet/broadcom/tg3.c 			*entry = NEXT_TX(*entry);
entry            7750 drivers/net/ethernet/broadcom/tg3.c 				tg3_tx_set_bd(&tnapi->tx_ring[*entry], map,
entry            7753 drivers/net/ethernet/broadcom/tg3.c 				*entry = NEXT_TX(*entry);
entry            7760 drivers/net/ethernet/broadcom/tg3.c 		tg3_tx_set_bd(&tnapi->tx_ring[*entry], map,
entry            7762 drivers/net/ethernet/broadcom/tg3.c 		*entry = NEXT_TX(*entry);
entry            7768 drivers/net/ethernet/broadcom/tg3.c static void tg3_tx_skb_unmap(struct tg3_napi *tnapi, u32 entry, int last)
entry            7772 drivers/net/ethernet/broadcom/tg3.c 	struct tg3_tx_ring_info *txb = &tnapi->tx_buffers[entry];
entry            7784 drivers/net/ethernet/broadcom/tg3.c 		entry = NEXT_TX(entry);
entry            7785 drivers/net/ethernet/broadcom/tg3.c 		txb = &tnapi->tx_buffers[entry];
entry            7791 drivers/net/ethernet/broadcom/tg3.c 		entry = NEXT_TX(entry);
entry            7792 drivers/net/ethernet/broadcom/tg3.c 		txb = &tnapi->tx_buffers[entry];
entry            7800 drivers/net/ethernet/broadcom/tg3.c 			entry = NEXT_TX(entry);
entry            7801 drivers/net/ethernet/broadcom/tg3.c 			txb = &tnapi->tx_buffers[entry];
entry            7809 drivers/net/ethernet/broadcom/tg3.c 				       u32 *entry, u32 *budget,
entry            7838 drivers/net/ethernet/broadcom/tg3.c 			u32 save_entry = *entry;
entry            7842 drivers/net/ethernet/broadcom/tg3.c 			tnapi->tx_buffers[*entry].skb = new_skb;
entry            7843 drivers/net/ethernet/broadcom/tg3.c 			dma_unmap_addr_set(&tnapi->tx_buffers[*entry],
entry            7846 drivers/net/ethernet/broadcom/tg3.c 			if (tg3_tx_frag_set(tnapi, entry, budget, new_addr,
entry            7918 drivers/net/ethernet/broadcom/tg3.c 	u32 len, entry, base_flags, mss, vlan = 0;
entry            7953 drivers/net/ethernet/broadcom/tg3.c 	entry = tnapi->tx_prod;
entry            8065 drivers/net/ethernet/broadcom/tg3.c 	tnapi->tx_buffers[entry].skb = skb;
entry            8066 drivers/net/ethernet/broadcom/tg3.c 	dma_unmap_addr_set(&tnapi->tx_buffers[entry], mapping, mapping);
entry            8073 drivers/net/ethernet/broadcom/tg3.c 	if (tg3_tx_frag_set(tnapi, &entry, &budget, mapping, len, base_flags |
entry            8096 drivers/net/ethernet/broadcom/tg3.c 			tnapi->tx_buffers[entry].skb = NULL;
entry            8097 drivers/net/ethernet/broadcom/tg3.c 			dma_unmap_addr_set(&tnapi->tx_buffers[entry], mapping,
entry            8103 drivers/net/ethernet/broadcom/tg3.c 			    tg3_tx_frag_set(tnapi, &entry, &budget, mapping,
entry            8131 drivers/net/ethernet/broadcom/tg3.c 		entry = tnapi->tx_prod;
entry            8133 drivers/net/ethernet/broadcom/tg3.c 		if (tigon3_dma_hwbug_workaround(tnapi, &skb, &entry, &budget,
entry            8144 drivers/net/ethernet/broadcom/tg3.c 	tnapi->tx_prod = entry;
entry            8160 drivers/net/ethernet/broadcom/tg3.c 		tw32_tx_mbox(tnapi->prodmbox, entry);
entry            11482 drivers/net/ethernet/broadcom/tg3.c 		msix_ent[i].entry  = i;
entry            2654 drivers/net/ethernet/brocade/bna/bnad.c 		bnad->msix_table[i].entry = i;
entry            3126 drivers/net/ethernet/brocade/bna/bnad.c 	int entry;
entry            3140 drivers/net/ethernet/brocade/bna/bnad.c 	entry = 0;
entry            3142 drivers/net/ethernet/brocade/bna/bnad.c 		ether_addr_copy(&mac_list[entry * ETH_ALEN], &ha->addr[0]);
entry            3143 drivers/net/ethernet/brocade/bna/bnad.c 		entry++;
entry            3146 drivers/net/ethernet/brocade/bna/bnad.c 	ret = bna_rx_ucast_listset(bnad->rx_info[0].rx, entry, mac_list);
entry             917 drivers/net/ethernet/cadence/macb_main.c 	unsigned int		entry;
entry             925 drivers/net/ethernet/cadence/macb_main.c 		entry = macb_rx_ring_wrap(bp, queue->rx_prepared_head);
entry             931 drivers/net/ethernet/cadence/macb_main.c 		desc = macb_rx_desc(queue, entry);
entry             933 drivers/net/ethernet/cadence/macb_main.c 		if (!queue->rx_skbuff[entry]) {
entry             951 drivers/net/ethernet/cadence/macb_main.c 			queue->rx_skbuff[entry] = skb;
entry             953 drivers/net/ethernet/cadence/macb_main.c 			if (entry == bp->rx_ring_size - 1)
entry            1004 drivers/net/ethernet/cadence/macb_main.c 	unsigned int		entry;
entry            1014 drivers/net/ethernet/cadence/macb_main.c 		entry = macb_rx_ring_wrap(bp, queue->rx_tail);
entry            1015 drivers/net/ethernet/cadence/macb_main.c 		desc = macb_rx_desc(queue, entry);
entry            1041 drivers/net/ethernet/cadence/macb_main.c 		skb = queue->rx_skbuff[entry];
entry            1050 drivers/net/ethernet/cadence/macb_main.c 		queue->rx_skbuff[entry] = NULL;
entry            1053 drivers/net/ethernet/cadence/macb_main.c 		netdev_vdbg(bp->dev, "gem_rx %u (len %u)\n", entry, len);
entry            1499 drivers/net/ethernet/cadence/macb_main.c 	unsigned int len, entry, i, tx_head = queue->tx_head;
entry            1525 drivers/net/ethernet/cadence/macb_main.c 		entry = macb_tx_ring_wrap(bp, tx_head);
entry            1526 drivers/net/ethernet/cadence/macb_main.c 		tx_skb = &queue->tx_skb[entry];
entry            1556 drivers/net/ethernet/cadence/macb_main.c 			entry = macb_tx_ring_wrap(bp, tx_head);
entry            1557 drivers/net/ethernet/cadence/macb_main.c 			tx_skb = &queue->tx_skb[entry];
entry            1594 drivers/net/ethernet/cadence/macb_main.c 	entry = macb_tx_ring_wrap(bp, i);
entry            1596 drivers/net/ethernet/cadence/macb_main.c 	desc = macb_tx_desc(queue, entry);
entry            1616 drivers/net/ethernet/cadence/macb_main.c 		entry = macb_tx_ring_wrap(bp, i);
entry            1617 drivers/net/ethernet/cadence/macb_main.c 		tx_skb = &queue->tx_skb[entry];
entry            1618 drivers/net/ethernet/cadence/macb_main.c 		desc = macb_tx_desc(queue, entry);
entry            1625 drivers/net/ethernet/cadence/macb_main.c 		if (unlikely(entry == (bp->tx_ring_size - 1)))
entry             684 drivers/net/ethernet/calxeda/xgmac.c 		int entry = priv->rx_head;
entry             687 drivers/net/ethernet/calxeda/xgmac.c 		p = priv->dma_rx + entry;
entry             689 drivers/net/ethernet/calxeda/xgmac.c 		if (priv->rx_skbuff[entry] == NULL) {
entry             701 drivers/net/ethernet/calxeda/xgmac.c 			priv->rx_skbuff[entry] = skb;
entry             869 drivers/net/ethernet/calxeda/xgmac.c 		unsigned int entry = priv->tx_tail;
entry             870 drivers/net/ethernet/calxeda/xgmac.c 		struct sk_buff *skb = priv->tx_skbuff[entry];
entry             871 drivers/net/ethernet/calxeda/xgmac.c 		struct xgmac_dma_desc *p = priv->dma_tx + entry;
entry             893 drivers/net/ethernet/calxeda/xgmac.c 		priv->tx_skbuff[entry] = NULL;
entry             894 drivers/net/ethernet/calxeda/xgmac.c 		priv->tx_tail = dma_ring_incr(entry, DMA_TX_RING_SZ);
entry            1077 drivers/net/ethernet/calxeda/xgmac.c 	unsigned int entry;
entry            1091 drivers/net/ethernet/calxeda/xgmac.c 	entry = priv->tx_head;
entry            1092 drivers/net/ethernet/calxeda/xgmac.c 	desc = priv->dma_tx + entry;
entry            1101 drivers/net/ethernet/calxeda/xgmac.c 	priv->tx_skbuff[entry] = skb;
entry            1114 drivers/net/ethernet/calxeda/xgmac.c 		entry = dma_ring_incr(entry, DMA_TX_RING_SZ);
entry            1115 drivers/net/ethernet/calxeda/xgmac.c 		desc = priv->dma_tx + entry;
entry            1116 drivers/net/ethernet/calxeda/xgmac.c 		priv->tx_skbuff[entry] = skb;
entry            1136 drivers/net/ethernet/calxeda/xgmac.c 	priv->tx_head = dma_ring_incr(entry, DMA_TX_RING_SZ);
entry            1150 drivers/net/ethernet/calxeda/xgmac.c 	entry = priv->tx_head;
entry            1152 drivers/net/ethernet/calxeda/xgmac.c 		entry = dma_ring_incr(entry, DMA_TX_RING_SZ);
entry            1153 drivers/net/ethernet/calxeda/xgmac.c 		desc = priv->dma_tx + entry;
entry            1154 drivers/net/ethernet/calxeda/xgmac.c 		priv->tx_skbuff[entry] = NULL;
entry            1168 drivers/net/ethernet/calxeda/xgmac.c 	unsigned int entry;
entry            1180 drivers/net/ethernet/calxeda/xgmac.c 		entry = priv->rx_tail;
entry            1181 drivers/net/ethernet/calxeda/xgmac.c 		p = priv->dma_rx + entry;
entry            1193 drivers/net/ethernet/calxeda/xgmac.c 		skb = priv->rx_skbuff[entry];
entry            1198 drivers/net/ethernet/calxeda/xgmac.c 		priv->rx_skbuff[entry] = NULL;
entry            1062 drivers/net/ethernet/cavium/liquidio/lio_core.c 				msix_entries[i].entry =
entry            1065 drivers/net/ethernet/cavium/liquidio/lio_core.c 			msix_entries[oct->num_msix_irqs - 1].entry =
entry            1069 drivers/net/ethernet/cavium/liquidio/lio_core.c 				msix_entries[i].entry = i;
entry            3125 drivers/net/ethernet/chelsio/cxgb3/cxgb3_main.c 		entries[i].entry = i;
entry              19 drivers/net/ethernet/chelsio/cxgb4/cudbg_entity.h 	struct mbox_cmd entry;
entry            2835 drivers/net/ethernet/chelsio/cxgb4/cudbg_lib.c 	struct mbox_cmd *entry;
entry            2855 drivers/net/ethernet/chelsio/cxgb4/cudbg_lib.c 		entry = mbox_cmd_log_entry(log, entry_idx);
entry            2857 drivers/net/ethernet/chelsio/cxgb4/cudbg_lib.c 		if (entry->timestamp == 0)
entry            2860 drivers/net/ethernet/chelsio/cxgb4/cudbg_lib.c 		memcpy(&mboxlog->entry, entry, sizeof(struct mbox_cmd));
entry            2862 drivers/net/ethernet/chelsio/cxgb4/cudbg_lib.c 			flit = entry->cmd[i];
entry             239 drivers/net/ethernet/chelsio/cxgb4/cudbg_lib.h 					struct cudbg_qdesc_entry *entry)
entry             241 drivers/net/ethernet/chelsio/cxgb4/cudbg_lib.h 	entry->qtype = type;
entry             242 drivers/net/ethernet/chelsio/cxgb4/cudbg_lib.h 	entry->qid = txq->cntxt_id;
entry             243 drivers/net/ethernet/chelsio/cxgb4/cudbg_lib.h 	entry->desc_size = sizeof(struct tx_desc);
entry             244 drivers/net/ethernet/chelsio/cxgb4/cudbg_lib.h 	entry->num_desc = txq->size;
entry             245 drivers/net/ethernet/chelsio/cxgb4/cudbg_lib.h 	entry->data_size = txq->size * sizeof(struct tx_desc);
entry             246 drivers/net/ethernet/chelsio/cxgb4/cudbg_lib.h 	memcpy(entry->data, txq->desc, entry->data_size);
entry             251 drivers/net/ethernet/chelsio/cxgb4/cudbg_lib.h 					struct cudbg_qdesc_entry *entry)
entry             253 drivers/net/ethernet/chelsio/cxgb4/cudbg_lib.h 	entry->qtype = type;
entry             254 drivers/net/ethernet/chelsio/cxgb4/cudbg_lib.h 	entry->qid = rxq->cntxt_id;
entry             255 drivers/net/ethernet/chelsio/cxgb4/cudbg_lib.h 	entry->desc_size = rxq->iqe_len;
entry             256 drivers/net/ethernet/chelsio/cxgb4/cudbg_lib.h 	entry->num_desc = rxq->size;
entry             257 drivers/net/ethernet/chelsio/cxgb4/cudbg_lib.h 	entry->data_size = rxq->size * rxq->iqe_len;
entry             258 drivers/net/ethernet/chelsio/cxgb4/cudbg_lib.h 	memcpy(entry->data, rxq->desc, entry->data_size);
entry             263 drivers/net/ethernet/chelsio/cxgb4/cudbg_lib.h 					struct cudbg_qdesc_entry *entry)
entry             265 drivers/net/ethernet/chelsio/cxgb4/cudbg_lib.h 	entry->qtype = type;
entry             266 drivers/net/ethernet/chelsio/cxgb4/cudbg_lib.h 	entry->qid = flq->cntxt_id;
entry             267 drivers/net/ethernet/chelsio/cxgb4/cudbg_lib.h 	entry->desc_size = sizeof(__be64);
entry             268 drivers/net/ethernet/chelsio/cxgb4/cudbg_lib.h 	entry->num_desc = flq->size;
entry             269 drivers/net/ethernet/chelsio/cxgb4/cudbg_lib.h 	entry->data_size = flq->size * sizeof(__be64);
entry             270 drivers/net/ethernet/chelsio/cxgb4/cudbg_lib.h 	memcpy(entry->data, flq->desc, entry->data_size);
entry            1157 drivers/net/ethernet/chelsio/cxgb4/cxgb4_debugfs.c 	struct mbox_cmd *entry;
entry            1171 drivers/net/ethernet/chelsio/cxgb4/cxgb4_debugfs.c 	entry = mbox_cmd_log_entry(log, entry_idx);
entry            1174 drivers/net/ethernet/chelsio/cxgb4/cxgb4_debugfs.c 	if (entry->timestamp == 0)
entry            1178 drivers/net/ethernet/chelsio/cxgb4/cxgb4_debugfs.c 		   entry->seqno, entry->timestamp,
entry            1179 drivers/net/ethernet/chelsio/cxgb4/cxgb4_debugfs.c 		   entry->access, entry->execute);
entry            1181 drivers/net/ethernet/chelsio/cxgb4/cxgb4_debugfs.c 		u64 flit = entry->cmd[i];
entry            1988 drivers/net/ethernet/chelsio/cxgb4/cxgb4_debugfs.c 	u16 *entry = v;
entry            1991 drivers/net/ethernet/chelsio/cxgb4/cxgb4_debugfs.c 		   idx * 8, entry[0], entry[1], entry[2], entry[3], entry[4],
entry            1992 drivers/net/ethernet/chelsio/cxgb4/cxgb4_debugfs.c 		   entry[5], entry[6], entry[7]);
entry             351 drivers/net/ethernet/chelsio/cxgb4/cxgb4_main.c 	struct hash_mac_addr *entry;
entry             354 drivers/net/ethernet/chelsio/cxgb4/cxgb4_main.c 	list_for_each_entry(entry, &adap->mac_hlist, list) {
entry             355 drivers/net/ethernet/chelsio/cxgb4/cxgb4_main.c 		ucast |= is_unicast_ether_addr(entry->addr);
entry             356 drivers/net/ethernet/chelsio/cxgb4/cxgb4_main.c 		vec |= (1ULL << hash_mac_addr(entry->addr));
entry             406 drivers/net/ethernet/chelsio/cxgb4/cxgb4_main.c 	struct hash_mac_addr *entry, *tmp;
entry             411 drivers/net/ethernet/chelsio/cxgb4/cxgb4_main.c 	list_for_each_entry_safe(entry, tmp, &adap->mac_hlist, list) {
entry             412 drivers/net/ethernet/chelsio/cxgb4/cxgb4_main.c 		if (ether_addr_equal(entry->addr, mac_addr)) {
entry             413 drivers/net/ethernet/chelsio/cxgb4/cxgb4_main.c 			list_del(&entry->list);
entry             414 drivers/net/ethernet/chelsio/cxgb4/cxgb4_main.c 			kfree(entry);
entry             463 drivers/net/ethernet/chelsio/cxgb4/cxgb4_main.c 	struct hash_mac_addr *entry, *new_entry;
entry             473 drivers/net/ethernet/chelsio/cxgb4/cxgb4_main.c 		list_for_each_entry(entry, &adapter->mac_hlist, list) {
entry             474 drivers/net/ethernet/chelsio/cxgb4/cxgb4_main.c 			if (entry->iface_mac) {
entry             475 drivers/net/ethernet/chelsio/cxgb4/cxgb4_main.c 				ether_addr_copy(entry->addr, addr);
entry            5298 drivers/net/ethernet/chelsio/cxgb4/cxgb4_main.c 		entries[i].entry = i;
entry            6133 drivers/net/ethernet/chelsio/cxgb4/cxgb4_main.c 	struct hash_mac_addr *entry, *tmp;
entry            6185 drivers/net/ethernet/chelsio/cxgb4/cxgb4_main.c 		list_for_each_entry_safe(entry, tmp, &adapter->mac_hlist,
entry            6187 drivers/net/ethernet/chelsio/cxgb4/cxgb4_main.c 			list_del(&entry->list);
entry            6188 drivers/net/ethernet/chelsio/cxgb4/cxgb4_main.c 			kfree(entry);
entry              46 drivers/net/ethernet/chelsio/cxgb4/cxgb4_tc_u32.c 			     const struct cxgb4_match_field *entry,
entry              71 drivers/net/ethernet/chelsio/cxgb4/cxgb4_tc_u32.c 		for (j = 0; entry[j].val; j++) {
entry              72 drivers/net/ethernet/chelsio/cxgb4/cxgb4_tc_u32.c 			if (off == entry[j].off) {
entry              74 drivers/net/ethernet/chelsio/cxgb4/cxgb4_tc_u32.c 				err = entry[j].val(fs, val, mask);
entry             242 drivers/net/ethernet/chelsio/cxgb4/t4_hw.c 	struct mbox_cmd *entry;
entry             245 drivers/net/ethernet/chelsio/cxgb4/t4_hw.c 	entry = mbox_cmd_log_entry(log, log->cursor++);
entry             250 drivers/net/ethernet/chelsio/cxgb4/t4_hw.c 		entry->cmd[i] = be64_to_cpu(cmd[i]);
entry             252 drivers/net/ethernet/chelsio/cxgb4/t4_hw.c 		entry->cmd[i++] = 0;
entry             253 drivers/net/ethernet/chelsio/cxgb4/t4_hw.c 	entry->timestamp = jiffies;
entry             254 drivers/net/ethernet/chelsio/cxgb4/t4_hw.c 	entry->seqno = log->seqno++;
entry             255 drivers/net/ethernet/chelsio/cxgb4/t4_hw.c 	entry->access = access;
entry             256 drivers/net/ethernet/chelsio/cxgb4/t4_hw.c 	entry->execute = execute;
entry             289 drivers/net/ethernet/chelsio/cxgb4/t4_hw.c 	struct mbox_list entry;
entry             323 drivers/net/ethernet/chelsio/cxgb4/t4_hw.c 	list_add_tail(&entry.list, &adap->mlist.list);
entry             338 drivers/net/ethernet/chelsio/cxgb4/t4_hw.c 			list_del(&entry.list);
entry             349 drivers/net/ethernet/chelsio/cxgb4/t4_hw.c 				     list) == &entry)
entry             371 drivers/net/ethernet/chelsio/cxgb4/t4_hw.c 		list_del(&entry.list);
entry             424 drivers/net/ethernet/chelsio/cxgb4/t4_hw.c 			list_del(&entry.list);
entry             436 drivers/net/ethernet/chelsio/cxgb4/t4_hw.c 	list_del(&entry.list);
entry             247 drivers/net/ethernet/chelsio/cxgb4vf/cxgb4vf_main.c 	struct hash_mac_addr *entry;
entry             250 drivers/net/ethernet/chelsio/cxgb4vf/cxgb4vf_main.c 	list_for_each_entry(entry, &adapter->mac_hlist, list) {
entry             251 drivers/net/ethernet/chelsio/cxgb4vf/cxgb4vf_main.c 		ucast |= is_unicast_ether_addr(entry->addr);
entry             252 drivers/net/ethernet/chelsio/cxgb4vf/cxgb4vf_main.c 		vec |= (1ULL << hash_mac_addr(entry->addr));
entry             277 drivers/net/ethernet/chelsio/cxgb4vf/cxgb4vf_main.c 	struct hash_mac_addr *new_entry, *entry;
entry             287 drivers/net/ethernet/chelsio/cxgb4vf/cxgb4vf_main.c 		list_for_each_entry(entry, &adapter->mac_hlist, list) {
entry             288 drivers/net/ethernet/chelsio/cxgb4vf/cxgb4vf_main.c 			if (entry->iface_mac) {
entry             289 drivers/net/ethernet/chelsio/cxgb4vf/cxgb4vf_main.c 				ether_addr_copy(entry->addr, addr);
entry             975 drivers/net/ethernet/chelsio/cxgb4vf/cxgb4vf_main.c 	struct hash_mac_addr *entry, *tmp;
entry             980 drivers/net/ethernet/chelsio/cxgb4vf/cxgb4vf_main.c 	list_for_each_entry_safe(entry, tmp, &adapter->mac_hlist, list) {
entry             981 drivers/net/ethernet/chelsio/cxgb4vf/cxgb4vf_main.c 		if (ether_addr_equal(entry->addr, mac_addr)) {
entry             982 drivers/net/ethernet/chelsio/cxgb4vf/cxgb4vf_main.c 			list_del(&entry->list);
entry             983 drivers/net/ethernet/chelsio/cxgb4vf/cxgb4vf_main.c 			kfree(entry);
entry            1964 drivers/net/ethernet/chelsio/cxgb4vf/cxgb4vf_main.c 	struct mbox_cmd *entry;
entry            1978 drivers/net/ethernet/chelsio/cxgb4vf/cxgb4vf_main.c 	entry = mbox_cmd_log_entry(log, entry_idx);
entry            1981 drivers/net/ethernet/chelsio/cxgb4vf/cxgb4vf_main.c 	if (entry->timestamp == 0)
entry            1985 drivers/net/ethernet/chelsio/cxgb4vf/cxgb4vf_main.c 		   entry->seqno, entry->timestamp,
entry            1986 drivers/net/ethernet/chelsio/cxgb4vf/cxgb4vf_main.c 		   entry->access, entry->execute);
entry            1988 drivers/net/ethernet/chelsio/cxgb4vf/cxgb4vf_main.c 		u64 flit = entry->cmd[i];
entry            2873 drivers/net/ethernet/chelsio/cxgb4vf/cxgb4vf_main.c 		entries[i].entry = i;
entry            3332 drivers/net/ethernet/chelsio/cxgb4vf/cxgb4vf_main.c 	struct hash_mac_addr *entry, *tmp;
entry            3384 drivers/net/ethernet/chelsio/cxgb4vf/cxgb4vf_main.c 		list_for_each_entry_safe(entry, tmp, &adapter->mac_hlist,
entry            3386 drivers/net/ethernet/chelsio/cxgb4vf/cxgb4vf_main.c 			list_del(&entry->list);
entry            3387 drivers/net/ethernet/chelsio/cxgb4vf/cxgb4vf_main.c 			kfree(entry);
entry              91 drivers/net/ethernet/chelsio/cxgb4vf/t4vf_hw.c 	struct mbox_cmd *entry;
entry              94 drivers/net/ethernet/chelsio/cxgb4vf/t4vf_hw.c 	entry = mbox_cmd_log_entry(log, log->cursor++);
entry              99 drivers/net/ethernet/chelsio/cxgb4vf/t4vf_hw.c 		entry->cmd[i] = be64_to_cpu(cmd[i]);
entry             101 drivers/net/ethernet/chelsio/cxgb4vf/t4vf_hw.c 		entry->cmd[i++] = 0;
entry             102 drivers/net/ethernet/chelsio/cxgb4vf/t4vf_hw.c 	entry->timestamp = jiffies;
entry             103 drivers/net/ethernet/chelsio/cxgb4vf/t4vf_hw.c 	entry->seqno = log->seqno++;
entry             104 drivers/net/ethernet/chelsio/cxgb4vf/t4vf_hw.c 	entry->access = access;
entry             105 drivers/net/ethernet/chelsio/cxgb4vf/t4vf_hw.c 	entry->execute = execute;
entry             142 drivers/net/ethernet/chelsio/cxgb4vf/t4vf_hw.c 	struct mbox_list entry;
entry             166 drivers/net/ethernet/chelsio/cxgb4vf/t4vf_hw.c 	list_add_tail(&entry.list, &adapter->mlist.list);
entry             180 drivers/net/ethernet/chelsio/cxgb4vf/t4vf_hw.c 			list_del(&entry.list);
entry             191 drivers/net/ethernet/chelsio/cxgb4vf/t4vf_hw.c 				     list) == &entry)
entry             214 drivers/net/ethernet/chelsio/cxgb4vf/t4vf_hw.c 		list_del(&entry.list);
entry             301 drivers/net/ethernet/chelsio/cxgb4vf/t4vf_hw.c 			list_del(&entry.list);
entry             311 drivers/net/ethernet/chelsio/cxgb4vf/t4vf_hw.c 	list_del(&entry.list);
entry             233 drivers/net/ethernet/cirrus/ep93xx_eth.c 		int entry;
entry             240 drivers/net/ethernet/cirrus/ep93xx_eth.c 		entry = ep->rx_pointer;
entry             241 drivers/net/ethernet/cirrus/ep93xx_eth.c 		rstat = ep->descs->rstat + entry;
entry             255 drivers/net/ethernet/cirrus/ep93xx_eth.c 		if ((rstat1 & RSTAT1_BUFFER_INDEX) >> 16 != entry)
entry             283 drivers/net/ethernet/cirrus/ep93xx_eth.c 			struct ep93xx_rdesc *rxd = &ep->descs->rdesc[entry];
entry             287 drivers/net/ethernet/cirrus/ep93xx_eth.c 			skb_copy_to_linear_data(skb, ep->rx_buf[entry], length);
entry             303 drivers/net/ethernet/cirrus/ep93xx_eth.c 		ep->rx_pointer = (entry + 1) & (RX_QUEUE_ENTRIES - 1);
entry             335 drivers/net/ethernet/cirrus/ep93xx_eth.c 	int entry;
entry             343 drivers/net/ethernet/cirrus/ep93xx_eth.c 	entry = ep->tx_pointer;
entry             346 drivers/net/ethernet/cirrus/ep93xx_eth.c 	txd = &ep->descs->tdesc[entry];
entry             348 drivers/net/ethernet/cirrus/ep93xx_eth.c 	txd->tdesc1 = TDESC1_EOF | (entry << 16) | (skb->len & 0xfff);
entry             351 drivers/net/ethernet/cirrus/ep93xx_eth.c 	skb_copy_and_csum_dev(skb, ep->tx_buf[entry]);
entry             376 drivers/net/ethernet/cirrus/ep93xx_eth.c 		int entry;
entry             380 drivers/net/ethernet/cirrus/ep93xx_eth.c 		entry = ep->tx_clean_pointer;
entry             381 drivers/net/ethernet/cirrus/ep93xx_eth.c 		tstat = ep->descs->tstat + entry;
entry             391 drivers/net/ethernet/cirrus/ep93xx_eth.c 		if ((tstat0 & TSTAT0_BUFFER_INDEX) != entry)
entry             395 drivers/net/ethernet/cirrus/ep93xx_eth.c 			int length = ep->descs->tdesc[entry].tdesc1 & 0xfff;
entry             409 drivers/net/ethernet/cirrus/ep93xx_eth.c 		ep->tx_clean_pointer = (entry + 1) & (TX_QUEUE_ENTRIES - 1);
entry            2396 drivers/net/ethernet/cisco/enic/enic_main.c 		enic->msix_entry[i].entry = i;
entry            1205 drivers/net/ethernet/cisco/enic/vnic_dev.c int vnic_dev_classifier(struct vnic_dev *vdev, u8 cmd, u16 *entry,
entry            1239 drivers/net/ethernet/cisco/enic/vnic_dev.c 		action->u.rq_idx = *entry;
entry            1242 drivers/net/ethernet/cisco/enic/vnic_dev.c 		*entry = (u16)a0;
entry            1245 drivers/net/ethernet/cisco/enic/vnic_dev.c 		a0 = *entry;
entry             179 drivers/net/ethernet/cisco/enic/vnic_dev.h int vnic_dev_classifier(struct vnic_dev *vdev, u8 cmd, u16 *entry,
entry             607 drivers/net/ethernet/dec/tulip/de2104x.c 	unsigned int entry, tx_free;
entry             621 drivers/net/ethernet/dec/tulip/de2104x.c 	entry = de->tx_head;
entry             623 drivers/net/ethernet/dec/tulip/de2104x.c 	txd = &de->tx_ring[entry];
entry             627 drivers/net/ethernet/dec/tulip/de2104x.c 	if (entry == (DE_TX_RING_SIZE - 1))
entry             635 drivers/net/ethernet/dec/tulip/de2104x.c 	de->tx_skb[entry].skb = skb;
entry             636 drivers/net/ethernet/dec/tulip/de2104x.c 	de->tx_skb[entry].mapping = mapping;
entry             642 drivers/net/ethernet/dec/tulip/de2104x.c 	de->tx_head = NEXT_TX(entry);
entry             644 drivers/net/ethernet/dec/tulip/de2104x.c 		  entry, skb->len);
entry             722 drivers/net/ethernet/dec/tulip/de2104x.c 	unsigned int entry;
entry             751 drivers/net/ethernet/dec/tulip/de2104x.c 	entry = de->tx_head;
entry             754 drivers/net/ethernet/dec/tulip/de2104x.c 	if (entry != 0) {
entry             755 drivers/net/ethernet/dec/tulip/de2104x.c 		de->tx_skb[entry].skb = DE_DUMMY_SKB;
entry             757 drivers/net/ethernet/dec/tulip/de2104x.c 		dummy_txd = &de->tx_ring[entry];
entry             758 drivers/net/ethernet/dec/tulip/de2104x.c 		dummy_txd->opts2 = (entry == (DE_TX_RING_SIZE - 1)) ?
entry             764 drivers/net/ethernet/dec/tulip/de2104x.c 		entry = NEXT_TX(entry);
entry             767 drivers/net/ethernet/dec/tulip/de2104x.c 	de->tx_skb[entry].skb = DE_SETUP_SKB;
entry             768 drivers/net/ethernet/dec/tulip/de2104x.c 	de->tx_skb[entry].mapping = mapping =
entry             773 drivers/net/ethernet/dec/tulip/de2104x.c 	txd = &de->tx_ring[entry];
entry             774 drivers/net/ethernet/dec/tulip/de2104x.c 	if (entry == (DE_TX_RING_SIZE - 1))
entry             789 drivers/net/ethernet/dec/tulip/de2104x.c 	de->tx_head = NEXT_TX(entry);
entry            1605 drivers/net/ethernet/dec/tulip/de4x5.c     int entry;
entry            1608 drivers/net/ethernet/dec/tulip/de4x5.c     for (entry=lp->rx_new; (s32)le32_to_cpu(lp->rx_ring[entry].status)>=0;
entry            1609 drivers/net/ethernet/dec/tulip/de4x5.c 	                                                    entry=lp->rx_new) {
entry            1610 drivers/net/ethernet/dec/tulip/de4x5.c 	status = (s32)le32_to_cpu(lp->rx_ring[entry].status);
entry            1620 drivers/net/ethernet/dec/tulip/de4x5.c 	    lp->rx_old = entry;
entry            1637 drivers/net/ethernet/dec/tulip/de4x5.c 		short pkt_len = (short)(le32_to_cpu(lp->rx_ring[entry].status)
entry            1640 drivers/net/ethernet/dec/tulip/de4x5.c 		if ((skb = de4x5_alloc_rx_buff(dev, entry, pkt_len)) == NULL) {
entry            1659 drivers/net/ethernet/dec/tulip/de4x5.c 	    for (;lp->rx_old!=entry;lp->rx_old=(lp->rx_old + 1)%lp->rxRingSize) {
entry            1663 drivers/net/ethernet/dec/tulip/de4x5.c 	    lp->rx_ring[entry].status = cpu_to_le32(R_OWN);
entry            1677 drivers/net/ethernet/dec/tulip/de4x5.c de4x5_free_tx_buff(struct de4x5_private *lp, int entry)
entry            1679 drivers/net/ethernet/dec/tulip/de4x5.c     dma_unmap_single(lp->gendev, le32_to_cpu(lp->tx_ring[entry].buf),
entry            1680 drivers/net/ethernet/dec/tulip/de4x5.c 		     le32_to_cpu(lp->tx_ring[entry].des1) & TD_TBS1,
entry            1682 drivers/net/ethernet/dec/tulip/de4x5.c     if ((u_long) lp->tx_skb[entry] > 1)
entry            1683 drivers/net/ethernet/dec/tulip/de4x5.c 	dev_kfree_skb_irq(lp->tx_skb[entry]);
entry            1684 drivers/net/ethernet/dec/tulip/de4x5.c     lp->tx_skb[entry] = NULL;
entry            1695 drivers/net/ethernet/dec/tulip/de4x5.c     int entry;
entry            1698 drivers/net/ethernet/dec/tulip/de4x5.c     for (entry = lp->tx_old; entry != lp->tx_new; entry = lp->tx_old) {
entry            1699 drivers/net/ethernet/dec/tulip/de4x5.c 	status = (s32)le32_to_cpu(lp->tx_ring[entry].status);
entry            1723 drivers/net/ethernet/dec/tulip/de4x5.c 	    if (lp->tx_skb[entry] != NULL)
entry            1724 drivers/net/ethernet/dec/tulip/de4x5.c 	    	de4x5_free_tx_buff(lp, entry);
entry            1899 drivers/net/ethernet/dec/tulip/de4x5.c     int entry = (lp->tx_new ? lp->tx_new-1 : lp->txRingSize-1);
entry            1906 drivers/net/ethernet/dec/tulip/de4x5.c     lp->tx_ring[entry].des1 &= cpu_to_le32(~TD_IC);
entry              62 drivers/net/ethernet/dec/tulip/interrupt.c 	int entry;
entry              67 drivers/net/ethernet/dec/tulip/interrupt.c 		entry = tp->dirty_rx % RX_RING_SIZE;
entry              68 drivers/net/ethernet/dec/tulip/interrupt.c 		if (tp->rx_buffers[entry].skb == NULL) {
entry              72 drivers/net/ethernet/dec/tulip/interrupt.c 			skb = tp->rx_buffers[entry].skb =
entry              81 drivers/net/ethernet/dec/tulip/interrupt.c 				tp->rx_buffers[entry].skb = NULL;
entry              85 drivers/net/ethernet/dec/tulip/interrupt.c 			tp->rx_buffers[entry].mapping = mapping;
entry              87 drivers/net/ethernet/dec/tulip/interrupt.c 			tp->rx_ring[entry].buffer1 = cpu_to_le32(mapping);
entry              90 drivers/net/ethernet/dec/tulip/interrupt.c 		tp->rx_ring[entry].status = cpu_to_le32(DescOwned);
entry             116 drivers/net/ethernet/dec/tulip/interrupt.c 	int entry = tp->cur_rx % RX_RING_SIZE;
entry             132 drivers/net/ethernet/dec/tulip/interrupt.c 			   entry, tp->rx_ring[entry].status);
entry             144 drivers/net/ethernet/dec/tulip/interrupt.c                while ( ! (tp->rx_ring[entry].status & cpu_to_le32(DescOwned))) {
entry             145 drivers/net/ethernet/dec/tulip/interrupt.c                        s32 status = le32_to_cpu(tp->rx_ring[entry].status);
entry             153 drivers/net/ethernet/dec/tulip/interrupt.c 					   entry, status);
entry             214 drivers/net/ethernet/dec/tulip/interrupt.c 								   tp->rx_buffers[entry].mapping,
entry             217 drivers/net/ethernet/dec/tulip/interrupt.c                                        skb_copy_to_linear_data(skb, tp->rx_buffers[entry].skb->data,
entry             222 drivers/net/ethernet/dec/tulip/interrupt.c                                                     tp->rx_buffers[entry].skb->data,
entry             226 drivers/net/ethernet/dec/tulip/interrupt.c 								      tp->rx_buffers[entry].mapping,
entry             229 drivers/net/ethernet/dec/tulip/interrupt.c                                        char *temp = skb_put(skb = tp->rx_buffers[entry].skb,
entry             233 drivers/net/ethernet/dec/tulip/interrupt.c                                        if (tp->rx_buffers[entry].mapping !=
entry             234 drivers/net/ethernet/dec/tulip/interrupt.c                                            le32_to_cpu(tp->rx_ring[entry].buffer1)) {
entry             237 drivers/net/ethernet/dec/tulip/interrupt.c 						       le32_to_cpu(tp->rx_ring[entry].buffer1),
entry             238 drivers/net/ethernet/dec/tulip/interrupt.c 						       (unsigned long long)tp->rx_buffers[entry].mapping,
entry             243 drivers/net/ethernet/dec/tulip/interrupt.c                                        pci_unmap_single(tp->pdev, tp->rx_buffers[entry].mapping,
entry             246 drivers/net/ethernet/dec/tulip/interrupt.c                                        tp->rx_buffers[entry].skb = NULL;
entry             247 drivers/net/ethernet/dec/tulip/interrupt.c                                        tp->rx_buffers[entry].mapping = 0;
entry             260 drivers/net/ethernet/dec/tulip/interrupt.c                        entry = (++tp->cur_rx) % RX_RING_SIZE;
entry             368 drivers/net/ethernet/dec/tulip/interrupt.c 	int entry = tp->cur_rx % RX_RING_SIZE;
entry             374 drivers/net/ethernet/dec/tulip/interrupt.c 			   entry, tp->rx_ring[entry].status);
entry             376 drivers/net/ethernet/dec/tulip/interrupt.c 	while ( ! (tp->rx_ring[entry].status & cpu_to_le32(DescOwned))) {
entry             377 drivers/net/ethernet/dec/tulip/interrupt.c 		s32 status = le32_to_cpu(tp->rx_ring[entry].status);
entry             382 drivers/net/ethernet/dec/tulip/interrupt.c 				   entry, status);
entry             440 drivers/net/ethernet/dec/tulip/interrupt.c 							    tp->rx_buffers[entry].mapping,
entry             443 drivers/net/ethernet/dec/tulip/interrupt.c 				skb_copy_to_linear_data(skb, tp->rx_buffers[entry].skb->data,
entry             448 drivers/net/ethernet/dec/tulip/interrupt.c 					     tp->rx_buffers[entry].skb->data,
entry             452 drivers/net/ethernet/dec/tulip/interrupt.c 							       tp->rx_buffers[entry].mapping,
entry             455 drivers/net/ethernet/dec/tulip/interrupt.c 				char *temp = skb_put(skb = tp->rx_buffers[entry].skb,
entry             459 drivers/net/ethernet/dec/tulip/interrupt.c 				if (tp->rx_buffers[entry].mapping !=
entry             460 drivers/net/ethernet/dec/tulip/interrupt.c 				    le32_to_cpu(tp->rx_ring[entry].buffer1)) {
entry             463 drivers/net/ethernet/dec/tulip/interrupt.c 						le32_to_cpu(tp->rx_ring[entry].buffer1),
entry             464 drivers/net/ethernet/dec/tulip/interrupt.c 						(long long)tp->rx_buffers[entry].mapping,
entry             469 drivers/net/ethernet/dec/tulip/interrupt.c 				pci_unmap_single(tp->pdev, tp->rx_buffers[entry].mapping,
entry             472 drivers/net/ethernet/dec/tulip/interrupt.c 				tp->rx_buffers[entry].skb = NULL;
entry             473 drivers/net/ethernet/dec/tulip/interrupt.c 				tp->rx_buffers[entry].mapping = 0;
entry             483 drivers/net/ethernet/dec/tulip/interrupt.c 		entry = (++tp->cur_rx) % RX_RING_SIZE;
entry             531 drivers/net/ethernet/dec/tulip/interrupt.c 	int entry;
entry             590 drivers/net/ethernet/dec/tulip/interrupt.c 				int entry = dirty_tx % TX_RING_SIZE;
entry             591 drivers/net/ethernet/dec/tulip/interrupt.c 				int status = le32_to_cpu(tp->tx_ring[entry].status);
entry             597 drivers/net/ethernet/dec/tulip/interrupt.c 				if (tp->tx_buffers[entry].skb == NULL) {
entry             599 drivers/net/ethernet/dec/tulip/interrupt.c 					if (tp->tx_buffers[entry].mapping)
entry             601 drivers/net/ethernet/dec/tulip/interrupt.c 							 tp->tx_buffers[entry].mapping,
entry             627 drivers/net/ethernet/dec/tulip/interrupt.c 						tp->tx_buffers[entry].skb->len;
entry             632 drivers/net/ethernet/dec/tulip/interrupt.c 				pci_unmap_single(tp->pdev, tp->tx_buffers[entry].mapping,
entry             633 drivers/net/ethernet/dec/tulip/interrupt.c 						 tp->tx_buffers[entry].skb->len,
entry             637 drivers/net/ethernet/dec/tulip/interrupt.c 				dev_kfree_skb_irq(tp->tx_buffers[entry].skb);
entry             638 drivers/net/ethernet/dec/tulip/interrupt.c 				tp->tx_buffers[entry].skb = NULL;
entry             639 drivers/net/ethernet/dec/tulip/interrupt.c 				tp->tx_buffers[entry].mapping = 0;
entry             780 drivers/net/ethernet/dec/tulip/interrupt.c 	entry = tp->dirty_rx % RX_RING_SIZE;
entry             781 drivers/net/ethernet/dec/tulip/interrupt.c 	if (tp->rx_buffers[entry].skb == NULL) {
entry             667 drivers/net/ethernet/dec/tulip/tulip_core.c 	int entry;
entry             675 drivers/net/ethernet/dec/tulip/tulip_core.c 	entry = tp->cur_tx % TX_RING_SIZE;
entry             677 drivers/net/ethernet/dec/tulip/tulip_core.c 	tp->tx_buffers[entry].skb = skb;
entry             680 drivers/net/ethernet/dec/tulip/tulip_core.c 	tp->tx_buffers[entry].mapping = mapping;
entry             681 drivers/net/ethernet/dec/tulip/tulip_core.c 	tp->tx_ring[entry].buffer1 = cpu_to_le32(mapping);
entry             693 drivers/net/ethernet/dec/tulip/tulip_core.c 	if (entry == TX_RING_SIZE-1)
entry             696 drivers/net/ethernet/dec/tulip/tulip_core.c 	tp->tx_ring[entry].length = cpu_to_le32(skb->len | flag);
entry             699 drivers/net/ethernet/dec/tulip/tulip_core.c 	tp->tx_ring[entry].status = cpu_to_le32(DescOwned);
entry             718 drivers/net/ethernet/dec/tulip/tulip_core.c 		int entry = dirty_tx % TX_RING_SIZE;
entry             719 drivers/net/ethernet/dec/tulip/tulip_core.c 		int status = le32_to_cpu(tp->tx_ring[entry].status);
entry             723 drivers/net/ethernet/dec/tulip/tulip_core.c 			tp->tx_ring[entry].status = 0;
entry             727 drivers/net/ethernet/dec/tulip/tulip_core.c 		if (tp->tx_buffers[entry].skb == NULL) {
entry             729 drivers/net/ethernet/dec/tulip/tulip_core.c 			if (tp->tx_buffers[entry].mapping)
entry             731 drivers/net/ethernet/dec/tulip/tulip_core.c 					tp->tx_buffers[entry].mapping,
entry             737 drivers/net/ethernet/dec/tulip/tulip_core.c 		pci_unmap_single(tp->pdev, tp->tx_buffers[entry].mapping,
entry             738 drivers/net/ethernet/dec/tulip/tulip_core.c 				tp->tx_buffers[entry].skb->len,
entry             742 drivers/net/ethernet/dec/tulip/tulip_core.c 		dev_kfree_skb_irq(tp->tx_buffers[entry].skb);
entry             743 drivers/net/ethernet/dec/tulip/tulip_core.c 		tp->tx_buffers[entry].skb = NULL;
entry             744 drivers/net/ethernet/dec/tulip/tulip_core.c 		tp->tx_buffers[entry].mapping = 0;
entry            1142 drivers/net/ethernet/dec/tulip/tulip_core.c 			unsigned int entry;
entry            1147 drivers/net/ethernet/dec/tulip/tulip_core.c 			entry = tp->cur_tx++ % TX_RING_SIZE;
entry            1149 drivers/net/ethernet/dec/tulip/tulip_core.c 			if (entry != 0) {
entry            1151 drivers/net/ethernet/dec/tulip/tulip_core.c 				tp->tx_buffers[entry].skb = NULL;
entry            1152 drivers/net/ethernet/dec/tulip/tulip_core.c 				tp->tx_buffers[entry].mapping = 0;
entry            1153 drivers/net/ethernet/dec/tulip/tulip_core.c 				tp->tx_ring[entry].length =
entry            1154 drivers/net/ethernet/dec/tulip/tulip_core.c 					(entry == TX_RING_SIZE-1) ? cpu_to_le32(DESC_RING_WRAP) : 0;
entry            1155 drivers/net/ethernet/dec/tulip/tulip_core.c 				tp->tx_ring[entry].buffer1 = 0;
entry            1157 drivers/net/ethernet/dec/tulip/tulip_core.c 				dummy = entry;
entry            1158 drivers/net/ethernet/dec/tulip/tulip_core.c 				entry = tp->cur_tx++ % TX_RING_SIZE;
entry            1162 drivers/net/ethernet/dec/tulip/tulip_core.c 			tp->tx_buffers[entry].skb = NULL;
entry            1163 drivers/net/ethernet/dec/tulip/tulip_core.c 			tp->tx_buffers[entry].mapping =
entry            1168 drivers/net/ethernet/dec/tulip/tulip_core.c 			if (entry == TX_RING_SIZE-1)
entry            1170 drivers/net/ethernet/dec/tulip/tulip_core.c 			tp->tx_ring[entry].length = cpu_to_le32(tx_flags);
entry            1171 drivers/net/ethernet/dec/tulip/tulip_core.c 			tp->tx_ring[entry].buffer1 =
entry            1172 drivers/net/ethernet/dec/tulip/tulip_core.c 				cpu_to_le32(tp->tx_buffers[entry].mapping);
entry            1173 drivers/net/ethernet/dec/tulip/tulip_core.c 			tp->tx_ring[entry].status = cpu_to_le32(DescOwned);
entry             999 drivers/net/ethernet/dec/tulip/winbond-840.c 	unsigned entry;
entry            1005 drivers/net/ethernet/dec/tulip/winbond-840.c 	entry = np->cur_tx % TX_RING_SIZE;
entry            1007 drivers/net/ethernet/dec/tulip/winbond-840.c 	np->tx_addr[entry] = pci_map_single(np->pci_dev,
entry            1009 drivers/net/ethernet/dec/tulip/winbond-840.c 	np->tx_skbuff[entry] = skb;
entry            1011 drivers/net/ethernet/dec/tulip/winbond-840.c 	np->tx_ring[entry].buffer1 = np->tx_addr[entry];
entry            1013 drivers/net/ethernet/dec/tulip/winbond-840.c 		np->tx_ring[entry].length = DescWholePkt | skb->len;
entry            1017 drivers/net/ethernet/dec/tulip/winbond-840.c 		np->tx_ring[entry].buffer2 = np->tx_addr[entry]+TX_BUFLIMIT;
entry            1018 drivers/net/ethernet/dec/tulip/winbond-840.c 		np->tx_ring[entry].length = DescWholePkt | (len << 11) | TX_BUFLIMIT;
entry            1020 drivers/net/ethernet/dec/tulip/winbond-840.c 	if(entry == TX_RING_SIZE-1)
entry            1021 drivers/net/ethernet/dec/tulip/winbond-840.c 		np->tx_ring[entry].length |= DescEndRing;
entry            1037 drivers/net/ethernet/dec/tulip/winbond-840.c 	np->tx_ring[entry].status = DescOwned;
entry            1053 drivers/net/ethernet/dec/tulip/winbond-840.c 			   np->cur_tx, entry);
entry            1062 drivers/net/ethernet/dec/tulip/winbond-840.c 		int entry = np->dirty_tx % TX_RING_SIZE;
entry            1063 drivers/net/ethernet/dec/tulip/winbond-840.c 		int tx_status = np->tx_ring[entry].status;
entry            1084 drivers/net/ethernet/dec/tulip/winbond-840.c 					   entry, tx_status);
entry            1086 drivers/net/ethernet/dec/tulip/winbond-840.c 			np->stats.tx_bytes += np->tx_skbuff[entry]->len;
entry            1091 drivers/net/ethernet/dec/tulip/winbond-840.c 		pci_unmap_single(np->pci_dev,np->tx_addr[entry],
entry            1092 drivers/net/ethernet/dec/tulip/winbond-840.c 					np->tx_skbuff[entry]->len,
entry            1094 drivers/net/ethernet/dec/tulip/winbond-840.c 		np->tx_q_bytes -= np->tx_skbuff[entry]->len;
entry            1095 drivers/net/ethernet/dec/tulip/winbond-840.c 		dev_kfree_skb_irq(np->tx_skbuff[entry]);
entry            1096 drivers/net/ethernet/dec/tulip/winbond-840.c 		np->tx_skbuff[entry] = NULL;
entry            1178 drivers/net/ethernet/dec/tulip/winbond-840.c 	int entry = np->cur_rx % RX_RING_SIZE;
entry            1183 drivers/net/ethernet/dec/tulip/winbond-840.c 			   entry, np->rx_ring[entry].status);
entry            1230 drivers/net/ethernet/dec/tulip/winbond-840.c 				pci_dma_sync_single_for_cpu(np->pci_dev,np->rx_addr[entry],
entry            1231 drivers/net/ethernet/dec/tulip/winbond-840.c 							    np->rx_skbuff[entry]->len,
entry            1233 drivers/net/ethernet/dec/tulip/winbond-840.c 				skb_copy_to_linear_data(skb, np->rx_skbuff[entry]->data, pkt_len);
entry            1235 drivers/net/ethernet/dec/tulip/winbond-840.c 				pci_dma_sync_single_for_device(np->pci_dev,np->rx_addr[entry],
entry            1236 drivers/net/ethernet/dec/tulip/winbond-840.c 							       np->rx_skbuff[entry]->len,
entry            1239 drivers/net/ethernet/dec/tulip/winbond-840.c 				pci_unmap_single(np->pci_dev,np->rx_addr[entry],
entry            1240 drivers/net/ethernet/dec/tulip/winbond-840.c 							np->rx_skbuff[entry]->len,
entry            1242 drivers/net/ethernet/dec/tulip/winbond-840.c 				skb_put(skb = np->rx_skbuff[entry], pkt_len);
entry            1243 drivers/net/ethernet/dec/tulip/winbond-840.c 				np->rx_skbuff[entry] = NULL;
entry            1258 drivers/net/ethernet/dec/tulip/winbond-840.c 		entry = (++np->cur_rx) % RX_RING_SIZE;
entry            1259 drivers/net/ethernet/dec/tulip/winbond-840.c 		np->rx_head_desc = &np->rx_ring[entry];
entry            1265 drivers/net/ethernet/dec/tulip/winbond-840.c 		entry = np->dirty_rx % RX_RING_SIZE;
entry            1266 drivers/net/ethernet/dec/tulip/winbond-840.c 		if (np->rx_skbuff[entry] == NULL) {
entry            1268 drivers/net/ethernet/dec/tulip/winbond-840.c 			np->rx_skbuff[entry] = skb;
entry            1271 drivers/net/ethernet/dec/tulip/winbond-840.c 			np->rx_addr[entry] = pci_map_single(np->pci_dev,
entry            1274 drivers/net/ethernet/dec/tulip/winbond-840.c 			np->rx_ring[entry].buffer1 = np->rx_addr[entry];
entry            1277 drivers/net/ethernet/dec/tulip/winbond-840.c 		np->rx_ring[entry].status = DescOwned;
entry             659 drivers/net/ethernet/dlink/dl2k.c 	unsigned int entry;
entry             670 drivers/net/ethernet/dlink/dl2k.c 			entry = np->old_rx % RX_RING_SIZE;
entry             672 drivers/net/ethernet/dlink/dl2k.c 			if (np->rx_skbuff[entry] == NULL) {
entry             676 drivers/net/ethernet/dlink/dl2k.c 					np->rx_ring[entry].fraginfo = 0;
entry             679 drivers/net/ethernet/dlink/dl2k.c 						dev->name, entry);
entry             682 drivers/net/ethernet/dlink/dl2k.c 				np->rx_skbuff[entry] = skb;
entry             683 drivers/net/ethernet/dlink/dl2k.c 				np->rx_ring[entry].fraginfo =
entry             688 drivers/net/ethernet/dlink/dl2k.c 			np->rx_ring[entry].fraginfo |=
entry             690 drivers/net/ethernet/dlink/dl2k.c 			np->rx_ring[entry].status = 0;
entry             717 drivers/net/ethernet/dlink/dl2k.c 	unsigned entry;
entry             724 drivers/net/ethernet/dlink/dl2k.c 	entry = np->cur_tx % TX_RING_SIZE;
entry             725 drivers/net/ethernet/dlink/dl2k.c 	np->tx_skbuff[entry] = skb;
entry             726 drivers/net/ethernet/dlink/dl2k.c 	txdesc = &np->tx_ring[entry];
entry             747 drivers/net/ethernet/dlink/dl2k.c 	if (entry % np->tx_coalesce == 0 || np->speed == 10)
entry             748 drivers/net/ethernet/dlink/dl2k.c 		txdesc->status = cpu_to_le64 (entry | tfc_vlan_tag |
entry             753 drivers/net/ethernet/dlink/dl2k.c 		txdesc->status = cpu_to_le64 (entry | tfc_vlan_tag |
entry             772 drivers/net/ethernet/dlink/dl2k.c 		     entry * sizeof (struct netdev_desc));
entry             823 drivers/net/ethernet/dlink/dl2k.c 	int entry = np->old_tx % TX_RING_SIZE;
entry             833 drivers/net/ethernet/dlink/dl2k.c 	while (entry != np->cur_tx) {
entry             836 drivers/net/ethernet/dlink/dl2k.c 		if (!(np->tx_ring[entry].status & cpu_to_le64(TFDDone)))
entry             838 drivers/net/ethernet/dlink/dl2k.c 		skb = np->tx_skbuff[entry];
entry             840 drivers/net/ethernet/dlink/dl2k.c 				  desc_to_dma(&np->tx_ring[entry]),
entry             847 drivers/net/ethernet/dlink/dl2k.c 		np->tx_skbuff[entry] = NULL;
entry             848 drivers/net/ethernet/dlink/dl2k.c 		entry = (entry + 1) % TX_RING_SIZE;
entry             855 drivers/net/ethernet/dlink/dl2k.c 	np->old_tx = entry;
entry             926 drivers/net/ethernet/dlink/dl2k.c 	int entry = np->cur_rx % RX_RING_SIZE;
entry             931 drivers/net/ethernet/dlink/dl2k.c 		struct netdev_desc *desc = &np->rx_ring[entry];
entry             965 drivers/net/ethernet/dlink/dl2k.c 				skb_put (skb = np->rx_skbuff[entry], pkt_len);
entry             966 drivers/net/ethernet/dlink/dl2k.c 				np->rx_skbuff[entry] = NULL;
entry             973 drivers/net/ethernet/dlink/dl2k.c 						  np->rx_skbuff[entry]->data,
entry             991 drivers/net/ethernet/dlink/dl2k.c 		entry = (entry + 1) % RX_RING_SIZE;
entry             994 drivers/net/ethernet/dlink/dl2k.c 	np->cur_rx = entry;
entry             996 drivers/net/ethernet/dlink/dl2k.c 	entry = np->old_rx;
entry             997 drivers/net/ethernet/dlink/dl2k.c 	while (entry != np->cur_rx) {
entry            1000 drivers/net/ethernet/dlink/dl2k.c 		if (np->rx_skbuff[entry] == NULL) {
entry            1003 drivers/net/ethernet/dlink/dl2k.c 				np->rx_ring[entry].fraginfo = 0;
entry            1007 drivers/net/ethernet/dlink/dl2k.c 					dev->name, entry);
entry            1010 drivers/net/ethernet/dlink/dl2k.c 			np->rx_skbuff[entry] = skb;
entry            1011 drivers/net/ethernet/dlink/dl2k.c 			np->rx_ring[entry].fraginfo =
entry            1016 drivers/net/ethernet/dlink/dl2k.c 		np->rx_ring[entry].fraginfo |=
entry            1018 drivers/net/ethernet/dlink/dl2k.c 		np->rx_ring[entry].status = 0;
entry            1019 drivers/net/ethernet/dlink/dl2k.c 		entry = (entry + 1) % RX_RING_SIZE;
entry            1021 drivers/net/ethernet/dlink/dl2k.c 	np->old_rx = entry;
entry            1082 drivers/net/ethernet/dlink/sundance.c 		int entry = np->cur_task % TX_RING_SIZE;
entry            1083 drivers/net/ethernet/dlink/sundance.c 		txdesc = &np->tx_ring[entry];
entry            1086 drivers/net/ethernet/dlink/sundance.c 				entry*sizeof(struct netdev_desc));
entry            1103 drivers/net/ethernet/dlink/sundance.c 	unsigned entry;
entry            1106 drivers/net/ethernet/dlink/sundance.c 	entry = np->cur_tx % TX_RING_SIZE;
entry            1107 drivers/net/ethernet/dlink/sundance.c 	np->tx_skbuff[entry] = skb;
entry            1108 drivers/net/ethernet/dlink/sundance.c 	txdesc = &np->tx_ring[entry];
entry            1111 drivers/net/ethernet/dlink/sundance.c 	txdesc->status = cpu_to_le32 ((entry << 2) | DisableAlign);
entry            1135 drivers/net/ethernet/dlink/sundance.c 			dev->name, np->cur_tx, entry);
entry            1141 drivers/net/ethernet/dlink/sundance.c 	np->tx_skbuff[entry] = NULL;
entry            1271 drivers/net/ethernet/dlink/sundance.c 				int entry = np->dirty_tx % TX_RING_SIZE;
entry            1275 drivers/net/ethernet/dlink/sundance.c 					np->tx_ring[entry].status) >> 2) & 0xff;
entry            1277 drivers/net/ethernet/dlink/sundance.c 					!(le32_to_cpu(np->tx_ring[entry].status)
entry            1283 drivers/net/ethernet/dlink/sundance.c 				skb = np->tx_skbuff[entry];
entry            1286 drivers/net/ethernet/dlink/sundance.c 					le32_to_cpu(np->tx_ring[entry].frag[0].addr),
entry            1288 drivers/net/ethernet/dlink/sundance.c 				dev_consume_skb_irq(np->tx_skbuff[entry]);
entry            1289 drivers/net/ethernet/dlink/sundance.c 				np->tx_skbuff[entry] = NULL;
entry            1290 drivers/net/ethernet/dlink/sundance.c 				np->tx_ring[entry].frag[0].addr = 0;
entry            1291 drivers/net/ethernet/dlink/sundance.c 				np->tx_ring[entry].frag[0].length = 0;
entry            1297 drivers/net/ethernet/dlink/sundance.c 				int entry = np->dirty_tx % TX_RING_SIZE;
entry            1299 drivers/net/ethernet/dlink/sundance.c 				if (!(le32_to_cpu(np->tx_ring[entry].status)
entry            1302 drivers/net/ethernet/dlink/sundance.c 				skb = np->tx_skbuff[entry];
entry            1305 drivers/net/ethernet/dlink/sundance.c 					le32_to_cpu(np->tx_ring[entry].frag[0].addr),
entry            1307 drivers/net/ethernet/dlink/sundance.c 				dev_consume_skb_irq(np->tx_skbuff[entry]);
entry            1308 drivers/net/ethernet/dlink/sundance.c 				np->tx_skbuff[entry] = NULL;
entry            1309 drivers/net/ethernet/dlink/sundance.c 				np->tx_ring[entry].frag[0].addr = 0;
entry            1310 drivers/net/ethernet/dlink/sundance.c 				np->tx_ring[entry].frag[0].length = 0;
entry            1334 drivers/net/ethernet/dlink/sundance.c 	int entry = np->cur_rx % RX_RING_SIZE;
entry            1341 drivers/net/ethernet/dlink/sundance.c 		struct netdev_desc *desc = &(np->rx_ring[entry]);
entry            1389 drivers/net/ethernet/dlink/sundance.c 				skb_copy_to_linear_data(skb, np->rx_skbuff[entry]->data, pkt_len);
entry            1398 drivers/net/ethernet/dlink/sundance.c 				skb_put(skb = np->rx_skbuff[entry], pkt_len);
entry            1399 drivers/net/ethernet/dlink/sundance.c 				np->rx_skbuff[entry] = NULL;
entry            1405 drivers/net/ethernet/dlink/sundance.c 		entry = (entry + 1) % RX_RING_SIZE;
entry            1408 drivers/net/ethernet/dlink/sundance.c 	np->cur_rx = entry;
entry            1415 drivers/net/ethernet/dlink/sundance.c 	np->cur_rx = entry;
entry            1428 drivers/net/ethernet/dlink/sundance.c 	int entry;
entry            1435 drivers/net/ethernet/dlink/sundance.c 		entry = np->dirty_rx % RX_RING_SIZE;
entry            1436 drivers/net/ethernet/dlink/sundance.c 		if (np->rx_skbuff[entry] == NULL) {
entry            1438 drivers/net/ethernet/dlink/sundance.c 			np->rx_skbuff[entry] = skb;
entry            1442 drivers/net/ethernet/dlink/sundance.c 			np->rx_ring[entry].frag[0].addr = cpu_to_le32(
entry            1446 drivers/net/ethernet/dlink/sundance.c 				    np->rx_ring[entry].frag[0].addr)) {
entry            1448 drivers/net/ethernet/dlink/sundance.c 			    np->rx_skbuff[entry] = NULL;
entry            1453 drivers/net/ethernet/dlink/sundance.c 		np->rx_ring[entry].frag[0].length =
entry            1455 drivers/net/ethernet/dlink/sundance.c 		np->rx_ring[entry].status = 0;
entry             641 drivers/net/ethernet/emulex/benet/be.h 	struct list_head entry;
entry            1428 drivers/net/ethernet/emulex/benet/be_main.c 	u32 *entry;
entry            1437 drivers/net/ethernet/emulex/benet/be_main.c 		entry = txo->q.dma_mem.va;
entry            1439 drivers/net/ethernet/emulex/benet/be_main.c 			if (entry[j] != 0 || entry[j + 1] != 0 ||
entry            1440 drivers/net/ethernet/emulex/benet/be_main.c 			    entry[j + 2] != 0 || entry[j + 3] != 0) {
entry            1442 drivers/net/ethernet/emulex/benet/be_main.c 					 j, entry[j], entry[j + 1],
entry            1443 drivers/net/ethernet/emulex/benet/be_main.c 					 entry[j + 2], entry[j + 3]);
entry            1447 drivers/net/ethernet/emulex/benet/be_main.c 		entry = txo->cq.dma_mem.va;
entry            1452 drivers/net/ethernet/emulex/benet/be_main.c 			if (entry[j] != 0 || entry[j + 1] != 0 ||
entry            1453 drivers/net/ethernet/emulex/benet/be_main.c 			    entry[j + 2] != 0 || entry[j + 3] != 0) {
entry            1455 drivers/net/ethernet/emulex/benet/be_main.c 					 j, entry[j], entry[j + 1],
entry            1456 drivers/net/ethernet/emulex/benet/be_main.c 					 entry[j + 2], entry[j + 3]);
entry            3448 drivers/net/ethernet/emulex/benet/be_main.c 		adapter->msix_entries[i].entry = i;
entry              85 drivers/net/ethernet/emulex/benet/be_roce.c 		INIT_LIST_HEAD(&adapter->entry);
entry              87 drivers/net/ethernet/emulex/benet/be_roce.c 		list_add_tail(&adapter->entry, &be_adapter_list);
entry             110 drivers/net/ethernet/emulex/benet/be_roce.c 		list_del(&adapter->entry);
entry             137 drivers/net/ethernet/emulex/benet/be_roce.c 	list_for_each_entry(dev, &be_adapter_list, entry) {
entry             150 drivers/net/ethernet/emulex/benet/be_roce.c 	list_for_each_entry(dev, &be_adapter_list, entry) {
entry             422 drivers/net/ethernet/ethoc.c 		unsigned int entry;
entry             425 drivers/net/ethernet/ethoc.c 		entry = priv->num_tx + priv->cur_rx;
entry             426 drivers/net/ethernet/ethoc.c 		ethoc_read_bd(priv, entry, &bd);
entry             436 drivers/net/ethernet/ethoc.c 			ethoc_read_bd(priv, entry, &bd);
entry             449 drivers/net/ethernet/ethoc.c 				void *src = priv->vma[entry];
entry             468 drivers/net/ethernet/ethoc.c 		ethoc_write_bd(priv, entry, &bd);
entry             515 drivers/net/ethernet/ethoc.c 		unsigned int entry;
entry             517 drivers/net/ethernet/ethoc.c 		entry = priv->dty_tx & (priv->num_tx-1);
entry             519 drivers/net/ethernet/ethoc.c 		ethoc_read_bd(priv, entry, &bd);
entry             530 drivers/net/ethernet/ethoc.c 			ethoc_read_bd(priv, entry, &bd);
entry             884 drivers/net/ethernet/ethoc.c 	unsigned int entry;
entry             897 drivers/net/ethernet/ethoc.c 	entry = priv->cur_tx % priv->num_tx;
entry             901 drivers/net/ethernet/ethoc.c 	ethoc_read_bd(priv, entry, &bd);
entry             907 drivers/net/ethernet/ethoc.c 	dest = priv->vma[entry];
entry             912 drivers/net/ethernet/ethoc.c 	ethoc_write_bd(priv, entry, &bd);
entry             915 drivers/net/ethernet/ethoc.c 	ethoc_write_bd(priv, entry, &bd);
entry             382 drivers/net/ethernet/faraday/ftgmac100.c static int ftgmac100_alloc_rx_buf(struct ftgmac100 *priv, unsigned int entry,
entry             410 drivers/net/ethernet/faraday/ftgmac100.c 	priv->rx_skbs[entry] = skb;
entry             419 drivers/net/ethernet/faraday/ftgmac100.c 	if (entry == (priv->rx_q_entries - 1))
entry            1236 drivers/net/ethernet/freescale/enetc/enetc.c 		int entry = ENETC_BDR_INT_BASE_IDX + i;
entry            1250 drivers/net/ethernet/freescale/enetc/enetc.c 		enetc_wr(hw, ENETC_SIMSIRRV(i), entry);
entry            1255 drivers/net/ethernet/freescale/enetc/enetc.c 			enetc_wr(hw, ENETC_SIMSITRV(idx), entry);
entry             946 drivers/net/ethernet/freescale/fman/fman_memac.c 	u32 entry;
entry             953 drivers/net/ethernet/freescale/fman/fman_memac.c 		for (entry = 0; entry < HASH_TABLE_SIZE; entry++)
entry             954 drivers/net/ethernet/freescale/fman/fman_memac.c 			iowrite32be(entry | HASH_CTRL_MCAST_EN,
entry             957 drivers/net/ethernet/freescale/fman/fman_memac.c 		for (entry = 0; entry < HASH_TABLE_SIZE; entry++)
entry             958 drivers/net/ethernet/freescale/fman/fman_memac.c 			iowrite32be(entry & ~HASH_CTRL_MCAST_EN,
entry             571 drivers/net/ethernet/freescale/fman/fman_tgec.c 	u32 entry;
entry             578 drivers/net/ethernet/freescale/fman/fman_tgec.c 		for (entry = 0; entry < TGEC_HASH_TABLE_SIZE; entry++)
entry             579 drivers/net/ethernet/freescale/fman/fman_tgec.c 			iowrite32be(entry | TGEC_HASH_MCAST_EN,
entry             582 drivers/net/ethernet/freescale/fman/fman_tgec.c 		for (entry = 0; entry < TGEC_HASH_TABLE_SIZE; entry++)
entry             583 drivers/net/ethernet/freescale/fman/fman_tgec.c 			iowrite32be(entry & ~TGEC_HASH_MCAST_EN,
entry             149 drivers/net/ethernet/google/gve/gve_main.c 		priv->msix_vectors[i].entry = i;
entry             191 drivers/net/ethernet/huawei/hinic/hinic_hw_dev.c 		hwdev->msix_entries[i].entry = i;
entry             391 drivers/net/ethernet/huawei/hinic/hinic_hw_eqs.c 	hinic_msix_attr_cnt_clear(aeq->hwif, aeq->msix_entry.entry);
entry             412 drivers/net/ethernet/huawei/hinic/hinic_hw_eqs.c 	hinic_msix_attr_cnt_clear(ceq->hwif, ceq->msix_entry.entry);
entry             436 drivers/net/ethernet/huawei/hinic/hinic_hw_eqs.c 		ctrl0 = HINIC_AEQ_CTRL_0_SET(msix_entry->entry, INT_IDX)     |
entry             457 drivers/net/ethernet/huawei/hinic/hinic_hw_eqs.c 		ctrl0 = HINIC_CEQ_CTRL_0_SET(msix_entry->entry, INTR_IDX)     |
entry             664 drivers/net/ethernet/huawei/hinic/hinic_hw_eqs.c 		   struct msix_entry entry)
entry             694 drivers/net/ethernet/huawei/hinic/hinic_hw_eqs.c 	eq->msix_entry = entry;
entry             725 drivers/net/ethernet/huawei/hinic/hinic_hw_eqs.c 	hinic_msix_attr_set(eq->hwif, eq->msix_entry.entry,
entry             733 drivers/net/ethernet/huawei/hinic/hinic_hw_eqs.c 		err = request_irq(entry.vector, aeq_interrupt, 0,
entry             736 drivers/net/ethernet/huawei/hinic/hinic_hw_eqs.c 		err = request_irq(entry.vector, ceq_interrupt, 0,
entry             757 drivers/net/ethernet/huawei/hinic/hinic_hw_eqs.c 	hinic_set_msix_state(eq->hwif, eq->msix_entry.entry,
entry             277 drivers/net/ethernet/huawei/hinic/hinic_hw_qp.c 		  struct hinic_wq *wq, struct msix_entry *entry,
entry             285 drivers/net/ethernet/huawei/hinic/hinic_hw_qp.c 	sq->irq = entry->vector;
entry             286 drivers/net/ethernet/huawei/hinic/hinic_hw_qp.c 	sq->msix_entry = entry->entry;
entry             380 drivers/net/ethernet/huawei/hinic/hinic_hw_qp.c 		  struct hinic_wq *wq, struct msix_entry *entry)
entry             390 drivers/net/ethernet/huawei/hinic/hinic_hw_qp.c 	rq->irq = entry->vector;
entry             391 drivers/net/ethernet/huawei/hinic/hinic_hw_qp.c 	rq->msix_entry = entry->entry;
entry             130 drivers/net/ethernet/huawei/hinic/hinic_hw_qp.h 		  struct hinic_wq *wq, struct msix_entry *entry, void *ci_addr,
entry             136 drivers/net/ethernet/huawei/hinic/hinic_hw_qp.h 		  struct hinic_wq *wq, struct msix_entry *entry);
entry             609 drivers/net/ethernet/huawei/hinic/hinic_port.c 		indir_tbl->entry[i] = indir_table[i];
entry             612 drivers/net/ethernet/huawei/hinic/hinic_port.c 			temp = (u32 *)&indir_tbl->entry[i - 3];
entry             633 drivers/net/ethernet/huawei/hinic/hinic_port.c 	memcpy(&indir_tbl->entry[0], &indir_tbl->entry[indir_size], indir_size);
entry             294 drivers/net/ethernet/huawei/hinic/hinic_port.h 	u8 entry[HINIC_RSS_INDIR_SIZE];
entry             930 drivers/net/ethernet/ibm/ehea/ehea_main.c 		qp_token = EHEA_BMASK_GET(EHEA_EQE_QP_TOKEN, eqe->entry);
entry             932 drivers/net/ethernet/ibm/ehea/ehea_main.c 		       eqe->entry, qp_token);
entry            1225 drivers/net/ethernet/ibm/ehea/ehea_main.c 		pr_debug("*eqe=%lx\n", (unsigned long) eqe->entry);
entry            1226 drivers/net/ethernet/ibm/ehea/ehea_main.c 		ehea_parse_eqe(adapter, eqe->entry);
entry             190 drivers/net/ethernet/ibm/ehea/ehea_qmr.h 	u64 entry;
entry            2136 drivers/net/ethernet/ibm/ibmvnic.c 	struct list_head *entry, *tmp_entry;
entry            2158 drivers/net/ethernet/ibm/ibmvnic.c 	list_for_each(entry, &adapter->rwi_list) {
entry            2159 drivers/net/ethernet/ibm/ibmvnic.c 		tmp = list_entry(entry, struct ibmvnic_rwi, list);
entry            2179 drivers/net/ethernet/ibm/ibmvnic.c 		list_for_each_safe(entry, tmp_entry, &adapter->rwi_list)
entry            2180 drivers/net/ethernet/ibm/ibmvnic.c 			list_del(entry);
entry            3292 drivers/net/ethernet/ibm/ibmvnic.c 	union sub_crq *entry = &scrq->msgs[scrq->cur];
entry            3294 drivers/net/ethernet/ibm/ibmvnic.c 	if (entry->generic.first & IBMVNIC_CRQ_CMD_RSP)
entry            3303 drivers/net/ethernet/ibm/ibmvnic.c 	union sub_crq *entry;
entry            3307 drivers/net/ethernet/ibm/ibmvnic.c 	entry = &scrq->msgs[scrq->cur];
entry            3308 drivers/net/ethernet/ibm/ibmvnic.c 	if (entry->generic.first & IBMVNIC_CRQ_CMD_RSP) {
entry            3312 drivers/net/ethernet/ibm/ibmvnic.c 		entry = NULL;
entry            3316 drivers/net/ethernet/ibm/ibmvnic.c 	return entry;
entry            2063 drivers/net/ethernet/intel/e1000e/netdev.c 					adapter->msix_entries[i].entry = i;
entry            1847 drivers/net/ethernet/intel/fm10k/fm10k_main.c 		interface->msix_entries[vector].entry = vector;
entry            1424 drivers/net/ethernet/intel/fm10k/fm10k_pci.c 	struct msix_entry *entry;
entry            1431 drivers/net/ethernet/intel/fm10k/fm10k_pci.c 	entry = &interface->msix_entries[FM10K_MBX_VECTOR];
entry            1454 drivers/net/ethernet/intel/fm10k/fm10k_pci.c 	free_irq(entry->vector, interface);
entry            1509 drivers/net/ethernet/intel/fm10k/fm10k_pci.c 	struct msix_entry *entry = &interface->msix_entries[FM10K_MBX_VECTOR];
entry            1515 drivers/net/ethernet/intel/fm10k/fm10k_pci.c 	u32 itr = entry->entry | FM10K_INT_MAP_TIMER0;
entry            1523 drivers/net/ethernet/intel/fm10k/fm10k_pci.c 	err = request_irq(entry->vector, fm10k_msix_mbx_vf, 0,
entry            1535 drivers/net/ethernet/intel/fm10k/fm10k_pci.c 	fm10k_write_reg(hw, FM10K_VFITR(entry->entry), FM10K_ITR_ENABLE);
entry            1642 drivers/net/ethernet/intel/fm10k/fm10k_pci.c 	struct msix_entry *entry = &interface->msix_entries[FM10K_MBX_VECTOR];
entry            1648 drivers/net/ethernet/intel/fm10k/fm10k_pci.c 	u32 mbx_itr = entry->entry | FM10K_INT_MAP_TIMER0;
entry            1649 drivers/net/ethernet/intel/fm10k/fm10k_pci.c 	u32 other_itr = entry->entry | FM10K_INT_MAP_IMMEDIATE;
entry            1657 drivers/net/ethernet/intel/fm10k/fm10k_pci.c 	err = request_irq(entry->vector, fm10k_msix_mbx_pf, 0,
entry            1686 drivers/net/ethernet/intel/fm10k/fm10k_pci.c 	fm10k_write_reg(hw, FM10K_ITR(entry->entry), FM10K_ITR_ENABLE);
entry            1723 drivers/net/ethernet/intel/fm10k/fm10k_pci.c 	struct msix_entry *entry;
entry            1725 drivers/net/ethernet/intel/fm10k/fm10k_pci.c 	entry = &interface->msix_entries[NON_Q_VECTORS + vector];
entry            1731 drivers/net/ethernet/intel/fm10k/fm10k_pci.c 		entry--;
entry            1738 drivers/net/ethernet/intel/fm10k/fm10k_pci.c 		irq_set_affinity_hint(entry->vector, NULL);
entry            1743 drivers/net/ethernet/intel/fm10k/fm10k_pci.c 		free_irq(entry->vector, q_vector);
entry            1758 drivers/net/ethernet/intel/fm10k/fm10k_pci.c 	struct msix_entry *entry;
entry            1762 drivers/net/ethernet/intel/fm10k/fm10k_pci.c 	entry = &interface->msix_entries[NON_Q_VECTORS];
entry            1785 drivers/net/ethernet/intel/fm10k/fm10k_pci.c 				&interface->uc_addr[FM10K_ITR(entry->entry)] :
entry            1786 drivers/net/ethernet/intel/fm10k/fm10k_pci.c 				&interface->uc_addr[FM10K_VFITR(entry->entry)];
entry            1789 drivers/net/ethernet/intel/fm10k/fm10k_pci.c 		err = request_irq(entry->vector, &fm10k_msix_clean_rings, 0,
entry            1799 drivers/net/ethernet/intel/fm10k/fm10k_pci.c 		irq_set_affinity_hint(entry->vector, &q_vector->affinity_mask);
entry            1804 drivers/net/ethernet/intel/fm10k/fm10k_pci.c 		entry++;
entry            1814 drivers/net/ethernet/intel/fm10k/fm10k_pci.c 		entry--;
entry            1822 drivers/net/ethernet/intel/fm10k/fm10k_pci.c 		irq_set_affinity_hint(entry->vector, NULL);
entry            1827 drivers/net/ethernet/intel/fm10k/fm10k_pci.c 		free_irq(entry->vector, q_vector);
entry             392 drivers/net/ethernet/intel/i40e/i40e_ddp.c 	struct i40e_ddp_old_profile_list *entry;
entry             397 drivers/net/ethernet/intel/i40e/i40e_ddp.c 		entry = list_first_entry(&pf->ddp_old_prof,
entry             400 drivers/net/ethernet/intel/i40e/i40e_ddp.c 		status = i40e_ddp_load(netdev, entry->old_ddp_buf,
entry             401 drivers/net/ethernet/intel/i40e/i40e_ddp.c 				       entry->old_ddp_size, false);
entry             402 drivers/net/ethernet/intel/i40e/i40e_ddp.c 		list_del(&entry->list);
entry             403 drivers/net/ethernet/intel/i40e/i40e_ddp.c 		kfree(entry);
entry            3576 drivers/net/ethernet/intel/i40e/i40e_ethtool.c 	struct i40e_flex_pit *entry, *tmp;
entry            3580 drivers/net/ethernet/intel/i40e/i40e_ethtool.c 	list_for_each_entry_safe(entry, tmp, &pf->l3_flex_pit_list, list) {
entry            3587 drivers/net/ethernet/intel/i40e/i40e_ethtool.c 			    rule->flex_offset == entry->src_offset) {
entry            3597 drivers/net/ethernet/intel/i40e/i40e_ethtool.c 			list_del(&entry->list);
entry            3598 drivers/net/ethernet/intel/i40e/i40e_ethtool.c 			kfree(entry);
entry            3603 drivers/net/ethernet/intel/i40e/i40e_ethtool.c 	list_for_each_entry_safe(entry, tmp, &pf->l4_flex_pit_list, list) {
entry            3613 drivers/net/ethernet/intel/i40e/i40e_ethtool.c 			    rule->flex_offset == entry->src_offset) {
entry            3623 drivers/net/ethernet/intel/i40e/i40e_ethtool.c 			list_del(&entry->list);
entry            3624 drivers/net/ethernet/intel/i40e/i40e_ethtool.c 			kfree(entry);
entry            3674 drivers/net/ethernet/intel/i40e/i40e_ethtool.c 	struct i40e_flex_pit *entry;
entry            3681 drivers/net/ethernet/intel/i40e/i40e_ethtool.c 	list_for_each_entry(entry, &pf->l4_flex_pit_list, list)
entry            3682 drivers/net/ethernet/intel/i40e/i40e_ethtool.c 		clear_bit(entry->pit_index, &available_index);
entry            3684 drivers/net/ethernet/intel/i40e/i40e_ethtool.c 	list_for_each_entry(entry, &pf->l3_flex_pit_list, list)
entry            3685 drivers/net/ethernet/intel/i40e/i40e_ethtool.c 		clear_bit(entry->pit_index, &available_index);
entry            3703 drivers/net/ethernet/intel/i40e/i40e_ethtool.c 	struct i40e_flex_pit *entry;
entry            3709 drivers/net/ethernet/intel/i40e/i40e_ethtool.c 	list_for_each_entry(entry, flex_pit_list, list) {
entry            3711 drivers/net/ethernet/intel/i40e/i40e_ethtool.c 		if (entry->src_offset == src_offset)
entry            3712 drivers/net/ethernet/intel/i40e/i40e_ethtool.c 			return entry;
entry            3743 drivers/net/ethernet/intel/i40e/i40e_ethtool.c 	struct i40e_flex_pit *new_pit, *entry;
entry            3745 drivers/net/ethernet/intel/i40e/i40e_ethtool.c 	new_pit = kzalloc(sizeof(*entry), GFP_KERNEL);
entry            3755 drivers/net/ethernet/intel/i40e/i40e_ethtool.c 	list_for_each_entry(entry, flex_pit_list, list) {
entry            3756 drivers/net/ethernet/intel/i40e/i40e_ethtool.c 		if (new_pit->src_offset < entry->src_offset) {
entry            3757 drivers/net/ethernet/intel/i40e/i40e_ethtool.c 			list_add_tail(&new_pit->list, &entry->list);
entry            3765 drivers/net/ethernet/intel/i40e/i40e_ethtool.c 		if (new_pit->src_offset == entry->src_offset) {
entry            3771 drivers/net/ethernet/intel/i40e/i40e_ethtool.c 			if (new_pit->pit_index != entry->pit_index)
entry            3811 drivers/net/ethernet/intel/i40e/i40e_ethtool.c 	struct i40e_flex_pit *entry = NULL;
entry            3818 drivers/net/ethernet/intel/i40e/i40e_ethtool.c 	list_for_each_entry(entry, flex_pit_list, list) {
entry            3834 drivers/net/ethernet/intel/i40e/i40e_ethtool.c 			u16 offset = entry->src_offset + j;
entry            3851 drivers/net/ethernet/intel/i40e/i40e_ethtool.c 				  I40E_FLEX_PREP_VAL(entry->pit_index + 50,
entry            3853 drivers/net/ethernet/intel/i40e/i40e_ethtool.c 						     entry->src_offset));
entry            3864 drivers/net/ethernet/intel/i40e/i40e_ethtool.c 		last_offset = list_prev_entry(entry, list)->src_offset + 1;
entry            11049 drivers/net/ethernet/intel/i40e/i40e_main.c 		pf->msix_entries[i].entry = i;
entry            1189 drivers/net/ethernet/intel/iavf/iavf_main.c 		adapter->msix_entries[vector].entry = vector;
entry            1508 drivers/net/ethernet/intel/ice/ice_adminq_cmd.h 	__le16 entry[1];
entry             487 drivers/net/ethernet/intel/ice/ice_common.c 	(((n) - 1) * sizeof(((struct ice_aqc_fw_logging_data *)0)->entry)))
entry             520 drivers/net/ethernet/intel/ice/ice_common.c 			v = le16_to_cpu(config->entry[i]);
entry             626 drivers/net/ethernet/intel/ice/ice_common.c 			data->entry[chgs++] = cpu_to_le16(val);
entry             675 drivers/net/ethernet/intel/ice/ice_common.c 			v = le16_to_cpu(data->entry[i]);
entry            1292 drivers/net/ethernet/intel/ice/ice_flex_pipe.c 			src = (u8 *)pid->entry;
entry             346 drivers/net/ethernet/intel/ice/ice_flex_type.h 	struct ice_prof_tcam_entry entry[1];
entry            2412 drivers/net/ethernet/intel/ice/ice_main.c 		pf->msix_entries[i].entry = i;
entry            1745 drivers/net/ethernet/intel/ice/ice_switch.c 		struct ice_aqc_sw_rules_elem *entry = r_iter;
entry            1749 drivers/net/ethernet/intel/ice/ice_switch.c 		status = ice_aq_sw_rules(hw, entry, elem_sent * s_rule_size,
entry            2044 drivers/net/ethernet/intel/ice/ice_switch.c 		struct ice_fltr_mgmt_list_entry *entry;
entry            2047 drivers/net/ethernet/intel/ice/ice_switch.c 		list_for_each_entry_safe(entry, tmp, rule_head, list_entry) {
entry            2048 drivers/net/ethernet/intel/ice/ice_switch.c 			list_del(&entry->list_entry);
entry            2049 drivers/net/ethernet/intel/ice/ice_switch.c 			devm_kfree(ice_hw_to_dev(hw), entry);
entry            1126 drivers/net/ethernet/intel/igb/igb_main.c 		adapter->msix_entries[i].entry = i;
entry            7189 drivers/net/ethernet/intel/igb/igb_main.c static bool igb_mac_entry_can_be_used(const struct igb_mac_addr *entry,
entry            7192 drivers/net/ethernet/intel/igb/igb_main.c 	if (!(entry->state & IGB_MAC_STATE_IN_USE))
entry            7195 drivers/net/ethernet/intel/igb/igb_main.c 	if ((entry->state & IGB_MAC_STATE_SRC_ADDR) !=
entry            7199 drivers/net/ethernet/intel/igb/igb_main.c 	if (!ether_addr_equal(addr, entry->addr))
entry            7354 drivers/net/ethernet/intel/igb/igb_main.c 	struct vf_mac_filter *entry = NULL;
entry            7361 drivers/net/ethernet/intel/igb/igb_main.c 			entry = list_entry(pos, struct vf_mac_filter, l);
entry            7362 drivers/net/ethernet/intel/igb/igb_main.c 			if (entry->vf == vf) {
entry            7363 drivers/net/ethernet/intel/igb/igb_main.c 				entry->vf = -1;
entry            7364 drivers/net/ethernet/intel/igb/igb_main.c 				entry->free = true;
entry            7365 drivers/net/ethernet/intel/igb/igb_main.c 				igb_del_mac_filter(adapter, entry->vf_mac, vf);
entry            7386 drivers/net/ethernet/intel/igb/igb_main.c 			entry = list_entry(pos, struct vf_mac_filter, l);
entry            7387 drivers/net/ethernet/intel/igb/igb_main.c 			if (entry->free)
entry            7391 drivers/net/ethernet/intel/igb/igb_main.c 		if (entry && entry->free) {
entry            7392 drivers/net/ethernet/intel/igb/igb_main.c 			entry->free = false;
entry            7393 drivers/net/ethernet/intel/igb/igb_main.c 			entry->vf = vf;
entry            7394 drivers/net/ethernet/intel/igb/igb_main.c 			ether_addr_copy(entry->vf_mac, addr);
entry            1025 drivers/net/ethernet/intel/igbvf/netdev.c 			adapter->msix_entries[i].entry = i;
entry            2405 drivers/net/ethernet/intel/igc/igc_main.c static bool igc_mac_entry_can_be_used(const struct igc_mac_addr *entry,
entry            2408 drivers/net/ethernet/intel/igc/igc_main.c 	if (!(entry->state & IGC_MAC_STATE_IN_USE))
entry            2411 drivers/net/ethernet/intel/igc/igc_main.c 	if ((entry->state & IGC_MAC_STATE_SRC_ADDR) !=
entry            2415 drivers/net/ethernet/intel/igc/igc_main.c 	if (!ether_addr_equal(addr, entry->addr))
entry            3460 drivers/net/ethernet/intel/igc/igc_main.c 		adapter->msix_entries[i].entry = i;
entry             773 drivers/net/ethernet/intel/ixgbe/ixgbe_lib.c 		adapter->msix_entries[i].entry = i;
entry            3233 drivers/net/ethernet/intel/ixgbe/ixgbe_main.c 		struct msix_entry *entry = &adapter->msix_entries[vector];
entry            3249 drivers/net/ethernet/intel/ixgbe/ixgbe_main.c 		err = request_irq(entry->vector, &ixgbe_msix_clean_rings, 0,
entry            3259 drivers/net/ethernet/intel/ixgbe/ixgbe_main.c 			irq_set_affinity_hint(entry->vector,
entry            3403 drivers/net/ethernet/intel/ixgbe/ixgbe_main.c 		struct msix_entry *entry = &adapter->msix_entries[vector];
entry            3410 drivers/net/ethernet/intel/ixgbe/ixgbe_main.c 		irq_set_affinity_hint(entry->vector, NULL);
entry            3412 drivers/net/ethernet/intel/ixgbe/ixgbe_main.c 		free_irq(entry->vector, q_vector);
entry            10582 drivers/net/ethernet/intel/ixgbe/ixgbe_main.c 	struct pci_dev *entry, *pdev = adapter->pdev;
entry            10592 drivers/net/ethernet/intel/ixgbe/ixgbe_main.c 	list_for_each_entry(entry, &adapter->pdev->bus->devices, bus_list) {
entry            10594 drivers/net/ethernet/intel/ixgbe/ixgbe_main.c 		if (entry->is_virtfn)
entry            10603 drivers/net/ethernet/intel/ixgbe/ixgbe_main.c 		if ((entry->vendor != pdev->vendor) ||
entry            10604 drivers/net/ethernet/intel/ixgbe/ixgbe_main.c 		    (entry->device != pdev->device))
entry             639 drivers/net/ethernet/intel/ixgbe/ixgbe_sriov.c 	struct vf_macvlans *entry;
entry             645 drivers/net/ethernet/intel/ixgbe/ixgbe_sriov.c 			entry = list_entry(pos, struct vf_macvlans, l);
entry             646 drivers/net/ethernet/intel/ixgbe/ixgbe_sriov.c 			if (entry->vf == vf) {
entry             647 drivers/net/ethernet/intel/ixgbe/ixgbe_sriov.c 				entry->vf = -1;
entry             648 drivers/net/ethernet/intel/ixgbe/ixgbe_sriov.c 				entry->free = true;
entry             649 drivers/net/ethernet/intel/ixgbe/ixgbe_sriov.c 				entry->is_macvlan = false;
entry             651 drivers/net/ethernet/intel/ixgbe/ixgbe_sriov.c 						     entry->vf_macvlan, vf);
entry             663 drivers/net/ethernet/intel/ixgbe/ixgbe_sriov.c 	entry = NULL;
entry             666 drivers/net/ethernet/intel/ixgbe/ixgbe_sriov.c 		entry = list_entry(pos, struct vf_macvlans, l);
entry             667 drivers/net/ethernet/intel/ixgbe/ixgbe_sriov.c 		if (entry->free)
entry             679 drivers/net/ethernet/intel/ixgbe/ixgbe_sriov.c 	if (!entry || !entry->free)
entry             686 drivers/net/ethernet/intel/ixgbe/ixgbe_sriov.c 	entry->free = false;
entry             687 drivers/net/ethernet/intel/ixgbe/ixgbe_sriov.c 	entry->is_macvlan = true;
entry             688 drivers/net/ethernet/intel/ixgbe/ixgbe_sriov.c 	entry->vf = vf;
entry             689 drivers/net/ethernet/intel/ixgbe/ixgbe_sriov.c 	memcpy(entry->vf_macvlan, mac_addr, ETH_ALEN);
entry            1540 drivers/net/ethernet/intel/ixgbevf/ixgbevf_main.c 		struct msix_entry *entry = &adapter->msix_entries[vector];
entry            1556 drivers/net/ethernet/intel/ixgbevf/ixgbevf_main.c 		err = request_irq(entry->vector, &ixgbevf_msix_clean_rings, 0,
entry            2664 drivers/net/ethernet/intel/ixgbevf/ixgbevf_main.c 		adapter->msix_entries[vector].entry = vector;
entry            1876 drivers/net/ethernet/marvell/mv643xx_eth.c 		u8 entry;
entry            1880 drivers/net/ethernet/marvell/mv643xx_eth.c 			entry = a[5];
entry            1883 drivers/net/ethernet/marvell/mv643xx_eth.c 			entry = addr_crc(a);
entry            1886 drivers/net/ethernet/marvell/mv643xx_eth.c 		table[entry >> 2] |= 1 << (8 * (entry & 3));
entry            1049 drivers/net/ethernet/marvell/mvpp2/mvpp2_cls.c static void mvpp22_port_c2_lookup_disable(struct mvpp2_port *port, int entry)
entry            1053 drivers/net/ethernet/marvell/mvpp2/mvpp2_cls.c 	mvpp2_cls_c2_read(port->priv, entry, &c2);
entry              60 drivers/net/ethernet/marvell/mvpp2/mvpp2_debugfs.c 	struct mvpp2_dbgfs_flow_tbl_entry *entry = s->private;
entry              62 drivers/net/ethernet/marvell/mvpp2/mvpp2_debugfs.c 	u32 hits = mvpp2_cls_flow_hits(entry->priv, entry->id);
entry              73 drivers/net/ethernet/marvell/mvpp2/mvpp2_debugfs.c 	struct mvpp2_dbgfs_flow_entry *entry = s->private;
entry              75 drivers/net/ethernet/marvell/mvpp2/mvpp2_debugfs.c 	u32 hits = mvpp2_cls_lookup_hits(entry->priv, entry->flow);
entry              86 drivers/net/ethernet/marvell/mvpp2/mvpp2_debugfs.c 	struct mvpp2_dbgfs_flow_entry *entry = s->private;
entry              90 drivers/net/ethernet/marvell/mvpp2/mvpp2_debugfs.c 	f = mvpp2_cls_flow_get(entry->flow);
entry             126 drivers/net/ethernet/marvell/mvpp2/mvpp2_debugfs.c 	const struct mvpp2_dbgfs_flow_entry *entry = s->private;
entry             129 drivers/net/ethernet/marvell/mvpp2/mvpp2_debugfs.c 	f = mvpp2_cls_flow_get(entry->flow);
entry             142 drivers/net/ethernet/marvell/mvpp2/mvpp2_debugfs.c 	struct mvpp2_dbgfs_port_flow_entry *entry = s->private;
entry             143 drivers/net/ethernet/marvell/mvpp2/mvpp2_debugfs.c 	struct mvpp2_port *port = entry->port;
entry             149 drivers/net/ethernet/marvell/mvpp2/mvpp2_debugfs.c 	f = mvpp2_cls_flow_get(entry->dbg_fe->flow);
entry             153 drivers/net/ethernet/marvell/mvpp2/mvpp2_debugfs.c 	flow_index = MVPP2_CLS_FLT_HASH_ENTRY(entry->port->id, f->flow_id);
entry             168 drivers/net/ethernet/marvell/mvpp2/mvpp2_debugfs.c 	struct mvpp2_dbgfs_port_flow_entry *entry = s->private;
entry             169 drivers/net/ethernet/marvell/mvpp2/mvpp2_debugfs.c 	struct mvpp2_port *port = entry->port;
entry             174 drivers/net/ethernet/marvell/mvpp2/mvpp2_debugfs.c 	f = mvpp2_cls_flow_get(entry->dbg_fe->flow);
entry             178 drivers/net/ethernet/marvell/mvpp2/mvpp2_debugfs.c 	flow_index = MVPP2_CLS_FLT_HASH_ENTRY(entry->port->id, f->flow_id);
entry             193 drivers/net/ethernet/marvell/mvpp2/mvpp2_debugfs.c 	struct mvpp2_dbgfs_c2_entry *entry = s->private;
entry             196 drivers/net/ethernet/marvell/mvpp2/mvpp2_debugfs.c 	hits = mvpp2_cls_c2_hit_count(entry->priv, entry->id);
entry             207 drivers/net/ethernet/marvell/mvpp2/mvpp2_debugfs.c 	struct mvpp2_dbgfs_c2_entry *entry = s->private;
entry             211 drivers/net/ethernet/marvell/mvpp2/mvpp2_debugfs.c 	mvpp2_cls_c2_read(entry->priv, entry->id, &c2);
entry             228 drivers/net/ethernet/marvell/mvpp2/mvpp2_debugfs.c 	struct mvpp2_dbgfs_c2_entry *entry = s->private;
entry             232 drivers/net/ethernet/marvell/mvpp2/mvpp2_debugfs.c 	mvpp2_cls_c2_read(entry->priv, entry->id, &c2);
entry             339 drivers/net/ethernet/marvell/mvpp2/mvpp2_debugfs.c 	struct mvpp2_dbgfs_prs_entry *entry = s->private;
entry             340 drivers/net/ethernet/marvell/mvpp2/mvpp2_debugfs.c 	struct mvpp2 *priv = entry->priv;
entry             342 drivers/net/ethernet/marvell/mvpp2/mvpp2_debugfs.c 	seq_printf(s, "%x\n", priv->prs_shadow[entry->tid].lu);
entry             351 drivers/net/ethernet/marvell/mvpp2/mvpp2_debugfs.c 	struct mvpp2_dbgfs_prs_entry *entry = s->private;
entry             355 drivers/net/ethernet/marvell/mvpp2/mvpp2_debugfs.c 	mvpp2_prs_init_from_hw(entry->priv, &pe, entry->tid);
entry             369 drivers/net/ethernet/marvell/mvpp2/mvpp2_debugfs.c 	struct mvpp2_dbgfs_prs_entry *entry = s->private;
entry             373 drivers/net/ethernet/marvell/mvpp2/mvpp2_debugfs.c 	mvpp2_prs_init_from_hw(entry->priv, &pe, entry->tid);
entry             387 drivers/net/ethernet/marvell/mvpp2/mvpp2_debugfs.c 	struct mvpp2_dbgfs_prs_entry *entry = s->private;
entry             392 drivers/net/ethernet/marvell/mvpp2/mvpp2_debugfs.c 	mvpp2_prs_init_from_hw(entry->priv, &pe, entry->tid);
entry             406 drivers/net/ethernet/marvell/mvpp2/mvpp2_debugfs.c 	struct mvpp2_dbgfs_prs_entry *entry = s->private;
entry             409 drivers/net/ethernet/marvell/mvpp2/mvpp2_debugfs.c 	mvpp2_prs_init_from_hw(entry->priv, &pe, entry->tid);
entry             420 drivers/net/ethernet/marvell/mvpp2/mvpp2_debugfs.c 	struct mvpp2_dbgfs_prs_entry *entry = s->private;
entry             423 drivers/net/ethernet/marvell/mvpp2/mvpp2_debugfs.c 	val = mvpp2_prs_hits(entry->priv, entry->tid);
entry             436 drivers/net/ethernet/marvell/mvpp2/mvpp2_debugfs.c 	struct mvpp2_dbgfs_prs_entry *entry = s->private;
entry             437 drivers/net/ethernet/marvell/mvpp2/mvpp2_debugfs.c 	struct mvpp2 *priv = entry->priv;
entry             438 drivers/net/ethernet/marvell/mvpp2/mvpp2_debugfs.c 	int tid = entry->tid;
entry             449 drivers/net/ethernet/marvell/mvpp2/mvpp2_debugfs.c 				      struct mvpp2_dbgfs_flow_entry *entry)
entry             459 drivers/net/ethernet/marvell/mvpp2/mvpp2_debugfs.c 	port_entry->dbg_fe = entry;
entry             473 drivers/net/ethernet/marvell/mvpp2/mvpp2_debugfs.c 	struct mvpp2_dbgfs_flow_entry *entry;
entry             482 drivers/net/ethernet/marvell/mvpp2/mvpp2_debugfs.c 	entry = &priv->dbgfs_entries->flow_entries[flow];
entry             484 drivers/net/ethernet/marvell/mvpp2/mvpp2_debugfs.c 	entry->flow = flow;
entry             485 drivers/net/ethernet/marvell/mvpp2/mvpp2_debugfs.c 	entry->priv = priv;
entry             487 drivers/net/ethernet/marvell/mvpp2/mvpp2_debugfs.c 	debugfs_create_file("dec_hits", 0444, flow_entry_dir, entry,
entry             490 drivers/net/ethernet/marvell/mvpp2/mvpp2_debugfs.c 	debugfs_create_file("type", 0444, flow_entry_dir, entry,
entry             493 drivers/net/ethernet/marvell/mvpp2/mvpp2_debugfs.c 	debugfs_create_file("id", 0444, flow_entry_dir, entry,
entry             499 drivers/net/ethernet/marvell/mvpp2/mvpp2_debugfs.c 						 priv->port_list[i], entry);
entry             526 drivers/net/ethernet/marvell/mvpp2/mvpp2_debugfs.c 	struct mvpp2_dbgfs_prs_entry *entry;
entry             537 drivers/net/ethernet/marvell/mvpp2/mvpp2_debugfs.c 	entry = &priv->dbgfs_entries->prs_entries[tid];
entry             539 drivers/net/ethernet/marvell/mvpp2/mvpp2_debugfs.c 	entry->tid = tid;
entry             540 drivers/net/ethernet/marvell/mvpp2/mvpp2_debugfs.c 	entry->priv = priv;
entry             543 drivers/net/ethernet/marvell/mvpp2/mvpp2_debugfs.c 	debugfs_create_file("sram", 0444, prs_entry_dir, entry,
entry             546 drivers/net/ethernet/marvell/mvpp2/mvpp2_debugfs.c 	debugfs_create_file("valid", 0644, prs_entry_dir, entry,
entry             549 drivers/net/ethernet/marvell/mvpp2/mvpp2_debugfs.c 	debugfs_create_file("lookup_id", 0644, prs_entry_dir, entry,
entry             552 drivers/net/ethernet/marvell/mvpp2/mvpp2_debugfs.c 	debugfs_create_file("ai", 0644, prs_entry_dir, entry,
entry             555 drivers/net/ethernet/marvell/mvpp2/mvpp2_debugfs.c 	debugfs_create_file("header_data", 0644, prs_entry_dir, entry,
entry             558 drivers/net/ethernet/marvell/mvpp2/mvpp2_debugfs.c 	debugfs_create_file("hits", 0444, prs_entry_dir, entry,
entry             561 drivers/net/ethernet/marvell/mvpp2/mvpp2_debugfs.c 	debugfs_create_file("pmap", 0444, prs_entry_dir, entry,
entry             586 drivers/net/ethernet/marvell/mvpp2/mvpp2_debugfs.c 	struct mvpp2_dbgfs_c2_entry *entry;
entry             599 drivers/net/ethernet/marvell/mvpp2/mvpp2_debugfs.c 	entry = &priv->dbgfs_entries->c2_entries[id];
entry             601 drivers/net/ethernet/marvell/mvpp2/mvpp2_debugfs.c 	entry->id = id;
entry             602 drivers/net/ethernet/marvell/mvpp2/mvpp2_debugfs.c 	entry->priv = priv;
entry             604 drivers/net/ethernet/marvell/mvpp2/mvpp2_debugfs.c 	debugfs_create_file("hits", 0444, c2_entry_dir, entry,
entry             607 drivers/net/ethernet/marvell/mvpp2/mvpp2_debugfs.c 	debugfs_create_file("default_rxq", 0444, c2_entry_dir, entry,
entry             610 drivers/net/ethernet/marvell/mvpp2/mvpp2_debugfs.c 	debugfs_create_file("rss_enable", 0444, c2_entry_dir, entry,
entry             619 drivers/net/ethernet/marvell/mvpp2/mvpp2_debugfs.c 	struct mvpp2_dbgfs_flow_tbl_entry *entry;
entry             632 drivers/net/ethernet/marvell/mvpp2/mvpp2_debugfs.c 	entry = &priv->dbgfs_entries->flt_entries[id];
entry             634 drivers/net/ethernet/marvell/mvpp2/mvpp2_debugfs.c 	entry->id = id;
entry             635 drivers/net/ethernet/marvell/mvpp2/mvpp2_debugfs.c 	entry->priv = priv;
entry             637 drivers/net/ethernet/marvell/mvpp2/mvpp2_debugfs.c 	debugfs_create_file("hits", 0444, flow_tbl_entry_dir, entry,
entry             676 drivers/net/ethernet/marvell/octeontx2/af/mbox.h 	u16 entry; /* Entry allocated or start index if contiguous.
entry             686 drivers/net/ethernet/marvell/octeontx2/af/mbox.h 	u16 entry; /* Entry index to be freed */
entry             701 drivers/net/ethernet/marvell/octeontx2/af/mbox.h 	u16 entry;	 /* MCAM entry to write this match key */
entry             711 drivers/net/ethernet/marvell/octeontx2/af/mbox.h 	u16 entry;
entry             756 drivers/net/ethernet/marvell/octeontx2/af/mbox.h 	u16 entry; /* Entry and counter to be unmapped */
entry             772 drivers/net/ethernet/marvell/octeontx2/af/mbox.h 	u16 entry;
entry             151 drivers/net/ethernet/marvell/octeontx2/af/rvu.h 	struct mcam_entry entry;
entry            2510 drivers/net/ethernet/marvell/octeontx2/af/rvu_nix.c 	free_req.entry = alloc_rsp.entry_list[0];
entry             123 drivers/net/ethernet/marvell/octeontx2/af/rvu_npc.c static void npc_get_keyword(struct mcam_entry *entry, int idx,
entry             138 drivers/net/ethernet/marvell/octeontx2/af/rvu_npc.c 		*cam1 = entry->kw[0];
entry             139 drivers/net/ethernet/marvell/octeontx2/af/rvu_npc.c 		kw_mask = entry->kw_mask[0];
entry             143 drivers/net/ethernet/marvell/octeontx2/af/rvu_npc.c 		*cam1 = entry->kw[1] & CAM_MASK(48);
entry             144 drivers/net/ethernet/marvell/octeontx2/af/rvu_npc.c 		kw_mask = entry->kw_mask[1] & CAM_MASK(48);
entry             150 drivers/net/ethernet/marvell/octeontx2/af/rvu_npc.c 		*cam1 = (entry->kw[1] >> 48) & CAM_MASK(16);
entry             151 drivers/net/ethernet/marvell/octeontx2/af/rvu_npc.c 		*cam1 |= ((entry->kw[2] & CAM_MASK(48)) << 16);
entry             152 drivers/net/ethernet/marvell/octeontx2/af/rvu_npc.c 		kw_mask = (entry->kw_mask[1] >> 48) & CAM_MASK(16);
entry             153 drivers/net/ethernet/marvell/octeontx2/af/rvu_npc.c 		kw_mask |= ((entry->kw_mask[2] & CAM_MASK(48)) << 16);
entry             159 drivers/net/ethernet/marvell/octeontx2/af/rvu_npc.c 		*cam1 = (entry->kw[2] >> 48) & CAM_MASK(16);
entry             160 drivers/net/ethernet/marvell/octeontx2/af/rvu_npc.c 		*cam1 |= ((entry->kw[3] & CAM_MASK(32)) << 16);
entry             161 drivers/net/ethernet/marvell/octeontx2/af/rvu_npc.c 		kw_mask = (entry->kw_mask[2] >> 48) & CAM_MASK(16);
entry             162 drivers/net/ethernet/marvell/octeontx2/af/rvu_npc.c 		kw_mask |= ((entry->kw_mask[3] & CAM_MASK(32)) << 16);
entry             168 drivers/net/ethernet/marvell/octeontx2/af/rvu_npc.c 		*cam1 = (entry->kw[3] >> 32) & CAM_MASK(32);
entry             169 drivers/net/ethernet/marvell/octeontx2/af/rvu_npc.c 		*cam1 |= ((entry->kw[4] & CAM_MASK(32)) << 32);
entry             170 drivers/net/ethernet/marvell/octeontx2/af/rvu_npc.c 		kw_mask = (entry->kw_mask[3] >> 32) & CAM_MASK(32);
entry             171 drivers/net/ethernet/marvell/octeontx2/af/rvu_npc.c 		kw_mask |= ((entry->kw_mask[4] & CAM_MASK(32)) << 32);
entry             177 drivers/net/ethernet/marvell/octeontx2/af/rvu_npc.c 		*cam1 = (entry->kw[4] >> 32) & CAM_MASK(32);
entry             178 drivers/net/ethernet/marvell/octeontx2/af/rvu_npc.c 		*cam1 |= ((entry->kw[5] & CAM_MASK(16)) << 32);
entry             179 drivers/net/ethernet/marvell/octeontx2/af/rvu_npc.c 		kw_mask = (entry->kw_mask[4] >> 32) & CAM_MASK(32);
entry             180 drivers/net/ethernet/marvell/octeontx2/af/rvu_npc.c 		kw_mask |= ((entry->kw_mask[5] & CAM_MASK(16)) << 32);
entry             186 drivers/net/ethernet/marvell/octeontx2/af/rvu_npc.c 		*cam1 = (entry->kw[5] >> 16) & CAM_MASK(48);
entry             187 drivers/net/ethernet/marvell/octeontx2/af/rvu_npc.c 		*cam1 |= ((entry->kw[6] & CAM_MASK(16)) << 48);
entry             188 drivers/net/ethernet/marvell/octeontx2/af/rvu_npc.c 		kw_mask = (entry->kw_mask[5] >> 16) & CAM_MASK(48);
entry             189 drivers/net/ethernet/marvell/octeontx2/af/rvu_npc.c 		kw_mask |= ((entry->kw_mask[6] & CAM_MASK(16)) << 48);
entry             193 drivers/net/ethernet/marvell/octeontx2/af/rvu_npc.c 		*cam1 = (entry->kw[6] >> 16) & CAM_MASK(48);
entry             194 drivers/net/ethernet/marvell/octeontx2/af/rvu_npc.c 		kw_mask = (entry->kw_mask[6] >> 16) & CAM_MASK(48);
entry             204 drivers/net/ethernet/marvell/octeontx2/af/rvu_npc.c 				  struct mcam_entry *entry, bool enable)
entry             230 drivers/net/ethernet/marvell/octeontx2/af/rvu_npc.c 		npc_get_keyword(entry, kw, &cam0, &cam1);
entry             236 drivers/net/ethernet/marvell/octeontx2/af/rvu_npc.c 		npc_get_keyword(entry, kw + 1, &cam0, &cam1);
entry             245 drivers/net/ethernet/marvell/octeontx2/af/rvu_npc.c 		    NPC_AF_MCAMEX_BANKX_ACTION(index, actbank), entry->action);
entry             249 drivers/net/ethernet/marvell/octeontx2/af/rvu_npc.c 		    entry->vtag_action);
entry             313 drivers/net/ethernet/marvell/octeontx2/af/rvu_npc.c 	struct mcam_entry entry = { {0} };
entry             333 drivers/net/ethernet/marvell/octeontx2/af/rvu_npc.c 	entry.kw[0] = chan;
entry             334 drivers/net/ethernet/marvell/octeontx2/af/rvu_npc.c 	entry.kw_mask[0] = 0xFFFULL;
entry             337 drivers/net/ethernet/marvell/octeontx2/af/rvu_npc.c 	entry.kw[kwi] = mac;
entry             338 drivers/net/ethernet/marvell/octeontx2/af/rvu_npc.c 	entry.kw_mask[kwi] = BIT_ULL(48) - 1;
entry             352 drivers/net/ethernet/marvell/octeontx2/af/rvu_npc.c 	entry.action = *(u64 *)&action;
entry             354 drivers/net/ethernet/marvell/octeontx2/af/rvu_npc.c 			      NIX_INTF_RX, &entry, true);
entry             357 drivers/net/ethernet/marvell/octeontx2/af/rvu_npc.c 	entry.kw[0] |= (NPC_LT_LB_STAG | NPC_LT_LB_CTAG) << 20;
entry             358 drivers/net/ethernet/marvell/octeontx2/af/rvu_npc.c 	entry.kw_mask[0] |= (NPC_LT_LB_STAG & NPC_LT_LB_CTAG) << 20;
entry             360 drivers/net/ethernet/marvell/octeontx2/af/rvu_npc.c 	entry.vtag_action = VTAG0_VALID_BIT |
entry             365 drivers/net/ethernet/marvell/octeontx2/af/rvu_npc.c 	memcpy(&pfvf->entry, &entry, sizeof(entry));
entry             373 drivers/net/ethernet/marvell/octeontx2/af/rvu_npc.c 	struct mcam_entry entry = { {0} };
entry             387 drivers/net/ethernet/marvell/octeontx2/af/rvu_npc.c 	entry.kw[0] = chan;
entry             388 drivers/net/ethernet/marvell/octeontx2/af/rvu_npc.c 	entry.kw_mask[0] = 0xFFFULL;
entry             392 drivers/net/ethernet/marvell/octeontx2/af/rvu_npc.c 		entry.kw[kwi] = BIT_ULL(40); /* LSB bit of 1st byte in DMAC */
entry             393 drivers/net/ethernet/marvell/octeontx2/af/rvu_npc.c 		entry.kw_mask[kwi] = BIT_ULL(40);
entry             412 drivers/net/ethernet/marvell/octeontx2/af/rvu_npc.c 	entry.action = *(u64 *)&action;
entry             414 drivers/net/ethernet/marvell/octeontx2/af/rvu_npc.c 			      NIX_INTF_RX, &entry, true);
entry             450 drivers/net/ethernet/marvell/octeontx2/af/rvu_npc.c 	struct mcam_entry entry = { {0} };
entry             491 drivers/net/ethernet/marvell/octeontx2/af/rvu_npc.c 	entry.kw[0] = BIT_ULL(13) | chan;
entry             492 drivers/net/ethernet/marvell/octeontx2/af/rvu_npc.c 	entry.kw_mask[0] = BIT_ULL(13) | 0xFFFULL;
entry             508 drivers/net/ethernet/marvell/octeontx2/af/rvu_npc.c 	entry.action = *(u64 *)&action;
entry             510 drivers/net/ethernet/marvell/octeontx2/af/rvu_npc.c 			      NIX_INTF_RX, &entry, true);
entry             842 drivers/net/ethernet/marvell/octeontx2/af/rvu_npc.c 				 int kpu, int entry, bool pkind)
entry             855 drivers/net/ethernet/marvell/octeontx2/af/rvu_npc.c 		reg = NPC_AF_PKINDX_ACTION1(entry);
entry             857 drivers/net/ethernet/marvell/octeontx2/af/rvu_npc.c 		reg = NPC_AF_KPUX_ENTRYX_ACTION1(kpu, entry);
entry             875 drivers/net/ethernet/marvell/octeontx2/af/rvu_npc.c 		reg = NPC_AF_PKINDX_ACTION0(entry);
entry             877 drivers/net/ethernet/marvell/octeontx2/af/rvu_npc.c 		reg = NPC_AF_KPUX_ENTRYX_ACTION0(kpu, entry);
entry             884 drivers/net/ethernet/marvell/octeontx2/af/rvu_npc.c 			      int kpu, int entry)
entry             900 drivers/net/ethernet/marvell/octeontx2/af/rvu_npc.c 		    NPC_AF_KPUX_ENTRYX_CAMX(kpu, entry, 0), *(u64 *)&cam0);
entry             902 drivers/net/ethernet/marvell/octeontx2/af/rvu_npc.c 		    NPC_AF_KPUX_ENTRYX_CAMX(kpu, entry, 1), *(u64 *)&cam1);
entry             913 drivers/net/ethernet/marvell/octeontx2/af/rvu_npc.c 	int entry, num_entries, max_entries;
entry             925 drivers/net/ethernet/marvell/octeontx2/af/rvu_npc.c 	for (entry = 0; entry < num_entries; entry++)
entry             927 drivers/net/ethernet/marvell/octeontx2/af/rvu_npc.c 				  &profile->cam[entry], kpu, entry);
entry             931 drivers/net/ethernet/marvell/octeontx2/af/rvu_npc.c 	for (entry = 0; entry < num_entries; entry++)
entry             932 drivers/net/ethernet/marvell/octeontx2/af/rvu_npc.c 		npc_config_kpuaction(rvu, blkaddr, &profile->action[entry],
entry             933 drivers/net/ethernet/marvell/octeontx2/af/rvu_npc.c 				     kpu, entry, false);
entry            1105 drivers/net/ethernet/marvell/octeontx2/af/rvu_npc.c 	int blkaddr, entry, bank, err;
entry            1117 drivers/net/ethernet/marvell/octeontx2/af/rvu_npc.c 		for (entry = 0; entry < ((cfg >> 28) & 0xFFFF); entry++)
entry            1119 drivers/net/ethernet/marvell/octeontx2/af/rvu_npc.c 				    NPC_AF_MCAMEX_BANKX_CFG(entry, bank), 0);
entry            1204 drivers/net/ethernet/marvell/octeontx2/af/rvu_npc.c 				 u16 pcifunc, int entry)
entry            1209 drivers/net/ethernet/marvell/octeontx2/af/rvu_npc.c 	if (entry >= mcam->bmap_entries)
entry            1212 drivers/net/ethernet/marvell/octeontx2/af/rvu_npc.c 	if (pcifunc != mcam->entry2pfvf_map[entry])
entry            1234 drivers/net/ethernet/marvell/octeontx2/af/rvu_npc.c 					int blkaddr, u16 entry, u16 cntr)
entry            1236 drivers/net/ethernet/marvell/octeontx2/af/rvu_npc.c 	u16 index = entry & (mcam->banksize - 1);
entry            1237 drivers/net/ethernet/marvell/octeontx2/af/rvu_npc.c 	u16 bank = npc_get_bank(mcam, entry);
entry            1240 drivers/net/ethernet/marvell/octeontx2/af/rvu_npc.c 	mcam->entry2cntr_map[entry] = cntr;
entry            1250 drivers/net/ethernet/marvell/octeontx2/af/rvu_npc.c 					  int blkaddr, u16 entry, u16 cntr)
entry            1252 drivers/net/ethernet/marvell/octeontx2/af/rvu_npc.c 	u16 index = entry & (mcam->banksize - 1);
entry            1253 drivers/net/ethernet/marvell/octeontx2/af/rvu_npc.c 	u16 bank = npc_get_bank(mcam, entry);
entry            1256 drivers/net/ethernet/marvell/octeontx2/af/rvu_npc.c 	mcam->entry2cntr_map[entry] = NPC_MCAM_INVALID_MAP;
entry            1269 drivers/net/ethernet/marvell/octeontx2/af/rvu_npc.c 	u16 entry, rentry;
entry            1271 drivers/net/ethernet/marvell/octeontx2/af/rvu_npc.c 	entry = index;
entry            1274 drivers/net/ethernet/marvell/octeontx2/af/rvu_npc.c 	__set_bit(entry, mcam->bmap);
entry            1285 drivers/net/ethernet/marvell/octeontx2/af/rvu_npc.c 	u16 entry, rentry;
entry            1287 drivers/net/ethernet/marvell/octeontx2/af/rvu_npc.c 	entry = index;
entry            1290 drivers/net/ethernet/marvell/octeontx2/af/rvu_npc.c 	__clear_bit(entry, mcam->bmap);
entry            1464 drivers/net/ethernet/marvell/octeontx2/af/rvu_npc.c 	int entry, next_start;
entry            1550 drivers/net/ethernet/marvell/octeontx2/af/rvu_npc.c 			rsp->entry = mcam->bmap_entries - index - max_contig;
entry            1552 drivers/net/ethernet/marvell/octeontx2/af/rvu_npc.c 			rsp->entry = index;
entry            1559 drivers/net/ethernet/marvell/octeontx2/af/rvu_npc.c 		for (entry = 0; entry < req->count; entry++) {
entry            1569 drivers/net/ethernet/marvell/octeontx2/af/rvu_npc.c 			entry_list[entry] = index;
entry            1610 drivers/net/ethernet/marvell/octeontx2/af/rvu_npc.c 		for (entry = rsp->count - 1; entry >= 0; entry--) {
entry            1612 drivers/net/ethernet/marvell/octeontx2/af/rvu_npc.c 				rsp->entry_list[index++] = entry_list[entry];
entry            1614 drivers/net/ethernet/marvell/octeontx2/af/rvu_npc.c 				rsp->entry_list[entry] = entry_list[entry];
entry            1619 drivers/net/ethernet/marvell/octeontx2/af/rvu_npc.c 	for (entry = 0; entry < rsp->count; entry++) {
entry            1621 drivers/net/ethernet/marvell/octeontx2/af/rvu_npc.c 			(rsp->entry + entry) : rsp->entry_list[entry];
entry            1646 drivers/net/ethernet/marvell/octeontx2/af/rvu_npc.c 	rsp->entry = NPC_MCAM_ENTRY_INVALID;
entry            1696 drivers/net/ethernet/marvell/octeontx2/af/rvu_npc.c 	rc = npc_mcam_verify_entry(mcam, pcifunc, req->entry);
entry            1700 drivers/net/ethernet/marvell/octeontx2/af/rvu_npc.c 	mcam->entry2pfvf_map[req->entry] = 0;
entry            1701 drivers/net/ethernet/marvell/octeontx2/af/rvu_npc.c 	npc_mcam_clear_bit(mcam, req->entry);
entry            1702 drivers/net/ethernet/marvell/octeontx2/af/rvu_npc.c 	npc_enable_mcam_entry(rvu, mcam, blkaddr, req->entry, false);
entry            1705 drivers/net/ethernet/marvell/octeontx2/af/rvu_npc.c 	cntr = mcam->entry2cntr_map[req->entry];
entry            1708 drivers/net/ethernet/marvell/octeontx2/af/rvu_npc.c 					      req->entry, cntr);
entry            1733 drivers/net/ethernet/marvell/octeontx2/af/rvu_npc.c 	rc = npc_mcam_verify_entry(mcam, pcifunc, req->entry);
entry            1748 drivers/net/ethernet/marvell/octeontx2/af/rvu_npc.c 	npc_config_mcam_entry(rvu, mcam, blkaddr, req->entry, req->intf,
entry            1753 drivers/net/ethernet/marvell/octeontx2/af/rvu_npc.c 					    req->entry, req->cntr);
entry            1774 drivers/net/ethernet/marvell/octeontx2/af/rvu_npc.c 	rc = npc_mcam_verify_entry(mcam, pcifunc, req->entry);
entry            1779 drivers/net/ethernet/marvell/octeontx2/af/rvu_npc.c 	npc_enable_mcam_entry(rvu, mcam, blkaddr, req->entry, true);
entry            1797 drivers/net/ethernet/marvell/octeontx2/af/rvu_npc.c 	rc = npc_mcam_verify_entry(mcam, pcifunc, req->entry);
entry            1802 drivers/net/ethernet/marvell/octeontx2/af/rvu_npc.c 	npc_enable_mcam_entry(rvu, mcam, blkaddr, req->entry, false);
entry            1944 drivers/net/ethernet/marvell/octeontx2/af/rvu_npc.c 	u16 index, entry = 0;
entry            1963 drivers/net/ethernet/marvell/octeontx2/af/rvu_npc.c 	while (entry < mcam->bmap_entries) {
entry            1967 drivers/net/ethernet/marvell/octeontx2/af/rvu_npc.c 		index = find_next_bit(mcam->bmap, mcam->bmap_entries, entry);
entry            1973 drivers/net/ethernet/marvell/octeontx2/af/rvu_npc.c 		entry = index + 1;
entry            1986 drivers/net/ethernet/marvell/octeontx2/af/rvu_npc.c 	u16 index, entry = 0;
entry            2000 drivers/net/ethernet/marvell/octeontx2/af/rvu_npc.c 		rc = npc_mcam_verify_entry(mcam, req->hdr.pcifunc, req->entry);
entry            2004 drivers/net/ethernet/marvell/octeontx2/af/rvu_npc.c 					      req->entry, req->cntr);
entry            2009 drivers/net/ethernet/marvell/octeontx2/af/rvu_npc.c 	while (entry < mcam->bmap_entries) {
entry            2013 drivers/net/ethernet/marvell/octeontx2/af/rvu_npc.c 		index = find_next_bit(mcam->bmap, mcam->bmap_entries, entry);
entry            2019 drivers/net/ethernet/marvell/octeontx2/af/rvu_npc.c 		entry = index + 1;
entry            2081 drivers/net/ethernet/marvell/octeontx2/af/rvu_npc.c 	u16 entry = NPC_MCAM_ENTRY_INVALID;
entry            2107 drivers/net/ethernet/marvell/octeontx2/af/rvu_npc.c 	entry = entry_rsp.entry;
entry            2121 drivers/net/ethernet/marvell/octeontx2/af/rvu_npc.c 		mcam->entry2pfvf_map[entry] = 0;
entry            2122 drivers/net/ethernet/marvell/octeontx2/af/rvu_npc.c 		npc_mcam_clear_bit(mcam, entry);
entry            2131 drivers/net/ethernet/marvell/octeontx2/af/rvu_npc.c 	npc_config_mcam_entry(rvu, mcam, blkaddr, entry, req->intf,
entry            2135 drivers/net/ethernet/marvell/octeontx2/af/rvu_npc.c 		npc_map_mcam_entry_and_cntr(rvu, mcam, blkaddr, entry, cntr);
entry            2138 drivers/net/ethernet/marvell/octeontx2/af/rvu_npc.c 	rsp->entry = entry;
entry            2206 drivers/net/ethernet/marvell/octeontx2/af/rvu_npc.c 	pfvf->entry.action = npc_get_mcam_action(rvu, mcam, blkaddr, index);
entry            2209 drivers/net/ethernet/marvell/octeontx2/af/rvu_npc.c 			      NIX_INTF_RX, &pfvf->entry, enable);
entry             440 drivers/net/ethernet/marvell/pxa168_eth.c 	struct addr_table_entry *entry, *start;
entry             467 drivers/net/ethernet/marvell/pxa168_eth.c 	entry = start + hash_function(mac_addr);
entry             469 drivers/net/ethernet/marvell/pxa168_eth.c 		if (!(le32_to_cpu(entry->lo) & HASH_ENTRY_VALID)) {
entry             473 drivers/net/ethernet/marvell/pxa168_eth.c 			if (((le32_to_cpu(entry->lo) & 0xfffffff8) ==
entry             475 drivers/net/ethernet/marvell/pxa168_eth.c 				(le32_to_cpu(entry->hi) == new_high)) {
entry             479 drivers/net/ethernet/marvell/pxa168_eth.c 		if (entry == start + 0x7ff)
entry             480 drivers/net/ethernet/marvell/pxa168_eth.c 			entry = start;
entry             482 drivers/net/ethernet/marvell/pxa168_eth.c 			entry++;
entry             485 drivers/net/ethernet/marvell/pxa168_eth.c 	if (((le32_to_cpu(entry->lo) & 0xfffffff8) != (new_low & 0xfffffff8)) &&
entry             486 drivers/net/ethernet/marvell/pxa168_eth.c 	    (le32_to_cpu(entry->hi) != new_high) && del)
entry             504 drivers/net/ethernet/marvell/pxa168_eth.c 		entry->hi = 0;
entry             505 drivers/net/ethernet/marvell/pxa168_eth.c 		entry->lo = 0;
entry             507 drivers/net/ethernet/marvell/pxa168_eth.c 		entry->hi = cpu_to_le32(new_high);
entry             508 drivers/net/ethernet/marvell/pxa168_eth.c 		entry->lo = cpu_to_le32(new_low);
entry             285 drivers/net/ethernet/mellanox/mlx4/alloc.c static void __mlx4_zone_remove_one_entry(struct mlx4_zone_entry *entry)
entry             287 drivers/net/ethernet/mellanox/mlx4/alloc.c 	struct mlx4_zone_allocator *zone_alloc = entry->allocator;
entry             289 drivers/net/ethernet/mellanox/mlx4/alloc.c 	if (!list_empty(&entry->prio_list)) {
entry             291 drivers/net/ethernet/mellanox/mlx4/alloc.c 		if (!list_is_last(&entry->list, &zone_alloc->entries)) {
entry             292 drivers/net/ethernet/mellanox/mlx4/alloc.c 			struct mlx4_zone_entry *next = list_first_entry(&entry->list,
entry             296 drivers/net/ethernet/mellanox/mlx4/alloc.c 			if (next->priority == entry->priority)
entry             297 drivers/net/ethernet/mellanox/mlx4/alloc.c 				list_add_tail(&next->prio_list, &entry->prio_list);
entry             300 drivers/net/ethernet/mellanox/mlx4/alloc.c 		list_del(&entry->prio_list);
entry             303 drivers/net/ethernet/mellanox/mlx4/alloc.c 	list_del(&entry->list);
entry             705 drivers/net/ethernet/mellanox/mlx4/en_netdev.c 		struct mlx4_mac_entry *entry;
entry             710 drivers/net/ethernet/mellanox/mlx4/en_netdev.c 		hlist_for_each_entry_safe(entry, tmp, bucket, hlist) {
entry             711 drivers/net/ethernet/mellanox/mlx4/en_netdev.c 			if (ether_addr_equal_64bits(entry->mac, prev_mac)) {
entry             712 drivers/net/ethernet/mellanox/mlx4/en_netdev.c 				mlx4_en_uc_steer_release(priv, entry->mac,
entry             713 drivers/net/ethernet/mellanox/mlx4/en_netdev.c 							 qpn, entry->reg_id);
entry             716 drivers/net/ethernet/mellanox/mlx4/en_netdev.c 				hlist_del_rcu(&entry->hlist);
entry             718 drivers/net/ethernet/mellanox/mlx4/en_netdev.c 				memcpy(entry->mac, new_mac, ETH_ALEN);
entry             719 drivers/net/ethernet/mellanox/mlx4/en_netdev.c 				entry->reg_id = 0;
entry             721 drivers/net/ethernet/mellanox/mlx4/en_netdev.c 				hlist_add_head_rcu(&entry->hlist,
entry             726 drivers/net/ethernet/mellanox/mlx4/en_netdev.c 							   &entry->reg_id);
entry            1135 drivers/net/ethernet/mellanox/mlx4/en_netdev.c 	struct mlx4_mac_entry *entry;
entry            1152 drivers/net/ethernet/mellanox/mlx4/en_netdev.c 		hlist_for_each_entry_safe(entry, tmp, bucket, hlist) {
entry            1155 drivers/net/ethernet/mellanox/mlx4/en_netdev.c 				if (ether_addr_equal_64bits(entry->mac,
entry            1163 drivers/net/ethernet/mellanox/mlx4/en_netdev.c 			if (ether_addr_equal_64bits(entry->mac,
entry            1168 drivers/net/ethernet/mellanox/mlx4/en_netdev.c 				mac = mlx4_mac_to_u64(entry->mac);
entry            1169 drivers/net/ethernet/mellanox/mlx4/en_netdev.c 				mlx4_en_uc_steer_release(priv, entry->mac,
entry            1171 drivers/net/ethernet/mellanox/mlx4/en_netdev.c 							 entry->reg_id);
entry            1174 drivers/net/ethernet/mellanox/mlx4/en_netdev.c 				hlist_del_rcu(&entry->hlist);
entry            1175 drivers/net/ethernet/mellanox/mlx4/en_netdev.c 				kfree_rcu(entry, rcu);
entry            1177 drivers/net/ethernet/mellanox/mlx4/en_netdev.c 				       entry->mac, priv->port);
entry            1196 drivers/net/ethernet/mellanox/mlx4/en_netdev.c 		hlist_for_each_entry(entry, bucket, hlist) {
entry            1197 drivers/net/ethernet/mellanox/mlx4/en_netdev.c 			if (ether_addr_equal_64bits(entry->mac, ha->addr)) {
entry            1204 drivers/net/ethernet/mellanox/mlx4/en_netdev.c 			entry = kmalloc(sizeof(*entry), GFP_KERNEL);
entry            1205 drivers/net/ethernet/mellanox/mlx4/en_netdev.c 			if (!entry) {
entry            1212 drivers/net/ethernet/mellanox/mlx4/en_netdev.c 			memcpy(entry->mac, ha->addr, ETH_ALEN);
entry            1217 drivers/net/ethernet/mellanox/mlx4/en_netdev.c 				kfree(entry);
entry            1223 drivers/net/ethernet/mellanox/mlx4/en_netdev.c 						   &entry->reg_id);
entry            1228 drivers/net/ethernet/mellanox/mlx4/en_netdev.c 				kfree(entry);
entry            1237 drivers/net/ethernet/mellanox/mlx4/en_netdev.c 				hlist_add_head_rcu(&entry->hlist, bucket);
entry            1302 drivers/net/ethernet/mellanox/mlx4/en_netdev.c 	struct mlx4_mac_entry *entry;
entry            1313 drivers/net/ethernet/mellanox/mlx4/en_netdev.c 	entry = kmalloc(sizeof(*entry), GFP_KERNEL);
entry            1314 drivers/net/ethernet/mellanox/mlx4/en_netdev.c 	if (!entry) {
entry            1319 drivers/net/ethernet/mellanox/mlx4/en_netdev.c 	memcpy(entry->mac, priv->dev->dev_addr, sizeof(entry->mac));
entry            1320 drivers/net/ethernet/mellanox/mlx4/en_netdev.c 	memcpy(priv->current_mac, entry->mac, sizeof(priv->current_mac));
entry            1321 drivers/net/ethernet/mellanox/mlx4/en_netdev.c 	entry->reg_id = reg_id;
entry            1322 drivers/net/ethernet/mellanox/mlx4/en_netdev.c 	hlist_add_head_rcu(&entry->hlist,
entry            1323 drivers/net/ethernet/mellanox/mlx4/en_netdev.c 			   &priv->mac_hash[entry->mac[MLX4_EN_MAC_HASH_IDX]]);
entry            1343 drivers/net/ethernet/mellanox/mlx4/en_netdev.c 	struct mlx4_mac_entry *entry;
entry            1347 drivers/net/ethernet/mellanox/mlx4/en_netdev.c 		hlist_for_each_entry_safe(entry, tmp, bucket, hlist) {
entry            1348 drivers/net/ethernet/mellanox/mlx4/en_netdev.c 			mac = mlx4_mac_to_u64(entry->mac);
entry            1350 drivers/net/ethernet/mellanox/mlx4/en_netdev.c 			       entry->mac);
entry            1351 drivers/net/ethernet/mellanox/mlx4/en_netdev.c 			mlx4_en_uc_steer_release(priv, entry->mac,
entry            1352 drivers/net/ethernet/mellanox/mlx4/en_netdev.c 						 qpn, entry->reg_id);
entry            1355 drivers/net/ethernet/mellanox/mlx4/en_netdev.c 			hlist_del_rcu(&entry->hlist);
entry            1356 drivers/net/ethernet/mellanox/mlx4/en_netdev.c 			kfree_rcu(entry, rcu);
entry              51 drivers/net/ethernet/mellanox/mlx4/en_port.c 	u32 entry;
entry              60 drivers/net/ethernet/mellanox/mlx4/en_port.c 		entry = 0;
entry              63 drivers/net/ethernet/mellanox/mlx4/en_port.c 				entry |= 1 << j;
entry              64 drivers/net/ethernet/mellanox/mlx4/en_port.c 		filter->entry[i] = cpu_to_be32(entry);
entry              45 drivers/net/ethernet/mellanox/mlx4/en_port.h 	__be32 entry[VLAN_FLTR_SIZE];
entry             740 drivers/net/ethernet/mellanox/mlx4/en_rx.c 				struct mlx4_mac_entry *entry;
entry             747 drivers/net/ethernet/mellanox/mlx4/en_rx.c 				hlist_for_each_entry_rcu(entry, bucket, hlist) {
entry             748 drivers/net/ethernet/mellanox/mlx4/en_rx.c 					if (ether_addr_equal_64bits(entry->mac,
entry             106 drivers/net/ethernet/mellanox/mlx4/eq.c static struct mlx4_eqe *get_eqe(struct mlx4_eq *eq, u32 entry, u8 eqe_factor,
entry             110 drivers/net/ethernet/mellanox/mlx4/eq.c 	unsigned long offset = (entry & (eq->nent - 1)) * eqe_size;
entry            2668 drivers/net/ethernet/mellanox/mlx4/main.c void mlx4_set_admin_guid(struct mlx4_dev *dev, __be64 guid, int entry, int port)
entry            2672 drivers/net/ethernet/mellanox/mlx4/main.c 	priv->mfunc.master.vf_admin[entry].vport[port].guid = guid;
entry            2676 drivers/net/ethernet/mellanox/mlx4/main.c __be64 mlx4_get_admin_guid(struct mlx4_dev *dev, int entry, int port)
entry            2680 drivers/net/ethernet/mellanox/mlx4/main.c 	return priv->mfunc.master.vf_admin[entry].vport[port].guid;
entry            2684 drivers/net/ethernet/mellanox/mlx4/main.c void mlx4_set_random_admin_guid(struct mlx4_dev *dev, int entry, int port)
entry            2690 drivers/net/ethernet/mellanox/mlx4/main.c 	if (entry == 0)
entry            2696 drivers/net/ethernet/mellanox/mlx4/main.c 	priv->mfunc.master.vf_admin[entry].vport[port].guid = guid;
entry            2952 drivers/net/ethernet/mellanox/mlx4/main.c 			entries[i].entry = i;
entry            3127 drivers/net/ethernet/mellanox/mlx4/main.c 	struct mlx4_steer_index *entry, *tmp_entry;
entry            3140 drivers/net/ethernet/mellanox/mlx4/main.c 			list_for_each_entry_safe(entry, tmp_entry,
entry            3143 drivers/net/ethernet/mellanox/mlx4/main.c 				list_del(&entry->list);
entry            3145 drivers/net/ethernet/mellanox/mlx4/main.c 							 &entry->duplicates,
entry            3150 drivers/net/ethernet/mellanox/mlx4/main.c 				kfree(entry);
entry             244 drivers/net/ethernet/mellanox/mlx4/mcg.c 	struct mlx4_steer_index *tmp_entry, *entry = NULL;
entry             259 drivers/net/ethernet/mellanox/mlx4/mcg.c 			entry = tmp_entry;
entry             263 drivers/net/ethernet/mellanox/mlx4/mcg.c 	if (unlikely(!entry)) {
entry             271 drivers/net/ethernet/mellanox/mlx4/mcg.c 	list_for_each_entry(dqp, &entry->duplicates, list) {
entry             281 drivers/net/ethernet/mellanox/mlx4/mcg.c 	list_add_tail(&dqp->list, &entry->duplicates);
entry             293 drivers/net/ethernet/mellanox/mlx4/mcg.c 	struct mlx4_steer_index *tmp_entry, *entry = NULL;
entry             309 drivers/net/ethernet/mellanox/mlx4/mcg.c 			entry = tmp_entry;
entry             313 drivers/net/ethernet/mellanox/mlx4/mcg.c 	if (unlikely(!entry)) {
entry             317 drivers/net/ethernet/mellanox/mlx4/mcg.c 	list_for_each_entry_safe(dqp, tmp_dqp, &entry->duplicates, list) {
entry             373 drivers/net/ethernet/mellanox/mlx4/mcg.c 	struct mlx4_steer_index *entry = NULL, *tmp_entry;
entry             389 drivers/net/ethernet/mellanox/mlx4/mcg.c 	list_for_each_entry_safe(entry, tmp_entry, &s_steer->steer_entries[steer], list) {
entry             390 drivers/net/ethernet/mellanox/mlx4/mcg.c 		if (entry->index == index) {
entry             391 drivers/net/ethernet/mellanox/mlx4/mcg.c 			if (list_empty(&entry->duplicates) ||
entry             398 drivers/net/ethernet/mellanox/mlx4/mcg.c 				list_del(&entry->list);
entry             400 drivers/net/ethernet/mellanox/mlx4/mcg.c 							 &entry->duplicates,
entry             405 drivers/net/ethernet/mellanox/mlx4/mcg.c 				kfree(entry);
entry             424 drivers/net/ethernet/mellanox/mlx4/mcg.c 	struct mlx4_steer_index *entry;
entry             465 drivers/net/ethernet/mellanox/mlx4/mcg.c 		list_for_each_entry(entry,
entry             468 drivers/net/ethernet/mellanox/mlx4/mcg.c 			err = mlx4_READ_ENTRY(dev, entry->index, mailbox);
entry             489 drivers/net/ethernet/mellanox/mlx4/mcg.c 						      &entry->duplicates);
entry             506 drivers/net/ethernet/mellanox/mlx4/mcg.c 				err = mlx4_WRITE_ENTRY(dev, entry->index,
entry             555 drivers/net/ethernet/mellanox/mlx4/mcg.c 	struct mlx4_steer_index *entry, *tmp_entry;
entry             600 drivers/net/ethernet/mellanox/mlx4/mcg.c 		list_for_each_entry_safe(entry, tmp_entry,
entry             604 drivers/net/ethernet/mellanox/mlx4/mcg.c 			list_for_each_entry(dqp, &entry->duplicates, list) {
entry             620 drivers/net/ethernet/mellanox/mlx4/mcg.c 						      entry->index,
entry             629 drivers/net/ethernet/mellanox/mlx4/mcg.c 						  qpn, entry->index);
entry             630 drivers/net/ethernet/mellanox/mlx4/mcg.c 					list_del(&entry->list);
entry             631 drivers/net/ethernet/mellanox/mlx4/mcg.c 					kfree(entry);
entry             644 drivers/net/ethernet/mellanox/mlx4/mcg.c 						 qpn, entry->index);
entry             659 drivers/net/ethernet/mellanox/mlx4/mcg.c 						       entry->index,
entry             457 drivers/net/ethernet/mellanox/mlx4/mlx4.h 	__be32 entry[VLAN_FLTR_SIZE];
entry              50 drivers/net/ethernet/mellanox/mlx5/core/lib/eq.h static inline struct mlx5_eqe *get_eqe(struct mlx5_eq *eq, u32 entry)
entry              52 drivers/net/ethernet/mellanox/mlx5/core/lib/eq.h 	return mlx5_buf_offset(&eq->buf, entry * MLX5_EQE_SIZE);
entry             165 drivers/net/ethernet/mellanox/mlx5/core/rl.c 	struct mlx5_rl_entry *entry;
entry             177 drivers/net/ethernet/mellanox/mlx5/core/rl.c 	entry = find_rl_entry(table, rl);
entry             178 drivers/net/ethernet/mellanox/mlx5/core/rl.c 	if (!entry) {
entry             184 drivers/net/ethernet/mellanox/mlx5/core/rl.c 	if (entry->refcount) {
entry             186 drivers/net/ethernet/mellanox/mlx5/core/rl.c 		entry->refcount++;
entry             189 drivers/net/ethernet/mellanox/mlx5/core/rl.c 		err = mlx5_set_pp_rate_limit_cmd(dev, entry->index, rl);
entry             196 drivers/net/ethernet/mellanox/mlx5/core/rl.c 		entry->rl = *rl;
entry             197 drivers/net/ethernet/mellanox/mlx5/core/rl.c 		entry->refcount = 1;
entry             199 drivers/net/ethernet/mellanox/mlx5/core/rl.c 	*index = entry->index;
entry             210 drivers/net/ethernet/mellanox/mlx5/core/rl.c 	struct mlx5_rl_entry *entry = NULL;
entry             218 drivers/net/ethernet/mellanox/mlx5/core/rl.c 	entry = find_rl_entry(table, rl);
entry             219 drivers/net/ethernet/mellanox/mlx5/core/rl.c 	if (!entry || !entry->refcount) {
entry             225 drivers/net/ethernet/mellanox/mlx5/core/rl.c 	entry->refcount--;
entry             226 drivers/net/ethernet/mellanox/mlx5/core/rl.c 	if (!entry->refcount) {
entry             228 drivers/net/ethernet/mellanox/mlx5/core/rl.c 		mlx5_set_pp_rate_limit_cmd(dev, entry->index, &reset_rl);
entry             229 drivers/net/ethernet/mellanox/mlx5/core/rl.c 		entry->rl = reset_rl;
entry             177 drivers/net/ethernet/mellanox/mlxsw/spectrum1_acl_tcam.c 	struct mlxsw_sp1_acl_tcam_entry *entry = entry_priv;
entry             180 drivers/net/ethernet/mellanox/mlxsw/spectrum1_acl_tcam.c 					    &chunk->cchunk, &entry->centry,
entry             190 drivers/net/ethernet/mellanox/mlxsw/spectrum1_acl_tcam.c 	struct mlxsw_sp1_acl_tcam_entry *entry = entry_priv;
entry             193 drivers/net/ethernet/mellanox/mlxsw/spectrum1_acl_tcam.c 				     &chunk->cchunk, &entry->centry);
entry             228 drivers/net/ethernet/mellanox/mlxsw/spectrum1_acl_tcam.c 	struct mlxsw_sp1_acl_tcam_entry *entry = entry_priv;
entry             231 drivers/net/ethernet/mellanox/mlxsw/spectrum1_acl_tcam.c 	offset = mlxsw_sp_acl_ctcam_entry_offset(&entry->centry);
entry             207 drivers/net/ethernet/mellanox/mlxsw/spectrum2_acl_tcam.c 	struct mlxsw_sp2_acl_tcam_entry *entry = entry_priv;
entry             209 drivers/net/ethernet/mellanox/mlxsw/spectrum2_acl_tcam.c 	entry->act_block = rulei->act_block;
entry             211 drivers/net/ethernet/mellanox/mlxsw/spectrum2_acl_tcam.c 					    &chunk->achunk, &entry->aentry,
entry             221 drivers/net/ethernet/mellanox/mlxsw/spectrum2_acl_tcam.c 	struct mlxsw_sp2_acl_tcam_entry *entry = entry_priv;
entry             224 drivers/net/ethernet/mellanox/mlxsw/spectrum2_acl_tcam.c 				     &entry->aentry);
entry             233 drivers/net/ethernet/mellanox/mlxsw/spectrum2_acl_tcam.c 	struct mlxsw_sp2_acl_tcam_entry *entry = entry_priv;
entry             235 drivers/net/ethernet/mellanox/mlxsw/spectrum2_acl_tcam.c 	entry->act_block = rulei->act_block;
entry             238 drivers/net/ethernet/mellanox/mlxsw/spectrum2_acl_tcam.c 						       &entry->aentry, rulei);
entry             246 drivers/net/ethernet/mellanox/mlxsw/spectrum2_acl_tcam.c 	struct mlxsw_sp2_acl_tcam_entry *entry = entry_priv;
entry             248 drivers/net/ethernet/mellanox/mlxsw/spectrum2_acl_tcam.c 	return mlxsw_afa_block_activity_get(entry->act_block, activity);
entry             251 drivers/net/ethernet/mellanox/mlxsw/spectrum_acl_tcam.c 	struct mlxsw_sp_acl_tcam_entry *entry;
entry            1107 drivers/net/ethernet/mellanox/mlxsw/spectrum_acl_tcam.c 	struct mlxsw_sp_acl_tcam_entry *entry;
entry            1110 drivers/net/ethernet/mellanox/mlxsw/spectrum_acl_tcam.c 	entry = kzalloc(sizeof(*entry) + ops->entry_priv_size, GFP_KERNEL);
entry            1111 drivers/net/ethernet/mellanox/mlxsw/spectrum_acl_tcam.c 	if (!entry)
entry            1113 drivers/net/ethernet/mellanox/mlxsw/spectrum_acl_tcam.c 	entry->ventry = ventry;
entry            1114 drivers/net/ethernet/mellanox/mlxsw/spectrum_acl_tcam.c 	entry->chunk = chunk;
entry            1117 drivers/net/ethernet/mellanox/mlxsw/spectrum_acl_tcam.c 			     entry->priv, ventry->rulei);
entry            1121 drivers/net/ethernet/mellanox/mlxsw/spectrum_acl_tcam.c 	return entry;
entry            1124 drivers/net/ethernet/mellanox/mlxsw/spectrum_acl_tcam.c 	kfree(entry);
entry            1129 drivers/net/ethernet/mellanox/mlxsw/spectrum_acl_tcam.c 					    struct mlxsw_sp_acl_tcam_entry *entry)
entry            1133 drivers/net/ethernet/mellanox/mlxsw/spectrum_acl_tcam.c 	ops->entry_del(mlxsw_sp, entry->chunk->region->priv,
entry            1134 drivers/net/ethernet/mellanox/mlxsw/spectrum_acl_tcam.c 		       entry->chunk->priv, entry->priv);
entry            1135 drivers/net/ethernet/mellanox/mlxsw/spectrum_acl_tcam.c 	kfree(entry);
entry            1141 drivers/net/ethernet/mellanox/mlxsw/spectrum_acl_tcam.c 				       struct mlxsw_sp_acl_tcam_entry *entry,
entry            1147 drivers/net/ethernet/mellanox/mlxsw/spectrum_acl_tcam.c 					 entry->priv, rulei);
entry            1152 drivers/net/ethernet/mellanox/mlxsw/spectrum_acl_tcam.c 				     struct mlxsw_sp_acl_tcam_entry *entry,
entry            1157 drivers/net/ethernet/mellanox/mlxsw/spectrum_acl_tcam.c 	return ops->entry_activity_get(mlxsw_sp, entry->chunk->region->priv,
entry            1158 drivers/net/ethernet/mellanox/mlxsw/spectrum_acl_tcam.c 				       entry->priv, activity);
entry            1180 drivers/net/ethernet/mellanox/mlxsw/spectrum_acl_tcam.c 	ventry->entry = mlxsw_sp_acl_tcam_entry_create(mlxsw_sp, ventry,
entry            1182 drivers/net/ethernet/mellanox/mlxsw/spectrum_acl_tcam.c 	if (IS_ERR(ventry->entry)) {
entry            1184 drivers/net/ethernet/mellanox/mlxsw/spectrum_acl_tcam.c 		err = PTR_ERR(ventry->entry);
entry            1208 drivers/net/ethernet/mellanox/mlxsw/spectrum_acl_tcam.c 	mlxsw_sp_acl_tcam_entry_destroy(mlxsw_sp, ventry->entry);
entry            1222 drivers/net/ethernet/mellanox/mlxsw/spectrum_acl_tcam.c 						      ventry->entry, rulei);
entry            1231 drivers/net/ethernet/mellanox/mlxsw/spectrum_acl_tcam.c 						    ventry->entry, activity);
entry            1243 drivers/net/ethernet/mellanox/mlxsw/spectrum_acl_tcam.c 	if (ventry->entry->chunk == chunk)
entry            1252 drivers/net/ethernet/mellanox/mlxsw/spectrum_acl_tcam.c 	mlxsw_sp_acl_tcam_entry_destroy(mlxsw_sp, ventry->entry);
entry            1253 drivers/net/ethernet/mellanox/mlxsw/spectrum_acl_tcam.c 	ventry->entry = new_entry;
entry             123 drivers/net/ethernet/mellanox/mlxsw/spectrum_dpipe.c static int mlxsw_sp_erif_entry_prepare(struct devlink_dpipe_entry *entry,
entry             129 drivers/net/ethernet/mellanox/mlxsw/spectrum_dpipe.c 	entry->match_values = match_value;
entry             130 drivers/net/ethernet/mellanox/mlxsw/spectrum_dpipe.c 	entry->match_values_count = 1;
entry             132 drivers/net/ethernet/mellanox/mlxsw/spectrum_dpipe.c 	entry->action_values = action_value;
entry             133 drivers/net/ethernet/mellanox/mlxsw/spectrum_dpipe.c 	entry->action_values_count = 1;
entry             154 drivers/net/ethernet/mellanox/mlxsw/spectrum_dpipe.c 				   struct devlink_dpipe_entry *entry,
entry             164 drivers/net/ethernet/mellanox/mlxsw/spectrum_dpipe.c 	rif_value = entry->match_values->value;
entry             166 drivers/net/ethernet/mellanox/mlxsw/spectrum_dpipe.c 	entry->match_values->mapping_value = mlxsw_sp_rif_dev_ifindex(rif);
entry             167 drivers/net/ethernet/mellanox/mlxsw/spectrum_dpipe.c 	entry->match_values->mapping_valid = true;
entry             170 drivers/net/ethernet/mellanox/mlxsw/spectrum_dpipe.c 	action_value = entry->action_values->value;
entry             173 drivers/net/ethernet/mellanox/mlxsw/spectrum_dpipe.c 	entry->counter_valid = false;
entry             174 drivers/net/ethernet/mellanox/mlxsw/spectrum_dpipe.c 	entry->counter = 0;
entry             175 drivers/net/ethernet/mellanox/mlxsw/spectrum_dpipe.c 	entry->index = mlxsw_sp_rif_index(rif);
entry             184 drivers/net/ethernet/mellanox/mlxsw/spectrum_dpipe.c 		entry->counter = cnt;
entry             185 drivers/net/ethernet/mellanox/mlxsw/spectrum_dpipe.c 		entry->counter_valid = true;
entry             197 drivers/net/ethernet/mellanox/mlxsw/spectrum_dpipe.c 	struct devlink_dpipe_entry entry = {0};
entry             207 drivers/net/ethernet/mellanox/mlxsw/spectrum_dpipe.c 	err = mlxsw_sp_erif_entry_prepare(&entry, &match_value, &match,
entry             225 drivers/net/ethernet/mellanox/mlxsw/spectrum_dpipe.c 		err = mlxsw_sp_erif_entry_get(mlxsw_sp, &entry, rif,
entry             229 drivers/net/ethernet/mellanox/mlxsw/spectrum_dpipe.c 		err = devlink_dpipe_entry_ctx_append(dump_ctx, &entry);
entry             246 drivers/net/ethernet/mellanox/mlxsw/spectrum_dpipe.c 	devlink_dpipe_entry_clear(&entry);
entry             252 drivers/net/ethernet/mellanox/mlxsw/spectrum_dpipe.c 	devlink_dpipe_entry_clear(&entry);
entry             400 drivers/net/ethernet/mellanox/mlxsw/spectrum_dpipe.c mlxsw_sp_dpipe_table_host_entry_prepare(struct devlink_dpipe_entry *entry,
entry             410 drivers/net/ethernet/mellanox/mlxsw/spectrum_dpipe.c 	entry->match_values = match_values;
entry             411 drivers/net/ethernet/mellanox/mlxsw/spectrum_dpipe.c 	entry->match_values_count = MLXSW_SP_DPIPE_TABLE_HOST_MATCH_COUNT;
entry             413 drivers/net/ethernet/mellanox/mlxsw/spectrum_dpipe.c 	entry->action_values = action_value;
entry             414 drivers/net/ethernet/mellanox/mlxsw/spectrum_dpipe.c 	entry->action_values_count = 1;
entry             455 drivers/net/ethernet/mellanox/mlxsw/spectrum_dpipe.c __mlxsw_sp_dpipe_table_host_entry_fill(struct devlink_dpipe_entry *entry,
entry             464 drivers/net/ethernet/mellanox/mlxsw/spectrum_dpipe.c 	value = &entry->match_values[MLXSW_SP_DPIPE_TABLE_HOST_MATCH_RIF];
entry             472 drivers/net/ethernet/mellanox/mlxsw/spectrum_dpipe.c 	value = &entry->match_values[MLXSW_SP_DPIPE_TABLE_HOST_MATCH_DIP];
entry             476 drivers/net/ethernet/mellanox/mlxsw/spectrum_dpipe.c 	value = entry->action_values;
entry             482 drivers/net/ethernet/mellanox/mlxsw/spectrum_dpipe.c mlxsw_sp_dpipe_table_host4_entry_fill(struct devlink_dpipe_entry *entry,
entry             491 drivers/net/ethernet/mellanox/mlxsw/spectrum_dpipe.c 	__mlxsw_sp_dpipe_table_host_entry_fill(entry, rif, ha, &dip);
entry             495 drivers/net/ethernet/mellanox/mlxsw/spectrum_dpipe.c mlxsw_sp_dpipe_table_host6_entry_fill(struct devlink_dpipe_entry *entry,
entry             505 drivers/net/ethernet/mellanox/mlxsw/spectrum_dpipe.c 	__mlxsw_sp_dpipe_table_host_entry_fill(entry, rif, ha, dip);
entry             510 drivers/net/ethernet/mellanox/mlxsw/spectrum_dpipe.c 				     struct devlink_dpipe_entry *entry,
entry             519 drivers/net/ethernet/mellanox/mlxsw/spectrum_dpipe.c 		mlxsw_sp_dpipe_table_host4_entry_fill(entry, neigh_entry, rif);
entry             522 drivers/net/ethernet/mellanox/mlxsw/spectrum_dpipe.c 		mlxsw_sp_dpipe_table_host6_entry_fill(entry, neigh_entry, rif);
entry             530 drivers/net/ethernet/mellanox/mlxsw/spectrum_dpipe.c 					 &entry->counter);
entry             532 drivers/net/ethernet/mellanox/mlxsw/spectrum_dpipe.c 		entry->counter_valid = true;
entry             537 drivers/net/ethernet/mellanox/mlxsw/spectrum_dpipe.c 				      struct devlink_dpipe_entry *entry,
entry             579 drivers/net/ethernet/mellanox/mlxsw/spectrum_dpipe.c 			mlxsw_sp_dpipe_table_host_entry_fill(mlxsw_sp, entry,
entry             582 drivers/net/ethernet/mellanox/mlxsw/spectrum_dpipe.c 			entry->index = neigh_count;
entry             583 drivers/net/ethernet/mellanox/mlxsw/spectrum_dpipe.c 			err = devlink_dpipe_entry_ctx_append(dump_ctx, entry);
entry             624 drivers/net/ethernet/mellanox/mlxsw/spectrum_dpipe.c 	struct devlink_dpipe_entry entry = {0};
entry             634 drivers/net/ethernet/mellanox/mlxsw/spectrum_dpipe.c 	err = mlxsw_sp_dpipe_table_host_entry_prepare(&entry, match_values,
entry             640 drivers/net/ethernet/mellanox/mlxsw/spectrum_dpipe.c 	err = mlxsw_sp_dpipe_table_host_entries_get(mlxsw_sp, &entry,
entry             644 drivers/net/ethernet/mellanox/mlxsw/spectrum_dpipe.c 	devlink_dpipe_entry_clear(&entry);
entry             967 drivers/net/ethernet/mellanox/mlxsw/spectrum_dpipe.c mlxsw_sp_dpipe_table_adj_entry_prepare(struct devlink_dpipe_entry *entry,
entry             977 drivers/net/ethernet/mellanox/mlxsw/spectrum_dpipe.c 	entry->match_values = match_values;
entry             978 drivers/net/ethernet/mellanox/mlxsw/spectrum_dpipe.c 	entry->match_values_count = MLXSW_SP_DPIPE_TABLE_ADJ_MATCH_COUNT;
entry             980 drivers/net/ethernet/mellanox/mlxsw/spectrum_dpipe.c 	entry->action_values = action_values;
entry             981 drivers/net/ethernet/mellanox/mlxsw/spectrum_dpipe.c 	entry->action_values_count = MLXSW_SP_DPIPE_TABLE_ADJ_ACTION_COUNT;
entry            1032 drivers/net/ethernet/mellanox/mlxsw/spectrum_dpipe.c __mlxsw_sp_dpipe_table_adj_entry_fill(struct devlink_dpipe_entry *entry,
entry            1041 drivers/net/ethernet/mellanox/mlxsw/spectrum_dpipe.c 	value = &entry->match_values[MLXSW_SP_DPIPE_TABLE_ADJ_MATCH_INDEX];
entry            1045 drivers/net/ethernet/mellanox/mlxsw/spectrum_dpipe.c 	value = &entry->match_values[MLXSW_SP_DPIPE_TABLE_ADJ_MATCH_SIZE];
entry            1049 drivers/net/ethernet/mellanox/mlxsw/spectrum_dpipe.c 	value = &entry->match_values[MLXSW_SP_DPIPE_TABLE_ADJ_MATCH_HASH_INDEX];
entry            1053 drivers/net/ethernet/mellanox/mlxsw/spectrum_dpipe.c 	value = &entry->action_values[MLXSW_SP_DPIPE_TABLE_ADJ_ACTION_DST_MAC];
entry            1056 drivers/net/ethernet/mellanox/mlxsw/spectrum_dpipe.c 	value = &entry->action_values[MLXSW_SP_DPIPE_TABLE_ADJ_ACTION_ERIF_PORT];
entry            1065 drivers/net/ethernet/mellanox/mlxsw/spectrum_dpipe.c 						struct devlink_dpipe_entry *entry)
entry            1075 drivers/net/ethernet/mellanox/mlxsw/spectrum_dpipe.c 	__mlxsw_sp_dpipe_table_adj_entry_fill(entry, adj_index, adj_size,
entry            1077 drivers/net/ethernet/mellanox/mlxsw/spectrum_dpipe.c 	err = mlxsw_sp_nexthop_counter_get(mlxsw_sp, nh, &entry->counter);
entry            1079 drivers/net/ethernet/mellanox/mlxsw/spectrum_dpipe.c 		entry->counter_valid = true;
entry            1084 drivers/net/ethernet/mellanox/mlxsw/spectrum_dpipe.c 				     struct devlink_dpipe_entry *entry,
entry            1113 drivers/net/ethernet/mellanox/mlxsw/spectrum_dpipe.c 		mlxsw_sp_dpipe_table_adj_entry_fill(mlxsw_sp, nh, entry);
entry            1114 drivers/net/ethernet/mellanox/mlxsw/spectrum_dpipe.c 		entry->index = entry_index;
entry            1115 drivers/net/ethernet/mellanox/mlxsw/spectrum_dpipe.c 		err = devlink_dpipe_entry_ctx_append(dump_ctx, entry);
entry            1151 drivers/net/ethernet/mellanox/mlxsw/spectrum_dpipe.c 	struct devlink_dpipe_entry entry = {0};
entry            1165 drivers/net/ethernet/mellanox/mlxsw/spectrum_dpipe.c 	err = mlxsw_sp_dpipe_table_adj_entry_prepare(&entry,
entry            1171 drivers/net/ethernet/mellanox/mlxsw/spectrum_dpipe.c 	err = mlxsw_sp_dpipe_table_adj_entries_get(mlxsw_sp, &entry,
entry            1174 drivers/net/ethernet/mellanox/mlxsw/spectrum_dpipe.c 	devlink_dpipe_entry_clear(&entry);
entry            2641 drivers/net/ethernet/micrel/ksz884x.c 	int entry;
entry            2645 drivers/net/ethernet/micrel/ksz884x.c 	for (entry = 0; entry < VLAN_TABLE_ENTRIES; entry++) {
entry            2646 drivers/net/ethernet/micrel/ksz884x.c 		sw_r_vlan_table(hw, entry,
entry            2647 drivers/net/ethernet/micrel/ksz884x.c 			&sw->vlan_table[entry].vid,
entry            2648 drivers/net/ethernet/micrel/ksz884x.c 			&sw->vlan_table[entry].fid,
entry            2649 drivers/net/ethernet/micrel/ksz884x.c 			&sw->vlan_table[entry].member);
entry            2828 drivers/net/ethernet/micrel/ksz884x.c 	struct ksz_mac_table *entry;
entry            2832 drivers/net/ethernet/micrel/ksz884x.c 		entry = &hw->ksz_switch->mac_table[i];
entry            2834 drivers/net/ethernet/micrel/ksz884x.c 			entry->mac_addr, entry->ports,
entry            2835 drivers/net/ethernet/micrel/ksz884x.c 			entry->override, 0,
entry            2836 drivers/net/ethernet/micrel/ksz884x.c 			entry->use_fid, entry->fid);
entry            2848 drivers/net/ethernet/micrel/ksz884x.c 	struct ksz_mac_table *entry;
entry            2850 drivers/net/ethernet/micrel/ksz884x.c 	entry = &hw->ksz_switch->mac_table[STP_ENTRY];
entry            2851 drivers/net/ethernet/micrel/ksz884x.c 	entry->mac_addr[0] = 0x01;
entry            2852 drivers/net/ethernet/micrel/ksz884x.c 	entry->mac_addr[1] = 0x80;
entry            2853 drivers/net/ethernet/micrel/ksz884x.c 	entry->mac_addr[2] = 0xC2;
entry            2854 drivers/net/ethernet/micrel/ksz884x.c 	entry->mac_addr[3] = 0x00;
entry            2855 drivers/net/ethernet/micrel/ksz884x.c 	entry->mac_addr[4] = 0x00;
entry            2856 drivers/net/ethernet/micrel/ksz884x.c 	entry->mac_addr[5] = 0x00;
entry            2857 drivers/net/ethernet/micrel/ksz884x.c 	entry->ports = HOST_MASK;
entry            2858 drivers/net/ethernet/micrel/ksz884x.c 	entry->override = 1;
entry            2859 drivers/net/ethernet/micrel/ksz884x.c 	entry->valid = 1;
entry            2861 drivers/net/ethernet/micrel/ksz884x.c 		entry->mac_addr, entry->ports,
entry            2862 drivers/net/ethernet/micrel/ksz884x.c 		entry->override, entry->valid,
entry            2863 drivers/net/ethernet/micrel/ksz884x.c 		entry->use_fid, entry->fid);
entry            2874 drivers/net/ethernet/micrel/ksz884x.c 	struct ksz_mac_table *entry;
entry            2878 drivers/net/ethernet/micrel/ksz884x.c 		entry = &hw->ksz_switch->mac_table[i];
entry            2879 drivers/net/ethernet/micrel/ksz884x.c 		entry->valid = 0;
entry            2881 drivers/net/ethernet/micrel/ksz884x.c 			entry->mac_addr, entry->ports,
entry            2882 drivers/net/ethernet/micrel/ksz884x.c 			entry->override, entry->valid,
entry            2883 drivers/net/ethernet/micrel/ksz884x.c 			entry->use_fid, entry->fid);
entry             469 drivers/net/ethernet/microchip/lan743x_main.c 		msix_entries[index].entry = index;
entry             811 drivers/net/ethernet/mscc/ocelot.c static int ocelot_fdb_do_dump(struct ocelot_mact_entry *entry,
entry             836 drivers/net/ethernet/mscc/ocelot.c 	if (nla_put(dump->skb, NDA_LLADDR, ETH_ALEN, entry->mac))
entry             839 drivers/net/ethernet/mscc/ocelot.c 	if (entry->vid && nla_put_u16(dump->skb, NDA_VLAN, entry->vid))
entry             854 drivers/net/ethernet/mscc/ocelot.c 				   struct ocelot_mact_entry *entry)
entry             895 drivers/net/ethernet/mscc/ocelot.c 	entry->vid = (mach >> 16) & 0xfff;
entry             896 drivers/net/ethernet/mscc/ocelot.c 	ether_addr_copy(entry->mac, mac);
entry             914 drivers/net/ethernet/mscc/ocelot.c 	struct ocelot_mact_entry entry;
entry             921 drivers/net/ethernet/mscc/ocelot.c 			ret = ocelot_mact_read(port, i, j, &entry);
entry             930 drivers/net/ethernet/mscc/ocelot.c 			ret = ocelot_fdb_do_dump(&entry, &dump);
entry            2213 drivers/net/ethernet/mscc/ocelot.c 	struct ocelot_skb *entry;
entry            2227 drivers/net/ethernet/mscc/ocelot.c 			entry = list_entry(pos, struct ocelot_skb, head);
entry            2230 drivers/net/ethernet/mscc/ocelot.c 			dev_kfree_skb_any(entry->skb);
entry            2231 drivers/net/ethernet/mscc/ocelot.c 			kfree(entry);
entry              83 drivers/net/ethernet/mscc/ocelot_ace.c 	u32 entry[VCAP_ENTRY_WIDTH]; /* ENTRY_DAT */
entry             138 drivers/net/ethernet/mscc/ocelot_ace.c 		ocelot_write_rix(oc, data->entry[i], S2_CACHE_ENTRY_DAT, i);
entry             149 drivers/net/ethernet/mscc/ocelot_ace.c 		data->entry[i] = ocelot_read_rix(oc, S2_CACHE_ENTRY_DAT, i);
entry             248 drivers/net/ethernet/mscc/ocelot_ace.c 	vcap_data_set(data->entry, offset + data->key_offset, width, value);
entry             203 drivers/net/ethernet/mscc/ocelot_board.c 		struct ocelot_skb *entry;
entry             224 drivers/net/ethernet/mscc/ocelot_board.c 			entry = list_entry(pos, struct ocelot_skb, head);
entry             225 drivers/net/ethernet/mscc/ocelot_board.c 			if (entry->id != id)
entry             228 drivers/net/ethernet/mscc/ocelot_board.c 			skb = entry->skb;
entry             231 drivers/net/ethernet/mscc/ocelot_board.c 			kfree(entry);
entry             157 drivers/net/ethernet/myricom/myri10ge/myri10ge.c 	struct mcp_slot *entry;
entry             943 drivers/net/ethernet/myricom/myri10ge/myri10ge.c 	bytes = mgp->max_intr_slots * sizeof(*mgp->ss[0].rx_done.entry);
entry            1048 drivers/net/ethernet/myricom/myri10ge/myri10ge.c 		memset(ss->rx_done.entry, 0, bytes);
entry            1469 drivers/net/ethernet/myricom/myri10ge/myri10ge.c 	while (rx_done->entry[idx].length != 0 && work_done < budget) {
entry            1470 drivers/net/ethernet/myricom/myri10ge/myri10ge.c 		length = ntohs(rx_done->entry[idx].length);
entry            1471 drivers/net/ethernet/myricom/myri10ge/myri10ge.c 		rx_done->entry[idx].length = 0;
entry            1472 drivers/net/ethernet/myricom/myri10ge/myri10ge.c 		checksum = csum_unfold(rx_done->entry[idx].checksum);
entry            3566 drivers/net/ethernet/myricom/myri10ge/myri10ge.c 		if (ss->rx_done.entry != NULL) {
entry            3568 drivers/net/ethernet/myricom/myri10ge/myri10ge.c 			    sizeof(*ss->rx_done.entry);
entry            3570 drivers/net/ethernet/myricom/myri10ge/myri10ge.c 					  ss->rx_done.entry, ss->rx_done.bus);
entry            3571 drivers/net/ethernet/myricom/myri10ge/myri10ge.c 			ss->rx_done.entry = NULL;
entry            3603 drivers/net/ethernet/myricom/myri10ge/myri10ge.c 		bytes = mgp->max_intr_slots * sizeof(*ss->rx_done.entry);
entry            3604 drivers/net/ethernet/myricom/myri10ge/myri10ge.c 		ss->rx_done.entry = dma_alloc_coherent(&pdev->dev, bytes,
entry            3607 drivers/net/ethernet/myricom/myri10ge/myri10ge.c 		if (ss->rx_done.entry == NULL)
entry            3717 drivers/net/ethernet/myricom/myri10ge/myri10ge.c 		mgp->msix_vectors[i].entry = i;
entry            1935 drivers/net/ethernet/natsemi/natsemi.c 		int entry = np->dirty_rx % RX_RING_SIZE;
entry            1936 drivers/net/ethernet/natsemi/natsemi.c 		if (np->rx_skbuff[entry] == NULL) {
entry            1939 drivers/net/ethernet/natsemi/natsemi.c 			np->rx_skbuff[entry] = skb;
entry            1942 drivers/net/ethernet/natsemi/natsemi.c 			np->rx_dma[entry] = pci_map_single(np->pci_dev,
entry            1945 drivers/net/ethernet/natsemi/natsemi.c 						  np->rx_dma[entry])) {
entry            1947 drivers/net/ethernet/natsemi/natsemi.c 				np->rx_skbuff[entry] = NULL;
entry            1950 drivers/net/ethernet/natsemi/natsemi.c 			np->rx_ring[entry].addr = cpu_to_le32(np->rx_dma[entry]);
entry            1952 drivers/net/ethernet/natsemi/natsemi.c 		np->rx_ring[entry].cmd_status = cpu_to_le32(np->rx_buf_sz);
entry            2094 drivers/net/ethernet/natsemi/natsemi.c 	unsigned entry;
entry            2101 drivers/net/ethernet/natsemi/natsemi.c 	entry = np->cur_tx % TX_RING_SIZE;
entry            2103 drivers/net/ethernet/natsemi/natsemi.c 	np->tx_skbuff[entry] = skb;
entry            2104 drivers/net/ethernet/natsemi/natsemi.c 	np->tx_dma[entry] = pci_map_single(np->pci_dev,
entry            2106 drivers/net/ethernet/natsemi/natsemi.c 	if (pci_dma_mapping_error(np->pci_dev, np->tx_dma[entry])) {
entry            2107 drivers/net/ethernet/natsemi/natsemi.c 		np->tx_skbuff[entry] = NULL;
entry            2113 drivers/net/ethernet/natsemi/natsemi.c 	np->tx_ring[entry].addr = cpu_to_le32(np->tx_dma[entry]);
entry            2118 drivers/net/ethernet/natsemi/natsemi.c 		np->tx_ring[entry].cmd_status = cpu_to_le32(DescOwn | skb->len);
entry            2138 drivers/net/ethernet/natsemi/natsemi.c 			dev->name, np->cur_tx, entry);
entry            2148 drivers/net/ethernet/natsemi/natsemi.c 		int entry = np->dirty_tx % TX_RING_SIZE;
entry            2149 drivers/net/ethernet/natsemi/natsemi.c 		if (np->tx_ring[entry].cmd_status & cpu_to_le32(DescOwn))
entry            2155 drivers/net/ethernet/natsemi/natsemi.c 					le32_to_cpu(np->tx_ring[entry].cmd_status));
entry            2156 drivers/net/ethernet/natsemi/natsemi.c 		if (np->tx_ring[entry].cmd_status & cpu_to_le32(DescPktOK)) {
entry            2158 drivers/net/ethernet/natsemi/natsemi.c 			dev->stats.tx_bytes += np->tx_skbuff[entry]->len;
entry            2161 drivers/net/ethernet/natsemi/natsemi.c 				le32_to_cpu(np->tx_ring[entry].cmd_status);
entry            2172 drivers/net/ethernet/natsemi/natsemi.c 		pci_unmap_single(np->pci_dev,np->tx_dma[entry],
entry            2173 drivers/net/ethernet/natsemi/natsemi.c 					np->tx_skbuff[entry]->len,
entry            2176 drivers/net/ethernet/natsemi/natsemi.c 		dev_consume_skb_irq(np->tx_skbuff[entry]);
entry            2177 drivers/net/ethernet/natsemi/natsemi.c 		np->tx_skbuff[entry] = NULL;
entry            2285 drivers/net/ethernet/natsemi/natsemi.c 	int entry = np->cur_rx % RX_RING_SIZE;
entry            2297 drivers/net/ethernet/natsemi/natsemi.c 				entry, desc_status);
entry            2363 drivers/net/ethernet/natsemi/natsemi.c 					np->rx_dma[entry],
entry            2367 drivers/net/ethernet/natsemi/natsemi.c 					np->rx_skbuff[entry]->data, pkt_len);
entry            2370 drivers/net/ethernet/natsemi/natsemi.c 					np->rx_dma[entry],
entry            2374 drivers/net/ethernet/natsemi/natsemi.c 				pci_unmap_single(np->pci_dev, np->rx_dma[entry],
entry            2377 drivers/net/ethernet/natsemi/natsemi.c 				skb_put(skb = np->rx_skbuff[entry], pkt_len);
entry            2378 drivers/net/ethernet/natsemi/natsemi.c 				np->rx_skbuff[entry] = NULL;
entry            2385 drivers/net/ethernet/natsemi/natsemi.c 		entry = (++np->cur_rx) % RX_RING_SIZE;
entry            2386 drivers/net/ethernet/natsemi/natsemi.c 		np->rx_head_desc = &np->rx_ring[entry];
entry             242 drivers/net/ethernet/natsemi/sonic.c 	int entry;
entry             267 drivers/net/ethernet/natsemi/sonic.c 	entry = lp->next_tx;
entry             269 drivers/net/ethernet/natsemi/sonic.c 	sonic_tda_put(dev, entry, SONIC_TD_STATUS, 0);       /* clear status */
entry             270 drivers/net/ethernet/natsemi/sonic.c 	sonic_tda_put(dev, entry, SONIC_TD_FRAG_COUNT, 1);   /* single fragment */
entry             271 drivers/net/ethernet/natsemi/sonic.c 	sonic_tda_put(dev, entry, SONIC_TD_PKTSIZE, length); /* length of packet */
entry             272 drivers/net/ethernet/natsemi/sonic.c 	sonic_tda_put(dev, entry, SONIC_TD_FRAG_PTR_L, laddr & 0xffff);
entry             273 drivers/net/ethernet/natsemi/sonic.c 	sonic_tda_put(dev, entry, SONIC_TD_FRAG_PTR_H, laddr >> 16);
entry             274 drivers/net/ethernet/natsemi/sonic.c 	sonic_tda_put(dev, entry, SONIC_TD_FRAG_SIZE, length);
entry             275 drivers/net/ethernet/natsemi/sonic.c 	sonic_tda_put(dev, entry, SONIC_TD_LINK,
entry             276 drivers/net/ethernet/natsemi/sonic.c 		sonic_tda_get(dev, entry, SONIC_TD_LINK) | SONIC_EOL);
entry             279 drivers/net/ethernet/natsemi/sonic.c 	lp->tx_len[entry] = length;
entry             280 drivers/net/ethernet/natsemi/sonic.c 	lp->tx_laddr[entry] = laddr;
entry             281 drivers/net/ethernet/natsemi/sonic.c 	lp->tx_skb[entry] = skb;
entry             286 drivers/net/ethernet/natsemi/sonic.c 	lp->eol_tx = entry;
entry             288 drivers/net/ethernet/natsemi/sonic.c 	lp->next_tx = (entry + 1) & SONIC_TDS_MASK;
entry             339 drivers/net/ethernet/natsemi/sonic.c 			int entry = lp->cur_tx;
entry             353 drivers/net/ethernet/natsemi/sonic.c 			while (lp->tx_skb[entry] != NULL) {
entry             354 drivers/net/ethernet/natsemi/sonic.c 				if ((td_status = sonic_tda_get(dev, entry, SONIC_TD_STATUS)) == 0)
entry             359 drivers/net/ethernet/natsemi/sonic.c 					lp->stats.tx_bytes += sonic_tda_get(dev, entry, SONIC_TD_PKTSIZE);
entry             374 drivers/net/ethernet/natsemi/sonic.c 				dev_consume_skb_irq(lp->tx_skb[entry]);
entry             375 drivers/net/ethernet/natsemi/sonic.c 				lp->tx_skb[entry] = NULL;
entry             377 drivers/net/ethernet/natsemi/sonic.c 				dma_unmap_single(lp->device, lp->tx_laddr[entry], lp->tx_len[entry], DMA_TO_DEVICE);
entry             378 drivers/net/ethernet/natsemi/sonic.c 				lp->tx_laddr[entry] = (dma_addr_t)0;
entry             381 drivers/net/ethernet/natsemi/sonic.c 				if (sonic_tda_get(dev, entry, SONIC_TD_LINK) & SONIC_EOL) {
entry             382 drivers/net/ethernet/natsemi/sonic.c 					entry = (entry + 1) & SONIC_TDS_MASK;
entry             385 drivers/net/ethernet/natsemi/sonic.c 				entry = (entry + 1) & SONIC_TDS_MASK;
entry             388 drivers/net/ethernet/natsemi/sonic.c 			if (freed_some || lp->tx_skb[entry] == NULL)
entry             390 drivers/net/ethernet/natsemi/sonic.c 			lp->cur_tx = entry;
entry             490 drivers/net/ethernet/natsemi/sonic.c 	unsigned int entry = sonic_rr_entry(dev, SONIC_READ(SONIC_RWP));
entry             498 drivers/net/ethernet/natsemi/sonic.c 		buf = (sonic_rra_get(dev, entry, SONIC_RR_BUFADR_H) << 16) |
entry             499 drivers/net/ethernet/natsemi/sonic.c 		      sonic_rra_get(dev, entry, SONIC_RR_BUFADR_L);
entry             504 drivers/net/ethernet/natsemi/sonic.c 		entry = (entry + 1) & SONIC_RRS_MASK;
entry             505 drivers/net/ethernet/natsemi/sonic.c 	} while (entry != end);
entry             509 drivers/net/ethernet/natsemi/sonic.c 	sonic_rra_put(dev, entry, SONIC_RR_BUFADR_H, new_addr >> 16);
entry             510 drivers/net/ethernet/natsemi/sonic.c 	sonic_rra_put(dev, entry, SONIC_RR_BUFADR_L, new_addr & 0xffff);
entry             512 drivers/net/ethernet/natsemi/sonic.c 	entry = (entry + 1) & SONIC_RRS_MASK;
entry             514 drivers/net/ethernet/natsemi/sonic.c 	SONIC_WRITE(SONIC_RWP, sonic_rr_addr(dev, entry));
entry             523 drivers/net/ethernet/natsemi/sonic.c 	int entry = lp->cur_rx;
entry             527 drivers/net/ethernet/natsemi/sonic.c 	while (sonic_rda_get(dev, entry, SONIC_RD_IN_USE) == 0) {
entry             528 drivers/net/ethernet/natsemi/sonic.c 		u16 status = sonic_rda_get(dev, entry, SONIC_RD_STATUS);
entry             534 drivers/net/ethernet/natsemi/sonic.c 			u32 addr = (sonic_rda_get(dev, entry,
entry             536 drivers/net/ethernet/natsemi/sonic.c 				   sonic_rda_get(dev, entry, SONIC_RD_PKTPTR_L);
entry             537 drivers/net/ethernet/natsemi/sonic.c 			int i = index_from_addr(lp, addr, entry);
entry             552 drivers/net/ethernet/natsemi/sonic.c 				pkt_len = sonic_rda_get(dev, entry,
entry             577 drivers/net/ethernet/natsemi/sonic.c 		sonic_rda_put(dev, entry, SONIC_RD_STATUS, 0);
entry             578 drivers/net/ethernet/natsemi/sonic.c 		sonic_rda_put(dev, entry, SONIC_RD_IN_USE, 1);
entry             580 drivers/net/ethernet/natsemi/sonic.c 		prev_entry = entry;
entry             581 drivers/net/ethernet/natsemi/sonic.c 		entry = (entry + 1) & SONIC_RDS_MASK;
entry             584 drivers/net/ethernet/natsemi/sonic.c 	lp->cur_rx = entry;
entry             378 drivers/net/ethernet/natsemi/sonic.h static inline void sonic_cda_put(struct net_device* dev, int entry,
entry             383 drivers/net/ethernet/natsemi/sonic.h 		      (entry * SIZEOF_SONIC_CD) + offset, val);
entry             386 drivers/net/ethernet/natsemi/sonic.h static inline __u16 sonic_cda_get(struct net_device* dev, int entry,
entry             391 drivers/net/ethernet/natsemi/sonic.h 			     (entry * SIZEOF_SONIC_CD) + offset);
entry             406 drivers/net/ethernet/natsemi/sonic.h static inline void sonic_tda_put(struct net_device* dev, int entry,
entry             411 drivers/net/ethernet/natsemi/sonic.h 		      (entry * SIZEOF_SONIC_TD) + offset, val);
entry             414 drivers/net/ethernet/natsemi/sonic.h static inline __u16 sonic_tda_get(struct net_device* dev, int entry,
entry             419 drivers/net/ethernet/natsemi/sonic.h 			     (entry * SIZEOF_SONIC_TD) + offset);
entry             422 drivers/net/ethernet/natsemi/sonic.h static inline void sonic_rda_put(struct net_device* dev, int entry,
entry             427 drivers/net/ethernet/natsemi/sonic.h 		      (entry * SIZEOF_SONIC_RD) + offset, val);
entry             430 drivers/net/ethernet/natsemi/sonic.h static inline __u16 sonic_rda_get(struct net_device* dev, int entry,
entry             435 drivers/net/ethernet/natsemi/sonic.h 			     (entry * SIZEOF_SONIC_RD) + offset);
entry             438 drivers/net/ethernet/natsemi/sonic.h static inline void sonic_rra_put(struct net_device* dev, int entry,
entry             443 drivers/net/ethernet/natsemi/sonic.h 		      (entry * SIZEOF_SONIC_RR) + offset, val);
entry             446 drivers/net/ethernet/natsemi/sonic.h static inline __u16 sonic_rra_get(struct net_device* dev, int entry,
entry             451 drivers/net/ethernet/natsemi/sonic.h 			     (entry * SIZEOF_SONIC_RR) + offset);
entry             454 drivers/net/ethernet/natsemi/sonic.h static inline u16 sonic_rr_addr(struct net_device *dev, int entry)
entry             459 drivers/net/ethernet/natsemi/sonic.h 	       entry * SIZEOF_SONIC_RR * SONIC_BUS_SCALE(lp->dma_bitmode);
entry            3749 drivers/net/ethernet/neterion/s2io.c 	nic->entries[0].entry = 0;
entry            3750 drivers/net/ethernet/neterion/s2io.c 	nic->s2io_entries[0].entry = 0;
entry            3756 drivers/net/ethernet/neterion/s2io.c 		nic->entries[i].entry = ((i - 1) * 8) + 1;
entry            3757 drivers/net/ethernet/neterion/s2io.c 		nic->s2io_entries[i].entry = ((i - 1) * 8) + 1;
entry             833 drivers/net/ethernet/neterion/s2io.h 	u16 entry;
entry            1137 drivers/net/ethernet/neterion/vxge/vxge-config.c 	struct __vxge_hw_blockpool_entry *entry = NULL;
entry            1159 drivers/net/ethernet/neterion/vxge/vxge-config.c 		entry = kzalloc(sizeof(struct __vxge_hw_blockpool_entry),
entry            1161 drivers/net/ethernet/neterion/vxge/vxge-config.c 		if (entry == NULL) {
entry            1166 drivers/net/ethernet/neterion/vxge/vxge-config.c 		list_add(&entry->item, &blockpool->free_entry_list);
entry            1192 drivers/net/ethernet/neterion/vxge/vxge-config.c 			entry = (struct __vxge_hw_blockpool_entry *)
entry            1197 drivers/net/ethernet/neterion/vxge/vxge-config.c 		if (entry == NULL)
entry            1198 drivers/net/ethernet/neterion/vxge/vxge-config.c 			entry =
entry            1201 drivers/net/ethernet/neterion/vxge/vxge-config.c 		if (entry != NULL) {
entry            1202 drivers/net/ethernet/neterion/vxge/vxge-config.c 			list_del(&entry->item);
entry            1203 drivers/net/ethernet/neterion/vxge/vxge-config.c 			entry->length = VXGE_HW_BLOCK_SIZE;
entry            1204 drivers/net/ethernet/neterion/vxge/vxge-config.c 			entry->memblock = memblock;
entry            1205 drivers/net/ethernet/neterion/vxge/vxge-config.c 			entry->dma_addr = dma_addr;
entry            1206 drivers/net/ethernet/neterion/vxge/vxge-config.c 			entry->acc_handle = acc_handle;
entry            1207 drivers/net/ethernet/neterion/vxge/vxge-config.c 			entry->dma_handle = dma_handle;
entry            1208 drivers/net/ethernet/neterion/vxge/vxge-config.c 			list_add(&entry->item,
entry            2257 drivers/net/ethernet/neterion/vxge/vxge-config.c 	struct __vxge_hw_blockpool_entry *entry = NULL;
entry            2277 drivers/net/ethernet/neterion/vxge/vxge-config.c 		entry = (struct __vxge_hw_blockpool_entry *)
entry            2282 drivers/net/ethernet/neterion/vxge/vxge-config.c 	if (entry == NULL)
entry            2283 drivers/net/ethernet/neterion/vxge/vxge-config.c 		entry =	vmalloc(sizeof(struct __vxge_hw_blockpool_entry));
entry            2285 drivers/net/ethernet/neterion/vxge/vxge-config.c 		list_del(&entry->item);
entry            2287 drivers/net/ethernet/neterion/vxge/vxge-config.c 	if (entry) {
entry            2288 drivers/net/ethernet/neterion/vxge/vxge-config.c 		entry->length = length;
entry            2289 drivers/net/ethernet/neterion/vxge/vxge-config.c 		entry->memblock = block_addr;
entry            2290 drivers/net/ethernet/neterion/vxge/vxge-config.c 		entry->dma_addr = dma_addr;
entry            2291 drivers/net/ethernet/neterion/vxge/vxge-config.c 		entry->acc_handle = acc_handle;
entry            2292 drivers/net/ethernet/neterion/vxge/vxge-config.c 		entry->dma_handle = dma_h;
entry            2293 drivers/net/ethernet/neterion/vxge/vxge-config.c 		list_add(&entry->item, &blockpool->free_block_list);
entry            2347 drivers/net/ethernet/neterion/vxge/vxge-config.c 	struct __vxge_hw_blockpool_entry *entry = NULL;
entry            2376 drivers/net/ethernet/neterion/vxge/vxge-config.c 			entry = (struct __vxge_hw_blockpool_entry *)
entry            2381 drivers/net/ethernet/neterion/vxge/vxge-config.c 		if (entry != NULL) {
entry            2382 drivers/net/ethernet/neterion/vxge/vxge-config.c 			list_del(&entry->item);
entry            2383 drivers/net/ethernet/neterion/vxge/vxge-config.c 			dma_object->addr = entry->dma_addr;
entry            2384 drivers/net/ethernet/neterion/vxge/vxge-config.c 			dma_object->handle = entry->dma_handle;
entry            2385 drivers/net/ethernet/neterion/vxge/vxge-config.c 			dma_object->acc_handle = entry->acc_handle;
entry            2386 drivers/net/ethernet/neterion/vxge/vxge-config.c 			memblock = entry->memblock;
entry            2388 drivers/net/ethernet/neterion/vxge/vxge-config.c 			list_add(&entry->item,
entry            2441 drivers/net/ethernet/neterion/vxge/vxge-config.c 	struct __vxge_hw_blockpool_entry *entry = NULL;
entry            2454 drivers/net/ethernet/neterion/vxge/vxge-config.c 			entry = (struct __vxge_hw_blockpool_entry *)
entry            2459 drivers/net/ethernet/neterion/vxge/vxge-config.c 		if (entry == NULL)
entry            2460 drivers/net/ethernet/neterion/vxge/vxge-config.c 			entry =	vmalloc(sizeof(
entry            2463 drivers/net/ethernet/neterion/vxge/vxge-config.c 			list_del(&entry->item);
entry            2465 drivers/net/ethernet/neterion/vxge/vxge-config.c 		if (entry != NULL) {
entry            2466 drivers/net/ethernet/neterion/vxge/vxge-config.c 			entry->length = size;
entry            2467 drivers/net/ethernet/neterion/vxge/vxge-config.c 			entry->memblock = memblock;
entry            2468 drivers/net/ethernet/neterion/vxge/vxge-config.c 			entry->dma_addr = dma_object->addr;
entry            2469 drivers/net/ethernet/neterion/vxge/vxge-config.c 			entry->acc_handle = dma_object->acc_handle;
entry            2470 drivers/net/ethernet/neterion/vxge/vxge-config.c 			entry->dma_handle = dma_object->handle;
entry            2471 drivers/net/ethernet/neterion/vxge/vxge-config.c 			list_add(&entry->item,
entry            4763 drivers/net/ethernet/neterion/vxge/vxge-config.c 	struct __vxge_hw_blockpool_entry *entry = NULL;
entry            4771 drivers/net/ethernet/neterion/vxge/vxge-config.c 			entry = (struct __vxge_hw_blockpool_entry *)
entry            4776 drivers/net/ethernet/neterion/vxge/vxge-config.c 		if (entry != NULL) {
entry            4777 drivers/net/ethernet/neterion/vxge/vxge-config.c 			list_del(&entry->item);
entry            4782 drivers/net/ethernet/neterion/vxge/vxge-config.c 	if (entry != NULL)
entry            4785 drivers/net/ethernet/neterion/vxge/vxge-config.c 	return entry;
entry            4944 drivers/net/ethernet/neterion/vxge/vxge-config.c 			       struct __vxge_hw_blockpool_entry *entry)
entry            4950 drivers/net/ethernet/neterion/vxge/vxge-config.c 	if (entry->length == blockpool->block_size) {
entry            4951 drivers/net/ethernet/neterion/vxge/vxge-config.c 		list_add(&entry->item, &blockpool->free_block_list);
entry             658 drivers/net/ethernet/neterion/vxge/vxge-main.c 	struct list_head *entry, *next;
entry             659 drivers/net/ethernet/neterion/vxge/vxge-main.c 	list_for_each_safe(entry, next, &vpath->mac_addr_list) {
entry             660 drivers/net/ethernet/neterion/vxge/vxge-main.c 		if (((struct vxge_mac_addrs *)entry)->macaddr == del_mac)
entry            1068 drivers/net/ethernet/neterion/vxge/vxge-main.c 	struct list_head *entry, *next;
entry            1075 drivers/net/ethernet/neterion/vxge/vxge-main.c 	list_for_each_safe(entry, next, &vpath->mac_addr_list) {
entry            1076 drivers/net/ethernet/neterion/vxge/vxge-main.c 		if (((struct vxge_mac_addrs *)entry)->macaddr == del_mac) {
entry            1077 drivers/net/ethernet/neterion/vxge/vxge-main.c 			list_del(entry);
entry            1078 drivers/net/ethernet/neterion/vxge/vxge-main.c 			kfree((struct vxge_mac_addrs *)entry);
entry            1131 drivers/net/ethernet/neterion/vxge/vxge-main.c 	struct list_head *entry, *next;
entry            1195 drivers/net/ethernet/neterion/vxge/vxge-main.c 			list_for_each_safe(entry, next, list_head) {
entry            1196 drivers/net/ethernet/neterion/vxge/vxge-main.c 				mac_entry = (struct vxge_mac_addrs *)entry;
entry            1237 drivers/net/ethernet/neterion/vxge/vxge-main.c 			list_for_each_safe(entry, next, list_head) {
entry            1238 drivers/net/ethernet/neterion/vxge/vxge-main.c 				mac_entry = (struct vxge_mac_addrs *)entry;
entry            1446 drivers/net/ethernet/neterion/vxge/vxge-main.c 	struct list_head *entry, *next;
entry            1451 drivers/net/ethernet/neterion/vxge/vxge-main.c 		list_for_each_safe(entry, next, &vpath->mac_addr_list) {
entry            1454 drivers/net/ethernet/neterion/vxge/vxge-main.c 				((struct vxge_mac_addrs *)entry)->macaddr;
entry            1456 drivers/net/ethernet/neterion/vxge/vxge-main.c 			((struct vxge_mac_addrs *)entry)->state =
entry            1471 drivers/net/ethernet/neterion/vxge/vxge-main.c 					((struct vxge_mac_addrs *)entry)->state
entry            2325 drivers/net/ethernet/neterion/vxge/vxge-main.c 		vdev->entries[j].entry = msix_intr_vect;
entry            2326 drivers/net/ethernet/neterion/vxge/vxge-main.c 		vdev->vxge_entries[j].entry = msix_intr_vect;
entry            2331 drivers/net/ethernet/neterion/vxge/vxge-main.c 		vdev->entries[j].entry = msix_intr_vect + 1;
entry            2332 drivers/net/ethernet/neterion/vxge/vxge-main.c 		vdev->vxge_entries[j].entry = msix_intr_vect + 1;
entry            2338 drivers/net/ethernet/neterion/vxge/vxge-main.c 	vdev->entries[j].entry = VXGE_ALARM_MSIX_ID;
entry            2339 drivers/net/ethernet/neterion/vxge/vxge-main.c 	vdev->vxge_entries[j].entry = VXGE_ALARM_MSIX_ID;
entry            2474 drivers/net/ethernet/neterion/vxge/vxge-main.c 					vdev->entries[intr_cnt].entry,
entry            2489 drivers/net/ethernet/neterion/vxge/vxge-main.c 					vdev->entries[intr_cnt].entry,
entry            2535 drivers/net/ethernet/neterion/vxge/vxge-main.c 			vdev->entries[intr_cnt].entry,
entry            2909 drivers/net/ethernet/neterion/vxge/vxge-main.c 	struct list_head *entry, *next;
entry            2913 drivers/net/ethernet/neterion/vxge/vxge-main.c 	list_for_each_safe(entry, next, &vpath->mac_addr_list) {
entry            2914 drivers/net/ethernet/neterion/vxge/vxge-main.c 		list_del(entry);
entry            2915 drivers/net/ethernet/neterion/vxge/vxge-main.c 		kfree((struct vxge_mac_addrs *)entry);
entry            4311 drivers/net/ethernet/neterion/vxge/vxge-main.c 	struct vxge_mac_addrs *entry;
entry            4685 drivers/net/ethernet/neterion/vxge/vxge-main.c 		entry =	kzalloc(sizeof(struct vxge_mac_addrs), GFP_KERNEL);
entry            4686 drivers/net/ethernet/neterion/vxge/vxge-main.c 		if (NULL == entry) {
entry            4693 drivers/net/ethernet/neterion/vxge/vxge-main.c 		macaddr = (u8 *)&entry->macaddr;
entry            4695 drivers/net/ethernet/neterion/vxge/vxge-main.c 		list_add(&entry->item, &vdev->vpaths[i].mac_addr_list);
entry             198 drivers/net/ethernet/neterion/vxge/vxge-main.h 	u16 entry;
entry             145 drivers/net/ethernet/netronome/nfp/flower/lag_conf.c 	struct nfp_fl_lag_group *entry;
entry             150 drivers/net/ethernet/netronome/nfp/flower/lag_conf.c 	list_for_each_entry(entry, &lag->group_list, list)
entry             151 drivers/net/ethernet/netronome/nfp/flower/lag_conf.c 		if (entry->master_ndev == master)
entry             152 drivers/net/ethernet/netronome/nfp/flower/lag_conf.c 			return entry;
entry             267 drivers/net/ethernet/netronome/nfp/flower/lag_conf.c 	struct nfp_fl_lag_group *entry, *storage;
entry             278 drivers/net/ethernet/netronome/nfp/flower/lag_conf.c 	list_for_each_entry_safe(entry, storage, &lag->group_list, list) {
entry             285 drivers/net/ethernet/netronome/nfp/flower/lag_conf.c 		if (entry->to_remove) {
entry             287 drivers/net/ethernet/netronome/nfp/flower/lag_conf.c 			err = nfp_fl_lag_config_group(lag, entry, NULL, 0,
entry             290 drivers/net/ethernet/netronome/nfp/flower/lag_conf.c 				entry->to_remove = false;
entry             291 drivers/net/ethernet/netronome/nfp/flower/lag_conf.c 				entry->offloaded = false;
entry             300 drivers/net/ethernet/netronome/nfp/flower/lag_conf.c 			if (entry->to_destroy) {
entry             302 drivers/net/ethernet/netronome/nfp/flower/lag_conf.c 						  entry->group_id);
entry             303 drivers/net/ethernet/netronome/nfp/flower/lag_conf.c 				list_del(&entry->list);
entry             304 drivers/net/ethernet/netronome/nfp/flower/lag_conf.c 				kfree(entry);
entry             309 drivers/net/ethernet/netronome/nfp/flower/lag_conf.c 		acti_netdevs = kmalloc_array(entry->slave_cnt,
entry             319 drivers/net/ethernet/netronome/nfp/flower/lag_conf.c 		for_each_netdev_in_bond_rcu(entry->master_ndev, iter_netdev) {
entry             333 drivers/net/ethernet/netronome/nfp/flower/lag_conf.c 			if (slaves > entry->slave_cnt)
entry             342 drivers/net/ethernet/netronome/nfp/flower/lag_conf.c 				entry->dirty = true;
entry             351 drivers/net/ethernet/netronome/nfp/flower/lag_conf.c 		if (slaves != entry->slave_cnt || !entry->dirty) {
entry             356 drivers/net/ethernet/netronome/nfp/flower/lag_conf.c 		err = nfp_fl_lag_config_group(lag, entry, acti_netdevs,
entry             359 drivers/net/ethernet/netronome/nfp/flower/lag_conf.c 			entry->offloaded = true;
entry             360 drivers/net/ethernet/netronome/nfp/flower/lag_conf.c 			entry->dirty = false;
entry             678 drivers/net/ethernet/netronome/nfp/flower/lag_conf.c 	struct nfp_fl_lag_group *entry, *storage;
entry             686 drivers/net/ethernet/netronome/nfp/flower/lag_conf.c 	list_for_each_entry_safe(entry, storage, &lag->group_list, list) {
entry             687 drivers/net/ethernet/netronome/nfp/flower/lag_conf.c 		list_del(&entry->list);
entry             688 drivers/net/ethernet/netronome/nfp/flower/lag_conf.c 		kfree(entry);
entry              42 drivers/net/ethernet/netronome/nfp/flower/main.c 	struct net_device *entry;
entry              46 drivers/net/ethernet/netronome/nfp/flower/main.c 	idr_for_each_entry(&priv->internal_ports.port_ids, entry, i)
entry              47 drivers/net/ethernet/netronome/nfp/flower/main.c 		if (entry == netdev) {
entry             149 drivers/net/ethernet/netronome/nfp/flower/main.c 	struct nfp_flower_non_repr_priv *entry;
entry             153 drivers/net/ethernet/netronome/nfp/flower/main.c 	list_for_each_entry(entry, &priv->non_repr_priv, list)
entry             154 drivers/net/ethernet/netronome/nfp/flower/main.c 		if (entry->netdev == netdev)
entry             155 drivers/net/ethernet/netronome/nfp/flower/main.c 			return entry;
entry             170 drivers/net/ethernet/netronome/nfp/flower/main.c 	struct nfp_flower_non_repr_priv *entry;
entry             172 drivers/net/ethernet/netronome/nfp/flower/main.c 	entry = nfp_flower_non_repr_priv_lookup(app, netdev);
entry             173 drivers/net/ethernet/netronome/nfp/flower/main.c 	if (entry)
entry             176 drivers/net/ethernet/netronome/nfp/flower/main.c 	entry = kzalloc(sizeof(*entry), GFP_KERNEL);
entry             177 drivers/net/ethernet/netronome/nfp/flower/main.c 	if (!entry)
entry             180 drivers/net/ethernet/netronome/nfp/flower/main.c 	entry->netdev = netdev;
entry             181 drivers/net/ethernet/netronome/nfp/flower/main.c 	list_add(&entry->list, &priv->non_repr_priv);
entry             184 drivers/net/ethernet/netronome/nfp/flower/main.c 	__nfp_flower_non_repr_priv_get(entry);
entry             185 drivers/net/ethernet/netronome/nfp/flower/main.c 	return entry;
entry             201 drivers/net/ethernet/netronome/nfp/flower/main.c 	struct nfp_flower_non_repr_priv *entry;
entry             203 drivers/net/ethernet/netronome/nfp/flower/main.c 	entry = nfp_flower_non_repr_priv_lookup(app, netdev);
entry             204 drivers/net/ethernet/netronome/nfp/flower/main.c 	if (!entry)
entry             207 drivers/net/ethernet/netronome/nfp/flower/main.c 	__nfp_flower_non_repr_priv_put(entry);
entry             230 drivers/net/ethernet/netronome/nfp/flower/tunnel_conf.c 	struct nfp_ipv4_route_entry *entry;
entry             235 drivers/net/ethernet/netronome/nfp/flower/tunnel_conf.c 		entry = list_entry(ptr, struct nfp_ipv4_route_entry, list);
entry             236 drivers/net/ethernet/netronome/nfp/flower/tunnel_conf.c 		if (entry->ipv4_addr == ipv4_addr) {
entry             248 drivers/net/ethernet/netronome/nfp/flower/tunnel_conf.c 	struct nfp_ipv4_route_entry *entry;
entry             253 drivers/net/ethernet/netronome/nfp/flower/tunnel_conf.c 		entry = list_entry(ptr, struct nfp_ipv4_route_entry, list);
entry             254 drivers/net/ethernet/netronome/nfp/flower/tunnel_conf.c 		if (entry->ipv4_addr == ipv4_addr) {
entry             259 drivers/net/ethernet/netronome/nfp/flower/tunnel_conf.c 	entry = kmalloc(sizeof(*entry), GFP_ATOMIC);
entry             260 drivers/net/ethernet/netronome/nfp/flower/tunnel_conf.c 	if (!entry) {
entry             266 drivers/net/ethernet/netronome/nfp/flower/tunnel_conf.c 	entry->ipv4_addr = ipv4_addr;
entry             267 drivers/net/ethernet/netronome/nfp/flower/tunnel_conf.c 	list_add_tail(&entry->list, &priv->tun.neigh_off_list);
entry             274 drivers/net/ethernet/netronome/nfp/flower/tunnel_conf.c 	struct nfp_ipv4_route_entry *entry;
entry             279 drivers/net/ethernet/netronome/nfp/flower/tunnel_conf.c 		entry = list_entry(ptr, struct nfp_ipv4_route_entry, list);
entry             280 drivers/net/ethernet/netronome/nfp/flower/tunnel_conf.c 		if (entry->ipv4_addr == ipv4_addr) {
entry             281 drivers/net/ethernet/netronome/nfp/flower/tunnel_conf.c 			list_del(&entry->list);
entry             282 drivers/net/ethernet/netronome/nfp/flower/tunnel_conf.c 			kfree(entry);
entry             427 drivers/net/ethernet/netronome/nfp/flower/tunnel_conf.c 	struct nfp_ipv4_addr_entry *entry;
entry             441 drivers/net/ethernet/netronome/nfp/flower/tunnel_conf.c 		entry = list_entry(ptr, struct nfp_ipv4_addr_entry, list);
entry             442 drivers/net/ethernet/netronome/nfp/flower/tunnel_conf.c 		payload.ipv4_addr[count++] = entry->ipv4_addr;
entry             455 drivers/net/ethernet/netronome/nfp/flower/tunnel_conf.c 	struct nfp_ipv4_addr_entry *entry;
entry             460 drivers/net/ethernet/netronome/nfp/flower/tunnel_conf.c 		entry = list_entry(ptr, struct nfp_ipv4_addr_entry, list);
entry             461 drivers/net/ethernet/netronome/nfp/flower/tunnel_conf.c 		if (entry->ipv4_addr == ipv4) {
entry             462 drivers/net/ethernet/netronome/nfp/flower/tunnel_conf.c 			entry->ref_count++;
entry             468 drivers/net/ethernet/netronome/nfp/flower/tunnel_conf.c 	entry = kmalloc(sizeof(*entry), GFP_KERNEL);
entry             469 drivers/net/ethernet/netronome/nfp/flower/tunnel_conf.c 	if (!entry) {
entry             474 drivers/net/ethernet/netronome/nfp/flower/tunnel_conf.c 	entry->ipv4_addr = ipv4;
entry             475 drivers/net/ethernet/netronome/nfp/flower/tunnel_conf.c 	entry->ref_count = 1;
entry             476 drivers/net/ethernet/netronome/nfp/flower/tunnel_conf.c 	list_add_tail(&entry->list, &priv->tun.ipv4_off_list);
entry             485 drivers/net/ethernet/netronome/nfp/flower/tunnel_conf.c 	struct nfp_ipv4_addr_entry *entry;
entry             490 drivers/net/ethernet/netronome/nfp/flower/tunnel_conf.c 		entry = list_entry(ptr, struct nfp_ipv4_addr_entry, list);
entry             491 drivers/net/ethernet/netronome/nfp/flower/tunnel_conf.c 		if (entry->ipv4_addr == ipv4) {
entry             492 drivers/net/ethernet/netronome/nfp/flower/tunnel_conf.c 			entry->ref_count--;
entry             493 drivers/net/ethernet/netronome/nfp/flower/tunnel_conf.c 			if (!entry->ref_count) {
entry             494 drivers/net/ethernet/netronome/nfp/flower/tunnel_conf.c 				list_del(&entry->list);
entry             495 drivers/net/ethernet/netronome/nfp/flower/tunnel_conf.c 				kfree(entry);
entry             564 drivers/net/ethernet/netronome/nfp/flower/tunnel_conf.c nfp_tunnel_offloaded_macs_inc_ref_and_link(struct nfp_tun_offloaded_mac *entry,
entry             578 drivers/net/ethernet/netronome/nfp/flower/tunnel_conf.c 		list_add_tail(&repr_priv->mac_list, &entry->repr_list);
entry             580 drivers/net/ethernet/netronome/nfp/flower/tunnel_conf.c 		entry->bridge_count++;
entry             583 drivers/net/ethernet/netronome/nfp/flower/tunnel_conf.c 	entry->ref_count++;
entry             592 drivers/net/ethernet/netronome/nfp/flower/tunnel_conf.c 	struct nfp_tun_offloaded_mac *entry;
entry             595 drivers/net/ethernet/netronome/nfp/flower/tunnel_conf.c 	entry = nfp_tunnel_lookup_offloaded_macs(app, netdev->dev_addr);
entry             596 drivers/net/ethernet/netronome/nfp/flower/tunnel_conf.c 	if (entry && nfp_tunnel_is_mac_idx_global(entry->index)) {
entry             597 drivers/net/ethernet/netronome/nfp/flower/tunnel_conf.c 		if (entry->bridge_count ||
entry             599 drivers/net/ethernet/netronome/nfp/flower/tunnel_conf.c 			nfp_tunnel_offloaded_macs_inc_ref_and_link(entry,
entry             605 drivers/net/ethernet/netronome/nfp/flower/tunnel_conf.c 		nfp_mac_idx = entry->index | NFP_TUN_PRE_TUN_IDX_BIT;
entry             610 drivers/net/ethernet/netronome/nfp/flower/tunnel_conf.c 		if (entry || !port) {
entry             628 drivers/net/ethernet/netronome/nfp/flower/tunnel_conf.c 	if (!entry) {
entry             629 drivers/net/ethernet/netronome/nfp/flower/tunnel_conf.c 		entry = kzalloc(sizeof(*entry), GFP_KERNEL);
entry             630 drivers/net/ethernet/netronome/nfp/flower/tunnel_conf.c 		if (!entry) {
entry             635 drivers/net/ethernet/netronome/nfp/flower/tunnel_conf.c 		ether_addr_copy(entry->addr, netdev->dev_addr);
entry             636 drivers/net/ethernet/netronome/nfp/flower/tunnel_conf.c 		INIT_LIST_HEAD(&entry->repr_list);
entry             639 drivers/net/ethernet/netronome/nfp/flower/tunnel_conf.c 					   &entry->ht_node,
entry             650 drivers/net/ethernet/netronome/nfp/flower/tunnel_conf.c 		if (!entry->ref_count)
entry             655 drivers/net/ethernet/netronome/nfp/flower/tunnel_conf.c 	entry->index = nfp_mac_idx;
entry             656 drivers/net/ethernet/netronome/nfp/flower/tunnel_conf.c 	nfp_tunnel_offloaded_macs_inc_ref_and_link(entry, netdev, mod);
entry             661 drivers/net/ethernet/netronome/nfp/flower/tunnel_conf.c 	rhashtable_remove_fast(&priv->tun.offloaded_macs, &entry->ht_node,
entry             664 drivers/net/ethernet/netronome/nfp/flower/tunnel_conf.c 	kfree(entry);
entry             678 drivers/net/ethernet/netronome/nfp/flower/tunnel_conf.c 	struct nfp_tun_offloaded_mac *entry;
entry             682 drivers/net/ethernet/netronome/nfp/flower/tunnel_conf.c 	entry = nfp_tunnel_lookup_offloaded_macs(app, mac);
entry             683 drivers/net/ethernet/netronome/nfp/flower/tunnel_conf.c 	if (!entry)
entry             686 drivers/net/ethernet/netronome/nfp/flower/tunnel_conf.c 	entry->ref_count--;
entry             695 drivers/net/ethernet/netronome/nfp/flower/tunnel_conf.c 		entry->bridge_count--;
entry             697 drivers/net/ethernet/netronome/nfp/flower/tunnel_conf.c 		if (!entry->bridge_count && entry->ref_count) {
entry             700 drivers/net/ethernet/netronome/nfp/flower/tunnel_conf.c 			nfp_mac_idx = entry->index & ~NFP_TUN_PRE_TUN_IDX_BIT;
entry             708 drivers/net/ethernet/netronome/nfp/flower/tunnel_conf.c 			entry->index = nfp_mac_idx;
entry             714 drivers/net/ethernet/netronome/nfp/flower/tunnel_conf.c 	if (entry->ref_count == 1 && list_is_singular(&entry->repr_list)) {
entry             718 drivers/net/ethernet/netronome/nfp/flower/tunnel_conf.c 		repr_priv = list_first_entry(&entry->repr_list,
entry             731 drivers/net/ethernet/netronome/nfp/flower/tunnel_conf.c 		ida_idx = nfp_tunnel_get_ida_from_global_mac_idx(entry->index);
entry             733 drivers/net/ethernet/netronome/nfp/flower/tunnel_conf.c 		entry->index = nfp_mac_idx;
entry             737 drivers/net/ethernet/netronome/nfp/flower/tunnel_conf.c 	if (entry->ref_count)
entry             741 drivers/net/ethernet/netronome/nfp/flower/tunnel_conf.c 					    &entry->ht_node,
entry             744 drivers/net/ethernet/netronome/nfp/flower/tunnel_conf.c 	if (nfp_tunnel_is_mac_idx_global(entry->index)) {
entry             745 drivers/net/ethernet/netronome/nfp/flower/tunnel_conf.c 		ida_idx = nfp_tunnel_get_ida_from_global_mac_idx(entry->index);
entry             749 drivers/net/ethernet/netronome/nfp/flower/tunnel_conf.c 	kfree(entry);
entry             406 drivers/net/ethernet/netronome/nfp/nfp_net_common.c 		irq_entries[i].entry = i;
entry             537 drivers/net/ethernet/netronome/nfp/nfp_net_common.c 	struct msix_entry *entry;
entry             539 drivers/net/ethernet/netronome/nfp/nfp_net_common.c 	entry = &nn->irq_entries[NFP_NET_IRQ_LSC_IDX];
entry             543 drivers/net/ethernet/netronome/nfp/nfp_net_common.c 	nfp_net_irq_unmask(nn, entry->entry);
entry             622 drivers/net/ethernet/netronome/nfp/nfp_net_common.c 	struct msix_entry *entry;
entry             625 drivers/net/ethernet/netronome/nfp/nfp_net_common.c 	entry = &nn->irq_entries[vector_idx];
entry             628 drivers/net/ethernet/netronome/nfp/nfp_net_common.c 	err = request_irq(entry->vector, handler, 0, name, nn);
entry             631 drivers/net/ethernet/netronome/nfp/nfp_net_common.c 		       entry->vector, err);
entry             634 drivers/net/ethernet/netronome/nfp/nfp_net_common.c 	nn_writeb(nn, ctrl_offset, entry->entry);
entry             635 drivers/net/ethernet/netronome/nfp/nfp_net_common.c 	nfp_net_irq_unmask(nn, entry->entry);
entry            2318 drivers/net/ethernet/netronome/nfp/nfp_net_common.c 		struct msix_entry *entry;
entry            2320 drivers/net/ethernet/netronome/nfp/nfp_net_common.c 		entry = &nn->irq_entries[NFP_NET_NON_Q_VECTORS + r];
entry            2324 drivers/net/ethernet/netronome/nfp/nfp_net_common.c 		r_vec->irq_entry = entry->entry;
entry            2325 drivers/net/ethernet/netronome/nfp/nfp_net_common.c 		r_vec->irq_vector = entry->vector;
entry              82 drivers/net/ethernet/netronome/nfp/nfpcore/nfp_cppcore.c 	struct list_head entry;
entry             167 drivers/net/ethernet/netronome/nfp/nfpcore/nfp_cppcore.c 	list_for_each_entry_safe(cache, ctmp, &cpp->area_cache_list, entry) {
entry             168 drivers/net/ethernet/netronome/nfp/nfpcore/nfp_cppcore.c 		list_del(&cache->entry);
entry             814 drivers/net/ethernet/netronome/nfp/nfpcore/nfp_cppcore.c 	list_add_tail(&cache->entry, &cpp->area_cache_list);
entry             849 drivers/net/ethernet/netronome/nfp/nfpcore/nfp_cppcore.c 	list_for_each_entry(cache, &cpp->area_cache_list, entry) {
entry             858 drivers/net/ethernet/netronome/nfp/nfpcore/nfp_cppcore.c 			   struct nfp_cpp_area_cache, entry);
entry             908 drivers/net/ethernet/netronome/nfp/nfpcore/nfp_cppcore.c 	list_del(&cache->entry);
entry             909 drivers/net/ethernet/netronome/nfp/nfpcore/nfp_cppcore.c 	list_add(&cache->entry, &cpp->area_cache_list);
entry             204 drivers/net/ethernet/netronome/nfp/nfpcore/nfp_nsp_eth.c nfp_eth_calc_port_type(struct nfp_cpp *cpp, struct nfp_eth_table_port *entry)
entry             206 drivers/net/ethernet/netronome/nfp/nfpcore/nfp_nsp_eth.c 	if (entry->interface == NFP_INTERFACE_NONE) {
entry             207 drivers/net/ethernet/netronome/nfp/nfpcore/nfp_nsp_eth.c 		entry->port_type = PORT_NONE;
entry             209 drivers/net/ethernet/netronome/nfp/nfpcore/nfp_nsp_eth.c 	} else if (entry->interface == NFP_INTERFACE_RJ45) {
entry             210 drivers/net/ethernet/netronome/nfp/nfpcore/nfp_nsp_eth.c 		entry->port_type = PORT_TP;
entry             214 drivers/net/ethernet/netronome/nfp/nfpcore/nfp_nsp_eth.c 	if (entry->media == NFP_MEDIA_FIBRE)
entry             215 drivers/net/ethernet/netronome/nfp/nfpcore/nfp_nsp_eth.c 		entry->port_type = PORT_FIBRE;
entry             217 drivers/net/ethernet/netronome/nfp/nfpcore/nfp_nsp_eth.c 		entry->port_type = PORT_DA;
entry              71 drivers/net/ethernet/netronome/nfp/nfpcore/nfp_resource.c 	struct nfp_resource_entry entry;
entry              88 drivers/net/ethernet/netronome/nfp/nfpcore/nfp_resource.c 		ret = nfp_cpp_read(cpp, cpp_id, addr, &entry, sizeof(entry));
entry              89 drivers/net/ethernet/netronome/nfp/nfpcore/nfp_resource.c 		if (ret != sizeof(entry))
entry              92 drivers/net/ethernet/netronome/nfp/nfpcore/nfp_resource.c 		if (entry.mutex.key != key)
entry              99 drivers/net/ethernet/netronome/nfp/nfpcore/nfp_resource.c 		res->cpp_id = NFP_CPP_ID(entry.region.cpp_target,
entry             100 drivers/net/ethernet/netronome/nfp/nfpcore/nfp_resource.c 					 entry.region.cpp_action,
entry             101 drivers/net/ethernet/netronome/nfp/nfpcore/nfp_resource.c 					 entry.region.cpp_token);
entry             102 drivers/net/ethernet/netronome/nfp/nfpcore/nfp_resource.c 		res->addr = (u64)entry.region.page_offset << 8;
entry             103 drivers/net/ethernet/netronome/nfp/nfpcore/nfp_resource.c 		res->size = (u64)entry.region.page_size << 8;
entry            3996 drivers/net/ethernet/nvidia/forcedeth.c 			np->msi_x_entry[i].entry = i;
entry             996 drivers/net/ethernet/packetengines/hamachi.c 		int entry = hmp->dirty_tx % TX_RING_SIZE;
entry             999 drivers/net/ethernet/packetengines/hamachi.c 		if (hmp->tx_ring[entry].status_n_length & cpu_to_le32(DescOwn))
entry            1002 drivers/net/ethernet/packetengines/hamachi.c 		skb = hmp->tx_skbuff[entry];
entry            1005 drivers/net/ethernet/packetengines/hamachi.c 				leXX_to_cpu(hmp->tx_ring[entry].addr),
entry            1008 drivers/net/ethernet/packetengines/hamachi.c 			hmp->tx_skbuff[entry] = NULL;
entry            1010 drivers/net/ethernet/packetengines/hamachi.c 		hmp->tx_ring[entry].status_n_length = 0;
entry            1011 drivers/net/ethernet/packetengines/hamachi.c 		if (entry >= TX_RING_SIZE-1)
entry            1208 drivers/net/ethernet/packetengines/hamachi.c 	unsigned entry;
entry            1232 drivers/net/ethernet/packetengines/hamachi.c 	entry = hmp->cur_tx % TX_RING_SIZE;
entry            1234 drivers/net/ethernet/packetengines/hamachi.c 	hmp->tx_skbuff[entry] = skb;
entry            1236 drivers/net/ethernet/packetengines/hamachi.c         hmp->tx_ring[entry].addr = cpu_to_leXX(pci_map_single(hmp->pci_dev,
entry            1248 drivers/net/ethernet/packetengines/hamachi.c 	if (entry >= TX_RING_SIZE-1)		 /* Wrap ring */
entry            1249 drivers/net/ethernet/packetengines/hamachi.c 		hmp->tx_ring[entry].status_n_length = cpu_to_le32(DescOwn |
entry            1252 drivers/net/ethernet/packetengines/hamachi.c 		hmp->tx_ring[entry].status_n_length = cpu_to_le32(DescOwn |
entry            1282 drivers/net/ethernet/packetengines/hamachi.c 			   dev->name, hmp->cur_tx, entry);
entry            1328 drivers/net/ethernet/packetengines/hamachi.c 					int entry = hmp->dirty_tx % TX_RING_SIZE;
entry            1331 drivers/net/ethernet/packetengines/hamachi.c 					if (hmp->tx_ring[entry].status_n_length & cpu_to_le32(DescOwn))
entry            1333 drivers/net/ethernet/packetengines/hamachi.c 					skb = hmp->tx_skbuff[entry];
entry            1337 drivers/net/ethernet/packetengines/hamachi.c 							leXX_to_cpu(hmp->tx_ring[entry].addr),
entry            1341 drivers/net/ethernet/packetengines/hamachi.c 						hmp->tx_skbuff[entry] = NULL;
entry            1343 drivers/net/ethernet/packetengines/hamachi.c 					hmp->tx_ring[entry].status_n_length = 0;
entry            1344 drivers/net/ethernet/packetengines/hamachi.c 					if (entry >= TX_RING_SIZE-1)
entry            1398 drivers/net/ethernet/packetengines/hamachi.c 	int entry = hmp->cur_rx % RX_RING_SIZE;
entry            1403 drivers/net/ethernet/packetengines/hamachi.c 			   entry, hmp->rx_ring[entry].status_n_length);
entry            1408 drivers/net/ethernet/packetengines/hamachi.c 		struct hamachi_desc *desc = &(hmp->rx_ring[entry]);
entry            1420 drivers/net/ethernet/packetengines/hamachi.c 		buf_addr = (u8 *) hmp->rx_skbuff[entry]->data;
entry            1487 drivers/net/ethernet/packetengines/hamachi.c 							    leXX_to_cpu(hmp->rx_ring[entry].addr),
entry            1493 drivers/net/ethernet/packetengines/hamachi.c 					hmp->rx_skbuff[entry]->data, pkt_len);
entry            1497 drivers/net/ethernet/packetengines/hamachi.c 					     + entry*sizeof(*desc), pkt_len);
entry            1500 drivers/net/ethernet/packetengines/hamachi.c 							       leXX_to_cpu(hmp->rx_ring[entry].addr),
entry            1505 drivers/net/ethernet/packetengines/hamachi.c 						 leXX_to_cpu(hmp->rx_ring[entry].addr),
entry            1507 drivers/net/ethernet/packetengines/hamachi.c 				skb_put(skb = hmp->rx_skbuff[entry], pkt_len);
entry            1508 drivers/net/ethernet/packetengines/hamachi.c 				hmp->rx_skbuff[entry] = NULL;
entry            1573 drivers/net/ethernet/packetengines/hamachi.c 		entry = (++hmp->cur_rx) % RX_RING_SIZE;
entry            1580 drivers/net/ethernet/packetengines/hamachi.c 		entry = hmp->dirty_rx % RX_RING_SIZE;
entry            1581 drivers/net/ethernet/packetengines/hamachi.c 		desc = &(hmp->rx_ring[entry]);
entry            1582 drivers/net/ethernet/packetengines/hamachi.c 		if (hmp->rx_skbuff[entry] == NULL) {
entry            1585 drivers/net/ethernet/packetengines/hamachi.c 			hmp->rx_skbuff[entry] = skb;
entry            1593 drivers/net/ethernet/packetengines/hamachi.c 		if (entry >= RX_RING_SIZE-1)
entry             808 drivers/net/ethernet/packetengines/yellowfin.c 	unsigned entry;
entry             817 drivers/net/ethernet/packetengines/yellowfin.c 	entry = yp->cur_tx % TX_RING_SIZE;
entry             825 drivers/net/ethernet/packetengines/yellowfin.c 				yp->tx_skbuff[entry] = NULL;
entry             831 drivers/net/ethernet/packetengines/yellowfin.c 	yp->tx_skbuff[entry] = skb;
entry             834 drivers/net/ethernet/packetengines/yellowfin.c 	yp->tx_ring[entry].addr = cpu_to_le32(pci_map_single(yp->pci_dev,
entry             836 drivers/net/ethernet/packetengines/yellowfin.c 	yp->tx_ring[entry].result_status = 0;
entry             837 drivers/net/ethernet/packetengines/yellowfin.c 	if (entry >= TX_RING_SIZE-1) {
entry             843 drivers/net/ethernet/packetengines/yellowfin.c 		yp->tx_ring[entry+1].dbdma_cmd = cpu_to_le32(CMD_STOP);
entry             844 drivers/net/ethernet/packetengines/yellowfin.c 		yp->tx_ring[entry].dbdma_cmd =
entry             849 drivers/net/ethernet/packetengines/yellowfin.c 	yp->tx_ring[entry<<1].request_cnt = len;
entry             850 drivers/net/ethernet/packetengines/yellowfin.c 	yp->tx_ring[entry<<1].addr = cpu_to_le32(pci_map_single(yp->pci_dev,
entry             862 drivers/net/ethernet/packetengines/yellowfin.c 	yp->tx_ring[entry<<1].dbdma_cmd =
entry             863 drivers/net/ethernet/packetengines/yellowfin.c 		cpu_to_le32( ((entry % 6) == 0 ? CMD_TX_PKT|INTR_ALWAYS|BRANCH_IFTRUE :
entry             879 drivers/net/ethernet/packetengines/yellowfin.c 			      yp->cur_tx, entry);
entry             917 drivers/net/ethernet/packetengines/yellowfin.c 			int entry = yp->dirty_tx % TX_RING_SIZE;
entry             920 drivers/net/ethernet/packetengines/yellowfin.c 			if (yp->tx_ring[entry].result_status == 0)
entry             922 drivers/net/ethernet/packetengines/yellowfin.c 			skb = yp->tx_skbuff[entry];
entry             926 drivers/net/ethernet/packetengines/yellowfin.c 			pci_unmap_single(yp->pci_dev, le32_to_cpu(yp->tx_ring[entry].addr),
entry             929 drivers/net/ethernet/packetengines/yellowfin.c 			yp->tx_skbuff[entry] = NULL;
entry             944 drivers/net/ethernet/packetengines/yellowfin.c 				int entry = dirty_tx % TX_RING_SIZE;
entry             945 drivers/net/ethernet/packetengines/yellowfin.c 				u16 tx_errs = yp->tx_status[entry].tx_errs;
entry             951 drivers/net/ethernet/packetengines/yellowfin.c 						      entry,
entry             952 drivers/net/ethernet/packetengines/yellowfin.c 						      yp->tx_status[entry].tx_cnt,
entry             953 drivers/net/ethernet/packetengines/yellowfin.c 						      yp->tx_status[entry].tx_errs,
entry             954 drivers/net/ethernet/packetengines/yellowfin.c 						      yp->tx_status[entry].total_tx_cnt,
entry             955 drivers/net/ethernet/packetengines/yellowfin.c 						      yp->tx_status[entry].paused);
entry             959 drivers/net/ethernet/packetengines/yellowfin.c 				skb = yp->tx_skbuff[entry];
entry             984 drivers/net/ethernet/packetengines/yellowfin.c 					yp->tx_ring[entry<<1].addr, skb->len,
entry             987 drivers/net/ethernet/packetengines/yellowfin.c 				yp->tx_skbuff[entry] = 0;
entry             989 drivers/net/ethernet/packetengines/yellowfin.c 				yp->tx_status[entry].tx_errs = 0;
entry            1036 drivers/net/ethernet/packetengines/yellowfin.c 	int entry = yp->cur_rx % RX_RING_SIZE;
entry            1041 drivers/net/ethernet/packetengines/yellowfin.c 			   entry, yp->rx_ring[entry].result_status);
entry            1043 drivers/net/ethernet/packetengines/yellowfin.c 			   entry, yp->rx_ring[entry].dbdma_cmd, yp->rx_ring[entry].addr,
entry            1044 drivers/net/ethernet/packetengines/yellowfin.c 			   yp->rx_ring[entry].result_status);
entry            1049 drivers/net/ethernet/packetengines/yellowfin.c 		struct yellowfin_desc *desc = &yp->rx_ring[entry];
entry            1050 drivers/net/ethernet/packetengines/yellowfin.c 		struct sk_buff *rx_skb = yp->rx_skbuff[entry];
entry            1100 drivers/net/ethernet/packetengines/yellowfin.c 						      entry * yf_size),
entry            1103 drivers/net/ethernet/packetengines/yellowfin.c 						      entry * yf_size),
entry            1125 drivers/net/ethernet/packetengines/yellowfin.c 					le32_to_cpu(yp->rx_ring[entry].addr),
entry            1128 drivers/net/ethernet/packetengines/yellowfin.c 				yp->rx_skbuff[entry] = NULL;
entry            1146 drivers/net/ethernet/packetengines/yellowfin.c 		entry = (++yp->cur_rx) % RX_RING_SIZE;
entry            1151 drivers/net/ethernet/packetengines/yellowfin.c 		entry = yp->dirty_rx % RX_RING_SIZE;
entry            1152 drivers/net/ethernet/packetengines/yellowfin.c 		if (yp->rx_skbuff[entry] == NULL) {
entry            1156 drivers/net/ethernet/packetengines/yellowfin.c 			yp->rx_skbuff[entry] = skb;
entry            1158 drivers/net/ethernet/packetengines/yellowfin.c 			yp->rx_ring[entry].addr = cpu_to_le32(pci_map_single(yp->pci_dev,
entry            1161 drivers/net/ethernet/packetengines/yellowfin.c 		yp->rx_ring[entry].dbdma_cmd = cpu_to_le32(CMD_STOP);
entry            1162 drivers/net/ethernet/packetengines/yellowfin.c 		yp->rx_ring[entry].result_status = 0;	/* Clear complete bit. */
entry            1163 drivers/net/ethernet/packetengines/yellowfin.c 		if (entry != 0)
entry            1164 drivers/net/ethernet/packetengines/yellowfin.c 			yp->rx_ring[entry - 1].dbdma_cmd =
entry            2312 drivers/net/ethernet/qlogic/netxen/netxen_nic_hw.c 				struct netxen_minidump_entry *entry, int esize)
entry            2315 drivers/net/ethernet/qlogic/netxen/netxen_nic_hw.c 		entry->hdr.driver_flags |= NX_DUMP_SKIP;
entry            2318 drivers/net/ethernet/qlogic/netxen/netxen_nic_hw.c 	if (esize != entry->hdr.entry_capture_size) {
entry            2319 drivers/net/ethernet/qlogic/netxen/netxen_nic_hw.c 		entry->hdr.entry_capture_size = esize;
entry            2320 drivers/net/ethernet/qlogic/netxen/netxen_nic_hw.c 		entry->hdr.driver_flags |= NX_DUMP_SIZE_ERR;
entry            2323 drivers/net/ethernet/qlogic/netxen/netxen_nic_hw.c 			entry->hdr.entry_type, entry->hdr.entry_capture_mask,
entry            2324 drivers/net/ethernet/qlogic/netxen/netxen_nic_hw.c 			esize, entry->hdr.entry_capture_size);
entry            2339 drivers/net/ethernet/qlogic/netxen/netxen_nic_hw.c 	struct netxen_minidump_entry *entry;
entry            2348 drivers/net/ethernet/qlogic/netxen/netxen_nic_hw.c 	entry = (struct netxen_minidump_entry *) ((char *) template_buff +
entry            2357 drivers/net/ethernet/qlogic/netxen/netxen_nic_hw.c 		if (!(entry->hdr.entry_capture_mask & capture_mask)) {
entry            2358 drivers/net/ethernet/qlogic/netxen/netxen_nic_hw.c 			entry->hdr.driver_flags |= NX_DUMP_SKIP;
entry            2359 drivers/net/ethernet/qlogic/netxen/netxen_nic_hw.c 			entry = (struct netxen_minidump_entry *)
entry            2360 drivers/net/ethernet/qlogic/netxen/netxen_nic_hw.c 				((char *) entry + entry->hdr.entry_size);
entry            2363 drivers/net/ethernet/qlogic/netxen/netxen_nic_hw.c 		switch (entry->hdr.entry_type) {
entry            2365 drivers/net/ethernet/qlogic/netxen/netxen_nic_hw.c 			entry->hdr.driver_flags |= NX_DUMP_SKIP;
entry            2368 drivers/net/ethernet/qlogic/netxen/netxen_nic_hw.c 			entry->hdr.driver_flags |= NX_DUMP_SKIP;
entry            2373 drivers/net/ethernet/qlogic/netxen/netxen_nic_hw.c 				template_hdr, (void *)entry);
entry            2375 drivers/net/ethernet/qlogic/netxen/netxen_nic_hw.c 				entry->hdr.driver_flags |= NX_DUMP_SKIP;
entry            2380 drivers/net/ethernet/qlogic/netxen/netxen_nic_hw.c 					(void *) entry, (void *) dbuff);
entry            2382 drivers/net/ethernet/qlogic/netxen/netxen_nic_hw.c 				(adapter, entry, esize);
entry            2391 drivers/net/ethernet/qlogic/netxen/netxen_nic_hw.c 				(void *) entry, (void *) dbuff);
entry            2393 drivers/net/ethernet/qlogic/netxen/netxen_nic_hw.c 				(adapter, entry, esize);
entry            2402 drivers/net/ethernet/qlogic/netxen/netxen_nic_hw.c 				(void *) entry, (void *) dbuff);
entry            2404 drivers/net/ethernet/qlogic/netxen/netxen_nic_hw.c 				(adapter, entry, esize);
entry            2415 drivers/net/ethernet/qlogic/netxen/netxen_nic_hw.c 				(void *) entry, (void *) dbuff);
entry            2417 drivers/net/ethernet/qlogic/netxen/netxen_nic_hw.c 				(adapter, entry, esize);
entry            2426 drivers/net/ethernet/qlogic/netxen/netxen_nic_hw.c 				(void *) entry, (void *) dbuff);
entry            2428 drivers/net/ethernet/qlogic/netxen/netxen_nic_hw.c 				(adapter, entry, esize);
entry            2436 drivers/net/ethernet/qlogic/netxen/netxen_nic_hw.c 				(void *) entry, (void *) dbuff);
entry            2438 drivers/net/ethernet/qlogic/netxen/netxen_nic_hw.c 				(adapter, entry, esize);
entry            2446 drivers/net/ethernet/qlogic/netxen/netxen_nic_hw.c 				(void *) entry, (void *) dbuff);
entry            2448 drivers/net/ethernet/qlogic/netxen/netxen_nic_hw.c 				(adapter, entry, esize);
entry            2456 drivers/net/ethernet/qlogic/netxen/netxen_nic_hw.c 				(void *) entry, (void *) dbuff);
entry            2458 drivers/net/ethernet/qlogic/netxen/netxen_nic_hw.c 				(adapter, entry, esize);
entry            2464 drivers/net/ethernet/qlogic/netxen/netxen_nic_hw.c 			entry->hdr.driver_flags |= NX_DUMP_SKIP;
entry            2468 drivers/net/ethernet/qlogic/netxen/netxen_nic_hw.c 		entry = (struct netxen_minidump_entry *)
entry            2469 drivers/net/ethernet/qlogic/netxen/netxen_nic_hw.c 			((char *) entry + entry->hdr.entry_size);
entry             455 drivers/net/ethernet/qlogic/netxen/netxen_nic_main.c 		adapter->msix_entries[i].entry = i;
entry             294 drivers/net/ethernet/qlogic/qed/qed_dcbx.c 		protocol_id = QED_MFW_GET_FIELD(p_tbl[i].entry,
entry             296 drivers/net/ethernet/qlogic/qed/qed_dcbx.c 		priority_map = QED_MFW_GET_FIELD(p_tbl[i].entry,
entry             305 drivers/net/ethernet/qlogic/qed/qed_dcbx.c 		if (qed_dcbx_get_app_protocol_type(p_hwfn, p_tbl[i].entry,
entry             486 drivers/net/ethernet/qlogic/qed/qed_dcbx.c 	struct qed_app_entry *entry;
entry             497 drivers/net/ethernet/qlogic/qed/qed_dcbx.c 		entry = &p_params->app_entry[i];
entry             502 drivers/net/ethernet/qlogic/qed/qed_dcbx.c 			sf_ieee = QED_MFW_GET_FIELD(p_tbl[i].entry,
entry             507 drivers/net/ethernet/qlogic/qed/qed_dcbx.c 				val = QED_MFW_GET_FIELD(p_tbl[i].entry,
entry             509 drivers/net/ethernet/qlogic/qed/qed_dcbx.c 				entry->sf_ieee = val ?
entry             514 drivers/net/ethernet/qlogic/qed/qed_dcbx.c 				entry->sf_ieee = QED_DCBX_SF_IEEE_ETHTYPE;
entry             517 drivers/net/ethernet/qlogic/qed/qed_dcbx.c 				entry->sf_ieee = QED_DCBX_SF_IEEE_TCP_PORT;
entry             520 drivers/net/ethernet/qlogic/qed/qed_dcbx.c 				entry->sf_ieee = QED_DCBX_SF_IEEE_UDP_PORT;
entry             523 drivers/net/ethernet/qlogic/qed/qed_dcbx.c 				entry->sf_ieee = QED_DCBX_SF_IEEE_TCP_UDP_PORT;
entry             527 drivers/net/ethernet/qlogic/qed/qed_dcbx.c 			entry->ethtype = !(QED_MFW_GET_FIELD(p_tbl[i].entry,
entry             531 drivers/net/ethernet/qlogic/qed/qed_dcbx.c 		pri_map = QED_MFW_GET_FIELD(p_tbl[i].entry, DCBX_APP_PRI_MAP);
entry             532 drivers/net/ethernet/qlogic/qed/qed_dcbx.c 		entry->prio = ffs(pri_map) - 1;
entry             533 drivers/net/ethernet/qlogic/qed/qed_dcbx.c 		entry->proto_id = QED_MFW_GET_FIELD(p_tbl[i].entry,
entry             535 drivers/net/ethernet/qlogic/qed/qed_dcbx.c 		qed_dcbx_get_app_protocol_type(p_hwfn, p_tbl[i].entry,
entry             536 drivers/net/ethernet/qlogic/qed/qed_dcbx.c 					       entry->proto_id,
entry             537 drivers/net/ethernet/qlogic/qed/qed_dcbx.c 					       &entry->proto_type, ieee);
entry            1128 drivers/net/ethernet/qlogic/qed/qed_dcbx.c 	u32 *entry;
entry            1146 drivers/net/ethernet/qlogic/qed/qed_dcbx.c 		entry = &p_app->app_pri_tbl[i].entry;
entry            1147 drivers/net/ethernet/qlogic/qed/qed_dcbx.c 		*entry = 0;
entry            1149 drivers/net/ethernet/qlogic/qed/qed_dcbx.c 			*entry &= ~(DCBX_APP_SF_IEEE_MASK | DCBX_APP_SF_MASK);
entry            1152 drivers/net/ethernet/qlogic/qed/qed_dcbx.c 				*entry |= ((u32)DCBX_APP_SF_IEEE_ETHTYPE <<
entry            1154 drivers/net/ethernet/qlogic/qed/qed_dcbx.c 				*entry |= ((u32)DCBX_APP_SF_ETHTYPE <<
entry            1158 drivers/net/ethernet/qlogic/qed/qed_dcbx.c 				*entry |= ((u32)DCBX_APP_SF_IEEE_TCP_PORT <<
entry            1160 drivers/net/ethernet/qlogic/qed/qed_dcbx.c 				*entry |= ((u32)DCBX_APP_SF_PORT <<
entry            1164 drivers/net/ethernet/qlogic/qed/qed_dcbx.c 				*entry |= ((u32)DCBX_APP_SF_IEEE_UDP_PORT <<
entry            1166 drivers/net/ethernet/qlogic/qed/qed_dcbx.c 				*entry |= ((u32)DCBX_APP_SF_PORT <<
entry            1170 drivers/net/ethernet/qlogic/qed/qed_dcbx.c 				*entry |= ((u32)DCBX_APP_SF_IEEE_TCP_UDP_PORT <<
entry            1172 drivers/net/ethernet/qlogic/qed/qed_dcbx.c 				*entry |= ((u32)DCBX_APP_SF_PORT <<
entry            1177 drivers/net/ethernet/qlogic/qed/qed_dcbx.c 			*entry &= ~DCBX_APP_SF_MASK;
entry            1179 drivers/net/ethernet/qlogic/qed/qed_dcbx.c 				*entry |= ((u32)DCBX_APP_SF_ETHTYPE <<
entry            1182 drivers/net/ethernet/qlogic/qed/qed_dcbx.c 				*entry |= ((u32)DCBX_APP_SF_PORT <<
entry            1186 drivers/net/ethernet/qlogic/qed/qed_dcbx.c 		*entry &= ~DCBX_APP_PROTOCOL_ID_MASK;
entry            1187 drivers/net/ethernet/qlogic/qed/qed_dcbx.c 		*entry |= ((u32)p_params->app_entry[i].proto_id <<
entry            1189 drivers/net/ethernet/qlogic/qed/qed_dcbx.c 		*entry &= ~DCBX_APP_PRI_MAP_MASK;
entry            1190 drivers/net/ethernet/qlogic/qed/qed_dcbx.c 		*entry |= ((u32)(p_params->app_entry[i].prio) <<
entry            1756 drivers/net/ethernet/qlogic/qed/qed_dcbx.c 	struct qed_app_entry *entry;
entry            1767 drivers/net/ethernet/qlogic/qed/qed_dcbx.c 		entry = &dcbx_info->operational.params.app_entry[i];
entry            1768 drivers/net/ethernet/qlogic/qed/qed_dcbx.c 		if ((entry->ethtype == ethtype) && (entry->proto_id == idval)) {
entry            1769 drivers/net/ethernet/qlogic/qed/qed_dcbx.c 			prio = entry->prio;
entry            1790 drivers/net/ethernet/qlogic/qed/qed_dcbx.c 	struct qed_app_entry *entry;
entry            1802 drivers/net/ethernet/qlogic/qed/qed_dcbx.c 		entry = &dcbx_set.config.params.app_entry[i];
entry            1803 drivers/net/ethernet/qlogic/qed/qed_dcbx.c 		if ((entry->ethtype == ethtype) && (entry->proto_id == idval))
entry            1806 drivers/net/ethernet/qlogic/qed/qed_dcbx.c 		if (!entry->proto_id) {
entry            2282 drivers/net/ethernet/qlogic/qed/qed_dcbx.c 	struct qed_app_entry *entry;
entry            2307 drivers/net/ethernet/qlogic/qed/qed_dcbx.c 		entry = &dcbx_info->operational.params.app_entry[i];
entry            2308 drivers/net/ethernet/qlogic/qed/qed_dcbx.c 		if ((entry->sf_ieee == sf_ieee) &&
entry            2309 drivers/net/ethernet/qlogic/qed/qed_dcbx.c 		    (entry->proto_id == app->protocol)) {
entry            2310 drivers/net/ethernet/qlogic/qed/qed_dcbx.c 			prio = entry->prio;
entry            2334 drivers/net/ethernet/qlogic/qed/qed_dcbx.c 	struct qed_app_entry *entry;
entry            2370 drivers/net/ethernet/qlogic/qed/qed_dcbx.c 		entry = &dcbx_set.config.params.app_entry[i];
entry            2371 drivers/net/ethernet/qlogic/qed/qed_dcbx.c 		if ((entry->sf_ieee == sf_ieee) &&
entry            2372 drivers/net/ethernet/qlogic/qed/qed_dcbx.c 		    (entry->proto_id == app->protocol))
entry            2375 drivers/net/ethernet/qlogic/qed/qed_dcbx.c 		if (!entry->proto_id) {
entry            2530 drivers/net/ethernet/qlogic/qed/qed_dev.c 			p_block = &p_igu_info->entry[igu_sb_id];
entry            1005 drivers/net/ethernet/qlogic/qed/qed_hsi.h 	struct event_ring_entry entry;
entry            11999 drivers/net/ethernet/qlogic/qed/qed_hsi.h 	u32 entry;
entry            1580 drivers/net/ethernet/qlogic/qed/qed_int.c 		p_block = &p_hwfn->hw_info.p_igu_info->entry[igu_id];
entry            1600 drivers/net/ethernet/qlogic/qed/qed_int.c 		p_block = &p_hwfn->hw_info.p_igu_info->entry[igu_id];
entry            1651 drivers/net/ethernet/qlogic/qed/qed_int.c 			p_block = &p_info->entry[sb_info->igu_sb_id];
entry            1703 drivers/net/ethernet/qlogic/qed/qed_int.c 	p_block = &p_info->entry[sb_info->igu_sb_id];
entry            1936 drivers/net/ethernet/qlogic/qed/qed_int.c 	p_block = &p_hwfn->hw_info.p_igu_info->entry[igu_sb_id];
entry            1989 drivers/net/ethernet/qlogic/qed/qed_int.c 		p_block = &p_info->entry[igu_sb_id];
entry            2076 drivers/net/ethernet/qlogic/qed/qed_int.c 		p_block = &p_info->entry[igu_sb_id];
entry            2149 drivers/net/ethernet/qlogic/qed/qed_int.c 	p_block = &p_hwfn->hw_info.p_igu_info->entry[igu_sb_id];
entry            2186 drivers/net/ethernet/qlogic/qed/qed_int.c 		p_block = &p_igu_info->entry[igu_sb_id];
entry             232 drivers/net/ethernet/qlogic/qed/qed_int.h 	struct qed_igu_block entry[MAX_TOT_SB_PER_PATH];
entry             538 drivers/net/ethernet/qlogic/qed/qed_main.c 		int_params->msix_table[i].entry = i;
entry             949 drivers/net/ethernet/qlogic/qed/qed_sriov.c 		p_info->entry[igu_id].status |= QED_IGU_STATUS_FREE;
entry             163 drivers/net/ethernet/qlogic/qede/qede.h 	struct list_head entry;
entry             125 drivers/net/ethernet/qlogic/qede/qede_rdma.c 	INIT_LIST_HEAD(&edev->rdma_info.entry);
entry             127 drivers/net/ethernet/qlogic/qede/qede_rdma.c 	list_add_tail(&edev->rdma_info.entry, &qedr_dev_list);
entry             152 drivers/net/ethernet/qlogic/qede/qede_rdma.c 		list_del(&edev->rdma_info.entry);
entry             219 drivers/net/ethernet/qlogic/qede/qede_rdma.c 	list_for_each_entry(edev, &qedr_dev_list, rdma_info.entry) {
entry             242 drivers/net/ethernet/qlogic/qede/qede_rdma.c 	list_for_each_entry(edev, &qedr_dev_list, rdma_info.entry) {
entry            1798 drivers/net/ethernet/qlogic/qlcnic/qlcnic_83xx_init.c 	struct qlc_83xx_entry *entry;
entry            1800 drivers/net/ethernet/qlogic/qlcnic/qlcnic_83xx_init.c 	entry = (struct qlc_83xx_entry *)((char *)p_hdr +
entry            1803 drivers/net/ethernet/qlogic/qlcnic/qlcnic_83xx_init.c 	for (i = 0; i < p_hdr->count; i++, entry++) {
entry            1804 drivers/net/ethernet/qlogic/qlcnic/qlcnic_83xx_init.c 		qlcnic_83xx_wrt_reg_indirect(p_dev, entry->arg1,
entry            1805 drivers/net/ethernet/qlogic/qlcnic/qlcnic_83xx_init.c 					     entry->arg2);
entry            1816 drivers/net/ethernet/qlogic/qlcnic/qlcnic_83xx_init.c 	struct qlc_83xx_entry *entry;
entry            1818 drivers/net/ethernet/qlogic/qlcnic/qlcnic_83xx_init.c 	entry = (struct qlc_83xx_entry *)((char *)p_hdr +
entry            1821 drivers/net/ethernet/qlogic/qlcnic/qlcnic_83xx_init.c 	for (i = 0; i < p_hdr->count; i++, entry++) {
entry            1822 drivers/net/ethernet/qlogic/qlcnic/qlcnic_83xx_init.c 		qlcnic_83xx_read_write_crb_reg(p_dev, entry->arg1,
entry            1823 drivers/net/ethernet/qlogic/qlcnic/qlcnic_83xx_init.c 					       entry->arg2);
entry            1834 drivers/net/ethernet/qlogic/qlcnic/qlcnic_83xx_init.c 	struct qlc_83xx_entry *entry;
entry            1842 drivers/net/ethernet/qlogic/qlcnic/qlcnic_83xx_init.c 	entry = (struct qlc_83xx_entry *)((char *)poll +
entry            1847 drivers/net/ethernet/qlogic/qlcnic/qlcnic_83xx_init.c 		for (i = 0; i < p_hdr->count; i++, entry++)
entry            1848 drivers/net/ethernet/qlogic/qlcnic/qlcnic_83xx_init.c 			qlcnic_83xx_poll_reg(p_dev, entry->arg1,
entry            1852 drivers/net/ethernet/qlogic/qlcnic/qlcnic_83xx_init.c 		for (i = 0; i < p_hdr->count; i++, entry++) {
entry            1853 drivers/net/ethernet/qlogic/qlcnic/qlcnic_83xx_init.c 			arg1 = entry->arg1;
entry            1854 drivers/net/ethernet/qlogic/qlcnic/qlcnic_83xx_init.c 			arg2 = entry->arg2;
entry            1878 drivers/net/ethernet/qlogic/qlcnic/qlcnic_83xx_init.c 	struct qlc_83xx_quad_entry *entry;
entry            1883 drivers/net/ethernet/qlogic/qlcnic/qlcnic_83xx_init.c 	entry = (struct qlc_83xx_quad_entry *)((char *)poll +
entry            1887 drivers/net/ethernet/qlogic/qlcnic/qlcnic_83xx_init.c 	for (i = 0; i < p_hdr->count; i++, entry++) {
entry            1888 drivers/net/ethernet/qlogic/qlcnic/qlcnic_83xx_init.c 		qlcnic_83xx_wrt_reg_indirect(p_dev, entry->dr_addr,
entry            1889 drivers/net/ethernet/qlogic/qlcnic/qlcnic_83xx_init.c 					     entry->dr_value);
entry            1890 drivers/net/ethernet/qlogic/qlcnic/qlcnic_83xx_init.c 		qlcnic_83xx_wrt_reg_indirect(p_dev, entry->ar_addr,
entry            1891 drivers/net/ethernet/qlogic/qlcnic/qlcnic_83xx_init.c 					     entry->ar_value);
entry            1893 drivers/net/ethernet/qlogic/qlcnic/qlcnic_83xx_init.c 			qlcnic_83xx_poll_reg(p_dev, entry->ar_addr, delay,
entry            1903 drivers/net/ethernet/qlogic/qlcnic/qlcnic_83xx_init.c 	struct qlc_83xx_entry *entry;
entry            1909 drivers/net/ethernet/qlogic/qlcnic/qlcnic_83xx_init.c 	entry = (struct qlc_83xx_entry *)((char *)rmw_hdr +
entry            1912 drivers/net/ethernet/qlogic/qlcnic/qlcnic_83xx_init.c 	for (i = 0; i < p_hdr->count; i++, entry++) {
entry            1913 drivers/net/ethernet/qlogic/qlcnic/qlcnic_83xx_init.c 		qlcnic_83xx_rmw_crb_reg(p_dev, entry->arg1,
entry            1914 drivers/net/ethernet/qlogic/qlcnic/qlcnic_83xx_init.c 					entry->arg2, rmw_hdr);
entry            1932 drivers/net/ethernet/qlogic/qlcnic/qlcnic_83xx_init.c 	struct qlc_83xx_quad_entry *entry;
entry            1939 drivers/net/ethernet/qlogic/qlcnic/qlcnic_83xx_init.c 	entry = (struct qlc_83xx_quad_entry *)((char *)poll +
entry            1943 drivers/net/ethernet/qlogic/qlcnic/qlcnic_83xx_init.c 	for (i = 0; i < p_hdr->count; i++, entry++) {
entry            1944 drivers/net/ethernet/qlogic/qlcnic/qlcnic_83xx_init.c 		qlcnic_83xx_wrt_reg_indirect(p_dev, entry->ar_addr,
entry            1945 drivers/net/ethernet/qlogic/qlcnic/qlcnic_83xx_init.c 					     entry->ar_value);
entry            1947 drivers/net/ethernet/qlogic/qlcnic/qlcnic_83xx_init.c 			if (!qlcnic_83xx_poll_reg(p_dev, entry->ar_addr, delay,
entry            1950 drivers/net/ethernet/qlogic/qlcnic/qlcnic_83xx_init.c 				addr = entry->dr_addr;
entry            1999 drivers/net/ethernet/qlogic/qlcnic/qlcnic_83xx_init.c 	char *entry = p_buff;
entry            2007 drivers/net/ethernet/qlogic/qlcnic/qlcnic_83xx_init.c 		p_hdr = (struct qlc_83xx_entry_hdr *)entry;
entry            2045 drivers/net/ethernet/qlogic/qlcnic/qlcnic_83xx_init.c 		entry += p_hdr->size;
entry             715 drivers/net/ethernet/qlogic/qlcnic/qlcnic_main.c 		adapter->msix_entries[vector].entry = vector;
entry             774 drivers/net/ethernet/qlogic/qlcnic/qlcnic_main.c 			adapter->msix_entries[vector].entry = vector;
entry             386 drivers/net/ethernet/qlogic/qlcnic/qlcnic_minidump.c 			   struct qlcnic_dump_entry *entry, __le32 *buffer)
entry             390 drivers/net/ethernet/qlogic/qlcnic/qlcnic_minidump.c 	struct __crb *crb = &entry->region.crb;
entry             404 drivers/net/ethernet/qlogic/qlcnic/qlcnic_minidump.c 			    struct qlcnic_dump_entry *entry, __le32 *buffer)
entry             407 drivers/net/ethernet/qlogic/qlcnic/qlcnic_minidump.c 	struct __ctrl *ctr = &entry->region.ctrl;
entry             502 drivers/net/ethernet/qlogic/qlcnic/qlcnic_minidump.c 			   struct qlcnic_dump_entry *entry, __le32 *buffer)
entry             506 drivers/net/ethernet/qlogic/qlcnic/qlcnic_minidump.c 	struct __mux *mux = &entry->region.mux;
entry             520 drivers/net/ethernet/qlogic/qlcnic/qlcnic_minidump.c 			   struct qlcnic_dump_entry *entry, __le32 *buffer)
entry             524 drivers/net/ethernet/qlogic/qlcnic/qlcnic_minidump.c 	struct __queue *que = &entry->region.que;
entry             543 drivers/net/ethernet/qlogic/qlcnic/qlcnic_minidump.c 			   struct qlcnic_dump_entry *entry, __le32 *buffer)
entry             548 drivers/net/ethernet/qlogic/qlcnic/qlcnic_minidump.c 	struct __ocm *ocm = &entry->region.ocm;
entry             560 drivers/net/ethernet/qlogic/qlcnic/qlcnic_minidump.c 			   struct qlcnic_dump_entry *entry, __le32 *buffer)
entry             564 drivers/net/ethernet/qlogic/qlcnic/qlcnic_minidump.c 	struct __mem *rom = &entry->region.mem;
entry             590 drivers/net/ethernet/qlogic/qlcnic/qlcnic_minidump.c 				struct qlcnic_dump_entry *entry, __le32 *buffer)
entry             594 drivers/net/ethernet/qlogic/qlcnic/qlcnic_minidump.c 	struct __cache *l1 = &entry->region.cache;
entry             615 drivers/net/ethernet/qlogic/qlcnic/qlcnic_minidump.c 				struct qlcnic_dump_entry *entry, __le32 *buffer)
entry             620 drivers/net/ethernet/qlogic/qlcnic/qlcnic_minidump.c 	struct __cache *l2 = &entry->region.cache;
entry             856 drivers/net/ethernet/qlogic/qlcnic/qlcnic_minidump.c 			      struct qlcnic_dump_entry *entry, __le32 *buffer)
entry             860 drivers/net/ethernet/qlogic/qlcnic/qlcnic_minidump.c 	struct __mem *mem = &entry->region.mem;
entry             870 drivers/net/ethernet/qlogic/qlcnic/qlcnic_minidump.c 				 entry->hdr.mask);
entry             879 drivers/net/ethernet/qlogic/qlcnic/qlcnic_minidump.c 			 entry->hdr.mask);
entry             887 drivers/net/ethernet/qlogic/qlcnic/qlcnic_minidump.c 			   struct qlcnic_dump_entry *entry, __le32 *buffer)
entry             889 drivers/net/ethernet/qlogic/qlcnic/qlcnic_minidump.c 	entry->hdr.flags |= QLCNIC_DUMP_SKIP;
entry             894 drivers/net/ethernet/qlogic/qlcnic/qlcnic_minidump.c 				   struct qlcnic_dump_entry *entry, u32 size)
entry             897 drivers/net/ethernet/qlogic/qlcnic/qlcnic_minidump.c 	if (size != entry->hdr.cap_size) {
entry             900 drivers/net/ethernet/qlogic/qlcnic/qlcnic_minidump.c 			entry->hdr.type, entry->hdr.mask, size,
entry             901 drivers/net/ethernet/qlogic/qlcnic/qlcnic_minidump.c 			entry->hdr.cap_size);
entry             908 drivers/net/ethernet/qlogic/qlcnic/qlcnic_minidump.c 				 struct qlcnic_dump_entry *entry,
entry             911 drivers/net/ethernet/qlogic/qlcnic/qlcnic_minidump.c 	struct __pollrdmwr *poll = &entry->region.pollrdmwr;
entry             953 drivers/net/ethernet/qlogic/qlcnic/qlcnic_minidump.c 			      struct qlcnic_dump_entry *entry, __le32 *buffer)
entry             955 drivers/net/ethernet/qlogic/qlcnic/qlcnic_minidump.c 	struct __pollrd *pollrd = &entry->region.pollrd;
entry             988 drivers/net/ethernet/qlogic/qlcnic/qlcnic_minidump.c 			    struct qlcnic_dump_entry *entry, __le32 *buffer)
entry             990 drivers/net/ethernet/qlogic/qlcnic/qlcnic_minidump.c 	struct __mux2 *mux2 = &entry->region.mux2;
entry            1019 drivers/net/ethernet/qlogic/qlcnic/qlcnic_minidump.c 				struct qlcnic_dump_entry *entry, __le32 *buffer)
entry            1022 drivers/net/ethernet/qlogic/qlcnic/qlcnic_minidump.c 	struct __mem *rom = &entry->region.mem;
entry            1295 drivers/net/ethernet/qlogic/qlcnic/qlcnic_minidump.c 	struct qlcnic_dump_entry *entry;
entry            1353 drivers/net/ethernet/qlogic/qlcnic/qlcnic_minidump.c 		entry = tmpl_hdr + entry_offset;
entry            1354 drivers/net/ethernet/qlogic/qlcnic/qlcnic_minidump.c 		if (!(entry->hdr.mask & fw_dump->cap_mask)) {
entry            1355 drivers/net/ethernet/qlogic/qlcnic/qlcnic_minidump.c 			entry->hdr.flags |= QLCNIC_DUMP_SKIP;
entry            1356 drivers/net/ethernet/qlogic/qlcnic/qlcnic_minidump.c 			entry_offset += entry->hdr.offset;
entry            1363 drivers/net/ethernet/qlogic/qlcnic/qlcnic_minidump.c 			if (entry->hdr.type == fw_dump_ops[ops_index].opcode)
entry            1370 drivers/net/ethernet/qlogic/qlcnic/qlcnic_minidump.c 				 entry->hdr.type);
entry            1371 drivers/net/ethernet/qlogic/qlcnic/qlcnic_minidump.c 			entry->hdr.flags |= QLCNIC_DUMP_SKIP;
entry            1372 drivers/net/ethernet/qlogic/qlcnic/qlcnic_minidump.c 			entry_offset += entry->hdr.offset;
entry            1377 drivers/net/ethernet/qlogic/qlcnic/qlcnic_minidump.c 		dump = fw_dump_ops[ops_index].handler(adapter, entry, buffer);
entry            1378 drivers/net/ethernet/qlogic/qlcnic/qlcnic_minidump.c 		if (!qlcnic_valid_dump_entry(dev, entry, dump)) {
entry            1379 drivers/net/ethernet/qlogic/qlcnic/qlcnic_minidump.c 			entry->hdr.flags |= QLCNIC_DUMP_SKIP;
entry            1380 drivers/net/ethernet/qlogic/qlcnic/qlcnic_minidump.c 			entry_offset += entry->hdr.offset;
entry            1384 drivers/net/ethernet/qlogic/qlcnic/qlcnic_minidump.c 		buf_offset += entry->hdr.cap_size;
entry            1385 drivers/net/ethernet/qlogic/qlcnic/qlcnic_minidump.c 		entry_offset += entry->hdr.offset;
entry            1525 drivers/net/ethernet/qlogic/qlcnic/qlcnic_sriov_common.c 	struct qlcnic_async_cmd *entry;
entry            1532 drivers/net/ethernet/qlogic/qlcnic/qlcnic_sriov_common.c 		entry = list_entry(head->next, struct qlcnic_async_cmd,
entry            1534 drivers/net/ethernet/qlogic/qlcnic/qlcnic_sriov_common.c 		list_del(&entry->list);
entry            1535 drivers/net/ethernet/qlogic/qlcnic/qlcnic_sriov_common.c 		kfree(entry->cmd);
entry            1536 drivers/net/ethernet/qlogic/qlcnic/qlcnic_sriov_common.c 		kfree(entry);
entry            1598 drivers/net/ethernet/qlogic/qlcnic/qlcnic_sriov_common.c 	struct qlcnic_async_cmd *entry, *tmp;
entry            1611 drivers/net/ethernet/qlogic/qlcnic/qlcnic_sriov_common.c 	list_for_each_entry_safe(entry, tmp, &del_list, list) {
entry            1612 drivers/net/ethernet/qlogic/qlcnic/qlcnic_sriov_common.c 		list_del(&entry->list);
entry            1613 drivers/net/ethernet/qlogic/qlcnic/qlcnic_sriov_common.c 		cmd = entry->cmd;
entry            1615 drivers/net/ethernet/qlogic/qlcnic/qlcnic_sriov_common.c 		kfree(entry);
entry            1628 drivers/net/ethernet/qlogic/qlcnic/qlcnic_sriov_common.c 	struct qlcnic_async_cmd *entry = NULL;
entry            1630 drivers/net/ethernet/qlogic/qlcnic/qlcnic_sriov_common.c 	entry = kzalloc(sizeof(*entry), GFP_ATOMIC);
entry            1631 drivers/net/ethernet/qlogic/qlcnic/qlcnic_sriov_common.c 	if (!entry)
entry            1634 drivers/net/ethernet/qlogic/qlcnic/qlcnic_sriov_common.c 	entry->cmd = cmd;
entry            1637 drivers/net/ethernet/qlogic/qlcnic/qlcnic_sriov_common.c 	list_add_tail(&entry->list, &bc->async_cmd_list);
entry            1640 drivers/net/ethernet/qlogic/qlcnic/qlcnic_sriov_common.c 	return entry;
entry            1646 drivers/net/ethernet/qlogic/qlcnic/qlcnic_sriov_common.c 	struct qlcnic_async_cmd *entry = NULL;
entry            1648 drivers/net/ethernet/qlogic/qlcnic/qlcnic_sriov_common.c 	entry = qlcnic_sriov_alloc_async_cmd(bc, cmd);
entry            1649 drivers/net/ethernet/qlogic/qlcnic/qlcnic_sriov_common.c 	if (!entry) {
entry              53 drivers/net/ethernet/qualcomm/rmnet/rmnet_config.c 	int rc, entry;
entry              71 drivers/net/ethernet/qualcomm/rmnet/rmnet_config.c 	for (entry = 0; entry < RMNET_MAX_LOGICAL_EP; entry++)
entry              72 drivers/net/ethernet/qualcomm/rmnet/rmnet_config.c 		INIT_HLIST_HEAD(&port->muxed_ep[entry]);
entry             735 drivers/net/ethernet/realtek/8139cp.c 	unsigned entry;
entry             751 drivers/net/ethernet/realtek/8139cp.c 	entry = cp->tx_head;
entry             752 drivers/net/ethernet/realtek/8139cp.c 	eor = (entry == (CP_TX_RING_SIZE - 1)) ? RingEnd : 0;
entry             779 drivers/net/ethernet/realtek/8139cp.c 		struct cp_desc *txd = &cp->tx_ring[entry];
entry             797 drivers/net/ethernet/realtek/8139cp.c 		cp->tx_skb[entry] = skb;
entry             798 drivers/net/ethernet/realtek/8139cp.c 		cp->tx_opts[entry] = opts1;
entry             800 drivers/net/ethernet/realtek/8139cp.c 			  entry, skb->len);
entry             805 drivers/net/ethernet/realtek/8139cp.c 		int frag, first_entry = entry;
entry             817 drivers/net/ethernet/realtek/8139cp.c 		cp->tx_skb[entry] = skb;
entry             824 drivers/net/ethernet/realtek/8139cp.c 			entry = NEXT_TX(entry);
entry             831 drivers/net/ethernet/realtek/8139cp.c 				unwind_tx_frag_mapping(cp, skb, first_entry, entry);
entry             835 drivers/net/ethernet/realtek/8139cp.c 			eor = (entry == (CP_TX_RING_SIZE - 1)) ? RingEnd : 0;
entry             842 drivers/net/ethernet/realtek/8139cp.c 			txd = &cp->tx_ring[entry];
entry             850 drivers/net/ethernet/realtek/8139cp.c 			cp->tx_opts[entry] = ctrl;
entry             851 drivers/net/ethernet/realtek/8139cp.c 			cp->tx_skb[entry] = skb;
entry             865 drivers/net/ethernet/realtek/8139cp.c 			  first_entry, entry, skb->len);
entry             867 drivers/net/ethernet/realtek/8139cp.c 	cp->tx_head = NEXT_TX(entry);
entry            1719 drivers/net/ethernet/realtek/8139too.c 	unsigned int entry;
entry            1724 drivers/net/ethernet/realtek/8139too.c 	entry = tp->cur_tx % NUM_TX_DESC;
entry            1729 drivers/net/ethernet/realtek/8139too.c 			memset(tp->tx_buf[entry], 0, ETH_ZLEN);
entry            1730 drivers/net/ethernet/realtek/8139too.c 		skb_copy_and_csum_dev(skb, tp->tx_buf[entry]);
entry            1745 drivers/net/ethernet/realtek/8139too.c 	RTL_W32_F (TxStatus0 + (entry * sizeof (u32)),
entry            1755 drivers/net/ethernet/realtek/8139too.c 		  len, entry);
entry            1773 drivers/net/ethernet/realtek/8139too.c 		int entry = dirty_tx % NUM_TX_DESC;
entry            1776 drivers/net/ethernet/realtek/8139too.c 		txstatus = RTL_R32 (TxStatus0 + (entry * sizeof (u32)));
entry            5633 drivers/net/ethernet/realtek/r8169_main.c 		unsigned int entry = (start + i) % NUM_TX_DESC;
entry            5634 drivers/net/ethernet/realtek/r8169_main.c 		struct ring_info *tx_skb = tp->tx_skb + entry;
entry            5641 drivers/net/ethernet/realtek/r8169_main.c 					     tp->TxDescArray + entry);
entry            5686 drivers/net/ethernet/realtek/r8169_main.c static __le32 rtl8169_get_txd_opts1(u32 opts0, u32 len, unsigned int entry)
entry            5690 drivers/net/ethernet/realtek/r8169_main.c 	if (entry == NUM_TX_DESC - 1)
entry            5700 drivers/net/ethernet/realtek/r8169_main.c 	unsigned int cur_frag, entry;
entry            5704 drivers/net/ethernet/realtek/r8169_main.c 	entry = tp->cur_tx;
entry            5711 drivers/net/ethernet/realtek/r8169_main.c 		entry = (entry + 1) % NUM_TX_DESC;
entry            5713 drivers/net/ethernet/realtek/r8169_main.c 		txd = tp->TxDescArray + entry;
entry            5724 drivers/net/ethernet/realtek/r8169_main.c 		txd->opts1 = rtl8169_get_txd_opts1(opts[0], len, entry);
entry            5728 drivers/net/ethernet/realtek/r8169_main.c 		tp->tx_skb[entry].len = len;
entry            5732 drivers/net/ethernet/realtek/r8169_main.c 		tp->tx_skb[entry].skb = skb;
entry            5884 drivers/net/ethernet/realtek/r8169_main.c 	unsigned int entry = tp->cur_tx % NUM_TX_DESC;
entry            5885 drivers/net/ethernet/realtek/r8169_main.c 	struct TxDesc *txd = tp->TxDescArray + entry;
entry            5919 drivers/net/ethernet/realtek/r8169_main.c 	tp->tx_skb[entry].len = len;
entry            5929 drivers/net/ethernet/realtek/r8169_main.c 		tp->tx_skb[entry].skb = skb;
entry            5941 drivers/net/ethernet/realtek/r8169_main.c 	txd->opts1 = rtl8169_get_txd_opts1(opts[0], len, entry);
entry            5977 drivers/net/ethernet/realtek/r8169_main.c 	rtl8169_unmap_tx_skb(d, tp->tx_skb + entry, txd);
entry            6067 drivers/net/ethernet/realtek/r8169_main.c 		unsigned int entry = dirty_tx % NUM_TX_DESC;
entry            6068 drivers/net/ethernet/realtek/r8169_main.c 		struct ring_info *tx_skb = tp->tx_skb + entry;
entry            6071 drivers/net/ethernet/realtek/r8169_main.c 		status = le32_to_cpu(tp->TxDescArray[entry].opts1);
entry            6082 drivers/net/ethernet/realtek/r8169_main.c 				     tp->TxDescArray + entry);
entry            6149 drivers/net/ethernet/realtek/r8169_main.c 		unsigned int entry = cur_rx % NUM_RX_DESC;
entry            6150 drivers/net/ethernet/realtek/r8169_main.c 		const void *rx_buf = page_address(tp->Rx_databuff[entry]);
entry            6151 drivers/net/ethernet/realtek/r8169_main.c 		struct RxDesc *desc = tp->RxDescArray + entry;
entry             181 drivers/net/ethernet/renesas/ravb_main.c 	int entry;
entry             187 drivers/net/ethernet/renesas/ravb_main.c 		entry = priv->dirty_tx[q] % (priv->num_tx_ring[q] *
entry             189 drivers/net/ethernet/renesas/ravb_main.c 		desc = &priv->tx_ring[q][entry];
entry             197 drivers/net/ethernet/renesas/ravb_main.c 		if (priv->tx_skb[q][entry / num_tx_desc]) {
entry             201 drivers/net/ethernet/renesas/ravb_main.c 			if (entry % num_tx_desc == num_tx_desc - 1) {
entry             202 drivers/net/ethernet/renesas/ravb_main.c 				entry /= num_tx_desc;
entry             203 drivers/net/ethernet/renesas/ravb_main.c 				dev_kfree_skb_any(priv->tx_skb[q][entry]);
entry             204 drivers/net/ethernet/renesas/ravb_main.c 				priv->tx_skb[q][entry] = NULL;
entry             536 drivers/net/ethernet/renesas/ravb_main.c 	int entry = priv->cur_rx[q] % priv->num_rx_ring[q];
entry             550 drivers/net/ethernet/renesas/ravb_main.c 	desc = &priv->rx_ring[q][entry];
entry             581 drivers/net/ethernet/renesas/ravb_main.c 			skb = priv->rx_skb[q][entry];
entry             582 drivers/net/ethernet/renesas/ravb_main.c 			priv->rx_skb[q][entry] = NULL;
entry             609 drivers/net/ethernet/renesas/ravb_main.c 		entry = (++priv->cur_rx[q]) % priv->num_rx_ring[q];
entry             610 drivers/net/ethernet/renesas/ravb_main.c 		desc = &priv->rx_ring[q][entry];
entry             615 drivers/net/ethernet/renesas/ravb_main.c 		entry = priv->dirty_rx[q] % priv->num_rx_ring[q];
entry             616 drivers/net/ethernet/renesas/ravb_main.c 		desc = &priv->rx_ring[q][entry];
entry             619 drivers/net/ethernet/renesas/ravb_main.c 		if (!priv->rx_skb[q][entry]) {
entry             636 drivers/net/ethernet/renesas/ravb_main.c 			priv->rx_skb[q][entry] = skb;
entry            1482 drivers/net/ethernet/renesas/ravb_main.c 	u32 entry;
entry            1498 drivers/net/ethernet/renesas/ravb_main.c 	entry = priv->cur_tx[q] % (priv->num_tx_ring[q] * num_tx_desc);
entry            1499 drivers/net/ethernet/renesas/ravb_main.c 	priv->tx_skb[q][entry / num_tx_desc] = skb;
entry            1503 drivers/net/ethernet/renesas/ravb_main.c 			 entry / num_tx_desc * DPTR_ALIGN;
entry            1527 drivers/net/ethernet/renesas/ravb_main.c 		desc = &priv->tx_ring[q][entry];
entry            1540 drivers/net/ethernet/renesas/ravb_main.c 		desc = &priv->tx_ring[q][entry];
entry            1599 drivers/net/ethernet/renesas/ravb_main.c 	priv->tx_skb[q][entry / num_tx_desc] = NULL;
entry            1271 drivers/net/ethernet/renesas/sh_eth.c 	int entry;
entry            1275 drivers/net/ethernet/renesas/sh_eth.c 		entry = mdp->dirty_tx % mdp->num_tx_ring;
entry            1276 drivers/net/ethernet/renesas/sh_eth.c 		txdesc = &mdp->tx_ring[entry];
entry            1284 drivers/net/ethernet/renesas/sh_eth.c 			   entry, le32_to_cpu(txdesc->status));
entry            1286 drivers/net/ethernet/renesas/sh_eth.c 		if (mdp->tx_skbuff[entry]) {
entry            1291 drivers/net/ethernet/renesas/sh_eth.c 			dev_kfree_skb_irq(mdp->tx_skbuff[entry]);
entry            1292 drivers/net/ethernet/renesas/sh_eth.c 			mdp->tx_skbuff[entry] = NULL;
entry            1296 drivers/net/ethernet/renesas/sh_eth.c 		if (entry >= mdp->num_tx_ring - 1)
entry            1624 drivers/net/ethernet/renesas/sh_eth.c 	int entry = mdp->cur_rx % mdp->num_rx_ring;
entry            1636 drivers/net/ethernet/renesas/sh_eth.c 	rxdesc = &mdp->rx_ring[entry];
entry            1648 drivers/net/ethernet/renesas/sh_eth.c 			   entry, desc_status, pkt_len);
entry            1662 drivers/net/ethernet/renesas/sh_eth.c 		skb = mdp->rx_skbuff[entry];
entry            1684 drivers/net/ethernet/renesas/sh_eth.c 			mdp->rx_skbuff[entry] = NULL;
entry            1700 drivers/net/ethernet/renesas/sh_eth.c 		entry = (++mdp->cur_rx) % mdp->num_rx_ring;
entry            1701 drivers/net/ethernet/renesas/sh_eth.c 		rxdesc = &mdp->rx_ring[entry];
entry            1706 drivers/net/ethernet/renesas/sh_eth.c 		entry = mdp->dirty_rx % mdp->num_rx_ring;
entry            1707 drivers/net/ethernet/renesas/sh_eth.c 		rxdesc = &mdp->rx_ring[entry];
entry            1712 drivers/net/ethernet/renesas/sh_eth.c 		if (mdp->rx_skbuff[entry] == NULL) {
entry            1723 drivers/net/ethernet/renesas/sh_eth.c 			mdp->rx_skbuff[entry] = skb;
entry            1729 drivers/net/ethernet/renesas/sh_eth.c 		if (entry >= mdp->num_rx_ring - 1)
entry            2525 drivers/net/ethernet/renesas/sh_eth.c 	u32 entry;
entry            2542 drivers/net/ethernet/renesas/sh_eth.c 	entry = mdp->cur_tx % mdp->num_tx_ring;
entry            2543 drivers/net/ethernet/renesas/sh_eth.c 	mdp->tx_skbuff[entry] = skb;
entry            2544 drivers/net/ethernet/renesas/sh_eth.c 	txdesc = &mdp->tx_ring[entry];
entry            2558 drivers/net/ethernet/renesas/sh_eth.c 	if (entry >= mdp->num_tx_ring - 1)
entry            2676 drivers/net/ethernet/renesas/sh_eth.c static u32 sh_eth_tsu_get_post_mask(int entry)
entry            2678 drivers/net/ethernet/renesas/sh_eth.c 	return 0x0f << (28 - ((entry % 8) * 4));
entry            2681 drivers/net/ethernet/renesas/sh_eth.c static u32 sh_eth_tsu_get_post_bit(struct sh_eth_private *mdp, int entry)
entry            2683 drivers/net/ethernet/renesas/sh_eth.c 	return (0x08 >> (mdp->port << 1)) << (28 - ((entry % 8) * 4));
entry            2687 drivers/net/ethernet/renesas/sh_eth.c 					     int entry)
entry            2690 drivers/net/ethernet/renesas/sh_eth.c 	int reg = TSU_POST1 + entry / 8;
entry            2694 drivers/net/ethernet/renesas/sh_eth.c 	sh_eth_tsu_write(mdp, tmp | sh_eth_tsu_get_post_bit(mdp, entry), reg);
entry            2698 drivers/net/ethernet/renesas/sh_eth.c 					      int entry)
entry            2701 drivers/net/ethernet/renesas/sh_eth.c 	int reg = TSU_POST1 + entry / 8;
entry            2704 drivers/net/ethernet/renesas/sh_eth.c 	post_mask = sh_eth_tsu_get_post_mask(entry);
entry            2705 drivers/net/ethernet/renesas/sh_eth.c 	ref_mask = sh_eth_tsu_get_post_bit(mdp, entry) & ~post_mask;
entry            2785 drivers/net/ethernet/renesas/sh_eth.c 	int entry;
entry            2788 drivers/net/ethernet/renesas/sh_eth.c 	entry = sh_eth_tsu_find_entry(ndev, blank);
entry            2789 drivers/net/ethernet/renesas/sh_eth.c 	return (entry < 0) ? -ENOMEM : entry;
entry            2793 drivers/net/ethernet/renesas/sh_eth.c 					      int entry)
entry            2801 drivers/net/ethernet/renesas/sh_eth.c 			 ~(1 << (31 - entry)), TSU_TEN);
entry            2804 drivers/net/ethernet/renesas/sh_eth.c 	ret = sh_eth_tsu_write_entry(ndev, reg_offset + entry * 8, blank);
entry            2698 drivers/net/ethernet/rocker/rocker_main.c 		rocker->msix_entries[i].entry = i;
entry              93 drivers/net/ethernet/rocker/rocker_ofdpa.c 	struct hlist_node entry;
entry             103 drivers/net/ethernet/rocker/rocker_ofdpa.c 	struct hlist_node entry;
entry             129 drivers/net/ethernet/rocker/rocker_ofdpa.c 	struct hlist_node entry;
entry             141 drivers/net/ethernet/rocker/rocker_ofdpa.c 	struct hlist_node entry;
entry             148 drivers/net/ethernet/rocker/rocker_ofdpa.c 	struct hlist_node entry;
entry             306 drivers/net/ethernet/rocker/rocker_ofdpa.c 			       const struct ofdpa_flow_tbl_entry *entry)
entry             309 drivers/net/ethernet/rocker/rocker_ofdpa.c 			       entry->key.ig_port.in_pport))
entry             312 drivers/net/ethernet/rocker/rocker_ofdpa.c 			       entry->key.ig_port.in_pport_mask))
entry             315 drivers/net/ethernet/rocker/rocker_ofdpa.c 			       entry->key.ig_port.goto_tbl))
entry             323 drivers/net/ethernet/rocker/rocker_ofdpa.c 			    const struct ofdpa_flow_tbl_entry *entry)
entry             326 drivers/net/ethernet/rocker/rocker_ofdpa.c 			       entry->key.vlan.in_pport))
entry             329 drivers/net/ethernet/rocker/rocker_ofdpa.c 				entry->key.vlan.vlan_id))
entry             332 drivers/net/ethernet/rocker/rocker_ofdpa.c 				entry->key.vlan.vlan_id_mask))
entry             335 drivers/net/ethernet/rocker/rocker_ofdpa.c 			       entry->key.vlan.goto_tbl))
entry             337 drivers/net/ethernet/rocker/rocker_ofdpa.c 	if (entry->key.vlan.untagged &&
entry             339 drivers/net/ethernet/rocker/rocker_ofdpa.c 				entry->key.vlan.new_vlan_id))
entry             347 drivers/net/ethernet/rocker/rocker_ofdpa.c 				const struct ofdpa_flow_tbl_entry *entry)
entry             350 drivers/net/ethernet/rocker/rocker_ofdpa.c 			       entry->key.term_mac.in_pport))
entry             353 drivers/net/ethernet/rocker/rocker_ofdpa.c 			       entry->key.term_mac.in_pport_mask))
entry             356 drivers/net/ethernet/rocker/rocker_ofdpa.c 				entry->key.term_mac.eth_type))
entry             359 drivers/net/ethernet/rocker/rocker_ofdpa.c 			   ETH_ALEN, entry->key.term_mac.eth_dst))
entry             362 drivers/net/ethernet/rocker/rocker_ofdpa.c 			   ETH_ALEN, entry->key.term_mac.eth_dst_mask))
entry             365 drivers/net/ethernet/rocker/rocker_ofdpa.c 				entry->key.term_mac.vlan_id))
entry             368 drivers/net/ethernet/rocker/rocker_ofdpa.c 				entry->key.term_mac.vlan_id_mask))
entry             371 drivers/net/ethernet/rocker/rocker_ofdpa.c 			       entry->key.term_mac.goto_tbl))
entry             373 drivers/net/ethernet/rocker/rocker_ofdpa.c 	if (entry->key.term_mac.copy_to_cpu &&
entry             375 drivers/net/ethernet/rocker/rocker_ofdpa.c 			      entry->key.term_mac.copy_to_cpu))
entry             383 drivers/net/ethernet/rocker/rocker_ofdpa.c 				     const struct ofdpa_flow_tbl_entry *entry)
entry             386 drivers/net/ethernet/rocker/rocker_ofdpa.c 				entry->key.ucast_routing.eth_type))
entry             389 drivers/net/ethernet/rocker/rocker_ofdpa.c 				entry->key.ucast_routing.dst4))
entry             392 drivers/net/ethernet/rocker/rocker_ofdpa.c 				entry->key.ucast_routing.dst4_mask))
entry             395 drivers/net/ethernet/rocker/rocker_ofdpa.c 			       entry->key.ucast_routing.goto_tbl))
entry             398 drivers/net/ethernet/rocker/rocker_ofdpa.c 			       entry->key.ucast_routing.group_id))
entry             406 drivers/net/ethernet/rocker/rocker_ofdpa.c 			      const struct ofdpa_flow_tbl_entry *entry)
entry             408 drivers/net/ethernet/rocker/rocker_ofdpa.c 	if (entry->key.bridge.has_eth_dst &&
entry             410 drivers/net/ethernet/rocker/rocker_ofdpa.c 			   ETH_ALEN, entry->key.bridge.eth_dst))
entry             412 drivers/net/ethernet/rocker/rocker_ofdpa.c 	if (entry->key.bridge.has_eth_dst_mask &&
entry             414 drivers/net/ethernet/rocker/rocker_ofdpa.c 			   ETH_ALEN, entry->key.bridge.eth_dst_mask))
entry             416 drivers/net/ethernet/rocker/rocker_ofdpa.c 	if (entry->key.bridge.vlan_id &&
entry             418 drivers/net/ethernet/rocker/rocker_ofdpa.c 				entry->key.bridge.vlan_id))
entry             420 drivers/net/ethernet/rocker/rocker_ofdpa.c 	if (entry->key.bridge.tunnel_id &&
entry             422 drivers/net/ethernet/rocker/rocker_ofdpa.c 			       entry->key.bridge.tunnel_id))
entry             425 drivers/net/ethernet/rocker/rocker_ofdpa.c 			       entry->key.bridge.goto_tbl))
entry             428 drivers/net/ethernet/rocker/rocker_ofdpa.c 			       entry->key.bridge.group_id))
entry             430 drivers/net/ethernet/rocker/rocker_ofdpa.c 	if (entry->key.bridge.copy_to_cpu &&
entry             432 drivers/net/ethernet/rocker/rocker_ofdpa.c 			      entry->key.bridge.copy_to_cpu))
entry             440 drivers/net/ethernet/rocker/rocker_ofdpa.c 			   const struct ofdpa_flow_tbl_entry *entry)
entry             443 drivers/net/ethernet/rocker/rocker_ofdpa.c 			       entry->key.acl.in_pport))
entry             446 drivers/net/ethernet/rocker/rocker_ofdpa.c 			       entry->key.acl.in_pport_mask))
entry             449 drivers/net/ethernet/rocker/rocker_ofdpa.c 			   ETH_ALEN, entry->key.acl.eth_src))
entry             452 drivers/net/ethernet/rocker/rocker_ofdpa.c 			   ETH_ALEN, entry->key.acl.eth_src_mask))
entry             455 drivers/net/ethernet/rocker/rocker_ofdpa.c 			   ETH_ALEN, entry->key.acl.eth_dst))
entry             458 drivers/net/ethernet/rocker/rocker_ofdpa.c 			   ETH_ALEN, entry->key.acl.eth_dst_mask))
entry             461 drivers/net/ethernet/rocker/rocker_ofdpa.c 				entry->key.acl.eth_type))
entry             464 drivers/net/ethernet/rocker/rocker_ofdpa.c 				entry->key.acl.vlan_id))
entry             467 drivers/net/ethernet/rocker/rocker_ofdpa.c 				entry->key.acl.vlan_id_mask))
entry             470 drivers/net/ethernet/rocker/rocker_ofdpa.c 	switch (ntohs(entry->key.acl.eth_type)) {
entry             474 drivers/net/ethernet/rocker/rocker_ofdpa.c 				      entry->key.acl.ip_proto))
entry             478 drivers/net/ethernet/rocker/rocker_ofdpa.c 				      entry->key.acl.ip_proto_mask))
entry             481 drivers/net/ethernet/rocker/rocker_ofdpa.c 				      entry->key.acl.ip_tos & 0x3f))
entry             485 drivers/net/ethernet/rocker/rocker_ofdpa.c 				      entry->key.acl.ip_tos_mask & 0x3f))
entry             488 drivers/net/ethernet/rocker/rocker_ofdpa.c 				      (entry->key.acl.ip_tos & 0xc0) >> 6))
entry             492 drivers/net/ethernet/rocker/rocker_ofdpa.c 				      (entry->key.acl.ip_tos_mask & 0xc0) >> 6))
entry             497 drivers/net/ethernet/rocker/rocker_ofdpa.c 	if (entry->key.acl.group_id != ROCKER_GROUP_NONE &&
entry             499 drivers/net/ethernet/rocker/rocker_ofdpa.c 			       entry->key.acl.group_id))
entry             509 drivers/net/ethernet/rocker/rocker_ofdpa.c 	const struct ofdpa_flow_tbl_entry *entry = priv;
entry             513 drivers/net/ethernet/rocker/rocker_ofdpa.c 	if (rocker_tlv_put_u16(desc_info, ROCKER_TLV_CMD_TYPE, entry->cmd))
entry             519 drivers/net/ethernet/rocker/rocker_ofdpa.c 			       entry->key.tbl_id))
entry             522 drivers/net/ethernet/rocker/rocker_ofdpa.c 			       entry->key.priority))
entry             527 drivers/net/ethernet/rocker/rocker_ofdpa.c 			       entry->cookie))
entry             530 drivers/net/ethernet/rocker/rocker_ofdpa.c 	switch (entry->key.tbl_id) {
entry             532 drivers/net/ethernet/rocker/rocker_ofdpa.c 		err = ofdpa_cmd_flow_tbl_add_ig_port(desc_info, entry);
entry             535 drivers/net/ethernet/rocker/rocker_ofdpa.c 		err = ofdpa_cmd_flow_tbl_add_vlan(desc_info, entry);
entry             538 drivers/net/ethernet/rocker/rocker_ofdpa.c 		err = ofdpa_cmd_flow_tbl_add_term_mac(desc_info, entry);
entry             541 drivers/net/ethernet/rocker/rocker_ofdpa.c 		err = ofdpa_cmd_flow_tbl_add_ucast_routing(desc_info, entry);
entry             544 drivers/net/ethernet/rocker/rocker_ofdpa.c 		err = ofdpa_cmd_flow_tbl_add_bridge(desc_info, entry);
entry             547 drivers/net/ethernet/rocker/rocker_ofdpa.c 		err = ofdpa_cmd_flow_tbl_add_acl(desc_info, entry);
entry             566 drivers/net/ethernet/rocker/rocker_ofdpa.c 	const struct ofdpa_flow_tbl_entry *entry = priv;
entry             569 drivers/net/ethernet/rocker/rocker_ofdpa.c 	if (rocker_tlv_put_u16(desc_info, ROCKER_TLV_CMD_TYPE, entry->cmd))
entry             575 drivers/net/ethernet/rocker/rocker_ofdpa.c 			       entry->cookie))
entry             584 drivers/net/ethernet/rocker/rocker_ofdpa.c 				     struct ofdpa_group_tbl_entry *entry)
entry             587 drivers/net/ethernet/rocker/rocker_ofdpa.c 			       ROCKER_GROUP_PORT_GET(entry->group_id)))
entry             590 drivers/net/ethernet/rocker/rocker_ofdpa.c 			      entry->l2_interface.pop_vlan))
entry             598 drivers/net/ethernet/rocker/rocker_ofdpa.c 				   const struct ofdpa_group_tbl_entry *entry)
entry             601 drivers/net/ethernet/rocker/rocker_ofdpa.c 			       entry->l2_rewrite.group_id))
entry             603 drivers/net/ethernet/rocker/rocker_ofdpa.c 	if (!is_zero_ether_addr(entry->l2_rewrite.eth_src) &&
entry             605 drivers/net/ethernet/rocker/rocker_ofdpa.c 			   ETH_ALEN, entry->l2_rewrite.eth_src))
entry             607 drivers/net/ethernet/rocker/rocker_ofdpa.c 	if (!is_zero_ether_addr(entry->l2_rewrite.eth_dst) &&
entry             609 drivers/net/ethernet/rocker/rocker_ofdpa.c 			   ETH_ALEN, entry->l2_rewrite.eth_dst))
entry             611 drivers/net/ethernet/rocker/rocker_ofdpa.c 	if (entry->l2_rewrite.vlan_id &&
entry             613 drivers/net/ethernet/rocker/rocker_ofdpa.c 				entry->l2_rewrite.vlan_id))
entry             621 drivers/net/ethernet/rocker/rocker_ofdpa.c 				  const struct ofdpa_group_tbl_entry *entry)
entry             627 drivers/net/ethernet/rocker/rocker_ofdpa.c 			       entry->group_count))
entry             635 drivers/net/ethernet/rocker/rocker_ofdpa.c 	for (i = 0; i < entry->group_count; i++)
entry             637 drivers/net/ethernet/rocker/rocker_ofdpa.c 		if (rocker_tlv_put_u32(desc_info, i + 1, entry->group_ids[i]))
entry             647 drivers/net/ethernet/rocker/rocker_ofdpa.c 				   const struct ofdpa_group_tbl_entry *entry)
entry             649 drivers/net/ethernet/rocker/rocker_ofdpa.c 	if (!is_zero_ether_addr(entry->l3_unicast.eth_src) &&
entry             651 drivers/net/ethernet/rocker/rocker_ofdpa.c 			   ETH_ALEN, entry->l3_unicast.eth_src))
entry             653 drivers/net/ethernet/rocker/rocker_ofdpa.c 	if (!is_zero_ether_addr(entry->l3_unicast.eth_dst) &&
entry             655 drivers/net/ethernet/rocker/rocker_ofdpa.c 			   ETH_ALEN, entry->l3_unicast.eth_dst))
entry             657 drivers/net/ethernet/rocker/rocker_ofdpa.c 	if (entry->l3_unicast.vlan_id &&
entry             659 drivers/net/ethernet/rocker/rocker_ofdpa.c 				entry->l3_unicast.vlan_id))
entry             662 drivers/net/ethernet/rocker/rocker_ofdpa.c 			      entry->l3_unicast.ttl_check))
entry             665 drivers/net/ethernet/rocker/rocker_ofdpa.c 			       entry->l3_unicast.group_id))
entry             675 drivers/net/ethernet/rocker/rocker_ofdpa.c 	struct ofdpa_group_tbl_entry *entry = priv;
entry             679 drivers/net/ethernet/rocker/rocker_ofdpa.c 	if (rocker_tlv_put_u16(desc_info, ROCKER_TLV_CMD_TYPE, entry->cmd))
entry             686 drivers/net/ethernet/rocker/rocker_ofdpa.c 			       entry->group_id))
entry             689 drivers/net/ethernet/rocker/rocker_ofdpa.c 	switch (ROCKER_GROUP_TYPE_GET(entry->group_id)) {
entry             691 drivers/net/ethernet/rocker/rocker_ofdpa.c 		err = ofdpa_cmd_group_tbl_add_l2_interface(desc_info, entry);
entry             694 drivers/net/ethernet/rocker/rocker_ofdpa.c 		err = ofdpa_cmd_group_tbl_add_l2_rewrite(desc_info, entry);
entry             698 drivers/net/ethernet/rocker/rocker_ofdpa.c 		err = ofdpa_cmd_group_tbl_add_group_ids(desc_info, entry);
entry             701 drivers/net/ethernet/rocker/rocker_ofdpa.c 		err = ofdpa_cmd_group_tbl_add_l3_unicast(desc_info, entry);
entry             720 drivers/net/ethernet/rocker/rocker_ofdpa.c 	const struct ofdpa_group_tbl_entry *entry = priv;
entry             723 drivers/net/ethernet/rocker/rocker_ofdpa.c 	if (rocker_tlv_put_u16(desc_info, ROCKER_TLV_CMD_TYPE, entry->cmd))
entry             729 drivers/net/ethernet/rocker/rocker_ofdpa.c 			       entry->group_id))
entry             748 drivers/net/ethernet/rocker/rocker_ofdpa.c 			       entry, match->key_crc32) {
entry             772 drivers/net/ethernet/rocker/rocker_ofdpa.c 		hash_del(&found->entry);
entry             782 drivers/net/ethernet/rocker/rocker_ofdpa.c 	hash_add(ofdpa->flow_tbl, &found->entry, found->key_crc32);
entry             807 drivers/net/ethernet/rocker/rocker_ofdpa.c 		hash_del(&found->entry);
entry             827 drivers/net/ethernet/rocker/rocker_ofdpa.c 			     struct ofdpa_flow_tbl_entry *entry)
entry             830 drivers/net/ethernet/rocker/rocker_ofdpa.c 		return ofdpa_flow_tbl_del(ofdpa_port, flags, entry);
entry             832 drivers/net/ethernet/rocker/rocker_ofdpa.c 		return ofdpa_flow_tbl_add(ofdpa_port, flags, entry);
entry             839 drivers/net/ethernet/rocker/rocker_ofdpa.c 	struct ofdpa_flow_tbl_entry *entry;
entry             841 drivers/net/ethernet/rocker/rocker_ofdpa.c 	entry = kzalloc(sizeof(*entry), GFP_KERNEL);
entry             842 drivers/net/ethernet/rocker/rocker_ofdpa.c 	if (!entry)
entry             845 drivers/net/ethernet/rocker/rocker_ofdpa.c 	entry->key.priority = OFDPA_PRIORITY_IG_PORT;
entry             846 drivers/net/ethernet/rocker/rocker_ofdpa.c 	entry->key.tbl_id = ROCKER_OF_DPA_TABLE_ID_INGRESS_PORT;
entry             847 drivers/net/ethernet/rocker/rocker_ofdpa.c 	entry->key.ig_port.in_pport = in_pport;
entry             848 drivers/net/ethernet/rocker/rocker_ofdpa.c 	entry->key.ig_port.in_pport_mask = in_pport_mask;
entry             849 drivers/net/ethernet/rocker/rocker_ofdpa.c 	entry->key.ig_port.goto_tbl = goto_tbl;
entry             851 drivers/net/ethernet/rocker/rocker_ofdpa.c 	return ofdpa_flow_tbl_do(ofdpa_port, flags, entry);
entry             861 drivers/net/ethernet/rocker/rocker_ofdpa.c 	struct ofdpa_flow_tbl_entry *entry;
entry             863 drivers/net/ethernet/rocker/rocker_ofdpa.c 	entry = kzalloc(sizeof(*entry), GFP_KERNEL);
entry             864 drivers/net/ethernet/rocker/rocker_ofdpa.c 	if (!entry)
entry             867 drivers/net/ethernet/rocker/rocker_ofdpa.c 	entry->key.priority = OFDPA_PRIORITY_VLAN;
entry             868 drivers/net/ethernet/rocker/rocker_ofdpa.c 	entry->key.tbl_id = ROCKER_OF_DPA_TABLE_ID_VLAN;
entry             869 drivers/net/ethernet/rocker/rocker_ofdpa.c 	entry->key.vlan.in_pport = in_pport;
entry             870 drivers/net/ethernet/rocker/rocker_ofdpa.c 	entry->key.vlan.vlan_id = vlan_id;
entry             871 drivers/net/ethernet/rocker/rocker_ofdpa.c 	entry->key.vlan.vlan_id_mask = vlan_id_mask;
entry             872 drivers/net/ethernet/rocker/rocker_ofdpa.c 	entry->key.vlan.goto_tbl = goto_tbl;
entry             874 drivers/net/ethernet/rocker/rocker_ofdpa.c 	entry->key.vlan.untagged = untagged;
entry             875 drivers/net/ethernet/rocker/rocker_ofdpa.c 	entry->key.vlan.new_vlan_id = new_vlan_id;
entry             877 drivers/net/ethernet/rocker/rocker_ofdpa.c 	return ofdpa_flow_tbl_do(ofdpa_port, flags, entry);
entry             887 drivers/net/ethernet/rocker/rocker_ofdpa.c 	struct ofdpa_flow_tbl_entry *entry;
entry             889 drivers/net/ethernet/rocker/rocker_ofdpa.c 	entry = kzalloc(sizeof(*entry), GFP_KERNEL);
entry             890 drivers/net/ethernet/rocker/rocker_ofdpa.c 	if (!entry)
entry             894 drivers/net/ethernet/rocker/rocker_ofdpa.c 		entry->key.priority = OFDPA_PRIORITY_TERM_MAC_MCAST;
entry             895 drivers/net/ethernet/rocker/rocker_ofdpa.c 		entry->key.term_mac.goto_tbl =
entry             898 drivers/net/ethernet/rocker/rocker_ofdpa.c 		entry->key.priority = OFDPA_PRIORITY_TERM_MAC_UCAST;
entry             899 drivers/net/ethernet/rocker/rocker_ofdpa.c 		entry->key.term_mac.goto_tbl =
entry             903 drivers/net/ethernet/rocker/rocker_ofdpa.c 	entry->key.tbl_id = ROCKER_OF_DPA_TABLE_ID_TERMINATION_MAC;
entry             904 drivers/net/ethernet/rocker/rocker_ofdpa.c 	entry->key.term_mac.in_pport = in_pport;
entry             905 drivers/net/ethernet/rocker/rocker_ofdpa.c 	entry->key.term_mac.in_pport_mask = in_pport_mask;
entry             906 drivers/net/ethernet/rocker/rocker_ofdpa.c 	entry->key.term_mac.eth_type = eth_type;
entry             907 drivers/net/ethernet/rocker/rocker_ofdpa.c 	ether_addr_copy(entry->key.term_mac.eth_dst, eth_dst);
entry             908 drivers/net/ethernet/rocker/rocker_ofdpa.c 	ether_addr_copy(entry->key.term_mac.eth_dst_mask, eth_dst_mask);
entry             909 drivers/net/ethernet/rocker/rocker_ofdpa.c 	entry->key.term_mac.vlan_id = vlan_id;
entry             910 drivers/net/ethernet/rocker/rocker_ofdpa.c 	entry->key.term_mac.vlan_id_mask = vlan_id_mask;
entry             911 drivers/net/ethernet/rocker/rocker_ofdpa.c 	entry->key.term_mac.copy_to_cpu = copy_to_cpu;
entry             913 drivers/net/ethernet/rocker/rocker_ofdpa.c 	return ofdpa_flow_tbl_do(ofdpa_port, flags, entry);
entry             923 drivers/net/ethernet/rocker/rocker_ofdpa.c 	struct ofdpa_flow_tbl_entry *entry;
entry             929 drivers/net/ethernet/rocker/rocker_ofdpa.c 	entry = kzalloc(sizeof(*entry), GFP_ATOMIC);
entry             930 drivers/net/ethernet/rocker/rocker_ofdpa.c 	if (!entry)
entry             933 drivers/net/ethernet/rocker/rocker_ofdpa.c 	entry->key.tbl_id = ROCKER_OF_DPA_TABLE_ID_BRIDGING;
entry             936 drivers/net/ethernet/rocker/rocker_ofdpa.c 		entry->key.bridge.has_eth_dst = 1;
entry             937 drivers/net/ethernet/rocker/rocker_ofdpa.c 		ether_addr_copy(entry->key.bridge.eth_dst, eth_dst);
entry             940 drivers/net/ethernet/rocker/rocker_ofdpa.c 		entry->key.bridge.has_eth_dst_mask = 1;
entry             941 drivers/net/ethernet/rocker/rocker_ofdpa.c 		ether_addr_copy(entry->key.bridge.eth_dst_mask, eth_dst_mask);
entry             960 drivers/net/ethernet/rocker/rocker_ofdpa.c 	entry->key.priority = priority;
entry             961 drivers/net/ethernet/rocker/rocker_ofdpa.c 	entry->key.bridge.vlan_id = vlan_id;
entry             962 drivers/net/ethernet/rocker/rocker_ofdpa.c 	entry->key.bridge.tunnel_id = tunnel_id;
entry             963 drivers/net/ethernet/rocker/rocker_ofdpa.c 	entry->key.bridge.goto_tbl = goto_tbl;
entry             964 drivers/net/ethernet/rocker/rocker_ofdpa.c 	entry->key.bridge.group_id = group_id;
entry             965 drivers/net/ethernet/rocker/rocker_ofdpa.c 	entry->key.bridge.copy_to_cpu = copy_to_cpu;
entry             967 drivers/net/ethernet/rocker/rocker_ofdpa.c 	return ofdpa_flow_tbl_do(ofdpa_port, flags, entry);
entry             977 drivers/net/ethernet/rocker/rocker_ofdpa.c 	struct ofdpa_flow_tbl_entry *entry;
entry             979 drivers/net/ethernet/rocker/rocker_ofdpa.c 	entry = kzalloc(sizeof(*entry), GFP_KERNEL);
entry             980 drivers/net/ethernet/rocker/rocker_ofdpa.c 	if (!entry)
entry             983 drivers/net/ethernet/rocker/rocker_ofdpa.c 	entry->key.tbl_id = ROCKER_OF_DPA_TABLE_ID_UNICAST_ROUTING;
entry             984 drivers/net/ethernet/rocker/rocker_ofdpa.c 	entry->key.priority = priority;
entry             985 drivers/net/ethernet/rocker/rocker_ofdpa.c 	entry->key.ucast_routing.eth_type = eth_type;
entry             986 drivers/net/ethernet/rocker/rocker_ofdpa.c 	entry->key.ucast_routing.dst4 = dst;
entry             987 drivers/net/ethernet/rocker/rocker_ofdpa.c 	entry->key.ucast_routing.dst4_mask = dst_mask;
entry             988 drivers/net/ethernet/rocker/rocker_ofdpa.c 	entry->key.ucast_routing.goto_tbl = goto_tbl;
entry             989 drivers/net/ethernet/rocker/rocker_ofdpa.c 	entry->key.ucast_routing.group_id = group_id;
entry             990 drivers/net/ethernet/rocker/rocker_ofdpa.c 	entry->key_len = offsetof(struct ofdpa_flow_tbl_key,
entry             992 drivers/net/ethernet/rocker/rocker_ofdpa.c 	entry->fi = fi;
entry             994 drivers/net/ethernet/rocker/rocker_ofdpa.c 	return ofdpa_flow_tbl_do(ofdpa_port, flags, entry);
entry            1007 drivers/net/ethernet/rocker/rocker_ofdpa.c 	struct ofdpa_flow_tbl_entry *entry;
entry            1009 drivers/net/ethernet/rocker/rocker_ofdpa.c 	entry = kzalloc(sizeof(*entry), GFP_KERNEL);
entry            1010 drivers/net/ethernet/rocker/rocker_ofdpa.c 	if (!entry)
entry            1021 drivers/net/ethernet/rocker/rocker_ofdpa.c 	entry->key.priority = priority;
entry            1022 drivers/net/ethernet/rocker/rocker_ofdpa.c 	entry->key.tbl_id = ROCKER_OF_DPA_TABLE_ID_ACL_POLICY;
entry            1023 drivers/net/ethernet/rocker/rocker_ofdpa.c 	entry->key.acl.in_pport = in_pport;
entry            1024 drivers/net/ethernet/rocker/rocker_ofdpa.c 	entry->key.acl.in_pport_mask = in_pport_mask;
entry            1027 drivers/net/ethernet/rocker/rocker_ofdpa.c 		ether_addr_copy(entry->key.acl.eth_src, eth_src);
entry            1029 drivers/net/ethernet/rocker/rocker_ofdpa.c 		ether_addr_copy(entry->key.acl.eth_src_mask, eth_src_mask);
entry            1031 drivers/net/ethernet/rocker/rocker_ofdpa.c 		ether_addr_copy(entry->key.acl.eth_dst, eth_dst);
entry            1033 drivers/net/ethernet/rocker/rocker_ofdpa.c 		ether_addr_copy(entry->key.acl.eth_dst_mask, eth_dst_mask);
entry            1035 drivers/net/ethernet/rocker/rocker_ofdpa.c 	entry->key.acl.eth_type = eth_type;
entry            1036 drivers/net/ethernet/rocker/rocker_ofdpa.c 	entry->key.acl.vlan_id = vlan_id;
entry            1037 drivers/net/ethernet/rocker/rocker_ofdpa.c 	entry->key.acl.vlan_id_mask = vlan_id_mask;
entry            1038 drivers/net/ethernet/rocker/rocker_ofdpa.c 	entry->key.acl.ip_proto = ip_proto;
entry            1039 drivers/net/ethernet/rocker/rocker_ofdpa.c 	entry->key.acl.ip_proto_mask = ip_proto_mask;
entry            1040 drivers/net/ethernet/rocker/rocker_ofdpa.c 	entry->key.acl.ip_tos = ip_tos;
entry            1041 drivers/net/ethernet/rocker/rocker_ofdpa.c 	entry->key.acl.ip_tos_mask = ip_tos_mask;
entry            1042 drivers/net/ethernet/rocker/rocker_ofdpa.c 	entry->key.acl.group_id = group_id;
entry            1044 drivers/net/ethernet/rocker/rocker_ofdpa.c 	return ofdpa_flow_tbl_do(ofdpa_port, flags, entry);
entry            1054 drivers/net/ethernet/rocker/rocker_ofdpa.c 			       entry, match->group_id) {
entry            1062 drivers/net/ethernet/rocker/rocker_ofdpa.c static void ofdpa_group_tbl_entry_free(struct ofdpa_group_tbl_entry *entry)
entry            1064 drivers/net/ethernet/rocker/rocker_ofdpa.c 	switch (ROCKER_GROUP_TYPE_GET(entry->group_id)) {
entry            1067 drivers/net/ethernet/rocker/rocker_ofdpa.c 		kfree(entry->group_ids);
entry            1072 drivers/net/ethernet/rocker/rocker_ofdpa.c 	kfree(entry);
entry            1087 drivers/net/ethernet/rocker/rocker_ofdpa.c 		hash_del(&found->entry);
entry            1096 drivers/net/ethernet/rocker/rocker_ofdpa.c 	hash_add(ofdpa->group_tbl, &found->entry, found->group_id);
entry            1119 drivers/net/ethernet/rocker/rocker_ofdpa.c 		hash_del(&found->entry);
entry            1139 drivers/net/ethernet/rocker/rocker_ofdpa.c 			      struct ofdpa_group_tbl_entry *entry)
entry            1142 drivers/net/ethernet/rocker/rocker_ofdpa.c 		return ofdpa_group_tbl_del(ofdpa_port, flags, entry);
entry            1144 drivers/net/ethernet/rocker/rocker_ofdpa.c 		return ofdpa_group_tbl_add(ofdpa_port, flags, entry);
entry            1151 drivers/net/ethernet/rocker/rocker_ofdpa.c 	struct ofdpa_group_tbl_entry *entry;
entry            1153 drivers/net/ethernet/rocker/rocker_ofdpa.c 	entry = kzalloc(sizeof(*entry), GFP_KERNEL);
entry            1154 drivers/net/ethernet/rocker/rocker_ofdpa.c 	if (!entry)
entry            1157 drivers/net/ethernet/rocker/rocker_ofdpa.c 	entry->group_id = ROCKER_GROUP_L2_INTERFACE(vlan_id, out_pport);
entry            1158 drivers/net/ethernet/rocker/rocker_ofdpa.c 	entry->l2_interface.pop_vlan = pop_vlan;
entry            1160 drivers/net/ethernet/rocker/rocker_ofdpa.c 	return ofdpa_group_tbl_do(ofdpa_port, flags, entry);
entry            1167 drivers/net/ethernet/rocker/rocker_ofdpa.c 	struct ofdpa_group_tbl_entry *entry;
entry            1169 drivers/net/ethernet/rocker/rocker_ofdpa.c 	entry = kzalloc(sizeof(*entry), GFP_KERNEL);
entry            1170 drivers/net/ethernet/rocker/rocker_ofdpa.c 	if (!entry)
entry            1173 drivers/net/ethernet/rocker/rocker_ofdpa.c 	entry->group_id = group_id;
entry            1174 drivers/net/ethernet/rocker/rocker_ofdpa.c 	entry->group_count = group_count;
entry            1176 drivers/net/ethernet/rocker/rocker_ofdpa.c 	entry->group_ids = kcalloc(group_count, sizeof(u32), GFP_KERNEL);
entry            1177 drivers/net/ethernet/rocker/rocker_ofdpa.c 	if (!entry->group_ids) {
entry            1178 drivers/net/ethernet/rocker/rocker_ofdpa.c 		kfree(entry);
entry            1181 drivers/net/ethernet/rocker/rocker_ofdpa.c 	memcpy(entry->group_ids, group_ids, group_count * sizeof(u32));
entry            1183 drivers/net/ethernet/rocker/rocker_ofdpa.c 	return ofdpa_group_tbl_do(ofdpa_port, flags, entry);
entry            1200 drivers/net/ethernet/rocker/rocker_ofdpa.c 	struct ofdpa_group_tbl_entry *entry;
entry            1202 drivers/net/ethernet/rocker/rocker_ofdpa.c 	entry = kzalloc(sizeof(*entry), GFP_KERNEL);
entry            1203 drivers/net/ethernet/rocker/rocker_ofdpa.c 	if (!entry)
entry            1206 drivers/net/ethernet/rocker/rocker_ofdpa.c 	entry->group_id = ROCKER_GROUP_L3_UNICAST(index);
entry            1208 drivers/net/ethernet/rocker/rocker_ofdpa.c 		ether_addr_copy(entry->l3_unicast.eth_src, src_mac);
entry            1210 drivers/net/ethernet/rocker/rocker_ofdpa.c 		ether_addr_copy(entry->l3_unicast.eth_dst, dst_mac);
entry            1211 drivers/net/ethernet/rocker/rocker_ofdpa.c 	entry->l3_unicast.vlan_id = vlan_id;
entry            1212 drivers/net/ethernet/rocker/rocker_ofdpa.c 	entry->l3_unicast.ttl_check = ttl_check;
entry            1213 drivers/net/ethernet/rocker/rocker_ofdpa.c 	entry->l3_unicast.group_id = ROCKER_GROUP_L2_INTERFACE(vlan_id, pport);
entry            1215 drivers/net/ethernet/rocker/rocker_ofdpa.c 	return ofdpa_group_tbl_do(ofdpa_port, flags, entry);
entry            1224 drivers/net/ethernet/rocker/rocker_ofdpa.c 			       entry, be32_to_cpu(ip_addr))
entry            1232 drivers/net/ethernet/rocker/rocker_ofdpa.c 			    struct ofdpa_neigh_tbl_entry *entry)
entry            1234 drivers/net/ethernet/rocker/rocker_ofdpa.c 	entry->index = ofdpa->neigh_tbl_next_index++;
entry            1235 drivers/net/ethernet/rocker/rocker_ofdpa.c 	entry->ref_count++;
entry            1236 drivers/net/ethernet/rocker/rocker_ofdpa.c 	hash_add(ofdpa->neigh_tbl, &entry->entry,
entry            1237 drivers/net/ethernet/rocker/rocker_ofdpa.c 		 be32_to_cpu(entry->ip_addr));
entry            1240 drivers/net/ethernet/rocker/rocker_ofdpa.c static void ofdpa_neigh_del(struct ofdpa_neigh_tbl_entry *entry)
entry            1242 drivers/net/ethernet/rocker/rocker_ofdpa.c 	if (--entry->ref_count == 0) {
entry            1243 drivers/net/ethernet/rocker/rocker_ofdpa.c 		hash_del(&entry->entry);
entry            1244 drivers/net/ethernet/rocker/rocker_ofdpa.c 		kfree(entry);
entry            1248 drivers/net/ethernet/rocker/rocker_ofdpa.c static void ofdpa_neigh_update(struct ofdpa_neigh_tbl_entry *entry,
entry            1252 drivers/net/ethernet/rocker/rocker_ofdpa.c 		ether_addr_copy(entry->eth_dst, eth_dst);
entry            1253 drivers/net/ethernet/rocker/rocker_ofdpa.c 		entry->ttl_check = ttl_check;
entry            1255 drivers/net/ethernet/rocker/rocker_ofdpa.c 		entry->ref_count++;
entry            1263 drivers/net/ethernet/rocker/rocker_ofdpa.c 	struct ofdpa_neigh_tbl_entry *entry;
entry            1276 drivers/net/ethernet/rocker/rocker_ofdpa.c 	entry = kzalloc(sizeof(*entry), GFP_KERNEL);
entry            1277 drivers/net/ethernet/rocker/rocker_ofdpa.c 	if (!entry)
entry            1289 drivers/net/ethernet/rocker/rocker_ofdpa.c 		entry->ip_addr = ip_addr;
entry            1290 drivers/net/ethernet/rocker/rocker_ofdpa.c 		entry->dev = ofdpa_port->dev;
entry            1291 drivers/net/ethernet/rocker/rocker_ofdpa.c 		ether_addr_copy(entry->eth_dst, eth_dst);
entry            1292 drivers/net/ethernet/rocker/rocker_ofdpa.c 		entry->ttl_check = true;
entry            1293 drivers/net/ethernet/rocker/rocker_ofdpa.c 		ofdpa_neigh_add(ofdpa, entry);
entry            1295 drivers/net/ethernet/rocker/rocker_ofdpa.c 		memcpy(entry, found, sizeof(*entry));
entry            1299 drivers/net/ethernet/rocker/rocker_ofdpa.c 		memcpy(entry, found, sizeof(*entry));
entry            1316 drivers/net/ethernet/rocker/rocker_ofdpa.c 				     entry->index,
entry            1318 drivers/net/ethernet/rocker/rocker_ofdpa.c 				     entry->eth_dst,
entry            1320 drivers/net/ethernet/rocker/rocker_ofdpa.c 				     entry->ttl_check,
entry            1324 drivers/net/ethernet/rocker/rocker_ofdpa.c 			   err, entry->index);
entry            1329 drivers/net/ethernet/rocker/rocker_ofdpa.c 		group_id = ROCKER_GROUP_L3_UNICAST(entry->index);
entry            1338 drivers/net/ethernet/rocker/rocker_ofdpa.c 				   err, &entry->ip_addr, group_id);
entry            1343 drivers/net/ethernet/rocker/rocker_ofdpa.c 		kfree(entry);
entry            1380 drivers/net/ethernet/rocker/rocker_ofdpa.c 	struct ofdpa_neigh_tbl_entry *entry;
entry            1389 drivers/net/ethernet/rocker/rocker_ofdpa.c 	entry = kzalloc(sizeof(*entry), GFP_KERNEL);
entry            1390 drivers/net/ethernet/rocker/rocker_ofdpa.c 	if (!entry)
entry            1402 drivers/net/ethernet/rocker/rocker_ofdpa.c 		entry->ip_addr = ip_addr;
entry            1403 drivers/net/ethernet/rocker/rocker_ofdpa.c 		entry->dev = ofdpa_port->dev;
entry            1404 drivers/net/ethernet/rocker/rocker_ofdpa.c 		ofdpa_neigh_add(ofdpa, entry);
entry            1405 drivers/net/ethernet/rocker/rocker_ofdpa.c 		*index = entry->index;
entry            1421 drivers/net/ethernet/rocker/rocker_ofdpa.c 		kfree(entry);
entry            1889 drivers/net/ethernet/rocker/rocker_ofdpa.c 	hash_for_each_possible(ofdpa->fdb_tbl, found, entry, match->key_crc32)
entry            1925 drivers/net/ethernet/rocker/rocker_ofdpa.c 			hash_del(&found->entry);
entry            1928 drivers/net/ethernet/rocker/rocker_ofdpa.c 		hash_add(ofdpa->fdb_tbl, &fdb->entry,
entry            1963 drivers/net/ethernet/rocker/rocker_ofdpa.c 	hash_for_each_safe(ofdpa->fdb_tbl, bkt, tmp, found, entry) {
entry            1973 drivers/net/ethernet/rocker/rocker_ofdpa.c 		hash_del(&found->entry);
entry            1986 drivers/net/ethernet/rocker/rocker_ofdpa.c 	struct ofdpa_fdb_tbl_entry *entry;
entry            1997 drivers/net/ethernet/rocker/rocker_ofdpa.c 	hash_for_each_safe(ofdpa->fdb_tbl, bkt, tmp, entry, entry) {
entry            1998 drivers/net/ethernet/rocker/rocker_ofdpa.c 		if (!entry->learned)
entry            2000 drivers/net/ethernet/rocker/rocker_ofdpa.c 		ofdpa_port = entry->key.ofdpa_port;
entry            2001 drivers/net/ethernet/rocker/rocker_ofdpa.c 		expires = entry->touched + ofdpa_port->ageing_time;
entry            2004 drivers/net/ethernet/rocker/rocker_ofdpa.c 					     entry->key.addr,
entry            2005 drivers/net/ethernet/rocker/rocker_ofdpa.c 					     entry->key.vlan_id);
entry            2006 drivers/net/ethernet/rocker/rocker_ofdpa.c 			hash_del(&entry->entry);
entry            2217 drivers/net/ethernet/rocker/rocker_ofdpa.c 			       entry, ifindex) {
entry            2229 drivers/net/ethernet/rocker/rocker_ofdpa.c 	struct ofdpa_internal_vlan_tbl_entry *entry;
entry            2234 drivers/net/ethernet/rocker/rocker_ofdpa.c 	entry = kzalloc(sizeof(*entry), GFP_KERNEL);
entry            2235 drivers/net/ethernet/rocker/rocker_ofdpa.c 	if (!entry)
entry            2238 drivers/net/ethernet/rocker/rocker_ofdpa.c 	entry->ifindex = ifindex;
entry            2244 drivers/net/ethernet/rocker/rocker_ofdpa.c 		kfree(entry);
entry            2248 drivers/net/ethernet/rocker/rocker_ofdpa.c 	found = entry;
entry            2249 drivers/net/ethernet/rocker/rocker_ofdpa.c 	hash_add(ofdpa->internal_vlan_tbl, &found->entry, found->ifindex);
entry            2334 drivers/net/ethernet/rocker/rocker_ofdpa.c 		hash_del(&found->entry);
entry            2392 drivers/net/ethernet/rocker/rocker_ofdpa.c 	hash_for_each_safe(ofdpa->flow_tbl, bkt, tmp, flow_entry, entry)
entry            2393 drivers/net/ethernet/rocker/rocker_ofdpa.c 		hash_del(&flow_entry->entry);
entry            2397 drivers/net/ethernet/rocker/rocker_ofdpa.c 	hash_for_each_safe(ofdpa->group_tbl, bkt, tmp, group_entry, entry)
entry            2398 drivers/net/ethernet/rocker/rocker_ofdpa.c 		hash_del(&group_entry->entry);
entry            2402 drivers/net/ethernet/rocker/rocker_ofdpa.c 	hash_for_each_safe(ofdpa->fdb_tbl, bkt, tmp, fdb_entry, entry)
entry            2403 drivers/net/ethernet/rocker/rocker_ofdpa.c 		hash_del(&fdb_entry->entry);
entry            2408 drivers/net/ethernet/rocker/rocker_ofdpa.c 			   tmp, internal_vlan_entry, entry)
entry            2409 drivers/net/ethernet/rocker/rocker_ofdpa.c 		hash_del(&internal_vlan_entry->entry);
entry            2413 drivers/net/ethernet/rocker/rocker_ofdpa.c 	hash_for_each_safe(ofdpa->neigh_tbl, bkt, tmp, neigh_entry, entry)
entry            2414 drivers/net/ethernet/rocker/rocker_ofdpa.c 		hash_del(&neigh_entry->entry);
entry            2787 drivers/net/ethernet/rocker/rocker_ofdpa.c 	hash_for_each_safe(ofdpa->flow_tbl, bkt, tmp, flow_entry, entry) {
entry             743 drivers/net/ethernet/samsung/sxgbe/sxgbe_main.c 		unsigned int entry = tqueue->dirty_tx % tx_rsize;
entry             744 drivers/net/ethernet/samsung/sxgbe/sxgbe_main.c 		struct sk_buff *skb = tqueue->tx_skbuff[entry];
entry             747 drivers/net/ethernet/samsung/sxgbe/sxgbe_main.c 		p = tqueue->dma_tx + entry;
entry             757 drivers/net/ethernet/samsung/sxgbe/sxgbe_main.c 		if (likely(tqueue->tx_skbuff_dma[entry])) {
entry             759 drivers/net/ethernet/samsung/sxgbe/sxgbe_main.c 					 tqueue->tx_skbuff_dma[entry],
entry             762 drivers/net/ethernet/samsung/sxgbe/sxgbe_main.c 			tqueue->tx_skbuff_dma[entry] = 0;
entry             767 drivers/net/ethernet/samsung/sxgbe/sxgbe_main.c 			tqueue->tx_skbuff[entry] = NULL;
entry            1268 drivers/net/ethernet/samsung/sxgbe/sxgbe_main.c 	unsigned int entry, frag_num;
entry            1306 drivers/net/ethernet/samsung/sxgbe/sxgbe_main.c 	entry = tqueue->cur_tx % tx_rsize;
entry            1307 drivers/net/ethernet/samsung/sxgbe/sxgbe_main.c 	tx_desc = tqueue->dma_tx + entry;
entry            1314 drivers/net/ethernet/samsung/sxgbe/sxgbe_main.c 	tqueue->tx_skbuff[entry] = skb;
entry            1331 drivers/net/ethernet/samsung/sxgbe/sxgbe_main.c 				entry = (++tqueue->cur_tx) % tx_rsize;
entry            1332 drivers/net/ethernet/samsung/sxgbe/sxgbe_main.c 				first_desc = tqueue->dma_tx + entry;
entry            1353 drivers/net/ethernet/samsung/sxgbe/sxgbe_main.c 		entry = (++tqueue->cur_tx) % tx_rsize;
entry            1354 drivers/net/ethernet/samsung/sxgbe/sxgbe_main.c 		tx_desc = tqueue->dma_tx + entry;
entry            1358 drivers/net/ethernet/samsung/sxgbe/sxgbe_main.c 		tqueue->tx_skbuff_dma[entry] = tx_desc->tdes01;
entry            1359 drivers/net/ethernet/samsung/sxgbe/sxgbe_main.c 		tqueue->tx_skbuff[entry] = NULL;
entry            1398 drivers/net/ethernet/samsung/sxgbe/sxgbe_main.c 		  tqueue->dirty_tx % tx_rsize, entry,
entry            1437 drivers/net/ethernet/samsung/sxgbe/sxgbe_main.c 		unsigned int entry = priv->rxq[qnum]->dirty_rx % rxsize;
entry            1440 drivers/net/ethernet/samsung/sxgbe/sxgbe_main.c 		p = priv->rxq[qnum]->dma_rx + entry;
entry            1442 drivers/net/ethernet/samsung/sxgbe/sxgbe_main.c 		if (likely(priv->rxq[qnum]->rx_skbuff[entry] == NULL)) {
entry            1450 drivers/net/ethernet/samsung/sxgbe/sxgbe_main.c 			priv->rxq[qnum]->rx_skbuff[entry] = skb;
entry            1451 drivers/net/ethernet/samsung/sxgbe/sxgbe_main.c 			priv->rxq[qnum]->rx_skbuff_dma[entry] =
entry            1456 drivers/net/ethernet/samsung/sxgbe/sxgbe_main.c 				priv->rxq[qnum]->rx_skbuff_dma[entry];
entry            1479 drivers/net/ethernet/samsung/sxgbe/sxgbe_main.c 	unsigned int entry = priv->rxq[qnum]->cur_rx;
entry            1490 drivers/net/ethernet/samsung/sxgbe/sxgbe_main.c 		p = priv->rxq[qnum]->dma_rx + entry;
entry            1507 drivers/net/ethernet/samsung/sxgbe/sxgbe_main.c 			entry = next_entry;
entry            1513 drivers/net/ethernet/samsung/sxgbe/sxgbe_main.c 		skb = priv->rxq[qnum]->rx_skbuff[entry];
entry            1519 drivers/net/ethernet/samsung/sxgbe/sxgbe_main.c 		priv->rxq[qnum]->rx_skbuff[entry] = NULL;
entry            1531 drivers/net/ethernet/samsung/sxgbe/sxgbe_main.c 		entry = next_entry;
entry             589 drivers/net/ethernet/seeq/sgiseeq.c 	int len, entry;
entry             604 drivers/net/ethernet/seeq/sgiseeq.c 	entry = sp->tx_new;
entry             605 drivers/net/ethernet/seeq/sgiseeq.c 	td = &sp->tx_desc[entry];
entry             104 drivers/net/ethernet/sfc/ef10.c 	} *entry;
entry            4028 drivers/net/ethernet/sfc/ef10.c 	return (struct efx_filter_spec *)(table->entry[filter_idx].spec &
entry            4036 drivers/net/ethernet/sfc/ef10.c 	return table->entry[filter_idx].spec & EFX_EF10_FILTER_FLAGS;
entry            4045 drivers/net/ethernet/sfc/ef10.c 	table->entry[filter_idx].spec =	(unsigned long)spec | flags;
entry            4400 drivers/net/ethernet/sfc/ef10.c 			table->entry[ins_index].spec &=
entry            4419 drivers/net/ethernet/sfc/ef10.c 	rc = efx_ef10_filter_push(efx, spec, &table->entry[ins_index].handle,
entry            4474 drivers/net/ethernet/sfc/ef10.c 					       table->entry[i].handle);
entry            4550 drivers/net/ethernet/sfc/ef10.c 		table->entry[filter_idx].spec &= ~EFX_EF10_FILTER_FLAG_AUTO_OLD;
entry            4569 drivers/net/ethernet/sfc/ef10.c 					  &table->entry[filter_idx].handle,
entry            4583 drivers/net/ethernet/sfc/ef10.c 			       table->entry[filter_idx].handle);
entry            4698 drivers/net/ethernet/sfc/ef10.c 		if (table->entry[filter_idx].spec &&
entry            4996 drivers/net/ethernet/sfc/ef10.c 	table->entry = vzalloc(array_size(HUNT_FILTER_TBL_ROWS,
entry            4997 drivers/net/ethernet/sfc/ef10.c 					  sizeof(*table->entry)));
entry            4998 drivers/net/ethernet/sfc/ef10.c 	if (!table->entry) {
entry            5090 drivers/net/ethernet/sfc/ef10.c 					  &table->entry[filter_idx].handle,
entry            5158 drivers/net/ethernet/sfc/ef10.c 			       table->entry[filter_idx].handle);
entry            5168 drivers/net/ethernet/sfc/ef10.c 	vfree(table->entry);
entry            5181 drivers/net/ethernet/sfc/ef10.c 		if (!table->entry[filter_idx].spec)
entry            5185 drivers/net/ethernet/sfc/ef10.c 		table->entry[filter_idx].spec |= EFX_EF10_FILTER_FLAG_AUTO_OLD;
entry            5505 drivers/net/ethernet/sfc/ef10.c 		if (READ_ONCE(table->entry[i].spec) &
entry            6364 drivers/net/ethernet/sfc/ef10.c 			efx_dword_t entry;
entry            6366 drivers/net/ethernet/sfc/ef10.c 			EFX_POPULATE_DWORD_2(entry,
entry            6373 drivers/net/ethernet/sfc/ef10.c 				num_entries++) = entry;
entry            1463 drivers/net/ethernet/sfc/efx.c 			xentries[i].entry = i;
entry            3554 drivers/net/ethernet/sfc/efx.c 			 const struct pci_device_id *entry)
entry            3566 drivers/net/ethernet/sfc/efx.c 	efx->type = (const struct efx_nic_type *) entry->driver_data;
entry            1385 drivers/net/ethernet/sfc/falcon/efx.c 			xentries[i].entry = i;
entry            2915 drivers/net/ethernet/sfc/falcon/efx.c 			 const struct pci_device_id *entry)
entry            2927 drivers/net/ethernet/sfc/falcon/efx.c 	efx->type = (const struct ef4_nic_type *) entry->driver_data;
entry             162 drivers/net/ethernet/sfc/mcdi_mon.c 				  efx_dword_t *entry)
entry             179 drivers/net/ethernet/sfc/mcdi_mon.c 	*entry = ((efx_dword_t *)hwmon->dma_buf.addr)[index];
entry             192 drivers/net/ethernet/sfc/mcdi_mon.c 	efx_dword_t entry;
entry             196 drivers/net/ethernet/sfc/mcdi_mon.c 	rc = efx_mcdi_mon_get_entry(dev, mon_attr->index, &entry);
entry             200 drivers/net/ethernet/sfc/mcdi_mon.c 	state = EFX_DWORD_FIELD(entry, MC_CMD_SENSOR_VALUE_ENTRY_TYPEDEF_STATE);
entry             204 drivers/net/ethernet/sfc/mcdi_mon.c 	value = EFX_DWORD_FIELD(entry, MC_CMD_SENSOR_VALUE_ENTRY_TYPEDEF_VALUE);
entry             256 drivers/net/ethernet/sfc/mcdi_mon.c 	efx_dword_t entry;
entry             260 drivers/net/ethernet/sfc/mcdi_mon.c 	rc = efx_mcdi_mon_get_entry(dev, mon_attr->index, &entry);
entry             264 drivers/net/ethernet/sfc/mcdi_mon.c 	state = EFX_DWORD_FIELD(entry, MC_CMD_SENSOR_VALUE_ENTRY_TYPEDEF_STATE);
entry             802 drivers/net/ethernet/sgi/ioc3-eth.c static inline void ioc3_tx_unmap(struct ioc3_private *ip, int entry)
entry             807 drivers/net/ethernet/sgi/ioc3-eth.c 	desc = &ip->txr[entry];
entry             652 drivers/net/ethernet/silan/sc92031.c 	unsigned entry;
entry             657 drivers/net/ethernet/silan/sc92031.c 		entry = priv->tx_tail % NUM_TX_DESC;
entry             658 drivers/net/ethernet/silan/sc92031.c 		tx_status = ioread32(port_base + TxStatus0 + entry * 4);
entry             937 drivers/net/ethernet/silan/sc92031.c 	unsigned entry;
entry             954 drivers/net/ethernet/silan/sc92031.c 	entry = priv->tx_head++ % NUM_TX_DESC;
entry             956 drivers/net/ethernet/silan/sc92031.c 	skb_copy_and_csum_dev(skb, priv->tx_bufs + entry * TX_BUF_SIZE);
entry             960 drivers/net/ethernet/silan/sc92031.c 		memset(priv->tx_bufs + entry * TX_BUF_SIZE + len,
entry             974 drivers/net/ethernet/silan/sc92031.c 	iowrite32(priv->tx_bufs_dma_addr + entry * TX_BUF_SIZE,
entry             975 drivers/net/ethernet/silan/sc92031.c 			port_base + TxAddr0 + entry * 4);
entry             976 drivers/net/ethernet/silan/sc92031.c 	iowrite32(tx_status, port_base + TxStatus0 + entry * 4);
entry             585 drivers/net/ethernet/sis/sis190.c 		unsigned int entry = cur_rx % NUM_RX_DESC;
entry             586 drivers/net/ethernet/sis/sis190.c 		struct RxDesc *desc = tp->RxDescRing + entry;
entry             599 drivers/net/ethernet/sis/sis190.c 			struct sk_buff *skb = tp->Rx_skbuff[entry];
entry             621 drivers/net/ethernet/sis/sis190.c 				tp->Rx_skbuff[entry] = NULL;
entry             699 drivers/net/ethernet/sis/sis190.c 		unsigned int entry = dirty_tx % NUM_TX_DESC;
entry             700 drivers/net/ethernet/sis/sis190.c 		struct TxDesc *txd = tp->TxDescRing + entry;
entry             707 drivers/net/ethernet/sis/sis190.c 		skb = tp->Tx_skbuff[entry];
entry             716 drivers/net/ethernet/sis/sis190.c 		tp->Tx_skbuff[entry] = NULL;
entry            1176 drivers/net/ethernet/sis/sis190.c 	u32 len, entry, dirty_tx;
entry            1190 drivers/net/ethernet/sis/sis190.c 	entry = tp->cur_tx % NUM_TX_DESC;
entry            1191 drivers/net/ethernet/sis/sis190.c 	desc = tp->TxDescRing + entry;
entry            1207 drivers/net/ethernet/sis/sis190.c 	tp->Tx_skbuff[entry] = skb;
entry            1213 drivers/net/ethernet/sis/sis190.c 	if (entry == (NUM_TX_DESC - 1))
entry            1603 drivers/net/ethernet/sis/sis900.c 	unsigned int  entry;
entry            1611 drivers/net/ethernet/sis/sis900.c 	entry = sis_priv->cur_tx % NUM_TX_DESC;
entry            1612 drivers/net/ethernet/sis/sis900.c 	sis_priv->tx_skbuff[entry] = skb;
entry            1615 drivers/net/ethernet/sis/sis900.c 	sis_priv->tx_ring[entry].bufptr = pci_map_single(sis_priv->pci_dev,
entry            1618 drivers/net/ethernet/sis/sis900.c 		sis_priv->tx_ring[entry].bufptr))) {
entry            1620 drivers/net/ethernet/sis/sis900.c 			sis_priv->tx_skbuff[entry] = NULL;
entry            1625 drivers/net/ethernet/sis/sis900.c 	sis_priv->tx_ring[entry].cmdsts = (OWN | INTR | skb->len);
entry            1653 drivers/net/ethernet/sis/sis900.c 		       net_dev->name, skb->data, (int)skb->len, entry);
entry            1734 drivers/net/ethernet/sis/sis900.c 	unsigned int entry = sis_priv->cur_rx % NUM_RX_DESC;
entry            1735 drivers/net/ethernet/sis/sis900.c 	u32 rx_status = sis_priv->rx_ring[entry].cmdsts;
entry            1776 drivers/net/ethernet/sis/sis900.c 			sis_priv->rx_ring[entry].cmdsts = RX_BUF_SIZE;
entry            1782 drivers/net/ethernet/sis/sis900.c 				sis_priv->rx_ring[entry].bufptr, RX_BUF_SIZE,
entry            1794 drivers/net/ethernet/sis/sis900.c 				skb = sis_priv->rx_skbuff[entry];
entry            1802 drivers/net/ethernet/sis/sis900.c 			if (sis_priv->rx_skbuff[entry] == NULL) {
entry            1814 drivers/net/ethernet/sis/sis900.c 			rx_skb = sis_priv->rx_skbuff[entry];
entry            1826 drivers/net/ethernet/sis/sis900.c 			sis_priv->rx_skbuff[entry] = skb;
entry            1827 drivers/net/ethernet/sis/sis900.c 			sis_priv->rx_ring[entry].cmdsts = RX_BUF_SIZE;
entry            1828 drivers/net/ethernet/sis/sis900.c 			sis_priv->rx_ring[entry].bufptr =
entry            1832 drivers/net/ethernet/sis/sis900.c 				sis_priv->rx_ring[entry].bufptr))) {
entry            1834 drivers/net/ethernet/sis/sis900.c 				sis_priv->rx_skbuff[entry] = NULL;
entry            1839 drivers/net/ethernet/sis/sis900.c 		entry = sis_priv->cur_rx % NUM_RX_DESC;
entry            1840 drivers/net/ethernet/sis/sis900.c 		rx_status = sis_priv->rx_ring[entry].cmdsts;
entry            1848 drivers/net/ethernet/sis/sis900.c 		entry = sis_priv->dirty_rx % NUM_RX_DESC;
entry            1850 drivers/net/ethernet/sis/sis900.c 		if (sis_priv->rx_skbuff[entry] == NULL) {
entry            1860 drivers/net/ethernet/sis/sis900.c 			sis_priv->rx_skbuff[entry] = skb;
entry            1861 drivers/net/ethernet/sis/sis900.c 			sis_priv->rx_ring[entry].cmdsts = RX_BUF_SIZE;
entry            1862 drivers/net/ethernet/sis/sis900.c 			sis_priv->rx_ring[entry].bufptr =
entry            1866 drivers/net/ethernet/sis/sis900.c 					sis_priv->rx_ring[entry].bufptr))) {
entry            1868 drivers/net/ethernet/sis/sis900.c 				sis_priv->rx_skbuff[entry] = NULL;
entry            1895 drivers/net/ethernet/sis/sis900.c 		unsigned int entry;
entry            1898 drivers/net/ethernet/sis/sis900.c 		entry = sis_priv->dirty_tx % NUM_TX_DESC;
entry            1899 drivers/net/ethernet/sis/sis900.c 		tx_status = sis_priv->tx_ring[entry].cmdsts;
entry            1930 drivers/net/ethernet/sis/sis900.c 		skb = sis_priv->tx_skbuff[entry];
entry            1932 drivers/net/ethernet/sis/sis900.c 			sis_priv->tx_ring[entry].bufptr, skb->len,
entry            1935 drivers/net/ethernet/sis/sis900.c 		sis_priv->tx_skbuff[entry] = NULL;
entry            1936 drivers/net/ethernet/sis/sis900.c 		sis_priv->tx_ring[entry].bufptr = 0;
entry            1937 drivers/net/ethernet/sis/sis900.c 		sis_priv->tx_ring[entry].cmdsts = 0;
entry             941 drivers/net/ethernet/smsc/epic100.c 	int entry, free_count;
entry             954 drivers/net/ethernet/smsc/epic100.c 	entry = ep->cur_tx % TX_RING_SIZE;
entry             956 drivers/net/ethernet/smsc/epic100.c 	ep->tx_skbuff[entry] = skb;
entry             957 drivers/net/ethernet/smsc/epic100.c 	ep->tx_ring[entry].bufaddr = pci_map_single(ep->pci_dev, skb->data,
entry             970 drivers/net/ethernet/smsc/epic100.c 	ep->tx_ring[entry].buflength = ctrl_word | skb->len;
entry             971 drivers/net/ethernet/smsc/epic100.c 	ep->tx_ring[entry].txstatus =
entry             985 drivers/net/ethernet/smsc/epic100.c 			   skb->len, entry, ctrl_word, er32(TxSTAT));
entry            1023 drivers/net/ethernet/smsc/epic100.c 		int entry = dirty_tx % TX_RING_SIZE;
entry            1024 drivers/net/ethernet/smsc/epic100.c 		int txstatus = ep->tx_ring[entry].txstatus;
entry            1032 drivers/net/ethernet/smsc/epic100.c 			dev->stats.tx_bytes += ep->tx_skbuff[entry]->len;
entry            1037 drivers/net/ethernet/smsc/epic100.c 		skb = ep->tx_skbuff[entry];
entry            1038 drivers/net/ethernet/smsc/epic100.c 		pci_unmap_single(ep->pci_dev, ep->tx_ring[entry].bufaddr,
entry            1041 drivers/net/ethernet/smsc/epic100.c 		ep->tx_skbuff[entry] = NULL;
entry            1133 drivers/net/ethernet/smsc/epic100.c 	int entry = ep->cur_rx % RX_RING_SIZE;
entry            1138 drivers/net/ethernet/smsc/epic100.c 		netdev_dbg(dev, " In epic_rx(), entry %d %8.8x.\n", entry,
entry            1139 drivers/net/ethernet/smsc/epic100.c 			   ep->rx_ring[entry].rxstatus);
entry            1145 drivers/net/ethernet/smsc/epic100.c 	while ((ep->rx_ring[entry].rxstatus & DescOwn) == 0) {
entry            1146 drivers/net/ethernet/smsc/epic100.c 		int status = ep->rx_ring[entry].rxstatus;
entry            1181 drivers/net/ethernet/smsc/epic100.c 							    ep->rx_ring[entry].bufaddr,
entry            1184 drivers/net/ethernet/smsc/epic100.c 				skb_copy_to_linear_data(skb, ep->rx_skbuff[entry]->data, pkt_len);
entry            1187 drivers/net/ethernet/smsc/epic100.c 							       ep->rx_ring[entry].bufaddr,
entry            1192 drivers/net/ethernet/smsc/epic100.c 					ep->rx_ring[entry].bufaddr,
entry            1194 drivers/net/ethernet/smsc/epic100.c 				skb_put(skb = ep->rx_skbuff[entry], pkt_len);
entry            1195 drivers/net/ethernet/smsc/epic100.c 				ep->rx_skbuff[entry] = NULL;
entry            1203 drivers/net/ethernet/smsc/epic100.c 		entry = (++ep->cur_rx) % RX_RING_SIZE;
entry            1208 drivers/net/ethernet/smsc/epic100.c 		entry = ep->dirty_rx % RX_RING_SIZE;
entry            1209 drivers/net/ethernet/smsc/epic100.c 		if (ep->rx_skbuff[entry] == NULL) {
entry            1211 drivers/net/ethernet/smsc/epic100.c 			skb = ep->rx_skbuff[entry] = netdev_alloc_skb(dev, ep->rx_buf_sz + 2);
entry            1215 drivers/net/ethernet/smsc/epic100.c 			ep->rx_ring[entry].bufaddr = pci_map_single(ep->pci_dev,
entry            1220 drivers/net/ethernet/smsc/epic100.c 		ep->rx_ring[entry].rxstatus = DescOwn;
entry             631 drivers/net/ethernet/socionext/netsec.c 	struct netsec_de *entry;
entry             639 drivers/net/ethernet/socionext/netsec.c 	entry = dring->vaddr + DESC_SZ * tail;
entry             641 drivers/net/ethernet/socionext/netsec.c 	while (!(entry->attr & (1U << NETSEC_TX_SHIFT_OWN_FIELD)) &&
entry             647 drivers/net/ethernet/socionext/netsec.c 		eop = (entry->attr >> NETSEC_TX_LAST) & 1;
entry             675 drivers/net/ethernet/socionext/netsec.c 		entry->attr = 1U << NETSEC_TX_SHIFT_OWN_FIELD;
entry             680 drivers/net/ethernet/socionext/netsec.c 		entry = dring->vaddr + DESC_SZ * tail;
entry             296 drivers/net/ethernet/socionext/sni_ave.c static u32 ave_desc_read(struct net_device *ndev, enum desc_id id, int entry,
entry             303 drivers/net/ethernet/socionext/sni_ave.c 		+ entry * priv->desc_size + offset;
entry             309 drivers/net/ethernet/socionext/sni_ave.c 				int entry)
entry             311 drivers/net/ethernet/socionext/sni_ave.c 	return ave_desc_read(ndev, id, entry, AVE_DESC_OFS_CMDSTS);
entry             315 drivers/net/ethernet/socionext/sni_ave.c 			   int entry, int offset, u32 val)
entry             321 drivers/net/ethernet/socionext/sni_ave.c 		+ entry * priv->desc_size + offset;
entry             327 drivers/net/ethernet/socionext/sni_ave.c 				  int entry, u32 val)
entry             329 drivers/net/ethernet/socionext/sni_ave.c 	ave_desc_write(ndev, id, entry, AVE_DESC_OFS_CMDSTS, val);
entry             333 drivers/net/ethernet/socionext/sni_ave.c 				int entry, dma_addr_t paddr)
entry             337 drivers/net/ethernet/socionext/sni_ave.c 	ave_desc_write(ndev, id, entry, AVE_DESC_OFS_ADDRL,
entry             341 drivers/net/ethernet/socionext/sni_ave.c 			       entry, AVE_DESC_OFS_ADDRU,
entry             578 drivers/net/ethernet/socionext/sni_ave.c static int ave_rxdesc_prepare(struct net_device *ndev, int entry)
entry             585 drivers/net/ethernet/socionext/sni_ave.c 	skb = priv->rx.desc[entry].skbs;
entry             597 drivers/net/ethernet/socionext/sni_ave.c 	ave_desc_write_cmdsts(ndev, AVE_DESCID_RX, entry,
entry             609 drivers/net/ethernet/socionext/sni_ave.c 	ret = ave_dma_map(ndev, &priv->rx.desc[entry],
entry             618 drivers/net/ethernet/socionext/sni_ave.c 	priv->rx.desc[entry].skbs = skb;
entry             621 drivers/net/ethernet/socionext/sni_ave.c 	ave_desc_write_addr(ndev, AVE_DESCID_RX, entry, paddr);
entry             624 drivers/net/ethernet/socionext/sni_ave.c 	ave_desc_write_cmdsts(ndev, AVE_DESCID_RX, entry,
entry             981 drivers/net/ethernet/socionext/sni_ave.c static int ave_pfsel_start(struct net_device *ndev, unsigned int entry)
entry             986 drivers/net/ethernet/socionext/sni_ave.c 	if (WARN_ON(entry > AVE_PF_SIZE))
entry             990 drivers/net/ethernet/socionext/sni_ave.c 	writel(val | BIT(entry), priv->base + AVE_PFEN);
entry             995 drivers/net/ethernet/socionext/sni_ave.c static int ave_pfsel_stop(struct net_device *ndev, unsigned int entry)
entry            1000 drivers/net/ethernet/socionext/sni_ave.c 	if (WARN_ON(entry > AVE_PF_SIZE))
entry            1004 drivers/net/ethernet/socionext/sni_ave.c 	writel(val & ~BIT(entry), priv->base + AVE_PFEN);
entry            1010 drivers/net/ethernet/socionext/sni_ave.c 				 unsigned int entry,
entry            1016 drivers/net/ethernet/socionext/sni_ave.c 	if (WARN_ON(entry > AVE_PF_SIZE))
entry            1021 drivers/net/ethernet/socionext/sni_ave.c 	ave_pfsel_stop(ndev, entry);
entry            1025 drivers/net/ethernet/socionext/sni_ave.c 			     AVE_PKTF(entry), AVE_PKTF(entry) + 4);
entry            1029 drivers/net/ethernet/socionext/sni_ave.c 	       priv->base + AVE_PFMBYTE(entry));
entry            1030 drivers/net/ethernet/socionext/sni_ave.c 	writel(AVE_PFMBYTE_MASK1, priv->base + AVE_PFMBYTE(entry) + 4);
entry            1033 drivers/net/ethernet/socionext/sni_ave.c 	writel(AVE_PFMBIT_MASK, priv->base + AVE_PFMBIT(entry));
entry            1036 drivers/net/ethernet/socionext/sni_ave.c 	writel(0, priv->base + AVE_PFSEL(entry));
entry            1039 drivers/net/ethernet/socionext/sni_ave.c 	ave_pfsel_start(ndev, entry);
entry            1045 drivers/net/ethernet/socionext/sni_ave.c 				  unsigned int entry, u32 rxring)
entry            1049 drivers/net/ethernet/socionext/sni_ave.c 	if (WARN_ON(entry > AVE_PF_SIZE))
entry            1052 drivers/net/ethernet/socionext/sni_ave.c 	ave_pfsel_stop(ndev, entry);
entry            1055 drivers/net/ethernet/socionext/sni_ave.c 	writel(AVE_PFMBYTE_MASK0, priv->base + AVE_PFMBYTE(entry));
entry            1056 drivers/net/ethernet/socionext/sni_ave.c 	writel(AVE_PFMBYTE_MASK1, priv->base + AVE_PFMBYTE(entry) + 4);
entry            1059 drivers/net/ethernet/socionext/sni_ave.c 	writel(AVE_PFMBIT_MASK, priv->base + AVE_PFMBIT(entry));
entry            1062 drivers/net/ethernet/socionext/sni_ave.c 	writel(rxring, priv->base + AVE_PFSEL(entry));
entry            1064 drivers/net/ethernet/socionext/sni_ave.c 	ave_pfsel_start(ndev, entry);
entry            1266 drivers/net/ethernet/socionext/sni_ave.c 	int entry;
entry            1293 drivers/net/ethernet/socionext/sni_ave.c 	for (entry = 0; entry < priv->tx.ndesc; entry++) {
entry            1294 drivers/net/ethernet/socionext/sni_ave.c 		ave_desc_write_cmdsts(ndev, AVE_DESCID_TX, entry, 0);
entry            1295 drivers/net/ethernet/socionext/sni_ave.c 		ave_desc_write_addr(ndev, AVE_DESCID_TX, entry, 0);
entry            1304 drivers/net/ethernet/socionext/sni_ave.c 	for (entry = 0; entry < priv->rx.ndesc; entry++) {
entry            1305 drivers/net/ethernet/socionext/sni_ave.c 		if (ave_rxdesc_prepare(ndev, entry))
entry            1354 drivers/net/ethernet/socionext/sni_ave.c 	int entry;
entry            1368 drivers/net/ethernet/socionext/sni_ave.c 	for (entry = 0; entry < priv->tx.ndesc; entry++) {
entry            1369 drivers/net/ethernet/socionext/sni_ave.c 		if (!priv->tx.desc[entry].skbs)
entry            1372 drivers/net/ethernet/socionext/sni_ave.c 		ave_dma_unmap(ndev, &priv->tx.desc[entry], DMA_TO_DEVICE);
entry            1373 drivers/net/ethernet/socionext/sni_ave.c 		dev_kfree_skb_any(priv->tx.desc[entry].skbs);
entry            1374 drivers/net/ethernet/socionext/sni_ave.c 		priv->tx.desc[entry].skbs = NULL;
entry            1380 drivers/net/ethernet/socionext/sni_ave.c 	for (entry = 0; entry < priv->rx.ndesc; entry++) {
entry            1381 drivers/net/ethernet/socionext/sni_ave.c 		if (!priv->rx.desc[entry].skbs)
entry            1384 drivers/net/ethernet/socionext/sni_ave.c 		ave_dma_unmap(ndev, &priv->rx.desc[entry], DMA_FROM_DEVICE);
entry            1385 drivers/net/ethernet/socionext/sni_ave.c 		dev_kfree_skb_any(priv->rx.desc[entry].skbs);
entry            1386 drivers/net/ethernet/socionext/sni_ave.c 		priv->rx.desc[entry].skbs = NULL;
entry              22 drivers/net/ethernet/stmicro/stmmac/chain_mode.c 	unsigned int entry = tx_q->cur_tx;
entry              27 drivers/net/ethernet/stmicro/stmmac/chain_mode.c 	desc = tx_q->dma_tx + entry;
entry              41 drivers/net/ethernet/stmicro/stmmac/chain_mode.c 	tx_q->tx_skbuff_dma[entry].buf = des2;
entry              42 drivers/net/ethernet/stmicro/stmmac/chain_mode.c 	tx_q->tx_skbuff_dma[entry].len = bmax;
entry              48 drivers/net/ethernet/stmicro/stmmac/chain_mode.c 		tx_q->tx_skbuff[entry] = NULL;
entry              49 drivers/net/ethernet/stmicro/stmmac/chain_mode.c 		entry = STMMAC_GET_ENTRY(entry, DMA_TX_SIZE);
entry              50 drivers/net/ethernet/stmicro/stmmac/chain_mode.c 		desc = tx_q->dma_tx + entry;
entry              59 drivers/net/ethernet/stmicro/stmmac/chain_mode.c 			tx_q->tx_skbuff_dma[entry].buf = des2;
entry              60 drivers/net/ethernet/stmicro/stmmac/chain_mode.c 			tx_q->tx_skbuff_dma[entry].len = bmax;
entry              72 drivers/net/ethernet/stmicro/stmmac/chain_mode.c 			tx_q->tx_skbuff_dma[entry].buf = des2;
entry              73 drivers/net/ethernet/stmicro/stmmac/chain_mode.c 			tx_q->tx_skbuff_dma[entry].len = len;
entry              81 drivers/net/ethernet/stmicro/stmmac/chain_mode.c 	tx_q->cur_tx = entry;
entry              83 drivers/net/ethernet/stmicro/stmmac/chain_mode.c 	return entry;
entry             148 drivers/net/ethernet/stmicro/stmmac/chain_mode.c 	unsigned int entry = tx_q->dirty_tx;
entry             150 drivers/net/ethernet/stmicro/stmmac/chain_mode.c 	if (tx_q->tx_skbuff_dma[entry].last_segment && !priv->extend_desc &&
entry             331 drivers/net/ethernet/stmicro/stmmac/dwmac5.c 					  struct stmmac_tc_entry *entry,
entry             336 drivers/net/ethernet/stmicro/stmmac/dwmac5.c 	for (i = 0; i < (sizeof(entry->val) / sizeof(u32)); i++) {
entry             337 drivers/net/ethernet/stmicro/stmmac/dwmac5.c 		int real_pos = pos * (sizeof(entry->val) / sizeof(u32)) + i;
entry             347 drivers/net/ethernet/stmicro/stmmac/dwmac5.c 		val = *((u32 *)&entry->val + i);
entry             376 drivers/net/ethernet/stmicro/stmmac/dwmac5.c 	struct stmmac_tc_entry *entry;
entry             382 drivers/net/ethernet/stmicro/stmmac/dwmac5.c 		entry = &entries[i];
entry             385 drivers/net/ethernet/stmicro/stmmac/dwmac5.c 		if (!entry->in_use)
entry             388 drivers/net/ethernet/stmicro/stmmac/dwmac5.c 		if (entry->in_hw)
entry             391 drivers/net/ethernet/stmicro/stmmac/dwmac5.c 		if (entry->is_last)
entry             394 drivers/net/ethernet/stmicro/stmmac/dwmac5.c 		if (entry->is_frag)
entry             397 drivers/net/ethernet/stmicro/stmmac/dwmac5.c 		if (entry->prio < curr_prio)
entry             400 drivers/net/ethernet/stmicro/stmmac/dwmac5.c 		if (entry->prio < min_prio) {
entry             401 drivers/net/ethernet/stmicro/stmmac/dwmac5.c 			min_prio = entry->prio;
entry             415 drivers/net/ethernet/stmicro/stmmac/dwmac5.c 	struct stmmac_tc_entry *entry, *frag;
entry             432 drivers/net/ethernet/stmicro/stmmac/dwmac5.c 		entry = &entries[i];
entry             433 drivers/net/ethernet/stmicro/stmmac/dwmac5.c 		entry->in_hw = false;
entry             438 drivers/net/ethernet/stmicro/stmmac/dwmac5.c 		entry = dwmac5_rxp_get_next_entry(entries, count, curr_prio);
entry             439 drivers/net/ethernet/stmicro/stmmac/dwmac5.c 		if (!entry)
entry             442 drivers/net/ethernet/stmicro/stmmac/dwmac5.c 		curr_prio = entry->prio;
entry             443 drivers/net/ethernet/stmicro/stmmac/dwmac5.c 		frag = entry->frag_ptr;
entry             447 drivers/net/ethernet/stmicro/stmmac/dwmac5.c 			entry->val.af = 0;
entry             448 drivers/net/ethernet/stmicro/stmmac/dwmac5.c 			entry->val.rf = 0;
entry             449 drivers/net/ethernet/stmicro/stmmac/dwmac5.c 			entry->val.nc = 1;
entry             450 drivers/net/ethernet/stmicro/stmmac/dwmac5.c 			entry->val.ok_index = nve + 2;
entry             453 drivers/net/ethernet/stmicro/stmmac/dwmac5.c 		ret = dwmac5_rxp_update_single_entry(ioaddr, entry, nve);
entry             457 drivers/net/ethernet/stmicro/stmmac/dwmac5.c 		entry->table_pos = nve++;
entry             458 drivers/net/ethernet/stmicro/stmmac/dwmac5.c 		entry->in_hw = true;
entry             474 drivers/net/ethernet/stmicro/stmmac/dwmac5.c 		entry = &entries[i];
entry             475 drivers/net/ethernet/stmicro/stmmac/dwmac5.c 		if (!entry->is_last)
entry             478 drivers/net/ethernet/stmicro/stmmac/dwmac5.c 		ret = dwmac5_rxp_update_single_entry(ioaddr, entry, nve);
entry             482 drivers/net/ethernet/stmicro/stmmac/dwmac5.c 		entry->table_pos = nve++;
entry             896 drivers/net/ethernet/stmicro/stmmac/dwxgmac2_core.c 					    struct stmmac_tc_entry *entry,
entry             901 drivers/net/ethernet/stmicro/stmmac/dwxgmac2_core.c 	for (i = 0; i < (sizeof(entry->val) / sizeof(u32)); i++) {
entry             902 drivers/net/ethernet/stmicro/stmmac/dwxgmac2_core.c 		int real_pos = pos * (sizeof(entry->val) / sizeof(u32)) + i;
entry             912 drivers/net/ethernet/stmicro/stmmac/dwxgmac2_core.c 		val = *((u32 *)&entry->val + i);
entry             941 drivers/net/ethernet/stmicro/stmmac/dwxgmac2_core.c 	struct stmmac_tc_entry *entry;
entry             947 drivers/net/ethernet/stmicro/stmmac/dwxgmac2_core.c 		entry = &entries[i];
entry             950 drivers/net/ethernet/stmicro/stmmac/dwxgmac2_core.c 		if (!entry->in_use)
entry             953 drivers/net/ethernet/stmicro/stmmac/dwxgmac2_core.c 		if (entry->in_hw)
entry             956 drivers/net/ethernet/stmicro/stmmac/dwxgmac2_core.c 		if (entry->is_last)
entry             959 drivers/net/ethernet/stmicro/stmmac/dwxgmac2_core.c 		if (entry->is_frag)
entry             962 drivers/net/ethernet/stmicro/stmmac/dwxgmac2_core.c 		if (entry->prio < curr_prio)
entry             965 drivers/net/ethernet/stmicro/stmmac/dwxgmac2_core.c 		if (entry->prio < min_prio) {
entry             966 drivers/net/ethernet/stmicro/stmmac/dwxgmac2_core.c 			min_prio = entry->prio;
entry             981 drivers/net/ethernet/stmicro/stmmac/dwxgmac2_core.c 	struct stmmac_tc_entry *entry, *frag;
entry             998 drivers/net/ethernet/stmicro/stmmac/dwxgmac2_core.c 		entry = &entries[i];
entry             999 drivers/net/ethernet/stmicro/stmmac/dwxgmac2_core.c 		entry->in_hw = false;
entry            1004 drivers/net/ethernet/stmicro/stmmac/dwxgmac2_core.c 		entry = dwxgmac3_rxp_get_next_entry(entries, count, curr_prio);
entry            1005 drivers/net/ethernet/stmicro/stmmac/dwxgmac2_core.c 		if (!entry)
entry            1008 drivers/net/ethernet/stmicro/stmmac/dwxgmac2_core.c 		curr_prio = entry->prio;
entry            1009 drivers/net/ethernet/stmicro/stmmac/dwxgmac2_core.c 		frag = entry->frag_ptr;
entry            1013 drivers/net/ethernet/stmicro/stmmac/dwxgmac2_core.c 			entry->val.af = 0;
entry            1014 drivers/net/ethernet/stmicro/stmmac/dwxgmac2_core.c 			entry->val.rf = 0;
entry            1015 drivers/net/ethernet/stmicro/stmmac/dwxgmac2_core.c 			entry->val.nc = 1;
entry            1016 drivers/net/ethernet/stmicro/stmmac/dwxgmac2_core.c 			entry->val.ok_index = nve + 2;
entry            1019 drivers/net/ethernet/stmicro/stmmac/dwxgmac2_core.c 		ret = dwxgmac3_rxp_update_single_entry(ioaddr, entry, nve);
entry            1023 drivers/net/ethernet/stmicro/stmmac/dwxgmac2_core.c 		entry->table_pos = nve++;
entry            1024 drivers/net/ethernet/stmicro/stmmac/dwxgmac2_core.c 		entry->in_hw = true;
entry            1040 drivers/net/ethernet/stmicro/stmmac/dwxgmac2_core.c 		entry = &entries[i];
entry            1041 drivers/net/ethernet/stmicro/stmmac/dwxgmac2_core.c 		if (!entry->is_last)
entry            1044 drivers/net/ethernet/stmicro/stmmac/dwxgmac2_core.c 		ret = dwxgmac3_rxp_update_single_entry(ioaddr, entry, nve);
entry            1048 drivers/net/ethernet/stmicro/stmmac/dwxgmac2_core.c 		entry->table_pos = nve++;
entry             223 drivers/net/ethernet/stmicro/stmmac/hwif.c 	const struct stmmac_hwif_entry *entry;
entry             259 drivers/net/ethernet/stmicro/stmmac/hwif.c 		entry = &stmmac_hw[i];
entry             261 drivers/net/ethernet/stmicro/stmmac/hwif.c 		if (needs_gmac ^ entry->gmac)
entry             263 drivers/net/ethernet/stmicro/stmmac/hwif.c 		if (needs_gmac4 ^ entry->gmac4)
entry             265 drivers/net/ethernet/stmicro/stmmac/hwif.c 		if (needs_xgmac ^ entry->xgmac)
entry             268 drivers/net/ethernet/stmicro/stmmac/hwif.c 		if (priv->synopsys_id < entry->min_id)
entry             272 drivers/net/ethernet/stmicro/stmmac/hwif.c 		mac->desc = mac->desc ? : entry->desc;
entry             273 drivers/net/ethernet/stmicro/stmmac/hwif.c 		mac->dma = mac->dma ? : entry->dma;
entry             274 drivers/net/ethernet/stmicro/stmmac/hwif.c 		mac->mac = mac->mac ? : entry->mac;
entry             275 drivers/net/ethernet/stmicro/stmmac/hwif.c 		mac->ptp = mac->ptp ? : entry->hwtimestamp;
entry             276 drivers/net/ethernet/stmicro/stmmac/hwif.c 		mac->mode = mac->mode ? : entry->mode;
entry             277 drivers/net/ethernet/stmicro/stmmac/hwif.c 		mac->tc = mac->tc ? : entry->tc;
entry             278 drivers/net/ethernet/stmicro/stmmac/hwif.c 		mac->mmc = mac->mmc ? : entry->mmc;
entry             281 drivers/net/ethernet/stmicro/stmmac/hwif.c 		priv->ptpaddr = priv->ioaddr + entry->regs.ptp_off;
entry             282 drivers/net/ethernet/stmicro/stmmac/hwif.c 		priv->mmcaddr = priv->ioaddr + entry->regs.mmc_off;
entry             286 drivers/net/ethernet/stmicro/stmmac/hwif.c 			ret = entry->setup(priv);
entry             292 drivers/net/ethernet/stmicro/stmmac/hwif.c 		priv->hwif_quirks = entry->quirks;
entry              22 drivers/net/ethernet/stmicro/stmmac/ring_mode.c 	unsigned int entry = tx_q->cur_tx;
entry              27 drivers/net/ethernet/stmicro/stmmac/ring_mode.c 		desc = (struct dma_desc *)(tx_q->dma_etx + entry);
entry              29 drivers/net/ethernet/stmicro/stmmac/ring_mode.c 		desc = tx_q->dma_tx + entry;
entry              46 drivers/net/ethernet/stmicro/stmmac/ring_mode.c 		tx_q->tx_skbuff_dma[entry].buf = des2;
entry              47 drivers/net/ethernet/stmicro/stmmac/ring_mode.c 		tx_q->tx_skbuff_dma[entry].len = bmax;
entry              48 drivers/net/ethernet/stmicro/stmmac/ring_mode.c 		tx_q->tx_skbuff_dma[entry].is_jumbo = true;
entry              53 drivers/net/ethernet/stmicro/stmmac/ring_mode.c 		tx_q->tx_skbuff[entry] = NULL;
entry              54 drivers/net/ethernet/stmicro/stmmac/ring_mode.c 		entry = STMMAC_GET_ENTRY(entry, DMA_TX_SIZE);
entry              57 drivers/net/ethernet/stmicro/stmmac/ring_mode.c 			desc = (struct dma_desc *)(tx_q->dma_etx + entry);
entry              59 drivers/net/ethernet/stmicro/stmmac/ring_mode.c 			desc = tx_q->dma_tx + entry;
entry              66 drivers/net/ethernet/stmicro/stmmac/ring_mode.c 		tx_q->tx_skbuff_dma[entry].buf = des2;
entry              67 drivers/net/ethernet/stmicro/stmmac/ring_mode.c 		tx_q->tx_skbuff_dma[entry].len = len;
entry              68 drivers/net/ethernet/stmicro/stmmac/ring_mode.c 		tx_q->tx_skbuff_dma[entry].is_jumbo = true;
entry              80 drivers/net/ethernet/stmicro/stmmac/ring_mode.c 		tx_q->tx_skbuff_dma[entry].buf = des2;
entry              81 drivers/net/ethernet/stmicro/stmmac/ring_mode.c 		tx_q->tx_skbuff_dma[entry].len = nopaged_len;
entry              82 drivers/net/ethernet/stmicro/stmmac/ring_mode.c 		tx_q->tx_skbuff_dma[entry].is_jumbo = true;
entry              89 drivers/net/ethernet/stmicro/stmmac/ring_mode.c 	tx_q->cur_tx = entry;
entry              91 drivers/net/ethernet/stmicro/stmmac/ring_mode.c 	return entry;
entry             124 drivers/net/ethernet/stmicro/stmmac/ring_mode.c 	unsigned int entry = tx_q->dirty_tx;
entry             127 drivers/net/ethernet/stmicro/stmmac/ring_mode.c 	if (unlikely(tx_q->tx_skbuff_dma[entry].is_jumbo ||
entry             128 drivers/net/ethernet/stmicro/stmmac/ring_mode.c 		     (tx_q->tx_skbuff_dma[entry].last_segment &&
entry            1874 drivers/net/ethernet/stmicro/stmmac/stmmac_main.c 	unsigned int entry, count = 0;
entry            1880 drivers/net/ethernet/stmicro/stmmac/stmmac_main.c 	entry = tx_q->dirty_tx;
entry            1881 drivers/net/ethernet/stmicro/stmmac/stmmac_main.c 	while ((entry != tx_q->cur_tx) && (count < budget)) {
entry            1882 drivers/net/ethernet/stmicro/stmmac/stmmac_main.c 		struct sk_buff *skb = tx_q->tx_skbuff[entry];
entry            1887 drivers/net/ethernet/stmicro/stmmac/stmmac_main.c 			p = (struct dma_desc *)(tx_q->dma_etx + entry);
entry            1889 drivers/net/ethernet/stmicro/stmmac/stmmac_main.c 			p = tx_q->dma_tx + entry;
entry            1916 drivers/net/ethernet/stmicro/stmmac/stmmac_main.c 		if (likely(tx_q->tx_skbuff_dma[entry].buf)) {
entry            1917 drivers/net/ethernet/stmicro/stmmac/stmmac_main.c 			if (tx_q->tx_skbuff_dma[entry].map_as_page)
entry            1919 drivers/net/ethernet/stmicro/stmmac/stmmac_main.c 					       tx_q->tx_skbuff_dma[entry].buf,
entry            1920 drivers/net/ethernet/stmicro/stmmac/stmmac_main.c 					       tx_q->tx_skbuff_dma[entry].len,
entry            1924 drivers/net/ethernet/stmicro/stmmac/stmmac_main.c 						 tx_q->tx_skbuff_dma[entry].buf,
entry            1925 drivers/net/ethernet/stmicro/stmmac/stmmac_main.c 						 tx_q->tx_skbuff_dma[entry].len,
entry            1927 drivers/net/ethernet/stmicro/stmmac/stmmac_main.c 			tx_q->tx_skbuff_dma[entry].buf = 0;
entry            1928 drivers/net/ethernet/stmicro/stmmac/stmmac_main.c 			tx_q->tx_skbuff_dma[entry].len = 0;
entry            1929 drivers/net/ethernet/stmicro/stmmac/stmmac_main.c 			tx_q->tx_skbuff_dma[entry].map_as_page = false;
entry            1934 drivers/net/ethernet/stmicro/stmmac/stmmac_main.c 		tx_q->tx_skbuff_dma[entry].last_segment = false;
entry            1935 drivers/net/ethernet/stmicro/stmmac/stmmac_main.c 		tx_q->tx_skbuff_dma[entry].is_jumbo = false;
entry            1941 drivers/net/ethernet/stmicro/stmmac/stmmac_main.c 			tx_q->tx_skbuff[entry] = NULL;
entry            1946 drivers/net/ethernet/stmicro/stmmac/stmmac_main.c 		entry = STMMAC_GET_ENTRY(entry, DMA_TX_SIZE);
entry            1948 drivers/net/ethernet/stmicro/stmmac/stmmac_main.c 	tx_q->dirty_tx = entry;
entry            3140 drivers/net/ethernet/stmicro/stmmac/stmmac_main.c 	int entry;
entry            3168 drivers/net/ethernet/stmicro/stmmac/stmmac_main.c 	entry = tx_q->cur_tx;
entry            3169 drivers/net/ethernet/stmicro/stmmac/stmmac_main.c 	first_entry = entry;
entry            3175 drivers/net/ethernet/stmicro/stmmac/stmmac_main.c 		desc = (struct dma_desc *)(tx_q->dma_etx + entry);
entry            3177 drivers/net/ethernet/stmicro/stmmac/stmmac_main.c 		desc = tx_q->dma_tx + entry;
entry            3190 drivers/net/ethernet/stmicro/stmmac/stmmac_main.c 		entry = stmmac_jumbo_frm(priv, tx_q, skb, csum_insertion);
entry            3191 drivers/net/ethernet/stmicro/stmmac/stmmac_main.c 		if (unlikely(entry < 0) && (entry != -EINVAL))
entry            3200 drivers/net/ethernet/stmicro/stmmac/stmmac_main.c 		entry = STMMAC_GET_ENTRY(entry, DMA_TX_SIZE);
entry            3201 drivers/net/ethernet/stmicro/stmmac/stmmac_main.c 		WARN_ON(tx_q->tx_skbuff[entry]);
entry            3204 drivers/net/ethernet/stmicro/stmmac/stmmac_main.c 			desc = (struct dma_desc *)(tx_q->dma_etx + entry);
entry            3206 drivers/net/ethernet/stmicro/stmmac/stmmac_main.c 			desc = tx_q->dma_tx + entry;
entry            3213 drivers/net/ethernet/stmicro/stmmac/stmmac_main.c 		tx_q->tx_skbuff_dma[entry].buf = des;
entry            3217 drivers/net/ethernet/stmicro/stmmac/stmmac_main.c 		tx_q->tx_skbuff_dma[entry].map_as_page = true;
entry            3218 drivers/net/ethernet/stmicro/stmmac/stmmac_main.c 		tx_q->tx_skbuff_dma[entry].len = len;
entry            3219 drivers/net/ethernet/stmicro/stmmac/stmmac_main.c 		tx_q->tx_skbuff_dma[entry].last_segment = last_segment;
entry            3227 drivers/net/ethernet/stmicro/stmmac/stmmac_main.c 	tx_q->tx_skbuff[entry] = skb;
entry            3241 drivers/net/ethernet/stmicro/stmmac/stmmac_main.c 			desc = &tx_q->dma_etx[entry].basic;
entry            3243 drivers/net/ethernet/stmicro/stmmac/stmmac_main.c 			desc = &tx_q->dma_tx[entry];
entry            3255 drivers/net/ethernet/stmicro/stmmac/stmmac_main.c 	entry = STMMAC_GET_ENTRY(entry, DMA_TX_SIZE);
entry            3256 drivers/net/ethernet/stmicro/stmmac/stmmac_main.c 	tx_q->cur_tx = entry;
entry            3264 drivers/net/ethernet/stmicro/stmmac/stmmac_main.c 			   entry, first, nfrags);
entry            3388 drivers/net/ethernet/stmicro/stmmac/stmmac_main.c 	unsigned int entry = rx_q->dirty_rx;
entry            3393 drivers/net/ethernet/stmicro/stmmac/stmmac_main.c 		struct stmmac_rx_buffer *buf = &rx_q->buf_pool[entry];
entry            3398 drivers/net/ethernet/stmicro/stmmac/stmmac_main.c 			p = (struct dma_desc *)(rx_q->dma_erx + entry);
entry            3400 drivers/net/ethernet/stmicro/stmmac/stmmac_main.c 			p = rx_q->dma_rx + entry;
entry            3440 drivers/net/ethernet/stmicro/stmmac/stmmac_main.c 		entry = STMMAC_GET_ENTRY(entry, DMA_RX_SIZE);
entry            3442 drivers/net/ethernet/stmicro/stmmac/stmmac_main.c 	rx_q->dirty_rx = entry;
entry            3482 drivers/net/ethernet/stmicro/stmmac/stmmac_main.c 		int entry;
entry            3501 drivers/net/ethernet/stmicro/stmmac/stmmac_main.c 		entry = next_entry;
entry            3502 drivers/net/ethernet/stmicro/stmmac/stmmac_main.c 		buf = &rx_q->buf_pool[entry];
entry            3505 drivers/net/ethernet/stmicro/stmmac/stmmac_main.c 			p = (struct dma_desc *)(rx_q->dma_erx + entry);
entry            3507 drivers/net/ethernet/stmicro/stmmac/stmmac_main.c 			p = rx_q->dma_rx + entry;
entry            3529 drivers/net/ethernet/stmicro/stmmac/stmmac_main.c 					&priv->xstats, rx_q->dma_erx + entry);
entry              14 drivers/net/ethernet/stmicro/stmmac/stmmac_tc.c static void tc_fill_all_pass_entry(struct stmmac_tc_entry *entry)
entry              16 drivers/net/ethernet/stmicro/stmmac/stmmac_tc.c 	memset(entry, 0, sizeof(*entry));
entry              17 drivers/net/ethernet/stmicro/stmmac/stmmac_tc.c 	entry->in_use = true;
entry              18 drivers/net/ethernet/stmicro/stmmac/stmmac_tc.c 	entry->is_last = true;
entry              19 drivers/net/ethernet/stmicro/stmmac/stmmac_tc.c 	entry->is_frag = false;
entry              20 drivers/net/ethernet/stmicro/stmmac/stmmac_tc.c 	entry->prio = ~0x0;
entry              21 drivers/net/ethernet/stmicro/stmmac/stmmac_tc.c 	entry->handle = 0;
entry              22 drivers/net/ethernet/stmicro/stmmac/stmmac_tc.c 	entry->val.match_data = 0x0;
entry              23 drivers/net/ethernet/stmicro/stmmac/stmmac_tc.c 	entry->val.match_en = 0x0;
entry              24 drivers/net/ethernet/stmicro/stmmac/stmmac_tc.c 	entry->val.af = 1;
entry              25 drivers/net/ethernet/stmicro/stmmac/stmmac_tc.c 	entry->val.dma_ch_no = 0x0;
entry              32 drivers/net/ethernet/stmicro/stmmac/stmmac_tc.c 	struct stmmac_tc_entry *entry, *first = NULL, *dup = NULL;
entry              37 drivers/net/ethernet/stmicro/stmmac/stmmac_tc.c 		entry = &priv->tc_entries[i];
entry              38 drivers/net/ethernet/stmicro/stmmac/stmmac_tc.c 		if (!entry->in_use && !first && free)
entry              39 drivers/net/ethernet/stmicro/stmmac/stmmac_tc.c 			first = entry;
entry              40 drivers/net/ethernet/stmicro/stmmac/stmmac_tc.c 		if ((entry->handle == loc) && !free && !entry->is_frag)
entry              41 drivers/net/ethernet/stmicro/stmmac/stmmac_tc.c 			dup = entry;
entry              57 drivers/net/ethernet/stmicro/stmmac/stmmac_tc.c static int tc_fill_actions(struct stmmac_tc_entry *entry,
entry              61 drivers/net/ethernet/stmicro/stmmac/stmmac_tc.c 	struct stmmac_tc_entry *action_entry = entry;
entry              94 drivers/net/ethernet/stmicro/stmmac/stmmac_tc.c 	struct stmmac_tc_entry *entry, *frag = NULL;
entry             124 drivers/net/ethernet/stmicro/stmmac/stmmac_tc.c 	entry = tc_find_entry(priv, cls, true);
entry             125 drivers/net/ethernet/stmicro/stmmac/stmmac_tc.c 	if (!entry)
entry             135 drivers/net/ethernet/stmicro/stmmac/stmmac_tc.c 		entry->frag_ptr = frag;
entry             136 drivers/net/ethernet/stmicro/stmmac/stmmac_tc.c 		entry->val.match_en = (mask << (rem * 8)) &
entry             138 drivers/net/ethernet/stmicro/stmmac/stmmac_tc.c 		entry->val.match_data = (data << (rem * 8)) &
entry             140 drivers/net/ethernet/stmicro/stmmac/stmmac_tc.c 		entry->val.frame_offset = real_off;
entry             141 drivers/net/ethernet/stmicro/stmmac/stmmac_tc.c 		entry->prio = prio;
entry             151 drivers/net/ethernet/stmicro/stmmac/stmmac_tc.c 		entry->frag_ptr = NULL;
entry             152 drivers/net/ethernet/stmicro/stmmac/stmmac_tc.c 		entry->val.match_en = mask;
entry             153 drivers/net/ethernet/stmicro/stmmac/stmmac_tc.c 		entry->val.match_data = data;
entry             154 drivers/net/ethernet/stmicro/stmmac/stmmac_tc.c 		entry->val.frame_offset = real_off;
entry             155 drivers/net/ethernet/stmicro/stmmac/stmmac_tc.c 		entry->prio = prio;
entry             158 drivers/net/ethernet/stmicro/stmmac/stmmac_tc.c 	ret = tc_fill_actions(entry, frag, cls);
entry             167 drivers/net/ethernet/stmicro/stmmac/stmmac_tc.c 	entry->in_use = false;
entry             174 drivers/net/ethernet/stmicro/stmmac/stmmac_tc.c 	struct stmmac_tc_entry *entry;
entry             176 drivers/net/ethernet/stmicro/stmmac/stmmac_tc.c 	entry = tc_find_entry(priv, cls, false);
entry             177 drivers/net/ethernet/stmicro/stmmac/stmmac_tc.c 	if (!entry)
entry             180 drivers/net/ethernet/stmicro/stmmac/stmmac_tc.c 	entry->in_use = false;
entry             181 drivers/net/ethernet/stmicro/stmmac/stmmac_tc.c 	if (entry->frag_ptr) {
entry             182 drivers/net/ethernet/stmicro/stmmac/stmmac_tc.c 		entry = entry->frag_ptr;
entry             183 drivers/net/ethernet/stmicro/stmmac/stmmac_tc.c 		entry->is_frag = false;
entry             184 drivers/net/ethernet/stmicro/stmmac/stmmac_tc.c 		entry->in_use = false;
entry             372 drivers/net/ethernet/stmicro/stmmac/stmmac_tc.c 				 struct stmmac_flow_entry *entry)
entry             383 drivers/net/ethernet/stmicro/stmmac/stmmac_tc.c 			entry->action |= STMMAC_FLOW_ACTION_DROP;
entry             396 drivers/net/ethernet/stmicro/stmmac/stmmac_tc.c 			     struct stmmac_flow_entry *entry)
entry             407 drivers/net/ethernet/stmicro/stmmac/stmmac_tc.c 	entry->ip_proto = match.key->ip_proto;
entry             413 drivers/net/ethernet/stmicro/stmmac/stmmac_tc.c 			   struct stmmac_flow_entry *entry)
entry             417 drivers/net/ethernet/stmicro/stmmac/stmmac_tc.c 	bool inv = entry->action & STMMAC_FLOW_ACTION_DROP;
entry             429 drivers/net/ethernet/stmicro/stmmac/stmmac_tc.c 		ret = stmmac_config_l3_filter(priv, priv->hw, entry->idx, true,
entry             437 drivers/net/ethernet/stmicro/stmmac/stmmac_tc.c 		ret = stmmac_config_l3_filter(priv, priv->hw, entry->idx, true,
entry             448 drivers/net/ethernet/stmicro/stmmac/stmmac_tc.c 			     struct stmmac_flow_entry *entry)
entry             452 drivers/net/ethernet/stmicro/stmmac/stmmac_tc.c 	bool inv = entry->action & STMMAC_FLOW_ACTION_DROP;
entry             462 drivers/net/ethernet/stmicro/stmmac/stmmac_tc.c 	switch (entry->ip_proto) {
entry             477 drivers/net/ethernet/stmicro/stmmac/stmmac_tc.c 		ret = stmmac_config_l4_filter(priv, priv->hw, entry->idx, true,
entry             485 drivers/net/ethernet/stmicro/stmmac/stmmac_tc.c 		ret = stmmac_config_l4_filter(priv, priv->hw, entry->idx, true,
entry             491 drivers/net/ethernet/stmicro/stmmac/stmmac_tc.c 	entry->is_l4 = true;
entry             502 drivers/net/ethernet/stmicro/stmmac/stmmac_tc.c 		struct stmmac_flow_entry *entry = &priv->flow_entries[i];
entry             504 drivers/net/ethernet/stmicro/stmmac/stmmac_tc.c 		if (entry->cookie == cls->cookie)
entry             505 drivers/net/ethernet/stmicro/stmmac/stmmac_tc.c 			return entry;
entry             506 drivers/net/ethernet/stmicro/stmmac/stmmac_tc.c 		if (get_free && (entry->in_use == false))
entry             507 drivers/net/ethernet/stmicro/stmmac/stmmac_tc.c 			return entry;
entry             515 drivers/net/ethernet/stmicro/stmmac/stmmac_tc.c 		  struct stmmac_flow_entry *entry);
entry             525 drivers/net/ethernet/stmicro/stmmac/stmmac_tc.c 	struct stmmac_flow_entry *entry = tc_find_flow(priv, cls, false);
entry             529 drivers/net/ethernet/stmicro/stmmac/stmmac_tc.c 	if (!entry) {
entry             530 drivers/net/ethernet/stmicro/stmmac/stmmac_tc.c 		entry = tc_find_flow(priv, cls, true);
entry             531 drivers/net/ethernet/stmicro/stmmac/stmmac_tc.c 		if (!entry)
entry             535 drivers/net/ethernet/stmicro/stmmac/stmmac_tc.c 	ret = tc_parse_flow_actions(priv, &rule->action, entry);
entry             540 drivers/net/ethernet/stmicro/stmmac/stmmac_tc.c 		ret = tc_flow_parsers[i].fn(priv, cls, entry);
entry             542 drivers/net/ethernet/stmicro/stmmac/stmmac_tc.c 			entry->in_use = true;
entry             547 drivers/net/ethernet/stmicro/stmmac/stmmac_tc.c 	if (!entry->in_use)
entry             550 drivers/net/ethernet/stmicro/stmmac/stmmac_tc.c 	entry->cookie = cls->cookie;
entry             557 drivers/net/ethernet/stmicro/stmmac/stmmac_tc.c 	struct stmmac_flow_entry *entry = tc_find_flow(priv, cls, false);
entry             560 drivers/net/ethernet/stmicro/stmmac/stmmac_tc.c 	if (!entry || !entry->in_use)
entry             563 drivers/net/ethernet/stmicro/stmmac/stmmac_tc.c 	if (entry->is_l4) {
entry             564 drivers/net/ethernet/stmicro/stmmac/stmmac_tc.c 		ret = stmmac_config_l4_filter(priv, priv->hw, entry->idx, false,
entry             567 drivers/net/ethernet/stmicro/stmmac/stmmac_tc.c 		ret = stmmac_config_l3_filter(priv, priv->hw, entry->idx, false,
entry             571 drivers/net/ethernet/stmicro/stmmac/stmmac_tc.c 	entry->in_use = false;
entry             572 drivers/net/ethernet/stmicro/stmmac/stmmac_tc.c 	entry->cookie = 0;
entry             573 drivers/net/ethernet/stmicro/stmmac/stmmac_tc.c 	entry->is_l4 = false;
entry             623 drivers/net/ethernet/sun/cassini.c 	struct list_head *entry;
entry             640 drivers/net/ethernet/sun/cassini.c 	entry = cp->rx_spare_list.next;
entry             641 drivers/net/ethernet/sun/cassini.c 	list_del(entry);
entry             656 drivers/net/ethernet/sun/cassini.c 	return list_entry(entry, cas_page_t, list);
entry            1848 drivers/net/ethernet/sun/cassini.c 	int entry, count;
entry            1853 drivers/net/ethernet/sun/cassini.c 	entry = cp->tx_old[ring];
entry            1855 drivers/net/ethernet/sun/cassini.c 	count = TX_BUFF_COUNT(ring, entry, limit);
entry            1856 drivers/net/ethernet/sun/cassini.c 	while (entry != limit) {
entry            1857 drivers/net/ethernet/sun/cassini.c 		struct sk_buff *skb = skbs[entry];
entry            1864 drivers/net/ethernet/sun/cassini.c 			entry = TX_DESC_NEXT(ring, entry);
entry            1870 drivers/net/ethernet/sun/cassini.c 			+ cp->tx_tiny_use[ring][entry].nbufs + 1;
entry            1875 drivers/net/ethernet/sun/cassini.c 			     "tx[%d] done, slot %d\n", ring, entry);
entry            1877 drivers/net/ethernet/sun/cassini.c 		skbs[entry] = NULL;
entry            1878 drivers/net/ethernet/sun/cassini.c 		cp->tx_tiny_use[ring][entry].nbufs = 0;
entry            1881 drivers/net/ethernet/sun/cassini.c 			struct cas_tx_desc *txd = txds + entry;
entry            1888 drivers/net/ethernet/sun/cassini.c 			entry = TX_DESC_NEXT(ring, entry);
entry            1891 drivers/net/ethernet/sun/cassini.c 			if (cp->tx_tiny_use[ring][entry].used) {
entry            1892 drivers/net/ethernet/sun/cassini.c 				cp->tx_tiny_use[ring][entry].used = 0;
entry            1893 drivers/net/ethernet/sun/cassini.c 				entry = TX_DESC_NEXT(ring, entry);
entry            1903 drivers/net/ethernet/sun/cassini.c 	cp->tx_old[ring] = entry;
entry            1942 drivers/net/ethernet/sun/cassini.c 			      int entry, const u64 *words,
entry            2181 drivers/net/ethernet/sun/cassini.c 	int entry;
entry            2183 drivers/net/ethernet/sun/cassini.c 	entry = cp->rx_old[ring];
entry            2186 drivers/net/ethernet/sun/cassini.c 	cp->init_rxds[ring][entry].buffer = cpu_to_le64(new->dma_addr);
entry            2187 drivers/net/ethernet/sun/cassini.c 	cp->init_rxds[ring][entry].index  =
entry            2191 drivers/net/ethernet/sun/cassini.c 	entry = RX_DESC_ENTRY(ring, entry + 1);
entry            2192 drivers/net/ethernet/sun/cassini.c 	cp->rx_old[ring] = entry;
entry            2194 drivers/net/ethernet/sun/cassini.c 	if (entry % 4)
entry            2198 drivers/net/ethernet/sun/cassini.c 		writel(entry, cp->regs + REG_RX_KICK);
entry            2201 drivers/net/ethernet/sun/cassini.c 		writel(entry, cp->regs + REG_PLUS_RX_KICK1);
entry            2208 drivers/net/ethernet/sun/cassini.c 	unsigned int entry, last, count, released;
entry            2212 drivers/net/ethernet/sun/cassini.c 	entry = cp->rx_old[ring];
entry            2215 drivers/net/ethernet/sun/cassini.c 		     "rxd[%d] interrupt, done: %d\n", ring, entry);
entry            2218 drivers/net/ethernet/sun/cassini.c 	count = entry & 0x3;
entry            2219 drivers/net/ethernet/sun/cassini.c 	last = RX_DESC_ENTRY(ring, num ? entry + num - 4: entry - 4);
entry            2221 drivers/net/ethernet/sun/cassini.c 	while (entry != last) {
entry            2223 drivers/net/ethernet/sun/cassini.c 		if (page_count(page[entry]->buffer) > 1) {
entry            2233 drivers/net/ethernet/sun/cassini.c 				cp->rx_old[ring]  = entry;
entry            2238 drivers/net/ethernet/sun/cassini.c 			list_add(&page[entry]->list, &cp->rx_inuse_list);
entry            2240 drivers/net/ethernet/sun/cassini.c 			cp->init_rxds[ring][entry].buffer =
entry            2242 drivers/net/ethernet/sun/cassini.c 			page[entry] = new;
entry            2247 drivers/net/ethernet/sun/cassini.c 			cluster = entry;
entry            2251 drivers/net/ethernet/sun/cassini.c 		entry = RX_DESC_ENTRY(ring, entry + 1);
entry            2253 drivers/net/ethernet/sun/cassini.c 	cp->rx_old[ring] = entry;
entry            2282 drivers/net/ethernet/sun/cassini.c 	int entry, drops;
entry            2290 drivers/net/ethernet/sun/cassini.c 	entry = cp->rx_new[ring];
entry            2293 drivers/net/ethernet/sun/cassini.c 		struct cas_rx_comp *rxc = rxcs + entry;
entry            2332 drivers/net/ethernet/sun/cassini.c 		len = cas_rx_process_pkt(cp, rxc, entry, words, &skb);
entry            2379 drivers/net/ethernet/sun/cassini.c 		entry = RX_COMP_ENTRY(ring, entry + 1 +
entry            2386 drivers/net/ethernet/sun/cassini.c 	cp->rx_new[ring] = entry;
entry            2399 drivers/net/ethernet/sun/cassini.c 	int last, entry;
entry            2402 drivers/net/ethernet/sun/cassini.c 	entry = cp->rx_new[ring];
entry            2405 drivers/net/ethernet/sun/cassini.c 		     ring, readl(cp->regs + REG_RX_COMP_HEAD), entry);
entry            2408 drivers/net/ethernet/sun/cassini.c 	while (last != entry) {
entry            2716 drivers/net/ethernet/sun/cassini.c static inline int cas_intme(int ring, int entry)
entry            2719 drivers/net/ethernet/sun/cassini.c 	if (!(entry & ((TX_DESC_RINGN_SIZE(ring) >> 1) - 1)))
entry            2725 drivers/net/ethernet/sun/cassini.c static void cas_write_txd(struct cas *cp, int ring, int entry,
entry            2728 drivers/net/ethernet/sun/cassini.c 	struct cas_tx_desc *txd = cp->init_txds[ring] + entry;
entry            2731 drivers/net/ethernet/sun/cassini.c 	if (cas_intme(ring, entry))
entry            2740 drivers/net/ethernet/sun/cassini.c 				const int entry)
entry            2742 drivers/net/ethernet/sun/cassini.c 	return cp->tx_tiny_bufs[ring] + TX_TINY_BUF_LEN*entry;
entry            2746 drivers/net/ethernet/sun/cassini.c 				     const int entry, const int tentry)
entry            2749 drivers/net/ethernet/sun/cassini.c 	cp->tx_tiny_use[ring][entry].used = 1;
entry            2750 drivers/net/ethernet/sun/cassini.c 	return cp->tx_tiny_dvma[ring] + TX_TINY_BUF_LEN*entry;
entry            2757 drivers/net/ethernet/sun/cassini.c 	int entry, nr_frags, frag, tabort, tentry;
entry            2784 drivers/net/ethernet/sun/cassini.c 	entry = cp->tx_new[ring];
entry            2785 drivers/net/ethernet/sun/cassini.c 	cp->tx_skbs[ring][entry] = skb;
entry            2793 drivers/net/ethernet/sun/cassini.c 	tentry = entry;
entry            2797 drivers/net/ethernet/sun/cassini.c 		cas_write_txd(cp, ring, entry, mapping, len - tabort,
entry            2799 drivers/net/ethernet/sun/cassini.c 		entry = TX_DESC_NEXT(ring, entry);
entry            2802 drivers/net/ethernet/sun/cassini.c 			      tx_tiny_buf(cp, ring, entry), tabort);
entry            2803 drivers/net/ethernet/sun/cassini.c 		mapping = tx_tiny_map(cp, ring, entry, tentry);
entry            2804 drivers/net/ethernet/sun/cassini.c 		cas_write_txd(cp, ring, entry, mapping, tabort, ctrl,
entry            2807 drivers/net/ethernet/sun/cassini.c 		cas_write_txd(cp, ring, entry, mapping, len, ctrl |
entry            2810 drivers/net/ethernet/sun/cassini.c 	entry = TX_DESC_NEXT(ring, entry);
entry            2824 drivers/net/ethernet/sun/cassini.c 			cas_write_txd(cp, ring, entry, mapping, len - tabort,
entry            2826 drivers/net/ethernet/sun/cassini.c 			entry = TX_DESC_NEXT(ring, entry);
entry            2829 drivers/net/ethernet/sun/cassini.c 			memcpy(tx_tiny_buf(cp, ring, entry),
entry            2833 drivers/net/ethernet/sun/cassini.c 			mapping = tx_tiny_map(cp, ring, entry, tentry);
entry            2837 drivers/net/ethernet/sun/cassini.c 		cas_write_txd(cp, ring, entry, mapping, len, ctrl,
entry            2839 drivers/net/ethernet/sun/cassini.c 		entry = TX_DESC_NEXT(ring, entry);
entry            2842 drivers/net/ethernet/sun/cassini.c 	cp->tx_new[ring] = entry;
entry            2848 drivers/net/ethernet/sun/cassini.c 		     ring, entry, skb->len, TX_BUFFS_AVAIL(cp, ring));
entry            2849 drivers/net/ethernet/sun/cassini.c 	writel(entry, cp->regs + REG_TX_KICKN(ring));
entry            9030 drivers/net/ethernet/sun/niu.c 		msi_vec[i].entry = i;
entry             957 drivers/net/ethernet/sun/sunbmac.c 	int len, entry;
entry             966 drivers/net/ethernet/sun/sunbmac.c 	entry = bp->tx_new;
entry             967 drivers/net/ethernet/sun/sunbmac.c 	DTX(("bigmac_start_xmit: len(%d) entry(%d)\n", len, entry));
entry             968 drivers/net/ethernet/sun/sunbmac.c 	bp->bmac_block->be_txd[entry].tx_flags = TXD_UPDATE;
entry             969 drivers/net/ethernet/sun/sunbmac.c 	bp->tx_skbs[entry] = skb;
entry             970 drivers/net/ethernet/sun/sunbmac.c 	bp->bmac_block->be_txd[entry].tx_addr = mapping;
entry             971 drivers/net/ethernet/sun/sunbmac.c 	bp->bmac_block->be_txd[entry].tx_flags =
entry             973 drivers/net/ethernet/sun/sunbmac.c 	bp->tx_new = NEXT_TX(entry);
entry             645 drivers/net/ethernet/sun/sungem.c 	int entry, limit;
entry             647 drivers/net/ethernet/sun/sungem.c 	entry = gp->tx_old;
entry             649 drivers/net/ethernet/sun/sungem.c 	while (entry != limit) {
entry             658 drivers/net/ethernet/sun/sungem.c 				gp->dev->name, entry);
entry             659 drivers/net/ethernet/sun/sungem.c 		skb = gp->tx_skbs[entry];
entry             661 drivers/net/ethernet/sun/sungem.c 			int last = entry + skb_shinfo(skb)->nr_frags;
entry             662 drivers/net/ethernet/sun/sungem.c 			int walk = entry;
entry             676 drivers/net/ethernet/sun/sungem.c 		gp->tx_skbs[entry] = NULL;
entry             680 drivers/net/ethernet/sun/sungem.c 			txd = &gp->init_block->txd[entry];
entry             686 drivers/net/ethernet/sun/sungem.c 			entry = NEXT_TX(entry);
entry             692 drivers/net/ethernet/sun/sungem.c 	gp->tx_old = entry;
entry             760 drivers/net/ethernet/sun/sungem.c 	int entry, drops, work_done = 0;
entry             767 drivers/net/ethernet/sun/sungem.c 	entry = gp->rx_new;
entry             771 drivers/net/ethernet/sun/sungem.c 		struct gem_rxd *rxd = &gp->init_block->rxd[entry];
entry             790 drivers/net/ethernet/sun/sungem.c 		if (entry == done) {
entry             792 drivers/net/ethernet/sun/sungem.c 			if (entry == done)
entry             799 drivers/net/ethernet/sun/sungem.c 		skb = gp->rx_skbs[entry];
entry             827 drivers/net/ethernet/sun/sungem.c 			gp->rx_skbs[entry] = new_skb;
entry             871 drivers/net/ethernet/sun/sungem.c 		entry = NEXT_RX(entry);
entry             874 drivers/net/ethernet/sun/sungem.c 	gem_post_rxds(gp, entry);
entry             876 drivers/net/ethernet/sun/sungem.c 	gp->rx_new = entry;
entry             991 drivers/net/ethernet/sun/sungem.c static __inline__ int gem_intme(int entry)
entry             994 drivers/net/ethernet/sun/sungem.c 	if (!(entry & ((TX_RING_SIZE>>1)-1)))
entry            1004 drivers/net/ethernet/sun/sungem.c 	int entry;
entry            1026 drivers/net/ethernet/sun/sungem.c 	entry = gp->tx_new;
entry            1027 drivers/net/ethernet/sun/sungem.c 	gp->tx_skbs[entry] = skb;
entry            1030 drivers/net/ethernet/sun/sungem.c 		struct gem_txd *txd = &gp->init_block->txd[entry];
entry            1040 drivers/net/ethernet/sun/sungem.c 		if (gem_intme(entry))
entry            1045 drivers/net/ethernet/sun/sungem.c 		entry = NEXT_TX(entry);
entry            1051 drivers/net/ethernet/sun/sungem.c 		int frag, first_entry = entry;
entry            1054 drivers/net/ethernet/sun/sungem.c 		if (gem_intme(entry))
entry            1064 drivers/net/ethernet/sun/sungem.c 		entry = NEXT_TX(entry);
entry            1079 drivers/net/ethernet/sun/sungem.c 			txd = &gp->init_block->txd[entry];
entry            1084 drivers/net/ethernet/sun/sungem.c 			if (gem_intme(entry))
entry            1087 drivers/net/ethernet/sun/sungem.c 			entry = NEXT_TX(entry);
entry            1096 drivers/net/ethernet/sun/sungem.c 	gp->tx_new = entry;
entry            1111 drivers/net/ethernet/sun/sungem.c 		       dev->name, entry, skb->len);
entry            2268 drivers/net/ethernet/sun/sunhme.c 				 u32 first_len, u32 first_entry, u32 entry)
entry            2275 drivers/net/ethernet/sun/sunhme.c 	while (first_entry != entry) {
entry            2290 drivers/net/ethernet/sun/sunhme.c  	int entry;
entry            2313 drivers/net/ethernet/sun/sunhme.c 	entry = hp->tx_new;
entry            2314 drivers/net/ethernet/sun/sunhme.c 	SXD(("SX<l[%d]e[%d]>", len, entry));
entry            2315 drivers/net/ethernet/sun/sunhme.c 	hp->tx_skbs[entry] = skb;
entry            2325 drivers/net/ethernet/sun/sunhme.c 		hme_write_txd(hp, &hp->happy_block->happy_meal_txd[entry],
entry            2328 drivers/net/ethernet/sun/sunhme.c 		entry = NEXT_TX(entry);
entry            2331 drivers/net/ethernet/sun/sunhme.c 		int frag, first_entry = entry;
entry            2341 drivers/net/ethernet/sun/sunhme.c 		entry = NEXT_TX(entry);
entry            2352 drivers/net/ethernet/sun/sunhme.c 						     first_entry, entry);
entry            2358 drivers/net/ethernet/sun/sunhme.c 			hme_write_txd(hp, &hp->happy_block->happy_meal_txd[entry],
entry            2361 drivers/net/ethernet/sun/sunhme.c 			entry = NEXT_TX(entry);
entry            2368 drivers/net/ethernet/sun/sunhme.c 	hp->tx_new = entry;
entry             579 drivers/net/ethernet/sun/sunqe.c 	int len, entry;
entry             586 drivers/net/ethernet/sun/sunqe.c 	entry = qep->tx_new;
entry             588 drivers/net/ethernet/sun/sunqe.c 	txbuf = &qbufs->tx_buf[entry & (TX_RING_SIZE - 1)][0];
entry             590 drivers/net/ethernet/sun/sunqe.c 		qebuf_offset(tx_buf, (entry & (TX_RING_SIZE - 1)));
entry             593 drivers/net/ethernet/sun/sunqe.c 	qep->qe_block->qe_txd[entry].tx_flags = TXD_UPDATE;
entry             597 drivers/net/ethernet/sun/sunqe.c 	qep->qe_block->qe_txd[entry].tx_addr = txbuf_dvma;
entry             598 drivers/net/ethernet/sun/sunqe.c 	qep->qe_block->qe_txd[entry].tx_flags =
entry             600 drivers/net/ethernet/sun/sunqe.c 	qep->tx_new = NEXT_TX(entry);
entry             474 drivers/net/ethernet/ti/netcp_core.c 	struct netcp_hook_list *entry;
entry             478 drivers/net/ethernet/ti/netcp_core.c 	entry = devm_kzalloc(netcp_priv->dev, sizeof(*entry), GFP_KERNEL);
entry             479 drivers/net/ethernet/ti/netcp_core.c 	if (!entry)
entry             482 drivers/net/ethernet/ti/netcp_core.c 	entry->hook_rtn  = hook_rtn;
entry             483 drivers/net/ethernet/ti/netcp_core.c 	entry->hook_data = hook_data;
entry             484 drivers/net/ethernet/ti/netcp_core.c 	entry->order     = order;
entry             491 drivers/net/ethernet/ti/netcp_core.c 	__list_add(&entry->list, next->list.prev, &next->list);
entry             523 drivers/net/ethernet/ti/netcp_core.c 	struct netcp_hook_list *entry;
entry             527 drivers/net/ethernet/ti/netcp_core.c 	entry = devm_kzalloc(netcp_priv->dev, sizeof(*entry), GFP_KERNEL);
entry             528 drivers/net/ethernet/ti/netcp_core.c 	if (!entry)
entry             531 drivers/net/ethernet/ti/netcp_core.c 	entry->hook_rtn  = hook_rtn;
entry             532 drivers/net/ethernet/ti/netcp_core.c 	entry->hook_data = hook_data;
entry             533 drivers/net/ethernet/ti/netcp_core.c 	entry->order     = order;
entry             540 drivers/net/ethernet/ti/netcp_core.c 	__list_add(&entry->list, next->list.prev, &next->list);
entry            1246 drivers/net/ethernet/via/via-rhine.c 					   struct rhine_skb_dma *sd, int entry)
entry            1248 drivers/net/ethernet/via/via-rhine.c 	rp->rx_skbuff_dma[entry] = sd->dma;
entry            1249 drivers/net/ethernet/via/via-rhine.c 	rp->rx_skbuff[entry] = sd->skb;
entry            1251 drivers/net/ethernet/via/via-rhine.c 	rp->rx_ring[entry].addr = cpu_to_le32(sd->dma);
entry            1787 drivers/net/ethernet/via/via-rhine.c 	unsigned entry;
entry            1793 drivers/net/ethernet/via/via-rhine.c 	entry = rp->cur_tx % TX_RING_SIZE;
entry            1798 drivers/net/ethernet/via/via-rhine.c 	rp->tx_skbuff[entry] = skb;
entry            1806 drivers/net/ethernet/via/via-rhine.c 			rp->tx_skbuff[entry] = NULL;
entry            1812 drivers/net/ethernet/via/via-rhine.c 		skb_copy_and_csum_dev(skb, rp->tx_buf[entry]);
entry            1814 drivers/net/ethernet/via/via-rhine.c 			memset(rp->tx_buf[entry] + skb->len, 0,
entry            1816 drivers/net/ethernet/via/via-rhine.c 		rp->tx_skbuff_dma[entry] = 0;
entry            1817 drivers/net/ethernet/via/via-rhine.c 		rp->tx_ring[entry].addr = cpu_to_le32(rp->tx_bufs_dma +
entry            1818 drivers/net/ethernet/via/via-rhine.c 						      (rp->tx_buf[entry] -
entry            1821 drivers/net/ethernet/via/via-rhine.c 		rp->tx_skbuff_dma[entry] =
entry            1824 drivers/net/ethernet/via/via-rhine.c 		if (dma_mapping_error(hwdev, rp->tx_skbuff_dma[entry])) {
entry            1826 drivers/net/ethernet/via/via-rhine.c 			rp->tx_skbuff_dma[entry] = 0;
entry            1830 drivers/net/ethernet/via/via-rhine.c 		rp->tx_ring[entry].addr = cpu_to_le32(rp->tx_skbuff_dma[entry]);
entry            1833 drivers/net/ethernet/via/via-rhine.c 	rp->tx_ring[entry].desc_length =
entry            1842 drivers/net/ethernet/via/via-rhine.c 		rp->tx_ring[entry].tx_status = cpu_to_le32((vid_pcp) << 16);
entry            1844 drivers/net/ethernet/via/via-rhine.c 		rp->tx_ring[entry].desc_length |= cpu_to_le32(0x020000);
entry            1847 drivers/net/ethernet/via/via-rhine.c 		rp->tx_ring[entry].tx_status = 0;
entry            1852 drivers/net/ethernet/via/via-rhine.c 	rp->tx_ring[entry].tx_status |= cpu_to_le32(DescOwn);
entry            1884 drivers/net/ethernet/via/via-rhine.c 		  rp->cur_tx - 1, entry);
entry            1943 drivers/net/ethernet/via/via-rhine.c 		unsigned int entry = dirty_tx % TX_RING_SIZE;
entry            1944 drivers/net/ethernet/via/via-rhine.c 		u32 txstatus = le32_to_cpu(rp->tx_ring[entry].tx_status);
entry            1947 drivers/net/ethernet/via/via-rhine.c 			  entry, txstatus);
entry            1950 drivers/net/ethernet/via/via-rhine.c 		skb = rp->tx_skbuff[entry];
entry            1966 drivers/net/ethernet/via/via-rhine.c 				rp->tx_ring[entry].tx_status = cpu_to_le32(DescOwn);
entry            1984 drivers/net/ethernet/via/via-rhine.c 		if (rp->tx_skbuff_dma[entry]) {
entry            1986 drivers/net/ethernet/via/via-rhine.c 					 rp->tx_skbuff_dma[entry],
entry            1993 drivers/net/ethernet/via/via-rhine.c 		rp->tx_skbuff[entry] = NULL;
entry            2045 drivers/net/ethernet/via/via-rhine.c 	int entry = rp->cur_rx % RX_RING_SIZE;
entry            2049 drivers/net/ethernet/via/via-rhine.c 		  entry, le32_to_cpu(rp->rx_ring[entry].rx_status));
entry            2053 drivers/net/ethernet/via/via-rhine.c 		struct rx_desc *desc = rp->rx_ring + entry;
entry            2068 drivers/net/ethernet/via/via-rhine.c 					    entry, data_size,
entry            2103 drivers/net/ethernet/via/via-rhine.c 							rp->rx_skbuff_dma[entry],
entry            2108 drivers/net/ethernet/via/via-rhine.c 						 rp->rx_skbuff[entry]->data,
entry            2112 drivers/net/ethernet/via/via-rhine.c 							   rp->rx_skbuff_dma[entry],
entry            2121 drivers/net/ethernet/via/via-rhine.c 				skb = rp->rx_skbuff[entry];
entry            2124 drivers/net/ethernet/via/via-rhine.c 						 rp->rx_skbuff_dma[entry],
entry            2127 drivers/net/ethernet/via/via-rhine.c 				rhine_skb_dma_nic_store(rp, &sd, entry);
entry            2145 drivers/net/ethernet/via/via-rhine.c 		entry = (++rp->cur_rx) % RX_RING_SIZE;
entry            2158 drivers/net/ethernet/via/via-rhine.c 	int entry = rp->dirty_tx % TX_RING_SIZE;
entry            2170 drivers/net/ethernet/via/via-rhine.c 		iowrite32(rp->tx_ring_dma + entry * sizeof(struct tx_desc),
entry            2176 drivers/net/ethernet/via/via-rhine.c 		if (rp->tx_ring[entry].desc_length & cpu_to_le32(0x020000))
entry            2480 drivers/net/fddi/defxx.c 	p_addr = &bp->cmd_req_virt->addr_filter_set.entry[0];
entry            3107 drivers/net/fddi/defxx.c 		int entry;
entry            3109 drivers/net/fddi/defxx.c 		entry = bp->rcv_xmt_reg.index.rcv_comp;
entry            3111 drivers/net/fddi/defxx.c 		p_buff = (char *) (((struct sk_buff *)bp->p_rcv_buff_va[entry])->data);
entry            3113 drivers/net/fddi/defxx.c 		p_buff = bp->p_rcv_buff_va[entry];
entry            3115 drivers/net/fddi/defxx.c 		dma_addr = bp->descr_block_virt->rcv_data[entry].long_1;
entry            3165 drivers/net/fddi/defxx.c 						skb = (struct sk_buff *)bp->p_rcv_buff_va[entry];
entry            3171 drivers/net/fddi/defxx.c 						bp->p_rcv_buff_va[entry] = (char *)newskb;
entry            3172 drivers/net/fddi/defxx.c 						bp->descr_block_virt->rcv_data[entry].long_1 = (u32)new_dma_addr;
entry             643 drivers/net/fddi/defxx.h 	PI_LAN_ADDR	entry[PI_CMD_ADDR_FILTER_K_SIZE];
entry             665 drivers/net/fddi/defxx.h 	PI_LAN_ADDR		entry[PI_CMD_ADDR_FILTER_K_SIZE];
entry             137 drivers/net/macvlan.c 	struct macvlan_source_entry *entry;
entry             141 drivers/net/macvlan.c 	hlist_for_each_entry_rcu(entry, h, hlist) {
entry             142 drivers/net/macvlan.c 		if (ether_addr_equal_64bits(entry->addr, addr) &&
entry             143 drivers/net/macvlan.c 		    entry->vlan == vlan)
entry             144 drivers/net/macvlan.c 			return entry;
entry             153 drivers/net/macvlan.c 	struct macvlan_source_entry *entry;
entry             156 drivers/net/macvlan.c 	entry = macvlan_hash_lookup_source(vlan, addr);
entry             157 drivers/net/macvlan.c 	if (entry)
entry             160 drivers/net/macvlan.c 	entry = kmalloc(sizeof(*entry), GFP_KERNEL);
entry             161 drivers/net/macvlan.c 	if (!entry)
entry             164 drivers/net/macvlan.c 	ether_addr_copy(entry->addr, addr);
entry             165 drivers/net/macvlan.c 	entry->vlan = vlan;
entry             167 drivers/net/macvlan.c 	hlist_add_head_rcu(&entry->hlist, h);
entry             182 drivers/net/macvlan.c static void macvlan_hash_del_source(struct macvlan_source_entry *entry)
entry             184 drivers/net/macvlan.c 	hlist_del_rcu(&entry->hlist);
entry             185 drivers/net/macvlan.c 	kfree_rcu(entry, rcu);
entry             386 drivers/net/macvlan.c 			struct macvlan_source_entry *entry;
entry             388 drivers/net/macvlan.c 			entry = hlist_entry(h, struct macvlan_source_entry,
entry             390 drivers/net/macvlan.c 			if (entry->vlan == vlan)
entry             391 drivers/net/macvlan.c 				macvlan_hash_del_source(entry);
entry             427 drivers/net/macvlan.c 	struct macvlan_source_entry *entry;
entry             431 drivers/net/macvlan.c 	hlist_for_each_entry_rcu(entry, h, hlist) {
entry             432 drivers/net/macvlan.c 		if (ether_addr_equal_64bits(entry->addr, addr))
entry             433 drivers/net/macvlan.c 			macvlan_forward_source_one(skb, entry->vlan);
entry            1323 drivers/net/macvlan.c 	struct macvlan_source_entry *entry;
entry            1338 drivers/net/macvlan.c 		entry = macvlan_hash_lookup_source(vlan, addr);
entry            1339 drivers/net/macvlan.c 		if (entry) {
entry            1340 drivers/net/macvlan.c 			macvlan_hash_del_source(entry);
entry            1586 drivers/net/macvlan.c 	struct macvlan_source_entry *entry;
entry            1588 drivers/net/macvlan.c 	hlist_for_each_entry_rcu(entry, h, hlist) {
entry            1589 drivers/net/macvlan.c 		if (entry->vlan != vlan)
entry            1591 drivers/net/macvlan.c 		if (nla_put(skb, IFLA_MACVLAN_MACADDR, ETH_ALEN, entry->addr))
entry              47 drivers/net/netdevsim/bpf.c 	} entry[NSIM_BPF_MAX_KEYS];
entry             338 drivers/net/netdevsim/bpf.c 	for (i = 0; i < ARRAY_SIZE(nmap->entry); i++)
entry             339 drivers/net/netdevsim/bpf.c 		if (nsim_map_key_match(&offmap->map, &nmap->entry[i], key))
entry             350 drivers/net/netdevsim/bpf.c 	nmap->entry[idx].key = kmalloc(offmap->map.key_size, GFP_USER);
entry             351 drivers/net/netdevsim/bpf.c 	if (!nmap->entry[idx].key)
entry             353 drivers/net/netdevsim/bpf.c 	nmap->entry[idx].value = kmalloc(offmap->map.value_size, GFP_USER);
entry             354 drivers/net/netdevsim/bpf.c 	if (!nmap->entry[idx].value) {
entry             355 drivers/net/netdevsim/bpf.c 		kfree(nmap->entry[idx].key);
entry             356 drivers/net/netdevsim/bpf.c 		nmap->entry[idx].key = NULL;
entry             379 drivers/net/netdevsim/bpf.c 	for (; idx < ARRAY_SIZE(nmap->entry); idx++) {
entry             380 drivers/net/netdevsim/bpf.c 		if (nmap->entry[idx].key) {
entry             381 drivers/net/netdevsim/bpf.c 			memcpy(next_key, nmap->entry[idx].key,
entry             389 drivers/net/netdevsim/bpf.c 	if (idx == ARRAY_SIZE(nmap->entry))
entry             404 drivers/net/netdevsim/bpf.c 		memcpy(value, nmap->entry[idx].value, offmap->map.value_size);
entry             431 drivers/net/netdevsim/bpf.c 		for (idx = 0; idx < ARRAY_SIZE(nmap->entry); idx++)
entry             432 drivers/net/netdevsim/bpf.c 			if (!nmap->entry[idx].key)
entry             434 drivers/net/netdevsim/bpf.c 		if (idx == ARRAY_SIZE(nmap->entry)) {
entry             444 drivers/net/netdevsim/bpf.c 	memcpy(nmap->entry[idx].key, key, offmap->map.key_size);
entry             445 drivers/net/netdevsim/bpf.c 	memcpy(nmap->entry[idx].value, value, offmap->map.value_size);
entry             464 drivers/net/netdevsim/bpf.c 		kfree(nmap->entry[idx].key);
entry             465 drivers/net/netdevsim/bpf.c 		kfree(nmap->entry[idx].value);
entry             466 drivers/net/netdevsim/bpf.c 		memset(&nmap->entry[idx], 0, sizeof(nmap->entry[idx]));
entry             505 drivers/net/netdevsim/bpf.c 		for (i = 0; i < ARRAY_SIZE(nmap->entry); i++) {
entry             511 drivers/net/netdevsim/bpf.c 			key = nmap->entry[i].key;
entry             523 drivers/net/netdevsim/bpf.c 		kfree(nmap->entry[i].key);
entry             524 drivers/net/netdevsim/bpf.c 		kfree(nmap->entry[i].value);
entry             535 drivers/net/netdevsim/bpf.c 	for (i = 0; i < ARRAY_SIZE(nmap->entry); i++) {
entry             536 drivers/net/netdevsim/bpf.c 		kfree(nmap->entry[i].key);
entry             537 drivers/net/netdevsim/bpf.c 		kfree(nmap->entry[i].value);
entry              45 drivers/net/netdevsim/fib.c 	struct nsim_fib_entry *entry;
entry              49 drivers/net/netdevsim/fib.c 		entry = &fib_data->ipv4.fib;
entry              52 drivers/net/netdevsim/fib.c 		entry = &fib_data->ipv4.rules;
entry              55 drivers/net/netdevsim/fib.c 		entry = &fib_data->ipv6.fib;
entry              58 drivers/net/netdevsim/fib.c 		entry = &fib_data->ipv6.rules;
entry              64 drivers/net/netdevsim/fib.c 	return max ? entry->max : entry->num;
entry              71 drivers/net/netdevsim/fib.c 	struct nsim_fib_entry *entry;
entry              76 drivers/net/netdevsim/fib.c 		entry = &fib_data->ipv4.fib;
entry              79 drivers/net/netdevsim/fib.c 		entry = &fib_data->ipv4.rules;
entry              82 drivers/net/netdevsim/fib.c 		entry = &fib_data->ipv6.fib;
entry              85 drivers/net/netdevsim/fib.c 		entry = &fib_data->ipv6.rules;
entry              94 drivers/net/netdevsim/fib.c 	if (val < entry->num) {
entry              98 drivers/net/netdevsim/fib.c 		entry->max = val;
entry             104 drivers/net/netdevsim/fib.c static int nsim_fib_rule_account(struct nsim_fib_entry *entry, bool add,
entry             110 drivers/net/netdevsim/fib.c 		if (entry->num < entry->max) {
entry             111 drivers/net/netdevsim/fib.c 			entry->num++;
entry             117 drivers/net/netdevsim/fib.c 		entry->num--;
entry             141 drivers/net/netdevsim/fib.c static int nsim_fib_account(struct nsim_fib_entry *entry, bool add,
entry             147 drivers/net/netdevsim/fib.c 		if (entry->num < entry->max) {
entry             148 drivers/net/netdevsim/fib.c 			entry->num++;
entry             154 drivers/net/netdevsim/fib.c 		entry->num--;
entry            2239 drivers/net/usb/lan78xx.c 		struct skb_data	*entry;
entry            2244 drivers/net/usb/lan78xx.c 			entry = (struct skb_data *)skb->cb;
entry            2245 drivers/net/usb/lan78xx.c 			if (entry->state != unlink_start)
entry            2250 drivers/net/usb/lan78xx.c 		entry->state = unlink_start;
entry            2251 drivers/net/usb/lan78xx.c 		urb = entry->urb;
entry            2775 drivers/net/usb/lan78xx.c 	struct skb_data *entry = (struct skb_data *)skb->cb;
entry            2778 drivers/net/usb/lan78xx.c 	old_state = entry->state;
entry            2779 drivers/net/usb/lan78xx.c 	entry->state = state;
entry            2796 drivers/net/usb/lan78xx.c 	struct skb_data *entry = (struct skb_data *)skb->cb;
entry            2797 drivers/net/usb/lan78xx.c 	struct lan78xx_net *dev = entry->dev;
entry            2800 drivers/net/usb/lan78xx.c 		dev->net->stats.tx_packets += entry->num_of_packet;
entry            2801 drivers/net/usb/lan78xx.c 		dev->net->stats.tx_bytes += entry->length;
entry            2822 drivers/net/usb/lan78xx.c 				  "tx err %d\n", entry->urb->status);
entry            2835 drivers/net/usb/lan78xx.c 	struct skb_data *entry = (struct skb_data *)newsk->cb;
entry            2838 drivers/net/usb/lan78xx.c 	entry->state = state;
entry            3190 drivers/net/usb/lan78xx.c 	struct skb_data *entry;
entry            3201 drivers/net/usb/lan78xx.c 	entry = (struct skb_data *)skb->cb;
entry            3202 drivers/net/usb/lan78xx.c 	entry->urb = urb;
entry            3203 drivers/net/usb/lan78xx.c 	entry->dev = dev;
entry            3204 drivers/net/usb/lan78xx.c 	entry->length = 0;
entry            3250 drivers/net/usb/lan78xx.c 	struct skb_data	*entry = (struct skb_data *)skb->cb;
entry            3251 drivers/net/usb/lan78xx.c 	struct lan78xx_net *dev = entry->dev;
entry            3257 drivers/net/usb/lan78xx.c 	entry->urb = NULL;
entry            3279 drivers/net/usb/lan78xx.c 		entry->urb = urb;
entry            3287 drivers/net/usb/lan78xx.c 		entry->urb = urb;
entry            3321 drivers/net/usb/lan78xx.c 	struct skb_data *entry;
entry            3376 drivers/net/usb/lan78xx.c 	entry = (struct skb_data *)skb->cb;
entry            3377 drivers/net/usb/lan78xx.c 	entry->urb = urb;
entry            3378 drivers/net/usb/lan78xx.c 	entry->dev = dev;
entry            3379 drivers/net/usb/lan78xx.c 	entry->length = length;
entry            3380 drivers/net/usb/lan78xx.c 	entry->num_of_packet = count;
entry            3471 drivers/net/usb/lan78xx.c 	struct skb_data *entry;
entry            3474 drivers/net/usb/lan78xx.c 		entry = (struct skb_data *)(skb->cb);
entry            3475 drivers/net/usb/lan78xx.c 		switch (entry->state) {
entry            3477 drivers/net/usb/lan78xx.c 			entry->state = rx_cleanup;
entry            3481 drivers/net/usb/lan78xx.c 			usb_free_urb(entry->urb);
entry            3485 drivers/net/usb/lan78xx.c 			usb_free_urb(entry->urb);
entry            3489 drivers/net/usb/lan78xx.c 			netdev_dbg(dev->net, "skb state %d\n", entry->state);
entry             411 drivers/net/usb/usbnet.c 	struct skb_data *entry = (struct skb_data *) newsk->cb;
entry             414 drivers/net/usb/usbnet.c 	entry->state = state;
entry             428 drivers/net/usb/usbnet.c 	struct skb_data *entry = (struct skb_data *) skb->cb;
entry             431 drivers/net/usb/usbnet.c 	old_state = entry->state;
entry             432 drivers/net/usb/usbnet.c 	entry->state = state;
entry             471 drivers/net/usb/usbnet.c 	struct skb_data		*entry;
entry             493 drivers/net/usb/usbnet.c 	entry = (struct skb_data *) skb->cb;
entry             494 drivers/net/usb/usbnet.c 	entry->urb = urb;
entry             495 drivers/net/usb/usbnet.c 	entry->dev = dev;
entry             496 drivers/net/usb/usbnet.c 	entry->length = 0;
entry             578 drivers/net/usb/usbnet.c 	struct skb_data		*entry = (struct skb_data *) skb->cb;
entry             579 drivers/net/usb/usbnet.c 	struct usbnet		*dev = entry->dev;
entry             585 drivers/net/usb/usbnet.c 	entry->urb = NULL;
entry             624 drivers/net/usb/usbnet.c 		entry->urb = urb;
entry             712 drivers/net/usb/usbnet.c 		struct skb_data		*entry;
entry             717 drivers/net/usb/usbnet.c 			entry = (struct skb_data *) skb->cb;
entry             718 drivers/net/usb/usbnet.c 			if (entry->state != unlink_start)
entry             723 drivers/net/usb/usbnet.c 		entry->state = unlink_start;
entry             724 drivers/net/usb/usbnet.c 		urb = entry->urb;
entry            1244 drivers/net/usb/usbnet.c 	struct skb_data		*entry = (struct skb_data *) skb->cb;
entry            1245 drivers/net/usb/usbnet.c 	struct usbnet		*dev = entry->dev;
entry            1252 drivers/net/usb/usbnet.c 		stats64->tx_packets += entry->packets;
entry            1253 drivers/net/usb/usbnet.c 		stats64->tx_bytes += entry->length;
entry            1285 drivers/net/usb/usbnet.c 				  "tx err %d\n", entry->urb->status);
entry            1352 drivers/net/usb/usbnet.c 	struct skb_data		*entry;
entry            1378 drivers/net/usb/usbnet.c 	entry = (struct skb_data *) skb->cb;
entry            1379 drivers/net/usb/usbnet.c 	entry->urb = urb;
entry            1380 drivers/net/usb/usbnet.c 	entry->dev = dev;
entry            1418 drivers/net/usb/usbnet.c 		entry->length += length;
entry            1419 drivers/net/usb/usbnet.c 		if (WARN_ON_ONCE(entry->length <= 0))
entry            1420 drivers/net/usb/usbnet.c 			entry->length = length;
entry            1521 drivers/net/usb/usbnet.c 	struct skb_data		*entry;
entry            1524 drivers/net/usb/usbnet.c 		entry = (struct skb_data *) skb->cb;
entry            1525 drivers/net/usb/usbnet.c 		switch (entry->state) {
entry            1527 drivers/net/usb/usbnet.c 			entry->state = rx_cleanup;
entry            1531 drivers/net/usb/usbnet.c 			kfree(entry->urb->sg);
entry            1534 drivers/net/usb/usbnet.c 			usb_free_urb (entry->urb);
entry            1538 drivers/net/usb/usbnet.c 			netdev_dbg(dev->net, "bogus skb state %d\n", entry->state);
entry            3138 drivers/net/vmxnet3/vmxnet3_drv.c 			adapter->intr.msix_entries[i].entry = i;
entry             426 drivers/net/wan/lapbether.c 	struct list_head *entry, *tmp;
entry             432 drivers/net/wan/lapbether.c 	list_for_each_safe(entry, tmp, &lapbeth_devices) {
entry             433 drivers/net/wan/lapbether.c 		lapbeth = list_entry(entry, struct lapbethdev, node);
entry            1401 drivers/net/wan/lmc/lmc_main.c     int entry;
entry            1410 drivers/net/wan/lmc/lmc_main.c     entry = sc->lmc_next_tx % LMC_TXDESCS;
entry            1412 drivers/net/wan/lmc/lmc_main.c     sc->lmc_txq[entry] = skb;
entry            1413 drivers/net/wan/lmc/lmc_main.c     sc->lmc_txring[entry].buffer1 = virt_to_bus (skb->data);
entry            1452 drivers/net/wan/lmc/lmc_main.c         LMC_EVENT_LOG(LMC_EVENT_TBUSY1, entry, 0);
entry            1457 drivers/net/wan/lmc/lmc_main.c     if (entry == LMC_TXDESCS - 1)	/* last descriptor in ring */
entry            1461 drivers/net/wan/lmc/lmc_main.c     flag = sc->lmc_txring[entry].length = (skb->len) | flag |
entry            1472 drivers/net/wan/lmc/lmc_main.c     LMC_EVENT_LOG(LMC_EVENT_XMT, flag, entry);
entry            1473 drivers/net/wan/lmc/lmc_main.c     sc->lmc_txring[entry].status = 0x80000000;
entry             546 drivers/net/wimax/i2400m/rx.c 	} entry[I2400M_ROQ_LOG_LENGTH];
entry             601 drivers/net/wimax/i2400m/rx.c 	e = &roq->log->entry[cnt_idx];
entry             626 drivers/net/wimax/i2400m/rx.c 		e = &roq->log->entry[cnt_idx];
entry             311 drivers/net/wireless/admtek/adm8211.c 		unsigned int entry = dirty_tx % priv->tx_ring_size;
entry             312 drivers/net/wireless/admtek/adm8211.c 		u32 status = le32_to_cpu(priv->tx_ring[entry].status);
entry             321 drivers/net/wireless/admtek/adm8211.c 		info = &priv->tx_buffers[entry];
entry             354 drivers/net/wireless/admtek/adm8211.c 	unsigned int entry = priv->cur_rx % priv->rx_ring_size;
entry             361 drivers/net/wireless/admtek/adm8211.c 	while (!(priv->rx_ring[entry].status & cpu_to_le32(RDES0_STATUS_OWN))) {
entry             365 drivers/net/wireless/admtek/adm8211.c 		status = le32_to_cpu(priv->rx_ring[entry].status);
entry             367 drivers/net/wireless/admtek/adm8211.c 		rssi = le32_to_cpu(priv->rx_ring[entry].length) &
entry             387 drivers/net/wireless/admtek/adm8211.c 					priv->rx_buffers[entry].mapping,
entry             390 drivers/net/wireless/admtek/adm8211.c 					     skb_tail_pointer(priv->rx_buffers[entry].skb),
entry             394 drivers/net/wireless/admtek/adm8211.c 					priv->rx_buffers[entry].mapping,
entry             400 drivers/net/wireless/admtek/adm8211.c 				skb = priv->rx_buffers[entry].skb;
entry             404 drivers/net/wireless/admtek/adm8211.c 					priv->rx_buffers[entry].mapping,
entry             406 drivers/net/wireless/admtek/adm8211.c 				priv->rx_buffers[entry].skb = newskb;
entry             407 drivers/net/wireless/admtek/adm8211.c 				priv->rx_buffers[entry].mapping =
entry             413 drivers/net/wireless/admtek/adm8211.c 					   priv->rx_buffers[entry].mapping)) {
entry             414 drivers/net/wireless/admtek/adm8211.c 					priv->rx_buffers[entry].skb = NULL;
entry             424 drivers/net/wireless/admtek/adm8211.c 			priv->rx_ring[entry].buffer1 =
entry             425 drivers/net/wireless/admtek/adm8211.c 				cpu_to_le32(priv->rx_buffers[entry].mapping);
entry             428 drivers/net/wireless/admtek/adm8211.c 		priv->rx_ring[entry].status = cpu_to_le32(RDES0_STATUS_OWN |
entry             430 drivers/net/wireless/admtek/adm8211.c 		priv->rx_ring[entry].length =
entry             432 drivers/net/wireless/admtek/adm8211.c 				    (entry == priv->rx_ring_size - 1 ?
entry             452 drivers/net/wireless/admtek/adm8211.c 		entry = (++priv->cur_rx) % priv->rx_ring_size;
entry            1632 drivers/net/wireless/admtek/adm8211.c 	unsigned int entry;
entry            1650 drivers/net/wireless/admtek/adm8211.c 	entry = priv->cur_tx % priv->tx_ring_size;
entry            1652 drivers/net/wireless/admtek/adm8211.c 	priv->tx_buffers[entry].skb = skb;
entry            1653 drivers/net/wireless/admtek/adm8211.c 	priv->tx_buffers[entry].mapping = mapping;
entry            1654 drivers/net/wireless/admtek/adm8211.c 	priv->tx_buffers[entry].hdrlen = hdrlen;
entry            1655 drivers/net/wireless/admtek/adm8211.c 	priv->tx_ring[entry].buffer1 = cpu_to_le32(mapping);
entry            1657 drivers/net/wireless/admtek/adm8211.c 	if (entry == priv->tx_ring_size - 1)
entry            1659 drivers/net/wireless/admtek/adm8211.c 	priv->tx_ring[entry].length = cpu_to_le32(flag | skb->len);
entry            1663 drivers/net/wireless/admtek/adm8211.c 	priv->tx_ring[entry].status = cpu_to_le32(flag);
entry             205 drivers/net/wireless/ath/ath.h bool ath_hw_keyreset(struct ath_common *common, u16 entry);
entry            1823 drivers/net/wireless/ath/ath10k/pci.c 	const struct service_to_pipe *entry;
entry            1830 drivers/net/wireless/ath/ath10k/pci.c 		entry = &target_service_to_ce_map_wlan[i];
entry            1832 drivers/net/wireless/ath/ath10k/pci.c 		if (__le32_to_cpu(entry->service_id) != service_id)
entry            1835 drivers/net/wireless/ath/ath10k/pci.c 		switch (__le32_to_cpu(entry->pipedir)) {
entry            1840 drivers/net/wireless/ath/ath10k/pci.c 			*dl_pipe = __le32_to_cpu(entry->pipenum);
entry            1845 drivers/net/wireless/ath/ath10k/pci.c 			*ul_pipe = __le32_to_cpu(entry->pipenum);
entry            1851 drivers/net/wireless/ath/ath10k/pci.c 			*dl_pipe = __le32_to_cpu(entry->pipenum);
entry            1852 drivers/net/wireless/ath/ath10k/pci.c 			*ul_pipe = __le32_to_cpu(entry->pipenum);
entry             770 drivers/net/wireless/ath/ath10k/snoc.c 	const struct service_to_pipe *entry;
entry             777 drivers/net/wireless/ath/ath10k/snoc.c 		entry = &target_service_to_ce_map_wlan[i];
entry             779 drivers/net/wireless/ath/ath10k/snoc.c 		if (__le32_to_cpu(entry->service_id) != service_id)
entry             782 drivers/net/wireless/ath/ath10k/snoc.c 		switch (__le32_to_cpu(entry->pipedir)) {
entry             787 drivers/net/wireless/ath/ath10k/snoc.c 			*dl_pipe = __le32_to_cpu(entry->pipenum);
entry             792 drivers/net/wireless/ath/ath10k/snoc.c 			*ul_pipe = __le32_to_cpu(entry->pipenum);
entry             798 drivers/net/wireless/ath/ath10k/snoc.c 			*dl_pipe = __le32_to_cpu(entry->pipenum);
entry             799 drivers/net/wireless/ath/ath10k/snoc.c 			*ul_pipe = __le32_to_cpu(entry->pipenum);
entry             198 drivers/net/wireless/ath/ath5k/phy.c 	u16 entry;
entry             237 drivers/net/wireless/ath/ath5k/phy.c 	entry = ((first_bit - 1) / 8) + offset;
entry             244 drivers/net/wireless/ath/ath5k/phy.c 	     position = 0, entry++) {
entry             253 drivers/net/wireless/ath/ath5k/phy.c 			rfb[entry] &= ~mask;
entry             254 drivers/net/wireless/ath/ath5k/phy.c 			rfb[entry] |= ((data << position) << (col * 8)) & mask;
entry             257 drivers/net/wireless/ath/ath5k/phy.c 			data |= (((rfb[entry] & mask) >> (col * 8)) >> position)
entry             551 drivers/net/wireless/ath/ath6kl/htc_pipe.c 	struct htc_pipe_txcredit_alloc *entry;
entry             561 drivers/net/wireless/ath/ath6kl/htc_pipe.c 	entry = &target->pipe.txcredit_alloc[0];
entry             570 drivers/net/wireless/ath/ath6kl/htc_pipe.c 		entry++;
entry             571 drivers/net/wireless/ath/ath6kl/htc_pipe.c 		entry++;
entry             573 drivers/net/wireless/ath/ath6kl/htc_pipe.c 		entry->service_id = WMI_DATA_VO_SVC;
entry             574 drivers/net/wireless/ath/ath6kl/htc_pipe.c 		entry->credit_alloc = (credits - 6);
entry             575 drivers/net/wireless/ath/ath6kl/htc_pipe.c 		if (entry->credit_alloc == 0)
entry             576 drivers/net/wireless/ath/ath6kl/htc_pipe.c 			entry->credit_alloc++;
entry             578 drivers/net/wireless/ath/ath6kl/htc_pipe.c 		credits -= (int) entry->credit_alloc;
entry             582 drivers/net/wireless/ath/ath6kl/htc_pipe.c 		entry++;
entry             583 drivers/net/wireless/ath/ath6kl/htc_pipe.c 		entry->service_id = WMI_CONTROL_SVC;
entry             584 drivers/net/wireless/ath/ath6kl/htc_pipe.c 		entry->credit_alloc = credit_per_maxmsg;
entry             585 drivers/net/wireless/ath/ath6kl/htc_pipe.c 		credits -= (int) entry->credit_alloc;
entry             590 drivers/net/wireless/ath/ath6kl/htc_pipe.c 		entry++;
entry             591 drivers/net/wireless/ath/ath6kl/htc_pipe.c 		entry++;
entry             592 drivers/net/wireless/ath/ath6kl/htc_pipe.c 		entry->service_id = WMI_DATA_BE_SVC;
entry             593 drivers/net/wireless/ath/ath6kl/htc_pipe.c 		entry->credit_alloc = (u8) credits;
entry             596 drivers/net/wireless/ath/ath6kl/htc_pipe.c 		entry++;
entry             597 drivers/net/wireless/ath/ath6kl/htc_pipe.c 		entry->service_id = WMI_DATA_VI_SVC;
entry             598 drivers/net/wireless/ath/ath6kl/htc_pipe.c 		entry->credit_alloc = credits / 4;
entry             599 drivers/net/wireless/ath/ath6kl/htc_pipe.c 		if (entry->credit_alloc == 0)
entry             600 drivers/net/wireless/ath/ath6kl/htc_pipe.c 			entry->credit_alloc++;
entry             602 drivers/net/wireless/ath/ath6kl/htc_pipe.c 		credits -= (int) entry->credit_alloc;
entry             606 drivers/net/wireless/ath/ath6kl/htc_pipe.c 		entry++;
entry             607 drivers/net/wireless/ath/ath6kl/htc_pipe.c 		entry->service_id = WMI_DATA_VO_SVC;
entry             608 drivers/net/wireless/ath/ath6kl/htc_pipe.c 		entry->credit_alloc = credits / 4;
entry             609 drivers/net/wireless/ath/ath6kl/htc_pipe.c 		if (entry->credit_alloc == 0)
entry             610 drivers/net/wireless/ath/ath6kl/htc_pipe.c 			entry->credit_alloc++;
entry             612 drivers/net/wireless/ath/ath6kl/htc_pipe.c 		credits -= (int) entry->credit_alloc;
entry             616 drivers/net/wireless/ath/ath6kl/htc_pipe.c 		entry++;
entry             617 drivers/net/wireless/ath/ath6kl/htc_pipe.c 		entry->service_id = WMI_CONTROL_SVC;
entry             618 drivers/net/wireless/ath/ath6kl/htc_pipe.c 		entry->credit_alloc = credit_per_maxmsg;
entry             619 drivers/net/wireless/ath/ath6kl/htc_pipe.c 		credits -= (int) entry->credit_alloc;
entry             623 drivers/net/wireless/ath/ath6kl/htc_pipe.c 		entry++;
entry             624 drivers/net/wireless/ath/ath6kl/htc_pipe.c 		entry->service_id = WMI_DATA_BK_SVC;
entry             625 drivers/net/wireless/ath/ath6kl/htc_pipe.c 		entry->credit_alloc = credit_per_maxmsg;
entry             626 drivers/net/wireless/ath/ath6kl/htc_pipe.c 		credits -= (int) entry->credit_alloc;
entry             631 drivers/net/wireless/ath/ath6kl/htc_pipe.c 		entry++;
entry             632 drivers/net/wireless/ath/ath6kl/htc_pipe.c 		entry->service_id = WMI_DATA_BE_SVC;
entry             633 drivers/net/wireless/ath/ath6kl/htc_pipe.c 		entry->credit_alloc = (u8) credits;
entry              89 drivers/net/wireless/ath/ath6kl/main.c 	struct ath6kl_mgmt_buff *entry, *tmp;
entry              97 drivers/net/wireless/ath/ath6kl/main.c 		list_for_each_entry_safe(entry, tmp, &sta->mgmt_psq, list) {
entry              98 drivers/net/wireless/ath/ath6kl/main.c 			kfree(entry);
entry             293 drivers/net/wireless/ath/ath9k/ar9003_paprd.c 	u32 *entry = ah->paprd_gain_table_entries;
entry             299 drivers/net/wireless/ath/ath9k/ar9003_paprd.c 		entry[i] = REG_READ(ah, reg);
entry             300 drivers/net/wireless/ath/ath9k/ar9003_paprd.c 		index[i] = (entry[i] >> 24) & 0xff;
entry              29 drivers/net/wireless/ath/ath9k/mci.c 	struct ath_mci_profile_info *entry;
entry              34 drivers/net/wireless/ath/ath9k/mci.c 	list_for_each_entry(entry, &mci->info, list) {
entry              35 drivers/net/wireless/ath/ath9k/mci.c 		if (entry->conn_handle == info->conn_handle)
entry              36 drivers/net/wireless/ath/ath9k/mci.c 			return entry;
entry              45 drivers/net/wireless/ath/ath9k/mci.c 	struct ath_mci_profile_info *entry;
entry              56 drivers/net/wireless/ath/ath9k/mci.c 	entry = kzalloc(sizeof(*entry), GFP_ATOMIC);
entry              57 drivers/net/wireless/ath/ath9k/mci.c 	if (!entry)
entry              60 drivers/net/wireless/ath/ath9k/mci.c 	memcpy(entry, info, 10);
entry              62 drivers/net/wireless/ath/ath9k/mci.c 	list_add_tail(&entry->list, &mci->info);
entry              75 drivers/net/wireless/ath/ath9k/mci.c 				struct ath_mci_profile_info *entry)
entry              77 drivers/net/wireless/ath/ath9k/mci.c 	if (!entry)
entry              80 drivers/net/wireless/ath/ath9k/mci.c 	DEC_PROF(mci, entry);
entry              81 drivers/net/wireless/ath/ath9k/mci.c 	list_del(&entry->list);
entry              82 drivers/net/wireless/ath/ath9k/mci.c 	kfree(entry);
entry             301 drivers/net/wireless/ath/ath9k/mci.c 	struct ath_mci_profile_info *entry = NULL;
entry             303 drivers/net/wireless/ath/ath9k/mci.c 	entry = ath_mci_find_profile(mci, info);
entry             304 drivers/net/wireless/ath/ath9k/mci.c 	if (entry) {
entry             313 drivers/net/wireless/ath/ath9k/mci.c 		if (entry->type != info->type) {
entry             314 drivers/net/wireless/ath/ath9k/mci.c 			DEC_PROF(mci, entry);
entry             317 drivers/net/wireless/ath/ath9k/mci.c 		memcpy(entry, info, 10);
entry             321 drivers/net/wireless/ath/ath9k/mci.c 		if (!entry && !ath_mci_add_profile(common, mci, info))
entry             324 drivers/net/wireless/ath/ath9k/mci.c 		ath_mci_del_profile(common, mci, entry);
entry             576 drivers/net/wireless/ath/carl9170/rx.c 	struct carl9170_bar_list_entry *entry;
entry             590 drivers/net/wireless/ath/carl9170/rx.c 	list_for_each_entry_rcu(entry, &ar->bar_list[queue], list) {
entry             591 drivers/net/wireless/ath/carl9170/rx.c 		struct sk_buff *entry_skb = entry->skb;
entry             609 drivers/net/wireless/ath/carl9170/rx.c 			list_del_rcu(&entry->list);
entry             611 drivers/net/wireless/ath/carl9170/rx.c 			kfree_rcu(entry, head);
entry             454 drivers/net/wireless/ath/carl9170/tx.c 		struct carl9170_bar_list_entry *entry;
entry             458 drivers/net/wireless/ath/carl9170/tx.c 		list_for_each_entry_rcu(entry, &ar->bar_list[queue], list) {
entry             459 drivers/net/wireless/ath/carl9170/tx.c 			if (entry->skb == skb) {
entry             461 drivers/net/wireless/ath/carl9170/tx.c 				list_del_rcu(&entry->list);
entry             463 drivers/net/wireless/ath/carl9170/tx.c 				kfree_rcu(entry, head);
entry            1326 drivers/net/wireless/ath/carl9170/tx.c 		struct carl9170_bar_list_entry *entry;
entry            1329 drivers/net/wireless/ath/carl9170/tx.c 		entry = kmalloc(sizeof(*entry), GFP_ATOMIC);
entry            1330 drivers/net/wireless/ath/carl9170/tx.c 		if (!WARN_ON_ONCE(!entry)) {
entry            1331 drivers/net/wireless/ath/carl9170/tx.c 			entry->skb = skb;
entry            1333 drivers/net/wireless/ath/carl9170/tx.c 			list_add_tail_rcu(&entry->list, &ar->bar_list[queue]);
entry              42 drivers/net/wireless/ath/key.c bool ath_hw_keyreset(struct ath_common *common, u16 entry)
entry              47 drivers/net/wireless/ath/key.c 	if (entry >= common->keymax) {
entry              49 drivers/net/wireless/ath/key.c 			entry);
entry              53 drivers/net/wireless/ath/key.c 	keyType = REG_READ(ah, AR_KEYTABLE_TYPE(entry));
entry              57 drivers/net/wireless/ath/key.c 	REG_WRITE(ah, AR_KEYTABLE_KEY0(entry), 0);
entry              58 drivers/net/wireless/ath/key.c 	REG_WRITE(ah, AR_KEYTABLE_KEY1(entry), 0);
entry              59 drivers/net/wireless/ath/key.c 	REG_WRITE(ah, AR_KEYTABLE_KEY2(entry), 0);
entry              60 drivers/net/wireless/ath/key.c 	REG_WRITE(ah, AR_KEYTABLE_KEY3(entry), 0);
entry              61 drivers/net/wireless/ath/key.c 	REG_WRITE(ah, AR_KEYTABLE_KEY4(entry), 0);
entry              62 drivers/net/wireless/ath/key.c 	REG_WRITE(ah, AR_KEYTABLE_TYPE(entry), AR_KEYTABLE_TYPE_CLR);
entry              63 drivers/net/wireless/ath/key.c 	REG_WRITE(ah, AR_KEYTABLE_MAC0(entry), 0);
entry              64 drivers/net/wireless/ath/key.c 	REG_WRITE(ah, AR_KEYTABLE_MAC1(entry), 0);
entry              67 drivers/net/wireless/ath/key.c 		u16 micentry = entry + 64;
entry              88 drivers/net/wireless/ath/key.c 			     u16 entry, const u8 *mac)
entry              94 drivers/net/wireless/ath/key.c 	if (entry >= common->keymax) {
entry              96 drivers/net/wireless/ath/key.c 			entry);
entry             121 drivers/net/wireless/ath/key.c 	REG_WRITE(ah, AR_KEYTABLE_MAC0(entry), macLo);
entry             122 drivers/net/wireless/ath/key.c 	REG_WRITE(ah, AR_KEYTABLE_MAC1(entry), macHi | unicast_flag);
entry             129 drivers/net/wireless/ath/key.c static bool ath_hw_set_keycache_entry(struct ath_common *common, u16 entry,
entry             137 drivers/net/wireless/ath/key.c 	if (entry >= common->keymax) {
entry             139 drivers/net/wireless/ath/key.c 			entry);
entry             157 drivers/net/wireless/ath/key.c 		if (entry + 64 >= common->keymax) {
entry             159 drivers/net/wireless/ath/key.c 				"entry %u inappropriate for TKIP\n", entry);
entry             200 drivers/net/wireless/ath/key.c 		u16 micentry = entry + 64;
entry             208 drivers/net/wireless/ath/key.c 		REG_WRITE(ah, AR_KEYTABLE_KEY0(entry), ~key0);
entry             209 drivers/net/wireless/ath/key.c 		REG_WRITE(ah, AR_KEYTABLE_KEY1(entry), ~key1);
entry             212 drivers/net/wireless/ath/key.c 		REG_WRITE(ah, AR_KEYTABLE_KEY2(entry), key2);
entry             213 drivers/net/wireless/ath/key.c 		REG_WRITE(ah, AR_KEYTABLE_KEY3(entry), key3);
entry             216 drivers/net/wireless/ath/key.c 		REG_WRITE(ah, AR_KEYTABLE_KEY4(entry), key4);
entry             217 drivers/net/wireless/ath/key.c 		REG_WRITE(ah, AR_KEYTABLE_TYPE(entry), keyType);
entry             220 drivers/net/wireless/ath/key.c 		(void) ath_hw_keysetmac(common, entry, mac);
entry             311 drivers/net/wireless/ath/key.c 		REG_WRITE(ah, AR_KEYTABLE_KEY0(entry), key0);
entry             312 drivers/net/wireless/ath/key.c 		REG_WRITE(ah, AR_KEYTABLE_KEY1(entry), key1);
entry             319 drivers/net/wireless/ath/key.c 		REG_WRITE(ah, AR_KEYTABLE_KEY0(entry), key0);
entry             320 drivers/net/wireless/ath/key.c 		REG_WRITE(ah, AR_KEYTABLE_KEY1(entry), key1);
entry             323 drivers/net/wireless/ath/key.c 		REG_WRITE(ah, AR_KEYTABLE_KEY2(entry), key2);
entry             324 drivers/net/wireless/ath/key.c 		REG_WRITE(ah, AR_KEYTABLE_KEY3(entry), key3);
entry             327 drivers/net/wireless/ath/key.c 		REG_WRITE(ah, AR_KEYTABLE_KEY4(entry), key4);
entry             328 drivers/net/wireless/ath/key.c 		REG_WRITE(ah, AR_KEYTABLE_TYPE(entry), keyType);
entry             333 drivers/net/wireless/ath/key.c 		(void) ath_hw_keysetmac(common, entry, mac);
entry              84 drivers/net/wireless/ath/wcn36xx/smd.c 	struct wcn36xx_hal_cfg *entry;
entry              87 drivers/net/wireless/ath/wcn36xx/smd.c 	if (*len + sizeof(*entry) + sizeof(u32) >= WCN36XX_HAL_BUF_SIZE) {
entry              92 drivers/net/wireless/ath/wcn36xx/smd.c 	entry = (struct wcn36xx_hal_cfg *) (wcn->hal_buf + *len);
entry              93 drivers/net/wireless/ath/wcn36xx/smd.c 	entry->id = id;
entry              94 drivers/net/wireless/ath/wcn36xx/smd.c 	entry->len = sizeof(u32);
entry              95 drivers/net/wireless/ath/wcn36xx/smd.c 	entry->pad_bytes = 0;
entry              96 drivers/net/wireless/ath/wcn36xx/smd.c 	entry->reserve = 0;
entry              98 drivers/net/wireless/ath/wcn36xx/smd.c 	val = (u32 *) (entry + 1);
entry             101 drivers/net/wireless/ath/wcn36xx/smd.c 	*len += sizeof(*entry) + sizeof(u32);
entry            2333 drivers/net/wireless/broadcom/b43/phy_g.c static s8 b43_tssi2dbm_entry(s8 entry[], u8 index,
entry            2350 drivers/net/wireless/broadcom/b43/phy_g.c 	entry[index] = clamp_val(b43_tssi2dbm_ad(m1 * f, 8192), -127, 128);
entry            4929 drivers/net/wireless/broadcom/b43/phy_n.c 	u16 scale, entry;
entry            4938 drivers/net/wireless/broadcom/b43/phy_n.c 		entry = ((scale & 0xFF) << 8) | ladder_lo[i].g_env;
entry            4939 drivers/net/wireless/broadcom/b43/phy_n.c 		b43_ntab_write(dev, B43_NTAB16(15, i), entry);
entry            4942 drivers/net/wireless/broadcom/b43/phy_n.c 		entry = ((scale & 0xFF) << 8) | ladder_iq[i].g_env;
entry            4943 drivers/net/wireless/broadcom/b43/phy_n.c 		b43_ntab_write(dev, B43_NTAB16(15, i + 32), entry);
entry              12 drivers/net/wireless/broadcom/b43/ppr.c #define ppr_for_each_entry(ppr, i, entry)				\
entry              13 drivers/net/wireless/broadcom/b43/ppr.c 	for (i = 0, entry = &(ppr)->__all_rates[i];			\
entry              15 drivers/net/wireless/broadcom/b43/ppr.c 	     i++, entry++)
entry            1942 drivers/net/wireless/broadcom/b43legacy/phy.c s8 b43legacy_tssi2dbm_entry(s8 entry [], u8 index, s16 pab0, s16 pab1, s16 pab2)
entry            1963 drivers/net/wireless/broadcom/b43legacy/phy.c 	entry[index] = clamp_val(b43legacy_tssi2dbm_ad(m1 * f, 8192),
entry              56 drivers/net/wireless/broadcom/brcm80211/brcmfmac/firmware.c 	u32 entry;
entry              95 drivers/net/wireless/broadcom/brcm80211/brcmfmac/firmware.c 		nvp->entry = nvp->pos;
entry             114 drivers/net/wireless/broadcom/brcm80211/brcmfmac/firmware.c 		if (strncmp(&nvp->data[nvp->entry], "RAW1", 4) == 0)
entry             118 drivers/net/wireless/broadcom/brcm80211/brcmfmac/firmware.c 		if (strncmp(&nvp->data[nvp->entry], "devpath", 7) == 0)
entry             120 drivers/net/wireless/broadcom/brcm80211/brcmfmac/firmware.c 		if (strncmp(&nvp->data[nvp->entry], "pcie/", 5) == 0)
entry             122 drivers/net/wireless/broadcom/brcm80211/brcmfmac/firmware.c 		if (strncmp(&nvp->data[nvp->entry], "boardrev", 8) == 0)
entry             147 drivers/net/wireless/broadcom/brcm80211/brcmfmac/firmware.c 		skv = (u8 *)&nvp->data[nvp->entry];
entry             748 drivers/net/wireless/broadcom/brcm80211/brcmfmac/fwsignal.c 	struct brcmf_fws_mac_descriptor *entry;
entry             754 drivers/net/wireless/broadcom/brcm80211/brcmfmac/fwsignal.c 	entry = &fws->desc.nodes[0];
entry             756 drivers/net/wireless/broadcom/brcm80211/brcmfmac/fwsignal.c 		if (entry->occupied && !memcmp(entry->ea, ea, ETH_ALEN))
entry             757 drivers/net/wireless/broadcom/brcm80211/brcmfmac/fwsignal.c 			return entry;
entry             758 drivers/net/wireless/broadcom/brcm80211/brcmfmac/fwsignal.c 		entry++;
entry             767 drivers/net/wireless/broadcom/brcm80211/brcmfmac/fwsignal.c 	struct brcmf_fws_mac_descriptor *entry = &fws->desc.other;
entry             777 drivers/net/wireless/broadcom/brcm80211/brcmfmac/fwsignal.c 		entry = ifp->fws_desc;
entry             781 drivers/net/wireless/broadcom/brcm80211/brcmfmac/fwsignal.c 	entry = brcmf_fws_macdesc_lookup(fws, da);
entry             782 drivers/net/wireless/broadcom/brcm80211/brcmfmac/fwsignal.c 	if (IS_ERR(entry))
entry             783 drivers/net/wireless/broadcom/brcm80211/brcmfmac/fwsignal.c 		entry = ifp->fws_desc;
entry             786 drivers/net/wireless/broadcom/brcm80211/brcmfmac/fwsignal.c 	return entry;
entry             790 drivers/net/wireless/broadcom/brcm80211/brcmfmac/fwsignal.c 				     struct brcmf_fws_mac_descriptor *entry,
entry             799 drivers/net/wireless/broadcom/brcm80211/brcmfmac/fwsignal.c 	if (entry->mac_handle) {
entry             800 drivers/net/wireless/broadcom/brcm80211/brcmfmac/fwsignal.c 		if_entry = &fws->desc.iface[entry->interface_id];
entry             807 drivers/net/wireless/broadcom/brcm80211/brcmfmac/fwsignal.c 	closed = entry->state == BRCMF_FWS_STATE_CLOSE &&
entry             808 drivers/net/wireless/broadcom/brcm80211/brcmfmac/fwsignal.c 		 !entry->requested_credit && !entry->requested_packet;
entry             811 drivers/net/wireless/broadcom/brcm80211/brcmfmac/fwsignal.c 	return closed || !(entry->ac_bitmap & BIT(fifo));
entry             815 drivers/net/wireless/broadcom/brcm80211/brcmfmac/fwsignal.c 				      struct brcmf_fws_mac_descriptor *entry,
entry             818 drivers/net/wireless/broadcom/brcm80211/brcmfmac/fwsignal.c 	if (entry->occupied && (ifidx == -1 || ifidx == entry->interface_id)) {
entry             819 drivers/net/wireless/broadcom/brcm80211/brcmfmac/fwsignal.c 		brcmf_fws_psq_flush(fws, &entry->psq, ifidx);
entry             820 drivers/net/wireless/broadcom/brcm80211/brcmfmac/fwsignal.c 		entry->occupied = !!(entry->psq.len);
entry             877 drivers/net/wireless/broadcom/brcm80211/brcmfmac/fwsignal.c 	struct brcmf_fws_mac_descriptor *entry = brcmf_skbcb(skb)->mac;
entry             885 drivers/net/wireless/broadcom/brcm80211/brcmfmac/fwsignal.c 		  entry->name, brcmf_skb_if_flags_get_field(skb, INDEX),
entry             888 drivers/net/wireless/broadcom/brcm80211/brcmfmac/fwsignal.c 	if (entry->send_tim_signal)
entry             910 drivers/net/wireless/broadcom/brcm80211/brcmfmac/fwsignal.c 	if (entry->send_tim_signal) {
entry             911 drivers/net/wireless/broadcom/brcm80211/brcmfmac/fwsignal.c 		entry->send_tim_signal = 0;
entry             914 drivers/net/wireless/broadcom/brcm80211/brcmfmac/fwsignal.c 		wlh[2] = entry->mac_handle;
entry             915 drivers/net/wireless/broadcom/brcm80211/brcmfmac/fwsignal.c 		wlh[3] = entry->traffic_pending_bmp;
entry             917 drivers/net/wireless/broadcom/brcm80211/brcmfmac/fwsignal.c 			  entry->mac_handle, entry->traffic_pending_bmp);
entry             919 drivers/net/wireless/broadcom/brcm80211/brcmfmac/fwsignal.c 		entry->traffic_lastreported_bmp = entry->traffic_pending_bmp;
entry             928 drivers/net/wireless/broadcom/brcm80211/brcmfmac/fwsignal.c 				 struct brcmf_fws_mac_descriptor *entry,
entry             939 drivers/net/wireless/broadcom/brcm80211/brcmfmac/fwsignal.c 	if (brcmu_pktq_mlen(&entry->psq, 3 << (fifo * 2)) == 0)
entry             940 drivers/net/wireless/broadcom/brcm80211/brcmfmac/fwsignal.c 		entry->traffic_pending_bmp &= ~NBITVAL(fifo);
entry             942 drivers/net/wireless/broadcom/brcm80211/brcmfmac/fwsignal.c 		entry->traffic_pending_bmp |= NBITVAL(fifo);
entry             944 drivers/net/wireless/broadcom/brcm80211/brcmfmac/fwsignal.c 	entry->send_tim_signal = false;
entry             945 drivers/net/wireless/broadcom/brcm80211/brcmfmac/fwsignal.c 	if (entry->traffic_lastreported_bmp != entry->traffic_pending_bmp)
entry             946 drivers/net/wireless/broadcom/brcm80211/brcmfmac/fwsignal.c 		entry->send_tim_signal = true;
entry             947 drivers/net/wireless/broadcom/brcm80211/brcmfmac/fwsignal.c 	if (send_immediately && entry->send_tim_signal &&
entry             948 drivers/net/wireless/broadcom/brcm80211/brcmfmac/fwsignal.c 	    entry->state == BRCMF_FWS_STATE_CLOSE) {
entry             960 drivers/net/wireless/broadcom/brcm80211/brcmfmac/fwsignal.c 		skcb->mac = entry;
entry            1006 drivers/net/wireless/broadcom/brcm80211/brcmfmac/fwsignal.c 	struct brcmf_fws_mac_descriptor *entry, *existing;
entry            1015 drivers/net/wireless/broadcom/brcm80211/brcmfmac/fwsignal.c 	entry = &fws->desc.nodes[mac_handle & 0x1F];
entry            1017 drivers/net/wireless/broadcom/brcm80211/brcmfmac/fwsignal.c 		if (entry->occupied) {
entry            1019 drivers/net/wireless/broadcom/brcm80211/brcmfmac/fwsignal.c 				  entry->name, addr);
entry            1021 drivers/net/wireless/broadcom/brcm80211/brcmfmac/fwsignal.c 			brcmf_fws_macdesc_cleanup(fws, entry, -1);
entry            1022 drivers/net/wireless/broadcom/brcm80211/brcmfmac/fwsignal.c 			brcmf_fws_macdesc_deinit(entry);
entry            1031 drivers/net/wireless/broadcom/brcm80211/brcmfmac/fwsignal.c 		if (!entry->occupied) {
entry            1033 drivers/net/wireless/broadcom/brcm80211/brcmfmac/fwsignal.c 			entry->mac_handle = mac_handle;
entry            1034 drivers/net/wireless/broadcom/brcm80211/brcmfmac/fwsignal.c 			brcmf_fws_macdesc_init(entry, addr, ifidx);
entry            1035 drivers/net/wireless/broadcom/brcm80211/brcmfmac/fwsignal.c 			brcmf_fws_macdesc_set_name(fws, entry);
entry            1036 drivers/net/wireless/broadcom/brcm80211/brcmfmac/fwsignal.c 			brcmu_pktq_init(&entry->psq, BRCMF_FWS_PSQ_PREC_COUNT,
entry            1039 drivers/net/wireless/broadcom/brcm80211/brcmfmac/fwsignal.c 			brcmf_dbg(TRACE, "add %s mac %pM\n", entry->name, addr);
entry            1044 drivers/net/wireless/broadcom/brcm80211/brcmfmac/fwsignal.c 		if (entry != existing) {
entry            1047 drivers/net/wireless/broadcom/brcm80211/brcmfmac/fwsignal.c 			memcpy(entry, existing,
entry            1049 drivers/net/wireless/broadcom/brcm80211/brcmfmac/fwsignal.c 			entry->mac_handle = mac_handle;
entry            1051 drivers/net/wireless/broadcom/brcm80211/brcmfmac/fwsignal.c 			brcmf_fws_macdesc_set_name(fws, entry);
entry            1053 drivers/net/wireless/broadcom/brcm80211/brcmfmac/fwsignal.c 			brcmf_dbg(TRACE, "relocate %s mac %pM\n", entry->name,
entry            1057 drivers/net/wireless/broadcom/brcm80211/brcmfmac/fwsignal.c 			WARN_ON(entry->mac_handle != mac_handle);
entry            1067 drivers/net/wireless/broadcom/brcm80211/brcmfmac/fwsignal.c 	struct brcmf_fws_mac_descriptor *entry;
entry            1072 drivers/net/wireless/broadcom/brcm80211/brcmfmac/fwsignal.c 	entry = &fws->desc.nodes[mac_handle & 0x1F];
entry            1073 drivers/net/wireless/broadcom/brcm80211/brcmfmac/fwsignal.c 	if (!entry->occupied) {
entry            1079 drivers/net/wireless/broadcom/brcm80211/brcmfmac/fwsignal.c 	entry->requested_credit = 0;
entry            1080 drivers/net/wireless/broadcom/brcm80211/brcmfmac/fwsignal.c 	entry->requested_packet = 0;
entry            1082 drivers/net/wireless/broadcom/brcm80211/brcmfmac/fwsignal.c 		entry->state = BRCMF_FWS_STATE_OPEN;
entry            1085 drivers/net/wireless/broadcom/brcm80211/brcmfmac/fwsignal.c 		entry->state = BRCMF_FWS_STATE_CLOSE;
entry            1086 drivers/net/wireless/broadcom/brcm80211/brcmfmac/fwsignal.c 		brcmf_fws_tim_update(fws, entry, BRCMF_FWS_FIFO_AC_BK, false);
entry            1087 drivers/net/wireless/broadcom/brcm80211/brcmfmac/fwsignal.c 		brcmf_fws_tim_update(fws, entry, BRCMF_FWS_FIFO_AC_BE, false);
entry            1088 drivers/net/wireless/broadcom/brcm80211/brcmfmac/fwsignal.c 		brcmf_fws_tim_update(fws, entry, BRCMF_FWS_FIFO_AC_VI, false);
entry            1089 drivers/net/wireless/broadcom/brcm80211/brcmfmac/fwsignal.c 		brcmf_fws_tim_update(fws, entry, BRCMF_FWS_FIFO_AC_VO, true);
entry            1099 drivers/net/wireless/broadcom/brcm80211/brcmfmac/fwsignal.c 	struct brcmf_fws_mac_descriptor *entry;
entry            1110 drivers/net/wireless/broadcom/brcm80211/brcmfmac/fwsignal.c 	entry = &fws->desc.iface[ifidx];
entry            1111 drivers/net/wireless/broadcom/brcm80211/brcmfmac/fwsignal.c 	if (!entry->occupied) {
entry            1117 drivers/net/wireless/broadcom/brcm80211/brcmfmac/fwsignal.c 		  entry->name);
entry            1121 drivers/net/wireless/broadcom/brcm80211/brcmfmac/fwsignal.c 		entry->state = BRCMF_FWS_STATE_OPEN;
entry            1125 drivers/net/wireless/broadcom/brcm80211/brcmfmac/fwsignal.c 		entry->state = BRCMF_FWS_STATE_CLOSE;
entry            1144 drivers/net/wireless/broadcom/brcm80211/brcmfmac/fwsignal.c 	struct brcmf_fws_mac_descriptor *entry;
entry            1146 drivers/net/wireless/broadcom/brcm80211/brcmfmac/fwsignal.c 	entry = &fws->desc.nodes[data[1] & 0x1F];
entry            1147 drivers/net/wireless/broadcom/brcm80211/brcmfmac/fwsignal.c 	if (!entry->occupied) {
entry            1156 drivers/net/wireless/broadcom/brcm80211/brcmfmac/fwsignal.c 		  brcmf_fws_get_tlv_name(type), type, entry->name,
entry            1160 drivers/net/wireless/broadcom/brcm80211/brcmfmac/fwsignal.c 		entry->requested_credit = data[0];
entry            1162 drivers/net/wireless/broadcom/brcm80211/brcmfmac/fwsignal.c 		entry->requested_packet = data[0];
entry            1164 drivers/net/wireless/broadcom/brcm80211/brcmfmac/fwsignal.c 	entry->ac_bitmap = data[2];
entry            1170 drivers/net/wireless/broadcom/brcm80211/brcmfmac/fwsignal.c brcmf_fws_macdesc_use_req_credit(struct brcmf_fws_mac_descriptor *entry,
entry            1173 drivers/net/wireless/broadcom/brcm80211/brcmfmac/fwsignal.c 	if (entry->requested_credit > 0) {
entry            1174 drivers/net/wireless/broadcom/brcm80211/brcmfmac/fwsignal.c 		entry->requested_credit--;
entry            1177 drivers/net/wireless/broadcom/brcm80211/brcmfmac/fwsignal.c 		if (entry->state != BRCMF_FWS_STATE_CLOSE)
entry            1179 drivers/net/wireless/broadcom/brcm80211/brcmfmac/fwsignal.c 	} else if (entry->requested_packet > 0) {
entry            1180 drivers/net/wireless/broadcom/brcm80211/brcmfmac/fwsignal.c 		entry->requested_packet--;
entry            1183 drivers/net/wireless/broadcom/brcm80211/brcmfmac/fwsignal.c 		if (entry->state != BRCMF_FWS_STATE_CLOSE)
entry            1193 drivers/net/wireless/broadcom/brcm80211/brcmfmac/fwsignal.c 	struct brcmf_fws_mac_descriptor *entry = brcmf_skbcb(skb)->mac;
entry            1196 drivers/net/wireless/broadcom/brcm80211/brcmfmac/fwsignal.c 	    (entry->state == BRCMF_FWS_STATE_CLOSE))
entry            1197 drivers/net/wireless/broadcom/brcm80211/brcmfmac/fwsignal.c 		entry->requested_credit++;
entry            1254 drivers/net/wireless/broadcom/brcm80211/brcmfmac/fwsignal.c 	struct brcmf_fws_mac_descriptor *entry;
entry            1262 drivers/net/wireless/broadcom/brcm80211/brcmfmac/fwsignal.c 	entry = brcmf_skbcb(p)->mac;
entry            1263 drivers/net/wireless/broadcom/brcm80211/brcmfmac/fwsignal.c 	if (entry == NULL) {
entry            1275 drivers/net/wireless/broadcom/brcm80211/brcmfmac/fwsignal.c 		pq = &entry->psq;
entry            1320 drivers/net/wireless/broadcom/brcm80211/brcmfmac/fwsignal.c 	} else if (brcmu_pktq_penq(&entry->psq, prec, p) == NULL) {
entry            1336 drivers/net/wireless/broadcom/brcm80211/brcmfmac/fwsignal.c 	brcmf_fws_tim_update(fws, entry, fifo, true);
entry            1337 drivers/net/wireless/broadcom/brcm80211/brcmfmac/fwsignal.c 	brcmf_fws_flow_control_check(fws, &entry->psq,
entry            1345 drivers/net/wireless/broadcom/brcm80211/brcmfmac/fwsignal.c 	struct brcmf_fws_mac_descriptor *entry;
entry            1358 drivers/net/wireless/broadcom/brcm80211/brcmfmac/fwsignal.c 		entry = &table[(node_pos + i) % num_nodes];
entry            1359 drivers/net/wireless/broadcom/brcm80211/brcmfmac/fwsignal.c 		if (!entry->occupied ||
entry            1360 drivers/net/wireless/broadcom/brcm80211/brcmfmac/fwsignal.c 		    brcmf_fws_macdesc_closed(fws, entry, fifo))
entry            1363 drivers/net/wireless/broadcom/brcm80211/brcmfmac/fwsignal.c 		if (entry->suppressed)
entry            1367 drivers/net/wireless/broadcom/brcm80211/brcmfmac/fwsignal.c 		p = brcmu_pktq_mdeq(&entry->psq, pmsk << (fifo * 2), &prec_out);
entry            1369 drivers/net/wireless/broadcom/brcm80211/brcmfmac/fwsignal.c 			if (entry->suppressed) {
entry            1370 drivers/net/wireless/broadcom/brcm80211/brcmfmac/fwsignal.c 				if (entry->suppr_transit_count)
entry            1372 drivers/net/wireless/broadcom/brcm80211/brcmfmac/fwsignal.c 				entry->suppressed = false;
entry            1373 drivers/net/wireless/broadcom/brcm80211/brcmfmac/fwsignal.c 				p = brcmu_pktq_mdeq(&entry->psq,
entry            1380 drivers/net/wireless/broadcom/brcm80211/brcmfmac/fwsignal.c 		brcmf_fws_macdesc_use_req_credit(entry, p);
entry            1384 drivers/net/wireless/broadcom/brcm80211/brcmfmac/fwsignal.c 		brcmf_fws_flow_control_check(fws, &entry->psq,
entry            1392 drivers/net/wireless/broadcom/brcm80211/brcmfmac/fwsignal.c 		brcmf_fws_tim_update(fws, entry, fifo, false);
entry            1413 drivers/net/wireless/broadcom/brcm80211/brcmfmac/fwsignal.c 	struct brcmf_fws_mac_descriptor *entry = brcmf_skbcb(skb)->mac;
entry            1420 drivers/net/wireless/broadcom/brcm80211/brcmfmac/fwsignal.c 	if (!entry->suppressed) {
entry            1421 drivers/net/wireless/broadcom/brcm80211/brcmfmac/fwsignal.c 		entry->suppressed = true;
entry            1422 drivers/net/wireless/broadcom/brcm80211/brcmfmac/fwsignal.c 		entry->suppr_transit_count = entry->transit_count;
entry            1424 drivers/net/wireless/broadcom/brcm80211/brcmfmac/fwsignal.c 			  entry->name, entry->transit_count);
entry            1427 drivers/net/wireless/broadcom/brcm80211/brcmfmac/fwsignal.c 	entry->generation = genbit;
entry            1461 drivers/net/wireless/broadcom/brcm80211/brcmfmac/fwsignal.c 	struct brcmf_fws_mac_descriptor *entry = NULL;
entry            1491 drivers/net/wireless/broadcom/brcm80211/brcmfmac/fwsignal.c 		entry = skcb->mac;
entry            1492 drivers/net/wireless/broadcom/brcm80211/brcmfmac/fwsignal.c 		if (WARN_ON(!entry)) {
entry            1496 drivers/net/wireless/broadcom/brcm80211/brcmfmac/fwsignal.c 		entry->transit_count--;
entry            1497 drivers/net/wireless/broadcom/brcm80211/brcmfmac/fwsignal.c 		if (entry->suppressed && entry->suppr_transit_count)
entry            1498 drivers/net/wireless/broadcom/brcm80211/brcmfmac/fwsignal.c 			entry->suppr_transit_count--;
entry            1500 drivers/net/wireless/broadcom/brcm80211/brcmfmac/fwsignal.c 		brcmf_dbg(DATA, "%s flags %d htod %X seq %X\n", entry->name,
entry            1976 drivers/net/wireless/broadcom/brcm80211/brcmfmac/fwsignal.c 	struct brcmf_fws_mac_descriptor *entry = skcb->mac;
entry            1980 drivers/net/wireless/broadcom/brcm80211/brcmfmac/fwsignal.c 		brcmf_skb_htod_tag_set_field(p, GENERATION, entry->generation);
entry            1997 drivers/net/wireless/broadcom/brcm80211/brcmfmac/fwsignal.c 	struct brcmf_fws_mac_descriptor *entry;
entry            2002 drivers/net/wireless/broadcom/brcm80211/brcmfmac/fwsignal.c 	entry = brcmf_skbcb(skb)->mac;
entry            2003 drivers/net/wireless/broadcom/brcm80211/brcmfmac/fwsignal.c 	if (entry->occupied) {
entry            2008 drivers/net/wireless/broadcom/brcm80211/brcmfmac/fwsignal.c 		pktout = brcmu_pktq_penq_head(&entry->psq, qidx, skb);
entry            2010 drivers/net/wireless/broadcom/brcm80211/brcmfmac/fwsignal.c 			bphy_err(drvr, "%s queue %d full\n", entry->name, qidx);
entry            2014 drivers/net/wireless/broadcom/brcm80211/brcmfmac/fwsignal.c 		bphy_err(drvr, "%s entry removed\n", entry->name);
entry            2058 drivers/net/wireless/broadcom/brcm80211/brcmfmac/fwsignal.c 	struct brcmf_fws_mac_descriptor *entry;
entry            2063 drivers/net/wireless/broadcom/brcm80211/brcmfmac/fwsignal.c 	entry = skcb->mac;
entry            2064 drivers/net/wireless/broadcom/brcm80211/brcmfmac/fwsignal.c 	if (IS_ERR(entry))
entry            2065 drivers/net/wireless/broadcom/brcm80211/brcmfmac/fwsignal.c 		return PTR_ERR(entry);
entry            2068 drivers/net/wireless/broadcom/brcm80211/brcmfmac/fwsignal.c 	entry->transit_count++;
entry            2069 drivers/net/wireless/broadcom/brcm80211/brcmfmac/fwsignal.c 	if (entry->suppressed)
entry            2070 drivers/net/wireless/broadcom/brcm80211/brcmfmac/fwsignal.c 		entry->suppr_transit_count++;
entry            2075 drivers/net/wireless/broadcom/brcm80211/brcmfmac/fwsignal.c 	brcmf_dbg(DATA, "%s flags %X htod %X bus_tx %d\n", entry->name,
entry            2078 drivers/net/wireless/broadcom/brcm80211/brcmfmac/fwsignal.c 		entry->transit_count--;
entry            2079 drivers/net/wireless/broadcom/brcm80211/brcmfmac/fwsignal.c 		if (entry->suppressed)
entry            2080 drivers/net/wireless/broadcom/brcm80211/brcmfmac/fwsignal.c 			entry->suppr_transit_count--;
entry            2159 drivers/net/wireless/broadcom/brcm80211/brcmfmac/fwsignal.c 	struct brcmf_fws_mac_descriptor *entry = ifp->fws_desc;
entry            2162 drivers/net/wireless/broadcom/brcm80211/brcmfmac/fwsignal.c 	if (!entry)
entry            2165 drivers/net/wireless/broadcom/brcm80211/brcmfmac/fwsignal.c 	brcmf_fws_macdesc_init(entry, ifp->mac_addr, ifp->ifidx);
entry            2171 drivers/net/wireless/broadcom/brcm80211/brcmfmac/fwsignal.c 	struct brcmf_fws_mac_descriptor *entry;
entry            2176 drivers/net/wireless/broadcom/brcm80211/brcmfmac/fwsignal.c 	entry = &fws->desc.iface[ifp->ifidx];
entry            2177 drivers/net/wireless/broadcom/brcm80211/brcmfmac/fwsignal.c 	ifp->fws_desc = entry;
entry            2178 drivers/net/wireless/broadcom/brcm80211/brcmfmac/fwsignal.c 	brcmf_fws_macdesc_init(entry, ifp->mac_addr, ifp->ifidx);
entry            2179 drivers/net/wireless/broadcom/brcm80211/brcmfmac/fwsignal.c 	brcmf_fws_macdesc_set_name(fws, entry);
entry            2180 drivers/net/wireless/broadcom/brcm80211/brcmfmac/fwsignal.c 	brcmu_pktq_init(&entry->psq, BRCMF_FWS_PSQ_PREC_COUNT,
entry            2182 drivers/net/wireless/broadcom/brcm80211/brcmfmac/fwsignal.c 	brcmf_dbg(TRACE, "added %s\n", entry->name);
entry            2187 drivers/net/wireless/broadcom/brcm80211/brcmfmac/fwsignal.c 	struct brcmf_fws_mac_descriptor *entry = ifp->fws_desc;
entry            2190 drivers/net/wireless/broadcom/brcm80211/brcmfmac/fwsignal.c 	if (!entry)
entry            2195 drivers/net/wireless/broadcom/brcm80211/brcmfmac/fwsignal.c 	brcmf_dbg(TRACE, "deleting %s\n", entry->name);
entry            2198 drivers/net/wireless/broadcom/brcm80211/brcmfmac/fwsignal.c 	brcmf_fws_macdesc_deinit(entry);
entry             176 drivers/net/wireless/broadcom/brcm80211/brcmsmac/debug.c 	struct brcms_debugfs_entry *entry = inode->i_private;
entry             178 drivers/net/wireless/broadcom/brcm80211/brcmsmac/debug.c 	return single_open(f, entry->read, entry->drvr);
entry             195 drivers/net/wireless/broadcom/brcm80211/brcmsmac/debug.c 	struct brcms_debugfs_entry *entry;
entry             197 drivers/net/wireless/broadcom/brcm80211/brcmsmac/debug.c 	entry = devm_kzalloc(dev, sizeof(*entry), GFP_KERNEL);
entry             198 drivers/net/wireless/broadcom/brcm80211/brcmsmac/debug.c 	if (!entry)
entry             201 drivers/net/wireless/broadcom/brcm80211/brcmsmac/debug.c 	entry->read = read_fn;
entry             202 drivers/net/wireless/broadcom/brcm80211/brcmsmac/debug.c 	entry->drvr = drvr;
entry             204 drivers/net/wireless/broadcom/brcm80211/brcmsmac/debug.c 	debugfs_create_file(fn, 0444, dentry, entry, &brcms_debugfs_def_ops);
entry            1596 drivers/net/wireless/broadcom/brcm80211/brcmsmac/mac80211_if.c 	int i, entry;
entry            1601 drivers/net/wireless/broadcom/brcm80211/brcmsmac/mac80211_if.c 		for (entry = 0; entry < wl->fw.hdr_num_entries[i];
entry            1602 drivers/net/wireless/broadcom/brcm80211/brcmsmac/mac80211_if.c 		     entry++, hdr++) {
entry            1628 drivers/net/wireless/broadcom/brcm80211/brcmsmac/mac80211_if.c 	int i, entry;
entry            1633 drivers/net/wireless/broadcom/brcm80211/brcmsmac/mac80211_if.c 		for (entry = 0; entry < wl->fw.hdr_num_entries[i];
entry            1634 drivers/net/wireless/broadcom/brcm80211/brcmsmac/mac80211_if.c 		     entry++, hdr++) {
entry            1670 drivers/net/wireless/broadcom/brcm80211/brcmsmac/mac80211_if.c 	int entry;
entry            1696 drivers/net/wireless/broadcom/brcm80211/brcmsmac/mac80211_if.c 			for (entry = 0; entry < wl->fw.hdr_num_entries[i] &&
entry            1697 drivers/net/wireless/broadcom/brcm80211/brcmsmac/mac80211_if.c 			     !rc; entry++, ucode_hdr++) {
entry            4518 drivers/net/wireless/cisco/airo.c 	struct proc_dir_entry *entry;
entry            4529 drivers/net/wireless/cisco/airo.c 	entry = proc_create_data("StatsDelta", 0444 & proc_perm,
entry            4531 drivers/net/wireless/cisco/airo.c 	if (!entry)
entry            4533 drivers/net/wireless/cisco/airo.c 	proc_set_user(entry, proc_kuid, proc_kgid);
entry            4536 drivers/net/wireless/cisco/airo.c 	entry = proc_create_data("Stats", 0444 & proc_perm,
entry            4538 drivers/net/wireless/cisco/airo.c 	if (!entry)
entry            4540 drivers/net/wireless/cisco/airo.c 	proc_set_user(entry, proc_kuid, proc_kgid);
entry            4543 drivers/net/wireless/cisco/airo.c 	entry = proc_create_data("Status", 0444 & proc_perm,
entry            4545 drivers/net/wireless/cisco/airo.c 	if (!entry)
entry            4547 drivers/net/wireless/cisco/airo.c 	proc_set_user(entry, proc_kuid, proc_kgid);
entry            4550 drivers/net/wireless/cisco/airo.c 	entry = proc_create_data("Config", proc_perm,
entry            4552 drivers/net/wireless/cisco/airo.c 	if (!entry)
entry            4554 drivers/net/wireless/cisco/airo.c 	proc_set_user(entry, proc_kuid, proc_kgid);
entry            4557 drivers/net/wireless/cisco/airo.c 	entry = proc_create_data("SSID", proc_perm,
entry            4559 drivers/net/wireless/cisco/airo.c 	if (!entry)
entry            4561 drivers/net/wireless/cisco/airo.c 	proc_set_user(entry, proc_kuid, proc_kgid);
entry            4564 drivers/net/wireless/cisco/airo.c 	entry = proc_create_data("APList", proc_perm,
entry            4566 drivers/net/wireless/cisco/airo.c 	if (!entry)
entry            4568 drivers/net/wireless/cisco/airo.c 	proc_set_user(entry, proc_kuid, proc_kgid);
entry            4571 drivers/net/wireless/cisco/airo.c 	entry = proc_create_data("BSSList", proc_perm,
entry            4573 drivers/net/wireless/cisco/airo.c 	if (!entry)
entry            4575 drivers/net/wireless/cisco/airo.c 	proc_set_user(entry, proc_kuid, proc_kgid);
entry            4578 drivers/net/wireless/cisco/airo.c 	entry = proc_create_data("WepKey", proc_perm,
entry            4580 drivers/net/wireless/cisco/airo.c 	if (!entry)
entry            4582 drivers/net/wireless/cisco/airo.c 	proc_set_user(entry, proc_kuid, proc_kgid);
entry            3896 drivers/net/wireless/intel/ipw2x00/ipw2200.c 	struct ipw_station_entry entry;
entry            3916 drivers/net/wireless/intel/ipw2x00/ipw2200.c 	entry.reserved = 0;
entry            3917 drivers/net/wireless/intel/ipw2x00/ipw2200.c 	entry.support_mode = 0;
entry            3918 drivers/net/wireless/intel/ipw2x00/ipw2200.c 	memcpy(entry.mac_addr, bssid, ETH_ALEN);
entry            3920 drivers/net/wireless/intel/ipw2x00/ipw2200.c 	ipw_write_direct(priv, IPW_STATION_TABLE_LOWER + i * sizeof(entry),
entry            3921 drivers/net/wireless/intel/ipw2x00/ipw2200.c 			 &entry, sizeof(entry));
entry            8138 drivers/net/wireless/intel/ipw2x00/ipw2200.c 			struct ipw_ibss_seq *entry = NULL;
entry            8143 drivers/net/wireless/intel/ipw2x00/ipw2200.c 				entry =
entry            8145 drivers/net/wireless/intel/ipw2x00/ipw2200.c 				if (ether_addr_equal(entry->mac, mac))
entry            8149 drivers/net/wireless/intel/ipw2x00/ipw2200.c 				entry = kmalloc(sizeof(*entry), GFP_ATOMIC);
entry            8150 drivers/net/wireless/intel/ipw2x00/ipw2200.c 				if (!entry) {
entry            8155 drivers/net/wireless/intel/ipw2x00/ipw2200.c 				memcpy(entry->mac, mac, ETH_ALEN);
entry            8156 drivers/net/wireless/intel/ipw2x00/ipw2200.c 				entry->seq_num = seq;
entry            8157 drivers/net/wireless/intel/ipw2x00/ipw2200.c 				entry->frag_num = frag;
entry            8158 drivers/net/wireless/intel/ipw2x00/ipw2200.c 				entry->packet_time = jiffies;
entry            8159 drivers/net/wireless/intel/ipw2x00/ipw2200.c 				list_add(&entry->list,
entry            8163 drivers/net/wireless/intel/ipw2x00/ipw2200.c 			last_seq = &entry->seq_num;
entry            8164 drivers/net/wireless/intel/ipw2x00/ipw2200.c 			last_frag = &entry->frag_num;
entry            8165 drivers/net/wireless/intel/ipw2x00/ipw2200.c 			last_time = &entry->packet_time;
entry              60 drivers/net/wireless/intel/ipw2x00/libipw_rx.c 	struct libipw_frag_entry *entry;
entry              64 drivers/net/wireless/intel/ipw2x00/libipw_rx.c 		entry = &ieee->frag_cache[i];
entry              65 drivers/net/wireless/intel/ipw2x00/libipw_rx.c 		if (entry->skb != NULL &&
entry              66 drivers/net/wireless/intel/ipw2x00/libipw_rx.c 		    time_after(jiffies, entry->first_frag_time + 2 * HZ)) {
entry              69 drivers/net/wireless/intel/ipw2x00/libipw_rx.c 					     entry->seq, entry->last_frag);
entry              70 drivers/net/wireless/intel/ipw2x00/libipw_rx.c 			dev_kfree_skb_any(entry->skb);
entry              71 drivers/net/wireless/intel/ipw2x00/libipw_rx.c 			entry->skb = NULL;
entry              74 drivers/net/wireless/intel/ipw2x00/libipw_rx.c 		if (entry->skb != NULL && entry->seq == seq &&
entry              75 drivers/net/wireless/intel/ipw2x00/libipw_rx.c 		    (entry->last_frag + 1 == frag || frag == -1) &&
entry              76 drivers/net/wireless/intel/ipw2x00/libipw_rx.c 		    ether_addr_equal(entry->src_addr, src) &&
entry              77 drivers/net/wireless/intel/ipw2x00/libipw_rx.c 		    ether_addr_equal(entry->dst_addr, dst))
entry              78 drivers/net/wireless/intel/ipw2x00/libipw_rx.c 			return entry;
entry              91 drivers/net/wireless/intel/ipw2x00/libipw_rx.c 	struct libipw_frag_entry *entry;
entry             107 drivers/net/wireless/intel/ipw2x00/libipw_rx.c 		entry = &ieee->frag_cache[ieee->frag_next_idx];
entry             112 drivers/net/wireless/intel/ipw2x00/libipw_rx.c 		if (entry->skb != NULL)
entry             113 drivers/net/wireless/intel/ipw2x00/libipw_rx.c 			dev_kfree_skb_any(entry->skb);
entry             115 drivers/net/wireless/intel/ipw2x00/libipw_rx.c 		entry->first_frag_time = jiffies;
entry             116 drivers/net/wireless/intel/ipw2x00/libipw_rx.c 		entry->seq = seq;
entry             117 drivers/net/wireless/intel/ipw2x00/libipw_rx.c 		entry->last_frag = frag;
entry             118 drivers/net/wireless/intel/ipw2x00/libipw_rx.c 		entry->skb = skb;
entry             119 drivers/net/wireless/intel/ipw2x00/libipw_rx.c 		memcpy(entry->src_addr, hdr->addr2, ETH_ALEN);
entry             120 drivers/net/wireless/intel/ipw2x00/libipw_rx.c 		memcpy(entry->dst_addr, hdr->addr1, ETH_ALEN);
entry             124 drivers/net/wireless/intel/ipw2x00/libipw_rx.c 		entry = libipw_frag_cache_find(ieee, seq, frag, hdr->addr2,
entry             126 drivers/net/wireless/intel/ipw2x00/libipw_rx.c 		if (entry != NULL) {
entry             127 drivers/net/wireless/intel/ipw2x00/libipw_rx.c 			entry->last_frag = frag;
entry             128 drivers/net/wireless/intel/ipw2x00/libipw_rx.c 			skb = entry->skb;
entry             141 drivers/net/wireless/intel/ipw2x00/libipw_rx.c 	struct libipw_frag_entry *entry;
entry             146 drivers/net/wireless/intel/ipw2x00/libipw_rx.c 	entry = libipw_frag_cache_find(ieee, seq, -1, hdr->addr2,
entry             149 drivers/net/wireless/intel/ipw2x00/libipw_rx.c 	if (entry == NULL) {
entry             155 drivers/net/wireless/intel/ipw2x00/libipw_rx.c 	entry->skb = NULL;
entry            1652 drivers/net/wireless/intel/iwlwifi/fw/dbg.c 	struct iwl_fw_ini_dump_entry *entry;
entry            1666 drivers/net/wireless/intel/iwlwifi/fw/dbg.c 	entry = kmalloc(sizeof(*entry) + sizeof(*tlv) + size, GFP_KERNEL);
entry            1667 drivers/net/wireless/intel/iwlwifi/fw/dbg.c 	if (!entry)
entry            1670 drivers/net/wireless/intel/iwlwifi/fw/dbg.c 	entry->size = sizeof(*tlv) + size;
entry            1672 drivers/net/wireless/intel/iwlwifi/fw/dbg.c 	tlv = (void *)entry->data;
entry            1708 drivers/net/wireless/intel/iwlwifi/fw/dbg.c 	list_add_tail(&entry->list, list);
entry            1710 drivers/net/wireless/intel/iwlwifi/fw/dbg.c 	return entry->size;
entry            1713 drivers/net/wireless/intel/iwlwifi/fw/dbg.c 	kfree(entry);
entry            1722 drivers/net/wireless/intel/iwlwifi/fw/dbg.c 	struct iwl_fw_ini_dump_entry *entry;
entry            1728 drivers/net/wireless/intel/iwlwifi/fw/dbg.c 	entry = kmalloc(sizeof(*entry) + size, GFP_KERNEL);
entry            1729 drivers/net/wireless/intel/iwlwifi/fw/dbg.c 	if (!entry)
entry            1732 drivers/net/wireless/intel/iwlwifi/fw/dbg.c 	entry->size = size;
entry            1734 drivers/net/wireless/intel/iwlwifi/fw/dbg.c 	tlv = (void *)entry->data;
entry            1786 drivers/net/wireless/intel/iwlwifi/fw/dbg.c 	list_add(&entry->list, list);
entry            1788 drivers/net/wireless/intel/iwlwifi/fw/dbg.c 	return entry->size;
entry            1904 drivers/net/wireless/intel/iwlwifi/fw/dbg.c 	struct iwl_fw_ini_dump_entry *entry;
entry            1916 drivers/net/wireless/intel/iwlwifi/fw/dbg.c 	entry = kmalloc(sizeof(*entry) + sizeof(*hdr), GFP_KERNEL);
entry            1917 drivers/net/wireless/intel/iwlwifi/fw/dbg.c 	if (!entry)
entry            1920 drivers/net/wireless/intel/iwlwifi/fw/dbg.c 	entry->size = sizeof(*hdr);
entry            1924 drivers/net/wireless/intel/iwlwifi/fw/dbg.c 		kfree(entry);
entry            1928 drivers/net/wireless/intel/iwlwifi/fw/dbg.c 	hdr = (void *)entry->data;
entry            1930 drivers/net/wireless/intel/iwlwifi/fw/dbg.c 	hdr->file_len = cpu_to_le32(size + entry->size);
entry            1932 drivers/net/wireless/intel/iwlwifi/fw/dbg.c 	list_add(&entry->list, list);
entry            1986 drivers/net/wireless/intel/iwlwifi/fw/dbg.c 		struct iwl_fw_ini_dump_entry *entry =
entry            1987 drivers/net/wireless/intel/iwlwifi/fw/dbg.c 			list_entry(list->next, typeof(*entry), list);
entry            1989 drivers/net/wireless/intel/iwlwifi/fw/dbg.c 		list_del(&entry->list);
entry            1990 drivers/net/wireless/intel/iwlwifi/fw/dbg.c 		kfree(entry);
entry            2007 drivers/net/wireless/intel/iwlwifi/fw/dbg.c 		struct iwl_fw_ini_dump_entry *entry;
entry            2011 drivers/net/wireless/intel/iwlwifi/fw/dbg.c 		list_for_each_entry(entry, &dump_list, list) {
entry            2013 drivers/net/wireless/intel/iwlwifi/fw/dbg.c 					     entry->data, entry->size, offs);
entry            2014 drivers/net/wireless/intel/iwlwifi/fw/dbg.c 			offs += entry->size;
entry             160 drivers/net/wireless/intel/iwlwifi/iwl-devtrace-io.h 		__field(u32, entry)
entry             167 drivers/net/wireless/intel/iwlwifi/iwl-devtrace-io.h 		__entry->entry = msix_entry->entry;
entry             173 drivers/net/wireless/intel/iwlwifi/iwl-devtrace-io.h 		  __entry->entry, __entry->defirq,
entry             169 drivers/net/wireless/intel/iwlwifi/iwl-phy-db.c 	struct iwl_phy_db_entry *entry =
entry             171 drivers/net/wireless/intel/iwlwifi/iwl-phy-db.c 	if (!entry)
entry             174 drivers/net/wireless/intel/iwlwifi/iwl-phy-db.c 	kfree(entry->data);
entry             175 drivers/net/wireless/intel/iwlwifi/iwl-phy-db.c 	entry->data = NULL;
entry             176 drivers/net/wireless/intel/iwlwifi/iwl-phy-db.c 	entry->size = 0;
entry             208 drivers/net/wireless/intel/iwlwifi/iwl-phy-db.c 	struct iwl_phy_db_entry *entry;
entry             244 drivers/net/wireless/intel/iwlwifi/iwl-phy-db.c 	entry = iwl_phy_db_get_section(phy_db, type, chg_id);
entry             245 drivers/net/wireless/intel/iwlwifi/iwl-phy-db.c 	if (!entry)
entry             248 drivers/net/wireless/intel/iwlwifi/iwl-phy-db.c 	kfree(entry->data);
entry             249 drivers/net/wireless/intel/iwlwifi/iwl-phy-db.c 	entry->data = kmemdup(phy_db_notif->data, size, GFP_ATOMIC);
entry             250 drivers/net/wireless/intel/iwlwifi/iwl-phy-db.c 	if (!entry->data) {
entry             251 drivers/net/wireless/intel/iwlwifi/iwl-phy-db.c 		entry->size = 0;
entry             255 drivers/net/wireless/intel/iwlwifi/iwl-phy-db.c 	entry->size = size;
entry             329 drivers/net/wireless/intel/iwlwifi/iwl-phy-db.c 	struct iwl_phy_db_entry *entry;
entry             341 drivers/net/wireless/intel/iwlwifi/iwl-phy-db.c 	entry = iwl_phy_db_get_section(phy_db, type, ch_group_id);
entry             342 drivers/net/wireless/intel/iwlwifi/iwl-phy-db.c 	if (!entry)
entry             345 drivers/net/wireless/intel/iwlwifi/iwl-phy-db.c 	*data = entry->data;
entry             346 drivers/net/wireless/intel/iwlwifi/iwl-phy-db.c 	*size = entry->size;
entry             388 drivers/net/wireless/intel/iwlwifi/iwl-phy-db.c 	struct iwl_phy_db_entry *entry;
entry             392 drivers/net/wireless/intel/iwlwifi/iwl-phy-db.c 		entry = iwl_phy_db_get_section(phy_db,
entry             395 drivers/net/wireless/intel/iwlwifi/iwl-phy-db.c 		if (!entry)
entry             398 drivers/net/wireless/intel/iwlwifi/iwl-phy-db.c 		if (!entry->size)
entry             404 drivers/net/wireless/intel/iwlwifi/iwl-phy-db.c 					  entry->size,
entry             405 drivers/net/wireless/intel/iwlwifi/iwl-phy-db.c 					  entry->data);
entry             436 drivers/net/wireless/intel/iwlwifi/mvm/ftm-initiator.c 	struct iwl_mvm_loc_entry *entry;
entry             438 drivers/net/wireless/intel/iwlwifi/mvm/ftm-initiator.c 	list_for_each_entry(entry, &mvm->ftm_initiator.loc_list, list) {
entry             439 drivers/net/wireless/intel/iwlwifi/mvm/ftm-initiator.c 		if (!ether_addr_equal_unaligned(res->addr, entry->addr))
entry             442 drivers/net/wireless/intel/iwlwifi/mvm/ftm-initiator.c 		if (entry->lci_len) {
entry             443 drivers/net/wireless/intel/iwlwifi/mvm/ftm-initiator.c 			res->ftm.lci_len = entry->lci_len;
entry             444 drivers/net/wireless/intel/iwlwifi/mvm/ftm-initiator.c 			res->ftm.lci = entry->buf;
entry             447 drivers/net/wireless/intel/iwlwifi/mvm/ftm-initiator.c 		if (entry->civic_len) {
entry             448 drivers/net/wireless/intel/iwlwifi/mvm/ftm-initiator.c 			res->ftm.civicloc_len = entry->civic_len;
entry             449 drivers/net/wireless/intel/iwlwifi/mvm/ftm-initiator.c 			res->ftm.civicloc = entry->buf + entry->lci_len;
entry             635 drivers/net/wireless/intel/iwlwifi/mvm/ftm-initiator.c 	struct iwl_mvm_loc_entry *entry;
entry             665 drivers/net/wireless/intel/iwlwifi/mvm/ftm-initiator.c 	entry = kmalloc(sizeof(*entry) + lci_len + civic_len, GFP_KERNEL);
entry             666 drivers/net/wireless/intel/iwlwifi/mvm/ftm-initiator.c 	if (!entry)
entry             669 drivers/net/wireless/intel/iwlwifi/mvm/ftm-initiator.c 	memcpy(entry->addr, mgmt->bssid, ETH_ALEN);
entry             671 drivers/net/wireless/intel/iwlwifi/mvm/ftm-initiator.c 	entry->lci_len = lci_len;
entry             673 drivers/net/wireless/intel/iwlwifi/mvm/ftm-initiator.c 		memcpy(entry->buf, lci, lci_len);
entry             675 drivers/net/wireless/intel/iwlwifi/mvm/ftm-initiator.c 	entry->civic_len = civic_len;
entry             677 drivers/net/wireless/intel/iwlwifi/mvm/ftm-initiator.c 		memcpy(entry->buf + lci_len, civic, civic_len);
entry             679 drivers/net/wireless/intel/iwlwifi/mvm/ftm-initiator.c 	list_add_tail(&entry->list, &mvm->ftm_initiator.loc_list);
entry             824 drivers/net/wireless/intel/iwlwifi/mvm/fw.c 			union acpi_object *entry;
entry             826 drivers/net/wireless/intel/iwlwifi/mvm/fw.c 			entry = &wifi_pkg->package.elements[idx++];
entry             827 drivers/net/wireless/intel/iwlwifi/mvm/fw.c 			if ((entry->type != ACPI_TYPE_INTEGER) ||
entry             828 drivers/net/wireless/intel/iwlwifi/mvm/fw.c 			    (entry->integer.value > U8_MAX)) {
entry             833 drivers/net/wireless/intel/iwlwifi/mvm/fw.c 			mvm->geo_profiles[i].values[j] = entry->integer.value;
entry             916 drivers/net/wireless/intel/iwlwifi/mvm/ops.c 	struct iwl_async_handler_entry *entry, *tmp;
entry             919 drivers/net/wireless/intel/iwlwifi/mvm/ops.c 	list_for_each_entry_safe(entry, tmp, &mvm->async_handlers_list, list) {
entry             920 drivers/net/wireless/intel/iwlwifi/mvm/ops.c 		iwl_free_rxb(&entry->rxb);
entry             921 drivers/net/wireless/intel/iwlwifi/mvm/ops.c 		list_del(&entry->list);
entry             922 drivers/net/wireless/intel/iwlwifi/mvm/ops.c 		kfree(entry);
entry             931 drivers/net/wireless/intel/iwlwifi/mvm/ops.c 	struct iwl_async_handler_entry *entry, *tmp;
entry             944 drivers/net/wireless/intel/iwlwifi/mvm/ops.c 	list_for_each_entry_safe(entry, tmp, &local_list, list) {
entry             945 drivers/net/wireless/intel/iwlwifi/mvm/ops.c 		if (entry->context == RX_HANDLER_ASYNC_LOCKED)
entry             947 drivers/net/wireless/intel/iwlwifi/mvm/ops.c 		entry->fn(mvm, &entry->rxb);
entry             948 drivers/net/wireless/intel/iwlwifi/mvm/ops.c 		iwl_free_rxb(&entry->rxb);
entry             949 drivers/net/wireless/intel/iwlwifi/mvm/ops.c 		list_del(&entry->list);
entry             950 drivers/net/wireless/intel/iwlwifi/mvm/ops.c 		if (entry->context == RX_HANDLER_ASYNC_LOCKED)
entry             952 drivers/net/wireless/intel/iwlwifi/mvm/ops.c 		kfree(entry);
entry            1006 drivers/net/wireless/intel/iwlwifi/mvm/ops.c 		struct iwl_async_handler_entry *entry;
entry            1016 drivers/net/wireless/intel/iwlwifi/mvm/ops.c 		entry = kzalloc(sizeof(*entry), GFP_ATOMIC);
entry            1018 drivers/net/wireless/intel/iwlwifi/mvm/ops.c 		if (!entry)
entry            1021 drivers/net/wireless/intel/iwlwifi/mvm/ops.c 		entry->rxb._page = rxb_steal_page(rxb);
entry            1022 drivers/net/wireless/intel/iwlwifi/mvm/ops.c 		entry->rxb._offset = rxb->_offset;
entry            1023 drivers/net/wireless/intel/iwlwifi/mvm/ops.c 		entry->rxb._rx_page_order = rxb->_rx_page_order;
entry            1024 drivers/net/wireless/intel/iwlwifi/mvm/ops.c 		entry->fn = rx_h->fn;
entry            1025 drivers/net/wireless/intel/iwlwifi/mvm/ops.c 		entry->context = rx_h->context;
entry            1027 drivers/net/wireless/intel/iwlwifi/mvm/ops.c 		list_add_tail(&entry->list, &mvm->async_handlers_list);
entry             616 drivers/net/wireless/intel/iwlwifi/pcie/internal.h 				      struct msix_entry *entry)
entry             626 drivers/net/wireless/intel/iwlwifi/pcie/internal.h 	iwl_write32(trans, CSR_MSIX_AUTOMASK_ST_AD, BIT(entry->entry));
entry            1552 drivers/net/wireless/intel/iwlwifi/pcie/rx.c static struct iwl_trans_pcie *iwl_pcie_get_trans_pcie(struct msix_entry *entry)
entry            1554 drivers/net/wireless/intel/iwlwifi/pcie/rx.c 	u8 queue = entry->entry;
entry            1555 drivers/net/wireless/intel/iwlwifi/pcie/rx.c 	struct msix_entry *entries = entry - queue;
entry            1566 drivers/net/wireless/intel/iwlwifi/pcie/rx.c 	struct msix_entry *entry = dev_id;
entry            1567 drivers/net/wireless/intel/iwlwifi/pcie/rx.c 	struct iwl_trans_pcie *trans_pcie = iwl_pcie_get_trans_pcie(entry);
entry            1570 drivers/net/wireless/intel/iwlwifi/pcie/rx.c 	trace_iwlwifi_dev_irq_msix(trans->dev, entry, false, 0, 0);
entry            1572 drivers/net/wireless/intel/iwlwifi/pcie/rx.c 	if (WARN_ON(entry->entry >= trans->num_rx_queues))
entry            1578 drivers/net/wireless/intel/iwlwifi/pcie/rx.c 	iwl_pcie_rx_handle(trans, entry->entry);
entry            1581 drivers/net/wireless/intel/iwlwifi/pcie/rx.c 	iwl_pcie_clear_irq(trans, entry);
entry            2097 drivers/net/wireless/intel/iwlwifi/pcie/rx.c 	struct msix_entry *entry = dev_id;
entry            2098 drivers/net/wireless/intel/iwlwifi/pcie/rx.c 	struct iwl_trans_pcie *trans_pcie = iwl_pcie_get_trans_pcie(entry);
entry            2115 drivers/net/wireless/intel/iwlwifi/pcie/rx.c 	trace_iwlwifi_dev_irq_msix(trans->dev, entry, true, inta_fh, inta_hw);
entry            2257 drivers/net/wireless/intel/iwlwifi/pcie/rx.c 	iwl_pcie_clear_irq(trans, entry);
entry            1644 drivers/net/wireless/intel/iwlwifi/pcie/trans.c 		trans_pcie->msix_entries[i].entry = i;
entry             224 drivers/net/wireless/intersil/hostap/hostap_80211_rx.c 	struct prism2_frag_entry *entry;
entry             228 drivers/net/wireless/intersil/hostap/hostap_80211_rx.c 		entry = &local->frag_cache[i];
entry             229 drivers/net/wireless/intersil/hostap/hostap_80211_rx.c 		if (entry->skb != NULL &&
entry             230 drivers/net/wireless/intersil/hostap/hostap_80211_rx.c 		    time_after(jiffies, entry->first_frag_time + 2 * HZ)) {
entry             233 drivers/net/wireless/intersil/hostap/hostap_80211_rx.c 			       local->dev->name, entry->seq, entry->last_frag);
entry             234 drivers/net/wireless/intersil/hostap/hostap_80211_rx.c 			dev_kfree_skb(entry->skb);
entry             235 drivers/net/wireless/intersil/hostap/hostap_80211_rx.c 			entry->skb = NULL;
entry             238 drivers/net/wireless/intersil/hostap/hostap_80211_rx.c 		if (entry->skb != NULL && entry->seq == seq &&
entry             239 drivers/net/wireless/intersil/hostap/hostap_80211_rx.c 		    (entry->last_frag + 1 == frag || frag == -1) &&
entry             240 drivers/net/wireless/intersil/hostap/hostap_80211_rx.c 		    memcmp(entry->src_addr, src, ETH_ALEN) == 0 &&
entry             241 drivers/net/wireless/intersil/hostap/hostap_80211_rx.c 		    memcmp(entry->dst_addr, dst, ETH_ALEN) == 0)
entry             242 drivers/net/wireless/intersil/hostap/hostap_80211_rx.c 			return entry;
entry             256 drivers/net/wireless/intersil/hostap/hostap_80211_rx.c 	struct prism2_frag_entry *entry;
entry             272 drivers/net/wireless/intersil/hostap/hostap_80211_rx.c 		entry = &local->frag_cache[local->frag_next_idx];
entry             277 drivers/net/wireless/intersil/hostap/hostap_80211_rx.c 		if (entry->skb != NULL)
entry             278 drivers/net/wireless/intersil/hostap/hostap_80211_rx.c 			dev_kfree_skb(entry->skb);
entry             280 drivers/net/wireless/intersil/hostap/hostap_80211_rx.c 		entry->first_frag_time = jiffies;
entry             281 drivers/net/wireless/intersil/hostap/hostap_80211_rx.c 		entry->seq = seq;
entry             282 drivers/net/wireless/intersil/hostap/hostap_80211_rx.c 		entry->last_frag = frag;
entry             283 drivers/net/wireless/intersil/hostap/hostap_80211_rx.c 		entry->skb = skb;
entry             284 drivers/net/wireless/intersil/hostap/hostap_80211_rx.c 		memcpy(entry->src_addr, hdr->addr2, ETH_ALEN);
entry             285 drivers/net/wireless/intersil/hostap/hostap_80211_rx.c 		memcpy(entry->dst_addr, hdr->addr1, ETH_ALEN);
entry             289 drivers/net/wireless/intersil/hostap/hostap_80211_rx.c 		entry = prism2_frag_cache_find(local, seq, frag, hdr->addr2,
entry             291 drivers/net/wireless/intersil/hostap/hostap_80211_rx.c 		if (entry != NULL) {
entry             292 drivers/net/wireless/intersil/hostap/hostap_80211_rx.c 			entry->last_frag = frag;
entry             293 drivers/net/wireless/intersil/hostap/hostap_80211_rx.c 			skb = entry->skb;
entry             307 drivers/net/wireless/intersil/hostap/hostap_80211_rx.c 	struct prism2_frag_entry *entry;
entry             312 drivers/net/wireless/intersil/hostap/hostap_80211_rx.c 	entry = prism2_frag_cache_find(local, seq, -1, hdr->addr2, hdr->addr1);
entry             314 drivers/net/wireless/intersil/hostap/hostap_80211_rx.c 	if (entry == NULL) {
entry             321 drivers/net/wireless/intersil/hostap/hostap_80211_rx.c 	entry->skb = NULL;
entry             325 drivers/net/wireless/intersil/hostap/hostap_ap.c 	struct mac_entry *entry;
entry             348 drivers/net/wireless/intersil/hostap/hostap_ap.c 	entry = v;
entry             349 drivers/net/wireless/intersil/hostap/hostap_ap.c 	seq_printf(m, "%pM\n", entry->addr);
entry             381 drivers/net/wireless/intersil/hostap/hostap_ap.c 	struct mac_entry *entry;
entry             383 drivers/net/wireless/intersil/hostap/hostap_ap.c 	entry = kmalloc(sizeof(struct mac_entry), GFP_KERNEL);
entry             384 drivers/net/wireless/intersil/hostap/hostap_ap.c 	if (entry == NULL)
entry             387 drivers/net/wireless/intersil/hostap/hostap_ap.c 	memcpy(entry->addr, mac, ETH_ALEN);
entry             390 drivers/net/wireless/intersil/hostap/hostap_ap.c 	list_add_tail(&entry->list, &mac_restrictions->mac_list);
entry             401 drivers/net/wireless/intersil/hostap/hostap_ap.c 	struct mac_entry *entry;
entry             406 drivers/net/wireless/intersil/hostap/hostap_ap.c 		entry = list_entry(ptr, struct mac_entry, list);
entry             408 drivers/net/wireless/intersil/hostap/hostap_ap.c 		if (ether_addr_equal(entry->addr, mac)) {
entry             410 drivers/net/wireless/intersil/hostap/hostap_ap.c 			kfree(entry);
entry             424 drivers/net/wireless/intersil/hostap/hostap_ap.c 	struct mac_entry *entry;
entry             431 drivers/net/wireless/intersil/hostap/hostap_ap.c 	list_for_each_entry(entry, &mac_restrictions->mac_list, list) {
entry             432 drivers/net/wireless/intersil/hostap/hostap_ap.c 		if (ether_addr_equal(entry->addr, mac)) {
entry             449 drivers/net/wireless/intersil/hostap/hostap_ap.c 	struct mac_entry *entry;
entry             458 drivers/net/wireless/intersil/hostap/hostap_ap.c 		entry = list_entry(ptr, struct mac_entry, list);
entry             460 drivers/net/wireless/intersil/hostap/hostap_ap.c 		kfree(entry);
entry            1069 drivers/net/wireless/intersil/hostap/hostap_ap.c 	struct add_sta_proc_data *entry, *prev;
entry            1071 drivers/net/wireless/intersil/hostap/hostap_ap.c 	entry = ap->add_sta_proc_entries;
entry            1074 drivers/net/wireless/intersil/hostap/hostap_ap.c 	while (entry) {
entry            1076 drivers/net/wireless/intersil/hostap/hostap_ap.c 		sta = ap_get_sta(ap, entry->addr);
entry            1090 drivers/net/wireless/intersil/hostap/hostap_ap.c 		prev = entry;
entry            1091 drivers/net/wireless/intersil/hostap/hostap_ap.c 		entry = entry->next;
entry            1120 drivers/net/wireless/intersil/hostap/hostap_ap.c 		struct add_sta_proc_data *entry;
entry            1123 drivers/net/wireless/intersil/hostap/hostap_ap.c 		entry = kmalloc(sizeof(*entry), GFP_ATOMIC);
entry            1124 drivers/net/wireless/intersil/hostap/hostap_ap.c 		if (entry) {
entry            1125 drivers/net/wireless/intersil/hostap/hostap_ap.c 			memcpy(entry->addr, sta->addr, ETH_ALEN);
entry            1126 drivers/net/wireless/intersil/hostap/hostap_ap.c 			entry->next = ap->add_sta_proc_entries;
entry            1127 drivers/net/wireless/intersil/hostap/hostap_ap.c 			ap->add_sta_proc_entries = entry;
entry            1956 drivers/net/wireless/intersil/hostap/hostap_ap.c 	struct wds_oper_data *entry, *prev;
entry            1959 drivers/net/wireless/intersil/hostap/hostap_ap.c 	entry = local->ap->wds_oper_entries;
entry            1963 drivers/net/wireless/intersil/hostap/hostap_ap.c 	while (entry) {
entry            1967 drivers/net/wireless/intersil/hostap/hostap_ap.c 		       entry->type == WDS_ADD ? "adding" : "removing",
entry            1968 drivers/net/wireless/intersil/hostap/hostap_ap.c 		       entry->addr);
entry            1969 drivers/net/wireless/intersil/hostap/hostap_ap.c 		if (entry->type == WDS_ADD)
entry            1970 drivers/net/wireless/intersil/hostap/hostap_ap.c 			prism2_wds_add(local, entry->addr, 0);
entry            1971 drivers/net/wireless/intersil/hostap/hostap_ap.c 		else if (entry->type == WDS_DEL)
entry            1972 drivers/net/wireless/intersil/hostap/hostap_ap.c 			prism2_wds_del(local, entry->addr, 0, 1);
entry            1974 drivers/net/wireless/intersil/hostap/hostap_ap.c 		prev = entry;
entry            1975 drivers/net/wireless/intersil/hostap/hostap_ap.c 		entry = entry->next;
entry            3255 drivers/net/wireless/intersil/hostap/hostap_ap.c 	struct wds_oper_data *entry;
entry            3257 drivers/net/wireless/intersil/hostap/hostap_ap.c 	entry = kmalloc(sizeof(*entry), GFP_ATOMIC);
entry            3258 drivers/net/wireless/intersil/hostap/hostap_ap.c 	if (!entry)
entry            3260 drivers/net/wireless/intersil/hostap/hostap_ap.c 	memcpy(entry->addr, addr, ETH_ALEN);
entry            3261 drivers/net/wireless/intersil/hostap/hostap_ap.c 	entry->type = type;
entry            3263 drivers/net/wireless/intersil/hostap/hostap_ap.c 	entry->next = local->ap->wds_oper_entries;
entry            3264 drivers/net/wireless/intersil/hostap/hostap_ap.c 	local->ap->wds_oper_entries = entry;
entry             171 drivers/net/wireless/intersil/hostap/hostap_hw.c 					   struct hostap_cmd_queue *entry,
entry             175 drivers/net/wireless/intersil/hostap/hostap_hw.c 		entry->del_req = 1;
entry             176 drivers/net/wireless/intersil/hostap/hostap_hw.c 		if (!list_empty(&entry->list)) {
entry             177 drivers/net/wireless/intersil/hostap/hostap_hw.c 			list_del_init(&entry->list);
entry             182 drivers/net/wireless/intersil/hostap/hostap_hw.c 	if (refcount_dec_and_test(&entry->usecnt) && entry->del_req)
entry             183 drivers/net/wireless/intersil/hostap/hostap_hw.c 		kfree(entry);
entry             196 drivers/net/wireless/intersil/hostap/hostap_hw.c 					 struct hostap_cmd_queue *entry,
entry             202 drivers/net/wireless/intersil/hostap/hostap_hw.c 	__hostap_cmd_queue_free(local, entry, del_req);
entry             215 drivers/net/wireless/intersil/hostap/hostap_hw.c 	struct hostap_cmd_queue *entry;
entry             219 drivers/net/wireless/intersil/hostap/hostap_hw.c 		entry = list_entry(ptr, struct hostap_cmd_queue, list);
entry             220 drivers/net/wireless/intersil/hostap/hostap_hw.c 		refcount_inc(&entry->usecnt);
entry             223 drivers/net/wireless/intersil/hostap/hostap_hw.c 		       local->dev->name, entry->type, entry->cmd,
entry             224 drivers/net/wireless/intersil/hostap/hostap_hw.c 		       entry->param0);
entry             225 drivers/net/wireless/intersil/hostap/hostap_hw.c 		__hostap_cmd_queue_free(local, entry, 1);
entry             244 drivers/net/wireless/intersil/hostap/hostap_hw.c 				    struct hostap_cmd_queue *entry)
entry             258 drivers/net/wireless/intersil/hostap/hostap_hw.c 	if (entry->issued) {
entry             260 drivers/net/wireless/intersil/hostap/hostap_hw.c 		       dev->name, entry);
entry             288 drivers/net/wireless/intersil/hostap/hostap_hw.c 	HFA384X_OUTW(entry->param0, HFA384X_PARAM0_OFF);
entry             289 drivers/net/wireless/intersil/hostap/hostap_hw.c 	HFA384X_OUTW(entry->param1, HFA384X_PARAM1_OFF);
entry             290 drivers/net/wireless/intersil/hostap/hostap_hw.c 	HFA384X_OUTW(entry->cmd, HFA384X_CMD_OFF);
entry             291 drivers/net/wireless/intersil/hostap/hostap_hw.c 	entry->issued = 1;
entry             317 drivers/net/wireless/intersil/hostap/hostap_hw.c 	struct hostap_cmd_queue *entry;
entry             338 drivers/net/wireless/intersil/hostap/hostap_hw.c 	entry = kzalloc(sizeof(*entry), GFP_ATOMIC);
entry             339 drivers/net/wireless/intersil/hostap/hostap_hw.c 	if (entry == NULL)
entry             342 drivers/net/wireless/intersil/hostap/hostap_hw.c 	refcount_set(&entry->usecnt, 1);
entry             343 drivers/net/wireless/intersil/hostap/hostap_hw.c 	entry->type = CMD_SLEEP;
entry             344 drivers/net/wireless/intersil/hostap/hostap_hw.c 	entry->cmd = cmd;
entry             345 drivers/net/wireless/intersil/hostap/hostap_hw.c 	entry->param0 = param0;
entry             347 drivers/net/wireless/intersil/hostap/hostap_hw.c 		entry->param1 = *param1;
entry             348 drivers/net/wireless/intersil/hostap/hostap_hw.c 	init_waitqueue_head(&entry->compl);
entry             352 drivers/net/wireless/intersil/hostap/hostap_hw.c 	add_wait_queue(&entry->compl, &wait);
entry             358 drivers/net/wireless/intersil/hostap/hostap_hw.c 		entry->issuing = 1;
entry             359 drivers/net/wireless/intersil/hostap/hostap_hw.c 	list_add_tail(&entry->list, &local->cmd_queue);
entry             371 drivers/net/wireless/intersil/hostap/hostap_hw.c 		if (hfa384x_cmd_issue(dev, entry))
entry             378 drivers/net/wireless/intersil/hostap/hostap_hw.c 	if (!err && entry->type != CMD_COMPLETED) {
entry             397 drivers/net/wireless/intersil/hostap/hostap_hw.c 	remove_wait_queue(&entry->compl, &wait);
entry             417 drivers/net/wireless/intersil/hostap/hostap_hw.c 	if (!list_empty(&entry->list)) {
entry             419 drivers/net/wireless/intersil/hostap/hostap_hw.c 		       "(entry=%p, type=%d, res=%d)\n", dev->name, entry,
entry             420 drivers/net/wireless/intersil/hostap/hostap_hw.c 		       entry->type, res);
entry             421 drivers/net/wireless/intersil/hostap/hostap_hw.c 		list_del_init(&entry->list);
entry             433 drivers/net/wireless/intersil/hostap/hostap_hw.c 	if (entry->type != CMD_COMPLETED) {
entry             438 drivers/net/wireless/intersil/hostap/hostap_hw.c 		       res, entry, entry->type, entry->cmd, entry->param0, reg,
entry             453 drivers/net/wireless/intersil/hostap/hostap_hw.c 		*resp0 = entry->resp0;
entry             455 drivers/net/wireless/intersil/hostap/hostap_hw.c 	if (entry->res) {
entry             458 drivers/net/wireless/intersil/hostap/hostap_hw.c 		       dev->name, cmd, entry->res, entry->resp0);
entry             462 drivers/net/wireless/intersil/hostap/hostap_hw.c 	res = entry->res;
entry             464 drivers/net/wireless/intersil/hostap/hostap_hw.c 	hostap_cmd_queue_free(local, entry, 1);
entry             493 drivers/net/wireless/intersil/hostap/hostap_hw.c 	struct hostap_cmd_queue *entry;
entry             504 drivers/net/wireless/intersil/hostap/hostap_hw.c 	entry = kzalloc(sizeof(*entry), GFP_ATOMIC);
entry             505 drivers/net/wireless/intersil/hostap/hostap_hw.c 	if (entry == NULL)
entry             508 drivers/net/wireless/intersil/hostap/hostap_hw.c 	refcount_set(&entry->usecnt, 1);
entry             509 drivers/net/wireless/intersil/hostap/hostap_hw.c 	entry->type = CMD_CALLBACK;
entry             510 drivers/net/wireless/intersil/hostap/hostap_hw.c 	entry->cmd = cmd;
entry             511 drivers/net/wireless/intersil/hostap/hostap_hw.c 	entry->param0 = param0;
entry             512 drivers/net/wireless/intersil/hostap/hostap_hw.c 	entry->callback = callback;
entry             513 drivers/net/wireless/intersil/hostap/hostap_hw.c 	entry->context = context;
entry             518 drivers/net/wireless/intersil/hostap/hostap_hw.c 		entry->issuing = 1;
entry             519 drivers/net/wireless/intersil/hostap/hostap_hw.c 	list_add_tail(&entry->list, &local->cmd_queue);
entry             523 drivers/net/wireless/intersil/hostap/hostap_hw.c 	if (issue && hfa384x_cmd_issue(dev, entry))
entry             528 drivers/net/wireless/intersil/hostap/hostap_hw.c 	hostap_cmd_queue_free(local, entry, ret);
entry             649 drivers/net/wireless/intersil/hostap/hostap_hw.c 	struct hostap_cmd_queue *entry = NULL;
entry             656 drivers/net/wireless/intersil/hostap/hostap_hw.c 		entry = list_entry(local->cmd_queue.next,
entry             658 drivers/net/wireless/intersil/hostap/hostap_hw.c 		refcount_inc(&entry->usecnt);
entry             659 drivers/net/wireless/intersil/hostap/hostap_hw.c 		list_del_init(&entry->list);
entry             662 drivers/net/wireless/intersil/hostap/hostap_hw.c 		if (!entry->issued) {
entry             665 drivers/net/wireless/intersil/hostap/hostap_hw.c 			__hostap_cmd_queue_free(local, entry, 1);
entry             666 drivers/net/wireless/intersil/hostap/hostap_hw.c 			entry = NULL;
entry             671 drivers/net/wireless/intersil/hostap/hostap_hw.c 	if (!entry) {
entry             678 drivers/net/wireless/intersil/hostap/hostap_hw.c 	entry->resp0 = HFA384X_INW(HFA384X_RESP0_OFF);
entry             679 drivers/net/wireless/intersil/hostap/hostap_hw.c 	entry->res = (HFA384X_INW(HFA384X_STATUS_OFF) &
entry             685 drivers/net/wireless/intersil/hostap/hostap_hw.c 	if (entry->type == CMD_SLEEP) {
entry             686 drivers/net/wireless/intersil/hostap/hostap_hw.c 		entry->type = CMD_COMPLETED;
entry             687 drivers/net/wireless/intersil/hostap/hostap_hw.c 		wake_up_interruptible(&entry->compl);
entry             688 drivers/net/wireless/intersil/hostap/hostap_hw.c 	} else if (entry->type == CMD_CALLBACK) {
entry             689 drivers/net/wireless/intersil/hostap/hostap_hw.c 		if (entry->callback)
entry             690 drivers/net/wireless/intersil/hostap/hostap_hw.c 			entry->callback(dev, entry->context, entry->resp0,
entry             691 drivers/net/wireless/intersil/hostap/hostap_hw.c 					entry->res);
entry             694 drivers/net/wireless/intersil/hostap/hostap_hw.c 		       dev->name, entry->type);
entry             696 drivers/net/wireless/intersil/hostap/hostap_hw.c 	hostap_cmd_queue_free(local, entry, 1);
entry             699 drivers/net/wireless/intersil/hostap/hostap_hw.c 	entry = NULL;
entry             702 drivers/net/wireless/intersil/hostap/hostap_hw.c 		entry = list_entry(local->cmd_queue.next,
entry             704 drivers/net/wireless/intersil/hostap/hostap_hw.c 		if (entry->issuing) {
entry             707 drivers/net/wireless/intersil/hostap/hostap_hw.c 			entry = NULL;
entry             709 drivers/net/wireless/intersil/hostap/hostap_hw.c 		if (entry)
entry             710 drivers/net/wireless/intersil/hostap/hostap_hw.c 			refcount_inc(&entry->usecnt);
entry             714 drivers/net/wireless/intersil/hostap/hostap_hw.c 	if (entry) {
entry             717 drivers/net/wireless/intersil/hostap/hostap_hw.c 		int res = hfa384x_cmd_issue(dev, entry);
entry             719 drivers/net/wireless/intersil/hostap/hostap_hw.c 		__hostap_cmd_queue_free(local, entry, res);
entry            2975 drivers/net/wireless/intersil/hostap/hostap_hw.c 		struct set_tim_data *entry =
entry            2977 drivers/net/wireless/intersil/hostap/hostap_hw.c 		if (entry->aid == aid) {
entry            2980 drivers/net/wireless/intersil/hostap/hostap_hw.c 			       local->dev->name, aid, entry->set, set);
entry            2981 drivers/net/wireless/intersil/hostap/hostap_hw.c 			entry->set = set;
entry            3000 drivers/net/wireless/intersil/hostap/hostap_hw.c 	struct set_tim_data *entry;
entry            3004 drivers/net/wireless/intersil/hostap/hostap_hw.c 		entry = NULL;
entry            3007 drivers/net/wireless/intersil/hostap/hostap_hw.c 			entry = list_entry(local->set_tim_list.next,
entry            3009 drivers/net/wireless/intersil/hostap/hostap_hw.c 			list_del(&entry->list);
entry            3012 drivers/net/wireless/intersil/hostap/hostap_hw.c 		if (!entry)
entry            3016 drivers/net/wireless/intersil/hostap/hostap_hw.c 		       local->dev->name, entry->aid, entry->set);
entry            3018 drivers/net/wireless/intersil/hostap/hostap_hw.c 		val = entry->aid;
entry            3019 drivers/net/wireless/intersil/hostap/hostap_hw.c 		if (entry->set)
entry            3024 drivers/net/wireless/intersil/hostap/hostap_hw.c 			       local->dev->name, entry->aid, entry->set);
entry            3027 drivers/net/wireless/intersil/hostap/hostap_hw.c 		kfree(entry);
entry            3037 drivers/net/wireless/intersil/hostap/hostap_hw.c 		struct set_tim_data *entry;
entry            3038 drivers/net/wireless/intersil/hostap/hostap_hw.c 		entry = list_entry(ptr, struct set_tim_data, list);
entry            3039 drivers/net/wireless/intersil/hostap/hostap_hw.c 		list_del(&entry->list);
entry            3040 drivers/net/wireless/intersil/hostap/hostap_hw.c 		kfree(entry);
entry             172 drivers/net/wireless/intersil/hostap/hostap_info.c 	struct hfa384x_hostscan_result *selected, *entry;
entry             210 drivers/net/wireless/intersil/hostap/hostap_info.c 			entry = &local->last_scan_results[i];
entry             211 drivers/net/wireless/intersil/hostap/hostap_info.c 			if (memcmp(local->preferred_ap, entry->bssid, 6) == 0)
entry             215 drivers/net/wireless/intersil/hostap/hostap_info.c 				selected = entry;
entry             646 drivers/net/wireless/intersil/hostap/hostap_ioctl.c 	struct hfa384x_hostscan_result *entry;
entry             658 drivers/net/wireless/intersil/hostap/hostap_ioctl.c 		entry = &local->last_scan_results[i];
entry             659 drivers/net/wireless/intersil/hostap/hostap_ioctl.c 		if (ether_addr_equal(local->preferred_ap, entry->bssid)) {
entry             660 drivers/net/wireless/intersil/hostap/hostap_ioctl.c 			req.channel = entry->chid;
entry            1958 drivers/net/wireless/intersil/hostap/hostap_ioctl.c 	int entry, hostscan;
entry            1972 drivers/net/wireless/intersil/hostap/hostap_ioctl.c 	for (entry = 0; entry < local->last_scan_results_count; entry++) {
entry            1974 drivers/net/wireless/intersil/hostap/hostap_ioctl.c 		scan = &local->last_scan_results[entry];
entry             239 drivers/net/wireless/intersil/hostap/hostap_main.c 	struct hostap_tx_callback_info *entry;
entry             241 drivers/net/wireless/intersil/hostap/hostap_main.c 	entry = kmalloc(sizeof(*entry), GFP_KERNEL);
entry             242 drivers/net/wireless/intersil/hostap/hostap_main.c 	if (entry == NULL)
entry             245 drivers/net/wireless/intersil/hostap/hostap_main.c 	entry->func = func;
entry             246 drivers/net/wireless/intersil/hostap/hostap_main.c 	entry->data = data;
entry             249 drivers/net/wireless/intersil/hostap/hostap_main.c 	entry->idx = local->tx_callback ? local->tx_callback->idx + 1 : 1;
entry             250 drivers/net/wireless/intersil/hostap/hostap_main.c 	entry->next = local->tx_callback;
entry             251 drivers/net/wireless/intersil/hostap/hostap_main.c 	local->tx_callback = entry;
entry             254 drivers/net/wireless/intersil/hostap/hostap_main.c 	return entry->idx;
entry             273 drivers/net/wireless/intersil/hostap/hostap_proc.c 	unsigned long entry;
entry             284 drivers/net/wireless/intersil/hostap/hostap_proc.c 	entry = (unsigned long)v - 2;
entry             285 drivers/net/wireless/intersil/hostap/hostap_proc.c 	scanres = &local->last_scan_results[entry];
entry             236 drivers/net/wireless/intersil/p54/eeprom.c 	struct p54_channel_entry *entry = NULL;
entry             245 drivers/net/wireless/intersil/p54/eeprom.c 			entry = &list->channels[i];
entry             262 drivers/net/wireless/intersil/p54/eeprom.c 			entry = &list->channels[i];
entry             263 drivers/net/wireless/intersil/p54/eeprom.c 			entry->freq = freq;
entry             264 drivers/net/wireless/intersil/p54/eeprom.c 			entry->band = band;
entry             265 drivers/net/wireless/intersil/p54/eeprom.c 			entry->index = ieee80211_frequency_to_channel(freq);
entry             266 drivers/net/wireless/intersil/p54/eeprom.c 			entry->max_power = 0;
entry             267 drivers/net/wireless/intersil/p54/eeprom.c 			entry->data = 0;
entry             271 drivers/net/wireless/intersil/p54/eeprom.c 	if (entry)
entry             272 drivers/net/wireless/intersil/p54/eeprom.c 		entry->data |= data;
entry             274 drivers/net/wireless/intersil/p54/eeprom.c 	return entry;
entry             521 drivers/net/wireless/intersil/p54/eeprom.c 	struct p54_rssi_db_entry *entry;
entry             550 drivers/net/wireless/intersil/p54/eeprom.c 	db_len = sizeof(*entry) * entries;
entry             557 drivers/net/wireless/intersil/p54/eeprom.c 	priv->rssi_db->entry_size = sizeof(*entry);
entry             560 drivers/net/wireless/intersil/p54/eeprom.c 	entry = (void *)((unsigned long)priv->rssi_db->data + priv->rssi_db->offset);
entry             565 drivers/net/wireless/intersil/p54/eeprom.c 			entry[i].freq = le16_to_cpu(cal[i].freq);
entry             566 drivers/net/wireless/intersil/p54/eeprom.c 			entry[i].mul = (s16) le16_to_cpu(cal[i].mul);
entry             567 drivers/net/wireless/intersil/p54/eeprom.c 			entry[i].add = (s16) le16_to_cpu(cal[i].add);
entry             583 drivers/net/wireless/intersil/p54/eeprom.c 			entry[i].freq = freq;
entry             584 drivers/net/wireless/intersil/p54/eeprom.c 			entry[i].mul = (s16) le16_to_cpu(cal[i].mul);
entry             585 drivers/net/wireless/intersil/p54/eeprom.c 			entry[i].add = (s16) le16_to_cpu(cal[i].add);
entry             590 drivers/net/wireless/intersil/p54/eeprom.c 	sort(entry, entries, sizeof(*entry), p54_compare_rssichan, NULL);
entry             606 drivers/net/wireless/intersil/p54/eeprom.c 	struct p54_rssi_db_entry *entry;
entry             612 drivers/net/wireless/intersil/p54/eeprom.c 	entry = (void *)(priv->rssi_db->data + priv->rssi_db->offset);
entry             614 drivers/net/wireless/intersil/p54/eeprom.c 		if (!same_band(freq, entry[i].freq))
entry             623 drivers/net/wireless/intersil/p54/eeprom.c 		if (abs(freq - entry[i].freq) <
entry             624 drivers/net/wireless/intersil/p54/eeprom.c 		    abs(freq - entry[found].freq)) {
entry             632 drivers/net/wireless/intersil/p54/eeprom.c 	return found < 0 ? &p54_rssi_default : &entry[found];
entry             733 drivers/net/wireless/intersil/p54/eeprom.c 	struct pda_entry *entry;
entry             742 drivers/net/wireless/intersil/p54/eeprom.c 	entry = (void *)wrap->data + le16_to_cpu(wrap->len);
entry             745 drivers/net/wireless/intersil/p54/eeprom.c 	while ((u8 *)entry <= end - sizeof(*entry)) {
entry             746 drivers/net/wireless/intersil/p54/eeprom.c 		entry_len = le16_to_cpu(entry->len);
entry             750 drivers/net/wireless/intersil/p54/eeprom.c 		if ((u8 *)entry + sizeof(*entry) + data_len > end)
entry             753 drivers/net/wireless/intersil/p54/eeprom.c 		switch (le16_to_cpu(entry->code)) {
entry             757 drivers/net/wireless/intersil/p54/eeprom.c 			SET_IEEE80211_PERM_ADDR(dev, entry->data);
entry             762 drivers/net/wireless/intersil/p54/eeprom.c 			err = p54_convert_output_limits(dev, entry->data,
entry             769 drivers/net/wireless/intersil/p54/eeprom.c 				(struct pda_pa_curve_data *)entry->data;
entry             794 drivers/net/wireless/intersil/p54/eeprom.c 			priv->iq_autocal = kmemdup(entry->data, data_len,
entry             804 drivers/net/wireless/intersil/p54/eeprom.c 			p54_parse_default_country(dev, entry->data, data_len);
entry             807 drivers/net/wireless/intersil/p54/eeprom.c 			tmp = entry->data;
entry             808 drivers/net/wireless/intersil/p54/eeprom.c 			while ((u8 *)tmp < entry->data + data_len) {
entry             818 drivers/net/wireless/intersil/p54/eeprom.c 			priv->version = *(u8 *)(entry->data + 1);
entry             823 drivers/net/wireless/intersil/p54/eeprom.c 			err = p54_parse_rssical(dev, entry->data, data_len,
entry             824 drivers/net/wireless/intersil/p54/eeprom.c 						le16_to_cpu(entry->code));
entry             829 drivers/net/wireless/intersil/p54/eeprom.c 			struct pda_custom_wrapper *pda = (void *) entry->data;
entry             850 drivers/net/wireless/intersil/p54/eeprom.c 			struct pda_custom_wrapper *pda = (void *) entry->data;
entry             857 drivers/net/wireless/intersil/p54/eeprom.c 			struct pda_custom_wrapper *pda = (void *) entry->data;
entry             864 drivers/net/wireless/intersil/p54/eeprom.c 			crc16 = ~crc_ccitt(crc16, (u8 *) entry, sizeof(*entry));
entry             865 drivers/net/wireless/intersil/p54/eeprom.c 			if (crc16 != le16_to_cpup((__le16 *)entry->data)) {
entry             878 drivers/net/wireless/intersil/p54/eeprom.c 		crc16 = crc_ccitt(crc16, (u8 *)entry, (entry_len + 1) * 2);
entry             879 drivers/net/wireless/intersil/p54/eeprom.c 		entry = (void *)entry + (entry_len + 1) * 2;
entry             274 drivers/net/wireless/intersil/p54/fwio.c 	tim->entry[0] = cpu_to_le16(set ? (aid | 0x8000) : aid);
entry             402 drivers/net/wireless/intersil/p54/fwio.c 	void *entry;
entry             470 drivers/net/wireless/intersil/p54/fwio.c 	entry = (void *)(priv->curve_data->data + priv->curve_data->offset);
entry             472 drivers/net/wireless/intersil/p54/fwio.c 		if (*((__le16 *)entry) != freq) {
entry             473 drivers/net/wireless/intersil/p54/fwio.c 			entry += priv->curve_data->entry_size;
entry             479 drivers/net/wireless/intersil/p54/fwio.c 				entry + sizeof(__le16),
entry             486 drivers/net/wireless/intersil/p54/fwio.c 			entry += sizeof(__le16);
entry             489 drivers/net/wireless/intersil/p54/fwio.c 			memcpy(chan->curve_data, entry,
entry             664 drivers/net/wireless/intersil/p54/fwio.c 	rxkey->entry = slot;
entry             437 drivers/net/wireless/intersil/p54/lmac.h 	u8 entry;
entry             503 drivers/net/wireless/intersil/p54/lmac.h 	__le16 entry[8];
entry             422 drivers/net/wireless/intersil/p54/p54spi.c 	struct p54s_tx_info *entry;
entry             433 drivers/net/wireless/intersil/p54/p54spi.c 		entry = list_entry(priv->tx_pending.next,
entry             436 drivers/net/wireless/intersil/p54/p54spi.c 		list_del_init(&entry->tx_list);
entry             440 drivers/net/wireless/intersil/p54/p54spi.c 		dinfo = container_of((void *) entry, struct p54s_tx_info,
entry             225 drivers/net/wireless/intersil/p54/p54usb.c 	struct urb *entry = NULL;
entry             236 drivers/net/wireless/intersil/p54/p54usb.c 		entry = usb_alloc_urb(0, GFP_KERNEL);
entry             237 drivers/net/wireless/intersil/p54/p54usb.c 		if (!entry) {
entry             242 drivers/net/wireless/intersil/p54/p54usb.c 		usb_fill_bulk_urb(entry, priv->udev,
entry             247 drivers/net/wireless/intersil/p54/p54usb.c 		info->urb = entry;
entry             251 drivers/net/wireless/intersil/p54/p54usb.c 		usb_anchor_urb(entry, &priv->submitted);
entry             252 drivers/net/wireless/intersil/p54/p54usb.c 		ret = usb_submit_urb(entry, GFP_KERNEL);
entry             255 drivers/net/wireless/intersil/p54/p54usb.c 			usb_unanchor_urb(entry);
entry             258 drivers/net/wireless/intersil/p54/p54usb.c 		usb_free_urb(entry);
entry             259 drivers/net/wireless/intersil/p54/p54usb.c 		entry = NULL;
entry             265 drivers/net/wireless/intersil/p54/p54usb.c 	usb_free_urb(entry);
entry              81 drivers/net/wireless/intersil/p54/txrx.c 	struct sk_buff *entry, *target_skb = NULL;
entry             105 drivers/net/wireless/intersil/p54/txrx.c 	skb_queue_walk(&priv->tx_queue, entry) {
entry             107 drivers/net/wireless/intersil/p54/txrx.c 		info = IEEE80211_SKB_CB(entry);
entry             112 drivers/net/wireless/intersil/p54/txrx.c 			target_skb = entry->prev;
entry             251 drivers/net/wireless/intersil/p54/txrx.c 	struct sk_buff *entry;
entry             255 drivers/net/wireless/intersil/p54/txrx.c 	skb_queue_walk(&priv->tx_queue, entry) {
entry             256 drivers/net/wireless/intersil/p54/txrx.c 		struct p54_hdr *hdr = (struct p54_hdr *) entry->data;
entry             259 drivers/net/wireless/intersil/p54/txrx.c 			__skb_unlink(entry, &priv->tx_queue);
entry             261 drivers/net/wireless/intersil/p54/txrx.c 			p54_tx_qos_accounting_free(priv, entry);
entry             262 drivers/net/wireless/intersil/p54/txrx.c 			return entry;
entry             406 drivers/net/wireless/intersil/p54/txrx.c 	struct sk_buff *entry;
entry             410 drivers/net/wireless/intersil/p54/txrx.c 	entry = p54_find_and_unlink_skb(priv, hdr->req_id);
entry             411 drivers/net/wireless/intersil/p54/txrx.c 	if (unlikely(!entry))
entry             414 drivers/net/wireless/intersil/p54/txrx.c 	frame_len = entry->len;
entry             415 drivers/net/wireless/intersil/p54/txrx.c 	info = IEEE80211_SKB_CB(entry);
entry             416 drivers/net/wireless/intersil/p54/txrx.c 	entry_hdr = (struct p54_hdr *) entry->data;
entry             426 drivers/net/wireless/intersil/p54/txrx.c 		dev_kfree_skb_any(entry);
entry             487 drivers/net/wireless/intersil/p54/txrx.c 	skb_trim(entry, frame_len);
entry             488 drivers/net/wireless/intersil/p54/txrx.c 	skb_pull(entry, sizeof(*hdr) + pad + sizeof(*entry_data));
entry             489 drivers/net/wireless/intersil/p54/txrx.c 	ieee80211_tx_status_irqsafe(priv->hw, entry);
entry            1796 drivers/net/wireless/intersil/prism54/isl_ioctl.c 	struct mac_entry *entry;
entry            1807 drivers/net/wireless/intersil/prism54/isl_ioctl.c 		entry = list_entry(ptr, struct mac_entry, _list);
entry            1809 drivers/net/wireless/intersil/prism54/isl_ioctl.c 		kfree(entry);
entry            1827 drivers/net/wireless/intersil/prism54/isl_ioctl.c 	struct mac_entry *entry;
entry            1833 drivers/net/wireless/intersil/prism54/isl_ioctl.c 	entry = kmalloc(sizeof (struct mac_entry), GFP_KERNEL);
entry            1834 drivers/net/wireless/intersil/prism54/isl_ioctl.c 	if (entry == NULL)
entry            1837 drivers/net/wireless/intersil/prism54/isl_ioctl.c 	memcpy(entry->addr, addr->sa_data, ETH_ALEN);
entry            1840 drivers/net/wireless/intersil/prism54/isl_ioctl.c 		kfree(entry);
entry            1843 drivers/net/wireless/intersil/prism54/isl_ioctl.c 	list_add_tail(&entry->_list, &acl->mac_list);
entry            1856 drivers/net/wireless/intersil/prism54/isl_ioctl.c 	struct mac_entry *entry;
entry            1864 drivers/net/wireless/intersil/prism54/isl_ioctl.c 	list_for_each_entry(entry, &acl->mac_list, _list) {
entry            1865 drivers/net/wireless/intersil/prism54/isl_ioctl.c 		if (ether_addr_equal(entry->addr, addr->sa_data)) {
entry            1866 drivers/net/wireless/intersil/prism54/isl_ioctl.c 			list_del(&entry->_list);
entry            1868 drivers/net/wireless/intersil/prism54/isl_ioctl.c 			kfree(entry);
entry            1883 drivers/net/wireless/intersil/prism54/isl_ioctl.c 	struct mac_entry *entry;
entry            1891 drivers/net/wireless/intersil/prism54/isl_ioctl.c 	list_for_each_entry(entry, &acl->mac_list, _list) {
entry            1892 drivers/net/wireless/intersil/prism54/isl_ioctl.c 		memcpy(dst->sa_data, entry->addr, ETH_ALEN);
entry            1958 drivers/net/wireless/intersil/prism54/isl_ioctl.c 	struct mac_entry *entry;
entry            1969 drivers/net/wireless/intersil/prism54/isl_ioctl.c 	list_for_each_entry(entry, &acl->mac_list, _list) {
entry            1970 drivers/net/wireless/intersil/prism54/isl_ioctl.c 		if (memcmp(entry->addr, mac, ETH_ALEN) == 0) {
entry            3783 drivers/net/wireless/mac80211_hwsim.c 	struct mac80211_hwsim_data *entry, *tmp;
entry            3787 drivers/net/wireless/mac80211_hwsim.c 	list_for_each_entry_safe(entry, tmp, &hwsim_radios, list) {
entry            3788 drivers/net/wireless/mac80211_hwsim.c 		if (entry->destroy_on_close && entry->portid == portid) {
entry            3789 drivers/net/wireless/mac80211_hwsim.c 			list_move(&entry->list, &list);
entry            3790 drivers/net/wireless/mac80211_hwsim.c 			rhashtable_remove_fast(&hwsim_radios_rht, &entry->rht,
entry            3797 drivers/net/wireless/mac80211_hwsim.c 	list_for_each_entry_safe(entry, tmp, &list, list) {
entry            3798 drivers/net/wireless/mac80211_hwsim.c 		list_del(&entry->list);
entry            3799 drivers/net/wireless/mac80211_hwsim.c 		mac80211_hwsim_del_radio(entry, wiphy_name(entry->hw->wiphy),
entry            1204 drivers/net/wireless/marvell/mwifiex/main.c 		struct memory_type_mapping *entry =
entry            1207 drivers/net/wireless/marvell/mwifiex/main.c 		if (entry->mem_ptr) {
entry            1209 drivers/net/wireless/marvell/mwifiex/main.c 					strlen(entry->mem_name) +
entry            1211 drivers/net/wireless/marvell/mwifiex/main.c 					(entry->mem_size + 1) +
entry            1236 drivers/net/wireless/marvell/mwifiex/main.c 		struct memory_type_mapping *entry =
entry            1239 drivers/net/wireless/marvell/mwifiex/main.c 		if (entry->mem_ptr) {
entry            1243 drivers/net/wireless/marvell/mwifiex/main.c 			strcpy(fw_dump_ptr, entry->mem_name);
entry            1244 drivers/net/wireless/marvell/mwifiex/main.c 			fw_dump_ptr += strlen(entry->mem_name);
entry            1249 drivers/net/wireless/marvell/mwifiex/main.c 			memcpy(fw_dump_ptr, entry->mem_ptr, entry->mem_size);
entry            1250 drivers/net/wireless/marvell/mwifiex/main.c 			fw_dump_ptr += entry->mem_size;
entry            1260 drivers/net/wireless/marvell/mwifiex/main.c 		struct memory_type_mapping *entry =
entry            1263 drivers/net/wireless/marvell/mwifiex/main.c 		vfree(entry->mem_ptr);
entry            1264 drivers/net/wireless/marvell/mwifiex/main.c 		entry->mem_ptr = NULL;
entry            1265 drivers/net/wireless/marvell/mwifiex/main.c 		entry->mem_size = 0;
entry            2673 drivers/net/wireless/marvell/mwifiex/pcie.c 		struct memory_type_mapping *entry =
entry            2676 drivers/net/wireless/marvell/mwifiex/pcie.c 		if (entry->mem_ptr) {
entry            2677 drivers/net/wireless/marvell/mwifiex/pcie.c 			vfree(entry->mem_ptr);
entry            2678 drivers/net/wireless/marvell/mwifiex/pcie.c 			entry->mem_ptr = NULL;
entry            2680 drivers/net/wireless/marvell/mwifiex/pcie.c 		entry->mem_size = 0;
entry            2701 drivers/net/wireless/marvell/mwifiex/pcie.c 		struct memory_type_mapping *entry =
entry            2731 drivers/net/wireless/marvell/mwifiex/pcie.c 			    "%s_SIZE=0x%x\n", entry->mem_name, memory_size);
entry            2732 drivers/net/wireless/marvell/mwifiex/pcie.c 		entry->mem_ptr = vmalloc(memory_size + 1);
entry            2733 drivers/net/wireless/marvell/mwifiex/pcie.c 		entry->mem_size = memory_size;
entry            2734 drivers/net/wireless/marvell/mwifiex/pcie.c 		if (!entry->mem_ptr) {
entry            2736 drivers/net/wireless/marvell/mwifiex/pcie.c 				    "Vmalloc %s failed\n", entry->mem_name);
entry            2739 drivers/net/wireless/marvell/mwifiex/pcie.c 		dbg_ptr = entry->mem_ptr;
entry            2742 drivers/net/wireless/marvell/mwifiex/pcie.c 		doneflag = entry->done_flag;
entry            2744 drivers/net/wireless/marvell/mwifiex/pcie.c 			    entry->mem_name);
entry            2765 drivers/net/wireless/marvell/mwifiex/pcie.c 				memcpy(tmp_ptr, entry->mem_ptr, memory_size);
entry            2766 drivers/net/wireless/marvell/mwifiex/pcie.c 				vfree(entry->mem_ptr);
entry            2767 drivers/net/wireless/marvell/mwifiex/pcie.c 				entry->mem_ptr = tmp_ptr;
entry            2769 drivers/net/wireless/marvell/mwifiex/pcie.c 				dbg_ptr = entry->mem_ptr + memory_size;
entry            2771 drivers/net/wireless/marvell/mwifiex/pcie.c 				end_ptr = entry->mem_ptr + memory_size;
entry            2779 drivers/net/wireless/marvell/mwifiex/pcie.c 				    entry->mem_name, dbg_ptr - entry->mem_ptr);
entry            3027 drivers/net/wireless/marvell/mwifiex/pcie.c 			card->msix_entries[i].entry = i;
entry            2313 drivers/net/wireless/marvell/mwifiex/sdio.c 		struct memory_type_mapping *entry = &mem_type_mapping_tbl[idx];
entry            2315 drivers/net/wireless/marvell/mwifiex/sdio.c 		if (entry->mem_ptr) {
entry            2316 drivers/net/wireless/marvell/mwifiex/sdio.c 			vfree(entry->mem_ptr);
entry            2317 drivers/net/wireless/marvell/mwifiex/sdio.c 			entry->mem_ptr = NULL;
entry            2319 drivers/net/wireless/marvell/mwifiex/sdio.c 		entry->mem_size = 0;
entry            2341 drivers/net/wireless/marvell/mwifiex/sdio.c 		struct memory_type_mapping *entry = &mem_type_mapping_tbl[idx];
entry            2372 drivers/net/wireless/marvell/mwifiex/sdio.c 			    "%s_SIZE=0x%x\n", entry->mem_name, memory_size);
entry            2373 drivers/net/wireless/marvell/mwifiex/sdio.c 		entry->mem_ptr = vmalloc(memory_size + 1);
entry            2374 drivers/net/wireless/marvell/mwifiex/sdio.c 		entry->mem_size = memory_size;
entry            2375 drivers/net/wireless/marvell/mwifiex/sdio.c 		if (!entry->mem_ptr) {
entry            2377 drivers/net/wireless/marvell/mwifiex/sdio.c 				    entry->mem_name);
entry            2380 drivers/net/wireless/marvell/mwifiex/sdio.c 		dbg_ptr = entry->mem_ptr;
entry            2383 drivers/net/wireless/marvell/mwifiex/sdio.c 		doneflag = entry->done_flag;
entry            2386 drivers/net/wireless/marvell/mwifiex/sdio.c 			    entry->mem_name);
entry            2413 drivers/net/wireless/marvell/mwifiex/sdio.c 				    entry->mem_name, dbg_ptr - entry->mem_ptr);
entry            2426 drivers/net/wireless/marvell/mwifiex/sdio.c 	struct memory_type_mapping *entry = &generic_mem_type_map[0];
entry            2436 drivers/net/wireless/marvell/mwifiex/sdio.c 	if (entry->mem_ptr) {
entry            2437 drivers/net/wireless/marvell/mwifiex/sdio.c 		vfree(entry->mem_ptr);
entry            2438 drivers/net/wireless/marvell/mwifiex/sdio.c 		entry->mem_ptr = NULL;
entry            2440 drivers/net/wireless/marvell/mwifiex/sdio.c 	entry->mem_size = 0;
entry            2473 drivers/net/wireless/marvell/mwifiex/sdio.c 	entry->mem_ptr = vmalloc(0xf0000 + 1);
entry            2474 drivers/net/wireless/marvell/mwifiex/sdio.c 	if (!entry->mem_ptr) {
entry            2478 drivers/net/wireless/marvell/mwifiex/sdio.c 	dbg_ptr = entry->mem_ptr;
entry            2479 drivers/net/wireless/marvell/mwifiex/sdio.c 	entry->mem_size = 0xf0000;
entry            2480 drivers/net/wireless/marvell/mwifiex/sdio.c 	end_ptr = dbg_ptr + entry->mem_size;
entry            2482 drivers/net/wireless/marvell/mwifiex/sdio.c 	done_flag = entry->done_flag;
entry            2484 drivers/net/wireless/marvell/mwifiex/sdio.c 		    "Start %s output, please wait...\n", entry->mem_name);
entry            2501 drivers/net/wireless/marvell/mwifiex/sdio.c 				tmp_ptr = vmalloc(entry->mem_size + 0x4000 + 1);
entry            2505 drivers/net/wireless/marvell/mwifiex/sdio.c 				memcpy(tmp_ptr, entry->mem_ptr,
entry            2506 drivers/net/wireless/marvell/mwifiex/sdio.c 				       entry->mem_size);
entry            2507 drivers/net/wireless/marvell/mwifiex/sdio.c 				vfree(entry->mem_ptr);
entry            2508 drivers/net/wireless/marvell/mwifiex/sdio.c 				entry->mem_ptr = tmp_ptr;
entry            2510 drivers/net/wireless/marvell/mwifiex/sdio.c 				dbg_ptr = entry->mem_ptr + entry->mem_size;
entry            2511 drivers/net/wireless/marvell/mwifiex/sdio.c 				entry->mem_size += 0x4000;
entry            2512 drivers/net/wireless/marvell/mwifiex/sdio.c 				end_ptr = entry->mem_ptr + entry->mem_size;
entry            2516 drivers/net/wireless/marvell/mwifiex/sdio.c 			entry->mem_size = dbg_ptr - entry->mem_ptr;
entry            2518 drivers/net/wireless/marvell/mwifiex/sdio.c 				    entry->mem_name, entry->mem_size);
entry            2528 drivers/net/wireless/marvell/mwifiex/sdio.c 		if (entry->mem_ptr) {
entry            2529 drivers/net/wireless/marvell/mwifiex/sdio.c 			vfree(entry->mem_ptr);
entry            2530 drivers/net/wireless/marvell/mwifiex/sdio.c 			entry->mem_ptr = NULL;
entry            2532 drivers/net/wireless/marvell/mwifiex/sdio.c 		entry->mem_size = 0;
entry              30 drivers/net/wireless/mediatek/mt76/dma.c 	size = q->ndesc * sizeof(*q->entry);
entry              31 drivers/net/wireless/mediatek/mt76/dma.c 	q->entry = devm_kzalloc(dev->dev, size, GFP_KERNEL);
entry              32 drivers/net/wireless/mediatek/mt76/dma.c 	if (!q->entry)
entry              57 drivers/net/wireless/mediatek/mt76/dma.c 		q->entry[q->head].txwi = DMA_DUMMY_DATA;
entry              58 drivers/net/wireless/mediatek/mt76/dma.c 		q->entry[q->head].skip_buf0 = true;
entry              88 drivers/net/wireless/mediatek/mt76/dma.c 	q->entry[idx].txwi = txwi;
entry              89 drivers/net/wireless/mediatek/mt76/dma.c 	q->entry[idx].skb = skb;
entry              98 drivers/net/wireless/mediatek/mt76/dma.c 	struct mt76_queue_entry *e = &q->entry[idx];
entry             143 drivers/net/wireless/mediatek/mt76/dma.c 	struct mt76_queue_entry entry;
entry             158 drivers/net/wireless/mediatek/mt76/dma.c 		mt76_dma_tx_cleanup_idx(dev, q, q->tail, &entry);
entry             159 drivers/net/wireless/mediatek/mt76/dma.c 		if (entry.schedule)
entry             160 drivers/net/wireless/mediatek/mt76/dma.c 			n_swq_queued[entry.qid]++;
entry             165 drivers/net/wireless/mediatek/mt76/dma.c 		if (entry.skb)
entry             166 drivers/net/wireless/mediatek/mt76/dma.c 			dev->drv->tx_complete_skb(dev, qid, &entry);
entry             168 drivers/net/wireless/mediatek/mt76/dma.c 		if (entry.txwi) {
entry             170 drivers/net/wireless/mediatek/mt76/dma.c 				mt76_put_txwi(dev, entry.txwi);
entry             209 drivers/net/wireless/mediatek/mt76/dma.c 	struct mt76_queue_entry *e = &q->entry[idx];
entry             112 drivers/net/wireless/mediatek/mt76/mt76.h 	struct mt76_queue_entry *entry;
entry             462 drivers/net/wireless/mediatek/mt76/tx.c 		hwq->entry[idx].qid = sq - dev->q_tx;
entry             463 drivers/net/wireless/mediatek/mt76/tx.c 		hwq->entry[idx].schedule = true;
entry             399 drivers/net/wireless/mediatek/mt76/usb.c 		urb = q->entry[q->head].urb;
entry             517 drivers/net/wireless/mediatek/mt76/usb.c 	if (WARN_ONCE(q->entry[q->tail].urb != urb, "rx urb mismatch"))
entry             571 drivers/net/wireless/mediatek/mt76/usb.c 		err = mt76u_submit_rx_buf(dev, q->entry[i].urb);
entry             593 drivers/net/wireless/mediatek/mt76/usb.c 	q->entry = devm_kcalloc(dev->dev,
entry             594 drivers/net/wireless/mediatek/mt76/usb.c 				MT_NUM_RX_ENTRIES, sizeof(*q->entry),
entry             596 drivers/net/wireless/mediatek/mt76/usb.c 	if (!q->entry)
entry             603 drivers/net/wireless/mediatek/mt76/usb.c 		err = mt76u_rx_urb_alloc(dev, &q->entry[i]);
entry             618 drivers/net/wireless/mediatek/mt76/usb.c 		mt76u_urb_free(q->entry[i].urb);
entry             634 drivers/net/wireless/mediatek/mt76/usb.c 		usb_poison_urb(q->entry[i].urb);
entry             646 drivers/net/wireless/mediatek/mt76/usb.c 		usb_unpoison_urb(q->entry[i].urb);
entry             655 drivers/net/wireless/mediatek/mt76/usb.c 	struct mt76_queue_entry entry;
entry             668 drivers/net/wireless/mediatek/mt76/usb.c 			if (!q->entry[q->head].done)
entry             671 drivers/net/wireless/mediatek/mt76/usb.c 			if (q->entry[q->head].schedule) {
entry             672 drivers/net/wireless/mediatek/mt76/usb.c 				q->entry[q->head].schedule = false;
entry             676 drivers/net/wireless/mediatek/mt76/usb.c 			entry = q->entry[q->head];
entry             677 drivers/net/wireless/mediatek/mt76/usb.c 			q->entry[q->head].done = false;
entry             681 drivers/net/wireless/mediatek/mt76/usb.c 			dev->drv->tx_complete_skb(dev, i, &entry);
entry             787 drivers/net/wireless/mediatek/mt76/usb.c 	err = mt76u_tx_setup_buffers(dev, tx_info.skb, q->entry[idx].urb);
entry             792 drivers/net/wireless/mediatek/mt76/usb.c 			    q->entry[idx].urb, mt76u_complete_tx,
entry             793 drivers/net/wireless/mediatek/mt76/usb.c 			    &q->entry[idx]);
entry             796 drivers/net/wireless/mediatek/mt76/usb.c 	q->entry[idx].skb = tx_info.skb;
entry             808 drivers/net/wireless/mediatek/mt76/usb.c 		urb = q->entry[q->first].urb;
entry             845 drivers/net/wireless/mediatek/mt76/usb.c 		q->entry = devm_kcalloc(dev->dev,
entry             846 drivers/net/wireless/mediatek/mt76/usb.c 					MT_NUM_TX_ENTRIES, sizeof(*q->entry),
entry             848 drivers/net/wireless/mediatek/mt76/usb.c 		if (!q->entry)
entry             853 drivers/net/wireless/mediatek/mt76/usb.c 			err = mt76u_urb_alloc(dev, &q->entry[j],
entry             870 drivers/net/wireless/mediatek/mt76/usb.c 			usb_free_urb(q->entry[j].urb);
entry             876 drivers/net/wireless/mediatek/mt76/usb.c 	struct mt76_queue_entry entry;
entry             888 drivers/net/wireless/mediatek/mt76/usb.c 				usb_kill_urb(q->entry[j].urb);
entry             902 drivers/net/wireless/mediatek/mt76/usb.c 				entry = q->entry[q->head];
entry             906 drivers/net/wireless/mediatek/mt76/usb.c 				dev->drv->tx_complete_skb(dev, i, &entry);
entry             714 drivers/net/wireless/ralink/rt2x00/rt2400pci.c static bool rt2400pci_get_entry_state(struct queue_entry *entry)
entry             716 drivers/net/wireless/ralink/rt2x00/rt2400pci.c 	struct queue_entry_priv_mmio *entry_priv = entry->priv_data;
entry             719 drivers/net/wireless/ralink/rt2x00/rt2400pci.c 	if (entry->queue->qid == QID_RX) {
entry             731 drivers/net/wireless/ralink/rt2x00/rt2400pci.c static void rt2400pci_clear_entry(struct queue_entry *entry)
entry             733 drivers/net/wireless/ralink/rt2x00/rt2400pci.c 	struct queue_entry_priv_mmio *entry_priv = entry->priv_data;
entry             734 drivers/net/wireless/ralink/rt2x00/rt2400pci.c 	struct skb_frame_desc *skbdesc = get_skb_frame_desc(entry->skb);
entry             737 drivers/net/wireless/ralink/rt2x00/rt2400pci.c 	if (entry->queue->qid == QID_RX) {
entry             739 drivers/net/wireless/ralink/rt2x00/rt2400pci.c 		rt2x00_set_field32(&word, RXD_W2_BUFFER_LENGTH, entry->skb->len);
entry            1094 drivers/net/wireless/ralink/rt2x00/rt2400pci.c static void rt2400pci_write_tx_desc(struct queue_entry *entry,
entry            1097 drivers/net/wireless/ralink/rt2x00/rt2400pci.c 	struct skb_frame_desc *skbdesc = get_skb_frame_desc(entry->skb);
entry            1098 drivers/net/wireless/ralink/rt2x00/rt2400pci.c 	struct queue_entry_priv_mmio *entry_priv = entry->priv_data;
entry            1165 drivers/net/wireless/ralink/rt2x00/rt2400pci.c static void rt2400pci_write_beacon(struct queue_entry *entry,
entry            1168 drivers/net/wireless/ralink/rt2x00/rt2400pci.c 	struct rt2x00_dev *rt2x00dev = entry->queue->rt2x00dev;
entry            1179 drivers/net/wireless/ralink/rt2x00/rt2400pci.c 	if (rt2x00queue_map_txskb(entry)) {
entry            1190 drivers/net/wireless/ralink/rt2x00/rt2400pci.c 	rt2400pci_write_tx_desc(entry, txdesc);
entry            1195 drivers/net/wireless/ralink/rt2x00/rt2400pci.c 	rt2x00debug_dump_frame(rt2x00dev, DUMP_FRAME_BEACON, entry);
entry            1207 drivers/net/wireless/ralink/rt2x00/rt2400pci.c static void rt2400pci_fill_rxdone(struct queue_entry *entry,
entry            1210 drivers/net/wireless/ralink/rt2x00/rt2400pci.c 	struct rt2x00_dev *rt2x00dev = entry->queue->rt2x00dev;
entry            1211 drivers/net/wireless/ralink/rt2x00/rt2400pci.c 	struct queue_entry_priv_mmio *entry_priv = entry->priv_data;
entry            1254 drivers/net/wireless/ralink/rt2x00/rt2400pci.c 	    entry->queue->rt2x00dev->rssi_offset;
entry            1270 drivers/net/wireless/ralink/rt2x00/rt2400pci.c 	struct queue_entry *entry;
entry            1275 drivers/net/wireless/ralink/rt2x00/rt2400pci.c 		entry = rt2x00queue_get_entry(queue, Q_INDEX_DONE);
entry            1276 drivers/net/wireless/ralink/rt2x00/rt2400pci.c 		entry_priv = entry->priv_data;
entry            1300 drivers/net/wireless/ralink/rt2x00/rt2400pci.c 		rt2x00lib_txdone(entry, &txdesc);
entry             803 drivers/net/wireless/ralink/rt2x00/rt2500pci.c static bool rt2500pci_get_entry_state(struct queue_entry *entry)
entry             805 drivers/net/wireless/ralink/rt2x00/rt2500pci.c 	struct queue_entry_priv_mmio *entry_priv = entry->priv_data;
entry             808 drivers/net/wireless/ralink/rt2x00/rt2500pci.c 	if (entry->queue->qid == QID_RX) {
entry             820 drivers/net/wireless/ralink/rt2x00/rt2500pci.c static void rt2500pci_clear_entry(struct queue_entry *entry)
entry             822 drivers/net/wireless/ralink/rt2x00/rt2500pci.c 	struct queue_entry_priv_mmio *entry_priv = entry->priv_data;
entry             823 drivers/net/wireless/ralink/rt2x00/rt2500pci.c 	struct skb_frame_desc *skbdesc = get_skb_frame_desc(entry->skb);
entry             826 drivers/net/wireless/ralink/rt2x00/rt2500pci.c 	if (entry->queue->qid == QID_RX) {
entry            1247 drivers/net/wireless/ralink/rt2x00/rt2500pci.c static void rt2500pci_write_tx_desc(struct queue_entry *entry,
entry            1250 drivers/net/wireless/ralink/rt2x00/rt2500pci.c 	struct skb_frame_desc *skbdesc = get_skb_frame_desc(entry->skb);
entry            1251 drivers/net/wireless/ralink/rt2x00/rt2500pci.c 	struct queue_entry_priv_mmio *entry_priv = entry->priv_data;
entry            1264 drivers/net/wireless/ralink/rt2x00/rt2500pci.c 	rt2x00_set_field32(&word, TXD_W2_AIFS, entry->queue->aifs);
entry            1265 drivers/net/wireless/ralink/rt2x00/rt2500pci.c 	rt2x00_set_field32(&word, TXD_W2_CWMIN, entry->queue->cw_min);
entry            1266 drivers/net/wireless/ralink/rt2x00/rt2500pci.c 	rt2x00_set_field32(&word, TXD_W2_CWMAX, entry->queue->cw_max);
entry            1317 drivers/net/wireless/ralink/rt2x00/rt2500pci.c static void rt2500pci_write_beacon(struct queue_entry *entry,
entry            1320 drivers/net/wireless/ralink/rt2x00/rt2500pci.c 	struct rt2x00_dev *rt2x00dev = entry->queue->rt2x00dev;
entry            1331 drivers/net/wireless/ralink/rt2x00/rt2500pci.c 	if (rt2x00queue_map_txskb(entry)) {
entry            1339 drivers/net/wireless/ralink/rt2x00/rt2500pci.c 	rt2500pci_write_tx_desc(entry, txdesc);
entry            1344 drivers/net/wireless/ralink/rt2x00/rt2500pci.c 	rt2x00debug_dump_frame(rt2x00dev, DUMP_FRAME_BEACON, entry);
entry            1356 drivers/net/wireless/ralink/rt2x00/rt2500pci.c static void rt2500pci_fill_rxdone(struct queue_entry *entry,
entry            1359 drivers/net/wireless/ralink/rt2x00/rt2500pci.c 	struct queue_entry_priv_mmio *entry_priv = entry->priv_data;
entry            1379 drivers/net/wireless/ralink/rt2x00/rt2500pci.c 	    entry->queue->rt2x00dev->rssi_offset;
entry            1398 drivers/net/wireless/ralink/rt2x00/rt2500pci.c 	struct queue_entry *entry;
entry            1403 drivers/net/wireless/ralink/rt2x00/rt2500pci.c 		entry = rt2x00queue_get_entry(queue, Q_INDEX_DONE);
entry            1404 drivers/net/wireless/ralink/rt2x00/rt2500pci.c 		entry_priv = entry->priv_data;
entry            1428 drivers/net/wireless/ralink/rt2x00/rt2500pci.c 		rt2x00lib_txdone(entry, &txdesc);
entry            1056 drivers/net/wireless/ralink/rt2x00/rt2500usb.c static void rt2500usb_write_tx_desc(struct queue_entry *entry,
entry            1059 drivers/net/wireless/ralink/rt2x00/rt2500usb.c 	struct skb_frame_desc *skbdesc = get_skb_frame_desc(entry->skb);
entry            1060 drivers/net/wireless/ralink/rt2x00/rt2500usb.c 	__le32 *txd = (__le32 *) entry->skb->data;
entry            1086 drivers/net/wireless/ralink/rt2x00/rt2500usb.c 	rt2x00_set_field32(&word, TXD_W1_AIFS, entry->queue->aifs);
entry            1087 drivers/net/wireless/ralink/rt2x00/rt2500usb.c 	rt2x00_set_field32(&word, TXD_W1_CWMIN, entry->queue->cw_min);
entry            1088 drivers/net/wireless/ralink/rt2x00/rt2500usb.c 	rt2x00_set_field32(&word, TXD_W1_CWMAX, entry->queue->cw_max);
entry            1118 drivers/net/wireless/ralink/rt2x00/rt2500usb.c static void rt2500usb_write_beacon(struct queue_entry *entry,
entry            1121 drivers/net/wireless/ralink/rt2x00/rt2500usb.c 	struct rt2x00_dev *rt2x00dev = entry->queue->rt2x00dev;
entry            1123 drivers/net/wireless/ralink/rt2x00/rt2500usb.c 	struct queue_entry_priv_usb_bcn *bcn_priv = entry->priv_data;
entry            1124 drivers/net/wireless/ralink/rt2x00/rt2500usb.c 	int pipe = usb_sndbulkpipe(usb_dev, entry->queue->usb_endpoint);
entry            1139 drivers/net/wireless/ralink/rt2x00/rt2500usb.c 	skb_push(entry->skb, TXD_DESC_SIZE);
entry            1140 drivers/net/wireless/ralink/rt2x00/rt2500usb.c 	memset(entry->skb->data, 0, TXD_DESC_SIZE);
entry            1145 drivers/net/wireless/ralink/rt2x00/rt2500usb.c 	rt2500usb_write_tx_desc(entry, txdesc);
entry            1150 drivers/net/wireless/ralink/rt2x00/rt2500usb.c 	rt2x00debug_dump_frame(rt2x00dev, DUMP_FRAME_BEACON, entry);
entry            1157 drivers/net/wireless/ralink/rt2x00/rt2500usb.c 	length = rt2x00dev->ops->lib->get_tx_data_len(entry);
entry            1160 drivers/net/wireless/ralink/rt2x00/rt2500usb.c 			  entry->skb->data, length, rt2500usb_beacondone,
entry            1161 drivers/net/wireless/ralink/rt2x00/rt2500usb.c 			  entry);
entry            1171 drivers/net/wireless/ralink/rt2x00/rt2500usb.c 			  entry);
entry            1199 drivers/net/wireless/ralink/rt2x00/rt2500usb.c static int rt2500usb_get_tx_data_len(struct queue_entry *entry)
entry            1207 drivers/net/wireless/ralink/rt2x00/rt2500usb.c 	length = roundup(entry->skb->len, 2);
entry            1208 drivers/net/wireless/ralink/rt2x00/rt2500usb.c 	length += (2 * !(length % entry->queue->usb_maxpacket));
entry            1216 drivers/net/wireless/ralink/rt2x00/rt2500usb.c static void rt2500usb_fill_rxdone(struct queue_entry *entry,
entry            1219 drivers/net/wireless/ralink/rt2x00/rt2500usb.c 	struct rt2x00_dev *rt2x00dev = entry->queue->rt2x00dev;
entry            1220 drivers/net/wireless/ralink/rt2x00/rt2500usb.c 	struct queue_entry_priv_usb *entry_priv = entry->priv_data;
entry            1221 drivers/net/wireless/ralink/rt2x00/rt2500usb.c 	struct skb_frame_desc *skbdesc = get_skb_frame_desc(entry->skb);
entry            1223 drivers/net/wireless/ralink/rt2x00/rt2500usb.c 	    (__le32 *)(entry->skb->data +
entry            1225 drivers/net/wireless/ralink/rt2x00/rt2500usb.c 			entry->queue->desc_size));
entry            1286 drivers/net/wireless/ralink/rt2x00/rt2500usb.c 	skb_trim(entry->skb, rxdesc->size);
entry            1294 drivers/net/wireless/ralink/rt2x00/rt2500usb.c 	struct queue_entry *entry = (struct queue_entry *)urb->context;
entry            1295 drivers/net/wireless/ralink/rt2x00/rt2500usb.c 	struct queue_entry_priv_usb_bcn *bcn_priv = entry->priv_data;
entry            1297 drivers/net/wireless/ralink/rt2x00/rt2500usb.c 	if (!test_bit(DEVICE_STATE_ENABLED_RADIO, &entry->queue->rt2x00dev->flags))
entry            1309 drivers/net/wireless/ralink/rt2x00/rt2500usb.c 		dev_kfree_skb(entry->skb);
entry            1310 drivers/net/wireless/ralink/rt2x00/rt2500usb.c 		entry->skb = NULL;
entry             773 drivers/net/wireless/ralink/rt2x00/rt2800lib.c void rt2800_write_tx_data(struct queue_entry *entry,
entry             776 drivers/net/wireless/ralink/rt2x00/rt2800lib.c 	__le32 *txwi = rt2800_drv_get_txwi(entry);
entry             816 drivers/net/wireless/ralink/rt2x00/rt2800lib.c 	rt2x00_set_field32(&word, TXWI_W1_PACKETID_QUEUE, entry->queue->qid);
entry             817 drivers/net/wireless/ralink/rt2x00/rt2800lib.c 	rt2x00_set_field32(&word, TXWI_W1_PACKETID_ENTRY, (entry->entry_idx % 3) + 1);
entry             829 drivers/net/wireless/ralink/rt2x00/rt2800lib.c 	for (i = 2; i < entry->queue->winfo_size / sizeof(__le32); i++)
entry             877 drivers/net/wireless/ralink/rt2x00/rt2800lib.c void rt2800_process_rxwi(struct queue_entry *entry,
entry             880 drivers/net/wireless/ralink/rt2x00/rt2800lib.c 	__le32 *rxwi = (__le32 *) entry->skb->data;
entry             914 drivers/net/wireless/ralink/rt2x00/rt2800lib.c 	rxdesc->rssi = rt2800_agc_to_rssi(entry->queue->rt2x00dev, word);
entry             918 drivers/net/wireless/ralink/rt2x00/rt2800lib.c 	skb_pull(entry->skb, entry->queue->winfo_size);
entry             955 drivers/net/wireless/ralink/rt2x00/rt2800lib.c static bool rt2800_txdone_entry_check(struct queue_entry *entry, u32 reg)
entry             967 drivers/net/wireless/ralink/rt2x00/rt2800lib.c 	if (test_bit(ENTRY_DATA_IO_FAILED, &entry->flags))
entry             979 drivers/net/wireless/ralink/rt2x00/rt2800lib.c 	txwi = rt2800_drv_get_txwi(entry);
entry             987 drivers/net/wireless/ralink/rt2x00/rt2800lib.c 		rt2x00_dbg(entry->queue->rt2x00dev,
entry             989 drivers/net/wireless/ralink/rt2x00/rt2800lib.c 			   entry->queue->qid, entry->entry_idx);
entry             996 drivers/net/wireless/ralink/rt2x00/rt2800lib.c void rt2800_txdone_entry(struct queue_entry *entry, u32 status, __le32 *txwi,
entry             999 drivers/net/wireless/ralink/rt2x00/rt2800lib.c 	struct rt2x00_dev *rt2x00dev = entry->queue->rt2x00dev;
entry            1001 drivers/net/wireless/ralink/rt2x00/rt2800lib.c 	struct skb_frame_desc *skbdesc = get_skb_frame_desc(entry->skb);
entry            1090 drivers/net/wireless/ralink/rt2x00/rt2800lib.c 		rt2x00lib_txdone_nomatch(entry, &txdesc);
entry            1093 drivers/net/wireless/ralink/rt2x00/rt2800lib.c 		rt2x00lib_txdone(entry, &txdesc);
entry            1101 drivers/net/wireless/ralink/rt2x00/rt2800lib.c 	struct queue_entry *entry;
entry            1120 drivers/net/wireless/ralink/rt2x00/rt2800lib.c 		entry = rt2x00queue_get_entry(queue, Q_INDEX_DONE);
entry            1122 drivers/net/wireless/ralink/rt2x00/rt2800lib.c 		if (unlikely(test_bit(ENTRY_OWNER_DEVICE_DATA, &entry->flags) ||
entry            1123 drivers/net/wireless/ralink/rt2x00/rt2800lib.c 			     !test_bit(ENTRY_DATA_STATUS_PENDING, &entry->flags))) {
entry            1125 drivers/net/wireless/ralink/rt2x00/rt2800lib.c 				    entry->entry_idx, qid);
entry            1129 drivers/net/wireless/ralink/rt2x00/rt2800lib.c 		match = rt2800_txdone_entry_check(entry, reg);
entry            1130 drivers/net/wireless/ralink/rt2x00/rt2800lib.c 		rt2800_txdone_entry(entry, reg, rt2800_drv_get_txwi(entry), match);
entry            1136 drivers/net/wireless/ralink/rt2x00/rt2800lib.c 						 struct queue_entry *entry)
entry            1141 drivers/net/wireless/ralink/rt2x00/rt2800lib.c 	if (!test_bit(ENTRY_DATA_STATUS_PENDING, &entry->flags))
entry            1149 drivers/net/wireless/ralink/rt2x00/rt2800lib.c 	ret = time_after(jiffies, entry->last_action + tout);
entry            1151 drivers/net/wireless/ralink/rt2x00/rt2800lib.c 		rt2x00_dbg(entry->queue->rt2x00dev,
entry            1153 drivers/net/wireless/ralink/rt2x00/rt2800lib.c 			   entry->entry_idx, entry->queue->qid);
entry            1160 drivers/net/wireless/ralink/rt2x00/rt2800lib.c 	struct queue_entry *entry;
entry            1163 drivers/net/wireless/ralink/rt2x00/rt2800lib.c 		entry = rt2x00queue_get_entry(queue, Q_INDEX_DONE);
entry            1164 drivers/net/wireless/ralink/rt2x00/rt2800lib.c 		if (rt2800_entry_txstatus_timeout(rt2x00dev, entry))
entry            1192 drivers/net/wireless/ralink/rt2x00/rt2800lib.c 	struct queue_entry *entry;
entry            1203 drivers/net/wireless/ralink/rt2x00/rt2800lib.c 			entry = rt2x00queue_get_entry(queue, Q_INDEX_DONE);
entry            1205 drivers/net/wireless/ralink/rt2x00/rt2800lib.c 			if (test_bit(ENTRY_OWNER_DEVICE_DATA, &entry->flags) ||
entry            1206 drivers/net/wireless/ralink/rt2x00/rt2800lib.c 			    !test_bit(ENTRY_DATA_STATUS_PENDING, &entry->flags))
entry            1209 drivers/net/wireless/ralink/rt2x00/rt2800lib.c 			if (test_bit(ENTRY_DATA_IO_FAILED, &entry->flags) ||
entry            1210 drivers/net/wireless/ralink/rt2x00/rt2800lib.c 			    rt2800_entry_txstatus_timeout(rt2x00dev, entry))
entry            1211 drivers/net/wireless/ralink/rt2x00/rt2800lib.c 				rt2x00lib_txdone_noinfo(entry, TXDONE_FAILURE);
entry            1291 drivers/net/wireless/ralink/rt2x00/rt2800lib.c 	struct queue_entry *entry;
entry            1300 drivers/net/wireless/ralink/rt2x00/rt2800lib.c 		entry = &queue->entries[i];
entry            1301 drivers/net/wireless/ralink/rt2x00/rt2800lib.c 		if (!test_bit(ENTRY_BCN_ENABLED, &entry->flags))
entry            1303 drivers/net/wireless/ralink/rt2x00/rt2800lib.c 		off = rt2800_get_beacon_offset(rt2x00dev, entry->entry_idx);
entry            1320 drivers/net/wireless/ralink/rt2x00/rt2800lib.c void rt2800_write_beacon(struct queue_entry *entry, struct txentry_desc *txdesc)
entry            1322 drivers/net/wireless/ralink/rt2x00/rt2800lib.c 	struct rt2x00_dev *rt2x00dev = entry->queue->rt2x00dev;
entry            1323 drivers/net/wireless/ralink/rt2x00/rt2800lib.c 	struct skb_frame_desc *skbdesc = get_skb_frame_desc(entry->skb);
entry            1327 drivers/net/wireless/ralink/rt2x00/rt2800lib.c 	const int txwi_desc_size = entry->queue->winfo_size;
entry            1341 drivers/net/wireless/ralink/rt2x00/rt2800lib.c 	memset(skb_push(entry->skb, txwi_desc_size), 0, txwi_desc_size);
entry            1347 drivers/net/wireless/ralink/rt2x00/rt2800lib.c 	skbdesc->desc = entry->skb->data;
entry            1353 drivers/net/wireless/ralink/rt2x00/rt2800lib.c 	rt2800_write_tx_data(entry, txdesc);
entry            1358 drivers/net/wireless/ralink/rt2x00/rt2800lib.c 	rt2x00debug_dump_frame(rt2x00dev, DUMP_FRAME_BEACON, entry);
entry            1363 drivers/net/wireless/ralink/rt2x00/rt2800lib.c 	padding_len = roundup(entry->skb->len, 4) - entry->skb->len;
entry            1364 drivers/net/wireless/ralink/rt2x00/rt2800lib.c 	if (padding_len && skb_pad(entry->skb, padding_len)) {
entry            1367 drivers/net/wireless/ralink/rt2x00/rt2800lib.c 		entry->skb = NULL;
entry            1372 drivers/net/wireless/ralink/rt2x00/rt2800lib.c 	beacon_base = rt2800_hw_beacon_base(rt2x00dev, entry->entry_idx);
entry            1374 drivers/net/wireless/ralink/rt2x00/rt2800lib.c 	rt2800_register_multiwrite(rt2x00dev, beacon_base, entry->skb->data,
entry            1375 drivers/net/wireless/ralink/rt2x00/rt2800lib.c 				   entry->skb->len + padding_len);
entry            1376 drivers/net/wireless/ralink/rt2x00/rt2800lib.c 	__set_bit(ENTRY_BCN_ENABLED, &entry->flags);
entry            1391 drivers/net/wireless/ralink/rt2x00/rt2800lib.c 	dev_kfree_skb_any(entry->skb);
entry            1392 drivers/net/wireless/ralink/rt2x00/rt2800lib.c 	entry->skb = NULL;
entry            1414 drivers/net/wireless/ralink/rt2x00/rt2800lib.c void rt2800_clear_beacon(struct queue_entry *entry)
entry            1416 drivers/net/wireless/ralink/rt2x00/rt2800lib.c 	struct rt2x00_dev *rt2x00dev = entry->queue->rt2x00dev;
entry            1431 drivers/net/wireless/ralink/rt2x00/rt2800lib.c 	rt2800_clear_beacon_register(rt2x00dev, entry->entry_idx);
entry            1432 drivers/net/wireless/ralink/rt2x00/rt2800lib.c 	__clear_bit(ENTRY_BCN_ENABLED, &entry->flags);
entry            1863 drivers/net/wireless/ralink/rt2x00/rt2800lib.c 	struct queue_entry *entry;
entry            1872 drivers/net/wireless/ralink/rt2x00/rt2800lib.c 		entry = &queue->entries[i];
entry            1873 drivers/net/wireless/ralink/rt2x00/rt2800lib.c 		clear_bit(ENTRY_BCN_ASSIGNED, &entry->flags);
entry              67 drivers/net/wireless/ralink/rt2x00/rt2800lib.h 	__le32 *(*drv_get_txwi)(struct queue_entry *entry);
entry             163 drivers/net/wireless/ralink/rt2x00/rt2800lib.h static inline __le32 *rt2800_drv_get_txwi(struct queue_entry *entry)
entry             165 drivers/net/wireless/ralink/rt2x00/rt2800lib.h 	const struct rt2800_ops *rt2800ops = entry->queue->rt2x00dev->ops->drv;
entry             167 drivers/net/wireless/ralink/rt2x00/rt2800lib.h 	return rt2800ops->drv_get_txwi(entry);
entry             189 drivers/net/wireless/ralink/rt2x00/rt2800lib.h void rt2800_write_tx_data(struct queue_entry *entry,
entry             191 drivers/net/wireless/ralink/rt2x00/rt2800lib.h void rt2800_process_rxwi(struct queue_entry *entry, struct rxdone_entry_desc *txdesc);
entry             193 drivers/net/wireless/ralink/rt2x00/rt2800lib.h void rt2800_txdone_entry(struct queue_entry *entry, u32 status, __le32 *txwi,
entry             202 drivers/net/wireless/ralink/rt2x00/rt2800lib.h void rt2800_write_beacon(struct queue_entry *entry, struct txentry_desc *txdesc);
entry             203 drivers/net/wireless/ralink/rt2x00/rt2800lib.h void rt2800_clear_beacon(struct queue_entry *entry);
entry              30 drivers/net/wireless/ralink/rt2x00/rt2800mmio.c 	struct queue_entry *entry;
entry              45 drivers/net/wireless/ralink/rt2x00/rt2800mmio.c 		entry = rt2x00queue_get_entry(queue, Q_INDEX_DMA_DONE);
entry              46 drivers/net/wireless/ralink/rt2x00/rt2800mmio.c 		idx = entry->entry_idx;
entry              61 drivers/net/wireless/ralink/rt2x00/rt2800mmio.c __le32 *rt2800mmio_get_txwi(struct queue_entry *entry)
entry              63 drivers/net/wireless/ralink/rt2x00/rt2800mmio.c 	return (__le32 *) entry->skb->data;
entry              67 drivers/net/wireless/ralink/rt2x00/rt2800mmio.c void rt2800mmio_write_tx_desc(struct queue_entry *entry,
entry              70 drivers/net/wireless/ralink/rt2x00/rt2800mmio.c 	struct skb_frame_desc *skbdesc = get_skb_frame_desc(entry->skb);
entry              71 drivers/net/wireless/ralink/rt2x00/rt2800mmio.c 	struct queue_entry_priv_mmio *entry_priv = entry->priv_data;
entry              74 drivers/net/wireless/ralink/rt2x00/rt2800mmio.c 	const unsigned int txwi_size = entry->queue->winfo_size;
entry              92 drivers/net/wireless/ralink/rt2x00/rt2800mmio.c 	rt2x00_set_field32(&word, TXD_W1_SD_LEN1, entry->skb->len);
entry             124 drivers/net/wireless/ralink/rt2x00/rt2800mmio.c void rt2800mmio_fill_rxdone(struct queue_entry *entry,
entry             127 drivers/net/wireless/ralink/rt2x00/rt2800mmio.c 	struct queue_entry_priv_mmio *entry_priv = entry->priv_data;
entry             182 drivers/net/wireless/ralink/rt2x00/rt2800mmio.c 	rt2800_process_rxwi(entry, rxdesc);
entry             454 drivers/net/wireless/ralink/rt2x00/rt2800mmio.c 	struct queue_entry *entry;
entry             462 drivers/net/wireless/ralink/rt2x00/rt2800mmio.c 		entry = rt2x00queue_get_entry(queue, Q_INDEX);
entry             464 drivers/net/wireless/ralink/rt2x00/rt2800mmio.c 					  entry->entry_idx);
entry             469 drivers/net/wireless/ralink/rt2x00/rt2800mmio.c 		entry = rt2x00queue_get_entry(queue, Q_INDEX);
entry             471 drivers/net/wireless/ralink/rt2x00/rt2800mmio.c 					  entry->entry_idx);
entry             607 drivers/net/wireless/ralink/rt2x00/rt2800mmio.c bool rt2800mmio_get_entry_state(struct queue_entry *entry)
entry             609 drivers/net/wireless/ralink/rt2x00/rt2800mmio.c 	struct queue_entry_priv_mmio *entry_priv = entry->priv_data;
entry             612 drivers/net/wireless/ralink/rt2x00/rt2800mmio.c 	if (entry->queue->qid == QID_RX) {
entry             624 drivers/net/wireless/ralink/rt2x00/rt2800mmio.c void rt2800mmio_clear_entry(struct queue_entry *entry)
entry             626 drivers/net/wireless/ralink/rt2x00/rt2800mmio.c 	struct queue_entry_priv_mmio *entry_priv = entry->priv_data;
entry             627 drivers/net/wireless/ralink/rt2x00/rt2800mmio.c 	struct skb_frame_desc *skbdesc = get_skb_frame_desc(entry->skb);
entry             628 drivers/net/wireless/ralink/rt2x00/rt2800mmio.c 	struct rt2x00_dev *rt2x00dev = entry->queue->rt2x00dev;
entry             631 drivers/net/wireless/ralink/rt2x00/rt2800mmio.c 	if (entry->queue->qid == QID_RX) {
entry             645 drivers/net/wireless/ralink/rt2x00/rt2800mmio.c 					  entry->entry_idx);
entry             652 drivers/net/wireless/ralink/rt2x00/rt2800mmio.c 		if (entry->queue->length == 1)
entry             120 drivers/net/wireless/ralink/rt2x00/rt2800mmio.h __le32 *rt2800mmio_get_txwi(struct queue_entry *entry);
entry             121 drivers/net/wireless/ralink/rt2x00/rt2800mmio.h void rt2800mmio_write_tx_desc(struct queue_entry *entry,
entry             125 drivers/net/wireless/ralink/rt2x00/rt2800mmio.h void rt2800mmio_fill_rxdone(struct queue_entry *entry,
entry             147 drivers/net/wireless/ralink/rt2x00/rt2800mmio.h bool rt2800mmio_get_entry_state(struct queue_entry *entry);
entry             148 drivers/net/wireless/ralink/rt2x00/rt2800mmio.h void rt2800mmio_clear_entry(struct queue_entry *entry);
entry             155 drivers/net/wireless/ralink/rt2x00/rt2800usb.c static void rt2800usb_tx_dma_done(struct queue_entry *entry)
entry             157 drivers/net/wireless/ralink/rt2x00/rt2800usb.c 	struct rt2x00_dev *rt2x00dev = entry->queue->rt2x00dev;
entry             384 drivers/net/wireless/ralink/rt2x00/rt2800usb.c 	struct queue_entry *entry;
entry             386 drivers/net/wireless/ralink/rt2x00/rt2800usb.c 	entry = rt2x00queue_get_entry(queue, Q_INDEX_DMA_DONE);
entry             387 drivers/net/wireless/ralink/rt2x00/rt2800usb.c 	return entry->entry_idx;
entry             393 drivers/net/wireless/ralink/rt2x00/rt2800usb.c static __le32 *rt2800usb_get_txwi(struct queue_entry *entry)
entry             395 drivers/net/wireless/ralink/rt2x00/rt2800usb.c 	if (entry->queue->qid == QID_BEACON)
entry             396 drivers/net/wireless/ralink/rt2x00/rt2800usb.c 		return (__le32 *) (entry->skb->data);
entry             398 drivers/net/wireless/ralink/rt2x00/rt2800usb.c 		return (__le32 *) (entry->skb->data + TXINFO_DESC_SIZE);
entry             401 drivers/net/wireless/ralink/rt2x00/rt2800usb.c static void rt2800usb_write_tx_desc(struct queue_entry *entry,
entry             404 drivers/net/wireless/ralink/rt2x00/rt2800usb.c 	struct skb_frame_desc *skbdesc = get_skb_frame_desc(entry->skb);
entry             405 drivers/net/wireless/ralink/rt2x00/rt2800usb.c 	__le32 *txi = (__le32 *) entry->skb->data;
entry             419 drivers/net/wireless/ralink/rt2x00/rt2800usb.c 			   roundup(entry->skb->len, 4) - TXINFO_DESC_SIZE);
entry             434 drivers/net/wireless/ralink/rt2x00/rt2800usb.c 	skbdesc->desc_len = TXINFO_DESC_SIZE + entry->queue->winfo_size;
entry             440 drivers/net/wireless/ralink/rt2x00/rt2800usb.c static int rt2800usb_get_tx_data_len(struct queue_entry *entry)
entry             450 drivers/net/wireless/ralink/rt2x00/rt2800usb.c 	return roundup(entry->skb->len, 4) + 4;
entry             481 drivers/net/wireless/ralink/rt2x00/rt2800usb.c static void rt2800usb_fill_rxdone(struct queue_entry *entry,
entry             484 drivers/net/wireless/ralink/rt2x00/rt2800usb.c 	struct skb_frame_desc *skbdesc = get_skb_frame_desc(entry->skb);
entry             485 drivers/net/wireless/ralink/rt2x00/rt2800usb.c 	__le32 *rxi = (__le32 *)entry->skb->data;
entry             507 drivers/net/wireless/ralink/rt2x00/rt2800usb.c 	skb_pull(entry->skb, RXINFO_DESC_SIZE);
entry             514 drivers/net/wireless/ralink/rt2x00/rt2800usb.c 			rx_pkt_len > entry->queue->data_size)) {
entry             515 drivers/net/wireless/ralink/rt2x00/rt2800usb.c 		rt2x00_err(entry->queue->rt2x00dev,
entry             520 drivers/net/wireless/ralink/rt2x00/rt2800usb.c 	rxd = (__le32 *)(entry->skb->data + rx_pkt_len);
entry             571 drivers/net/wireless/ralink/rt2x00/rt2800usb.c 	skb_trim(entry->skb, rx_pkt_len);
entry             576 drivers/net/wireless/ralink/rt2x00/rt2800usb.c 	rt2800_process_rxwi(entry, rxdesc);
entry             546 drivers/net/wireless/ralink/rt2x00/rt2x00.h 	bool (*get_entry_state) (struct queue_entry *entry);
entry             547 drivers/net/wireless/ralink/rt2x00/rt2x00.h 	void (*clear_entry) (struct queue_entry *entry);
entry             572 drivers/net/wireless/ralink/rt2x00/rt2x00.h 	void (*tx_dma_done) (struct queue_entry *entry);
entry             577 drivers/net/wireless/ralink/rt2x00/rt2x00.h 	void (*write_tx_desc) (struct queue_entry *entry,
entry             579 drivers/net/wireless/ralink/rt2x00/rt2x00.h 	void (*write_tx_data) (struct queue_entry *entry,
entry             581 drivers/net/wireless/ralink/rt2x00/rt2x00.h 	void (*write_beacon) (struct queue_entry *entry,
entry             583 drivers/net/wireless/ralink/rt2x00/rt2x00.h 	void (*clear_beacon) (struct queue_entry *entry);
entry             584 drivers/net/wireless/ralink/rt2x00/rt2x00.h 	int (*get_tx_data_len) (struct queue_entry *entry);
entry             589 drivers/net/wireless/ralink/rt2x00/rt2x00.h 	void (*fill_rxdone) (struct queue_entry *entry,
entry            1020 drivers/net/wireless/ralink/rt2x00/rt2x00.h 	struct queue_entry *entry;
entry            1285 drivers/net/wireless/ralink/rt2x00/rt2x00.h int rt2x00queue_map_txskb(struct queue_entry *entry);
entry            1291 drivers/net/wireless/ralink/rt2x00/rt2x00.h void rt2x00queue_unmap_skb(struct queue_entry *entry);
entry            1404 drivers/net/wireless/ralink/rt2x00/rt2x00.h 			    enum rt2x00_dump_type type, struct queue_entry *entry);
entry            1408 drivers/net/wireless/ralink/rt2x00/rt2x00.h 					  struct queue_entry *entry)
entry            1425 drivers/net/wireless/ralink/rt2x00/rt2x00.h void rt2x00lib_dmastart(struct queue_entry *entry);
entry            1426 drivers/net/wireless/ralink/rt2x00/rt2x00.h void rt2x00lib_dmadone(struct queue_entry *entry);
entry            1427 drivers/net/wireless/ralink/rt2x00/rt2x00.h void rt2x00lib_txdone(struct queue_entry *entry,
entry            1429 drivers/net/wireless/ralink/rt2x00/rt2x00.h void rt2x00lib_txdone_nomatch(struct queue_entry *entry,
entry            1431 drivers/net/wireless/ralink/rt2x00/rt2x00.h void rt2x00lib_txdone_noinfo(struct queue_entry *entry, u32 status);
entry            1432 drivers/net/wireless/ralink/rt2x00/rt2x00.h void rt2x00lib_rxdone(struct queue_entry *entry, gfp_t gfp);
entry             131 drivers/net/wireless/ralink/rt2x00/rt2x00debug.c 			    enum rt2x00_dump_type type, struct queue_entry *entry)
entry             134 drivers/net/wireless/ralink/rt2x00/rt2x00debug.c 	struct sk_buff *skb = entry->skb;
entry             171 drivers/net/wireless/ralink/rt2x00/rt2x00debug.c 	dump_hdr->queue_index = entry->queue->qid;
entry             172 drivers/net/wireless/ralink/rt2x00/rt2x00debug.c 	dump_hdr->entry_index = entry->entry_idx;
entry             246 drivers/net/wireless/ralink/rt2x00/rt2x00dev.c void rt2x00lib_dmastart(struct queue_entry *entry)
entry             248 drivers/net/wireless/ralink/rt2x00/rt2x00dev.c 	set_bit(ENTRY_OWNER_DEVICE_DATA, &entry->flags);
entry             249 drivers/net/wireless/ralink/rt2x00/rt2x00dev.c 	rt2x00queue_index_inc(entry, Q_INDEX);
entry             253 drivers/net/wireless/ralink/rt2x00/rt2x00dev.c void rt2x00lib_dmadone(struct queue_entry *entry)
entry             255 drivers/net/wireless/ralink/rt2x00/rt2x00dev.c 	set_bit(ENTRY_DATA_STATUS_PENDING, &entry->flags);
entry             256 drivers/net/wireless/ralink/rt2x00/rt2x00dev.c 	clear_bit(ENTRY_OWNER_DEVICE_DATA, &entry->flags);
entry             257 drivers/net/wireless/ralink/rt2x00/rt2x00dev.c 	rt2x00queue_index_inc(entry, Q_INDEX_DMA_DONE);
entry             261 drivers/net/wireless/ralink/rt2x00/rt2x00dev.c static inline int rt2x00lib_txdone_bar_status(struct queue_entry *entry)
entry             263 drivers/net/wireless/ralink/rt2x00/rt2x00dev.c 	struct rt2x00_dev *rt2x00dev = entry->queue->rt2x00dev;
entry             264 drivers/net/wireless/ralink/rt2x00/rt2x00dev.c 	struct ieee80211_bar *bar = (void *) entry->skb->data;
entry             287 drivers/net/wireless/ralink/rt2x00/rt2x00dev.c 		if (bar_entry->entry != entry)
entry             385 drivers/net/wireless/ralink/rt2x00/rt2x00dev.c 				  struct queue_entry *entry)
entry             390 drivers/net/wireless/ralink/rt2x00/rt2x00dev.c 	entry->skb = NULL;
entry             391 drivers/net/wireless/ralink/rt2x00/rt2x00dev.c 	entry->flags = 0;
entry             393 drivers/net/wireless/ralink/rt2x00/rt2x00dev.c 	rt2x00dev->ops->lib->clear_entry(entry);
entry             395 drivers/net/wireless/ralink/rt2x00/rt2x00dev.c 	rt2x00queue_index_inc(entry, Q_INDEX_DONE);
entry             404 drivers/net/wireless/ralink/rt2x00/rt2x00dev.c 	spin_lock_bh(&entry->queue->tx_lock);
entry             405 drivers/net/wireless/ralink/rt2x00/rt2x00dev.c 	if (!rt2x00queue_threshold(entry->queue))
entry             406 drivers/net/wireless/ralink/rt2x00/rt2x00dev.c 		rt2x00queue_unpause_queue(entry->queue);
entry             407 drivers/net/wireless/ralink/rt2x00/rt2x00dev.c 	spin_unlock_bh(&entry->queue->tx_lock);
entry             410 drivers/net/wireless/ralink/rt2x00/rt2x00dev.c void rt2x00lib_txdone_nomatch(struct queue_entry *entry,
entry             413 drivers/net/wireless/ralink/rt2x00/rt2x00dev.c 	struct rt2x00_dev *rt2x00dev = entry->queue->rt2x00dev;
entry             414 drivers/net/wireless/ralink/rt2x00/rt2x00dev.c 	struct skb_frame_desc *skbdesc = get_skb_frame_desc(entry->skb);
entry             421 drivers/net/wireless/ralink/rt2x00/rt2x00dev.c 	rt2x00queue_unmap_skb(entry);
entry             432 drivers/net/wireless/ralink/rt2x00/rt2x00dev.c 	rt2x00debug_dump_frame(rt2x00dev, DUMP_FRAME_TXDONE, entry);
entry             440 drivers/net/wireless/ralink/rt2x00/rt2x00dev.c 	    rt2x00lib_txdone_bar_status(entry) ||
entry             455 drivers/net/wireless/ralink/rt2x00/rt2x00dev.c 	dev_kfree_skb_any(entry->skb);
entry             456 drivers/net/wireless/ralink/rt2x00/rt2x00dev.c 	rt2x00lib_clear_entry(rt2x00dev, entry);
entry             460 drivers/net/wireless/ralink/rt2x00/rt2x00dev.c void rt2x00lib_txdone(struct queue_entry *entry,
entry             463 drivers/net/wireless/ralink/rt2x00/rt2x00dev.c 	struct rt2x00_dev *rt2x00dev = entry->queue->rt2x00dev;
entry             464 drivers/net/wireless/ralink/rt2x00/rt2x00dev.c 	struct ieee80211_tx_info *tx_info = IEEE80211_SKB_CB(entry->skb);
entry             465 drivers/net/wireless/ralink/rt2x00/rt2x00dev.c 	struct skb_frame_desc *skbdesc = get_skb_frame_desc(entry->skb);
entry             473 drivers/net/wireless/ralink/rt2x00/rt2x00dev.c 	rt2x00queue_unmap_skb(entry);
entry             478 drivers/net/wireless/ralink/rt2x00/rt2x00dev.c 	skb_pull(entry->skb, rt2x00dev->extra_tx_headroom);
entry             488 drivers/net/wireless/ralink/rt2x00/rt2x00dev.c 	header_length = ieee80211_get_hdrlen_from_skb(entry->skb);
entry             494 drivers/net/wireless/ralink/rt2x00/rt2x00dev.c 		rt2x00queue_remove_l2pad(entry->skb, header_length);
entry             503 drivers/net/wireless/ralink/rt2x00/rt2x00dev.c 		rt2x00crypto_tx_insert_iv(entry->skb, header_length);
entry             509 drivers/net/wireless/ralink/rt2x00/rt2x00dev.c 	rt2x00debug_dump_frame(rt2x00dev, DUMP_FRAME_TXDONE, entry);
entry             517 drivers/net/wireless/ralink/rt2x00/rt2x00dev.c 	    rt2x00lib_txdone_bar_status(entry) ||
entry             537 drivers/net/wireless/ralink/rt2x00/rt2x00dev.c 			ieee80211_tx_status(rt2x00dev->hw, entry->skb);
entry             539 drivers/net/wireless/ralink/rt2x00/rt2x00dev.c 			ieee80211_tx_status_ni(rt2x00dev->hw, entry->skb);
entry             541 drivers/net/wireless/ralink/rt2x00/rt2x00dev.c 		dev_kfree_skb_any(entry->skb);
entry             544 drivers/net/wireless/ralink/rt2x00/rt2x00dev.c 	rt2x00lib_clear_entry(rt2x00dev, entry);
entry             548 drivers/net/wireless/ralink/rt2x00/rt2x00dev.c void rt2x00lib_txdone_noinfo(struct queue_entry *entry, u32 status)
entry             556 drivers/net/wireless/ralink/rt2x00/rt2x00dev.c 	rt2x00lib_txdone(entry, &txdesc);
entry             601 drivers/net/wireless/ralink/rt2x00/rt2x00dev.c 	struct rt2x00_bar_list_entry *entry;
entry             611 drivers/net/wireless/ralink/rt2x00/rt2x00dev.c 	list_for_each_entry_rcu(entry, &rt2x00dev->bar_list, list) {
entry             613 drivers/net/wireless/ralink/rt2x00/rt2x00dev.c 		if (ba->start_seq_num != entry->start_seq_num)
entry             620 drivers/net/wireless/ralink/rt2x00/rt2x00dev.c 		if (!TID_CHECK(ba->control, entry->control))
entry             625 drivers/net/wireless/ralink/rt2x00/rt2x00dev.c 		if (!ether_addr_equal_64bits(ba->ra, entry->ta))
entry             628 drivers/net/wireless/ralink/rt2x00/rt2x00dev.c 		if (!ether_addr_equal_64bits(ba->ta, entry->ra))
entry             633 drivers/net/wireless/ralink/rt2x00/rt2x00dev.c 		entry->block_acked = 1;
entry             737 drivers/net/wireless/ralink/rt2x00/rt2x00dev.c void rt2x00lib_rxdone(struct queue_entry *entry, gfp_t gfp)
entry             739 drivers/net/wireless/ralink/rt2x00/rt2x00dev.c 	struct rt2x00_dev *rt2x00dev = entry->queue->rt2x00dev;
entry             750 drivers/net/wireless/ralink/rt2x00/rt2x00dev.c 	if (test_bit(ENTRY_DATA_IO_FAILED, &entry->flags))
entry             757 drivers/net/wireless/ralink/rt2x00/rt2x00dev.c 	skb = rt2x00queue_alloc_rxskb(entry, gfp);
entry             764 drivers/net/wireless/ralink/rt2x00/rt2x00dev.c 	rt2x00queue_unmap_skb(entry);
entry             770 drivers/net/wireless/ralink/rt2x00/rt2x00dev.c 	rt2x00dev->ops->lib->fill_rxdone(entry, &rxdesc);
entry             777 drivers/net/wireless/ralink/rt2x00/rt2x00dev.c 		     rxdesc.size > entry->queue->data_size)) {
entry             779 drivers/net/wireless/ralink/rt2x00/rt2x00dev.c 			   rxdesc.size, entry->queue->data_size);
entry             780 drivers/net/wireless/ralink/rt2x00/rt2x00dev.c 		dev_kfree_skb(entry->skb);
entry             788 drivers/net/wireless/ralink/rt2x00/rt2x00dev.c 	header_length = ieee80211_get_hdrlen_from_skb(entry->skb);
entry             798 drivers/net/wireless/ralink/rt2x00/rt2x00dev.c 		rt2x00crypto_rx_insert_iv(entry->skb, header_length,
entry             803 drivers/net/wireless/ralink/rt2x00/rt2x00dev.c 		rt2x00queue_remove_l2pad(entry->skb, header_length);
entry             806 drivers/net/wireless/ralink/rt2x00/rt2x00dev.c 	skb_trim(entry->skb, rxdesc.size);
entry             820 drivers/net/wireless/ralink/rt2x00/rt2x00dev.c 	rt2x00lib_rxdone_check_ps(rt2x00dev, entry->skb, &rxdesc);
entry             826 drivers/net/wireless/ralink/rt2x00/rt2x00dev.c 	rt2x00lib_rxdone_check_ba(rt2x00dev, entry->skb, &rxdesc);
entry             831 drivers/net/wireless/ralink/rt2x00/rt2x00dev.c 	rt2x00link_update_stats(rt2x00dev, entry->skb, &rxdesc);
entry             833 drivers/net/wireless/ralink/rt2x00/rt2x00dev.c 	rt2x00debug_dump_frame(rt2x00dev, DUMP_FRAME_RXDONE, entry);
entry             839 drivers/net/wireless/ralink/rt2x00/rt2x00dev.c 	rx_status = IEEE80211_SKB_RXCB(entry->skb);
entry             859 drivers/net/wireless/ralink/rt2x00/rt2x00dev.c 	ieee80211_rx_ni(rt2x00dev->hw, entry->skb);
entry             865 drivers/net/wireless/ralink/rt2x00/rt2x00dev.c 	entry->skb = skb;
entry             868 drivers/net/wireless/ralink/rt2x00/rt2x00dev.c 	entry->flags = 0;
entry             869 drivers/net/wireless/ralink/rt2x00/rt2x00dev.c 	rt2x00queue_index_inc(entry, Q_INDEX_DONE);
entry             872 drivers/net/wireless/ralink/rt2x00/rt2x00dev.c 		rt2x00dev->ops->lib->clear_entry(entry);
entry             966 drivers/net/wireless/ralink/rt2x00/rt2x00dev.c static void rt2x00lib_channel(struct ieee80211_channel *entry,
entry             971 drivers/net/wireless/ralink/rt2x00/rt2x00dev.c 	entry->band = channel <= 14 ? NL80211_BAND_2GHZ : NL80211_BAND_5GHZ;
entry             972 drivers/net/wireless/ralink/rt2x00/rt2x00dev.c 	entry->center_freq = ieee80211_channel_to_frequency(channel,
entry             973 drivers/net/wireless/ralink/rt2x00/rt2x00dev.c 							    entry->band);
entry             974 drivers/net/wireless/ralink/rt2x00/rt2x00dev.c 	entry->hw_value = value;
entry             975 drivers/net/wireless/ralink/rt2x00/rt2x00dev.c 	entry->max_power = tx_power;
entry             976 drivers/net/wireless/ralink/rt2x00/rt2x00dev.c 	entry->max_antenna_gain = 0xff;
entry             979 drivers/net/wireless/ralink/rt2x00/rt2x00dev.c static void rt2x00lib_rate(struct ieee80211_rate *entry,
entry             982 drivers/net/wireless/ralink/rt2x00/rt2x00dev.c 	entry->flags = 0;
entry             983 drivers/net/wireless/ralink/rt2x00/rt2x00dev.c 	entry->bitrate = rate->bitrate;
entry             984 drivers/net/wireless/ralink/rt2x00/rt2x00dev.c 	entry->hw_value = index;
entry             985 drivers/net/wireless/ralink/rt2x00/rt2x00dev.c 	entry->hw_value_short = index;
entry             988 drivers/net/wireless/ralink/rt2x00/rt2x00dev.c 		entry->flags |= IEEE80211_RATE_SHORT_PREAMBLE;
entry              94 drivers/net/wireless/ralink/rt2x00/rt2x00lib.h struct sk_buff *rt2x00queue_alloc_rxskb(struct queue_entry *entry, gfp_t gfp);
entry             100 drivers/net/wireless/ralink/rt2x00/rt2x00lib.h void rt2x00queue_free_skb(struct queue_entry *entry);
entry             174 drivers/net/wireless/ralink/rt2x00/rt2x00lib.h void rt2x00queue_index_inc(struct queue_entry *entry, enum queue_index index);
entry             189 drivers/net/wireless/ralink/rt2x00/rt2x00mac.c 	struct queue_entry *entry = NULL;
entry             207 drivers/net/wireless/ralink/rt2x00/rt2x00mac.c 		entry = &queue->entries[i];
entry             208 drivers/net/wireless/ralink/rt2x00/rt2x00mac.c 		if (!test_and_set_bit(ENTRY_BCN_ASSIGNED, &entry->flags))
entry             226 drivers/net/wireless/ralink/rt2x00/rt2x00mac.c 	intf->beacon = entry;
entry              52 drivers/net/wireless/ralink/rt2x00/rt2x00mmio.c 	struct queue_entry *entry;
entry              58 drivers/net/wireless/ralink/rt2x00/rt2x00mmio.c 		entry = rt2x00queue_get_entry(queue, Q_INDEX);
entry              59 drivers/net/wireless/ralink/rt2x00/rt2x00mmio.c 		entry_priv = entry->priv_data;
entry              61 drivers/net/wireless/ralink/rt2x00/rt2x00mmio.c 		if (rt2x00dev->ops->lib->get_entry_state(entry))
entry              67 drivers/net/wireless/ralink/rt2x00/rt2x00mmio.c 		skbdesc = get_skb_frame_desc(entry->skb);
entry              69 drivers/net/wireless/ralink/rt2x00/rt2x00mmio.c 		skbdesc->desc_len = entry->queue->desc_size;
entry              75 drivers/net/wireless/ralink/rt2x00/rt2x00mmio.c 		rt2x00lib_dmastart(entry);
entry              76 drivers/net/wireless/ralink/rt2x00/rt2x00mmio.c 		rt2x00lib_dmadone(entry);
entry              81 drivers/net/wireless/ralink/rt2x00/rt2x00mmio.c 		rt2x00lib_rxdone(entry, GFP_ATOMIC);
entry              23 drivers/net/wireless/ralink/rt2x00/rt2x00queue.c struct sk_buff *rt2x00queue_alloc_rxskb(struct queue_entry *entry, gfp_t gfp)
entry              25 drivers/net/wireless/ralink/rt2x00/rt2x00queue.c 	struct data_queue *queue = entry->queue;
entry              93 drivers/net/wireless/ralink/rt2x00/rt2x00queue.c int rt2x00queue_map_txskb(struct queue_entry *entry)
entry              95 drivers/net/wireless/ralink/rt2x00/rt2x00queue.c 	struct device *dev = entry->queue->rt2x00dev->dev;
entry              96 drivers/net/wireless/ralink/rt2x00/rt2x00queue.c 	struct skb_frame_desc *skbdesc = get_skb_frame_desc(entry->skb);
entry              99 drivers/net/wireless/ralink/rt2x00/rt2x00queue.c 	    dma_map_single(dev, entry->skb->data, entry->skb->len, DMA_TO_DEVICE);
entry             105 drivers/net/wireless/ralink/rt2x00/rt2x00queue.c 	rt2x00lib_dmadone(entry);
entry             110 drivers/net/wireless/ralink/rt2x00/rt2x00queue.c void rt2x00queue_unmap_skb(struct queue_entry *entry)
entry             112 drivers/net/wireless/ralink/rt2x00/rt2x00queue.c 	struct device *dev = entry->queue->rt2x00dev->dev;
entry             113 drivers/net/wireless/ralink/rt2x00/rt2x00queue.c 	struct skb_frame_desc *skbdesc = get_skb_frame_desc(entry->skb);
entry             116 drivers/net/wireless/ralink/rt2x00/rt2x00queue.c 		dma_unmap_single(dev, skbdesc->skb_dma, entry->skb->len,
entry             120 drivers/net/wireless/ralink/rt2x00/rt2x00queue.c 		dma_unmap_single(dev, skbdesc->skb_dma, entry->skb->len,
entry             127 drivers/net/wireless/ralink/rt2x00/rt2x00queue.c void rt2x00queue_free_skb(struct queue_entry *entry)
entry             129 drivers/net/wireless/ralink/rt2x00/rt2x00queue.c 	if (!entry->skb)
entry             132 drivers/net/wireless/ralink/rt2x00/rt2x00queue.c 	rt2x00queue_unmap_skb(entry);
entry             133 drivers/net/wireless/ralink/rt2x00/rt2x00queue.c 	dev_kfree_skb_any(entry->skb);
entry             134 drivers/net/wireless/ralink/rt2x00/rt2x00queue.c 	entry->skb = NULL;
entry             487 drivers/net/wireless/ralink/rt2x00/rt2x00queue.c static int rt2x00queue_write_tx_data(struct queue_entry *entry,
entry             490 drivers/net/wireless/ralink/rt2x00/rt2x00queue.c 	struct rt2x00_dev *rt2x00dev = entry->queue->rt2x00dev;
entry             498 drivers/net/wireless/ralink/rt2x00/rt2x00queue.c 		     rt2x00dev->ops->lib->get_entry_state(entry))) {
entry             502 drivers/net/wireless/ralink/rt2x00/rt2x00queue.c 			   entry->queue->qid, DRV_PROJECT);
entry             509 drivers/net/wireless/ralink/rt2x00/rt2x00queue.c 	skb_push(entry->skb, rt2x00dev->extra_tx_headroom);
entry             510 drivers/net/wireless/ralink/rt2x00/rt2x00queue.c 	memset(entry->skb->data, 0, rt2x00dev->extra_tx_headroom);
entry             516 drivers/net/wireless/ralink/rt2x00/rt2x00queue.c 		rt2x00dev->ops->lib->write_tx_data(entry, txdesc);
entry             522 drivers/net/wireless/ralink/rt2x00/rt2x00queue.c 	    rt2x00queue_map_txskb(entry))
entry             528 drivers/net/wireless/ralink/rt2x00/rt2x00queue.c static void rt2x00queue_write_tx_descriptor(struct queue_entry *entry,
entry             531 drivers/net/wireless/ralink/rt2x00/rt2x00queue.c 	struct data_queue *queue = entry->queue;
entry             533 drivers/net/wireless/ralink/rt2x00/rt2x00queue.c 	queue->rt2x00dev->ops->lib->write_tx_desc(entry, txdesc);
entry             539 drivers/net/wireless/ralink/rt2x00/rt2x00queue.c 	rt2x00debug_dump_frame(queue->rt2x00dev, DUMP_FRAME_TX, entry);
entry             559 drivers/net/wireless/ralink/rt2x00/rt2x00queue.c static void rt2x00queue_bar_check(struct queue_entry *entry)
entry             561 drivers/net/wireless/ralink/rt2x00/rt2x00queue.c 	struct rt2x00_dev *rt2x00dev = entry->queue->rt2x00dev;
entry             562 drivers/net/wireless/ralink/rt2x00/rt2x00queue.c 	struct ieee80211_bar *bar = (void *) (entry->skb->data +
entry             579 drivers/net/wireless/ralink/rt2x00/rt2x00queue.c 	bar_entry->entry = entry;
entry             605 drivers/net/wireless/ralink/rt2x00/rt2x00queue.c 	struct queue_entry *entry;
entry             672 drivers/net/wireless/ralink/rt2x00/rt2x00queue.c 	entry = rt2x00queue_get_entry(queue, Q_INDEX);
entry             675 drivers/net/wireless/ralink/rt2x00/rt2x00queue.c 				      &entry->flags))) {
entry             684 drivers/net/wireless/ralink/rt2x00/rt2x00queue.c 	entry->skb = skb;
entry             691 drivers/net/wireless/ralink/rt2x00/rt2x00queue.c 	if (unlikely(rt2x00queue_write_tx_data(entry, &txdesc))) {
entry             692 drivers/net/wireless/ralink/rt2x00/rt2x00queue.c 		clear_bit(ENTRY_OWNER_DEVICE_DATA, &entry->flags);
entry             693 drivers/net/wireless/ralink/rt2x00/rt2x00queue.c 		entry->skb = NULL;
entry             701 drivers/net/wireless/ralink/rt2x00/rt2x00queue.c 	rt2x00queue_bar_check(entry);
entry             703 drivers/net/wireless/ralink/rt2x00/rt2x00queue.c 	set_bit(ENTRY_DATA_PENDING, &entry->flags);
entry             705 drivers/net/wireless/ralink/rt2x00/rt2x00queue.c 	rt2x00queue_index_inc(entry, Q_INDEX);
entry             706 drivers/net/wireless/ralink/rt2x00/rt2x00queue.c 	rt2x00queue_write_tx_descriptor(entry, &txdesc);
entry             790 drivers/net/wireless/ralink/rt2x00/rt2x00queue.c 				bool (*fn)(struct queue_entry *entry,
entry             844 drivers/net/wireless/ralink/rt2x00/rt2x00queue.c 	struct queue_entry *entry;
entry             855 drivers/net/wireless/ralink/rt2x00/rt2x00queue.c 	entry = &queue->entries[queue->index[index]];
entry             859 drivers/net/wireless/ralink/rt2x00/rt2x00queue.c 	return entry;
entry             863 drivers/net/wireless/ralink/rt2x00/rt2x00queue.c void rt2x00queue_index_inc(struct queue_entry *entry, enum queue_index index)
entry             865 drivers/net/wireless/ralink/rt2x00/rt2x00queue.c 	struct data_queue *queue = entry->queue;
entry             880 drivers/net/wireless/ralink/rt2x00/rt2x00queue.c 	entry->last_action = jiffies;
entry             584 drivers/net/wireless/ralink/rt2x00/rt2x00queue.h 				bool (*fn)(struct queue_entry *entry,
entry             626 drivers/net/wireless/ralink/rt2x00/rt2x00queue.h static inline int rt2x00queue_dma_timeout(struct queue_entry *entry)
entry             628 drivers/net/wireless/ralink/rt2x00/rt2x00queue.h 	if (!test_bit(ENTRY_OWNER_DEVICE_DATA, &entry->flags))
entry             630 drivers/net/wireless/ralink/rt2x00/rt2x00queue.h 	return time_after(jiffies, entry->last_action + msecs_to_jiffies(100));
entry             229 drivers/net/wireless/ralink/rt2x00/rt2x00usb.c static void rt2x00usb_work_txdone_entry(struct queue_entry *entry)
entry             239 drivers/net/wireless/ralink/rt2x00/rt2x00usb.c 	if (test_bit(ENTRY_DATA_IO_FAILED, &entry->flags))
entry             240 drivers/net/wireless/ralink/rt2x00/rt2x00usb.c 		rt2x00lib_txdone_noinfo(entry, TXDONE_FAILURE);
entry             242 drivers/net/wireless/ralink/rt2x00/rt2x00usb.c 		rt2x00lib_txdone_noinfo(entry, TXDONE_UNKNOWN);
entry             250 drivers/net/wireless/ralink/rt2x00/rt2x00usb.c 	struct queue_entry *entry;
entry             254 drivers/net/wireless/ralink/rt2x00/rt2x00usb.c 			entry = rt2x00queue_get_entry(queue, Q_INDEX_DONE);
entry             256 drivers/net/wireless/ralink/rt2x00/rt2x00usb.c 			if (test_bit(ENTRY_OWNER_DEVICE_DATA, &entry->flags) ||
entry             257 drivers/net/wireless/ralink/rt2x00/rt2x00usb.c 			    !test_bit(ENTRY_DATA_STATUS_PENDING, &entry->flags))
entry             260 drivers/net/wireless/ralink/rt2x00/rt2x00usb.c 			rt2x00usb_work_txdone_entry(entry);
entry             267 drivers/net/wireless/ralink/rt2x00/rt2x00usb.c 	struct queue_entry *entry = (struct queue_entry *)urb->context;
entry             268 drivers/net/wireless/ralink/rt2x00/rt2x00usb.c 	struct rt2x00_dev *rt2x00dev = entry->queue->rt2x00dev;
entry             270 drivers/net/wireless/ralink/rt2x00/rt2x00usb.c 	if (!test_bit(ENTRY_OWNER_DEVICE_DATA, &entry->flags))
entry             276 drivers/net/wireless/ralink/rt2x00/rt2x00usb.c 		set_bit(ENTRY_DATA_IO_FAILED, &entry->flags);
entry             280 drivers/net/wireless/ralink/rt2x00/rt2x00usb.c 	rt2x00lib_dmadone(entry);
entry             283 drivers/net/wireless/ralink/rt2x00/rt2x00usb.c 		rt2x00dev->ops->lib->tx_dma_done(entry);
entry             293 drivers/net/wireless/ralink/rt2x00/rt2x00usb.c static bool rt2x00usb_kick_tx_entry(struct queue_entry *entry, void *data)
entry             295 drivers/net/wireless/ralink/rt2x00/rt2x00usb.c 	struct rt2x00_dev *rt2x00dev = entry->queue->rt2x00dev;
entry             297 drivers/net/wireless/ralink/rt2x00/rt2x00usb.c 	struct queue_entry_priv_usb *entry_priv = entry->priv_data;
entry             301 drivers/net/wireless/ralink/rt2x00/rt2x00usb.c 	if (!test_and_clear_bit(ENTRY_DATA_PENDING, &entry->flags) ||
entry             302 drivers/net/wireless/ralink/rt2x00/rt2x00usb.c 	    test_bit(ENTRY_DATA_STATUS_PENDING, &entry->flags))
entry             310 drivers/net/wireless/ralink/rt2x00/rt2x00usb.c 	length = rt2x00dev->ops->lib->get_tx_data_len(entry);
entry             312 drivers/net/wireless/ralink/rt2x00/rt2x00usb.c 	status = skb_padto(entry->skb, length);
entry             316 drivers/net/wireless/ralink/rt2x00/rt2x00usb.c 		set_bit(ENTRY_DATA_IO_FAILED, &entry->flags);
entry             317 drivers/net/wireless/ralink/rt2x00/rt2x00usb.c 		rt2x00lib_dmadone(entry);
entry             323 drivers/net/wireless/ralink/rt2x00/rt2x00usb.c 			  usb_sndbulkpipe(usb_dev, entry->queue->usb_endpoint),
entry             324 drivers/net/wireless/ralink/rt2x00/rt2x00usb.c 			  entry->skb->data, length,
entry             325 drivers/net/wireless/ralink/rt2x00/rt2x00usb.c 			  rt2x00usb_interrupt_txdone, entry);
entry             331 drivers/net/wireless/ralink/rt2x00/rt2x00usb.c 		set_bit(ENTRY_DATA_IO_FAILED, &entry->flags);
entry             332 drivers/net/wireless/ralink/rt2x00/rt2x00usb.c 		rt2x00lib_dmadone(entry);
entry             345 drivers/net/wireless/ralink/rt2x00/rt2x00usb.c 	struct queue_entry *entry;
entry             350 drivers/net/wireless/ralink/rt2x00/rt2x00usb.c 		entry = rt2x00queue_get_entry(rt2x00dev->rx, Q_INDEX_DONE);
entry             352 drivers/net/wireless/ralink/rt2x00/rt2x00usb.c 		if (test_bit(ENTRY_OWNER_DEVICE_DATA, &entry->flags))
entry             358 drivers/net/wireless/ralink/rt2x00/rt2x00usb.c 		skbdesc = get_skb_frame_desc(entry->skb);
entry             360 drivers/net/wireless/ralink/rt2x00/rt2x00usb.c 		skbdesc->desc_len = entry->queue->desc_size;
entry             365 drivers/net/wireless/ralink/rt2x00/rt2x00usb.c 		rt2x00lib_rxdone(entry, GFP_KERNEL);
entry             371 drivers/net/wireless/ralink/rt2x00/rt2x00usb.c 	struct queue_entry *entry = (struct queue_entry *)urb->context;
entry             372 drivers/net/wireless/ralink/rt2x00/rt2x00usb.c 	struct rt2x00_dev *rt2x00dev = entry->queue->rt2x00dev;
entry             374 drivers/net/wireless/ralink/rt2x00/rt2x00usb.c 	if (!test_bit(ENTRY_OWNER_DEVICE_DATA, &entry->flags))
entry             382 drivers/net/wireless/ralink/rt2x00/rt2x00usb.c 	if (urb->actual_length < entry->queue->desc_size || urb->status)
entry             383 drivers/net/wireless/ralink/rt2x00/rt2x00usb.c 		set_bit(ENTRY_DATA_IO_FAILED, &entry->flags);
entry             388 drivers/net/wireless/ralink/rt2x00/rt2x00usb.c 	rt2x00lib_dmadone(entry);
entry             396 drivers/net/wireless/ralink/rt2x00/rt2x00usb.c static bool rt2x00usb_kick_rx_entry(struct queue_entry *entry, void *data)
entry             398 drivers/net/wireless/ralink/rt2x00/rt2x00usb.c 	struct rt2x00_dev *rt2x00dev = entry->queue->rt2x00dev;
entry             400 drivers/net/wireless/ralink/rt2x00/rt2x00usb.c 	struct queue_entry_priv_usb *entry_priv = entry->priv_data;
entry             403 drivers/net/wireless/ralink/rt2x00/rt2x00usb.c 	if (test_and_set_bit(ENTRY_OWNER_DEVICE_DATA, &entry->flags))
entry             406 drivers/net/wireless/ralink/rt2x00/rt2x00usb.c 	rt2x00lib_dmastart(entry);
entry             409 drivers/net/wireless/ralink/rt2x00/rt2x00usb.c 			  usb_rcvbulkpipe(usb_dev, entry->queue->usb_endpoint),
entry             410 drivers/net/wireless/ralink/rt2x00/rt2x00usb.c 			  entry->skb->data, entry->skb->len,
entry             411 drivers/net/wireless/ralink/rt2x00/rt2x00usb.c 			  rt2x00usb_interrupt_rxdone, entry);
entry             417 drivers/net/wireless/ralink/rt2x00/rt2x00usb.c 		set_bit(ENTRY_DATA_IO_FAILED, &entry->flags);
entry             418 drivers/net/wireless/ralink/rt2x00/rt2x00usb.c 		rt2x00lib_dmadone(entry);
entry             452 drivers/net/wireless/ralink/rt2x00/rt2x00usb.c static bool rt2x00usb_flush_entry(struct queue_entry *entry, void *data)
entry             454 drivers/net/wireless/ralink/rt2x00/rt2x00usb.c 	struct rt2x00_dev *rt2x00dev = entry->queue->rt2x00dev;
entry             455 drivers/net/wireless/ralink/rt2x00/rt2x00usb.c 	struct queue_entry_priv_usb *entry_priv = entry->priv_data;
entry             456 drivers/net/wireless/ralink/rt2x00/rt2x00usb.c 	struct queue_entry_priv_usb_bcn *bcn_priv = entry->priv_data;
entry             458 drivers/net/wireless/ralink/rt2x00/rt2x00usb.c 	if (!test_bit(ENTRY_OWNER_DEVICE_DATA, &entry->flags))
entry             466 drivers/net/wireless/ralink/rt2x00/rt2x00usb.c 	if ((entry->queue->qid == QID_BEACON) &&
entry             535 drivers/net/wireless/ralink/rt2x00/rt2x00usb.c 	struct queue_entry *entry;
entry             537 drivers/net/wireless/ralink/rt2x00/rt2x00usb.c 	entry = rt2x00queue_get_entry(queue, Q_INDEX_DMA_DONE);
entry             538 drivers/net/wireless/ralink/rt2x00/rt2x00usb.c 	return rt2x00queue_dma_timeout(entry);
entry             567 drivers/net/wireless/ralink/rt2x00/rt2x00usb.c void rt2x00usb_clear_entry(struct queue_entry *entry)
entry             569 drivers/net/wireless/ralink/rt2x00/rt2x00usb.c 	entry->flags = 0;
entry             571 drivers/net/wireless/ralink/rt2x00/rt2x00usb.c 	if (entry->queue->qid == QID_RX)
entry             572 drivers/net/wireless/ralink/rt2x00/rt2x00usb.c 		rt2x00usb_kick_rx_entry(entry, NULL);
entry             391 drivers/net/wireless/ralink/rt2x00/rt2x00usb.h void rt2x00usb_clear_entry(struct queue_entry *entry);
entry            1287 drivers/net/wireless/ralink/rt2x00/rt61pci.c static bool rt61pci_get_entry_state(struct queue_entry *entry)
entry            1289 drivers/net/wireless/ralink/rt2x00/rt61pci.c 	struct queue_entry_priv_mmio *entry_priv = entry->priv_data;
entry            1292 drivers/net/wireless/ralink/rt2x00/rt61pci.c 	if (entry->queue->qid == QID_RX) {
entry            1304 drivers/net/wireless/ralink/rt2x00/rt61pci.c static void rt61pci_clear_entry(struct queue_entry *entry)
entry            1306 drivers/net/wireless/ralink/rt2x00/rt61pci.c 	struct queue_entry_priv_mmio *entry_priv = entry->priv_data;
entry            1307 drivers/net/wireless/ralink/rt2x00/rt61pci.c 	struct skb_frame_desc *skbdesc = get_skb_frame_desc(entry->skb);
entry            1310 drivers/net/wireless/ralink/rt2x00/rt61pci.c 	if (entry->queue->qid == QID_RX) {
entry            1775 drivers/net/wireless/ralink/rt2x00/rt61pci.c static void rt61pci_write_tx_desc(struct queue_entry *entry,
entry            1778 drivers/net/wireless/ralink/rt2x00/rt61pci.c 	struct skb_frame_desc *skbdesc = get_skb_frame_desc(entry->skb);
entry            1779 drivers/net/wireless/ralink/rt2x00/rt61pci.c 	struct queue_entry_priv_mmio *entry_priv = entry->priv_data;
entry            1787 drivers/net/wireless/ralink/rt2x00/rt61pci.c 	rt2x00_set_field32(&word, TXD_W1_HOST_Q_ID, entry->queue->qid);
entry            1788 drivers/net/wireless/ralink/rt2x00/rt61pci.c 	rt2x00_set_field32(&word, TXD_W1_AIFSN, entry->queue->aifs);
entry            1789 drivers/net/wireless/ralink/rt2x00/rt61pci.c 	rt2x00_set_field32(&word, TXD_W1_CWMIN, entry->queue->cw_min);
entry            1790 drivers/net/wireless/ralink/rt2x00/rt61pci.c 	rt2x00_set_field32(&word, TXD_W1_CWMAX, entry->queue->cw_max);
entry            1812 drivers/net/wireless/ralink/rt2x00/rt61pci.c 	rt2x00_set_field32(&word, TXD_W5_PID_TYPE, entry->queue->qid);
entry            1813 drivers/net/wireless/ralink/rt2x00/rt61pci.c 	rt2x00_set_field32(&word, TXD_W5_PID_SUBTYPE, entry->entry_idx);
entry            1815 drivers/net/wireless/ralink/rt2x00/rt61pci.c 			   TXPOWER_TO_DEV(entry->queue->rt2x00dev->tx_power));
entry            1819 drivers/net/wireless/ralink/rt2x00/rt61pci.c 	if (entry->queue->qid != QID_BEACON) {
entry            1865 drivers/net/wireless/ralink/rt2x00/rt61pci.c 	skbdesc->desc_len = (entry->queue->qid == QID_BEACON) ? TXINFO_SIZE :
entry            1872 drivers/net/wireless/ralink/rt2x00/rt61pci.c static void rt61pci_write_beacon(struct queue_entry *entry,
entry            1875 drivers/net/wireless/ralink/rt2x00/rt61pci.c 	struct rt2x00_dev *rt2x00dev = entry->queue->rt2x00dev;
entry            1876 drivers/net/wireless/ralink/rt2x00/rt61pci.c 	struct queue_entry_priv_mmio *entry_priv = entry->priv_data;
entry            1893 drivers/net/wireless/ralink/rt2x00/rt61pci.c 	rt61pci_write_tx_desc(entry, txdesc);
entry            1898 drivers/net/wireless/ralink/rt2x00/rt61pci.c 	rt2x00debug_dump_frame(rt2x00dev, DUMP_FRAME_BEACON, entry);
entry            1903 drivers/net/wireless/ralink/rt2x00/rt61pci.c 	padding_len = roundup(entry->skb->len, 4) - entry->skb->len;
entry            1904 drivers/net/wireless/ralink/rt2x00/rt61pci.c 	if (padding_len && skb_pad(entry->skb, padding_len)) {
entry            1907 drivers/net/wireless/ralink/rt2x00/rt61pci.c 		entry->skb = NULL;
entry            1912 drivers/net/wireless/ralink/rt2x00/rt61pci.c 	beacon_base = HW_BEACON_OFFSET(entry->entry_idx);
entry            1916 drivers/net/wireless/ralink/rt2x00/rt61pci.c 				       entry->skb->data,
entry            1917 drivers/net/wireless/ralink/rt2x00/rt61pci.c 				       entry->skb->len + padding_len);
entry            1933 drivers/net/wireless/ralink/rt2x00/rt61pci.c 	dev_kfree_skb_any(entry->skb);
entry            1934 drivers/net/wireless/ralink/rt2x00/rt61pci.c 	entry->skb = NULL;
entry            1937 drivers/net/wireless/ralink/rt2x00/rt61pci.c static void rt61pci_clear_beacon(struct queue_entry *entry)
entry            1939 drivers/net/wireless/ralink/rt2x00/rt61pci.c 	struct rt2x00_dev *rt2x00dev = entry->queue->rt2x00dev;
entry            1955 drivers/net/wireless/ralink/rt2x00/rt61pci.c 				  HW_BEACON_OFFSET(entry->entry_idx), 0);
entry            1994 drivers/net/wireless/ralink/rt2x00/rt61pci.c static void rt61pci_fill_rxdone(struct queue_entry *entry,
entry            1997 drivers/net/wireless/ralink/rt2x00/rt61pci.c 	struct rt2x00_dev *rt2x00dev = entry->queue->rt2x00dev;
entry            1998 drivers/net/wireless/ralink/rt2x00/rt61pci.c 	struct queue_entry_priv_mmio *entry_priv = entry->priv_data;
entry            2062 drivers/net/wireless/ralink/rt2x00/rt61pci.c 	struct queue_entry *entry;
entry            2103 drivers/net/wireless/ralink/rt2x00/rt61pci.c 		entry = &queue->entries[index];
entry            2104 drivers/net/wireless/ralink/rt2x00/rt61pci.c 		entry_priv = entry->priv_data;
entry            2112 drivers/net/wireless/ralink/rt2x00/rt61pci.c 		while (entry != entry_done) {
entry            2146 drivers/net/wireless/ralink/rt2x00/rt61pci.c 		rt2x00lib_txdone(entry, &txdesc);
entry            1444 drivers/net/wireless/ralink/rt2x00/rt73usb.c static void rt73usb_write_tx_desc(struct queue_entry *entry,
entry            1447 drivers/net/wireless/ralink/rt2x00/rt73usb.c 	struct skb_frame_desc *skbdesc = get_skb_frame_desc(entry->skb);
entry            1448 drivers/net/wireless/ralink/rt2x00/rt73usb.c 	__le32 *txd = (__le32 *) entry->skb->data;
entry            1481 drivers/net/wireless/ralink/rt2x00/rt73usb.c 	rt2x00_set_field32(&word, TXD_W1_HOST_Q_ID, entry->queue->qid);
entry            1482 drivers/net/wireless/ralink/rt2x00/rt73usb.c 	rt2x00_set_field32(&word, TXD_W1_AIFSN, entry->queue->aifs);
entry            1483 drivers/net/wireless/ralink/rt2x00/rt73usb.c 	rt2x00_set_field32(&word, TXD_W1_CWMIN, entry->queue->cw_min);
entry            1484 drivers/net/wireless/ralink/rt2x00/rt73usb.c 	rt2x00_set_field32(&word, TXD_W1_CWMAX, entry->queue->cw_max);
entry            1506 drivers/net/wireless/ralink/rt2x00/rt73usb.c 			   TXPOWER_TO_DEV(entry->queue->rt2x00dev->tx_power));
entry            1521 drivers/net/wireless/ralink/rt2x00/rt73usb.c static void rt73usb_write_beacon(struct queue_entry *entry,
entry            1524 drivers/net/wireless/ralink/rt2x00/rt73usb.c 	struct rt2x00_dev *rt2x00dev = entry->queue->rt2x00dev;
entry            1541 drivers/net/wireless/ralink/rt2x00/rt73usb.c 	skb_push(entry->skb, TXD_DESC_SIZE);
entry            1542 drivers/net/wireless/ralink/rt2x00/rt73usb.c 	memset(entry->skb->data, 0, TXD_DESC_SIZE);
entry            1547 drivers/net/wireless/ralink/rt2x00/rt73usb.c 	rt73usb_write_tx_desc(entry, txdesc);
entry            1552 drivers/net/wireless/ralink/rt2x00/rt73usb.c 	rt2x00debug_dump_frame(rt2x00dev, DUMP_FRAME_BEACON, entry);
entry            1557 drivers/net/wireless/ralink/rt2x00/rt73usb.c 	padding_len = roundup(entry->skb->len, 4) - entry->skb->len;
entry            1558 drivers/net/wireless/ralink/rt2x00/rt73usb.c 	if (padding_len && skb_pad(entry->skb, padding_len)) {
entry            1561 drivers/net/wireless/ralink/rt2x00/rt73usb.c 		entry->skb = NULL;
entry            1566 drivers/net/wireless/ralink/rt2x00/rt73usb.c 	beacon_base = HW_BEACON_OFFSET(entry->entry_idx);
entry            1567 drivers/net/wireless/ralink/rt2x00/rt73usb.c 	rt2x00usb_register_multiwrite(rt2x00dev, beacon_base, entry->skb->data,
entry            1568 drivers/net/wireless/ralink/rt2x00/rt73usb.c 				      entry->skb->len + padding_len);
entry            1584 drivers/net/wireless/ralink/rt2x00/rt73usb.c 	dev_kfree_skb(entry->skb);
entry            1585 drivers/net/wireless/ralink/rt2x00/rt73usb.c 	entry->skb = NULL;
entry            1588 drivers/net/wireless/ralink/rt2x00/rt73usb.c static void rt73usb_clear_beacon(struct queue_entry *entry)
entry            1590 drivers/net/wireless/ralink/rt2x00/rt73usb.c 	struct rt2x00_dev *rt2x00dev = entry->queue->rt2x00dev;
entry            1606 drivers/net/wireless/ralink/rt2x00/rt73usb.c 	beacon_base = HW_BEACON_OFFSET(entry->entry_idx);
entry            1615 drivers/net/wireless/ralink/rt2x00/rt73usb.c static int rt73usb_get_tx_data_len(struct queue_entry *entry)
entry            1623 drivers/net/wireless/ralink/rt2x00/rt73usb.c 	length = roundup(entry->skb->len, 4);
entry            1624 drivers/net/wireless/ralink/rt2x00/rt73usb.c 	length += (4 * !(length % entry->queue->usb_maxpacket));
entry            1667 drivers/net/wireless/ralink/rt2x00/rt73usb.c static void rt73usb_fill_rxdone(struct queue_entry *entry,
entry            1670 drivers/net/wireless/ralink/rt2x00/rt73usb.c 	struct rt2x00_dev *rt2x00dev = entry->queue->rt2x00dev;
entry            1671 drivers/net/wireless/ralink/rt2x00/rt73usb.c 	struct skb_frame_desc *skbdesc = get_skb_frame_desc(entry->skb);
entry            1672 drivers/net/wireless/ralink/rt2x00/rt73usb.c 	__le32 *rxd = (__le32 *)entry->skb->data;
entry            1742 drivers/net/wireless/ralink/rt2x00/rt73usb.c 	skb_pull(entry->skb, entry->queue->desc_size);
entry            1743 drivers/net/wireless/ralink/rt2x00/rt73usb.c 	skb_trim(entry->skb, rxdesc->size);
entry             221 drivers/net/wireless/realtek/rtl818x/rtl8180/dev.c 		void *entry = priv->rx_ring + priv->rx_idx * priv->rx_ring_sz;
entry             227 drivers/net/wireless/realtek/rtl818x/rtl8180/dev.c 			struct rtl8187se_rx_desc *desc = entry;
entry             240 drivers/net/wireless/realtek/rtl818x/rtl8180/dev.c 			struct rtl8180_rx_desc *desc = entry;
entry             331 drivers/net/wireless/realtek/rtl818x/rtl8180/dev.c 		cmd_desc = entry;
entry             348 drivers/net/wireless/realtek/rtl818x/rtl8180/dev.c 		struct rtl8180_tx_desc *entry = &ring->desc[ring->idx];
entry             351 drivers/net/wireless/realtek/rtl818x/rtl8180/dev.c 		u32 flags = le32_to_cpu(entry->flags);
entry             358 drivers/net/wireless/realtek/rtl818x/rtl8180/dev.c 		pci_unmap_single(priv->pdev, le32_to_cpu(entry->tx_buf),
entry             462 drivers/net/wireless/realtek/rtl818x/rtl8180/dev.c 	struct rtl8180_tx_desc *entry;
entry             545 drivers/net/wireless/realtek/rtl818x/rtl8180/dev.c 	entry = &ring->desc[idx];
entry             548 drivers/net/wireless/realtek/rtl818x/rtl8180/dev.c 		entry->frame_duration = cpu_to_le16(frame_duration);
entry             549 drivers/net/wireless/realtek/rtl818x/rtl8180/dev.c 		entry->frame_len_se = cpu_to_le16(skb->len);
entry             552 drivers/net/wireless/realtek/rtl818x/rtl8180/dev.c 		entry->flags3 = cpu_to_le16(1<<4);
entry             554 drivers/net/wireless/realtek/rtl818x/rtl8180/dev.c 		entry->frame_len = cpu_to_le32(skb->len);
entry             556 drivers/net/wireless/realtek/rtl818x/rtl8180/dev.c 	entry->rts_duration = rts_duration;
entry             557 drivers/net/wireless/realtek/rtl818x/rtl8180/dev.c 	entry->plcp_len = cpu_to_le16(plcp_len);
entry             558 drivers/net/wireless/realtek/rtl818x/rtl8180/dev.c 	entry->tx_buf = cpu_to_le32(mapping);
entry             560 drivers/net/wireless/realtek/rtl818x/rtl8180/dev.c 	entry->retry_limit = info->control.rates[0].count - 1;
entry             566 drivers/net/wireless/realtek/rtl818x/rtl8180/dev.c 	entry->flags = cpu_to_le32(tx_flags);
entry             999 drivers/net/wireless/realtek/rtl818x/rtl8180/dev.c 	struct rtl818x_rx_cmd_desc *entry;
entry            1019 drivers/net/wireless/realtek/rtl818x/rtl8180/dev.c 		entry = priv->rx_ring + priv->rx_ring_sz*i;
entry            1039 drivers/net/wireless/realtek/rtl818x/rtl8180/dev.c 		entry->rx_buf = cpu_to_le32(*mapping);
entry            1040 drivers/net/wireless/realtek/rtl818x/rtl8180/dev.c 		entry->flags = cpu_to_le32(RTL818X_RX_DESC_FLAG_OWN |
entry            1043 drivers/net/wireless/realtek/rtl818x/rtl8180/dev.c 	entry->flags |= cpu_to_le32(RTL818X_RX_DESC_FLAG_EOR);
entry            1103 drivers/net/wireless/realtek/rtl818x/rtl8180/dev.c 		struct rtl8180_tx_desc *entry = &ring->desc[ring->idx];
entry            1106 drivers/net/wireless/realtek/rtl818x/rtl8180/dev.c 		pci_unmap_single(priv->pdev, le32_to_cpu(entry->tx_buf),
entry             417 drivers/net/wireless/realtek/rtl818x/rtl8187/dev.c 	struct urb *entry = NULL;
entry             428 drivers/net/wireless/realtek/rtl818x/rtl8187/dev.c 		entry = usb_alloc_urb(0, GFP_KERNEL);
entry             429 drivers/net/wireless/realtek/rtl818x/rtl8187/dev.c 		if (!entry) {
entry             433 drivers/net/wireless/realtek/rtl818x/rtl8187/dev.c 		usb_fill_bulk_urb(entry, priv->udev,
entry             439 drivers/net/wireless/realtek/rtl818x/rtl8187/dev.c 		info->urb = entry;
entry             442 drivers/net/wireless/realtek/rtl818x/rtl8187/dev.c 		usb_anchor_urb(entry, &priv->anchored);
entry             443 drivers/net/wireless/realtek/rtl818x/rtl8187/dev.c 		ret = usb_submit_urb(entry, GFP_KERNEL);
entry             446 drivers/net/wireless/realtek/rtl818x/rtl8187/dev.c 			usb_unanchor_urb(entry);
entry             447 drivers/net/wireless/realtek/rtl818x/rtl8187/dev.c 			usb_put_urb(entry);
entry             450 drivers/net/wireless/realtek/rtl818x/rtl8187/dev.c 		usb_put_urb(entry);
entry             550 drivers/net/wireless/realtek/rtl818x/rtl8187/dev.c 	struct urb *entry;
entry             553 drivers/net/wireless/realtek/rtl818x/rtl8187/dev.c 	entry = usb_alloc_urb(0, GFP_KERNEL);
entry             554 drivers/net/wireless/realtek/rtl818x/rtl8187/dev.c 	if (!entry)
entry             557 drivers/net/wireless/realtek/rtl818x/rtl8187/dev.c 	usb_fill_bulk_urb(entry, priv->udev, usb_rcvbulkpipe(priv->udev, 9),
entry             561 drivers/net/wireless/realtek/rtl818x/rtl8187/dev.c 	usb_anchor_urb(entry, &priv->anchored);
entry             562 drivers/net/wireless/realtek/rtl818x/rtl8187/dev.c 	ret = usb_submit_urb(entry, GFP_KERNEL);
entry             564 drivers/net/wireless/realtek/rtl818x/rtl8187/dev.c 		usb_unanchor_urb(entry);
entry             565 drivers/net/wireless/realtek/rtl818x/rtl8187/dev.c 	usb_free_urb(entry);
entry            1936 drivers/net/wireless/realtek/rtlwifi/base.c 	struct rtl_bssid_entry *entry, *next;
entry            1938 drivers/net/wireless/realtek/rtlwifi/base.c 	list_for_each_entry_safe(entry, next, &rtlpriv->scan_list.list, list) {
entry            1939 drivers/net/wireless/realtek/rtlwifi/base.c 		list_del(&entry->list);
entry            1940 drivers/net/wireless/realtek/rtlwifi/base.c 		kfree(entry);
entry            1967 drivers/net/wireless/realtek/rtlwifi/base.c 	struct rtl_bssid_entry *entry, *next;
entry            1972 drivers/net/wireless/realtek/rtlwifi/base.c 	list_for_each_entry_safe(entry, next, &rtlpriv->scan_list.list, list) {
entry            1974 drivers/net/wireless/realtek/rtlwifi/base.c 		if (jiffies_to_msecs(jiffies - entry->age) < 180000)
entry            1977 drivers/net/wireless/realtek/rtlwifi/base.c 		list_del(&entry->list);
entry            1982 drivers/net/wireless/realtek/rtlwifi/base.c 			 entry->bssid, rtlpriv->scan_list.num);
entry            1983 drivers/net/wireless/realtek/rtlwifi/base.c 		kfree(entry);
entry            1998 drivers/net/wireless/realtek/rtlwifi/base.c 	struct rtl_bssid_entry *entry;
entry            2012 drivers/net/wireless/realtek/rtlwifi/base.c 	list_for_each_entry(entry, &rtlpriv->scan_list.list, list) {
entry            2013 drivers/net/wireless/realtek/rtlwifi/base.c 		if (memcmp(entry->bssid, hdr->addr3, ETH_ALEN) == 0) {
entry            2014 drivers/net/wireless/realtek/rtlwifi/base.c 			list_del_init(&entry->list);
entry            2024 drivers/net/wireless/realtek/rtlwifi/base.c 		entry = kmalloc(sizeof(*entry), GFP_ATOMIC);
entry            2026 drivers/net/wireless/realtek/rtlwifi/base.c 		if (!entry)
entry            2029 drivers/net/wireless/realtek/rtlwifi/base.c 		memcpy(entry->bssid, hdr->addr3, ETH_ALEN);
entry            2037 drivers/net/wireless/realtek/rtlwifi/base.c 	entry->age = jiffies;
entry            2039 drivers/net/wireless/realtek/rtlwifi/base.c 	list_add_tail(&entry->list, &rtlpriv->scan_list.list);
entry             538 drivers/net/wireless/realtek/rtlwifi/pci.c 		u8 *entry;
entry             541 drivers/net/wireless/realtek/rtlwifi/pci.c 			entry = (u8 *)(&ring->buffer_desc[ring->idx]);
entry             543 drivers/net/wireless/realtek/rtlwifi/pci.c 			entry = (u8 *)(&ring->desc[ring->idx]);
entry             552 drivers/net/wireless/realtek/rtlwifi/pci.c 					     get_desc(hw, (u8 *)entry, true,
entry             629 drivers/net/wireless/realtek/rtlwifi/pci.c 				    struct sk_buff *new_skb, u8 *entry,
entry             657 drivers/net/wireless/realtek/rtlwifi/pci.c 		rtlpriv->cfg->ops->set_desc(hw, (u8 *)entry, false,
entry             661 drivers/net/wireless/realtek/rtlwifi/pci.c 		rtlpriv->cfg->ops->set_desc(hw, (u8 *)entry, false,
entry             664 drivers/net/wireless/realtek/rtlwifi/pci.c 		rtlpriv->cfg->ops->set_desc(hw, (u8 *)entry, false,
entry             667 drivers/net/wireless/realtek/rtlwifi/pci.c 		rtlpriv->cfg->ops->set_desc(hw, (u8 *)entry, false,
entry            1085 drivers/net/wireless/realtek/rtlwifi/pci.c 	u8 *entry;
entry            1091 drivers/net/wireless/realtek/rtlwifi/pci.c 		entry = (u8 *)(&ring->buffer_desc[ring->idx]);
entry            1093 drivers/net/wireless/realtek/rtlwifi/pci.c 		entry = (u8 *)(&ring->desc[ring->idx]);
entry            1097 drivers/net/wireless/realtek/rtlwifi/pci.c 				 hw, (u8 *)entry, true, HW_DESC_TXBUFF_ADDR),
entry            1280 drivers/net/wireless/realtek/rtlwifi/pci.c 		struct rtl_rx_buffer_desc *entry = NULL;
entry            1297 drivers/net/wireless/realtek/rtlwifi/pci.c 			entry = &rtlpci->rx_ring[rxring_idx].buffer_desc[i];
entry            1298 drivers/net/wireless/realtek/rtlwifi/pci.c 			if (!_rtl_pci_init_one_rxdesc(hw, NULL, (u8 *)entry,
entry            1303 drivers/net/wireless/realtek/rtlwifi/pci.c 		struct rtl_rx_desc *entry = NULL;
entry            1321 drivers/net/wireless/realtek/rtlwifi/pci.c 			entry = &rtlpci->rx_ring[rxring_idx].desc[i];
entry            1322 drivers/net/wireless/realtek/rtlwifi/pci.c 			if (!_rtl_pci_init_one_rxdesc(hw, NULL, (u8 *)entry,
entry            1327 drivers/net/wireless/realtek/rtlwifi/pci.c 		rtlpriv->cfg->ops->set_desc(hw, (u8 *)entry, false,
entry            1342 drivers/net/wireless/realtek/rtlwifi/pci.c 		u8 *entry;
entry            1346 drivers/net/wireless/realtek/rtlwifi/pci.c 			entry = (u8 *)(&ring->buffer_desc[ring->idx]);
entry            1348 drivers/net/wireless/realtek/rtlwifi/pci.c 			entry = (u8 *)(&ring->desc[ring->idx]);
entry            1351 drivers/net/wireless/realtek/rtlwifi/pci.c 				 rtlpriv->cfg->ops->get_desc(hw, (u8 *)entry,
entry            1474 drivers/net/wireless/realtek/rtlwifi/pci.c 			struct rtl_rx_desc *entry = NULL;
entry            1478 drivers/net/wireless/realtek/rtlwifi/pci.c 				entry = &rtlpci->rx_ring[rxring_idx].desc[i];
entry            1480 drivers/net/wireless/realtek/rtlwifi/pci.c 				  rtlpriv->cfg->ops->get_desc(hw, (u8 *)entry,
entry            1482 drivers/net/wireless/realtek/rtlwifi/pci.c 				memset((u8 *)entry, 0,
entry            1487 drivers/net/wireless/realtek/rtlwifi/pci.c 					    (u8 *)entry, false,
entry            1492 drivers/net/wireless/realtek/rtlwifi/pci.c 					    (u8 *)entry, false,
entry            1496 drivers/net/wireless/realtek/rtlwifi/pci.c 					    (u8 *)entry, false,
entry            1500 drivers/net/wireless/realtek/rtlwifi/pci.c 					    (u8 *)entry, false,
entry            1505 drivers/net/wireless/realtek/rtlwifi/pci.c 			rtlpriv->cfg->ops->set_desc(hw, (u8 *)entry, false,
entry            1521 drivers/net/wireless/realtek/rtlwifi/pci.c 				u8 *entry;
entry            1525 drivers/net/wireless/realtek/rtlwifi/pci.c 					entry = (u8 *)(&ring->buffer_desc
entry            1528 drivers/net/wireless/realtek/rtlwifi/pci.c 					entry = (u8 *)(&ring->desc[ring->idx]);
entry            1533 drivers/net/wireless/realtek/rtlwifi/pci.c 							 entry,
entry              75 drivers/net/wireless/realtek/rtlwifi/rtl8188ee/hw.c 		struct rtl_tx_desc *entry = &ring->desc[ring->idx];
entry              81 drivers/net/wireless/realtek/rtlwifi/rtl8188ee/hw.c 				 (u8 *)entry, true, HW_DESC_TXBUFF_ADDR),
entry             814 drivers/net/wireless/realtek/rtlwifi/rtl8188ee/trx.c 	u8 *entry = (u8 *)(&ring->desc[ring->idx]);
entry             815 drivers/net/wireless/realtek/rtlwifi/rtl8188ee/trx.c 	u8 own = (u8)rtl88ee_get_desc(hw, entry, true, HW_DESC_OWN);
entry             712 drivers/net/wireless/realtek/rtlwifi/rtl8192ce/trx.c 	u8 *entry = (u8 *)(&ring->desc[ring->idx]);
entry             713 drivers/net/wireless/realtek/rtlwifi/rtl8192ce/trx.c 	u8 own = (u8)rtl92ce_get_desc(hw, entry, true, HW_DESC_OWN);
entry             831 drivers/net/wireless/realtek/rtlwifi/rtl8192de/trx.c 	u8 *entry = (u8 *)(&ring->desc[ring->idx]);
entry             832 drivers/net/wireless/realtek/rtlwifi/rtl8192de/trx.c 	u8 own = (u8)rtl92de_get_desc(hw, entry, true, HW_DESC_OWN);
entry             219 drivers/net/wireless/realtek/rtlwifi/rtl8192se/sw.c 	u8 *entry = (u8 *)(&ring->desc[ring->idx]);
entry             220 drivers/net/wireless/realtek/rtlwifi/rtl8192se/sw.c 	u8 own = (u8)rtl92se_get_desc(hw, entry, true, HW_DESC_OWN);
entry             673 drivers/net/wireless/realtek/rtlwifi/rtl8723ae/trx.c 	u8 *entry = (u8 *)(&ring->desc[ring->idx]);
entry             674 drivers/net/wireless/realtek/rtlwifi/rtl8723ae/trx.c 	u8 own = (u8)rtl8723e_get_desc(hw, entry, true, HW_DESC_OWN);
entry              37 drivers/net/wireless/realtek/rtlwifi/rtl8723be/hw.c 		struct rtl_tx_desc *entry = &ring->desc[ring->idx];
entry              43 drivers/net/wireless/realtek/rtlwifi/rtl8723be/hw.c 				 (u8 *)entry, true, HW_DESC_TXBUFF_ADDR),
entry             726 drivers/net/wireless/realtek/rtlwifi/rtl8723be/trx.c 	u8 *entry = (u8 *)(&ring->desc[ring->idx]);
entry             727 drivers/net/wireless/realtek/rtlwifi/rtl8723be/trx.c 	u8 own = (u8)rtl8723be_get_desc(hw, entry, true, HW_DESC_OWN);
entry              33 drivers/net/wireless/realtek/rtlwifi/rtl8821ae/hw.c 		struct rtl_tx_desc *entry = &ring->desc[ring->idx];
entry              39 drivers/net/wireless/realtek/rtlwifi/rtl8821ae/hw.c 				 (u8 *)entry, true, HW_DESC_TXBUFF_ADDR),
entry             966 drivers/net/wireless/realtek/rtlwifi/rtl8821ae/trx.c 	u8 *entry = (u8 *)(&ring->desc[ring->idx]);
entry             967 drivers/net/wireless/realtek/rtlwifi/rtl8821ae/trx.c 	u8 own = (u8)rtl8821ae_get_desc(hw, entry, true, HW_DESC_OWN);
entry             448 drivers/net/wireless/st/cw1200/sta.c 		.entry[0].ie_id = WLAN_EID_VENDOR_SPECIFIC,
entry             449 drivers/net/wireless/st/cw1200/sta.c 		.entry[0].flags = WSM_BEACON_FILTER_IE_HAS_CHANGED |
entry             452 drivers/net/wireless/st/cw1200/sta.c 		.entry[0].oui[0] = 0x50,
entry             453 drivers/net/wireless/st/cw1200/sta.c 		.entry[0].oui[1] = 0x6F,
entry             454 drivers/net/wireless/st/cw1200/sta.c 		.entry[0].oui[2] = 0x9A,
entry             455 drivers/net/wireless/st/cw1200/sta.c 		.entry[1].ie_id = WLAN_EID_HT_OPERATION,
entry             456 drivers/net/wireless/st/cw1200/sta.c 		.entry[1].flags = WSM_BEACON_FILTER_IE_HAS_CHANGED |
entry             459 drivers/net/wireless/st/cw1200/sta.c 		.entry[2].ie_id = WLAN_EID_ERP_INFO,
entry             460 drivers/net/wireless/st/cw1200/sta.c 		.entry[2].flags = WSM_BEACON_FILTER_IE_HAS_CHANGED |
entry            1545 drivers/net/wireless/st/cw1200/sta.c 	struct cw1200_link_entry *entry;
entry            1558 drivers/net/wireless/st/cw1200/sta.c 	entry = &priv->link_id_db[sta_priv->link_id - 1];
entry            1563 drivers/net/wireless/st/cw1200/sta.c 	entry->status = CW1200_LINK_HARD;
entry            1564 drivers/net/wireless/st/cw1200/sta.c 	while ((skb = skb_dequeue(&entry->rx_queue)))
entry            1576 drivers/net/wireless/st/cw1200/sta.c 	struct cw1200_link_entry *entry;
entry            1581 drivers/net/wireless/st/cw1200/sta.c 	entry = &priv->link_id_db[sta_priv->link_id - 1];
entry            1583 drivers/net/wireless/st/cw1200/sta.c 	entry->status = CW1200_LINK_RESERVE;
entry            1584 drivers/net/wireless/st/cw1200/sta.c 	entry->timestamp = jiffies;
entry             235 drivers/net/wireless/st/cw1200/txrx.c 				 struct tx_policy_cache_entry *entry)
entry             237 drivers/net/wireless/st/cw1200/txrx.c 	++entry->policy.usage_count;
entry             238 drivers/net/wireless/st/cw1200/txrx.c 	list_move(&entry->link, &cache->used);
entry             242 drivers/net/wireless/st/cw1200/txrx.c 				    struct tx_policy_cache_entry *entry)
entry             244 drivers/net/wireless/st/cw1200/txrx.c 	int ret = --entry->policy.usage_count;
entry             246 drivers/net/wireless/st/cw1200/txrx.c 		list_move(&entry->link, &cache->free);
entry             254 drivers/net/wireless/st/cw1200/txrx.c 	struct tx_policy_cache_entry *entry;
entry             261 drivers/net/wireless/st/cw1200/txrx.c 		entry = &cache->cache[idx];
entry             265 drivers/net/wireless/st/cw1200/txrx.c 		if (WARN_ON(entry->policy.usage_count)) {
entry             266 drivers/net/wireless/st/cw1200/txrx.c 			entry->policy.usage_count = 0;
entry             267 drivers/net/wireless/st/cw1200/txrx.c 			list_move(&entry->link, &cache->free);
entry             269 drivers/net/wireless/st/cw1200/txrx.c 		memset(&entry->policy, 0, sizeof(entry->policy));
entry             316 drivers/net/wireless/st/cw1200/txrx.c 		struct tx_policy_cache_entry *entry;
entry             321 drivers/net/wireless/st/cw1200/txrx.c 		entry = list_entry(cache->free.prev,
entry             323 drivers/net/wireless/st/cw1200/txrx.c 		entry->policy = wanted;
entry             324 drivers/net/wireless/st/cw1200/txrx.c 		idx = entry - cache->cache;
entry             326 drivers/net/wireless/st/cw1200/txrx.c 		tx_policy_dump(&entry->policy);
entry            1012 drivers/net/wireless/st/cw1200/txrx.c 	struct cw1200_link_entry *entry = NULL;
entry            1026 drivers/net/wireless/st/cw1200/txrx.c 		entry =	&priv->link_id_db[link_id - 1];
entry            1027 drivers/net/wireless/st/cw1200/txrx.c 		if (entry->status == CW1200_LINK_SOFT &&
entry            1030 drivers/net/wireless/st/cw1200/txrx.c 		entry->timestamp = jiffies;
entry            1209 drivers/net/wireless/st/cw1200/txrx.c 		if (entry->status == CW1200_LINK_SOFT)
entry            1210 drivers/net/wireless/st/cw1200/txrx.c 			skb_queue_tail(&entry->rx_queue, skb);
entry            1346 drivers/net/wireless/st/cw1200/txrx.c 		struct cw1200_link_entry *entry = &priv->link_id_db[ret - 1];
entry            1348 drivers/net/wireless/st/cw1200/txrx.c 		entry->status = CW1200_LINK_RESERVE;
entry            1349 drivers/net/wireless/st/cw1200/txrx.c 		memcpy(&entry->mac, mac, ETH_ALEN);
entry            1350 drivers/net/wireless/st/cw1200/txrx.c 		memset(&entry->buffered, 0, CW1200_MAX_TID);
entry            1351 drivers/net/wireless/st/cw1200/txrx.c 		skb_queue_head_init(&entry->rx_queue);
entry            1401 drivers/net/wireless/st/cw1200/wsm.h 	struct wsm_beacon_filter_table_entry entry[10];
entry             214 drivers/net/xen-netback/common.h 	struct list_head entry;
entry              38 drivers/net/xen-netback/hash.c 	struct xenvif_hash_cache_entry *new, *entry, *oldest;
entry              42 drivers/net/xen-netback/hash.c 	new = kmalloc(sizeof(*entry), GFP_ATOMIC);
entry              54 drivers/net/xen-netback/hash.c 	list_for_each_entry_rcu(entry, &vif->hash.cache.list, link) {
entry              56 drivers/net/xen-netback/hash.c 		if (entry->len == len &&
entry              57 drivers/net/xen-netback/hash.c 		    memcmp(entry->tag, tag, len) == 0)
entry              59 drivers/net/xen-netback/hash.c 		if (!oldest || entry->seq < oldest->seq)
entry              60 drivers/net/xen-netback/hash.c 			oldest = entry;
entry              97 drivers/net/xen-netback/hash.c 	struct xenvif_hash_cache_entry *entry;
entry             105 drivers/net/xen-netback/hash.c 	list_for_each_entry_rcu(entry, &vif->hash.cache.list, link) {
entry             106 drivers/net/xen-netback/hash.c 		list_del_rcu(&entry->link);
entry             108 drivers/net/xen-netback/hash.c 		kfree_rcu(entry, rcu);
entry             117 drivers/net/xen-netback/hash.c 	struct xenvif_hash_cache_entry *entry;
entry             131 drivers/net/xen-netback/hash.c 	list_for_each_entry_rcu(entry, &vif->hash.cache.list, link) {
entry             132 drivers/net/xen-netback/hash.c 		if (entry->len == len &&
entry             133 drivers/net/xen-netback/hash.c 		    memcmp(entry->tag, data, len) == 0) {
entry             134 drivers/net/xen-netback/hash.c 			val = entry->val;
entry             135 drivers/net/xen-netback/hash.c 			entry->seq = atomic_inc_return(&vif->hash.cache.seq);
entry             735 drivers/net/xen-netback/netback.c 	list_add_tail_rcu(&mcast->entry, &vif->fe_mcast_addr);
entry             745 drivers/net/xen-netback/netback.c 	list_for_each_entry_rcu(mcast, &vif->fe_mcast_addr, entry) {
entry             748 drivers/net/xen-netback/netback.c 			list_del_rcu(&mcast->entry);
entry             760 drivers/net/xen-netback/netback.c 	list_for_each_entry_rcu(mcast, &vif->fe_mcast_addr, entry) {
entry             781 drivers/net/xen-netback/netback.c 					 entry);
entry             783 drivers/net/xen-netback/netback.c 		list_del(&mcast->entry);
entry             606 drivers/ntb/hw/amd/ntb_hw_amd.c 		ndev->msix[i].entry = i;
entry             392 drivers/ntb/hw/intel/ntb_hw_gen1.c 		ndev->msix[i].entry = i;
entry             191 drivers/ntb/msi.c 	struct msi_desc *entry;
entry             195 drivers/ntb/msi.c static int ntb_msi_set_desc(struct ntb_dev *ntb, struct msi_desc *entry,
entry             200 drivers/ntb/msi.c 	addr = entry->msg.address_lo +
entry             201 drivers/ntb/msi.c 		((uint64_t)entry->msg.address_hi << 32);
entry             206 drivers/ntb/msi.c 			      entry->irq, addr, ntb->msi->base_addr,
entry             212 drivers/ntb/msi.c 	msi_desc->data = entry->msg.data;
entry             217 drivers/ntb/msi.c static void ntb_msi_write_msg(struct msi_desc *entry, void *data)
entry             221 drivers/ntb/msi.c 	WARN_ON(ntb_msi_set_desc(dr->ntb, entry, dr->msi_desc));
entry             231 drivers/ntb/msi.c 	dr->entry->write_msi_msg = NULL;
entry             232 drivers/ntb/msi.c 	dr->entry->write_msi_msg_data = NULL;
entry             235 drivers/ntb/msi.c static int ntbm_msi_setup_callback(struct ntb_dev *ntb, struct msi_desc *entry,
entry             246 drivers/ntb/msi.c 	dr->entry = entry;
entry             251 drivers/ntb/msi.c 	dr->entry->write_msi_msg = ntb_msi_write_msg;
entry             252 drivers/ntb/msi.c 	dr->entry->write_msi_msg_data = dr;
entry             284 drivers/ntb/msi.c 	struct msi_desc *entry;
entry             291 drivers/ntb/msi.c 	for_each_pci_msi_entry(entry, ntb->pdev) {
entry             292 drivers/ntb/msi.c 		desc = irq_to_desc(entry->irq);
entry             296 drivers/ntb/msi.c 		ret = devm_request_threaded_irq(&ntb->dev, entry->irq, handler,
entry             301 drivers/ntb/msi.c 		if (ntb_msi_set_desc(ntb, entry, msi_desc)) {
entry             302 drivers/ntb/msi.c 			devm_free_irq(&ntb->dev, entry->irq, dev_id);
entry             306 drivers/ntb/msi.c 		ret = ntbm_msi_setup_callback(ntb, entry, msi_desc);
entry             308 drivers/ntb/msi.c 			devm_free_irq(&ntb->dev, entry->irq, dev_id);
entry             313 drivers/ntb/msi.c 		return entry->irq;
entry             325 drivers/ntb/msi.c 	return dr->ntb == ntb && dr->entry == data;
entry             339 drivers/ntb/msi.c 	struct msi_desc *entry = irq_get_msi_desc(irq);
entry             341 drivers/ntb/msi.c 	entry->write_msi_msg = NULL;
entry             342 drivers/ntb/msi.c 	entry->write_msi_msg_data = NULL;
entry             345 drivers/ntb/msi.c 			       ntbm_msi_callback_match, entry));
entry             109 drivers/ntb/ntb_transport.c 	struct list_head entry;
entry             128 drivers/ntb/ntb_transport.c 	unsigned int entry;
entry             217 drivers/ntb/ntb_transport.c 	struct list_head entry;
entry             223 drivers/ntb/ntb_transport.c 	struct list_head entry;
entry             280 drivers/ntb/ntb_transport.c 			       struct ntb_queue_entry *entry);
entry             281 drivers/ntb/ntb_transport.c static void ntb_memcpy_tx(struct ntb_queue_entry *entry, void __iomem *offset);
entry             282 drivers/ntb/ntb_transport.c static int ntb_async_rx_submit(struct ntb_queue_entry *entry, void *offset);
entry             283 drivers/ntb/ntb_transport.c static void ntb_memcpy_rx(struct ntb_queue_entry *entry, void *offset);
entry             330 drivers/ntb/ntb_transport.c 	list_add_tail(&nt->entry, &ntb_transport_list);
entry             338 drivers/ntb/ntb_transport.c 	list_for_each_entry_safe(client_dev, cd, &nt->client_devs, entry) {
entry             341 drivers/ntb/ntb_transport.c 		list_del(&client_dev->entry);
entry             345 drivers/ntb/ntb_transport.c 	list_del(&nt->entry);
entry             367 drivers/ntb/ntb_transport.c 	list_for_each_entry(nt, &ntb_transport_list, entry)
entry             368 drivers/ntb/ntb_transport.c 		list_for_each_entry_safe(client, cd, &nt->client_devs, entry)
entry             371 drivers/ntb/ntb_transport.c 				list_del(&client->entry);
entry             393 drivers/ntb/ntb_transport.c 	list_for_each_entry(nt, &ntb_transport_list, entry) {
entry             419 drivers/ntb/ntb_transport.c 		list_add_tail(&client_dev->entry, &nt->client_devs);
entry             529 drivers/ntb/ntb_transport.c 			       qp->remote_rx_info->entry);
entry             564 drivers/ntb/ntb_transport.c static void ntb_list_add(spinlock_t *lock, struct list_head *entry,
entry             570 drivers/ntb/ntb_transport.c 	list_add_tail(entry, list);
entry             577 drivers/ntb/ntb_transport.c 	struct ntb_queue_entry *entry;
entry             582 drivers/ntb/ntb_transport.c 		entry = NULL;
entry             585 drivers/ntb/ntb_transport.c 	entry = list_first_entry(list, struct ntb_queue_entry, entry);
entry             586 drivers/ntb/ntb_transport.c 	list_del(&entry->entry);
entry             591 drivers/ntb/ntb_transport.c 	return entry;
entry             598 drivers/ntb/ntb_transport.c 	struct ntb_queue_entry *entry;
entry             604 drivers/ntb/ntb_transport.c 		entry = NULL;
entry             606 drivers/ntb/ntb_transport.c 		entry = list_first_entry(list, struct ntb_queue_entry, entry);
entry             607 drivers/ntb/ntb_transport.c 		list_move_tail(&entry->entry, to_list);
entry             612 drivers/ntb/ntb_transport.c 	return entry;
entry             621 drivers/ntb/ntb_transport.c 	struct ntb_queue_entry *entry;
entry             659 drivers/ntb/ntb_transport.c 		entry = kzalloc_node(sizeof(*entry), GFP_KERNEL, node);
entry             660 drivers/ntb/ntb_transport.c 		if (!entry)
entry             663 drivers/ntb/ntb_transport.c 		entry->qp = qp;
entry             664 drivers/ntb/ntb_transport.c 		ntb_list_add(&qp->ntb_rx_q_lock, &entry->entry,
entry             669 drivers/ntb/ntb_transport.c 	qp->remote_rx_info->entry = qp->rx_max_entry - 1;
entry            1441 drivers/ntb/ntb_transport.c 	struct ntb_queue_entry *entry;
entry            1449 drivers/ntb/ntb_transport.c 		entry = list_first_entry(&qp->rx_post_q,
entry            1450 drivers/ntb/ntb_transport.c 					 struct ntb_queue_entry, entry);
entry            1451 drivers/ntb/ntb_transport.c 		if (!(entry->flags & DESC_DONE_FLAG))
entry            1454 drivers/ntb/ntb_transport.c 		entry->rx_hdr->flags = 0;
entry            1455 drivers/ntb/ntb_transport.c 		iowrite32(entry->rx_index, &qp->rx_info->entry);
entry            1457 drivers/ntb/ntb_transport.c 		cb_data = entry->cb_data;
entry            1458 drivers/ntb/ntb_transport.c 		len = entry->len;
entry            1460 drivers/ntb/ntb_transport.c 		list_move_tail(&entry->entry, &qp->rx_free_q);
entry            1476 drivers/ntb/ntb_transport.c 	struct ntb_queue_entry *entry = data;
entry            1485 drivers/ntb/ntb_transport.c 			entry->errors++;
entry            1489 drivers/ntb/ntb_transport.c 			struct ntb_transport_qp *qp = entry->qp;
entry            1493 drivers/ntb/ntb_transport.c 			ntb_memcpy_rx(entry, offset);
entry            1504 drivers/ntb/ntb_transport.c 	entry->flags |= DESC_DONE_FLAG;
entry            1506 drivers/ntb/ntb_transport.c 	ntb_complete_rxc(entry->qp);
entry            1509 drivers/ntb/ntb_transport.c static void ntb_memcpy_rx(struct ntb_queue_entry *entry, void *offset)
entry            1511 drivers/ntb/ntb_transport.c 	void *buf = entry->buf;
entry            1512 drivers/ntb/ntb_transport.c 	size_t len = entry->len;
entry            1519 drivers/ntb/ntb_transport.c 	ntb_rx_copy_callback(entry, NULL);
entry            1522 drivers/ntb/ntb_transport.c static int ntb_async_rx_submit(struct ntb_queue_entry *entry, void *offset)
entry            1525 drivers/ntb/ntb_transport.c 	struct ntb_transport_qp *qp = entry->qp;
entry            1531 drivers/ntb/ntb_transport.c 	void *buf = entry->buf;
entry            1533 drivers/ntb/ntb_transport.c 	len = entry->len;
entry            1567 drivers/ntb/ntb_transport.c 	txd->callback_param = entry;
entry            1590 drivers/ntb/ntb_transport.c static void ntb_async_rx(struct ntb_queue_entry *entry, void *offset)
entry            1592 drivers/ntb/ntb_transport.c 	struct ntb_transport_qp *qp = entry->qp;
entry            1599 drivers/ntb/ntb_transport.c 	if (entry->len < copy_bytes)
entry            1602 drivers/ntb/ntb_transport.c 	res = ntb_async_rx_submit(entry, offset);
entry            1606 drivers/ntb/ntb_transport.c 	if (!entry->retries)
entry            1612 drivers/ntb/ntb_transport.c 	ntb_memcpy_rx(entry, offset);
entry            1619 drivers/ntb/ntb_transport.c 	struct ntb_queue_entry *entry;
entry            1649 drivers/ntb/ntb_transport.c 	entry = ntb_list_mv(&qp->ntb_rx_q_lock, &qp->rx_pend_q, &qp->rx_post_q);
entry            1650 drivers/ntb/ntb_transport.c 	if (!entry) {
entry            1656 drivers/ntb/ntb_transport.c 	entry->rx_hdr = hdr;
entry            1657 drivers/ntb/ntb_transport.c 	entry->rx_index = qp->rx_index;
entry            1659 drivers/ntb/ntb_transport.c 	if (hdr->len > entry->len) {
entry            1662 drivers/ntb/ntb_transport.c 			hdr->len, entry->len);
entry            1665 drivers/ntb/ntb_transport.c 		entry->len = -EIO;
entry            1666 drivers/ntb/ntb_transport.c 		entry->flags |= DESC_DONE_FLAG;
entry            1672 drivers/ntb/ntb_transport.c 			qp->rx_index, hdr->ver, hdr->len, entry->len);
entry            1677 drivers/ntb/ntb_transport.c 		entry->len = hdr->len;
entry            1679 drivers/ntb/ntb_transport.c 		ntb_async_rx(entry, offset);
entry            1730 drivers/ntb/ntb_transport.c 	struct ntb_queue_entry *entry = data;
entry            1731 drivers/ntb/ntb_transport.c 	struct ntb_transport_qp *qp = entry->qp;
entry            1732 drivers/ntb/ntb_transport.c 	struct ntb_payload_header __iomem *hdr = entry->tx_hdr;
entry            1741 drivers/ntb/ntb_transport.c 			entry->errors++;
entry            1747 drivers/ntb/ntb_transport.c 				entry->tx_index;
entry            1750 drivers/ntb/ntb_transport.c 			ntb_memcpy_tx(entry, offset);
entry            1761 drivers/ntb/ntb_transport.c 	iowrite32(entry->flags | DESC_DONE_FLAG, &hdr->flags);
entry            1772 drivers/ntb/ntb_transport.c 	if (entry->len > 0) {
entry            1773 drivers/ntb/ntb_transport.c 		qp->tx_bytes += entry->len;
entry            1776 drivers/ntb/ntb_transport.c 			qp->tx_handler(qp, qp->cb_data, entry->cb_data,
entry            1777 drivers/ntb/ntb_transport.c 				       entry->len);
entry            1780 drivers/ntb/ntb_transport.c 	ntb_list_add(&qp->ntb_tx_free_q_lock, &entry->entry, &qp->tx_free_q);
entry            1783 drivers/ntb/ntb_transport.c static void ntb_memcpy_tx(struct ntb_queue_entry *entry, void __iomem *offset)
entry            1790 drivers/ntb/ntb_transport.c 	__copy_from_user_inatomic_nocache(offset, entry->buf, entry->len);
entry            1792 drivers/ntb/ntb_transport.c 	memcpy_toio(offset, entry->buf, entry->len);
entry            1798 drivers/ntb/ntb_transport.c 	ntb_tx_copy_callback(entry, NULL);
entry            1802 drivers/ntb/ntb_transport.c 			       struct ntb_queue_entry *entry)
entry            1807 drivers/ntb/ntb_transport.c 	size_t len = entry->len;
entry            1808 drivers/ntb/ntb_transport.c 	void *buf = entry->buf;
entry            1815 drivers/ntb/ntb_transport.c 	dest = qp->tx_mw_dma_addr + qp->tx_max_frame * entry->tx_index;
entry            1840 drivers/ntb/ntb_transport.c 	txd->callback_param = entry;
entry            1861 drivers/ntb/ntb_transport.c 			 struct ntb_queue_entry *entry)
entry            1868 drivers/ntb/ntb_transport.c 	entry->tx_index = qp->tx_index;
entry            1869 drivers/ntb/ntb_transport.c 	offset = qp->tx_mw + qp->tx_max_frame * entry->tx_index;
entry            1871 drivers/ntb/ntb_transport.c 	entry->tx_hdr = hdr;
entry            1873 drivers/ntb/ntb_transport.c 	iowrite32(entry->len, &hdr->len);
entry            1879 drivers/ntb/ntb_transport.c 	if (entry->len < copy_bytes)
entry            1882 drivers/ntb/ntb_transport.c 	res = ntb_async_tx_submit(qp, entry);
entry            1886 drivers/ntb/ntb_transport.c 	if (!entry->retries)
entry            1892 drivers/ntb/ntb_transport.c 	ntb_memcpy_tx(entry, offset);
entry            1897 drivers/ntb/ntb_transport.c 			  struct ntb_queue_entry *entry)
entry            1899 drivers/ntb/ntb_transport.c 	if (qp->tx_index == qp->remote_rx_info->entry) {
entry            1904 drivers/ntb/ntb_transport.c 	if (entry->len > qp->tx_max_frame - sizeof(struct ntb_payload_header)) {
entry            1908 drivers/ntb/ntb_transport.c 		ntb_list_add(&qp->ntb_tx_free_q_lock, &entry->entry,
entry            1913 drivers/ntb/ntb_transport.c 	ntb_async_tx(qp, entry);
entry            1926 drivers/ntb/ntb_transport.c 	struct ntb_queue_entry *entry;
entry            1935 drivers/ntb/ntb_transport.c 		entry = ntb_list_rm(&qp->ntb_tx_free_q_lock, &qp->tx_free_q);
entry            1936 drivers/ntb/ntb_transport.c 		if (entry)
entry            1941 drivers/ntb/ntb_transport.c 	if (!entry)
entry            1944 drivers/ntb/ntb_transport.c 	entry->cb_data = NULL;
entry            1945 drivers/ntb/ntb_transport.c 	entry->buf = NULL;
entry            1946 drivers/ntb/ntb_transport.c 	entry->len = 0;
entry            1947 drivers/ntb/ntb_transport.c 	entry->flags = LINK_DOWN_FLAG;
entry            1949 drivers/ntb/ntb_transport.c 	rc = ntb_process_tx(qp, entry);
entry            1983 drivers/ntb/ntb_transport.c 	struct ntb_queue_entry *entry;
entry            2054 drivers/ntb/ntb_transport.c 		entry = kzalloc_node(sizeof(*entry), GFP_KERNEL, node);
entry            2055 drivers/ntb/ntb_transport.c 		if (!entry)
entry            2058 drivers/ntb/ntb_transport.c 		entry->qp = qp;
entry            2059 drivers/ntb/ntb_transport.c 		ntb_list_add(&qp->ntb_rx_q_lock, &entry->entry,
entry            2065 drivers/ntb/ntb_transport.c 		entry = kzalloc_node(sizeof(*entry), GFP_KERNEL, node);
entry            2066 drivers/ntb/ntb_transport.c 		if (!entry)
entry            2069 drivers/ntb/ntb_transport.c 		entry->qp = qp;
entry            2070 drivers/ntb/ntb_transport.c 		ntb_list_add(&qp->ntb_tx_free_q_lock, &entry->entry,
entry            2082 drivers/ntb/ntb_transport.c 	while ((entry = ntb_list_rm(&qp->ntb_tx_free_q_lock, &qp->tx_free_q)))
entry            2083 drivers/ntb/ntb_transport.c 		kfree(entry);
entry            2086 drivers/ntb/ntb_transport.c 	while ((entry = ntb_list_rm(&qp->ntb_rx_q_lock, &qp->rx_free_q)))
entry            2087 drivers/ntb/ntb_transport.c 		kfree(entry);
entry            2111 drivers/ntb/ntb_transport.c 	struct ntb_queue_entry *entry;
entry            2168 drivers/ntb/ntb_transport.c 	while ((entry = ntb_list_rm(&qp->ntb_rx_q_lock, &qp->rx_free_q)))
entry            2169 drivers/ntb/ntb_transport.c 		kfree(entry);
entry            2171 drivers/ntb/ntb_transport.c 	while ((entry = ntb_list_rm(&qp->ntb_rx_q_lock, &qp->rx_pend_q))) {
entry            2173 drivers/ntb/ntb_transport.c 		kfree(entry);
entry            2176 drivers/ntb/ntb_transport.c 	while ((entry = ntb_list_rm(&qp->ntb_rx_q_lock, &qp->rx_post_q))) {
entry            2178 drivers/ntb/ntb_transport.c 		kfree(entry);
entry            2181 drivers/ntb/ntb_transport.c 	while ((entry = ntb_list_rm(&qp->ntb_tx_free_q_lock, &qp->tx_free_q)))
entry            2182 drivers/ntb/ntb_transport.c 		kfree(entry);
entry            2202 drivers/ntb/ntb_transport.c 	struct ntb_queue_entry *entry;
entry            2208 drivers/ntb/ntb_transport.c 	entry = ntb_list_rm(&qp->ntb_rx_q_lock, &qp->rx_pend_q);
entry            2209 drivers/ntb/ntb_transport.c 	if (!entry)
entry            2212 drivers/ntb/ntb_transport.c 	buf = entry->cb_data;
entry            2213 drivers/ntb/ntb_transport.c 	*len = entry->len;
entry            2215 drivers/ntb/ntb_transport.c 	ntb_list_add(&qp->ntb_rx_q_lock, &entry->entry, &qp->rx_free_q);
entry            2236 drivers/ntb/ntb_transport.c 	struct ntb_queue_entry *entry;
entry            2241 drivers/ntb/ntb_transport.c 	entry = ntb_list_rm(&qp->ntb_rx_q_lock, &qp->rx_free_q);
entry            2242 drivers/ntb/ntb_transport.c 	if (!entry)
entry            2245 drivers/ntb/ntb_transport.c 	entry->cb_data = cb;
entry            2246 drivers/ntb/ntb_transport.c 	entry->buf = data;
entry            2247 drivers/ntb/ntb_transport.c 	entry->len = len;
entry            2248 drivers/ntb/ntb_transport.c 	entry->flags = 0;
entry            2249 drivers/ntb/ntb_transport.c 	entry->retries = 0;
entry            2250 drivers/ntb/ntb_transport.c 	entry->errors = 0;
entry            2251 drivers/ntb/ntb_transport.c 	entry->rx_index = 0;
entry            2253 drivers/ntb/ntb_transport.c 	ntb_list_add(&qp->ntb_rx_q_lock, &entry->entry, &qp->rx_pend_q);
entry            2278 drivers/ntb/ntb_transport.c 	struct ntb_queue_entry *entry;
entry            2284 drivers/ntb/ntb_transport.c 	entry = ntb_list_rm(&qp->ntb_tx_free_q_lock, &qp->tx_free_q);
entry            2285 drivers/ntb/ntb_transport.c 	if (!entry) {
entry            2290 drivers/ntb/ntb_transport.c 	entry->cb_data = cb;
entry            2291 drivers/ntb/ntb_transport.c 	entry->buf = data;
entry            2292 drivers/ntb/ntb_transport.c 	entry->len = len;
entry            2293 drivers/ntb/ntb_transport.c 	entry->flags = 0;
entry            2294 drivers/ntb/ntb_transport.c 	entry->errors = 0;
entry            2295 drivers/ntb/ntb_transport.c 	entry->retries = 0;
entry            2296 drivers/ntb/ntb_transport.c 	entry->tx_index = 0;
entry            2298 drivers/ntb/ntb_transport.c 	rc = ntb_process_tx(qp, entry);
entry            2300 drivers/ntb/ntb_transport.c 		ntb_list_add(&qp->ntb_tx_free_q_lock, &entry->entry,
entry            2421 drivers/ntb/ntb_transport.c 	unsigned int tail = qp->remote_rx_info->entry;
entry             441 drivers/nvme/host/core.c 	list_del_init(&head->entry);
entry            2502 drivers/nvme/host/core.c 	list_del(&subsys->entry);
entry            2532 drivers/nvme/host/core.c 	list_for_each_entry(subsys, &nvme_subsystems, entry) {
entry            2674 drivers/nvme/host/core.c 		list_add_tail(&subsys->entry, &nvme_subsystems);
entry            3301 drivers/nvme/host/core.c 	list_for_each_entry(h, &subsys->nsheads, entry) {
entry            3316 drivers/nvme/host/core.c 	list_for_each_entry(h, &subsys->nsheads, entry) {
entry            3367 drivers/nvme/host/core.c 	list_add_tail(&head->entry, &ctrl->subsys->nsheads);
entry             499 drivers/nvme/host/fabrics.c 	list_add_tail(&ops->entry, &nvmf_transports);
entry             518 drivers/nvme/host/fabrics.c 	list_del(&ops->entry);
entry             530 drivers/nvme/host/fabrics.c 	list_for_each_entry(ops, &nvmf_transports, entry) {
entry             142 drivers/nvme/host/fabrics.h 	struct list_head	entry;
entry              20 drivers/nvme/host/multipath.c 	list_for_each_entry(h, &subsys->nsheads, entry)
entry              30 drivers/nvme/host/multipath.c 	list_for_each_entry(h, &subsys->nsheads, entry)
entry              40 drivers/nvme/host/multipath.c 	list_for_each_entry(h, &subsys->nsheads, entry)
entry             300 drivers/nvme/host/nvme.h 	struct list_head	entry;
entry             339 drivers/nvme/host/nvme.h 	struct list_head	entry;
entry              41 drivers/nvme/host/rdma.c 	struct list_head	entry;
entry             336 drivers/nvme/host/rdma.c 	list_del(&ndev->entry);
entry             359 drivers/nvme/host/rdma.c 	list_for_each_entry(ndev, &device_list, entry) {
entry             386 drivers/nvme/host/rdma.c 	list_add(&ndev->entry, &device_list);
entry            2089 drivers/nvme/host/rdma.c 	list_for_each_entry(ndev, &device_list, entry) {
entry              38 drivers/nvme/host/tcp.c 	struct list_head	entry;
entry             255 drivers/nvme/host/tcp.c 	list_add_tail(&req->entry, &queue->send_list);
entry             268 drivers/nvme/host/tcp.c 			struct nvme_tcp_request, entry);
entry             270 drivers/nvme/host/tcp.c 		list_del(&req->entry);
entry             638 drivers/nvme/target/configfs.c 	list_for_each_entry(p, &port->subsystems, entry) {
entry             649 drivers/nvme/target/configfs.c 	list_add_tail(&link->entry, &port->subsystems);
entry             669 drivers/nvme/target/configfs.c 	list_for_each_entry(p, &port->subsystems, entry) {
entry             677 drivers/nvme/target/configfs.c 	list_del(&p->entry);
entry             724 drivers/nvme/target/configfs.c 	list_for_each_entry(p, &subsys->hosts, entry) {
entry             728 drivers/nvme/target/configfs.c 	list_add_tail(&link->entry, &subsys->hosts);
entry             747 drivers/nvme/target/configfs.c 	list_for_each_entry(p, &subsys->hosts, entry) {
entry             755 drivers/nvme/target/configfs.c 	list_del(&p->entry);
entry            1001 drivers/nvme/target/configfs.c 	INIT_LIST_HEAD(&port->entry);
entry            1207 drivers/nvme/target/configfs.c 	INIT_LIST_HEAD(&port->entry);
entry             159 drivers/nvme/target/core.c 				struct nvmet_async_event, entry);
entry             168 drivers/nvme/target/core.c 		list_del(&aen->entry);
entry             190 drivers/nvme/target/core.c 	list_add_tail(&aen->entry, &ctrl->async_events);
entry             258 drivers/nvme/target/core.c 	list_for_each_entry(p, &port->subsystems, entry)
entry            1157 drivers/nvme/target/core.c 	list_for_each_entry(p, &subsys->hosts, entry) {
entry            1378 drivers/nvme/target/core.c 	list_for_each_entry(p, &port->subsystems, entry) {
entry              75 drivers/nvme/target/discovery.c 		list_for_each_entry(s, &port->subsystems, entry) {
entry              85 drivers/nvme/target/discovery.c 	if (list_empty(&port->entry)) {
entry              86 drivers/nvme/target/discovery.c 		list_add_tail(&port->entry, &parent->referrals);
entry              96 drivers/nvme/target/discovery.c 	if (!list_empty(&port->entry)) {
entry              98 drivers/nvme/target/discovery.c 		list_del_init(&port->entry);
entry             150 drivers/nvme/target/discovery.c 	list_for_each_entry(p, &req->port->subsystems, entry) {
entry             155 drivers/nvme/target/discovery.c 	list_for_each_entry(r, &req->port->referrals, entry)
entry             195 drivers/nvme/target/discovery.c 	list_for_each_entry(p, &req->port->subsystems, entry) {
entry             208 drivers/nvme/target/discovery.c 	list_for_each_entry(r, &req->port->referrals, entry) {
entry             556 drivers/nvme/target/loop.c 	list_for_each_entry(p, &nvme_loop_ports, entry) {
entry             651 drivers/nvme/target/loop.c 	list_add_tail(&port->entry, &nvme_loop_ports);
entry             659 drivers/nvme/target/loop.c 	list_del_init(&port->entry);
entry             129 drivers/nvme/target/nvmet.h 	struct list_head		entry;
entry             259 drivers/nvme/target/nvmet.h 	struct list_head	entry;
entry             264 drivers/nvme/target/nvmet.h 	struct list_head	entry;
entry             340 drivers/nvme/target/nvmet.h 	struct list_head	entry;
entry             112 drivers/nvme/target/rdma.c 	struct list_head	entry;
entry             905 drivers/nvme/target/rdma.c 	list_del(&ndev->entry);
entry             924 drivers/nvme/target/rdma.c 	list_for_each_entry(ndev, &device_list, entry) {
entry             959 drivers/nvme/target/rdma.c 	list_add(&ndev->entry, &device_list);
entry            1624 drivers/nvme/target/rdma.c 	list_for_each_entry(ndev, &device_list, entry) {
entry              65 drivers/nvme/target/tcp.c 	struct list_head		entry;
entry             186 drivers/nvme/target/tcp.c 				struct nvmet_tcp_cmd, entry);
entry             189 drivers/nvme/target/tcp.c 	list_del_init(&cmd->entry);
entry             204 drivers/nvme/target/tcp.c 	list_add_tail(&cmd->entry, &cmd->queue->free_list);
entry             453 drivers/nvme/target/tcp.c 		list_add(&cmd->entry, &queue->resp_send_list);
entry             462 drivers/nvme/target/tcp.c 				struct nvmet_tcp_cmd, entry);
entry             467 drivers/nvme/target/tcp.c 					struct nvmet_tcp_cmd, entry);
entry             472 drivers/nvme/target/tcp.c 	list_del_init(&queue->snd_cmd->entry);
entry            1241 drivers/nvme/target/tcp.c 	list_add_tail(&c->entry, &queue->free_list);
entry             331 drivers/oprofile/buffer_sync.c static void add_data(struct op_entry *entry, struct mm_struct *mm)
entry             337 drivers/oprofile/buffer_sync.c 	if (!op_cpu_buffer_get_data(entry, &code))
entry             339 drivers/oprofile/buffer_sync.c 	if (!op_cpu_buffer_get_data(entry, &pc))
entry             341 drivers/oprofile/buffer_sync.c 	if (!op_cpu_buffer_get_size(entry))
entry             364 drivers/oprofile/buffer_sync.c 	while (op_cpu_buffer_get_data(entry, &val))
entry             505 drivers/oprofile/buffer_sync.c 	struct op_entry entry;
entry             516 drivers/oprofile/buffer_sync.c 		sample = op_cpu_buffer_read_entry(&entry, cpu);
entry             534 drivers/oprofile/buffer_sync.c 			    && op_cpu_buffer_get_data(&entry, &val)) {
entry             544 drivers/oprofile/buffer_sync.c 			if (op_cpu_buffer_get_size(&entry))
entry             545 drivers/oprofile/buffer_sync.c 				add_data(&entry, mm);
entry             144 drivers/oprofile/cpu_buffer.c *op_cpu_buffer_write_reserve(struct op_entry *entry, unsigned long size)
entry             146 drivers/oprofile/cpu_buffer.c 	entry->event = ring_buffer_lock_reserve
entry             148 drivers/oprofile/cpu_buffer.c 		 size * sizeof(entry->sample->data[0]));
entry             149 drivers/oprofile/cpu_buffer.c 	if (!entry->event)
entry             151 drivers/oprofile/cpu_buffer.c 	entry->sample = ring_buffer_event_data(entry->event);
entry             152 drivers/oprofile/cpu_buffer.c 	entry->size = size;
entry             153 drivers/oprofile/cpu_buffer.c 	entry->data = entry->sample->data;
entry             155 drivers/oprofile/cpu_buffer.c 	return entry->sample;
entry             158 drivers/oprofile/cpu_buffer.c int op_cpu_buffer_write_commit(struct op_entry *entry)
entry             160 drivers/oprofile/cpu_buffer.c 	return ring_buffer_unlock_commit(op_ring_buffer, entry->event);
entry             163 drivers/oprofile/cpu_buffer.c struct op_sample *op_cpu_buffer_read_entry(struct op_entry *entry, int cpu)
entry             170 drivers/oprofile/cpu_buffer.c 	entry->event = e;
entry             171 drivers/oprofile/cpu_buffer.c 	entry->sample = ring_buffer_event_data(e);
entry             172 drivers/oprofile/cpu_buffer.c 	entry->size = (ring_buffer_event_length(e) - sizeof(struct op_sample))
entry             173 drivers/oprofile/cpu_buffer.c 		/ sizeof(entry->sample->data[0]);
entry             174 drivers/oprofile/cpu_buffer.c 	entry->data = entry->sample->data;
entry             175 drivers/oprofile/cpu_buffer.c 	return entry->sample;
entry             187 drivers/oprofile/cpu_buffer.c 	struct op_entry entry;
entry             221 drivers/oprofile/cpu_buffer.c 	sample = op_cpu_buffer_write_reserve(&entry, size);
entry             229 drivers/oprofile/cpu_buffer.c 		op_cpu_buffer_add_data(&entry, (unsigned long)task);
entry             231 drivers/oprofile/cpu_buffer.c 	op_cpu_buffer_write_commit(&entry);
entry             240 drivers/oprofile/cpu_buffer.c 	struct op_entry entry;
entry             243 drivers/oprofile/cpu_buffer.c 	sample = op_cpu_buffer_write_reserve(&entry, 0);
entry             250 drivers/oprofile/cpu_buffer.c 	return op_cpu_buffer_write_commit(&entry);
entry             357 drivers/oprofile/cpu_buffer.c oprofile_write_reserve(struct op_entry *entry, struct pt_regs * const regs,
entry             370 drivers/oprofile/cpu_buffer.c 	sample = op_cpu_buffer_write_reserve(entry, size + 2);
entry             376 drivers/oprofile/cpu_buffer.c 	op_cpu_buffer_add_data(entry, code);
entry             377 drivers/oprofile/cpu_buffer.c 	op_cpu_buffer_add_data(entry, pc);
entry             382 drivers/oprofile/cpu_buffer.c 	entry->event = NULL;
entry             386 drivers/oprofile/cpu_buffer.c int oprofile_add_data(struct op_entry *entry, unsigned long val)
entry             388 drivers/oprofile/cpu_buffer.c 	if (!entry->event)
entry             390 drivers/oprofile/cpu_buffer.c 	return op_cpu_buffer_add_data(entry, val);
entry             393 drivers/oprofile/cpu_buffer.c int oprofile_add_data64(struct op_entry *entry, u64 val)
entry             395 drivers/oprofile/cpu_buffer.c 	if (!entry->event)
entry             397 drivers/oprofile/cpu_buffer.c 	if (op_cpu_buffer_get_size(entry) < 2)
entry             403 drivers/oprofile/cpu_buffer.c 	if (!op_cpu_buffer_add_data(entry, (u32)val))
entry             405 drivers/oprofile/cpu_buffer.c 	return op_cpu_buffer_add_data(entry, (u32)(val >> 32));
entry             408 drivers/oprofile/cpu_buffer.c int oprofile_write_commit(struct op_entry *entry)
entry             410 drivers/oprofile/cpu_buffer.c 	if (!entry->event)
entry             412 drivers/oprofile/cpu_buffer.c 	return op_cpu_buffer_write_commit(entry);
entry              78 drivers/oprofile/cpu_buffer.h *op_cpu_buffer_write_reserve(struct op_entry *entry, unsigned long size);
entry              79 drivers/oprofile/cpu_buffer.h int op_cpu_buffer_write_commit(struct op_entry *entry);
entry              80 drivers/oprofile/cpu_buffer.h struct op_sample *op_cpu_buffer_read_entry(struct op_entry *entry, int cpu);
entry              85 drivers/oprofile/cpu_buffer.h int op_cpu_buffer_add_data(struct op_entry *entry, unsigned long val)
entry              87 drivers/oprofile/cpu_buffer.h 	if (!entry->size)
entry              89 drivers/oprofile/cpu_buffer.h 	*entry->data = val;
entry              90 drivers/oprofile/cpu_buffer.h 	entry->size--;
entry              91 drivers/oprofile/cpu_buffer.h 	entry->data++;
entry              92 drivers/oprofile/cpu_buffer.h 	return entry->size;
entry              97 drivers/oprofile/cpu_buffer.h int op_cpu_buffer_get_size(struct op_entry *entry)
entry              99 drivers/oprofile/cpu_buffer.h 	return entry->size;
entry             104 drivers/oprofile/cpu_buffer.h int op_cpu_buffer_get_data(struct op_entry *entry, unsigned long *val)
entry             106 drivers/oprofile/cpu_buffer.h 	int size = entry->size;
entry             109 drivers/oprofile/cpu_buffer.h 	*val = *entry->data;
entry             110 drivers/oprofile/cpu_buffer.h 	entry->size--;
entry             111 drivers/oprofile/cpu_buffer.h 	entry->data++;
entry              98 drivers/parisc/pdc_stable.c 	ssize_t (*show)(struct pdcspath_entry *entry, char *buf);
entry              99 drivers/parisc/pdc_stable.c 	ssize_t (*store)(struct pdcspath_entry *entry, const char *buf, size_t count);
entry             139 drivers/parisc/pdc_stable.c pdcspath_fetch(struct pdcspath_entry *entry)
entry             143 drivers/parisc/pdc_stable.c 	if (!entry)
entry             146 drivers/parisc/pdc_stable.c 	devpath = &entry->devpath;
entry             149 drivers/parisc/pdc_stable.c 			entry, devpath, entry->addr);
entry             152 drivers/parisc/pdc_stable.c 	if (pdc_stable_read(entry->addr, devpath, sizeof(*devpath)) != PDC_OK)
entry             158 drivers/parisc/pdc_stable.c 	entry->dev = hwpath_to_device((struct hardware_path *)devpath);
entry             160 drivers/parisc/pdc_stable.c 	entry->ready = 1;
entry             162 drivers/parisc/pdc_stable.c 	DPRINTK("%s: device: 0x%p\n", __func__, entry->dev);
entry             180 drivers/parisc/pdc_stable.c pdcspath_store(struct pdcspath_entry *entry)
entry             184 drivers/parisc/pdc_stable.c 	BUG_ON(!entry);
entry             186 drivers/parisc/pdc_stable.c 	devpath = &entry->devpath;
entry             191 drivers/parisc/pdc_stable.c 	if (!entry->ready) {
entry             193 drivers/parisc/pdc_stable.c 		BUG_ON(!entry->dev);
entry             194 drivers/parisc/pdc_stable.c 		device_to_hwpath(entry->dev, (struct hardware_path *)devpath);
entry             199 drivers/parisc/pdc_stable.c 			entry, devpath, entry->addr);
entry             202 drivers/parisc/pdc_stable.c 	if (pdc_stable_write(entry->addr, devpath, sizeof(*devpath)) != PDC_OK)
entry             208 drivers/parisc/pdc_stable.c 	entry->ready = 2;
entry             210 drivers/parisc/pdc_stable.c 	DPRINTK("%s: device: 0x%p\n", __func__, entry->dev);
entry             221 drivers/parisc/pdc_stable.c pdcspath_hwpath_read(struct pdcspath_entry *entry, char *buf)
entry             227 drivers/parisc/pdc_stable.c 	if (!entry || !buf)
entry             230 drivers/parisc/pdc_stable.c 	read_lock(&entry->rw_lock);
entry             231 drivers/parisc/pdc_stable.c 	devpath = &entry->devpath;
entry             232 drivers/parisc/pdc_stable.c 	i = entry->ready;
entry             233 drivers/parisc/pdc_stable.c 	read_unlock(&entry->rw_lock);
entry             264 drivers/parisc/pdc_stable.c pdcspath_hwpath_write(struct pdcspath_entry *entry, const char *buf, size_t count)
entry             272 drivers/parisc/pdc_stable.c 	if (!entry || !buf || !count)
entry             309 drivers/parisc/pdc_stable.c 			"hardware path: %s\n", __func__, entry->name, buf);
entry             314 drivers/parisc/pdc_stable.c 	write_lock(&entry->rw_lock);
entry             315 drivers/parisc/pdc_stable.c 	entry->ready = 0;
entry             316 drivers/parisc/pdc_stable.c 	entry->dev = dev;
entry             319 drivers/parisc/pdc_stable.c 	pdcspath_store(entry);
entry             322 drivers/parisc/pdc_stable.c 	sysfs_remove_link(&entry->kobj, "device");
entry             323 drivers/parisc/pdc_stable.c 	write_unlock(&entry->rw_lock);
entry             325 drivers/parisc/pdc_stable.c 	ret = sysfs_create_link(&entry->kobj, &entry->dev->kobj, "device");
entry             329 drivers/parisc/pdc_stable.c 		entry->name, buf);
entry             342 drivers/parisc/pdc_stable.c pdcspath_layer_read(struct pdcspath_entry *entry, char *buf)
entry             348 drivers/parisc/pdc_stable.c 	if (!entry || !buf)
entry             351 drivers/parisc/pdc_stable.c 	read_lock(&entry->rw_lock);
entry             352 drivers/parisc/pdc_stable.c 	devpath = &entry->devpath;
entry             353 drivers/parisc/pdc_stable.c 	i = entry->ready;
entry             354 drivers/parisc/pdc_stable.c 	read_unlock(&entry->rw_lock);
entry             380 drivers/parisc/pdc_stable.c pdcspath_layer_write(struct pdcspath_entry *entry, const char *buf, size_t count)
entry             386 drivers/parisc/pdc_stable.c 	if (!entry || !buf || !count)
entry             412 drivers/parisc/pdc_stable.c 	write_lock(&entry->rw_lock);
entry             416 drivers/parisc/pdc_stable.c 	memcpy(&entry->devpath.layers, &layers, sizeof(layers));
entry             419 drivers/parisc/pdc_stable.c 	pdcspath_store(entry);
entry             420 drivers/parisc/pdc_stable.c 	write_unlock(&entry->rw_lock);
entry             423 drivers/parisc/pdc_stable.c 		entry->name, buf);
entry             437 drivers/parisc/pdc_stable.c 	struct pdcspath_entry *entry = to_pdcspath_entry(kobj);
entry             442 drivers/parisc/pdc_stable.c 		ret = pdcs_attr->show(entry, buf);
entry             458 drivers/parisc/pdc_stable.c 	struct pdcspath_entry *entry = to_pdcspath_entry(kobj);
entry             466 drivers/parisc/pdc_stable.c 		ret = pdcs_attr->store(entry, buf, count);
entry             964 drivers/parisc/pdc_stable.c 	struct pdcspath_entry *entry;
entry             968 drivers/parisc/pdc_stable.c 	for (i = 0; (entry = pdcspath_entries[i]); i++)
entry             969 drivers/parisc/pdc_stable.c 		rwlock_init(&entry->rw_lock);
entry             971 drivers/parisc/pdc_stable.c 	for (i = 0; (entry = pdcspath_entries[i]); i++) {
entry             972 drivers/parisc/pdc_stable.c 		write_lock(&entry->rw_lock);
entry             973 drivers/parisc/pdc_stable.c 		err = pdcspath_fetch(entry);
entry             974 drivers/parisc/pdc_stable.c 		write_unlock(&entry->rw_lock);
entry             979 drivers/parisc/pdc_stable.c 		entry->kobj.kset = paths_kset;
entry             980 drivers/parisc/pdc_stable.c 		err = kobject_init_and_add(&entry->kobj, &ktype_pdcspath, NULL,
entry             981 drivers/parisc/pdc_stable.c 					   "%s", entry->name);
entry             986 drivers/parisc/pdc_stable.c 		write_lock(&entry->rw_lock);
entry             987 drivers/parisc/pdc_stable.c 		entry->ready = 2;
entry             988 drivers/parisc/pdc_stable.c 		write_unlock(&entry->rw_lock);
entry             991 drivers/parisc/pdc_stable.c 		if (entry->dev) {
entry             992 drivers/parisc/pdc_stable.c 			err = sysfs_create_link(&entry->kobj, &entry->dev->kobj, "device");
entry             996 drivers/parisc/pdc_stable.c 		kobject_uevent(&entry->kobj, KOBJ_ADD);
entry            1009 drivers/parisc/pdc_stable.c 	struct pdcspath_entry *entry;
entry            1011 drivers/parisc/pdc_stable.c 	for (i = 0; (entry = pdcspath_entries[i]); i++) {
entry            1012 drivers/parisc/pdc_stable.c 		read_lock(&entry->rw_lock);
entry            1013 drivers/parisc/pdc_stable.c 		if (entry->ready >= 2)
entry            1014 drivers/parisc/pdc_stable.c 			kobject_put(&entry->kobj);
entry            1015 drivers/parisc/pdc_stable.c 		read_unlock(&entry->rw_lock);
entry              22 drivers/pci/bus.c 	struct resource_entry *entry;
entry              24 drivers/pci/bus.c 	entry = resource_list_create_entry(res, 0);
entry              25 drivers/pci/bus.c 	if (!entry) {
entry              30 drivers/pci/bus.c 	entry->offset = offset;
entry              31 drivers/pci/bus.c 	resource_list_add_tail(entry, resources);
entry             199 drivers/pci/controller/pci-xgene.c 	struct resource_entry *entry;
entry             220 drivers/pci/controller/pci-xgene.c 	entry = list_first_entry(&list, struct resource_entry, node);
entry             221 drivers/pci/controller/pci-xgene.c 	*res = *entry->res;
entry            1190 drivers/pci/controller/pcie-iproc.c 	struct resource_entry *entry, *tmp;
entry            1205 drivers/pci/controller/pcie-iproc.c 	entry = resource_list_create_entry(res, 0);
entry            1206 drivers/pci/controller/pcie-iproc.c 	if (!entry)
entry            1209 drivers/pci/controller/pcie-iproc.c 	entry->offset = res->start - range->cpu_addr;
entry            1210 drivers/pci/controller/pcie-iproc.c 	resource_list_add(entry, head);
entry             713 drivers/pci/hotplug/pnv_php.c 	struct msix_entry entry;
entry             724 drivers/pci/hotplug/pnv_php.c 	entry.entry = (pcie_flag & PCI_EXP_FLAGS_IRQ) >> 9;
entry             725 drivers/pci/hotplug/pnv_php.c 	if (entry.entry >= nr_entries)
entry             729 drivers/pci/hotplug/pnv_php.c 	ret = pci_enable_msix_exact(pdev, &entry, 1);
entry             735 drivers/pci/hotplug/pnv_php.c 	return entry.vector;
entry              93 drivers/pci/msi.c 	struct msi_desc *entry;
entry             105 drivers/pci/msi.c 	for_each_pci_msi_entry(entry, dev) {
entry             106 drivers/pci/msi.c 		ret = arch_setup_msi_irq(dev, entry);
entry             123 drivers/pci/msi.c 	struct msi_desc *entry;
entry             125 drivers/pci/msi.c 	for_each_pci_msi_entry(entry, dev)
entry             126 drivers/pci/msi.c 		if (entry->irq)
entry             127 drivers/pci/msi.c 			for (i = 0; i < entry->nvec_used; i++)
entry             128 drivers/pci/msi.c 				arch_teardown_msi_irq(entry->irq + i);
entry             138 drivers/pci/msi.c 	struct msi_desc *entry;
entry             140 drivers/pci/msi.c 	entry = NULL;
entry             142 drivers/pci/msi.c 		for_each_pci_msi_entry(entry, dev) {
entry             143 drivers/pci/msi.c 			if (irq == entry->irq)
entry             147 drivers/pci/msi.c 		entry = irq_get_msi_desc(irq);
entry             150 drivers/pci/msi.c 	if (entry)
entry             151 drivers/pci/msi.c 		__pci_write_msi_msg(entry, &entry->msg);
entry             270 drivers/pci/msi.c 	struct msi_desc *entry;
entry             272 drivers/pci/msi.c 	for_each_pci_msi_entry(entry, dev)
entry             273 drivers/pci/msi.c 		default_restore_msi_irq(dev, entry->irq);
entry             276 drivers/pci/msi.c void __pci_read_msi_msg(struct msi_desc *entry, struct msi_msg *msg)
entry             278 drivers/pci/msi.c 	struct pci_dev *dev = msi_desc_to_pci_dev(entry);
entry             282 drivers/pci/msi.c 	if (entry->msi_attrib.is_msix) {
entry             283 drivers/pci/msi.c 		void __iomem *base = pci_msix_desc_addr(entry);
entry             299 drivers/pci/msi.c 		if (entry->msi_attrib.is_64) {
entry             311 drivers/pci/msi.c void __pci_write_msi_msg(struct msi_desc *entry, struct msi_msg *msg)
entry             313 drivers/pci/msi.c 	struct pci_dev *dev = msi_desc_to_pci_dev(entry);
entry             317 drivers/pci/msi.c 	} else if (entry->msi_attrib.is_msix) {
entry             318 drivers/pci/msi.c 		void __iomem *base = pci_msix_desc_addr(entry);
entry             332 drivers/pci/msi.c 		msgctl |= entry->msi_attrib.multiple << 4;
entry             337 drivers/pci/msi.c 		if (entry->msi_attrib.is_64) {
entry             349 drivers/pci/msi.c 	entry->msg = *msg;
entry             351 drivers/pci/msi.c 	if (entry->write_msi_msg)
entry             352 drivers/pci/msi.c 		entry->write_msi_msg(entry, entry->write_msi_msg_data);
entry             358 drivers/pci/msi.c 	struct msi_desc *entry = irq_get_msi_desc(irq);
entry             360 drivers/pci/msi.c 	__pci_write_msi_msg(entry, msg);
entry             367 drivers/pci/msi.c 	struct msi_desc *entry, *tmp;
entry             372 drivers/pci/msi.c 	for_each_pci_msi_entry(entry, dev)
entry             373 drivers/pci/msi.c 		if (entry->irq)
entry             374 drivers/pci/msi.c 			for (i = 0; i < entry->nvec_used; i++)
entry             375 drivers/pci/msi.c 				BUG_ON(irq_has_action(entry->irq + i));
entry             379 drivers/pci/msi.c 	list_for_each_entry_safe(entry, tmp, msi_list, list) {
entry             380 drivers/pci/msi.c 		if (entry->msi_attrib.is_msix) {
entry             381 drivers/pci/msi.c 			if (list_is_last(&entry->list, msi_list))
entry             382 drivers/pci/msi.c 				iounmap(entry->mask_base);
entry             385 drivers/pci/msi.c 		list_del(&entry->list);
entry             386 drivers/pci/msi.c 		free_msi_entry(entry);
entry             415 drivers/pci/msi.c 	struct msi_desc *entry;
entry             420 drivers/pci/msi.c 	entry = irq_get_msi_desc(dev->irq);
entry             427 drivers/pci/msi.c 	msi_mask_irq(entry, msi_mask(entry->msi_attrib.multi_cap),
entry             428 drivers/pci/msi.c 		     entry->masked);
entry             430 drivers/pci/msi.c 	control |= (entry->msi_attrib.multiple << 4) | PCI_MSI_FLAGS_ENABLE;
entry             436 drivers/pci/msi.c 	struct msi_desc *entry;
entry             448 drivers/pci/msi.c 	for_each_pci_msi_entry(entry, dev)
entry             449 drivers/pci/msi.c 		msix_mask_irq(entry, entry->masked);
entry             464 drivers/pci/msi.c 	struct msi_desc *entry;
entry             472 drivers/pci/msi.c 	entry = irq_get_msi_desc(irq);
entry             473 drivers/pci/msi.c 	if (entry)
entry             475 drivers/pci/msi.c 				entry->msi_attrib.is_msix ? "msix" : "msi");
entry             487 drivers/pci/msi.c 	struct msi_desc *entry;
entry             494 drivers/pci/msi.c 	for_each_pci_msi_entry(entry, pdev)
entry             495 drivers/pci/msi.c 		num_msi += entry->nvec_used;
entry             503 drivers/pci/msi.c 	for_each_pci_msi_entry(entry, pdev) {
entry             504 drivers/pci/msi.c 		for (i = 0; i < entry->nvec_used; i++) {
entry             512 drivers/pci/msi.c 							    entry->irq + i);
entry             561 drivers/pci/msi.c 	struct msi_desc *entry;
entry             568 drivers/pci/msi.c 	entry = alloc_msi_entry(&dev->dev, nvec, masks);
entry             569 drivers/pci/msi.c 	if (!entry)
entry             574 drivers/pci/msi.c 	entry->msi_attrib.is_msix	= 0;
entry             575 drivers/pci/msi.c 	entry->msi_attrib.is_64		= !!(control & PCI_MSI_FLAGS_64BIT);
entry             576 drivers/pci/msi.c 	entry->msi_attrib.is_virtual    = 0;
entry             577 drivers/pci/msi.c 	entry->msi_attrib.entry_nr	= 0;
entry             578 drivers/pci/msi.c 	entry->msi_attrib.maskbit	= !!(control & PCI_MSI_FLAGS_MASKBIT);
entry             579 drivers/pci/msi.c 	entry->msi_attrib.default_irq	= dev->irq;	/* Save IOAPIC IRQ */
entry             580 drivers/pci/msi.c 	entry->msi_attrib.multi_cap	= (control & PCI_MSI_FLAGS_QMASK) >> 1;
entry             581 drivers/pci/msi.c 	entry->msi_attrib.multiple	= ilog2(__roundup_pow_of_two(nvec));
entry             584 drivers/pci/msi.c 		entry->mask_pos = dev->msi_cap + PCI_MSI_MASK_64;
entry             586 drivers/pci/msi.c 		entry->mask_pos = dev->msi_cap + PCI_MSI_MASK_32;
entry             589 drivers/pci/msi.c 	if (entry->msi_attrib.maskbit)
entry             590 drivers/pci/msi.c 		pci_read_config_dword(dev, entry->mask_pos, &entry->masked);
entry             594 drivers/pci/msi.c 	return entry;
entry             599 drivers/pci/msi.c 	struct msi_desc *entry;
entry             601 drivers/pci/msi.c 	for_each_pci_msi_entry(entry, dev) {
entry             602 drivers/pci/msi.c 		if (!dev->no_64bit_msi || !entry->msg.address_hi)
entry             626 drivers/pci/msi.c 	struct msi_desc *entry;
entry             632 drivers/pci/msi.c 	entry = msi_setup_entry(dev, nvec, affd);
entry             633 drivers/pci/msi.c 	if (!entry)
entry             637 drivers/pci/msi.c 	mask = msi_mask(entry->msi_attrib.multi_cap);
entry             638 drivers/pci/msi.c 	msi_mask_irq(entry, mask, mask);
entry             640 drivers/pci/msi.c 	list_add_tail(&entry->list, dev_to_msi_list(&dev->dev));
entry             645 drivers/pci/msi.c 		msi_mask_irq(entry, mask, ~mask);
entry             652 drivers/pci/msi.c 		msi_mask_irq(entry, mask, ~mask);
entry             659 drivers/pci/msi.c 		msi_mask_irq(entry, mask, ~mask);
entry             670 drivers/pci/msi.c 	dev->irq = entry->irq;
entry             699 drivers/pci/msi.c 	struct msi_desc *entry;
entry             707 drivers/pci/msi.c 		entry = alloc_msi_entry(&dev->dev, 1, curmsk);
entry             708 drivers/pci/msi.c 		if (!entry) {
entry             718 drivers/pci/msi.c 		entry->msi_attrib.is_msix	= 1;
entry             719 drivers/pci/msi.c 		entry->msi_attrib.is_64		= 1;
entry             721 drivers/pci/msi.c 			entry->msi_attrib.entry_nr = entries[i].entry;
entry             723 drivers/pci/msi.c 			entry->msi_attrib.entry_nr = i;
entry             725 drivers/pci/msi.c 		entry->msi_attrib.is_virtual =
entry             726 drivers/pci/msi.c 			entry->msi_attrib.entry_nr >= vec_count;
entry             728 drivers/pci/msi.c 		entry->msi_attrib.default_irq	= dev->irq;
entry             729 drivers/pci/msi.c 		entry->mask_base		= base;
entry             731 drivers/pci/msi.c 		list_add_tail(&entry->list, dev_to_msi_list(&dev->dev));
entry             744 drivers/pci/msi.c 	struct msi_desc *entry;
entry             748 drivers/pci/msi.c 	for_each_pci_msi_entry(entry, dev) {
entry             750 drivers/pci/msi.c 			entries[i++].vector = entry->irq;
entry             752 drivers/pci/msi.c 		desc_addr = pci_msix_desc_addr(entry);
entry             754 drivers/pci/msi.c 			entry->masked = readl(desc_addr +
entry             757 drivers/pci/msi.c 			entry->masked = 0;
entry             759 drivers/pci/msi.c 		msix_mask_irq(entry, 1);
entry             831 drivers/pci/msi.c 		struct msi_desc *entry;
entry             834 drivers/pci/msi.c 		for_each_pci_msi_entry(entry, dev) {
entry             835 drivers/pci/msi.c 			if (entry->irq != 0)
entry             988 drivers/pci/msi.c 			if (entries[i].entry >= nr_entries)
entry             991 drivers/pci/msi.c 				if (entries[i].entry == entries[j].entry)
entry            1007 drivers/pci/msi.c 	struct msi_desc *entry;
entry            1018 drivers/pci/msi.c 	for_each_pci_msi_entry(entry, dev) {
entry            1020 drivers/pci/msi.c 		__pci_msix_desc_mask_irq(entry, 1);
entry            1261 drivers/pci/msi.c 		struct msi_desc *entry;
entry            1264 drivers/pci/msi.c 		for_each_pci_msi_entry(entry, dev) {
entry            1266 drivers/pci/msi.c 				return entry->irq;
entry            1274 drivers/pci/msi.c 		struct msi_desc *entry = first_pci_msi_entry(dev);
entry            1276 drivers/pci/msi.c 		if (WARN_ON_ONCE(nr >= entry->nvec_used))
entry            1295 drivers/pci/msi.c 		struct msi_desc *entry;
entry            1298 drivers/pci/msi.c 		for_each_pci_msi_entry(entry, dev) {
entry            1300 drivers/pci/msi.c 				return &entry->affinity->mask;
entry            1306 drivers/pci/msi.c 		struct msi_desc *entry = first_pci_msi_entry(dev);
entry            1308 drivers/pci/msi.c 		if (WARN_ON_ONCE(!entry || !entry->affinity ||
entry            1309 drivers/pci/msi.c 				 nr >= entry->nvec_used))
entry            1312 drivers/pci/msi.c 		return &entry->affinity[nr].mask;
entry             299 drivers/pci/p2pdma.c 	const struct pci_p2pdma_whitelist_entry *entry;
entry             309 drivers/pci/p2pdma.c 	for (entry = pci_p2pdma_whitelist; entry->vendor; entry++) {
entry             310 drivers/pci/p2pdma.c 		if (vendor != entry->vendor || device != entry->device)
entry             312 drivers/pci/p2pdma.c 		if (entry->flags & REQ_SAME_HOST_BRIDGE && !same_host_bridge)
entry              34 drivers/pci/pci-acpi.c 	struct resource_entry *entry;
entry              55 drivers/pci/pci-acpi.c 	entry = list_first_entry(&list, struct resource_entry, node);
entry              56 drivers/pci/pci-acpi.c 	*res = *entry->res;
entry            1025 drivers/pci/pcie/aer.c 	struct aer_recover_entry entry;
entry            1028 drivers/pci/pcie/aer.c 	while (kfifo_get(&aer_recover_ring, &entry)) {
entry            1029 drivers/pci/pcie/aer.c 		pdev = pci_get_domain_bus_and_slot(entry.domain, entry.bus,
entry            1030 drivers/pci/pcie/aer.c 						   entry.devfn);
entry            1033 drivers/pci/pcie/aer.c 			       entry.domain, entry.bus,
entry            1034 drivers/pci/pcie/aer.c 			       PCI_SLOT(entry.devfn), PCI_FUNC(entry.devfn));
entry            1037 drivers/pci/pcie/aer.c 		cper_print_aer(pdev, entry.severity, entry.regs);
entry            1038 drivers/pci/pcie/aer.c 		if (entry.severity == AER_NONFATAL)
entry            1041 drivers/pci/pcie/aer.c 		else if (entry.severity == AER_FATAL)
entry            1059 drivers/pci/pcie/aer.c 	struct aer_recover_entry entry = {
entry            1067 drivers/pci/pcie/aer.c 	if (kfifo_in_spinlocked(&aer_recover_ring, &entry, 1,
entry             262 drivers/pci/xen-pcifront.c 	struct msi_desc *entry;
entry             271 drivers/pci/xen-pcifront.c 	for_each_pci_msi_entry(entry, dev) {
entry             272 drivers/pci/xen-pcifront.c 		op.msix_entries[i].entry = entry->msi_attrib.entry_nr;
entry            1078 drivers/pcmcia/cistpl.c 			       cistpl_cftable_entry_t *entry)
entry            1084 drivers/pcmcia/cistpl.c 	entry->index = *p & 0x3f;
entry            1085 drivers/pcmcia/cistpl.c 	entry->flags = 0;
entry            1087 drivers/pcmcia/cistpl.c 		entry->flags |= CISTPL_CFTABLE_DEFAULT;
entry            1092 drivers/pcmcia/cistpl.c 			entry->flags |= CISTPL_CFTABLE_BVDS;
entry            1094 drivers/pcmcia/cistpl.c 			entry->flags |= CISTPL_CFTABLE_WP;
entry            1096 drivers/pcmcia/cistpl.c 			entry->flags |= CISTPL_CFTABLE_RDYBSY;
entry            1098 drivers/pcmcia/cistpl.c 			entry->flags |= CISTPL_CFTABLE_MWAIT;
entry            1099 drivers/pcmcia/cistpl.c 		entry->interface = *p & 0x0f;
entry            1101 drivers/pcmcia/cistpl.c 		entry->interface = 0;
entry            1110 drivers/pcmcia/cistpl.c 		p = parse_power(p, q, &entry->vcc);
entry            1114 drivers/pcmcia/cistpl.c 		entry->vcc.present = 0;
entry            1116 drivers/pcmcia/cistpl.c 		p = parse_power(p, q, &entry->vpp1);
entry            1120 drivers/pcmcia/cistpl.c 		entry->vpp1.present = 0;
entry            1122 drivers/pcmcia/cistpl.c 		p = parse_power(p, q, &entry->vpp2);
entry            1126 drivers/pcmcia/cistpl.c 		entry->vpp2.present = 0;
entry            1130 drivers/pcmcia/cistpl.c 		p = parse_timing(p, q, &entry->timing);
entry            1134 drivers/pcmcia/cistpl.c 		entry->timing.wait = 0;
entry            1135 drivers/pcmcia/cistpl.c 		entry->timing.ready = 0;
entry            1136 drivers/pcmcia/cistpl.c 		entry->timing.reserved = 0;
entry            1141 drivers/pcmcia/cistpl.c 		p = parse_io(p, q, &entry->io);
entry            1145 drivers/pcmcia/cistpl.c 		entry->io.nwin = 0;
entry            1149 drivers/pcmcia/cistpl.c 		p = parse_irq(p, q, &entry->irq);
entry            1153 drivers/pcmcia/cistpl.c 		entry->irq.IRQInfo1 = 0;
entry            1157 drivers/pcmcia/cistpl.c 		entry->mem.nwin = 0;
entry            1160 drivers/pcmcia/cistpl.c 		entry->mem.nwin = 1;
entry            1161 drivers/pcmcia/cistpl.c 		entry->mem.win[0].len = get_unaligned_le16(p) << 8;
entry            1162 drivers/pcmcia/cistpl.c 		entry->mem.win[0].card_addr = 0;
entry            1163 drivers/pcmcia/cistpl.c 		entry->mem.win[0].host_addr = 0;
entry            1169 drivers/pcmcia/cistpl.c 		entry->mem.nwin = 1;
entry            1170 drivers/pcmcia/cistpl.c 		entry->mem.win[0].len = get_unaligned_le16(p) << 8;
entry            1171 drivers/pcmcia/cistpl.c 		entry->mem.win[0].card_addr = get_unaligned_le16(p + 2) << 8;
entry            1172 drivers/pcmcia/cistpl.c 		entry->mem.win[0].host_addr = 0;
entry            1178 drivers/pcmcia/cistpl.c 		p = parse_mem(p, q, &entry->mem);
entry            1188 drivers/pcmcia/cistpl.c 		entry->flags |= (*p << 8);
entry            1195 drivers/pcmcia/cistpl.c 	entry->subtuples = q-p;
entry              66 drivers/perf/thunderx2_pmu.c 	struct list_head  entry;
entry             603 drivers/perf/thunderx2_pmu.c 	list_add(&tx2_pmu->entry, &tx2_pmus);
entry             652 drivers/perf/thunderx2_pmu.c 	INIT_LIST_HEAD(&tx2_pmu->entry);
entry             802 drivers/perf/thunderx2_pmu.c 		list_for_each_entry_safe(tx2_pmu, temp, &tx2_pmus, entry) {
entry             808 drivers/perf/thunderx2_pmu.c 				list_del(&tx2_pmu->entry);
entry             538 drivers/perf/xgene_pmu.c 	XGENE_PMU_EVENT_ATTR(pd-entry-vld,			0x07),
entry             539 drivers/perf/xgene_pmu.c 	XGENE_PMU_EVENT_ATTR(sref-entry-vld,			0x08),
entry             507 drivers/pinctrl/pinctrl-rza1.c 	const struct rza1_bidir_entry *entry = &table[port];
entry             511 drivers/pinctrl/pinctrl-rza1.c 	for (i = 0; i < entry->npins; ++i) {
entry             512 drivers/pinctrl/pinctrl-rza1.c 		bidir_pin = &entry->pins[i];
entry             564 drivers/platform/x86/dell-wmi.c 		const struct key_entry *entry = &dell_wmi_keymap_type_0010[i];
entry             572 drivers/platform/x86/dell-wmi.c 		    have_scancode(entry->code | (0x0010 << 16),
entry             577 drivers/platform/x86/dell-wmi.c 		keymap[pos] = *entry;
entry            9934 drivers/platform/x86/thinkpad_acpi.c 	struct proc_dir_entry *entry;
entry            9986 drivers/platform/x86/thinkpad_acpi.c 		entry = proc_create_data(ibm->name, mode, proc_dir,
entry            9988 drivers/platform/x86/thinkpad_acpi.c 		if (!entry) {
entry            2476 drivers/power/supply/ab8500_fg.c 	struct ab8500_fg_sysfs_entry *entry;
entry            2479 drivers/power/supply/ab8500_fg.c 	entry = container_of(attr, struct ab8500_fg_sysfs_entry, attr);
entry            2482 drivers/power/supply/ab8500_fg.c 	if (!entry->show)
entry            2485 drivers/power/supply/ab8500_fg.c 	return entry->show(di, buf);
entry            2491 drivers/power/supply/ab8500_fg.c 	struct ab8500_fg_sysfs_entry *entry;
entry            2494 drivers/power/supply/ab8500_fg.c 	entry = container_of(attr, struct ab8500_fg_sysfs_entry, attr);
entry            2497 drivers/power/supply/ab8500_fg.c 	if (!entry->store)
entry            2500 drivers/power/supply/ab8500_fg.c 	return entry->store(di, buf, count);
entry            1841 drivers/power/supply/abx500_chargalg.c 	struct abx500_chargalg_sysfs_entry *entry = container_of(attr,
entry            1847 drivers/power/supply/abx500_chargalg.c 	if (!entry->show)
entry            1850 drivers/power/supply/abx500_chargalg.c 	return entry->show(di, buf);
entry            1856 drivers/power/supply/abx500_chargalg.c 	struct abx500_chargalg_sysfs_entry *entry = container_of(attr,
entry            1862 drivers/power/supply/abx500_chargalg.c 	if (!entry->store)
entry            1865 drivers/power/supply/abx500_chargalg.c 	return entry->store(di, buf, length);
entry             745 drivers/power/supply/charger-manager.c 	list_for_each_entry(cm, &cm_list, entry) {
entry             768 drivers/power/supply/charger-manager.c 	list_for_each_entry(cm, &cm_list, entry) {
entry            1071 drivers/power/supply/charger-manager.c 	list_for_each_entry(cm, &cm_list, entry) {
entry            1790 drivers/power/supply/charger-manager.c 	list_add(&cm->entry, &cm_list);
entry            1840 drivers/power/supply/charger-manager.c 	list_del(&cm->entry);
entry            1889 drivers/power/supply/charger-manager.c 	list_for_each_entry(cm, &cm_list, entry) {
entry            2022 drivers/power/supply/charger-manager.c 	list_for_each_entry(cm, &cm_list, entry) {
entry             727 drivers/rapidio/devices/tsi721.c 	entries[TSI721_VECT_IDB].entry = TSI721_MSIX_SR2PC_IDBQ_RCV(IDB_QUEUE);
entry             728 drivers/rapidio/devices/tsi721.c 	entries[TSI721_VECT_PWRX].entry = TSI721_MSIX_SRIO_MAC_INT;
entry             737 drivers/rapidio/devices/tsi721.c 		entries[TSI721_VECT_IMB0_RCV + i].entry =
entry             739 drivers/rapidio/devices/tsi721.c 		entries[TSI721_VECT_IMB0_INT + i].entry =
entry             741 drivers/rapidio/devices/tsi721.c 		entries[TSI721_VECT_OMB0_DONE + i].entry =
entry             743 drivers/rapidio/devices/tsi721.c 		entries[TSI721_VECT_OMB0_INT + i].entry =
entry             754 drivers/rapidio/devices/tsi721.c 		entries[TSI721_VECT_DMA0_DONE + i].entry =
entry             756 drivers/rapidio/devices/tsi721.c 		entries[TSI721_VECT_DMA0_INT + i].entry =
entry              41 drivers/rapidio/switches/idt_gen3.c 	u32 entry = route_port;
entry              45 drivers/rapidio/switches/idt_gen3.c 		 __func__, table, route_destid, entry);
entry              51 drivers/rapidio/switches/idt_gen3.c 		entry = RIO_RT_ENTRY_DROP_PKT;
entry              57 drivers/rapidio/switches/idt_gen3.c 				entry);
entry              74 drivers/rapidio/switches/idt_gen3.c 			entry);
entry            1174 drivers/remoteproc/remoteproc_core.c 	struct rproc_mem_entry *entry, *tmp;
entry            1180 drivers/remoteproc/remoteproc_core.c 	list_for_each_entry_safe(entry, tmp, &rproc->carveouts, node) {
entry            1181 drivers/remoteproc/remoteproc_core.c 		if (entry->alloc) {
entry            1182 drivers/remoteproc/remoteproc_core.c 			ret = entry->alloc(rproc, entry);
entry            1185 drivers/remoteproc/remoteproc_core.c 					entry->name, ret);
entry            1190 drivers/remoteproc/remoteproc_core.c 		if (entry->rsc_offset != FW_RSC_ADDR_ANY) {
entry            1192 drivers/remoteproc/remoteproc_core.c 			rsc = (void *)rproc->table_ptr + entry->rsc_offset;
entry            1213 drivers/remoteproc/remoteproc_core.c 			if (entry->va)
entry            1214 drivers/remoteproc/remoteproc_core.c 				pa = (u64)rproc_va_to_pa(entry->va);
entry            1216 drivers/remoteproc/remoteproc_core.c 				pa = (u64)entry->dma;
entry            1223 drivers/remoteproc/remoteproc_core.c 			rsc->da = entry->da;
entry            1224 drivers/remoteproc/remoteproc_core.c 			rsc->len = entry->len;
entry            1237 drivers/remoteproc/remoteproc_core.c 	struct rproc_dump_segment *entry, *tmp;
entry            1239 drivers/remoteproc/remoteproc_core.c 	list_for_each_entry_safe(entry, tmp, &rproc->dump_segments, node) {
entry            1240 drivers/remoteproc/remoteproc_core.c 		list_del(&entry->node);
entry            1241 drivers/remoteproc/remoteproc_core.c 		kfree(entry);
entry            1254 drivers/remoteproc/remoteproc_core.c 	struct rproc_mem_entry *entry, *tmp;
entry            1268 drivers/remoteproc/remoteproc_core.c 	list_for_each_entry_safe(entry, tmp, &rproc->mappings, node) {
entry            1271 drivers/remoteproc/remoteproc_core.c 		unmapped = iommu_unmap(rproc->domain, entry->da, entry->len);
entry            1272 drivers/remoteproc/remoteproc_core.c 		if (unmapped != entry->len) {
entry            1274 drivers/remoteproc/remoteproc_core.c 			dev_err(dev, "failed to unmap %u/%zu\n", entry->len,
entry            1278 drivers/remoteproc/remoteproc_core.c 		list_del(&entry->node);
entry            1279 drivers/remoteproc/remoteproc_core.c 		kfree(entry);
entry            1283 drivers/remoteproc/remoteproc_core.c 	list_for_each_entry_safe(entry, tmp, &rproc->carveouts, node) {
entry            1284 drivers/remoteproc/remoteproc_core.c 		if (entry->release)
entry            1285 drivers/remoteproc/remoteproc_core.c 			entry->release(rproc, entry);
entry            1286 drivers/remoteproc/remoteproc_core.c 		list_del(&entry->node);
entry            1287 drivers/remoteproc/remoteproc_core.c 		kfree(entry);
entry             170 drivers/reset/core.c 	struct reset_control_lookup *entry;
entry             175 drivers/reset/core.c 		entry = &lookup[i];
entry             177 drivers/reset/core.c 		if (!entry->dev_id || !entry->provider) {
entry             183 drivers/reset/core.c 		list_add_tail(&entry->list, &reset_lookup_list);
entry            1193 drivers/rpmsg/qcom_smd.c 	struct qcom_smd_alloc_entry *entry;
entry            1209 drivers/rpmsg/qcom_smd.c 			entry = &alloc_tbl[i];
entry            1210 drivers/rpmsg/qcom_smd.c 			eflags = le32_to_cpu(entry->flags);
entry            1214 drivers/rpmsg/qcom_smd.c 			if (entry->ref_count == 0)
entry            1217 drivers/rpmsg/qcom_smd.c 			if (!entry->name[0])
entry            1226 drivers/rpmsg/qcom_smd.c 			cid = le32_to_cpu(entry->cid);
entry            1230 drivers/rpmsg/qcom_smd.c 			channel = qcom_smd_create_channel(edge, info_id, fifo_id, entry->name);
entry            6031 drivers/s390/block/dasd_eckd.c 	struct dasd_ckd_path_group_entry *entry;
entry            6051 drivers/s390/block/dasd_eckd.c 		entry = (struct dasd_ckd_path_group_entry *)
entry            6052 drivers/s390/block/dasd_eckd.c 			(info->entry + i * info->entry_size);
entry            6053 drivers/s390/block/dasd_eckd.c 		if (entry->status_flags & DASD_ECKD_PG_GROUPED)
entry            6067 drivers/s390/block/dasd_eckd.c 	struct dasd_ckd_path_group_entry *entry;
entry            6087 drivers/s390/block/dasd_eckd.c 		entry = (struct dasd_ckd_path_group_entry *)
entry            6088 drivers/s390/block/dasd_eckd.c 			(info->entry + i * info->entry_size);
entry            6090 drivers/s390/block/dasd_eckd.c 		seq_printf(m, "pgid %*phN\n", 11, entry->pgid);
entry            6092 drivers/s390/block/dasd_eckd.c 		seq_printf(m, "status_flags %02x\n", entry->status_flags);
entry            6094 drivers/s390/block/dasd_eckd.c 		memcpy(&sysplex, &entry->sysplex_name, sizeof(sysplex) - 1);
entry            6098 drivers/s390/block/dasd_eckd.c 		seq_printf(m, "supported_cylinder %d\n", entry->cylinder);
entry            6101 drivers/s390/block/dasd_eckd.c 			   entry->timestamp);
entry             512 drivers/s390/block/dasd_eckd.h 	__u8 entry[16390];
entry             112 drivers/s390/block/dcssblk.c 	struct segment_info *entry, *temp;
entry             115 drivers/s390/block/dcssblk.c 	list_for_each_entry_safe(entry, temp, &dev_info->seg_list, lh) {
entry             116 drivers/s390/block/dcssblk.c 		list_del(&entry->lh);
entry             117 drivers/s390/block/dcssblk.c 		kfree(entry);
entry             133 drivers/s390/block/dcssblk.c 	struct dcssblk_dev_info *entry;
entry             140 drivers/s390/block/dcssblk.c 		list_for_each_entry(entry, &dcssblk_devices, lh)
entry             141 drivers/s390/block/dcssblk.c 			if (minor == entry->gd->first_minor)
entry             159 drivers/s390/block/dcssblk.c 	struct dcssblk_dev_info *entry;
entry             161 drivers/s390/block/dcssblk.c 	list_for_each_entry(entry, &dcssblk_devices, lh) {
entry             162 drivers/s390/block/dcssblk.c 		if (!strcmp(name, entry->segment_name)) {
entry             163 drivers/s390/block/dcssblk.c 			return entry;
entry             178 drivers/s390/block/dcssblk.c 	struct segment_info *entry;
entry             181 drivers/s390/block/dcssblk.c 		list_for_each_entry(entry, &dev_info->seg_list, lh) {
entry             182 drivers/s390/block/dcssblk.c 			if (!strcmp(name, entry->segment_name))
entry             183 drivers/s390/block/dcssblk.c 				return entry;
entry             196 drivers/s390/block/dcssblk.c 	struct segment_info *entry;
entry             199 drivers/s390/block/dcssblk.c 	list_for_each_entry(entry, &dev_info->seg_list, lh) {
entry             200 drivers/s390/block/dcssblk.c 		if (highest_addr < entry->end)
entry             201 drivers/s390/block/dcssblk.c 			highest_addr = entry->end;
entry             214 drivers/s390/block/dcssblk.c 	struct segment_info *entry;
entry             218 drivers/s390/block/dcssblk.c 	list_for_each_entry(entry, &dev_info->seg_list, lh) {
entry             220 drivers/s390/block/dcssblk.c 			lowest_addr = entry->start;
entry             223 drivers/s390/block/dcssblk.c 			if (lowest_addr > entry->start)
entry             224 drivers/s390/block/dcssblk.c 				lowest_addr = entry->start;
entry             237 drivers/s390/block/dcssblk.c 	struct segment_info *sort_list, *entry, temp;
entry             248 drivers/s390/block/dcssblk.c 	list_for_each_entry(entry, &dev_info->seg_list, lh) {
entry             249 drivers/s390/block/dcssblk.c 		memcpy(&sort_list[i], entry, sizeof(struct segment_info));
entry             348 drivers/s390/block/dcssblk.c 	struct segment_info *entry, *temp;
entry             361 drivers/s390/block/dcssblk.c 		list_for_each_entry(entry, &dev_info->seg_list, lh) {
entry             362 drivers/s390/block/dcssblk.c 			rc = segment_modify_shared(entry->segment_name,
entry             386 drivers/s390/block/dcssblk.c 		list_for_each_entry(entry, &dev_info->seg_list, lh) {
entry             387 drivers/s390/block/dcssblk.c 			rc = segment_modify_shared(entry->segment_name,
entry             407 drivers/s390/block/dcssblk.c 	temp = entry;
entry             408 drivers/s390/block/dcssblk.c 	list_for_each_entry(entry, &dev_info->seg_list, lh) {
entry             409 drivers/s390/block/dcssblk.c 		if (entry != temp)
entry             410 drivers/s390/block/dcssblk.c 			segment_unload(entry->segment_name);
entry             454 drivers/s390/block/dcssblk.c 	struct segment_info *entry;
entry             466 drivers/s390/block/dcssblk.c 			list_for_each_entry(entry, &dev_info->seg_list, lh) {
entry             467 drivers/s390/block/dcssblk.c 				if (entry->segment_type == SEG_TYPE_EN ||
entry             468 drivers/s390/block/dcssblk.c 				    entry->segment_type == SEG_TYPE_SN)
entry             471 drivers/s390/block/dcssblk.c 						entry->segment_name);
entry             473 drivers/s390/block/dcssblk.c 					segment_save(entry->segment_name);
entry             512 drivers/s390/block/dcssblk.c 	struct segment_info *entry;
entry             518 drivers/s390/block/dcssblk.c 	list_for_each_entry(entry, &dev_info->seg_list, lh) {
entry             519 drivers/s390/block/dcssblk.c 		strcpy(&buf[i], entry->segment_name);
entry             520 drivers/s390/block/dcssblk.c 		i += strlen(entry->segment_name);
entry             745 drivers/s390/block/dcssblk.c 	struct segment_info *entry;
entry             794 drivers/s390/block/dcssblk.c 	list_for_each_entry(entry, &dev_info->seg_list, lh)
entry             795 drivers/s390/block/dcssblk.c 		segment_unload(entry->segment_name);
entry             830 drivers/s390/block/dcssblk.c 	struct segment_info *entry;
entry             841 drivers/s390/block/dcssblk.c 		list_for_each_entry(entry, &dev_info->seg_list, lh) {
entry             842 drivers/s390/block/dcssblk.c 			if (entry->segment_type == SEG_TYPE_EN ||
entry             843 drivers/s390/block/dcssblk.c 			    entry->segment_type == SEG_TYPE_SN)
entry             845 drivers/s390/block/dcssblk.c 					" be saved\n", entry->segment_name);
entry             847 drivers/s390/block/dcssblk.c 				segment_save(entry->segment_name);
entry            1019 drivers/s390/block/dcssblk.c 	struct segment_info *entry;
entry            1024 drivers/s390/block/dcssblk.c 		list_for_each_entry(entry, &dev_info->seg_list, lh) {
entry            1025 drivers/s390/block/dcssblk.c 			segment_unload(entry->segment_name);
entry            1026 drivers/s390/block/dcssblk.c 			rc = segment_load(entry->segment_name, SEGMENT_SHARED,
entry            1030 drivers/s390/block/dcssblk.c 				segment_warning(rc, entry->segment_name);
entry            1033 drivers/s390/block/dcssblk.c 			if (start != entry->start || end != entry->end) {
entry            1036 drivers/s390/block/dcssblk.c 				       entry->segment_name);
entry              91 drivers/s390/char/monwriter.c 	struct mon_buf *entry, *next;
entry              93 drivers/s390/char/monwriter.c 	list_for_each_entry_safe(entry, next, &monpriv->list, list)
entry              94 drivers/s390/char/monwriter.c 		if ((entry->hdr.mon_function == monhdr->mon_function ||
entry              96 drivers/s390/char/monwriter.c 		    entry->hdr.applid == monhdr->applid &&
entry              97 drivers/s390/char/monwriter.c 		    entry->hdr.record_num == monhdr->record_num &&
entry              98 drivers/s390/char/monwriter.c 		    entry->hdr.version == monhdr->version &&
entry              99 drivers/s390/char/monwriter.c 		    entry->hdr.release == monhdr->release &&
entry             100 drivers/s390/char/monwriter.c 		    entry->hdr.mod_level == monhdr->mod_level)
entry             101 drivers/s390/char/monwriter.c 			return entry;
entry             209 drivers/s390/char/monwriter.c 	struct mon_buf *entry, *next;
entry             211 drivers/s390/char/monwriter.c 	list_for_each_entry_safe(entry, next, &monpriv->list, list) {
entry             212 drivers/s390/char/monwriter.c 		if (entry->hdr.mon_function != MONWRITE_GEN_EVENT)
entry             213 drivers/s390/char/monwriter.c 			monwrite_diag(&entry->hdr, entry->data,
entry             216 drivers/s390/char/monwriter.c 		list_del(&entry->list);
entry             217 drivers/s390/char/monwriter.c 		kfree(entry->data);
entry             218 drivers/s390/char/monwriter.c 		kfree(entry);
entry             413 drivers/s390/cio/blacklist.c 	struct proc_dir_entry *entry;
entry             415 drivers/s390/cio/blacklist.c 	entry = proc_create("cio_ignore", S_IFREG | S_IRUGO | S_IWUSR, NULL,
entry             417 drivers/s390/cio/blacklist.c 	if (!entry)
entry             161 drivers/s390/cio/chp.c 	struct cmg_entry *entry, reference_buf;
entry             171 drivers/s390/cio/chp.c 	entry = area + (idx * sizeof(struct cmg_entry));
entry             173 drivers/s390/cio/chp.c 		memcpy(buf, entry, sizeof(*entry));
entry             174 drivers/s390/cio/chp.c 		memcpy(&reference_buf, entry, sizeof(*entry));
entry            1383 drivers/s390/cio/css.c 	struct proc_dir_entry *entry;
entry            1385 drivers/s390/cio/css.c 	entry = proc_create("cio_settle", S_IWUSR, NULL,
entry            1387 drivers/s390/cio/css.c 	if (!entry)
entry             256 drivers/s390/cio/qdio.h 	struct list_head entry;
entry              34 drivers/s390/cio/qdio_debug.c 	struct qdio_dbf_entry *entry;
entry              38 drivers/s390/cio/qdio_debug.c 	list_for_each_entry(entry, &qdio_dbf_list, dbf_list) {
entry              39 drivers/s390/cio/qdio_debug.c 		if (strcmp(entry->dbf_name, name) == 0) {
entry              40 drivers/s390/cio/qdio_debug.c 			rc = entry->dbf_info;
entry              50 drivers/s390/cio/qdio_debug.c 	struct qdio_dbf_entry *entry, *tmp;
entry              53 drivers/s390/cio/qdio_debug.c 	list_for_each_entry_safe(entry, tmp, &qdio_dbf_list, dbf_list) {
entry              54 drivers/s390/cio/qdio_debug.c 		list_del(&entry->dbf_list);
entry              55 drivers/s390/cio/qdio_debug.c 		debug_unregister(entry->dbf_info);
entry              56 drivers/s390/cio/qdio_debug.c 		kfree(entry);
entry            1784 drivers/s390/cio/qdio_main.c 				void *entry),
entry             154 drivers/s390/cio/qdio_setup.c 		INIT_LIST_HEAD(&q->entry);
entry             183 drivers/s390/cio/qdio_setup.c 	INIT_LIST_HEAD(&q->entry);
entry              80 drivers/s390/cio/qdio_thinint.c 	list_add_rcu(&irq_ptr->input_qs[0]->entry, &tiq_list);
entry              93 drivers/s390/cio/qdio_thinint.c 	list_del_rcu(&q->entry);
entry              96 drivers/s390/cio/qdio_thinint.c 	INIT_LIST_HEAD(&q->entry);
entry             194 drivers/s390/cio/qdio_thinint.c 	list_for_each_entry_rcu(q, &tiq_list, entry) {
entry             298 drivers/s390/cio/trace.h 		memcpy(&entry->request, chsc,
entry             302 drivers/s390/cio/trace.h 		memcpy(&entry->response, chsc,
entry             181 drivers/s390/net/ism.h 	struct smcd_event entry[15];
entry             392 drivers/s390/net/ism_drv.c 	struct smcd_event *entry;
entry             395 drivers/s390/net/ism_drv.c 		if (++(ism->ieq_idx) == ARRAY_SIZE(ism->ieq->entry))
entry             398 drivers/s390/net/ism_drv.c 		entry = &ism->ieq->entry[ism->ieq_idx];
entry             399 drivers/s390/net/ism_drv.c 		debug_event(ism_debug_info, 2, entry, sizeof(*entry));
entry             400 drivers/s390/net/ism_drv.c 		smcd_handle_event(ism->smcd, entry);
entry             967 drivers/s390/net/qeth_core.h 		struct qeth_buffer_pool_entry *entry)
entry             969 drivers/s390/net/qeth_core.h 	list_add_tail(&entry->list, &card->qdio.in_buf_pool.entry_list);
entry            2571 drivers/s390/net/qeth_core_main.c 	struct qeth_buffer_pool_entry *entry;
entry            2575 drivers/s390/net/qeth_core_main.c 	list_for_each_entry(entry,
entry            2577 drivers/s390/net/qeth_core_main.c 		qeth_put_buffer_pool_entry(card, entry);
entry            2585 drivers/s390/net/qeth_core_main.c 	struct qeth_buffer_pool_entry *entry;
entry            2593 drivers/s390/net/qeth_core_main.c 		entry = list_entry(plh, struct qeth_buffer_pool_entry, list);
entry            2596 drivers/s390/net/qeth_core_main.c 			if (page_count(virt_to_page(entry->elements[i])) > 1) {
entry            2602 drivers/s390/net/qeth_core_main.c 			list_del_init(&entry->list);
entry            2603 drivers/s390/net/qeth_core_main.c 			return entry;
entry            2608 drivers/s390/net/qeth_core_main.c 	entry = list_entry(card->qdio.in_buf_pool.entry_list.next,
entry            2611 drivers/s390/net/qeth_core_main.c 		if (page_count(virt_to_page(entry->elements[i])) > 1) {
entry            2616 drivers/s390/net/qeth_core_main.c 				free_page((unsigned long)entry->elements[i]);
entry            2617 drivers/s390/net/qeth_core_main.c 				entry->elements[i] = page_address(page);
entry            2622 drivers/s390/net/qeth_core_main.c 	list_del_init(&entry->list);
entry            2623 drivers/s390/net/qeth_core_main.c 	return entry;
entry            5522 drivers/s390/net/qeth_core_main.c 	struct qeth_dbf_entry *entry;
entry            5526 drivers/s390/net/qeth_core_main.c 	list_for_each_entry(entry, &qeth_dbf_list, dbf_list) {
entry            5527 drivers/s390/net/qeth_core_main.c 		if (strcmp(entry->dbf_name, name) == 0) {
entry            5528 drivers/s390/net/qeth_core_main.c 			rc = entry->dbf_info;
entry            5566 drivers/s390/net/qeth_core_main.c 	struct qeth_dbf_entry *entry, *tmp;
entry            5569 drivers/s390/net/qeth_core_main.c 	list_for_each_entry_safe(entry, tmp, &qeth_dbf_list, dbf_list) {
entry            5570 drivers/s390/net/qeth_core_main.c 		list_del(&entry->dbf_list);
entry            5571 drivers/s390/net/qeth_core_main.c 		debug_unregister(entry->dbf_info);
entry            5572 drivers/s390/net/qeth_core_main.c 		kfree(entry);
entry             710 drivers/s390/net/qeth_core_mpc.h 	struct qeth_sbp_port_entry entry[];
entry             718 drivers/s390/net/qeth_core_mpc.h 	struct qeth_sbp_port_entry entry[];
entry             755 drivers/s390/net/qeth_core_mpc.h 	struct qeth_ipacmd_addr_change_entry entry[];
entry            1160 drivers/s390/net/qeth_l2_main.c 	struct qeth_sbp_port_entry *entry = &data->qports.entry[0];
entry            1174 drivers/s390/net/qeth_l2_main.c 	data->card->options.sbp.role = entry->role;
entry            1179 drivers/s390/net/qeth_l2_main.c 		(entry->role == QETH_SBP_ROLE_NONE) ? "none" :
entry            1180 drivers/s390/net/qeth_l2_main.c 		(entry->role == QETH_SBP_ROLE_PRIMARY) ? "primary" :
entry            1181 drivers/s390/net/qeth_l2_main.c 		(entry->role == QETH_SBP_ROLE_SECONDARY) ? "secondary" :
entry            1184 drivers/s390/net/qeth_l2_main.c 		(entry->state == QETH_SBP_STATE_INACTIVE) ? "inactive" :
entry            1185 drivers/s390/net/qeth_l2_main.c 		(entry->state == QETH_SBP_STATE_STANDBY) ? "standby" :
entry            1186 drivers/s390/net/qeth_l2_main.c 		(entry->state == QETH_SBP_STATE_ACTIVE) ? "active" :
entry            1248 drivers/s390/net/qeth_l2_main.c 			struct qeth_ipacmd_addr_change_entry *entry =
entry            1249 drivers/s390/net/qeth_l2_main.c 					&data->hostevs.entry[i];
entry            1252 drivers/s390/net/qeth_l2_main.c 					entry->change_code,
entry            1253 drivers/s390/net/qeth_l2_main.c 					&entry->token, &entry->addr_lnid);
entry            1490 drivers/s390/net/qeth_l2_main.c 			*cbctl->data.qports.role = qports->entry[0].role;
entry            1492 drivers/s390/net/qeth_l2_main.c 			*cbctl->data.qports.state = qports->entry[0].state;
entry            1616 drivers/s390/net/qeth_l2_main.c 		enum qdio_brinfo_entry_type type, void *entry)
entry            1627 drivers/s390/net/qeth_l2_main.c 	l2entry = (struct qdio_brinfo_entry_l2 *)entry;
entry             110 drivers/s390/net/qeth_l3.h 	struct list_head entry;
entry             123 drivers/s390/net/qeth_l3_main.c 	list_for_each_entry(ipatoe, &card->ipato.entries, entry) {
entry             579 drivers/s390/net/qeth_l3_main.c 	list_for_each_entry_safe(ipatoe, tmp, &card->ipato.entries, entry) {
entry             580 drivers/s390/net/qeth_l3_main.c 		list_del(&ipatoe->entry);
entry             598 drivers/s390/net/qeth_l3_main.c 	list_for_each_entry(ipatoe, &card->ipato.entries, entry) {
entry             610 drivers/s390/net/qeth_l3_main.c 		list_add_tail(&new->entry, &card->ipato.entries);
entry             630 drivers/s390/net/qeth_l3_main.c 	list_for_each_entry_safe(ipatoe, tmp, &card->ipato.entries, entry) {
entry             636 drivers/s390/net/qeth_l3_main.c 			list_del(&ipatoe->entry);
entry            1755 drivers/s390/net/qeth_l3_main.c 				    struct qeth_arp_cache_entry *entry,
entry            1785 drivers/s390/net/qeth_l3_main.c 	ether_addr_copy(cmd_entry->macaddr, entry->macaddr);
entry            1786 drivers/s390/net/qeth_l3_main.c 	memcpy(cmd_entry->ipaddr, entry->ipaddr, 4);
entry             456 drivers/s390/net/qeth_l3_sys.c 	list_for_each_entry(ipatoe, &card->ipato.entries, entry) {
entry             285 drivers/s390/scsi/zfcp_dbf.c 	struct list_head *entry;
entry             300 drivers/s390/scsi/zfcp_dbf.c 	list_for_each(entry, &adapter->erp_ready_head)
entry             303 drivers/s390/scsi/zfcp_dbf.c 	list_for_each(entry, &adapter->erp_running_head)
entry             298 drivers/sbus/char/bbc_i2c.c 	int entry;
entry             328 drivers/sbus/char/bbc_i2c.c 	entry = 0;
entry             330 drivers/sbus/char/bbc_i2c.c 	     dp && entry < 8;
entry             331 drivers/sbus/char/bbc_i2c.c 	     dp = dp->sibling, entry++) {
entry             335 drivers/sbus/char/bbc_i2c.c 		bp->devs[entry].device = child_op;
entry             336 drivers/sbus/char/bbc_i2c.c 		bp->devs[entry].client_claimed = 0;
entry             345 drivers/sbus/char/bbc_i2c.c 	       bp->index, bp->i2c_control_regs, entry, bp->own, bp->clock);
entry            1529 drivers/scsi/aacraid/aacraid.h 	struct list_head	entry;
entry            2689 drivers/scsi/aacraid/aacraid.h int aac_consumer_get(struct aac_dev * dev, struct aac_queue * q, struct aac_entry **entry);
entry             176 drivers/scsi/aacraid/commctrl.c 		struct list_head * entry;
entry             206 drivers/scsi/aacraid/commctrl.c 		entry = dev->fib_list.next;
entry             207 drivers/scsi/aacraid/commctrl.c 		while (entry != &dev->fib_list) {
entry             208 drivers/scsi/aacraid/commctrl.c 			context = list_entry(entry, struct aac_fib_context, next);
entry             212 drivers/scsi/aacraid/commctrl.c 				entry = dev->fib_list.next;
entry             214 drivers/scsi/aacraid/commctrl.c 				entry = entry->next;
entry             244 drivers/scsi/aacraid/commctrl.c 	struct list_head * entry;
entry             256 drivers/scsi/aacraid/commctrl.c 	entry = dev->fib_list.next;
entry             259 drivers/scsi/aacraid/commctrl.c 	while (entry != &dev->fib_list) {
entry             260 drivers/scsi/aacraid/commctrl.c 		fibctx = list_entry(entry, struct aac_fib_context, next);
entry             267 drivers/scsi/aacraid/commctrl.c 		entry = entry->next;
entry             292 drivers/scsi/aacraid/commctrl.c 		entry = fibctx->fib_list.next;
entry             293 drivers/scsi/aacraid/commctrl.c 		list_del(entry);
entry             295 drivers/scsi/aacraid/commctrl.c 		fib = list_entry(entry, struct fib, fiblink);
entry             346 drivers/scsi/aacraid/commctrl.c 		struct list_head * entry;
entry             350 drivers/scsi/aacraid/commctrl.c 		entry = fibctx->fib_list.next;
entry             351 drivers/scsi/aacraid/commctrl.c 		list_del(entry);
entry             352 drivers/scsi/aacraid/commctrl.c 		fib = list_entry(entry, struct fib, fiblink);
entry             388 drivers/scsi/aacraid/commctrl.c 	struct list_head * entry;
entry             397 drivers/scsi/aacraid/commctrl.c 	entry = dev->fib_list.next;
entry             400 drivers/scsi/aacraid/commctrl.c 	while(entry != &dev->fib_list) {
entry             401 drivers/scsi/aacraid/commctrl.c 		fibctx = list_entry(entry, struct aac_fib_context, next);
entry             407 drivers/scsi/aacraid/commctrl.c 		entry = entry->next;
entry             357 drivers/scsi/aacraid/commsup.c static int aac_get_entry (struct aac_dev * dev, u32 qid, struct aac_entry **entry, u32 * index, unsigned long *nonotify)
entry             398 drivers/scsi/aacraid/commsup.c 		*entry = q->base + *index;
entry             421 drivers/scsi/aacraid/commsup.c 	struct aac_entry * entry = NULL;
entry             426 drivers/scsi/aacraid/commsup.c 		while (!aac_get_entry(dev, qid, &entry, index, nonotify)) {
entry             432 drivers/scsi/aacraid/commsup.c 		entry->size = cpu_to_le32(le16_to_cpu(hw_fib->header.Size));
entry             435 drivers/scsi/aacraid/commsup.c 		while (!aac_get_entry(dev, qid, &entry, index, nonotify)) {
entry             441 drivers/scsi/aacraid/commsup.c 		entry->size = cpu_to_le32(le16_to_cpu(hw_fib->header.Size));
entry             442 drivers/scsi/aacraid/commsup.c 		entry->addr = hw_fib->header.SenderFibAddress;
entry             452 drivers/scsi/aacraid/commsup.c 		entry->addr = cpu_to_le32(fibptr->hw_fib_pa);
entry             798 drivers/scsi/aacraid/commsup.c int aac_consumer_get(struct aac_dev * dev, struct aac_queue * q, struct aac_entry **entry)
entry             814 drivers/scsi/aacraid/commsup.c 		*entry = q->base + index;
entry            1725 drivers/scsi/aacraid/commsup.c 	struct list_head * entry;
entry            1748 drivers/scsi/aacraid/commsup.c 	entry = aac->fib_list.next;
entry            1756 drivers/scsi/aacraid/commsup.c 	while (entry != &aac->fib_list) {
entry            1760 drivers/scsi/aacraid/commsup.c 		struct aac_fib_context *fibctx = list_entry(entry, struct aac_fib_context, next);
entry            1780 drivers/scsi/aacraid/commsup.c 				entry = entry->next;
entry            1824 drivers/scsi/aacraid/commsup.c 		entry = entry->next;
entry            2017 drivers/scsi/aacraid/commsup.c 	struct list_head *entry;
entry            2029 drivers/scsi/aacraid/commsup.c 	entry = dev->fib_list.next;
entry            2030 drivers/scsi/aacraid/commsup.c 	while (entry != &dev->fib_list) {
entry            2031 drivers/scsi/aacraid/commsup.c 		entry = entry->next;
entry            2077 drivers/scsi/aacraid/commsup.c 	struct list_head *entry;
entry            2087 drivers/scsi/aacraid/commsup.c 	entry = dev->fib_list.next;
entry            2097 drivers/scsi/aacraid/commsup.c 	while (entry != &dev->fib_list) {
entry            2101 drivers/scsi/aacraid/commsup.c 		fibctx = list_entry(entry, struct aac_fib_context,
entry            2120 drivers/scsi/aacraid/commsup.c 				entry = entry->next;
entry            2131 drivers/scsi/aacraid/commsup.c 			entry = entry->next;
entry            2157 drivers/scsi/aacraid/commsup.c 		entry = entry->next;
entry            2179 drivers/scsi/aacraid/commsup.c 		struct list_head *entry;
entry            2187 drivers/scsi/aacraid/commsup.c 		entry = dev->queues->queue[HostNormCmdQueue].cmdq.next;
entry            2188 drivers/scsi/aacraid/commsup.c 		list_del(entry);
entry            2193 drivers/scsi/aacraid/commsup.c 		fib = list_entry(entry, struct fib, fiblink);
entry              42 drivers/scsi/aacraid/dpcsup.c 	struct aac_entry *entry;
entry              55 drivers/scsi/aacraid/dpcsup.c 	while(aac_consumer_get(dev, q, &entry))
entry              58 drivers/scsi/aacraid/dpcsup.c 		u32 index = le32_to_cpu(entry->addr);
entry             160 drivers/scsi/aacraid/dpcsup.c 	struct aac_entry *entry;
entry             170 drivers/scsi/aacraid/dpcsup.c 	while(aac_consumer_get(dev, q, &entry))
entry             177 drivers/scsi/aacraid/dpcsup.c 		index = le32_to_cpu(entry->addr) / sizeof(struct hw_fib);
entry             333 drivers/scsi/aacraid/linit.c 		struct partition *entry = first;
entry             339 drivers/scsi/aacraid/linit.c 			end_head = entry->end_head;
entry             340 drivers/scsi/aacraid/linit.c 			end_sec = entry->end_sector & 0x3f;
entry             355 drivers/scsi/aacraid/linit.c 			entry++;
entry            1121 drivers/scsi/aacraid/linit.c 	list_for_each_entry(aac, &aac_devices, entry) {
entry            1606 drivers/scsi/aacraid/linit.c 	list_for_each_entry(aac, &aac_devices, entry) {
entry            1609 drivers/scsi/aacraid/linit.c 		insert = &aac->entry;
entry            1669 drivers/scsi/aacraid/linit.c 	INIT_LIST_HEAD(&aac->entry);
entry            1765 drivers/scsi/aacraid/linit.c 	list_add(&aac->entry, insert);
entry            1948 drivers/scsi/aacraid/linit.c 	list_del(&aac->entry);
entry              79 drivers/scsi/aacraid/src.c 		struct list_head *entry;
entry             101 drivers/scsi/aacraid/src.c 				entry = dev->sync_fib_list.next;
entry             102 drivers/scsi/aacraid/src.c 				dev->sync_fib = list_entry(entry,
entry             105 drivers/scsi/aacraid/src.c 				list_del(entry);
entry             107 drivers/scsi/aic7xxx/aic7770.c 	struct	aic7770_identity *entry;
entry             111 drivers/scsi/aic7xxx/aic7770.c 		entry = &aic7770_ident_table[i];
entry             112 drivers/scsi/aic7xxx/aic7770.c 		if (entry->full_id == (id & entry->id_mask))
entry             113 drivers/scsi/aic7xxx/aic7770.c 			return (entry);
entry             119 drivers/scsi/aic7xxx/aic7770.c aic7770_config(struct ahc_softc *ahc, struct aic7770_identity *entry, u_int io)
entry             127 drivers/scsi/aic7xxx/aic7770.c 	error = entry->setup(ahc);
entry             144 drivers/scsi/aic7xxx/aic7770.c 	ahc->description = entry->name;
entry            4341 drivers/scsi/aic7xxx/aic79xx_core.c 	const struct ahd_phase_table_entry *entry;
entry            4349 drivers/scsi/aic7xxx/aic79xx_core.c 	for (entry = ahd_phase_table; entry < last_entry; entry++) {
entry            4350 drivers/scsi/aic7xxx/aic79xx_core.c 		if (phase == entry->phase)
entry            4353 drivers/scsi/aic7xxx/aic79xx_core.c 	return (entry);
entry            9648 drivers/scsi/aic7xxx/aic79xx_core.c 		int entry;
entry            9650 drivers/scsi/aic7xxx/aic79xx_core.c 		for (entry = 0; entry < num_entries; entry++) {
entry            9651 drivers/scsi/aic7xxx/aic79xx_core.c 			if (((value & table[entry].mask)
entry            9652 drivers/scsi/aic7xxx/aic79xx_core.c 			  != table[entry].value)
entry            9653 drivers/scsi/aic7xxx/aic79xx_core.c 			 || ((printed_mask & table[entry].mask)
entry            9654 drivers/scsi/aic7xxx/aic79xx_core.c 			  == table[entry].mask))
entry            9659 drivers/scsi/aic7xxx/aic79xx_core.c 					  table[entry].name);
entry            9660 drivers/scsi/aic7xxx/aic79xx_core.c 			printed_mask |= table[entry].mask;
entry            9664 drivers/scsi/aic7xxx/aic79xx_core.c 		if (entry >= num_entries)
entry             162 drivers/scsi/aic7xxx/aic79xx_osm_pci.c 	const struct ahd_pci_identity *entry;
entry             168 drivers/scsi/aic7xxx/aic79xx_osm_pci.c 	entry = ahd_find_pci_device(pci);
entry             169 drivers/scsi/aic7xxx/aic79xx_osm_pci.c 	if (entry == NULL)
entry             208 drivers/scsi/aic7xxx/aic79xx_osm_pci.c 	error = ahd_pci_config(ahd, entry);
entry             258 drivers/scsi/aic7xxx/aic79xx_pci.c 	const struct ahd_pci_identity *entry;
entry             277 drivers/scsi/aic7xxx/aic79xx_pci.c 		entry = &ahd_pci_ident_table[i];
entry             278 drivers/scsi/aic7xxx/aic79xx_pci.c 		if (entry->full_id == (full_id & entry->id_mask)) {
entry             280 drivers/scsi/aic7xxx/aic79xx_pci.c 			if (entry->name == NULL)
entry             282 drivers/scsi/aic7xxx/aic79xx_pci.c 			return (entry);
entry             289 drivers/scsi/aic7xxx/aic79xx_pci.c ahd_pci_config(struct ahd_softc *ahd, const struct ahd_pci_identity *entry)
entry             296 drivers/scsi/aic7xxx/aic79xx_pci.c 	ahd->description = entry->name;
entry             305 drivers/scsi/aic7xxx/aic79xx_pci.c 	error = entry->setup(ahd);
entry            2797 drivers/scsi/aic7xxx/aic7xxx_core.c 	const struct ahc_phase_table_entry *entry;
entry            2805 drivers/scsi/aic7xxx/aic7xxx_core.c 	for (entry = ahc_phase_table; entry < last_entry; entry++) {
entry            2806 drivers/scsi/aic7xxx/aic7xxx_core.c 		if (phase == entry->phase)
entry            2809 drivers/scsi/aic7xxx/aic7xxx_core.c 	return (entry);
entry            7106 drivers/scsi/aic7xxx/aic7xxx_core.c 		int entry;
entry            7108 drivers/scsi/aic7xxx/aic7xxx_core.c 		for (entry = 0; entry < num_entries; entry++) {
entry            7109 drivers/scsi/aic7xxx/aic7xxx_core.c 			if (((value & table[entry].mask)
entry            7110 drivers/scsi/aic7xxx/aic7xxx_core.c 			  != table[entry].value)
entry            7111 drivers/scsi/aic7xxx/aic7xxx_core.c 			 || ((printed_mask & table[entry].mask)
entry            7112 drivers/scsi/aic7xxx/aic7xxx_core.c 			  == table[entry].mask))
entry            7117 drivers/scsi/aic7xxx/aic7xxx_core.c 					  table[entry].name);
entry            7118 drivers/scsi/aic7xxx/aic7xxx_core.c 			printed_mask |= table[entry].mask;
entry            7122 drivers/scsi/aic7xxx/aic7xxx_core.c 		if (entry >= num_entries)
entry             209 drivers/scsi/aic7xxx/aic7xxx_osm_pci.c 	const struct ahc_pci_identity *entry;
entry             215 drivers/scsi/aic7xxx/aic7xxx_osm_pci.c 	entry = ahc_find_pci_device(pci);
entry             216 drivers/scsi/aic7xxx/aic7xxx_osm_pci.c 	if (entry == NULL)
entry             254 drivers/scsi/aic7xxx/aic7xxx_osm_pci.c 	error = ahc_pci_config(ahc, entry);
entry             671 drivers/scsi/aic7xxx/aic7xxx_pci.c 	const struct ahc_pci_identity *entry;
entry             693 drivers/scsi/aic7xxx/aic7xxx_pci.c 		entry = &ahc_pci_ident_table[i];
entry             694 drivers/scsi/aic7xxx/aic7xxx_pci.c 		if (entry->full_id == (full_id & entry->id_mask)) {
entry             696 drivers/scsi/aic7xxx/aic7xxx_pci.c 			if (entry->name == NULL)
entry             698 drivers/scsi/aic7xxx/aic7xxx_pci.c 			return (entry);
entry             705 drivers/scsi/aic7xxx/aic7xxx_pci.c ahc_pci_config(struct ahc_softc *ahc, const struct ahc_pci_identity *entry)
entry             717 drivers/scsi/aic7xxx/aic7xxx_pci.c 	error = entry->setup(ahc);
entry             721 drivers/scsi/aic7xxx/aic7xxx_pci.c 	ahc->description = entry->name;
entry              34 drivers/scsi/aic94xx/aic94xx_sds.c 	struct asd_ocm_dir_ent entry[15];
entry             183 drivers/scsi/aic94xx/aic94xx_sds.c 		if (dir->entry[i].type == type)
entry             188 drivers/scsi/aic94xx/aic94xx_sds.c 	ent = &dir->entry[i];
entry            1126 drivers/scsi/bfa/bfad.c 	bfa_msix(&bfad->bfa, vec->msix.entry);
entry            1153 drivers/scsi/bfa/bfad.c 			bfad->msix_tab[bfad->nvec].msix.entry = i;
entry            1155 drivers/scsi/bfa/bfad.c 			msix_entries[bfad->nvec].entry = i;
entry             135 drivers/scsi/bfa/bfad_im.h static inline void bfad_im_post_vendor_event(struct bfa_aen_entry_s *entry,
entry             148 drivers/scsi/bfa/bfad_im.h 	entry->aen_tv_sec = ts.tv_sec;
entry             149 drivers/scsi/bfa/bfad_im.h 	entry->aen_tv_usec = ts.tv_nsec / NSEC_PER_USEC;
entry             150 drivers/scsi/bfa/bfad_im.h 	entry->bfad_num = drv->inst_no;
entry             151 drivers/scsi/bfa/bfad_im.h 	entry->seq_num = cnt;
entry             152 drivers/scsi/bfa/bfad_im.h 	entry->aen_category = cat;
entry             153 drivers/scsi/bfa/bfad_im.h 	entry->aen_type = evt;
entry            1381 drivers/scsi/cxlflash/main.c 	u64 entry,
entry            1388 drivers/scsi/cxlflash/main.c 		entry = *hrrq_curr;
entry            1390 drivers/scsi/cxlflash/main.c 		if ((entry & SISL_RESP_HANDLE_T_BIT) != toggle)
entry            1393 drivers/scsi/cxlflash/main.c 		entry &= ~SISL_RESP_HANDLE_T_BIT;
entry            1396 drivers/scsi/cxlflash/main.c 			ioasa = (struct sisl_ioasa *)entry;
entry            1399 drivers/scsi/cxlflash/main.c 			ioarcb = (struct sisl_ioarcb *)entry;
entry              94 drivers/scsi/device_handler/scsi_dh_alua.c 	struct list_head	entry;
entry             866 drivers/scsi/device_handler/scsi_dh_alua.c 	list_for_each_entry_safe(qdata, tmp, &qdata_list, entry) {
entry             867 drivers/scsi/device_handler/scsi_dh_alua.c 		list_del(&qdata->entry);
entry             901 drivers/scsi/device_handler/scsi_dh_alua.c 		list_add_tail(&qdata->entry, &pg->rtpg_list);
entry             228 drivers/scsi/device_handler/scsi_dh_rdac.c 	struct list_head	entry;
entry             300 drivers/scsi/device_handler/scsi_dh_rdac.c 	list_for_each_entry(qdata, list, entry) {
entry             574 drivers/scsi/device_handler/scsi_dh_rdac.c 	list_for_each_entry_safe(qdata, tmp, &list, entry) {
entry             575 drivers/scsi/device_handler/scsi_dh_rdac.c 		list_del(&qdata->entry);
entry             601 drivers/scsi/device_handler/scsi_dh_rdac.c 	list_add_tail(&qdata->entry, &ctlr->ms_head);
entry            1795 drivers/scsi/fcoe/fcoe.c 	struct dcb_app_type *entry = ptr;
entry            1801 drivers/scsi/fcoe/fcoe.c 	if (entry->app.selector != DCB_APP_IDTYPE_ETHTYPE)
entry            1804 drivers/scsi/fcoe/fcoe.c 	netdev = dev_get_by_index(&init_net, entry->ifindex);
entry            1815 drivers/scsi/fcoe/fcoe.c 	if (entry->dcbx & DCB_CAP_DCBX_VER_CEE)
entry            1816 drivers/scsi/fcoe/fcoe.c 		prio = ffs(entry->app.priority) - 1;
entry            1818 drivers/scsi/fcoe/fcoe.c 		prio = entry->app.priority;
entry            1823 drivers/scsi/fcoe/fcoe.c 	if (entry->app.protocol == ETH_P_FIP ||
entry            1824 drivers/scsi/fcoe/fcoe.c 	    entry->app.protocol == ETH_P_FCOE)
entry            1827 drivers/scsi/fcoe/fcoe.c 	if (entry->app.protocol == ETH_P_FCOE)
entry             652 drivers/scsi/gdth.h     gdth_hentry_str entry[MAX_HDRIVES];         /* entries */
entry             491 drivers/scsi/gdth_proc.c             phg->offset = GDTOFFSOF(gdth_hget_str, entry[0]); 
entry             498 drivers/scsi/gdth_proc.c                     k = phg->entry[j].host_drive;
entry             501 drivers/scsi/gdth_proc.c                     ha->hdr[k].ldr_no = phg->entry[j].log_drive;
entry             502 drivers/scsi/gdth_proc.c                     ha->hdr[k].rw_attribs = phg->entry[j].rw_attribs;
entry             503 drivers/scsi/gdth_proc.c                     ha->hdr[k].start_sec = phg->entry[j].start_sec;
entry             215 drivers/scsi/hisi_sas/hisi_sas.h 	struct list_head entry;
entry             251 drivers/scsi/hisi_sas/hisi_sas_main.c 	list_del_init(&slot->entry);
entry             498 drivers/scsi/hisi_sas/hisi_sas_main.c 	list_add_tail(&slot->entry, &sas_dev->list);
entry            1020 drivers/scsi/hisi_sas/hisi_sas_main.c 	list_for_each_entry_safe(slot, slot2, &sas_dev->list, entry)
entry            1970 drivers/scsi/hisi_sas/hisi_sas_main.c 	list_add_tail(&slot->entry, &sas_dev->list);
entry             898 drivers/scsi/hisi_sas/hisi_sas_v1_hw.c 		struct hisi_sas_sge *entry = &sge_page->sge[i];
entry             900 drivers/scsi/hisi_sas/hisi_sas_v1_hw.c 		entry->addr = cpu_to_le64(sg_dma_address(sg));
entry             901 drivers/scsi/hisi_sas/hisi_sas_v1_hw.c 		entry->page_ctrl_0 = entry->page_ctrl_1 = 0;
entry             902 drivers/scsi/hisi_sas/hisi_sas_v1_hw.c 		entry->data_len = cpu_to_le32(sg_dma_len(sg));
entry             903 drivers/scsi/hisi_sas/hisi_sas_v1_hw.c 		entry->data_off = 0;
entry            1678 drivers/scsi/hisi_sas/hisi_sas_v2_hw.c 		struct hisi_sas_sge *entry = &sge_page->sge[i];
entry            1680 drivers/scsi/hisi_sas/hisi_sas_v2_hw.c 		entry->addr = cpu_to_le64(sg_dma_address(sg));
entry            1681 drivers/scsi/hisi_sas/hisi_sas_v2_hw.c 		entry->page_ctrl_0 = entry->page_ctrl_1 = 0;
entry            1682 drivers/scsi/hisi_sas/hisi_sas_v2_hw.c 		entry->data_len = cpu_to_le32(sg_dma_len(sg));
entry            1683 drivers/scsi/hisi_sas/hisi_sas_v2_hw.c 		entry->data_off = 0;
entry             830 drivers/scsi/hisi_sas/hisi_sas_v3_hw.c 	list_for_each_entry_safe(slot, slot2, &sas_dev->list, entry) {
entry            1059 drivers/scsi/hisi_sas/hisi_sas_v3_hw.c 		struct hisi_sas_sge *entry = &sge_page->sge[i];
entry            1061 drivers/scsi/hisi_sas/hisi_sas_v3_hw.c 		entry->addr = cpu_to_le64(sg_dma_address(sg));
entry            1062 drivers/scsi/hisi_sas/hisi_sas_v3_hw.c 		entry->page_ctrl_0 = entry->page_ctrl_1 = 0;
entry            1063 drivers/scsi/hisi_sas/hisi_sas_v3_hw.c 		entry->data_len = cpu_to_le32(sg_dma_len(sg));
entry            1064 drivers/scsi/hisi_sas/hisi_sas_v3_hw.c 		entry->data_off = 0;
entry            1085 drivers/scsi/hisi_sas/hisi_sas_v3_hw.c 		struct hisi_sas_sge *entry = &sge_dif_page->sge[i];
entry            1087 drivers/scsi/hisi_sas/hisi_sas_v3_hw.c 		entry->addr = cpu_to_le64(sg_dma_address(sg));
entry            1088 drivers/scsi/hisi_sas/hisi_sas_v3_hw.c 		entry->page_ctrl_0 = 0;
entry            1089 drivers/scsi/hisi_sas/hisi_sas_v3_hw.c 		entry->page_ctrl_1 = 0;
entry            1090 drivers/scsi/hisi_sas/hisi_sas_v3_hw.c 		entry->data_len = cpu_to_le32(sg_dma_len(sg));
entry            1091 drivers/scsi/hisi_sas/hisi_sas_v3_hw.c 		entry->data_off = 0;
entry            1346 drivers/scsi/hpsa.c 	int entry, struct hpsa_scsi_dev_t *new_entry)
entry            1349 drivers/scsi/hpsa.c 	BUG_ON(entry < 0 || entry >= HPSA_MAX_DEVICES);
entry            1352 drivers/scsi/hpsa.c 	h->dev[entry]->raid_level = new_entry->raid_level;
entry            1357 drivers/scsi/hpsa.c 	h->dev[entry]->ioaccel_handle = new_entry->ioaccel_handle;
entry            1369 drivers/scsi/hpsa.c 		h->dev[entry]->raid_map = new_entry->raid_map;
entry            1370 drivers/scsi/hpsa.c 		h->dev[entry]->ioaccel_handle = new_entry->ioaccel_handle;
entry            1373 drivers/scsi/hpsa.c 		h->dev[entry]->ioaccel_handle = new_entry->ioaccel_handle;
entry            1376 drivers/scsi/hpsa.c 	h->dev[entry]->hba_ioaccel_enabled = new_entry->hba_ioaccel_enabled;
entry            1377 drivers/scsi/hpsa.c 	h->dev[entry]->offload_config = new_entry->offload_config;
entry            1378 drivers/scsi/hpsa.c 	h->dev[entry]->offload_to_mirror = new_entry->offload_to_mirror;
entry            1379 drivers/scsi/hpsa.c 	h->dev[entry]->queue_depth = new_entry->queue_depth;
entry            1386 drivers/scsi/hpsa.c 	h->dev[entry]->offload_to_be_enabled = new_entry->offload_to_be_enabled;
entry            1392 drivers/scsi/hpsa.c 		h->dev[entry]->offload_enabled = 0;
entry            1394 drivers/scsi/hpsa.c 	hpsa_show_dev_msg(KERN_INFO, h, h->dev[entry], "updated");
entry            1399 drivers/scsi/hpsa.c 	int entry, struct hpsa_scsi_dev_t *new_entry,
entry            1404 drivers/scsi/hpsa.c 	BUG_ON(entry < 0 || entry >= HPSA_MAX_DEVICES);
entry            1405 drivers/scsi/hpsa.c 	removed[*nremoved] = h->dev[entry];
entry            1413 drivers/scsi/hpsa.c 		new_entry->target = h->dev[entry]->target;
entry            1414 drivers/scsi/hpsa.c 		new_entry->lun = h->dev[entry]->lun;
entry            1417 drivers/scsi/hpsa.c 	h->dev[entry] = new_entry;
entry            1425 drivers/scsi/hpsa.c static void hpsa_scsi_remove_entry(struct ctlr_info *h, int entry,
entry            1432 drivers/scsi/hpsa.c 	BUG_ON(entry < 0 || entry >= HPSA_MAX_DEVICES);
entry            1434 drivers/scsi/hpsa.c 	sd = h->dev[entry];
entry            1435 drivers/scsi/hpsa.c 	removed[*nremoved] = h->dev[entry];
entry            1438 drivers/scsi/hpsa.c 	for (i = entry; i < h->ndevices-1; i++)
entry            1911 drivers/scsi/hpsa.c 	int i, entry, device_change, changes = 0;
entry            1952 drivers/scsi/hpsa.c 		device_change = hpsa_scsi_find_entry(csd, sd, nsds, &entry);
entry            1959 drivers/scsi/hpsa.c 			hpsa_scsi_replace_entry(h, i, sd[entry],
entry            1964 drivers/scsi/hpsa.c 			sd[entry] = NULL;
entry            1966 drivers/scsi/hpsa.c 			hpsa_scsi_update_entry(h, i, sd[entry]);
entry            1991 drivers/scsi/hpsa.c 					h->ndevices, &entry);
entry             150 drivers/scsi/ibmvscsi/ibmvfc.c 	struct ibmvfc_trace_entry *entry;
entry             152 drivers/scsi/ibmvscsi/ibmvfc.c 	entry = &vhost->trace[vhost->trace_index++];
entry             153 drivers/scsi/ibmvscsi/ibmvfc.c 	entry->evt = evt;
entry             154 drivers/scsi/ibmvscsi/ibmvfc.c 	entry->time = jiffies;
entry             155 drivers/scsi/ibmvscsi/ibmvfc.c 	entry->fmt = evt->crq.format;
entry             156 drivers/scsi/ibmvscsi/ibmvfc.c 	entry->type = IBMVFC_TRC_START;
entry             158 drivers/scsi/ibmvscsi/ibmvfc.c 	switch (entry->fmt) {
entry             160 drivers/scsi/ibmvscsi/ibmvfc.c 		entry->op_code = vfc_cmd->iu.cdb[0];
entry             161 drivers/scsi/ibmvscsi/ibmvfc.c 		entry->scsi_id = be64_to_cpu(vfc_cmd->tgt_scsi_id);
entry             162 drivers/scsi/ibmvscsi/ibmvfc.c 		entry->lun = scsilun_to_int(&vfc_cmd->iu.lun);
entry             163 drivers/scsi/ibmvscsi/ibmvfc.c 		entry->tmf_flags = vfc_cmd->iu.tmf_flags;
entry             164 drivers/scsi/ibmvscsi/ibmvfc.c 		entry->u.start.xfer_len = be32_to_cpu(vfc_cmd->iu.xfer_len);
entry             167 drivers/scsi/ibmvscsi/ibmvfc.c 		entry->op_code = be32_to_cpu(mad->opcode);
entry             184 drivers/scsi/ibmvscsi/ibmvfc.c 	struct ibmvfc_trace_entry *entry = &vhost->trace[vhost->trace_index++];
entry             186 drivers/scsi/ibmvscsi/ibmvfc.c 	entry->evt = evt;
entry             187 drivers/scsi/ibmvscsi/ibmvfc.c 	entry->time = jiffies;
entry             188 drivers/scsi/ibmvscsi/ibmvfc.c 	entry->fmt = evt->crq.format;
entry             189 drivers/scsi/ibmvscsi/ibmvfc.c 	entry->type = IBMVFC_TRC_END;
entry             191 drivers/scsi/ibmvscsi/ibmvfc.c 	switch (entry->fmt) {
entry             193 drivers/scsi/ibmvscsi/ibmvfc.c 		entry->op_code = vfc_cmd->iu.cdb[0];
entry             194 drivers/scsi/ibmvscsi/ibmvfc.c 		entry->scsi_id = be64_to_cpu(vfc_cmd->tgt_scsi_id);
entry             195 drivers/scsi/ibmvscsi/ibmvfc.c 		entry->lun = scsilun_to_int(&vfc_cmd->iu.lun);
entry             196 drivers/scsi/ibmvscsi/ibmvfc.c 		entry->tmf_flags = vfc_cmd->iu.tmf_flags;
entry             197 drivers/scsi/ibmvscsi/ibmvfc.c 		entry->u.end.status = be16_to_cpu(vfc_cmd->status);
entry             198 drivers/scsi/ibmvscsi/ibmvfc.c 		entry->u.end.error = be16_to_cpu(vfc_cmd->error);
entry             199 drivers/scsi/ibmvscsi/ibmvfc.c 		entry->u.end.fcp_rsp_flags = vfc_cmd->rsp.flags;
entry             200 drivers/scsi/ibmvscsi/ibmvfc.c 		entry->u.end.rsp_code = vfc_cmd->rsp.data.info.rsp_code;
entry             201 drivers/scsi/ibmvscsi/ibmvfc.c 		entry->u.end.scsi_status = vfc_cmd->rsp.scsi_status;
entry             204 drivers/scsi/ibmvscsi/ibmvfc.c 		entry->op_code = be32_to_cpu(mad->opcode);
entry             205 drivers/scsi/ibmvscsi/ibmvfc.c 		entry->u.end.status = be16_to_cpu(mad->status);
entry            3213 drivers/scsi/ipr.c 	bytes_to_copy = offsetof(struct ipr_sdt, entry) +
entry            3250 drivers/scsi/ipr.c 		if (sdt->entry[i].flags & IPR_SDT_VALID_ENTRY) {
entry            3251 drivers/scsi/ipr.c 			sdt_word = be32_to_cpu(sdt->entry[i].start_token);
entry            3253 drivers/scsi/ipr.c 				bytes_to_copy = be32_to_cpu(sdt->entry[i].end_token);
entry            3256 drivers/scsi/ipr.c 				end_off = be32_to_cpu(sdt->entry[i].end_token);
entry            3265 drivers/scsi/ipr.c 					sdt->entry[i].flags &= ~IPR_SDT_VALID_ENTRY;
entry            4303 drivers/scsi/ipr.c 		sdt_end = offsetof(struct ipr_ioa_dump, sdt.entry) +
entry            4307 drivers/scsi/ipr.c 		sdt_end = offsetof(struct ipr_ioa_dump, sdt.entry) +
entry            8548 drivers/scsi/ipr.c 	if (rc || !(sdt.entry[0].flags & IPR_SDT_VALID_ENTRY) ||
entry            8557 drivers/scsi/ipr.c 		length = be32_to_cpu(sdt.entry[0].end_token);
entry            8559 drivers/scsi/ipr.c 		length = (be32_to_cpu(sdt.entry[0].end_token) -
entry            8560 drivers/scsi/ipr.c 			  be32_to_cpu(sdt.entry[0].start_token)) &
entry            8569 drivers/scsi/ipr.c 					be32_to_cpu(sdt.entry[0].start_token),
entry            1262 drivers/scsi/ipr.h 	struct ipr_sdt_entry entry[IPR_FMT3_NUM_SDT_ENTRIES];
entry            1267 drivers/scsi/ipr.h 	struct ipr_sdt_entry entry[1];
entry              28 drivers/scsi/lpfc/lpfc_scsi.h #define list_remove_head(list, entry, type, member)		\
entry              30 drivers/scsi/lpfc/lpfc_scsi.h 	entry = NULL;						\
entry              32 drivers/scsi/lpfc/lpfc_scsi.h 		entry = list_entry((list)->next, type, member);	\
entry              33 drivers/scsi/lpfc/lpfc_scsi.h 		list_del_init(&entry->member);			\
entry            3352 drivers/scsi/lpfc/lpfc_sli.c 	IOCB_t *entry = NULL;
entry            3389 drivers/scsi/lpfc/lpfc_sli.c 		entry = lpfc_resp_iocb(phba, pring);
entry            3395 drivers/scsi/lpfc/lpfc_sli.c 		lpfc_sli_pcimem_bcopy((uint32_t *) entry,
entry            3736 drivers/scsi/lpfc/lpfc_sli.c 	IOCB_t *entry;
entry            3787 drivers/scsi/lpfc/lpfc_sli.c 		entry = lpfc_resp_iocb(phba, pring);
entry            3797 drivers/scsi/lpfc/lpfc_sli.c 		lpfc_sli_pcimem_bcopy(entry, &rspiocbp->iocb,
entry            13021 drivers/scsi/lpfc/lpfc_sli.c lpfc_cq_event_setup(struct lpfc_hba *phba, void *entry, int size)
entry            13034 drivers/scsi/lpfc/lpfc_sli.c 	memcpy(&cq_event->cqe, entry, size);
entry             139 drivers/scsi/mvsas/mv_init.c 	list_for_each_entry(mwq, &mvi->wq_list, entry)
entry             814 drivers/scsi/mvsas/mv_sas.c 	list_add_tail(&slot->entry, &tei.port->list);
entry             907 drivers/scsi/mvsas/mv_sas.c 	list_del_init(&slot->entry);
entry            1844 drivers/scsi/mvsas/mv_sas.c 	list_for_each_entry_safe(slot, slot2, &port->list, entry) {
entry            1915 drivers/scsi/mvsas/mv_sas.c 	list_del(&mwq->entry);
entry            1931 drivers/scsi/mvsas/mv_sas.c 		list_add_tail(&mwq->entry, &mvi->wq_list);
entry             307 drivers/scsi/mvsas/mv_sas.h 	struct list_head entry;
entry             414 drivers/scsi/mvsas/mv_sas.h 	struct list_head entry;
entry             150 drivers/scsi/mvumi.c 	INIT_LIST_HEAD(&res->entry);
entry             151 drivers/scsi/mvumi.c 	list_add_tail(&res->entry, &mhba->res_list);
entry             160 drivers/scsi/mvumi.c 	list_for_each_entry_safe(res, tmp, &mhba->res_list, entry) {
entry             174 drivers/scsi/mvumi.c 		list_del(&res->entry);
entry             212 drivers/scsi/mvumi.h 	struct list_head entry;
entry              51 drivers/scsi/myrb.c 	struct myrb_devstate_name_entry *entry = myrb_devstate_name_list;
entry              55 drivers/scsi/myrb.c 		if (entry[i].state == state)
entry              56 drivers/scsi/myrb.c 			return entry[i].name;
entry              75 drivers/scsi/myrb.c 	struct myrb_raidlevel_name_entry *entry = myrb_raidlevel_name_list;
entry              79 drivers/scsi/myrb.c 		if (entry[i].level == level)
entry              80 drivers/scsi/myrb.c 			return entry[i].name;
entry            3499 drivers/scsi/myrb.c 		const struct pci_device_id *entry)
entry            3502 drivers/scsi/myrb.c 		(struct myrb_privdata *)entry->driver_data;
entry            3560 drivers/scsi/myrb.c static int myrb_probe(struct pci_dev *dev, const struct pci_device_id *entry)
entry            3565 drivers/scsi/myrb.c 	cb = myrb_detect(dev, entry);
entry              49 drivers/scsi/myrs.c 	struct myrs_devstate_name_entry *entry = myrs_devstate_name_list;
entry              53 drivers/scsi/myrs.c 		if (entry[i].state == state)
entry              54 drivers/scsi/myrs.c 			return entry[i].name;
entry              80 drivers/scsi/myrs.c 	struct myrs_raid_level_name_entry *entry = myrs_raid_level_name_list;
entry              84 drivers/scsi/myrs.c 		if (entry[i].level == level)
entry              85 drivers/scsi/myrs.c 			return entry[i].name;
entry            1938 drivers/scsi/myrs.c 		const struct pci_device_id *entry)
entry            2289 drivers/scsi/myrs.c 		const struct pci_device_id *entry)
entry            2292 drivers/scsi/myrs.c 		(struct myrs_privdata *)entry->driver_data;
entry            2297 drivers/scsi/myrs.c 	cs = myrs_alloc_host(pdev, entry);
entry            3172 drivers/scsi/myrs.c myrs_probe(struct pci_dev *dev, const struct pci_device_id *entry)
entry            3177 drivers/scsi/myrs.c 	cs = myrs_detect(dev, entry);
entry            2225 drivers/scsi/nsp32.c 	int               entry;
entry            2287 drivers/scsi/nsp32.c 		entry = nsp32_search_period_entry(data, target, get_period);
entry            2289 drivers/scsi/nsp32.c 		if (entry < 0) {
entry            2300 drivers/scsi/nsp32.c 		nsp32_set_sync_entry(data, target, entry, get_offset);
entry            2318 drivers/scsi/nsp32.c 		entry = nsp32_search_period_entry(data, target, get_period);
entry            2320 drivers/scsi/nsp32.c 		if (get_offset == ASYNC_OFFSET || entry < 0) {
entry            2324 drivers/scsi/nsp32.c 			nsp32_set_sync_entry(data, target, entry, get_offset);
entry            2427 drivers/scsi/nsp32.c 				 int            entry,
entry            2432 drivers/scsi/nsp32.c 	period      = data->synct[entry].period_num;
entry            2433 drivers/scsi/nsp32.c 	ackwidth    = data->synct[entry].ackwidth;
entry            2434 drivers/scsi/nsp32.c 	sample_rate = data->synct[entry].sample_rate;
entry            2987 drivers/scsi/nsp32.c 	int           entry;
entry            3036 drivers/scsi/nsp32.c 			entry = nsp32_search_period_entry(data, target, ret);
entry            3037 drivers/scsi/nsp32.c 			if (entry < 0) {
entry            3039 drivers/scsi/nsp32.c 				entry = 0;
entry            3041 drivers/scsi/nsp32.c 			target->limit_entry = entry;
entry            3073 drivers/scsi/nsp32.c 	int           entry, val;
entry            3105 drivers/scsi/nsp32.c 		entry = nsp32_search_period_entry(data, target, val);
entry            3106 drivers/scsi/nsp32.c 		if (entry < 0 || trans_mode == ULTRA20M_MODE) {
entry            3108 drivers/scsi/nsp32.c 			entry = 0;
entry            3110 drivers/scsi/nsp32.c 		target->limit_entry = entry;
entry             298 drivers/scsi/pm8001/pm8001_sas.h 	struct list_head	entry;
entry              93 drivers/scsi/qla2xxx/qla_bsg.c 	pri_entry = &pri_cfg->entry[0];
entry             544 drivers/scsi/qla2xxx/qla_def.h 			struct	list_head   entry;
entry            2650 drivers/scsi/qla2xxx/qla_def.h 	struct ct_fdmi_hba_attr entry[FDMI_HBA_ATTR_COUNT];
entry            2679 drivers/scsi/qla2xxx/qla_def.h 	struct ct_fdmiv2_hba_attr entry[FDMIV2_HBA_ATTR_COUNT];
entry            2746 drivers/scsi/qla2xxx/qla_def.h 	struct ct_fdmiv2_port_attr entry[FDMIV2_PORT_ATTR_COUNT];
entry            2764 drivers/scsi/qla2xxx/qla_def.h 	struct ct_fdmi_port_attr entry[FDMI_PORT_ATTR_COUNT];
entry            3249 drivers/scsi/qla2xxx/qla_def.h 	uint16_t entry;
entry            2077 drivers/scsi/qla2xxx/qla_fw.h 	struct qla_fcp_prio_entry entry[1];     /* fcp priority entries  */
entry            1074 drivers/scsi/qla2xxx/qla_gs.c 	uint8_t		*entry;
entry            1105 drivers/scsi/qla2xxx/qla_gs.c 			entry = &sns_cmd->p.gid_data[(i * 4) + 16];
entry            1106 drivers/scsi/qla2xxx/qla_gs.c 			list[i].d_id.b.domain = entry[1];
entry            1107 drivers/scsi/qla2xxx/qla_gs.c 			list[i].d_id.b.area = entry[2];
entry            1108 drivers/scsi/qla2xxx/qla_gs.c 			list[i].d_id.b.al_pa = entry[3];
entry            1111 drivers/scsi/qla2xxx/qla_gs.c 			if (entry[0] & BIT_7) {
entry            1112 drivers/scsi/qla2xxx/qla_gs.c 				list[i].d_id.b.rsvd_1 = entry[0];
entry            3988 drivers/scsi/qla2xxx/qla_init.c 			    msix->entry);
entry            3989 drivers/scsi/qla2xxx/qla_init.c 			icb->msix = cpu_to_le16(msix->entry);
entry            8741 drivers/scsi/qla2xxx/qla_init.c 	pri_entry = &ha->fcp_prio_cfg->entry[0];
entry            3513 drivers/scsi/qla2xxx/qla_isr.c 		qentry->entry = i;
entry            4432 drivers/scsi/qla2xxx/qla_mbx.c 	mcp->mb[14] = rsp->msix->entry;
entry            1102 drivers/scsi/qla2xxx/qla_sup.c 	struct qla_npiv_entry *entry;
entry            1138 drivers/scsi/qla2xxx/qla_sup.c 	cnt = (sizeof(hdr) + le16_to_cpu(hdr.entries) * sizeof(*entry)) >> 1;
entry            1150 drivers/scsi/qla2xxx/qla_sup.c 	entry = data + sizeof(struct qla_npiv_header);
entry            1152 drivers/scsi/qla2xxx/qla_sup.c 	for (i = 0; cnt; cnt--, entry++, i++) {
entry            1157 drivers/scsi/qla2xxx/qla_sup.c 		memcpy(&ha->npiv_info[i], entry, sizeof(struct qla_npiv_entry));
entry            1159 drivers/scsi/qla2xxx/qla_sup.c 		flags = le16_to_cpu(entry->flags);
entry            1169 drivers/scsi/qla2xxx/qla_sup.c 		vid.port_name = wwn_to_u64(entry->port_name);
entry            1170 drivers/scsi/qla2xxx/qla_sup.c 		vid.node_name = wwn_to_u64(entry->node_name);
entry            1175 drivers/scsi/qla2xxx/qla_sup.c 		    le16_to_cpu(entry->vf_id),
entry            1176 drivers/scsi/qla2xxx/qla_sup.c 		    entry->q_qos, entry->f_qos);
entry            3623 drivers/scsi/qla2xxx/qla_sup.c 	ha->isp_ops->read_optrom(vha, &ha->fcp_prio_cfg->entry[0],
entry             377 drivers/scsi/qla2xxx/qla_target.c 		struct imm_ntfy_from_isp *entry =
entry             382 drivers/scsi/qla2xxx/qla_target.c 		if ((entry->u.isp24.vp_index != 0xFF) &&
entry             383 drivers/scsi/qla2xxx/qla_target.c 		    (entry->u.isp24.nport_handle != 0xFFFF)) {
entry             385 drivers/scsi/qla2xxx/qla_target.c 			    entry->u.isp24.vp_index);
entry             391 drivers/scsi/qla2xxx/qla_target.c 				    vha->vp_idx, entry->u.isp24.vp_index);
entry             406 drivers/scsi/qla2xxx/qla_target.c 		struct abts_recv_from_24xx *entry =
entry             409 drivers/scsi/qla2xxx/qla_target.c 			entry->vp_index);
entry             416 drivers/scsi/qla2xxx/qla_target.c 			    vha->vp_idx, entry->vp_index);
entry             450 drivers/scsi/qla2xxx/qla_target.c 		struct ctio7_from_24xx *entry = (struct ctio7_from_24xx *)pkt;
entry             452 drivers/scsi/qla2xxx/qla_target.c 		    entry->vp_index);
entry             457 drivers/scsi/qla2xxx/qla_target.c 			    vha->vp_idx, entry->vp_index);
entry             467 drivers/scsi/qla2xxx/qla_target.c 		struct imm_ntfy_from_isp *entry =
entry             470 drivers/scsi/qla2xxx/qla_target.c 		host = qlt_find_host_by_vp_idx(vha, entry->u.isp24.vp_index);
entry             475 drivers/scsi/qla2xxx/qla_target.c 			    vha->vp_idx, entry->u.isp24.vp_index);
entry             485 drivers/scsi/qla2xxx/qla_target.c 		struct nack_to_isp *entry = (struct nack_to_isp *)pkt;
entry             487 drivers/scsi/qla2xxx/qla_target.c 		if (0xFF != entry->u.isp24.vp_index) {
entry             489 drivers/scsi/qla2xxx/qla_target.c 			    entry->u.isp24.vp_index);
entry             496 drivers/scsi/qla2xxx/qla_target.c 				    entry->u.isp24.vp_index);
entry             506 drivers/scsi/qla2xxx/qla_target.c 		struct abts_recv_from_24xx *entry =
entry             509 drivers/scsi/qla2xxx/qla_target.c 		    entry->vp_index);
entry             514 drivers/scsi/qla2xxx/qla_target.c 			    "vp_index %d\n", vha->vp_idx, entry->vp_index);
entry             523 drivers/scsi/qla2xxx/qla_target.c 		struct abts_resp_to_24xx *entry =
entry             526 drivers/scsi/qla2xxx/qla_target.c 		    entry->vp_index);
entry             531 drivers/scsi/qla2xxx/qla_target.c 			    "vp_index %d\n", vha->vp_idx, entry->vp_index);
entry            1880 drivers/scsi/qla2xxx/qla_target.c 	struct abts_recv_from_24xx *entry;
entry            1892 drivers/scsi/qla2xxx/qla_target.c 		entry = &mcmd->orig_iocb.abts;
entry            1895 drivers/scsi/qla2xxx/qla_target.c 		entry = (struct abts_recv_from_24xx *)pkt;
entry            1904 drivers/scsi/qla2xxx/qla_target.c 	ctio->nport_handle = entry->nport_handle;
entry            1908 drivers/scsi/qla2xxx/qla_target.c 	ctio->exchange_addr = entry->exchange_addr_to_abort;
entry            1912 drivers/scsi/qla2xxx/qla_target.c 		ctio->initiator_id = entry->fcp_hdr_le.s_id;
entry            1919 drivers/scsi/qla2xxx/qla_target.c 		ctio->initiator_id = entry->fcp_hdr_le.d_id;
entry            1925 drivers/scsi/qla2xxx/qla_target.c 	ctio->u.status1.ox_id = entry->fcp_hdr_le.ox_id;
entry            1944 drivers/scsi/qla2xxx/qla_target.c 		    (struct abts_recv_from_24xx *)entry, FCP_TMF_CMPL, true);
entry            5655 drivers/scsi/qla2xxx/qla_target.c     struct qla_qpair *qpair, struct abts_resp_from_24xx_fw *entry)
entry            5666 drivers/scsi/qla2xxx/qla_target.c 	if (qpair->retry_term_exchg_addr == entry->exchange_addr_to_abort &&
entry            5676 drivers/scsi/qla2xxx/qla_target.c 			    vha, 0xffff, (uint8_t *)entry, sizeof(*entry));
entry            5687 drivers/scsi/qla2xxx/qla_target.c 		qpair->retry_term_exchg_addr = entry->exchange_addr_to_abort;
entry            5699 drivers/scsi/qla2xxx/qla_target.c 	struct abts_resp_from_24xx_fw *entry =
entry            5719 drivers/scsi/qla2xxx/qla_target.c 	    entry->compl_status);
entry            5721 drivers/scsi/qla2xxx/qla_target.c 	if (le16_to_cpu(entry->compl_status) != ABTS_RESP_COMPL_SUCCESS) {
entry            5722 drivers/scsi/qla2xxx/qla_target.c 		if ((entry->error_subcode1 == 0x1E) &&
entry            5723 drivers/scsi/qla2xxx/qla_target.c 		    (entry->error_subcode2 == 0)) {
entry            5724 drivers/scsi/qla2xxx/qla_target.c 			if (qlt_chk_unresolv_exchg(vha, rsp->qpair, entry)) {
entry            5733 drivers/scsi/qla2xxx/qla_target.c 			    vha->vp_idx, entry->compl_status,
entry            5734 drivers/scsi/qla2xxx/qla_target.c 			    entry->error_subcode1,
entry            5735 drivers/scsi/qla2xxx/qla_target.c 			    entry->error_subcode2);
entry            5766 drivers/scsi/qla2xxx/qla_target.c 		struct ctio7_from_24xx *entry = (struct ctio7_from_24xx *)pkt;
entry            5768 drivers/scsi/qla2xxx/qla_target.c 		qlt_do_ctio_completion(vha, rsp, entry->handle,
entry            5769 drivers/scsi/qla2xxx/qla_target.c 		    le16_to_cpu(entry->status)|(pkt->entry_status << 16),
entry            5770 drivers/scsi/qla2xxx/qla_target.c 		    entry);
entry            5826 drivers/scsi/qla2xxx/qla_target.c 		struct ctio_to_2xxx *entry = (struct ctio_to_2xxx *)pkt;
entry            5828 drivers/scsi/qla2xxx/qla_target.c 		qlt_do_ctio_completion(vha, rsp, entry->handle,
entry            5829 drivers/scsi/qla2xxx/qla_target.c 		    le16_to_cpu(entry->status)|(pkt->entry_status << 16),
entry            5830 drivers/scsi/qla2xxx/qla_target.c 		    entry);
entry            5836 drivers/scsi/qla2xxx/qla_target.c 		struct ctio_to_2xxx *entry = (struct ctio_to_2xxx *)pkt;
entry            5838 drivers/scsi/qla2xxx/qla_target.c 		qlt_do_ctio_completion(vha, rsp, entry->handle,
entry            5839 drivers/scsi/qla2xxx/qla_target.c 		    le16_to_cpu(entry->status)|(pkt->entry_status << 16),
entry            5840 drivers/scsi/qla2xxx/qla_target.c 		    entry);
entry            5851 drivers/scsi/qla2xxx/qla_target.c 			struct nack_to_isp *entry = (struct nack_to_isp *)pkt;
entry            5855 drivers/scsi/qla2xxx/qla_target.c 			    le16_to_cpu(entry->u.isp2x.seq_id),
entry            5856 drivers/scsi/qla2xxx/qla_target.c 			    le16_to_cpu(entry->u.isp2x.status));
entry            5858 drivers/scsi/qla2xxx/qla_target.c 			if (entry->u.isp2x.status !=
entry            5863 drivers/scsi/qla2xxx/qla_target.c 				    le16_to_cpu(entry->u.isp2x.status));
entry            6822 drivers/scsi/qla2xxx/qla_target.c 				icb->msix_atio = cpu_to_le16(msix->entry);
entry            6827 drivers/scsi/qla2xxx/qla_target.c 			    msix->entry);
entry             280 drivers/scsi/qla4xxx/ql4_def.h         struct aen entry[MAX_AEN_ENTRIES];
entry             700 drivers/scsi/qla4xxx/ql4_isr.c 				ha->aen_log.entry[ha->aen_log.count].mbox_sts[i] =
entry             120 drivers/scsi/sg.c 	struct list_head entry;	/* list entry */
entry             875 drivers/scsi/sg.c 	list_for_each_entry(srp, &sfp->rq_list, entry) {
entry            1002 drivers/scsi/sg.c 		list_for_each_entry(srp, &sfp->rq_list, entry) {
entry            1016 drivers/scsi/sg.c 		list_for_each_entry(srp, &sfp->rq_list, entry) {
entry            1188 drivers/scsi/sg.c 	list_for_each_entry(srp, &sfp->rq_list, entry) {
entry            2083 drivers/scsi/sg.c 	list_for_each_entry(resp, &sfp->rq_list, entry) {
entry            2119 drivers/scsi/sg.c 	list_add_tail(&rp->entry, &sfp->rq_list);
entry            2137 drivers/scsi/sg.c 	if (!list_empty(&srp->entry)) {
entry            2138 drivers/scsi/sg.c 		list_del(&srp->entry);
entry            2205 drivers/scsi/sg.c 		srp = list_first_entry(&sfp->rq_list, Sg_request, entry);
entry            2207 drivers/scsi/sg.c 		list_del(&srp->entry);
entry            2539 drivers/scsi/sg.c 		list_for_each_entry(srp, &fp->rq_list, entry) {
entry              26 drivers/sh/intc/virq-debugfs.c 		struct intc_map_entry *entry = intc_irq_xlate_get(i);
entry              27 drivers/sh/intc/virq-debugfs.c 		struct intc_desc_int *desc = entry->desc;
entry              33 drivers/sh/intc/virq-debugfs.c 		seq_printf(m, "0x%05x  ", entry->enum_id);
entry              27 drivers/sh/intc/virq.c #define for_each_virq(entry, head) \
entry              28 drivers/sh/intc/virq.c 	for (entry = head; entry; entry = entry->next)
entry              86 drivers/sh/intc/virq.c 	struct intc_virq_list *entry;
entry              90 drivers/sh/intc/virq.c 	for_each_virq(entry, irq_get_handler_data(irq)) {
entry              91 drivers/sh/intc/virq.c 		if (entry->irq == virq)
entry              93 drivers/sh/intc/virq.c 		last = &entry->next;
entry              96 drivers/sh/intc/virq.c 	entry = kzalloc(sizeof(struct intc_virq_list), GFP_ATOMIC);
entry              97 drivers/sh/intc/virq.c 	if (!entry)
entry             100 drivers/sh/intc/virq.c 	entry->irq = virq;
entry             103 drivers/sh/intc/virq.c 		*last = entry;
entry             105 drivers/sh/intc/virq.c 		irq_set_handler_data(irq, entry);
entry             115 drivers/sh/intc/virq.c 	struct intc_virq_list *entry, *vlist = irq_data_get_irq_handler_data(data);
entry             120 drivers/sh/intc/virq.c 	for_each_virq(entry, vlist) {
entry             122 drivers/sh/intc/virq.c 		struct irq_desc *vdesc = irq_to_desc(entry->irq);
entry             165 drivers/sh/intc/virq.c 		struct intc_subgroup_entry *entry;
entry             171 drivers/sh/intc/virq.c 		entry = kmalloc(sizeof(*entry), GFP_NOWAIT);
entry             172 drivers/sh/intc/virq.c 		if (!entry)
entry             175 drivers/sh/intc/virq.c 		entry->pirq = pirq;
entry             176 drivers/sh/intc/virq.c 		entry->enum_id = subgroup->enum_ids[i];
entry             177 drivers/sh/intc/virq.c 		entry->handle = intc_subgroup_data(subgroup, d, i);
entry             179 drivers/sh/intc/virq.c 		err = radix_tree_insert(&d->tree, entry->enum_id, entry);
entry             183 drivers/sh/intc/virq.c 		radix_tree_tag_set(&d->tree, entry->enum_id,
entry             216 drivers/sh/intc/virq.c 		struct intc_subgroup_entry *entry;
entry             219 drivers/sh/intc/virq.c 		entry = radix_tree_deref_slot((void **)entries[i]);
entry             220 drivers/sh/intc/virq.c 		if (unlikely(!entry))
entry             222 drivers/sh/intc/virq.c 		if (radix_tree_deref_retry(entry))
entry             234 drivers/sh/intc/virq.c 			irq, entry->pirq);
entry             236 drivers/sh/intc/virq.c 		intc_irq_xlate_set(irq, entry->enum_id, d);
entry             238 drivers/sh/intc/virq.c 		irq_set_chip_and_handler_name(irq, irq_get_chip(entry->pirq),
entry             240 drivers/sh/intc/virq.c 		irq_set_chip_data(irq, irq_get_chip_data(entry->pirq));
entry             242 drivers/sh/intc/virq.c 		irq_set_handler_data(irq, (void *)entry->handle);
entry             250 drivers/sh/intc/virq.c 		add_virq_to_pirq(entry->pirq, irq);
entry             251 drivers/sh/intc/virq.c 		irq_set_chained_handler(entry->pirq, intc_virq_handler);
entry             253 drivers/sh/intc/virq.c 		radix_tree_tag_clear(&d->tree, entry->enum_id,
entry             118 drivers/soc/fsl/qe/qe_common.c 	struct muram_block *entry;
entry             129 drivers/soc/fsl/qe/qe_common.c 	entry = kmalloc(sizeof(*entry), GFP_ATOMIC);
entry             130 drivers/soc/fsl/qe/qe_common.c 	if (!entry)
entry             132 drivers/soc/fsl/qe/qe_common.c 	entry->start = start;
entry             133 drivers/soc/fsl/qe/qe_common.c 	entry->size = size;
entry             134 drivers/soc/fsl/qe/qe_common.c 	list_add(&entry->head, &muram_block_list);
entry             170 drivers/soc/qcom/smem.c 	struct smem_ptable_entry entry[];
entry             403 drivers/soc/qcom/smem.c 	struct smem_global_entry *entry;
entry             407 drivers/soc/qcom/smem.c 	entry = &header->toc[item];
entry             408 drivers/soc/qcom/smem.c 	if (entry->allocated)
entry             415 drivers/soc/qcom/smem.c 	entry->offset = header->free_offset;
entry             416 drivers/soc/qcom/smem.c 	entry->size = cpu_to_le32(size);
entry             424 drivers/soc/qcom/smem.c 	entry->allocated = cpu_to_le32(1);
entry             487 drivers/soc/qcom/smem.c 	struct smem_global_entry *entry;
entry             492 drivers/soc/qcom/smem.c 	entry = &header->toc[item];
entry             493 drivers/soc/qcom/smem.c 	if (!entry->allocated)
entry             496 drivers/soc/qcom/smem.c 	aux_base = le32_to_cpu(entry->aux_base) & AUX_BASE_MASK;
entry             503 drivers/soc/qcom/smem.c 				*size = le32_to_cpu(entry->size);
entry             504 drivers/soc/qcom/smem.c 			return region->virt_base + le32_to_cpu(entry->offset);
entry             712 drivers/soc/qcom/smem.c 	info = (struct smem_info *)&ptable->entry[ptable->num_entries];
entry             726 drivers/soc/qcom/smem.c 		struct smem_ptable_entry *entry, u16 host0, u16 host1)
entry             731 drivers/soc/qcom/smem.c 	header = smem->regions[0].virt_base + le32_to_cpu(entry->offset);
entry             752 drivers/soc/qcom/smem.c 	if (size != le32_to_cpu(entry->size)) {
entry             754 drivers/soc/qcom/smem.c 			size, le32_to_cpu(entry->size));
entry             770 drivers/soc/qcom/smem.c 	struct smem_ptable_entry *entry;
entry             785 drivers/soc/qcom/smem.c 		entry = &ptable->entry[i];
entry             786 drivers/soc/qcom/smem.c 		if (!le32_to_cpu(entry->offset))
entry             788 drivers/soc/qcom/smem.c 		if (!le32_to_cpu(entry->size))
entry             791 drivers/soc/qcom/smem.c 		if (le16_to_cpu(entry->host0) != SMEM_GLOBAL_HOST)
entry             794 drivers/soc/qcom/smem.c 		if (le16_to_cpu(entry->host1) == SMEM_GLOBAL_HOST) {
entry             805 drivers/soc/qcom/smem.c 	header = qcom_smem_partition_header(smem, entry,
entry             811 drivers/soc/qcom/smem.c 	smem->global_cacheline = le32_to_cpu(entry->cacheline);
entry             820 drivers/soc/qcom/smem.c 	struct smem_ptable_entry *entry;
entry             831 drivers/soc/qcom/smem.c 		entry = &ptable->entry[i];
entry             832 drivers/soc/qcom/smem.c 		if (!le32_to_cpu(entry->offset))
entry             834 drivers/soc/qcom/smem.c 		if (!le32_to_cpu(entry->size))
entry             837 drivers/soc/qcom/smem.c 		host0 = le16_to_cpu(entry->host0);
entry             838 drivers/soc/qcom/smem.c 		host1 = le16_to_cpu(entry->host1);
entry             856 drivers/soc/qcom/smem.c 		header = qcom_smem_partition_header(smem, entry, host0, host1);
entry             861 drivers/soc/qcom/smem.c 		smem->cacheline[remote_host] = le32_to_cpu(entry->cacheline);
entry             175 drivers/soc/qcom/smp2p.c 	struct smp2p_entry *entry;
entry             202 drivers/soc/qcom/smp2p.c 		list_for_each_entry(entry, &smp2p->inbound, node) {
entry             204 drivers/soc/qcom/smp2p.c 			if (!strcmp(buf, entry->name)) {
entry             205 drivers/soc/qcom/smp2p.c 				entry->value = &in->entries[i].value;
entry             213 drivers/soc/qcom/smp2p.c 	list_for_each_entry(entry, &smp2p->inbound, node) {
entry             215 drivers/soc/qcom/smp2p.c 		if (!entry->value)
entry             218 drivers/soc/qcom/smp2p.c 		val = readl(entry->value);
entry             220 drivers/soc/qcom/smp2p.c 		status = val ^ entry->last_value;
entry             221 drivers/soc/qcom/smp2p.c 		entry->last_value = val;
entry             227 drivers/soc/qcom/smp2p.c 		for_each_set_bit(i, entry->irq_enabled, 32) {
entry             231 drivers/soc/qcom/smp2p.c 			if ((val & BIT(i) && test_bit(i, entry->irq_rising)) ||
entry             232 drivers/soc/qcom/smp2p.c 			    (!(val & BIT(i)) && test_bit(i, entry->irq_falling))) {
entry             233 drivers/soc/qcom/smp2p.c 				irq_pin = irq_find_mapping(entry->domain, i);
entry             244 drivers/soc/qcom/smp2p.c 	struct smp2p_entry *entry = irq_data_get_irq_chip_data(irqd);
entry             247 drivers/soc/qcom/smp2p.c 	clear_bit(irq, entry->irq_enabled);
entry             252 drivers/soc/qcom/smp2p.c 	struct smp2p_entry *entry = irq_data_get_irq_chip_data(irqd);
entry             255 drivers/soc/qcom/smp2p.c 	set_bit(irq, entry->irq_enabled);
entry             260 drivers/soc/qcom/smp2p.c 	struct smp2p_entry *entry = irq_data_get_irq_chip_data(irqd);
entry             267 drivers/soc/qcom/smp2p.c 		set_bit(irq, entry->irq_rising);
entry             269 drivers/soc/qcom/smp2p.c 		clear_bit(irq, entry->irq_rising);
entry             272 drivers/soc/qcom/smp2p.c 		set_bit(irq, entry->irq_falling);
entry             274 drivers/soc/qcom/smp2p.c 		clear_bit(irq, entry->irq_falling);
entry             290 drivers/soc/qcom/smp2p.c 	struct smp2p_entry *entry = d->host_data;
entry             293 drivers/soc/qcom/smp2p.c 	irq_set_chip_data(irq, entry);
entry             306 drivers/soc/qcom/smp2p.c 				    struct smp2p_entry *entry,
entry             309 drivers/soc/qcom/smp2p.c 	entry->domain = irq_domain_add_linear(node, 32, &smp2p_irq_ops, entry);
entry             310 drivers/soc/qcom/smp2p.c 	if (!entry->domain) {
entry             320 drivers/soc/qcom/smp2p.c 	struct smp2p_entry *entry = data;
entry             324 drivers/soc/qcom/smp2p.c 	spin_lock(&entry->lock);
entry             325 drivers/soc/qcom/smp2p.c 	val = orig = readl(entry->value);
entry             328 drivers/soc/qcom/smp2p.c 	writel(val, entry->value);
entry             329 drivers/soc/qcom/smp2p.c 	spin_unlock(&entry->lock);
entry             332 drivers/soc/qcom/smp2p.c 		qcom_smp2p_kick(entry->smp2p);
entry             342 drivers/soc/qcom/smp2p.c 				     struct smp2p_entry *entry,
entry             349 drivers/soc/qcom/smp2p.c 	strlcpy(buf, entry->name, SMP2P_MAX_ENTRY_NAME);
entry             353 drivers/soc/qcom/smp2p.c 	entry->value = &out->entries[out->valid_entries].value;
entry             357 drivers/soc/qcom/smp2p.c 	entry->state = qcom_smem_state_register(node, &smp2p_state_ops, entry);
entry             358 drivers/soc/qcom/smp2p.c 	if (IS_ERR(entry->state)) {
entry             360 drivers/soc/qcom/smp2p.c 		return PTR_ERR(entry->state);
entry             443 drivers/soc/qcom/smp2p.c 	struct smp2p_entry *entry;
entry             501 drivers/soc/qcom/smp2p.c 		entry = devm_kzalloc(&pdev->dev, sizeof(*entry), GFP_KERNEL);
entry             502 drivers/soc/qcom/smp2p.c 		if (!entry) {
entry             507 drivers/soc/qcom/smp2p.c 		entry->smp2p = smp2p;
entry             508 drivers/soc/qcom/smp2p.c 		spin_lock_init(&entry->lock);
entry             510 drivers/soc/qcom/smp2p.c 		ret = of_property_read_string(node, "qcom,entry-name", &entry->name);
entry             515 drivers/soc/qcom/smp2p.c 			ret = qcom_smp2p_inbound_entry(smp2p, entry, node);
entry             519 drivers/soc/qcom/smp2p.c 			list_add(&entry->node, &smp2p->inbound);
entry             521 drivers/soc/qcom/smp2p.c 			ret = qcom_smp2p_outbound_entry(smp2p, entry, node);
entry             525 drivers/soc/qcom/smp2p.c 			list_add(&entry->node, &smp2p->outbound);
entry             545 drivers/soc/qcom/smp2p.c 	list_for_each_entry(entry, &smp2p->inbound, node)
entry             546 drivers/soc/qcom/smp2p.c 		irq_domain_remove(entry->domain);
entry             548 drivers/soc/qcom/smp2p.c 	list_for_each_entry(entry, &smp2p->outbound, node)
entry             549 drivers/soc/qcom/smp2p.c 		qcom_smem_state_unregister(entry->state);
entry             566 drivers/soc/qcom/smp2p.c 	struct smp2p_entry *entry;
entry             568 drivers/soc/qcom/smp2p.c 	list_for_each_entry(entry, &smp2p->inbound, node)
entry             569 drivers/soc/qcom/smp2p.c 		irq_domain_remove(entry->domain);
entry             571 drivers/soc/qcom/smp2p.c 	list_for_each_entry(entry, &smp2p->outbound, node)
entry             572 drivers/soc/qcom/smp2p.c 		qcom_smem_state_unregister(entry->state);
entry             200 drivers/soc/qcom/smsm.c 	struct smsm_entry *entry = data;
entry             206 drivers/soc/qcom/smsm.c 	val = readl(entry->remote_state);
entry             207 drivers/soc/qcom/smsm.c 	changed = val ^ entry->last_value;
entry             208 drivers/soc/qcom/smsm.c 	entry->last_value = val;
entry             210 drivers/soc/qcom/smsm.c 	for_each_set_bit(i, entry->irq_enabled, 32) {
entry             215 drivers/soc/qcom/smsm.c 			if (test_bit(i, entry->irq_rising)) {
entry             216 drivers/soc/qcom/smsm.c 				irq_pin = irq_find_mapping(entry->domain, i);
entry             220 drivers/soc/qcom/smsm.c 			if (test_bit(i, entry->irq_falling)) {
entry             221 drivers/soc/qcom/smsm.c 				irq_pin = irq_find_mapping(entry->domain, i);
entry             239 drivers/soc/qcom/smsm.c 	struct smsm_entry *entry = irq_data_get_irq_chip_data(irqd);
entry             241 drivers/soc/qcom/smsm.c 	struct qcom_smsm *smsm = entry->smsm;
entry             244 drivers/soc/qcom/smsm.c 	if (entry->subscription) {
entry             245 drivers/soc/qcom/smsm.c 		val = readl(entry->subscription + smsm->local_host);
entry             247 drivers/soc/qcom/smsm.c 		writel(val, entry->subscription + smsm->local_host);
entry             250 drivers/soc/qcom/smsm.c 	clear_bit(irq, entry->irq_enabled);
entry             264 drivers/soc/qcom/smsm.c 	struct smsm_entry *entry = irq_data_get_irq_chip_data(irqd);
entry             266 drivers/soc/qcom/smsm.c 	struct qcom_smsm *smsm = entry->smsm;
entry             269 drivers/soc/qcom/smsm.c 	set_bit(irq, entry->irq_enabled);
entry             271 drivers/soc/qcom/smsm.c 	if (entry->subscription) {
entry             272 drivers/soc/qcom/smsm.c 		val = readl(entry->subscription + smsm->local_host);
entry             274 drivers/soc/qcom/smsm.c 		writel(val, entry->subscription + smsm->local_host);
entry             285 drivers/soc/qcom/smsm.c 	struct smsm_entry *entry = irq_data_get_irq_chip_data(irqd);
entry             292 drivers/soc/qcom/smsm.c 		set_bit(irq, entry->irq_rising);
entry             294 drivers/soc/qcom/smsm.c 		clear_bit(irq, entry->irq_rising);
entry             297 drivers/soc/qcom/smsm.c 		set_bit(irq, entry->irq_falling);
entry             299 drivers/soc/qcom/smsm.c 		clear_bit(irq, entry->irq_falling);
entry             321 drivers/soc/qcom/smsm.c 	struct smsm_entry *entry = d->host_data;
entry             324 drivers/soc/qcom/smsm.c 	irq_set_chip_data(irq, entry);
entry             382 drivers/soc/qcom/smsm.c 			      struct smsm_entry *entry,
entry             397 drivers/soc/qcom/smsm.c 					"smsm", (void *)entry);
entry             403 drivers/soc/qcom/smsm.c 	entry->domain = irq_domain_add_linear(node, 32, &smsm_irq_ops, entry);
entry             404 drivers/soc/qcom/smsm.c 	if (!entry->domain) {
entry             459 drivers/soc/qcom/smsm.c 	struct smsm_entry *entry;
entry             562 drivers/soc/qcom/smsm.c 		entry = &smsm->entries[id];
entry             564 drivers/soc/qcom/smsm.c 		entry->smsm = smsm;
entry             565 drivers/soc/qcom/smsm.c 		entry->remote_state = states + id;
entry             568 drivers/soc/qcom/smsm.c 		entry->subscription = intr_mask + id * smsm->num_hosts;
entry             569 drivers/soc/qcom/smsm.c 		writel(0, entry->subscription + smsm->local_host);
entry             571 drivers/soc/qcom/smsm.c 		ret = smsm_inbound_entry(smsm, entry, node);
entry            2701 drivers/spi/spi.c 	INIT_LIST_HEAD(&sres->entry);
entry            2720 drivers/spi/spi.c 	WARN_ON(!list_empty(&sres->entry));
entry            2734 drivers/spi/spi.c 	WARN_ON(!list_empty(&sres->entry));
entry            2735 drivers/spi/spi.c 	list_add_tail(&sres->entry, &message->resources);
entry            2748 drivers/spi/spi.c 	list_for_each_entry_safe_reverse(res, tmp, &message->resources, entry) {
entry            2752 drivers/spi/spi.c 		list_del(&res->entry);
entry             860 drivers/staging/android/vsoc.c 		vsoc_dev.msix_entries[i].entry = i;
entry             260 drivers/staging/comedi/drivers/daqboard2000.c static void db2k_write_acq_scan_list_entry(struct comedi_device *dev, u16 entry)
entry             262 drivers/staging/comedi/drivers/daqboard2000.c 	writew(entry & 0x00ff, dev->mmio + DB2K_REG_ACQ_SCAN_LIST_FIFO);
entry             263 drivers/staging/comedi/drivers/daqboard2000.c 	writew((entry >> 8) & 0x00ff,
entry             468 drivers/staging/comedi/drivers/me4000.c 	unsigned int entry;
entry             472 drivers/staging/comedi/drivers/me4000.c 	entry = chan | ME4000_AI_LIST_RANGE(range);
entry             491 drivers/staging/comedi/drivers/me4000.c 		entry |= ME4000_AI_LIST_INPUT_DIFFERENTIAL;
entry             494 drivers/staging/comedi/drivers/me4000.c 	entry |= ME4000_AI_LIST_LAST_ENTRY;
entry             501 drivers/staging/comedi/drivers/me4000.c 	outl(entry, dev->iobase + ME4000_AI_CHANNEL_LIST_REG);
entry             637 drivers/staging/comedi/drivers/me4000.c 		unsigned int entry;
entry             639 drivers/staging/comedi/drivers/me4000.c 		entry = chan | ME4000_AI_LIST_RANGE(range);
entry             642 drivers/staging/comedi/drivers/me4000.c 			entry |= ME4000_AI_LIST_INPUT_DIFFERENTIAL;
entry             645 drivers/staging/comedi/drivers/me4000.c 			entry |= ME4000_AI_LIST_LAST_ENTRY;
entry             647 drivers/staging/comedi/drivers/me4000.c 		outl(entry, dev->iobase + ME4000_AI_CHANNEL_LIST_REG);
entry             281 drivers/staging/exfat/exfat.h 	s32       entry;
entry             511 drivers/staging/exfat/exfat.h 	s32       entry;
entry             547 drivers/staging/exfat/exfat.h 				  s32 entry, u32 type, u32 start_clu, u64 size);
entry             549 drivers/staging/exfat/exfat.h 				  s32 entry, s32 num_entries,
entry             556 drivers/staging/exfat/exfat.h 				    struct chain_t *p_dir, s32 entry,
entry             559 drivers/staging/exfat/exfat.h 					       struct chain_t *p_dir, s32 entry,
entry             562 drivers/staging/exfat/exfat.h 				     struct chain_t *p_dir, s32 entry,
entry             845 drivers/staging/exfat/exfat.h s32 fat_init_dir_entry(struct super_block *sb, struct chain_t *p_dir, s32 entry,
entry             848 drivers/staging/exfat/exfat.h 			 s32 entry, u32 type, u32 start_clu, u64 size);
entry             850 drivers/staging/exfat/exfat.h 			   s32 entry, s32 num_entries,
entry             854 drivers/staging/exfat/exfat.h 			     s32 entry, s32 num_entries,
entry             865 drivers/staging/exfat/exfat.h 			  s32 entry, s32 order, s32 num_entries);
entry             867 drivers/staging/exfat/exfat.h 			    s32 entry, s32 order, s32 num_entries);
entry             869 drivers/staging/exfat/exfat.h s32 find_location(struct super_block *sb, struct chain_t *p_dir, s32 entry,
entry             874 drivers/staging/exfat/exfat.h 				  s32 entry, sector_t *sector);
entry             876 drivers/staging/exfat/exfat.h 					       struct chain_t *p_dir, s32 entry,
entry             895 drivers/staging/exfat/exfat.h 			  s32 entry, struct dentry_t *p_entry);
entry             897 drivers/staging/exfat/exfat.h 			    s32 entry, struct dentry_t *p_entry);
entry             901 drivers/staging/exfat/exfat.h 			 s32 entry);
entry             914 drivers/staging/exfat/exfat.h 				     struct chain_t *p_dir, s32 entry,
entry             917 drivers/staging/exfat/exfat.h 				       struct chain_t *p_dir, s32 entry,
entry             945 drivers/staging/exfat/exfat.h void remove_file(struct inode *inode, struct chain_t *p_dir, s32 entry);
entry            1239 drivers/staging/exfat/exfat_core.c s32 fat_init_dir_entry(struct super_block *sb, struct chain_t *p_dir, s32 entry,
entry            1245 drivers/staging/exfat/exfat_core.c 	dos_ep = (struct dos_dentry_t *)get_entry_in_dir(sb, p_dir, entry,
entry            1257 drivers/staging/exfat/exfat_core.c 			 s32 entry, u32 type, u32 start_clu, u64 size)
entry            1267 drivers/staging/exfat/exfat_core.c 	file_ep = (struct file_dentry_t *)get_entry_in_dir(sb, p_dir, entry,
entry            1272 drivers/staging/exfat/exfat_core.c 	strm_ep = (struct strm_dentry_t *)get_entry_in_dir(sb, p_dir, entry + 1,
entry            1287 drivers/staging/exfat/exfat_core.c 			      s32 entry, s32 num_entries,
entry            1298 drivers/staging/exfat/exfat_core.c 	dos_ep = (struct dos_dentry_t *)get_entry_in_dir(sb, p_dir, entry,
entry            1314 drivers/staging/exfat/exfat_core.c 									 entry - i,
entry            1325 drivers/staging/exfat/exfat_core.c 								 entry - i,
entry            1338 drivers/staging/exfat/exfat_core.c 				s32 entry, s32 num_entries,
entry            1349 drivers/staging/exfat/exfat_core.c 	file_ep = (struct file_dentry_t *)get_entry_in_dir(sb, p_dir, entry,
entry            1357 drivers/staging/exfat/exfat_core.c 	strm_ep = (struct strm_dentry_t *)get_entry_in_dir(sb, p_dir, entry + 1,
entry            1368 drivers/staging/exfat/exfat_core.c 								   entry + i,
entry            1378 drivers/staging/exfat/exfat_core.c 	update_dir_checksum(sb, p_dir, entry);
entry            1487 drivers/staging/exfat/exfat_core.c 		s32 entry, s32 order, s32 num_entries)
entry            1495 drivers/staging/exfat/exfat_core.c 		ep = get_entry_in_dir(sb, p_dir, entry - i, &sector);
entry            1505 drivers/staging/exfat/exfat_core.c 		s32 entry, s32 order, s32 num_entries)
entry            1513 drivers/staging/exfat/exfat_core.c 		ep = get_entry_in_dir(sb, p_dir, entry + i, &sector);
entry            1523 drivers/staging/exfat/exfat_core.c 			 s32 entry)
entry            1531 drivers/staging/exfat/exfat_core.c 	file_ep = (struct file_dentry_t *)get_entry_in_dir(sb, p_dir, entry,
entry            1543 drivers/staging/exfat/exfat_core.c 		ep = get_entry_in_dir(sb, p_dir, entry + i, NULL);
entry            1604 drivers/staging/exfat/exfat_core.c s32 find_location(struct super_block *sb, struct chain_t *p_dir, s32 entry,
entry            1612 drivers/staging/exfat/exfat_core.c 	off = entry << DENTRY_SIZE_BITS;
entry            1650 drivers/staging/exfat/exfat_core.c 				  s32 entry, sector_t *sector)
entry            1656 drivers/staging/exfat/exfat_core.c 	if (find_location(sb, p_dir, entry, &sec, &off) != FFS_SUCCESS)
entry            1688 drivers/staging/exfat/exfat_core.c 					       struct chain_t *p_dir, s32 entry,
entry            1708 drivers/staging/exfat/exfat_core.c 	byte_offset = entry << DENTRY_SIZE_BITS;
entry            1976 drivers/staging/exfat/exfat_core.c 		if (p_fs->hint_uentry.entry == -1)
entry            1983 drivers/staging/exfat/exfat_core.c 		dentry = p_fs->hint_uentry.entry;
entry            1985 drivers/staging/exfat/exfat_core.c 		p_fs->hint_uentry.entry = -1;
entry            2012 drivers/staging/exfat/exfat_core.c 				if (p_fs->hint_uentry.entry == -1) {
entry            2014 drivers/staging/exfat/exfat_core.c 					p_fs->hint_uentry.entry = dentry;
entry            2028 drivers/staging/exfat/exfat_core.c 				p_fs->hint_uentry.entry = -1;
entry            2101 drivers/staging/exfat/exfat_core.c 		if (p_fs->hint_uentry.entry == -1) {
entry            2103 drivers/staging/exfat/exfat_core.c 			p_fs->hint_uentry.entry = p_dir->size << (p_fs->cluster_size_bits - DENTRY_SIZE_BITS);
entry            2118 drivers/staging/exfat/exfat_core.c 						      fid->entry + 1, &sector);
entry            2126 drivers/staging/exfat/exfat_core.c 						    fid->entry);
entry            2278 drivers/staging/exfat/exfat_core.c 	p_fs->hint_uentry.entry = -1;
entry            2295 drivers/staging/exfat/exfat_core.c 				if (p_fs->hint_uentry.entry == -1) {
entry            2304 drivers/staging/exfat/exfat_core.c 						p_fs->hint_uentry.entry = dentry - (num_empty - 1);
entry            2352 drivers/staging/exfat/exfat_core.c 							p_fs->hint_uentry.entry = -1;
entry            2387 drivers/staging/exfat/exfat_core.c 			  s32 entry, struct dentry_t *p_entry)
entry            2397 drivers/staging/exfat/exfat_core.c 	for (entry--; entry >= 0; entry--) {
entry            2399 drivers/staging/exfat/exfat_core.c 								 entry, NULL);
entry            2417 drivers/staging/exfat/exfat_core.c 			    s32 entry, struct dentry_t *p_entry)
entry            2425 drivers/staging/exfat/exfat_core.c 	for (i = 0, entry++; i < file_ep->num_ext; i++, entry++) {
entry            2426 drivers/staging/exfat/exfat_core.c 		ext_ep = get_entry_in_dir(sb, p_dir, entry, NULL);
entry            2618 drivers/staging/exfat/exfat_core.c 				     struct chain_t *p_dir, s32 entry,
entry            2625 drivers/staging/exfat/exfat_core.c 	for (entry--, i = 1; entry >= 0; entry--, i++) {
entry            2626 drivers/staging/exfat/exfat_core.c 		ep = (struct ext_dentry_t *)get_entry_in_dir(sb, p_dir, entry,
entry            2645 drivers/staging/exfat/exfat_core.c 				       struct chain_t *p_dir, s32 entry,
entry            2653 drivers/staging/exfat/exfat_core.c 	es = get_entry_set_in_dir(sb, p_dir, entry, ES_ALL_ENTRIES, &ep);
entry            3291 drivers/staging/exfat/exfat_core.c 	fid->entry = dentry;
entry            3341 drivers/staging/exfat/exfat_core.c 	fid->entry = dentry;
entry            3355 drivers/staging/exfat/exfat_core.c void remove_file(struct inode *inode, struct chain_t *p_dir, s32 entry)
entry            3364 drivers/staging/exfat/exfat_core.c 	ep = get_entry_in_dir(sb, p_dir, entry, &sector);
entry            3371 drivers/staging/exfat/exfat_core.c 	num_entries = fs_func->count_ext_entries(sb, p_dir, entry, ep);
entry            3381 drivers/staging/exfat/exfat_core.c 	fs_func->delete_dir_entry(sb, p_dir, entry, 0, num_entries);
entry            3465 drivers/staging/exfat/exfat_core.c 		fid->entry = newentry;
entry            3591 drivers/staging/exfat/exfat_core.c 	fid->entry = newentry;
entry             588 drivers/staging/exfat/exfat_super.c 	fid->entry = dentry;
entry            1006 drivers/staging/exfat/exfat_super.c 		es = get_entry_set_in_dir(sb, &(fid->dir), fid->entry,
entry            1012 drivers/staging/exfat/exfat_super.c 		ep = get_entry_in_dir(sb, &(fid->dir), fid->entry, &sector);
entry            1137 drivers/staging/exfat/exfat_super.c 		es = get_entry_set_in_dir(sb, &fid->dir, fid->entry,
entry            1145 drivers/staging/exfat/exfat_super.c 		ep = get_entry_in_dir(sb, &(fid->dir), fid->entry, &sector);
entry            1250 drivers/staging/exfat/exfat_super.c 	dentry = fid->entry;
entry            1281 drivers/staging/exfat/exfat_super.c 		new_entry = new_fid->entry;
entry            1365 drivers/staging/exfat/exfat_super.c 	dentry = fid->entry;
entry            1429 drivers/staging/exfat/exfat_super.c 		    (fid->entry == -1)) {
entry            1441 drivers/staging/exfat/exfat_super.c 		es = get_entry_set_in_dir(sb, &(fid->dir), fid->entry,
entry            1448 drivers/staging/exfat/exfat_super.c 		ep = get_entry_in_dir(sb, &(fid->dir), fid->entry, &sector);
entry            1519 drivers/staging/exfat/exfat_super.c 		    (fid->entry == -1)) {
entry            1557 drivers/staging/exfat/exfat_super.c 		es = get_entry_set_in_dir(sb, &(fid->dir), fid->entry,
entry            1565 drivers/staging/exfat/exfat_super.c 		ep = get_entry_in_dir(sb, &(fid->dir), fid->entry, &sector);
entry            1601 drivers/staging/exfat/exfat_super.c 	p_fs->fs_func->get_uni_name_from_ext_entry(sb, &(fid->dir), fid->entry,
entry            1669 drivers/staging/exfat/exfat_super.c 		    (fid->entry == -1)) {
entry            1681 drivers/staging/exfat/exfat_super.c 		es = get_entry_set_in_dir(sb, &(fid->dir), fid->entry,
entry            1690 drivers/staging/exfat/exfat_super.c 		ep = get_entry_in_dir(sb, &(fid->dir), fid->entry, &sector);
entry            1835 drivers/staging/exfat/exfat_super.c 			es = get_entry_set_in_dir(sb, &fid->dir, fid->entry,
entry            1849 drivers/staging/exfat/exfat_super.c 						      fid->entry, &sector);
entry            1961 drivers/staging/exfat/exfat_super.c 	if (fid->entry == -1) {
entry            2149 drivers/staging/exfat/exfat_super.c 	dentry = fid->entry;
entry            2384 drivers/staging/exfat/exfat_super.c 	i_pos = ((loff_t)fid.dir.dir << 32) | (fid.entry & 0xffffffff);
entry            2449 drivers/staging/exfat/exfat_super.c 	i_pos = ((loff_t)fid.dir.dir << 32) | (fid.entry & 0xffffffff);
entry            2606 drivers/staging/exfat/exfat_super.c 	i_pos = ((loff_t)fid.dir.dir << 32) | (fid.entry & 0xffffffff);
entry            2666 drivers/staging/exfat/exfat_super.c 	i_pos = ((loff_t)fid.dir.dir << 32) | (fid.entry & 0xffffffff);
entry            2778 drivers/staging/exfat/exfat_super.c 			(EXFAT_I(old_inode)->fid.entry & 0xffffffff);
entry            3795 drivers/staging/exfat/exfat_super.c 	EXFAT_I(inode)->fid.entry = -1;
entry             236 drivers/staging/fsl-dpaa2/ethsw/ethsw.c 	struct dpsw_fdb_unicast_cfg entry = {0};
entry             239 drivers/staging/fsl-dpaa2/ethsw/ethsw.c 	entry.if_egress = port_priv->idx;
entry             240 drivers/staging/fsl-dpaa2/ethsw/ethsw.c 	entry.type = DPSW_FDB_ENTRY_STATIC;
entry             241 drivers/staging/fsl-dpaa2/ethsw/ethsw.c 	ether_addr_copy(entry.mac_addr, addr);
entry             245 drivers/staging/fsl-dpaa2/ethsw/ethsw.c 				   0, &entry);
entry             255 drivers/staging/fsl-dpaa2/ethsw/ethsw.c 	struct dpsw_fdb_unicast_cfg entry = {0};
entry             258 drivers/staging/fsl-dpaa2/ethsw/ethsw.c 	entry.if_egress = port_priv->idx;
entry             259 drivers/staging/fsl-dpaa2/ethsw/ethsw.c 	entry.type = DPSW_FDB_ENTRY_STATIC;
entry             260 drivers/staging/fsl-dpaa2/ethsw/ethsw.c 	ether_addr_copy(entry.mac_addr, addr);
entry             264 drivers/staging/fsl-dpaa2/ethsw/ethsw.c 				      0, &entry);
entry             275 drivers/staging/fsl-dpaa2/ethsw/ethsw.c 	struct dpsw_fdb_multicast_cfg entry = {0};
entry             278 drivers/staging/fsl-dpaa2/ethsw/ethsw.c 	ether_addr_copy(entry.mac_addr, addr);
entry             279 drivers/staging/fsl-dpaa2/ethsw/ethsw.c 	entry.type = DPSW_FDB_ENTRY_STATIC;
entry             280 drivers/staging/fsl-dpaa2/ethsw/ethsw.c 	entry.num_ifs = 1;
entry             281 drivers/staging/fsl-dpaa2/ethsw/ethsw.c 	entry.if_id[0] = port_priv->idx;
entry             285 drivers/staging/fsl-dpaa2/ethsw/ethsw.c 				     0, &entry);
entry             296 drivers/staging/fsl-dpaa2/ethsw/ethsw.c 	struct dpsw_fdb_multicast_cfg entry = {0};
entry             299 drivers/staging/fsl-dpaa2/ethsw/ethsw.c 	ether_addr_copy(entry.mac_addr, addr);
entry             300 drivers/staging/fsl-dpaa2/ethsw/ethsw.c 	entry.type = DPSW_FDB_ENTRY_STATIC;
entry             301 drivers/staging/fsl-dpaa2/ethsw/ethsw.c 	entry.num_ifs = 1;
entry             302 drivers/staging/fsl-dpaa2/ethsw/ethsw.c 	entry.if_id[0] = port_priv->idx;
entry             306 drivers/staging/fsl-dpaa2/ethsw/ethsw.c 					0, &entry);
entry             559 drivers/staging/fsl-dpaa2/ethsw/ethsw.c static int ethsw_fdb_do_dump(struct fdb_dump_entry *entry,
entry             562 drivers/staging/fsl-dpaa2/ethsw/ethsw.c 	int is_dynamic = entry->type & DPSW_FDB_ENTRY_DINAMIC;
entry             585 drivers/staging/fsl-dpaa2/ethsw/ethsw.c 	if (nla_put(dump->skb, NDA_LLADDR, ETH_ALEN, entry->mac_addr))
entry             599 drivers/staging/fsl-dpaa2/ethsw/ethsw.c static int port_fdb_valid_entry(struct fdb_dump_entry *entry,
entry             605 drivers/staging/fsl-dpaa2/ethsw/ethsw.c 	if (entry->type & DPSW_FDB_ENTRY_TYPE_UNICAST)
entry             606 drivers/staging/fsl-dpaa2/ethsw/ethsw.c 		valid = entry->if_info == port_priv->idx;
entry             608 drivers/staging/fsl-dpaa2/ethsw/ethsw.c 		valid = entry->if_mask[idx / 8] & BIT(idx % 8);
entry             164 drivers/staging/gasket/gasket_interrupt.c 			interrupt = interrupt_data->msix_entries[i].entry;
entry             189 drivers/staging/gasket/gasket_interrupt.c 		interrupt_data->msix_entries[i].entry = i;
entry            1146 drivers/staging/greybus/camera.c 		const struct gb_camera_debugfs_entry *entry =
entry            1149 drivers/staging/greybus/camera.c 		if (!strcmp(file->f_path.dentry->d_iname, entry->name)) {
entry            1150 drivers/staging/greybus/camera.c 			file->private_data = (void *)entry;
entry            1185 drivers/staging/greybus/camera.c 		const struct gb_camera_debugfs_entry *entry =
entry            1190 drivers/staging/greybus/camera.c 		debugfs_create_file(entry->name, entry->mask,
entry              30 drivers/staging/greybus/raw.c 	struct list_head entry;
entry              83 drivers/staging/greybus/raw.c 	list_add_tail(&raw_data->entry, &raw->list);
entry             239 drivers/staging/greybus/raw.c 	list_for_each_entry_safe(raw_data, temp, &raw->list, entry) {
entry             240 drivers/staging/greybus/raw.c 		list_del(&raw_data->entry);
entry             296 drivers/staging/greybus/raw.c 	raw_data = list_first_entry(&raw->list, struct raw_data, entry);
entry             307 drivers/staging/greybus/raw.c 	list_del(&raw_data->entry);
entry              30 drivers/staging/media/ipu3/ipu3-css-pool.c 		imgu_dmamap_free(imgu, &pool->entry[i].param);
entry              39 drivers/staging/media/ipu3/ipu3-css-pool.c 		pool->entry[i].valid = false;
entry              41 drivers/staging/media/ipu3/ipu3-css-pool.c 			pool->entry[i].param.vaddr = NULL;
entry              45 drivers/staging/media/ipu3/ipu3-css-pool.c 		if (!imgu_dmamap_alloc(imgu, &pool->entry[i].param, size))
entry              66 drivers/staging/media/ipu3/ipu3-css-pool.c 	pool->entry[n].valid = true;
entry              75 drivers/staging/media/ipu3/ipu3-css-pool.c 	pool->entry[pool->last].valid = false;
entry              96 drivers/staging/media/ipu3/ipu3-css-pool.c 	if (!pool->entry[i].valid)
entry              99 drivers/staging/media/ipu3/ipu3-css-pool.c 	return &pool->entry[i].param;
entry              40 drivers/staging/media/ipu3/ipu3-css-pool.h 	} entry[IPU3_CSS_POOL_SIZE];
entry              29 drivers/staging/media/tegra-vde/dmabuf-cache.c static void tegra_vde_release_entry(struct tegra_vde_cache_entry *entry)
entry              31 drivers/staging/media/tegra-vde/dmabuf-cache.c 	struct dma_buf *dmabuf = entry->a->dmabuf;
entry              33 drivers/staging/media/tegra-vde/dmabuf-cache.c 	WARN_ON_ONCE(entry->refcnt);
entry              35 drivers/staging/media/tegra-vde/dmabuf-cache.c 	if (entry->vde->domain)
entry              36 drivers/staging/media/tegra-vde/dmabuf-cache.c 		tegra_vde_iommu_unmap(entry->vde, entry->iova);
entry              38 drivers/staging/media/tegra-vde/dmabuf-cache.c 	dma_buf_unmap_attachment(entry->a, entry->sgt, entry->dma_dir);
entry              39 drivers/staging/media/tegra-vde/dmabuf-cache.c 	dma_buf_detach(dmabuf, entry->a);
entry              42 drivers/staging/media/tegra-vde/dmabuf-cache.c 	list_del(&entry->list);
entry              43 drivers/staging/media/tegra-vde/dmabuf-cache.c 	kfree(entry);
entry              48 drivers/staging/media/tegra-vde/dmabuf-cache.c 	struct tegra_vde_cache_entry *entry;
entry              51 drivers/staging/media/tegra-vde/dmabuf-cache.c 	entry = container_of(work, struct tegra_vde_cache_entry,
entry              53 drivers/staging/media/tegra-vde/dmabuf-cache.c 	vde = entry->vde;
entry              56 drivers/staging/media/tegra-vde/dmabuf-cache.c 	tegra_vde_release_entry(entry);
entry              68 drivers/staging/media/tegra-vde/dmabuf-cache.c 	struct tegra_vde_cache_entry *entry;
entry              75 drivers/staging/media/tegra-vde/dmabuf-cache.c 	list_for_each_entry(entry, &vde->map_list, list) {
entry              76 drivers/staging/media/tegra-vde/dmabuf-cache.c 		if (entry->a->dmabuf != dmabuf)
entry              79 drivers/staging/media/tegra-vde/dmabuf-cache.c 		if (!cancel_delayed_work(&entry->dwork))
entry              82 drivers/staging/media/tegra-vde/dmabuf-cache.c 		if (entry->dma_dir != dma_dir)
entry              83 drivers/staging/media/tegra-vde/dmabuf-cache.c 			entry->dma_dir = DMA_BIDIRECTIONAL;
entry              88 drivers/staging/media/tegra-vde/dmabuf-cache.c 			*addrp = iova_dma_addr(&vde->iova, entry->iova);
entry              90 drivers/staging/media/tegra-vde/dmabuf-cache.c 			*addrp = sg_dma_address(entry->sgt->sgl);
entry             115 drivers/staging/media/tegra-vde/dmabuf-cache.c 	entry = kzalloc(sizeof(*entry), GFP_KERNEL);
entry             116 drivers/staging/media/tegra-vde/dmabuf-cache.c 	if (!entry) {
entry             132 drivers/staging/media/tegra-vde/dmabuf-cache.c 	INIT_DELAYED_WORK(&entry->dwork, tegra_vde_delayed_unmap);
entry             133 drivers/staging/media/tegra-vde/dmabuf-cache.c 	list_add(&entry->list, &vde->map_list);
entry             135 drivers/staging/media/tegra-vde/dmabuf-cache.c 	entry->dma_dir = dma_dir;
entry             136 drivers/staging/media/tegra-vde/dmabuf-cache.c 	entry->iova = iova;
entry             137 drivers/staging/media/tegra-vde/dmabuf-cache.c 	entry->vde = vde;
entry             138 drivers/staging/media/tegra-vde/dmabuf-cache.c 	entry->sgt = sgt;
entry             139 drivers/staging/media/tegra-vde/dmabuf-cache.c 	entry->a = attachment;
entry             141 drivers/staging/media/tegra-vde/dmabuf-cache.c 	entry->refcnt++;
entry             143 drivers/staging/media/tegra-vde/dmabuf-cache.c 	*ap = entry->a;
entry             150 drivers/staging/media/tegra-vde/dmabuf-cache.c 	kfree(entry);
entry             165 drivers/staging/media/tegra-vde/dmabuf-cache.c 	struct tegra_vde_cache_entry *entry;
entry             169 drivers/staging/media/tegra-vde/dmabuf-cache.c 	list_for_each_entry(entry, &vde->map_list, list) {
entry             170 drivers/staging/media/tegra-vde/dmabuf-cache.c 		if (entry->a != a)
entry             173 drivers/staging/media/tegra-vde/dmabuf-cache.c 		WARN_ON_ONCE(!entry->refcnt);
entry             175 drivers/staging/media/tegra-vde/dmabuf-cache.c 		if (--entry->refcnt == 0) {
entry             177 drivers/staging/media/tegra-vde/dmabuf-cache.c 				tegra_vde_release_entry(entry);
entry             179 drivers/staging/media/tegra-vde/dmabuf-cache.c 				schedule_delayed_work(&entry->dwork, 5 * HZ);
entry             189 drivers/staging/media/tegra-vde/dmabuf-cache.c 	struct tegra_vde_cache_entry *entry, *tmp;
entry             193 drivers/staging/media/tegra-vde/dmabuf-cache.c 	list_for_each_entry_safe(entry, tmp, &vde->map_list, list) {
entry             194 drivers/staging/media/tegra-vde/dmabuf-cache.c 		if (entry->refcnt)
entry             197 drivers/staging/media/tegra-vde/dmabuf-cache.c 		if (!cancel_delayed_work(&entry->dwork))
entry             200 drivers/staging/media/tegra-vde/dmabuf-cache.c 		tegra_vde_release_entry(entry);
entry             208 drivers/staging/media/tegra-vde/dmabuf-cache.c 	struct tegra_vde_cache_entry *entry, *tmp;
entry             213 drivers/staging/media/tegra-vde/dmabuf-cache.c 		list_for_each_entry_safe(entry, tmp, &vde->map_list, list) {
entry             214 drivers/staging/media/tegra-vde/dmabuf-cache.c 			if (!cancel_delayed_work(&entry->dwork))
entry             217 drivers/staging/media/tegra-vde/dmabuf-cache.c 			tegra_vde_release_entry(entry);
entry             341 drivers/staging/octeon-usb/octeon-hcd.c 	} entry[MAX_CHANNELS + 1];
entry            1210 drivers/staging/octeon-usb/octeon-hcd.c 		const u32 *ptr = cvmx_phys_to_ptr(fifo->entry[i].address);
entry            1211 drivers/staging/octeon-usb/octeon-hcd.c 		u64 csr_address = USB_FIFO_ADDRESS(fifo->entry[i].channel,
entry            1216 drivers/staging/octeon-usb/octeon-hcd.c 		if (fifo->entry[i].size <= available) {
entry            1217 drivers/staging/octeon-usb/octeon-hcd.c 			words = fifo->entry[i].size;
entry            1225 drivers/staging/octeon-usb/octeon-hcd.c 		fifo->entry[i].address += words * 4;
entry            1226 drivers/staging/octeon-usb/octeon-hcd.c 		fifo->entry[i].size -= words;
entry            1327 drivers/staging/octeon-usb/octeon-hcd.c 	fifo->entry[fifo->head].channel = channel;
entry            1328 drivers/staging/octeon-usb/octeon-hcd.c 	fifo->entry[fifo->head].address =
entry            1331 drivers/staging/octeon-usb/octeon-hcd.c 	fifo->entry[fifo->head].size = (usbc_hctsiz.s.xfersize + 3) >> 2;
entry            1611 drivers/staging/qlge/qlge_dbg.c 			DUMP_QDEV_ARRAY(qdev, "%d", msi_x_entry, i, entry);
entry            3339 drivers/staging/qlge/qlge_main.c 			qdev->msi_x_entry[i].entry = i;
entry             584 drivers/staging/rtl8188eu/core/rtw_cmd.c u8 rtw_clearstakey_cmd(struct adapter *padapter, u8 *psta, u8 entry, u8 enqueue)
entry             594 drivers/staging/rtl8188eu/core/rtw_cmd.c 		clear_cam_entry(padapter, entry);
entry             625 drivers/staging/rtl8188eu/core/rtw_cmd.c 		psetstakey_para->id = entry;
entry             411 drivers/staging/rtl8188eu/core/rtw_wlan_util.c void write_cam(struct adapter *padapter, u8 entry, u16 ctrl, u8 *mac, u8 *key)
entry             417 drivers/staging/rtl8188eu/core/rtw_wlan_util.c 	addr = entry << 3;
entry             441 drivers/staging/rtl8188eu/core/rtw_wlan_util.c void clear_cam_entry(struct adapter *padapter, u8 entry)
entry             446 drivers/staging/rtl8188eu/core/rtw_wlan_util.c 	write_cam(padapter, entry, 0, null_sta, null_key);
entry            1378 drivers/staging/rtl8188eu/core/rtw_xmit.c struct xmit_frame *rtw_dequeue_xframe(struct xmit_priv *pxmitpriv, struct hw_xmit *phwxmit_i, int entry)
entry            1400 drivers/staging/rtl8188eu/core/rtw_xmit.c 	for (i = 0; i < entry; i++) {
entry            1538 drivers/staging/rtl8188eu/core/rtw_xmit.c void rtw_init_hwxmits(struct hw_xmit *phwxmit, int entry)
entry            1542 drivers/staging/rtl8188eu/core/rtw_xmit.c 	for (i = 0; i < entry; i++, phwxmit++)
entry             244 drivers/staging/rtl8188eu/hal/odm_rtl8188e.c 	struct sta_info *entry;
entry             251 drivers/staging/rtl8188eu/hal/odm_rtl8188e.c 		entry = dm_odm->pODM_StaInfo[i];
entry             252 drivers/staging/rtl8188eu/hal/odm_rtl8188e.c 		if (IS_STA_VALID(entry)) {
entry             305 drivers/staging/rtl8188eu/include/rtw_cmd.h u8 rtw_clearstakey_cmd(struct adapter *padapter, u8 *psta, u8 entry,
entry             477 drivers/staging/rtl8188eu/include/rtw_mlme_ext.h void write_cam(struct adapter *padapter, u8 entry, u16 ctrl, u8 *mac, u8 *key);
entry             478 drivers/staging/rtl8188eu/include/rtw_mlme_ext.h void clear_cam_entry(struct adapter *padapter, u8 entry);
entry             325 drivers/staging/rtl8188eu/include/rtw_xmit.h 				      struct hw_xmit *phwxmit_i, int entry);
entry             336 drivers/staging/rtl8188eu/include/rtw_xmit.h void rtw_init_hwxmits(struct hw_xmit *phwxmit, int entry);
entry            1281 drivers/staging/rtl8192e/rtl8192e/r8192E_dev.c void  rtl92e_fill_tx_cmd_desc(struct net_device *dev, struct tx_desc_cmd *entry,
entry            1290 drivers/staging/rtl8192e/rtl8192e/r8192E_dev.c 	memset(entry, 0, 12);
entry            1291 drivers/staging/rtl8192e/rtl8192e/r8192E_dev.c 	entry->LINIP = cb_desc->bLastIniPkt;
entry            1292 drivers/staging/rtl8192e/rtl8192e/r8192E_dev.c 	entry->FirstSeg = 1;
entry            1293 drivers/staging/rtl8192e/rtl8192e/r8192E_dev.c 	entry->LastSeg = 1;
entry            1295 drivers/staging/rtl8192e/rtl8192e/r8192E_dev.c 		entry->CmdInit = DESC_PACKET_TYPE_INIT;
entry            1297 drivers/staging/rtl8192e/rtl8192e/r8192E_dev.c 		struct tx_desc *entry_tmp = (struct tx_desc *)entry;
entry            1307 drivers/staging/rtl8192e/rtl8192e/r8192E_dev.c 	entry->TxBufferSize = skb->len;
entry            1308 drivers/staging/rtl8192e/rtl8192e/r8192E_dev.c 	entry->TxBuffAddr = mapping;
entry            1309 drivers/staging/rtl8192e/rtl8192e/r8192E_dev.c 	entry->OWN = 1;
entry              32 drivers/staging/rtl8192e/rtl8192e/r8192E_dev.h void  rtl92e_fill_tx_cmd_desc(struct net_device *dev, struct tx_desc_cmd *entry,
entry            1583 drivers/staging/rtl8192e/rtl8192e/rtl_core.c 		struct tx_desc *entry = &ring->desc[ring->idx];
entry            1586 drivers/staging/rtl8192e/rtl8192e/rtl_core.c 		pci_unmap_single(priv->pdev, entry->TxBuffAddr,
entry            1670 drivers/staging/rtl8192e/rtl8192e/rtl_core.c 		struct tx_desc *entry = &ring->desc[ring->idx];
entry            1674 drivers/staging/rtl8192e/rtl8192e/rtl_core.c 			if (entry->OWN)
entry            1680 drivers/staging/rtl8192e/rtl8192e/rtl_core.c 		pci_unmap_single(priv->pdev, entry->TxBuffAddr,
entry            1693 drivers/staging/rtl8192e/rtl8192e/rtl_core.c 	struct tx_desc_cmd *entry;
entry            1702 drivers/staging/rtl8192e/rtl8192e/rtl_core.c 	entry = (struct tx_desc_cmd *)&ring->desc[idx];
entry            1706 drivers/staging/rtl8192e/rtl8192e/rtl_core.c 	priv->ops->tx_fill_cmd_descriptor(dev, entry, tcb_desc, skb);
entry            1782 drivers/staging/rtl8192e/rtl8192e/rtl_core.c 	struct rx_desc *entry = NULL;
entry            1801 drivers/staging/rtl8192e/rtl8192e/rtl_core.c 			entry = &priv->rx_ring[rx_queue_idx][i];
entry            1815 drivers/staging/rtl8192e/rtl8192e/rtl_core.c 			entry->BufferAddress = *mapping;
entry            1817 drivers/staging/rtl8192e/rtl8192e/rtl_core.c 			entry->Length = priv->rxbuffersize;
entry            1818 drivers/staging/rtl8192e/rtl8192e/rtl_core.c 			entry->OWN = 1;
entry            1821 drivers/staging/rtl8192e/rtl8192e/rtl_core.c 		if (entry)
entry            1822 drivers/staging/rtl8192e/rtl8192e/rtl_core.c 			entry->EOR = 1;
entry            1889 drivers/staging/rtl8192e/rtl8192e/rtl_core.c 			struct rx_desc *entry = NULL;
entry            1892 drivers/staging/rtl8192e/rtl8192e/rtl_core.c 				entry = &priv->rx_ring[rx_queue_idx][i];
entry            1893 drivers/staging/rtl8192e/rtl8192e/rtl_core.c 				entry->OWN = 1;
entry            1905 drivers/staging/rtl8192e/rtl8192e/rtl_core.c 				struct tx_desc *entry = &ring->desc[ring->idx];
entry            1910 drivers/staging/rtl8192e/rtl8192e/rtl_core.c 						 entry->TxBuffAddr,
entry             283 drivers/staging/rtl8192e/rtl8192e/rtl_core.h 				       struct tx_desc_cmd *entry,
entry              61 drivers/staging/rtl8192e/rtllib_rx.c 	struct rtllib_frag_entry *entry;
entry              65 drivers/staging/rtl8192e/rtllib_rx.c 		entry = &ieee->frag_cache[tid][i];
entry              66 drivers/staging/rtl8192e/rtllib_rx.c 		if (entry->skb != NULL &&
entry              67 drivers/staging/rtl8192e/rtllib_rx.c 		    time_after(jiffies, entry->first_frag_time + 2 * HZ)) {
entry              70 drivers/staging/rtl8192e/rtllib_rx.c 				   entry->seq, entry->last_frag);
entry              71 drivers/staging/rtl8192e/rtllib_rx.c 			dev_kfree_skb_any(entry->skb);
entry              72 drivers/staging/rtl8192e/rtllib_rx.c 			entry->skb = NULL;
entry              75 drivers/staging/rtl8192e/rtllib_rx.c 		if (entry->skb != NULL && entry->seq == seq &&
entry              76 drivers/staging/rtl8192e/rtllib_rx.c 		    (entry->last_frag + 1 == frag || frag == -1) &&
entry              77 drivers/staging/rtl8192e/rtllib_rx.c 		    memcmp(entry->src_addr, src, ETH_ALEN) == 0 &&
entry              78 drivers/staging/rtl8192e/rtllib_rx.c 		    memcmp(entry->dst_addr, dst, ETH_ALEN) == 0)
entry              79 drivers/staging/rtl8192e/rtllib_rx.c 			return entry;
entry              95 drivers/staging/rtl8192e/rtllib_rx.c 	struct rtllib_frag_entry *entry;
entry             128 drivers/staging/rtl8192e/rtllib_rx.c 		entry = &ieee->frag_cache[tid][ieee->frag_next_idx[tid]];
entry             133 drivers/staging/rtl8192e/rtllib_rx.c 		if (entry->skb != NULL)
entry             134 drivers/staging/rtl8192e/rtllib_rx.c 			dev_kfree_skb_any(entry->skb);
entry             136 drivers/staging/rtl8192e/rtllib_rx.c 		entry->first_frag_time = jiffies;
entry             137 drivers/staging/rtl8192e/rtllib_rx.c 		entry->seq = seq;
entry             138 drivers/staging/rtl8192e/rtllib_rx.c 		entry->last_frag = frag;
entry             139 drivers/staging/rtl8192e/rtllib_rx.c 		entry->skb = skb;
entry             140 drivers/staging/rtl8192e/rtllib_rx.c 		ether_addr_copy(entry->src_addr, hdr->addr2);
entry             141 drivers/staging/rtl8192e/rtllib_rx.c 		ether_addr_copy(entry->dst_addr, hdr->addr1);
entry             146 drivers/staging/rtl8192e/rtllib_rx.c 		entry = rtllib_frag_cache_find(ieee, seq, frag, tid, hdr->addr2,
entry             148 drivers/staging/rtl8192e/rtllib_rx.c 		if (entry != NULL) {
entry             149 drivers/staging/rtl8192e/rtllib_rx.c 			entry->last_frag = frag;
entry             150 drivers/staging/rtl8192e/rtllib_rx.c 			skb = entry->skb;
entry             165 drivers/staging/rtl8192e/rtllib_rx.c 	struct rtllib_frag_entry *entry;
entry             185 drivers/staging/rtl8192e/rtllib_rx.c 	entry = rtllib_frag_cache_find(ieee, seq, -1, tid, hdr->addr2,
entry             188 drivers/staging/rtl8192e/rtllib_rx.c 	if (entry == NULL) {
entry             195 drivers/staging/rtl8192e/rtllib_rx.c 	entry->skb = NULL;
entry             393 drivers/staging/rtl8192e/rtllib_rx.c 		struct ieee_ibss_seq *entry = NULL;
entry             398 drivers/staging/rtl8192e/rtllib_rx.c 			entry = list_entry(p, struct ieee_ibss_seq, list);
entry             399 drivers/staging/rtl8192e/rtllib_rx.c 			if (!memcmp(entry->mac, mac, ETH_ALEN))
entry             403 drivers/staging/rtl8192e/rtllib_rx.c 			entry = kmalloc(sizeof(struct ieee_ibss_seq),
entry             405 drivers/staging/rtl8192e/rtllib_rx.c 			if (!entry)
entry             408 drivers/staging/rtl8192e/rtllib_rx.c 			ether_addr_copy(entry->mac, mac);
entry             409 drivers/staging/rtl8192e/rtllib_rx.c 			entry->seq_num[tid] = seq;
entry             410 drivers/staging/rtl8192e/rtllib_rx.c 			entry->frag_num[tid] = frag;
entry             411 drivers/staging/rtl8192e/rtllib_rx.c 			entry->packet_time[tid] = jiffies;
entry             412 drivers/staging/rtl8192e/rtllib_rx.c 			list_add(&entry->list, &ieee->ibss_mac_hash[index]);
entry             415 drivers/staging/rtl8192e/rtllib_rx.c 		last_seq = &entry->seq_num[tid];
entry             416 drivers/staging/rtl8192e/rtllib_rx.c 		last_frag = &entry->frag_num[tid];
entry             417 drivers/staging/rtl8192e/rtllib_rx.c 		last_time = &entry->packet_time[tid];
entry              38 drivers/staging/rtl8192u/ieee80211/ieee80211_crypt.c 	struct ieee80211_crypt_data *entry;
entry              42 drivers/staging/rtl8192u/ieee80211/ieee80211_crypt.c 		entry = list_entry(ptr, struct ieee80211_crypt_data, list);
entry              44 drivers/staging/rtl8192u/ieee80211/ieee80211_crypt.c 		if (atomic_read(&entry->refcnt) != 0 && !force)
entry              49 drivers/staging/rtl8192u/ieee80211/ieee80211_crypt.c 		if (entry->ops)
entry              50 drivers/staging/rtl8192u/ieee80211/ieee80211_crypt.c 			entry->ops->deinit(entry->priv);
entry              51 drivers/staging/rtl8192u/ieee80211/ieee80211_crypt.c 		kfree(entry);
entry              65 drivers/staging/rtl8192u/ieee80211/ieee80211_rx.c 	struct ieee80211_frag_entry *entry;
entry              69 drivers/staging/rtl8192u/ieee80211/ieee80211_rx.c 		entry = &ieee->frag_cache[tid][i];
entry              70 drivers/staging/rtl8192u/ieee80211/ieee80211_rx.c 		if (entry->skb &&
entry              71 drivers/staging/rtl8192u/ieee80211/ieee80211_rx.c 		    time_after(jiffies, entry->first_frag_time + 2 * HZ)) {
entry              75 drivers/staging/rtl8192u/ieee80211/ieee80211_rx.c 				entry->seq, entry->last_frag);
entry              76 drivers/staging/rtl8192u/ieee80211/ieee80211_rx.c 			dev_kfree_skb_any(entry->skb);
entry              77 drivers/staging/rtl8192u/ieee80211/ieee80211_rx.c 			entry->skb = NULL;
entry              80 drivers/staging/rtl8192u/ieee80211/ieee80211_rx.c 		if (entry->skb && entry->seq == seq &&
entry              81 drivers/staging/rtl8192u/ieee80211/ieee80211_rx.c 		    (entry->last_frag + 1 == frag || frag == -1) &&
entry              82 drivers/staging/rtl8192u/ieee80211/ieee80211_rx.c 		    memcmp(entry->src_addr, src, ETH_ALEN) == 0 &&
entry              83 drivers/staging/rtl8192u/ieee80211/ieee80211_rx.c 		    memcmp(entry->dst_addr, dst, ETH_ALEN) == 0)
entry              84 drivers/staging/rtl8192u/ieee80211/ieee80211_rx.c 			return entry;
entry             100 drivers/staging/rtl8192u/ieee80211/ieee80211_rx.c 	struct ieee80211_frag_entry *entry;
entry             131 drivers/staging/rtl8192u/ieee80211/ieee80211_rx.c 		entry = &ieee->frag_cache[tid][ieee->frag_next_idx[tid]];
entry             136 drivers/staging/rtl8192u/ieee80211/ieee80211_rx.c 		if (entry->skb)
entry             137 drivers/staging/rtl8192u/ieee80211/ieee80211_rx.c 			dev_kfree_skb_any(entry->skb);
entry             139 drivers/staging/rtl8192u/ieee80211/ieee80211_rx.c 		entry->first_frag_time = jiffies;
entry             140 drivers/staging/rtl8192u/ieee80211/ieee80211_rx.c 		entry->seq = seq;
entry             141 drivers/staging/rtl8192u/ieee80211/ieee80211_rx.c 		entry->last_frag = frag;
entry             142 drivers/staging/rtl8192u/ieee80211/ieee80211_rx.c 		entry->skb = skb;
entry             143 drivers/staging/rtl8192u/ieee80211/ieee80211_rx.c 		memcpy(entry->src_addr, hdr->addr2, ETH_ALEN);
entry             144 drivers/staging/rtl8192u/ieee80211/ieee80211_rx.c 		memcpy(entry->dst_addr, hdr->addr1, ETH_ALEN);
entry             148 drivers/staging/rtl8192u/ieee80211/ieee80211_rx.c 		entry = ieee80211_frag_cache_find(ieee, seq, frag, tid, hdr->addr2,
entry             150 drivers/staging/rtl8192u/ieee80211/ieee80211_rx.c 		if (entry) {
entry             151 drivers/staging/rtl8192u/ieee80211/ieee80211_rx.c 			entry->last_frag = frag;
entry             152 drivers/staging/rtl8192u/ieee80211/ieee80211_rx.c 			skb = entry->skb;
entry             167 drivers/staging/rtl8192u/ieee80211/ieee80211_rx.c 	struct ieee80211_frag_entry *entry;
entry             186 drivers/staging/rtl8192u/ieee80211/ieee80211_rx.c 	entry = ieee80211_frag_cache_find(ieee, seq, -1, tid, hdr->addr2,
entry             189 drivers/staging/rtl8192u/ieee80211/ieee80211_rx.c 	if (!entry) {
entry             196 drivers/staging/rtl8192u/ieee80211/ieee80211_rx.c 	entry->skb = NULL;
entry             445 drivers/staging/rtl8192u/ieee80211/ieee80211_rx.c 		struct ieee_ibss_seq *entry = NULL;
entry             450 drivers/staging/rtl8192u/ieee80211/ieee80211_rx.c 			entry = list_entry(p, struct ieee_ibss_seq, list);
entry             451 drivers/staging/rtl8192u/ieee80211/ieee80211_rx.c 			if (!memcmp(entry->mac, mac, ETH_ALEN))
entry             456 drivers/staging/rtl8192u/ieee80211/ieee80211_rx.c 			entry = kmalloc(sizeof(struct ieee_ibss_seq), GFP_ATOMIC);
entry             457 drivers/staging/rtl8192u/ieee80211/ieee80211_rx.c 			if (!entry)
entry             459 drivers/staging/rtl8192u/ieee80211/ieee80211_rx.c 			memcpy(entry->mac, mac, ETH_ALEN);
entry             460 drivers/staging/rtl8192u/ieee80211/ieee80211_rx.c 			entry->seq_num[tid] = seq;
entry             461 drivers/staging/rtl8192u/ieee80211/ieee80211_rx.c 			entry->frag_num[tid] = frag;
entry             462 drivers/staging/rtl8192u/ieee80211/ieee80211_rx.c 			entry->packet_time[tid] = jiffies;
entry             463 drivers/staging/rtl8192u/ieee80211/ieee80211_rx.c 			list_add(&entry->list, &ieee->ibss_mac_hash[index]);
entry             466 drivers/staging/rtl8192u/ieee80211/ieee80211_rx.c 		last_seq = &entry->seq_num[tid];
entry             467 drivers/staging/rtl8192u/ieee80211/ieee80211_rx.c 		last_frag = &entry->frag_num[tid];
entry             468 drivers/staging/rtl8192u/ieee80211/ieee80211_rx.c 		last_time = &entry->packet_time[tid];
entry             719 drivers/staging/rtl8192u/r8192U_core.c 	struct urb *entry;
entry             728 drivers/staging/rtl8192u/r8192U_core.c 		entry = usb_alloc_urb(0, GFP_KERNEL);
entry             729 drivers/staging/rtl8192u/r8192U_core.c 		if (!entry) {
entry             733 drivers/staging/rtl8192u/r8192U_core.c 		usb_fill_bulk_urb(entry, priv->udev,
entry             738 drivers/staging/rtl8192u/r8192U_core.c 		info->urb = entry;
entry             742 drivers/staging/rtl8192u/r8192U_core.c 		usb_submit_urb(entry, GFP_KERNEL);
entry             750 drivers/staging/rtl8192u/r8192U_core.c 		entry = usb_alloc_urb(0, GFP_KERNEL);
entry             751 drivers/staging/rtl8192u/r8192U_core.c 		if (!entry) {
entry             755 drivers/staging/rtl8192u/r8192U_core.c 		usb_fill_bulk_urb(entry, priv->udev,
entry             760 drivers/staging/rtl8192u/r8192U_core.c 		info->urb = entry;
entry             764 drivers/staging/rtl8192u/r8192U_core.c 		usb_submit_urb(entry, GFP_KERNEL);
entry             170 drivers/staging/rtl8712/rtl8712_xmit.c 				     struct hw_xmit *phwxmit_i, sint entry)
entry             195 drivers/staging/rtl8712/rtl8712_xmit.c 	for (i = 0; i < entry; i++) {
entry              28 drivers/staging/rtl8712/rtl871x_xmit.c static void init_hwxmits(struct hw_xmit *phwxmit, sint entry);
entry             984 drivers/staging/rtl8712/rtl871x_xmit.c static void init_hwxmits(struct hw_xmit *phwxmit, sint entry)
entry             988 drivers/staging/rtl8712/rtl871x_xmit.c 	for (i = 0; i < entry; i++, phwxmit++) {
entry             545 drivers/staging/rtl8723bs/core/rtw_wlan_util.c void read_cam(struct adapter *padapter, u8 entry, u8 *get_key)
entry             548 drivers/staging/rtl8723bs/core/rtw_wlan_util.c 	addr = entry << 3;
entry             560 drivers/staging/rtl8723bs/core/rtw_wlan_util.c void _write_cam(struct adapter *padapter, u8 entry, u16 ctrl, u8 *mac, u8 *key)
entry             566 drivers/staging/rtl8723bs/core/rtw_wlan_util.c 	addr = entry << 3;
entry             589 drivers/staging/rtl8723bs/core/rtw_wlan_util.c void _clear_cam_entry(struct adapter *padapter, u8 entry)
entry             594 drivers/staging/rtl8723bs/core/rtw_wlan_util.c 	_write_cam(padapter, entry, 0, null_sta, null_key);
entry            2218 drivers/staging/rtl8723bs/core/rtw_xmit.c void rtw_init_hwxmits(struct hw_xmit *phwxmit, sint entry)
entry            2222 drivers/staging/rtl8723bs/core/rtw_xmit.c 	for (i = 0; i < entry; i++, phwxmit++) {
entry             574 drivers/staging/rtl8723bs/include/rtw_mlme_ext.h void read_cam(struct adapter *padapter , u8 entry, u8 *get_key);
entry             577 drivers/staging/rtl8723bs/include/rtw_mlme_ext.h void _write_cam(struct adapter *padapter, u8 entry, u16 ctrl, u8 *mac, u8 *key);
entry             578 drivers/staging/rtl8723bs/include/rtw_mlme_ext.h void _clear_cam_entry(struct adapter *padapter, u8 entry);
entry             483 drivers/staging/rtl8723bs/include/rtw_xmit.h void rtw_init_hwxmits(struct hw_xmit *phwxmit, sint entry);
entry             150 drivers/staging/unisys/visorhba/visorhba_main.c 	struct scsipending *entry;
entry             163 drivers/staging/unisys/visorhba/visorhba_main.c 	entry = &devdata->pending[insert_location];
entry             164 drivers/staging/unisys/visorhba/visorhba_main.c 	memset(&entry->cmdrsp, 0, sizeof(entry->cmdrsp));
entry             165 drivers/staging/unisys/visorhba/visorhba_main.c 	entry->cmdtype = cmdtype;
entry             167 drivers/staging/unisys/visorhba/visorhba_main.c 		entry->sent = new;
entry             170 drivers/staging/unisys/visorhba/visorhba_main.c 		entry->sent = &entry->cmdrsp;
entry              38 drivers/staging/uwb/est.c 	const struct uwb_est_entry *entry;
entry             243 drivers/staging/uwb/est.c 		     const struct uwb_est_entry *entry, size_t entries)
entry             268 drivers/staging/uwb/est.c 	uwb_est[itr].entry = entry;
entry             292 drivers/staging/uwb/est.c 		       const struct uwb_est_entry *entry, size_t entries)
entry             300 drivers/staging/uwb/est.c 		.entry = entry,
entry             345 drivers/staging/uwb/est.c 	const struct uwb_est_entry *entry;
entry             355 drivers/staging/uwb/est.c 	entry = &est->entry[event_low];
entry             356 drivers/staging/uwb/est.c 	if (entry->size == 0 && entry->offset == 0) {	/* unknown? */
entry             362 drivers/staging/uwb/est.c 	offset = entry->offset;	/* extra fries with that? */
entry             364 drivers/staging/uwb/est.c 		size = entry->size;
entry             371 drivers/staging/uwb/est.c 		switch (entry->type) {
entry             383 drivers/staging/uwb/est.c 		size = entry->size;
entry             385 drivers/staging/uwb/est.c 		switch (entry->type) {
entry              48 drivers/staging/uwb/lc-dev.c int uwb_notifs_deregister(struct uwb_rc *rc, struct uwb_notifs_handler *entry)
entry              52 drivers/staging/uwb/lc-dev.c 	list_del(&entry->list_node);
entry              27 drivers/staging/vt6655/key.c 	u32 entry = 0;
entry              43 drivers/staging/vt6655/key.c 				entry = key->hw_key_idx;
entry              52 drivers/staging/vt6655/key.c 		entry = MAX_KEY_TABLE - 1;
entry              53 drivers/staging/vt6655/key.c 		key->hw_key_idx = entry;
entry              85 drivers/staging/vt6655/key.c 	MACvSetKeyEntry(priv, key_mode, entry, key_inx,
entry              43 drivers/staging/vt6656/key.c 	u32 entry = 0;
entry              59 drivers/staging/vt6656/key.c 				entry = key->hw_key_idx;
entry              68 drivers/staging/vt6656/key.c 		entry = MAX_KEY_TABLE - 1;
entry              69 drivers/staging/vt6656/key.c 		key->hw_key_idx = entry;
entry             101 drivers/staging/vt6656/key.c 	vnt_mac_set_keyentry(priv, key_mode, entry, key_inx, bssid, key->key);
entry             960 drivers/target/target_core_user.c 	struct tcmu_cmd_entry *entry;
entry            1023 drivers/target/target_core_user.c 		entry = (void *) mb + CMDR_OFF + cmd_head;
entry            1024 drivers/target/target_core_user.c 		tcmu_hdr_set_op(&entry->hdr.len_op, TCMU_OP_PAD);
entry            1025 drivers/target/target_core_user.c 		tcmu_hdr_set_len(&entry->hdr.len_op, pad_size);
entry            1026 drivers/target/target_core_user.c 		entry->hdr.cmd_id = 0; /* not used for PAD */
entry            1027 drivers/target/target_core_user.c 		entry->hdr.kflags = 0;
entry            1028 drivers/target/target_core_user.c 		entry->hdr.uflags = 0;
entry            1029 drivers/target/target_core_user.c 		tcmu_flush_dcache_range(entry, sizeof(*entry));
entry            1038 drivers/target/target_core_user.c 	entry = (void *) mb + CMDR_OFF + cmd_head;
entry            1039 drivers/target/target_core_user.c 	memset(entry, 0, command_size);
entry            1040 drivers/target/target_core_user.c 	tcmu_hdr_set_op(&entry->hdr.len_op, TCMU_OP_CMD);
entry            1044 drivers/target/target_core_user.c 	iov = &entry->req.iov[0];
entry            1051 drivers/target/target_core_user.c 	entry->req.iov_cnt = iov_cnt;
entry            1061 drivers/target/target_core_user.c 	entry->req.iov_bidi_cnt = iov_cnt;
entry            1071 drivers/target/target_core_user.c 	entry->hdr.cmd_id = tcmu_cmd->cmd_id;
entry            1077 drivers/target/target_core_user.c 	base_command_size = tcmu_cmd_get_base_cmd_size(entry->req.iov_cnt +
entry            1078 drivers/target/target_core_user.c 						       entry->req.iov_bidi_cnt);
entry            1081 drivers/target/target_core_user.c 	tcmu_hdr_set_len(&entry->hdr.len_op, command_size);
entry            1086 drivers/target/target_core_user.c 	entry->req.cdb_off = cdb_off;
entry            1087 drivers/target/target_core_user.c 	tcmu_flush_dcache_range(entry, sizeof(*entry));
entry            1130 drivers/target/target_core_user.c static void tcmu_handle_completion(struct tcmu_cmd *cmd, struct tcmu_cmd_entry *entry)
entry            1150 drivers/target/target_core_user.c 	if (entry->hdr.uflags & TCMU_UFLAG_UNKNOWN_OP) {
entry            1153 drivers/target/target_core_user.c 		entry->rsp.scsi_status = SAM_STAT_CHECK_CONDITION;
entry            1159 drivers/target/target_core_user.c 	    (entry->hdr.uflags & TCMU_UFLAG_READ_LEN) && entry->rsp.read_len) {
entry            1161 drivers/target/target_core_user.c 		if (entry->rsp.read_len < read_len)
entry            1162 drivers/target/target_core_user.c 			read_len = entry->rsp.read_len;
entry            1165 drivers/target/target_core_user.c 	if (entry->rsp.scsi_status == SAM_STAT_CHECK_CONDITION) {
entry            1166 drivers/target/target_core_user.c 		transport_copy_sense_to_cmd(se_cmd, entry->rsp.sense_buffer);
entry            1188 drivers/target/target_core_user.c 					entry->rsp.scsi_status, read_len);
entry            1190 drivers/target/target_core_user.c 		target_complete_cmd(cmd->se_cmd, entry->rsp.scsi_status);
entry            1233 drivers/target/target_core_user.c 		struct tcmu_cmd_entry *entry = (void *) mb + CMDR_OFF + udev->cmdr_last_cleaned;
entry            1235 drivers/target/target_core_user.c 		tcmu_flush_dcache_range(entry, sizeof(*entry));
entry            1237 drivers/target/target_core_user.c 		if (tcmu_hdr_get_op(entry->hdr.len_op) == TCMU_OP_PAD) {
entry            1239 drivers/target/target_core_user.c 				    tcmu_hdr_get_len(entry->hdr.len_op),
entry            1243 drivers/target/target_core_user.c 		WARN_ON(tcmu_hdr_get_op(entry->hdr.len_op) != TCMU_OP_CMD);
entry            1245 drivers/target/target_core_user.c 		cmd = idr_remove(&udev->commands, entry->hdr.cmd_id);
entry            1248 drivers/target/target_core_user.c 			       entry->hdr.cmd_id);
entry            1253 drivers/target/target_core_user.c 		tcmu_handle_completion(cmd, entry);
entry            1256 drivers/target/target_core_user.c 			    tcmu_hdr_get_len(entry->hdr.len_op),
entry             308 drivers/thunderbolt/eeprom.c 	const struct tb_drom_entry_generic *entry =
entry             314 drivers/thunderbolt/eeprom.c 		sw->vendor_name = kstrndup(entry->data,
entry             321 drivers/thunderbolt/eeprom.c 		sw->device_name = kstrndup(entry->data,
entry             358 drivers/thunderbolt/eeprom.c 		struct tb_drom_entry_port *entry = (void *) header;
entry             359 drivers/thunderbolt/eeprom.c 		if (header->len != sizeof(*entry)) {
entry             365 drivers/thunderbolt/eeprom.c 		port->link_nr = entry->link_nr;
entry             366 drivers/thunderbolt/eeprom.c 		if (entry->has_dual_link_port)
entry             368 drivers/thunderbolt/eeprom.c 				&port->sw->ports[entry->dual_link_port_nr];
entry             386 drivers/thunderbolt/eeprom.c 		struct tb_drom_entry_header *entry = (void *) (sw->drom + pos);
entry             387 drivers/thunderbolt/eeprom.c 		if (pos + 1 == drom_size || pos + entry->len > drom_size
entry             388 drivers/thunderbolt/eeprom.c 				|| !entry->len) {
entry             393 drivers/thunderbolt/eeprom.c 		switch (entry->type) {
entry             395 drivers/thunderbolt/eeprom.c 			res = tb_drom_parse_entry_generic(sw, entry);
entry             398 drivers/thunderbolt/eeprom.c 			res = tb_drom_parse_entry_port(sw, entry);
entry             404 drivers/thunderbolt/eeprom.c 		pos += entry->len;
entry              52 drivers/thunderbolt/property.c static bool tb_property_entry_valid(const struct tb_property_entry *entry,
entry              55 drivers/thunderbolt/property.c 	switch (entry->type) {
entry              59 drivers/thunderbolt/property.c 		if (entry->length > block_len)
entry              61 drivers/thunderbolt/property.c 		if (entry->value + entry->length > block_len)
entry              66 drivers/thunderbolt/property.c 		if (entry->length != 1)
entry              96 drivers/thunderbolt/property.c 					const struct tb_property_entry *entry)
entry             102 drivers/thunderbolt/property.c 	if (!tb_property_entry_valid(entry, block_len))
entry             105 drivers/thunderbolt/property.c 	parse_dwdata(key, entry, 2);
entry             108 drivers/thunderbolt/property.c 	property = tb_property_alloc(key, entry->type);
entry             112 drivers/thunderbolt/property.c 	property->length = entry->length;
entry             116 drivers/thunderbolt/property.c 		dir = __tb_property_parse_dir(block, block_len, entry->value,
entry             117 drivers/thunderbolt/property.c 					      entry->length, false);
entry             132 drivers/thunderbolt/property.c 		parse_dwdata(property->value.data, block + entry->value,
entry             133 drivers/thunderbolt/property.c 			     entry->length);
entry             143 drivers/thunderbolt/property.c 		parse_dwdata(property->value.text, block + entry->value,
entry             144 drivers/thunderbolt/property.c 			     entry->length);
entry             150 drivers/thunderbolt/property.c 		property->value.immediate = entry->value;
entry             353 drivers/thunderbolt/property.c 	struct tb_property_entry *entry;
entry             417 drivers/thunderbolt/property.c 		entry = pe->entries;
entry             424 drivers/thunderbolt/property.c 		entry = re->entries;
entry             430 drivers/thunderbolt/property.c 		format_dwdata(entry, property->key, 2);
entry             431 drivers/thunderbolt/property.c 		entry->type = property->type;
entry             440 drivers/thunderbolt/property.c 			entry->length = tb_property_dir_length(child, false,
entry             442 drivers/thunderbolt/property.c 			entry->value = dir_end;
entry             449 drivers/thunderbolt/property.c 			entry->length = property->length;
entry             450 drivers/thunderbolt/property.c 			entry->value = data_offset;
entry             451 drivers/thunderbolt/property.c 			data_offset += entry->length;
entry             457 drivers/thunderbolt/property.c 			entry->length = property->length;
entry             458 drivers/thunderbolt/property.c 			entry->value = data_offset;
entry             459 drivers/thunderbolt/property.c 			data_offset += entry->length;
entry             463 drivers/thunderbolt/property.c 			entry->length = property->length;
entry             464 drivers/thunderbolt/property.c 			entry->value = property->value.immediate;
entry             471 drivers/thunderbolt/property.c 		entry++;
entry              53 drivers/tty/hvc/hvc_xen.c 	struct xencons_info *entry, *n, *ret = NULL;
entry              58 drivers/tty/hvc/hvc_xen.c 	list_for_each_entry_safe(entry, n, &xenconsoles, list) {
entry              59 drivers/tty/hvc/hvc_xen.c 		if (entry->vtermno == vtermno) {
entry              60 drivers/tty/hvc/hvc_xen.c 			ret  = entry;
entry            1038 drivers/tty/serial/msm_serial.c 	const struct msm_baud_map *entry, *end, *best;
entry            1063 drivers/tty/serial/msm_serial.c 	entry = table;
entry            1064 drivers/tty/serial/msm_serial.c 	while (entry < end) {
entry            1065 drivers/tty/serial/msm_serial.c 		if (entry->divisor <= divisor) {
entry            1066 drivers/tty/serial/msm_serial.c 			result = target / entry->divisor / 16;
entry            1072 drivers/tty/serial/msm_serial.c 				best = entry;
entry            1078 drivers/tty/serial/msm_serial.c 		} else if (entry->divisor > divisor) {
entry            1089 drivers/tty/serial/msm_serial.c 			entry = table;
entry            1093 drivers/tty/serial/msm_serial.c 		entry++;
entry            1105 drivers/tty/serial/msm_serial.c 	const struct msm_baud_map *entry;
entry            1111 drivers/tty/serial/msm_serial.c 	entry = msm_find_best_baud(port, baud, &rate);
entry            1113 drivers/tty/serial/msm_serial.c 	baud = rate / 16 / entry->divisor;
entry            1119 drivers/tty/serial/msm_serial.c 	msm_write(port, entry->code, UART_CSR);
entry            1122 drivers/tty/serial/msm_serial.c 	rxstale = entry->rxstale;
entry             105 drivers/uio/uio.c 	struct map_sysfs_entry *entry;
entry             107 drivers/uio/uio.c 	entry = container_of(attr, struct map_sysfs_entry, attr);
entry             109 drivers/uio/uio.c 	if (!entry->show)
entry             112 drivers/uio/uio.c 	return entry->show(mem, buf);
entry             193 drivers/uio/uio.c 	struct portio_sysfs_entry *entry;
entry             195 drivers/uio/uio.c 	entry = container_of(attr, struct portio_sysfs_entry, attr);
entry             197 drivers/uio/uio.c 	if (!entry->show)
entry             200 drivers/uio/uio.c 	return entry->show(port, buf);
entry             229 drivers/usb/early/xhci-dbc.c 	struct xdbc_erst_entry *entry;
entry             249 drivers/usb/early/xhci-dbc.c 	entry = (struct xdbc_erst_entry *)xdbc.erst_base;
entry             251 drivers/usb/early/xhci-dbc.c 	entry->seg_addr		= cpu_to_le64(xdbc.evt_seg.dma);
entry             252 drivers/usb/early/xhci-dbc.c 	entry->seg_size		= cpu_to_le32(XDBC_TRBS_PER_SEGMENT);
entry             253 drivers/usb/early/xhci-dbc.c 	entry->__reserved_0	= 0;
entry            1537 drivers/usb/gadget/composite.c 			list_for_each_entry(ext_prop, &d->ext_prop, entry) {
entry            1062 drivers/usb/gadget/configfs.c 	list_add_tail(&ext_prop->entry, &desc->ext_prop);
entry            1077 drivers/usb/gadget/configfs.c 	list_del(&ext_prop->entry);
entry            3033 drivers/usb/gadget/function/f_fs.c 		list_add_tail(&ext_prop->entry, &t->os_desc->ext_prop);
entry            3452 drivers/usb/gadget/function/f_fs.c 	list_for_each_entry(dev, &ffs_devices, entry) {
entry            3468 drivers/usb/gadget/function/f_fs.c 		dev = list_first_entry(&ffs_devices, struct ffs_dev, entry);
entry            3657 drivers/usb/gadget/function/f_fs.c 	list_add(&dev->entry, &ffs_devices);
entry            3703 drivers/usb/gadget/function/f_fs.c 	list_del(&dev->entry);
entry              42 drivers/usb/gadget/function/u_fs.h 	struct list_head entry;
entry             850 drivers/usb/gadget/function/uvc_configfs.c 	struct list_head	entry;
entry             929 drivers/usb/gadget/function/uvc_configfs.c 	list_add_tail(&format_ptr->entry, &src_hdr->formats);
entry             961 drivers/usb/gadget/function/uvc_configfs.c 	list_for_each_entry_safe(format_ptr, tmp, &src_hdr->formats, entry)
entry             963 drivers/usb/gadget/function/uvc_configfs.c 			list_del(&format_ptr->entry);
entry            1999 drivers/usb/gadget/function/uvc_configfs.c 	list_for_each_entry(f, &h->formats, entry) {
entry            2098 drivers/usb/gadget/function/uvc_configfs.c 		list_for_each_entry(f, &h->formats, entry) {
entry             221 drivers/usb/gadget/legacy/hid.c 	struct hidg_func_node *entry;
entry             228 drivers/usb/gadget/legacy/hid.c 	entry = kzalloc(sizeof(*entry), GFP_KERNEL);
entry             229 drivers/usb/gadget/legacy/hid.c 	if (!entry)
entry             232 drivers/usb/gadget/legacy/hid.c 	entry->func = func;
entry             233 drivers/usb/gadget/legacy/hid.c 	list_add_tail(&entry->node, &hidg_func_list);
entry             375 drivers/usb/host/ehci-dbg.c 	struct list_head	*entry;
entry             416 drivers/usb/host/ehci-dbg.c 	list_for_each(entry, &qh->qtd_list) {
entry             419 drivers/usb/host/ehci-dbg.c 		td = list_entry(entry, struct ehci_qtd, qtd_list);
entry             292 drivers/usb/host/ehci-q.c 	struct list_head	*entry, *tmp;
entry             322 drivers/usb/host/ehci-q.c 	list_for_each_safe (entry, tmp, &qh->qtd_list) {
entry             327 drivers/usb/host/ehci-q.c 		qtd = list_entry (entry, struct ehci_qtd, qtd_list);
entry             560 drivers/usb/host/ehci-q.c 	struct list_head	*entry, *temp;
entry             562 drivers/usb/host/ehci-q.c 	list_for_each_safe (entry, temp, qtd_list) {
entry             565 drivers/usb/host/ehci-q.c 		qtd = list_entry (entry, struct ehci_qtd, qtd_list);
entry             415 drivers/usb/host/ohci-dbg.c 		struct list_head *entry;
entry             433 drivers/usb/host/ohci-dbg.c 		list_for_each (entry, &ed->td_list) {
entry             436 drivers/usb/host/ohci-dbg.c 			td = list_entry (entry, struct td, td_list);
entry             532 drivers/usb/host/ohci-dbg.c 				struct list_head	*entry;
entry             536 drivers/usb/host/ohci-dbg.c 				list_for_each (entry, &ed->td_list)
entry             982 drivers/usb/host/ohci-q.c 		struct list_head	*entry, *tmp;
entry            1037 drivers/usb/host/ohci-q.c 		list_for_each_safe (entry, tmp, &ed->td_list) {
entry            1044 drivers/usb/host/ohci-q.c 			td = list_entry (entry, struct td, td_list);
entry            1805 drivers/usb/host/xhci-mem.c 	struct xhci_erst_entry *entry;
entry            1817 drivers/usb/host/xhci-mem.c 		entry = &erst->entries[val];
entry            1818 drivers/usb/host/xhci-mem.c 		entry->seg_addr = cpu_to_le64(seg->dma);
entry            1819 drivers/usb/host/xhci-mem.c 		entry->seg_size = cpu_to_le32(TRBS_PER_SEGMENT);
entry            1820 drivers/usb/host/xhci-mem.c 		entry->rsvd = 0;
entry             334 drivers/usb/storage/ene_ub6250.c 	struct ms_bootblock_sysent_rec entry[MS_NUMBER_OF_SYSTEM_ENTRY];
entry             992 drivers/usb/storage/ene_ub6250.c 		EntryOffset = be32_to_cpu(SysEntry->entry[i].dwStart);
entry             996 drivers/usb/storage/ene_ub6250.c 		EntrySize = be32_to_cpu(SysEntry->entry[i].dwSize);
entry            1008 drivers/usb/storage/ene_ub6250.c 			if (SysEntry->entry[i].bType != MS_SYSENT_TYPE_INVALID_BLOCK)
entry            1038 drivers/usb/storage/ene_ub6250.c 			if (SysEntry->entry[i].bType != MS_SYSENT_TYPE_CIS_IDI)
entry             210 drivers/vfio/vfio_iommu_spapr_tce.c 			unsigned long entry = ioba >> tbl->it_page_shift;
entry             214 drivers/vfio/vfio_iommu_spapr_tce.c 			if ((start <= entry) && (entry < end)) {
entry             338 drivers/vfio/vfio_iommu_spapr_tce.c 		unsigned long entry, unsigned long pages);
entry             410 drivers/vfio/vfio_iommu_spapr_tce.c 		struct iommu_table *tbl, unsigned long entry)
entry             415 drivers/vfio/vfio_iommu_spapr_tce.c 	__be64 *pua = IOMMU_TABLE_USERSPACE_ENTRY_RO(tbl, entry);
entry             424 drivers/vfio/vfio_iommu_spapr_tce.c 				__func__, be64_to_cpu(*pua), entry, ret);
entry             433 drivers/vfio/vfio_iommu_spapr_tce.c 		unsigned long entry, unsigned long pages)
entry             438 drivers/vfio/vfio_iommu_spapr_tce.c 	unsigned long lastentry = entry + pages, firstentry = entry;
entry             440 drivers/vfio/vfio_iommu_spapr_tce.c 	for ( ; entry < lastentry; ++entry) {
entry             451 drivers/vfio/vfio_iommu_spapr_tce.c 					entry);
entry             454 drivers/vfio/vfio_iommu_spapr_tce.c 				entry |= tbl->it_level_size - 1;
entry             463 drivers/vfio/vfio_iommu_spapr_tce.c 		ret = iommu_tce_xchg_no_kill(container->mm, tbl, entry, &oldhpa,
entry             472 drivers/vfio/vfio_iommu_spapr_tce.c 			tce_iommu_unuse_page_v2(container, tbl, entry);
entry             501 drivers/vfio/vfio_iommu_spapr_tce.c 		unsigned long entry, unsigned long tce, unsigned long pages,
entry             523 drivers/vfio/vfio_iommu_spapr_tce.c 		ret = iommu_tce_xchg_no_kill(container->mm, tbl, entry + i,
entry             528 drivers/vfio/vfio_iommu_spapr_tce.c 					__func__, entry << tbl->it_page_shift,
entry             540 drivers/vfio/vfio_iommu_spapr_tce.c 		tce_iommu_clear(container, tbl, entry, i);
entry             542 drivers/vfio/vfio_iommu_spapr_tce.c 		iommu_tce_kill(tbl, entry, pages);
entry             549 drivers/vfio/vfio_iommu_spapr_tce.c 		unsigned long entry, unsigned long tce, unsigned long pages,
entry             558 drivers/vfio/vfio_iommu_spapr_tce.c 		__be64 *pua = IOMMU_TABLE_USERSPACE_ENTRY(tbl, entry + i);
entry             579 drivers/vfio/vfio_iommu_spapr_tce.c 		ret = iommu_tce_xchg_no_kill(container->mm, tbl, entry + i,
entry             583 drivers/vfio/vfio_iommu_spapr_tce.c 			tce_iommu_unuse_page_v2(container, tbl, entry + i);
entry             585 drivers/vfio/vfio_iommu_spapr_tce.c 					__func__, entry << tbl->it_page_shift,
entry             591 drivers/vfio/vfio_iommu_spapr_tce.c 			tce_iommu_unuse_page_v2(container, tbl, entry + i);
entry             599 drivers/vfio/vfio_iommu_spapr_tce.c 		tce_iommu_clear(container, tbl, entry, i);
entry             601 drivers/vfio/vfio_iommu_spapr_tce.c 		iommu_tce_kill(tbl, entry, pages);
entry             666 drivers/vfio/vfio_iommu_type1.c 	struct vfio_regions *entry, *next;
entry             670 drivers/vfio/vfio_iommu_type1.c 	list_for_each_entry_safe(entry, next, regions, list) {
entry             672 drivers/vfio/vfio_iommu_type1.c 						    entry->iova,
entry             673 drivers/vfio/vfio_iommu_type1.c 						    entry->phys >> PAGE_SHIFT,
entry             674 drivers/vfio/vfio_iommu_type1.c 						    entry->len >> PAGE_SHIFT,
entry             676 drivers/vfio/vfio_iommu_type1.c 		list_del(&entry->list);
entry             677 drivers/vfio/vfio_iommu_type1.c 		kfree(entry);
entry             702 drivers/vfio/vfio_iommu_type1.c 	struct vfio_regions *entry = kzalloc(sizeof(*entry), GFP_KERNEL);
entry             704 drivers/vfio/vfio_iommu_type1.c 	if (entry) {
entry             709 drivers/vfio/vfio_iommu_type1.c 			kfree(entry);
entry             711 drivers/vfio/vfio_iommu_type1.c 			entry->iova = *iova;
entry             712 drivers/vfio/vfio_iommu_type1.c 			entry->phys = phys;
entry             713 drivers/vfio/vfio_iommu_type1.c 			entry->len  = unmapped;
entry             714 drivers/vfio/vfio_iommu_type1.c 			list_add_tail(&entry->list, unmapped_list);
entry             179 drivers/video/backlight/arcxcnn_bl.c 	u32 prog_val, num_entry, entry, sources[ARCXCNN_LEDEN_BITS];
entry             236 drivers/video/backlight/arcxcnn_bl.c 		for (entry = 0; entry < num_entry; entry++) {
entry             237 drivers/video/backlight/arcxcnn_bl.c 			u8 onbit = 1 << sources[entry];
entry             407 drivers/video/backlight/backlight.c 	list_add(&new_bd->entry, &backlight_dev_list);
entry             423 drivers/video/backlight/backlight.c 	list_for_each_entry(bd, &backlight_dev_list, entry) {
entry             447 drivers/video/backlight/backlight.c 	list_del(&bd->entry);
entry              83 drivers/video/fbdev/i810/i810_main.c 			   const struct pci_device_id *entry);
entry            1839 drivers/video/fbdev/i810/i810_main.c 				      const struct pci_device_id *entry)
entry            1865 drivers/video/fbdev/i810/i810_main.c 				i810_pci_list[entry->driver_data])) {
entry            1880 drivers/video/fbdev/i810/i810_main.c 				i810_pci_list[entry->driver_data])) {
entry            2012 drivers/video/fbdev/i810/i810_main.c 			   const struct pci_device_id *entry)
entry            2035 drivers/video/fbdev/i810/i810_main.c 	if ((err = i810_allocate_pci_resource(par, entry))) {
entry            2085 drivers/video/fbdev/i810/i810_main.c 	       i810_pci_list[entry->driver_data],
entry              77 drivers/video/fbdev/omap/hwa742.c 	struct list_head entry;
entry             241 drivers/video/fbdev/omap/hwa742.c 			 struct hwa742_request, entry);
entry             242 drivers/video/fbdev/omap/hwa742.c 	list_del(&req->entry);
entry             245 drivers/video/fbdev/omap/hwa742.c 	INIT_LIST_HEAD(&req->entry);
entry             257 drivers/video/fbdev/omap/hwa742.c 	list_move(&req->entry, &hwa742.free_req_list);
entry             276 drivers/video/fbdev/omap/hwa742.c 				 struct hwa742_request, entry);
entry             412 drivers/video/fbdev/omap/hwa742.c 	list_add_tail(&req->entry, req_head);	\
entry             470 drivers/video/fbdev/omap/hwa742.c 	last = list_entry(req_list.prev, struct hwa742_request, entry);
entry             501 drivers/video/fbdev/omap/hwa742.c 	last = list_entry(req_list.prev, struct hwa742_request, entry);
entry             554 drivers/video/fbdev/omap/hwa742.c 	list_add(&req->entry, &req_list);
entry            1002 drivers/video/fbdev/omap/hwa742.c 		list_add(&hwa742.req_pool[i].entry, &hwa742.free_req_list);
entry              77 drivers/video/fbdev/smscufx.c 	struct list_head entry;
entry            1796 drivers/video/fbdev/smscufx.c 	list_add_tail(&unode->entry, &dev->urbs.list);
entry            1833 drivers/video/fbdev/smscufx.c 		unode = list_entry(node, struct urb_node, entry);
entry            1885 drivers/video/fbdev/smscufx.c 		list_add_tail(&unode->entry, &dev->urbs.list);
entry            1902 drivers/video/fbdev/smscufx.c 	struct list_head *entry;
entry            1919 drivers/video/fbdev/smscufx.c 	entry = dev->urbs.list.next;
entry            1920 drivers/video/fbdev/smscufx.c 	list_del_init(entry);
entry            1925 drivers/video/fbdev/smscufx.c 	unode = list_entry(entry, struct urb_node, entry);
entry            1837 drivers/video/fbdev/udlfb.c 	list_add_tail(&unode->entry, &dlfb->urbs.list);
entry            1862 drivers/video/fbdev/udlfb.c 		unode = list_entry(node, struct urb_node, entry);
entry            1923 drivers/video/fbdev/udlfb.c 		list_add_tail(&unode->entry, &dlfb->urbs.list);
entry            1936 drivers/video/fbdev/udlfb.c 	struct list_head *entry;
entry            1952 drivers/video/fbdev/udlfb.c 	entry = dlfb->urbs.list.next;
entry            1953 drivers/video/fbdev/udlfb.c 	list_del_init(entry);
entry            1958 drivers/video/fbdev/udlfb.c 	unode = list_entry(entry, struct urb_node, entry);
entry             991 drivers/video/fbdev/uvesafb.c 	struct uvesafb_pal_entry entry;
entry             999 drivers/video/fbdev/uvesafb.c 		entry.red   = red   >> shift;
entry            1000 drivers/video/fbdev/uvesafb.c 		entry.green = green >> shift;
entry            1001 drivers/video/fbdev/uvesafb.c 		entry.blue  = blue  >> shift;
entry            1002 drivers/video/fbdev/uvesafb.c 		entry.pad   = 0;
entry            1004 drivers/video/fbdev/uvesafb.c 		err = uvesafb_setpalette(&entry, 1, regno, info);
entry            1070 drivers/video/fbdev/vermilion/vermilion.c 	struct vml_info *entry;
entry            1089 drivers/video/fbdev/vermilion/vermilion.c 		entry = list_entry(list, struct vml_info, head);
entry            1096 drivers/video/fbdev/vermilion/vermilion.c 		if (!vmlfb_check_var_locked(&entry->info.var, entry)) {
entry            1097 drivers/video/fbdev/vermilion/vermilion.c 			vmlfb_set_par_locked(entry);
entry            1107 drivers/video/fbdev/vermilion/vermilion.c 			save_activate = entry->info.var.activate;
entry            1108 drivers/video/fbdev/vermilion/vermilion.c 			entry->info.var.bits_per_pixel = 16;
entry            1109 drivers/video/fbdev/vermilion/vermilion.c 			vmlfb_set_pref_pixel_format(&entry->info.var);
entry            1110 drivers/video/fbdev/vermilion/vermilion.c 			if (fb_find_mode(&entry->info.var,
entry            1111 drivers/video/fbdev/vermilion/vermilion.c 					 &entry->info,
entry            1113 drivers/video/fbdev/vermilion/vermilion.c 				entry->info.var.activate |=
entry            1115 drivers/video/fbdev/vermilion/vermilion.c 				fb_set_var(&entry->info, &entry->info.var);
entry            1120 drivers/video/fbdev/vermilion/vermilion.c 			entry->info.var.activate = save_activate;
entry            1123 drivers/video/fbdev/vermilion/vermilion.c 		vmlfb_blank_locked(entry);
entry            1137 drivers/video/fbdev/vermilion/vermilion.c 	struct vml_info *entry, *next;
entry            1146 drivers/video/fbdev/vermilion/vermilion.c 	list_for_each_entry_safe(entry, next, &global_has_mode, head) {
entry            1148 drivers/video/fbdev/vermilion/vermilion.c 		vmlfb_disable_pipe(entry);
entry            1149 drivers/video/fbdev/vermilion/vermilion.c 		list_move_tail(&entry->head, &global_no_mode);
entry             108 drivers/video/fbdev/vesafb.c 		struct { u_char blue, green, red, pad; } entry;
entry             110 drivers/video/fbdev/vesafb.c 		entry.red   = red   >> shift;
entry             111 drivers/video/fbdev/vesafb.c 		entry.green = green >> shift;
entry             112 drivers/video/fbdev/vesafb.c 		entry.blue  = blue  >> shift;
entry             113 drivers/video/fbdev/vesafb.c 		entry.pad   = 0;
entry             121 drivers/video/fbdev/vesafb.c                   "D" (&entry),         /* EDI */
entry             146 drivers/video/of_display_timing.c 	struct device_node *entry;
entry             165 drivers/video/of_display_timing.c 	entry = of_parse_phandle(timings_np, "native-mode", 0);
entry             167 drivers/video/of_display_timing.c 	if (!entry)
entry             168 drivers/video/of_display_timing.c 		entry = of_get_next_child(timings_np, NULL);
entry             170 drivers/video/of_display_timing.c 	if (!entry) {
entry             175 drivers/video/of_display_timing.c 	pr_debug("%pOF: using %pOFn as default timing\n", np, entry);
entry             177 drivers/video/of_display_timing.c 	native_mode = entry;
entry             197 drivers/video/of_display_timing.c 	for_each_child_of_node(timings_np, entry) {
entry             208 drivers/video/of_display_timing.c 		r = of_parse_display_timing(entry, dt);
entry             220 drivers/video/of_display_timing.c 		if (native_mode == entry)
entry            1021 drivers/vme/bridges/vme_ca91cx42.c 	struct ca91cx42_dma_entry *entry, *prev;
entry            1031 drivers/vme/bridges/vme_ca91cx42.c 	entry = kmalloc(sizeof(*entry), GFP_KERNEL);
entry            1032 drivers/vme/bridges/vme_ca91cx42.c 	if (!entry) {
entry            1038 drivers/vme/bridges/vme_ca91cx42.c 	if ((unsigned long)&entry->descriptor & CA91CX42_DCPP_M) {
entry            1040 drivers/vme/bridges/vme_ca91cx42.c 			"required: %p\n", &entry->descriptor);
entry            1045 drivers/vme/bridges/vme_ca91cx42.c 	memset(&entry->descriptor, 0, sizeof(entry->descriptor));
entry            1048 drivers/vme/bridges/vme_ca91cx42.c 		entry->descriptor.dctl |= CA91CX42_DCTL_L2V;
entry            1085 drivers/vme/bridges/vme_ca91cx42.c 		entry->descriptor.dctl |= CA91CX42_DCTL_VCT_BLT;
entry            1090 drivers/vme/bridges/vme_ca91cx42.c 		entry->descriptor.dctl |= CA91CX42_DCTL_VDW_D8;
entry            1093 drivers/vme/bridges/vme_ca91cx42.c 		entry->descriptor.dctl |= CA91CX42_DCTL_VDW_D16;
entry            1096 drivers/vme/bridges/vme_ca91cx42.c 		entry->descriptor.dctl |= CA91CX42_DCTL_VDW_D32;
entry            1099 drivers/vme/bridges/vme_ca91cx42.c 		entry->descriptor.dctl |= CA91CX42_DCTL_VDW_D64;
entry            1109 drivers/vme/bridges/vme_ca91cx42.c 		entry->descriptor.dctl |= CA91CX42_DCTL_VAS_A16;
entry            1112 drivers/vme/bridges/vme_ca91cx42.c 		entry->descriptor.dctl |= CA91CX42_DCTL_VAS_A24;
entry            1115 drivers/vme/bridges/vme_ca91cx42.c 		entry->descriptor.dctl |= CA91CX42_DCTL_VAS_A32;
entry            1118 drivers/vme/bridges/vme_ca91cx42.c 		entry->descriptor.dctl |= CA91CX42_DCTL_VAS_USER1;
entry            1121 drivers/vme/bridges/vme_ca91cx42.c 		entry->descriptor.dctl |= CA91CX42_DCTL_VAS_USER2;
entry            1130 drivers/vme/bridges/vme_ca91cx42.c 		entry->descriptor.dctl |= CA91CX42_DCTL_SUPER_SUPR;
entry            1132 drivers/vme/bridges/vme_ca91cx42.c 		entry->descriptor.dctl |= CA91CX42_DCTL_PGM_PGM;
entry            1134 drivers/vme/bridges/vme_ca91cx42.c 	entry->descriptor.dtbc = count;
entry            1135 drivers/vme/bridges/vme_ca91cx42.c 	entry->descriptor.dla = pci_attr->address;
entry            1136 drivers/vme/bridges/vme_ca91cx42.c 	entry->descriptor.dva = vme_attr->address;
entry            1137 drivers/vme/bridges/vme_ca91cx42.c 	entry->descriptor.dcpp = CA91CX42_DCPP_NULL;
entry            1140 drivers/vme/bridges/vme_ca91cx42.c 	list_add_tail(&entry->list, &list->entries);
entry            1143 drivers/vme/bridges/vme_ca91cx42.c 	if (entry->list.prev != &list->entries) {
entry            1144 drivers/vme/bridges/vme_ca91cx42.c 		prev = list_entry(entry->list.prev, struct ca91cx42_dma_entry,
entry            1147 drivers/vme/bridges/vme_ca91cx42.c 		desc_ptr = virt_to_bus(&entry->descriptor);
entry            1157 drivers/vme/bridges/vme_ca91cx42.c 	kfree(entry);
entry            1180 drivers/vme/bridges/vme_ca91cx42.c 	struct ca91cx42_dma_entry *entry;
entry            1208 drivers/vme/bridges/vme_ca91cx42.c 	entry = list_first_entry(&list->entries, struct ca91cx42_dma_entry,
entry            1211 drivers/vme/bridges/vme_ca91cx42.c 	bus_addr = virt_to_bus(&entry->descriptor);
entry            1274 drivers/vme/bridges/vme_ca91cx42.c 	struct ca91cx42_dma_entry *entry;
entry            1279 drivers/vme/bridges/vme_ca91cx42.c 		entry = list_entry(pos, struct ca91cx42_dma_entry, list);
entry            1280 drivers/vme/bridges/vme_ca91cx42.c 		kfree(entry);
entry            1627 drivers/vme/bridges/vme_tsi148.c 	struct tsi148_dma_entry *entry, *prev;
entry            1638 drivers/vme/bridges/vme_tsi148.c 	entry = kmalloc(sizeof(*entry), GFP_KERNEL);
entry            1639 drivers/vme/bridges/vme_tsi148.c 	if (!entry) {
entry            1645 drivers/vme/bridges/vme_tsi148.c 	if ((unsigned long)&entry->descriptor & 0x7) {
entry            1648 drivers/vme/bridges/vme_tsi148.c 			&entry->descriptor);
entry            1656 drivers/vme/bridges/vme_tsi148.c 	memset(&entry->descriptor, 0, sizeof(entry->descriptor));
entry            1663 drivers/vme/bridges/vme_tsi148.c 		entry->descriptor.dsal = cpu_to_be32(pattern_attr->pattern);
entry            1674 drivers/vme/bridges/vme_tsi148.c 		entry->descriptor.dsat = cpu_to_be32(val);
entry            1681 drivers/vme/bridges/vme_tsi148.c 		entry->descriptor.dsau = cpu_to_be32(address_high);
entry            1682 drivers/vme/bridges/vme_tsi148.c 		entry->descriptor.dsal = cpu_to_be32(address_low);
entry            1683 drivers/vme/bridges/vme_tsi148.c 		entry->descriptor.dsat = cpu_to_be32(TSI148_LCSR_DSAT_TYP_PCI);
entry            1690 drivers/vme/bridges/vme_tsi148.c 		entry->descriptor.dsau = cpu_to_be32(address_high);
entry            1691 drivers/vme/bridges/vme_tsi148.c 		entry->descriptor.dsal = cpu_to_be32(address_low);
entry            1692 drivers/vme/bridges/vme_tsi148.c 		entry->descriptor.dsat = cpu_to_be32(TSI148_LCSR_DSAT_TYP_VME);
entry            1695 drivers/vme/bridges/vme_tsi148.c 			tsi148_bridge->parent, &entry->descriptor.dsat,
entry            1708 drivers/vme/bridges/vme_tsi148.c 	entry->descriptor.dnlau = cpu_to_be32(0);
entry            1709 drivers/vme/bridges/vme_tsi148.c 	entry->descriptor.dnlal = cpu_to_be32(TSI148_LCSR_DNLAL_LLA);
entry            1718 drivers/vme/bridges/vme_tsi148.c 		entry->descriptor.ddau = cpu_to_be32(address_high);
entry            1719 drivers/vme/bridges/vme_tsi148.c 		entry->descriptor.ddal = cpu_to_be32(address_low);
entry            1720 drivers/vme/bridges/vme_tsi148.c 		entry->descriptor.ddat = cpu_to_be32(TSI148_LCSR_DDAT_TYP_PCI);
entry            1727 drivers/vme/bridges/vme_tsi148.c 		entry->descriptor.ddau = cpu_to_be32(address_high);
entry            1728 drivers/vme/bridges/vme_tsi148.c 		entry->descriptor.ddal = cpu_to_be32(address_low);
entry            1729 drivers/vme/bridges/vme_tsi148.c 		entry->descriptor.ddat = cpu_to_be32(TSI148_LCSR_DDAT_TYP_VME);
entry            1732 drivers/vme/bridges/vme_tsi148.c 			tsi148_bridge->parent, &entry->descriptor.ddat,
entry            1745 drivers/vme/bridges/vme_tsi148.c 	entry->descriptor.dcnt = cpu_to_be32((u32)count);
entry            1748 drivers/vme/bridges/vme_tsi148.c 	list_add_tail(&entry->list, &list->entries);
entry            1750 drivers/vme/bridges/vme_tsi148.c 	entry->dma_handle = dma_map_single(tsi148_bridge->parent,
entry            1751 drivers/vme/bridges/vme_tsi148.c 					   &entry->descriptor,
entry            1752 drivers/vme/bridges/vme_tsi148.c 					   sizeof(entry->descriptor),
entry            1754 drivers/vme/bridges/vme_tsi148.c 	if (dma_mapping_error(tsi148_bridge->parent, entry->dma_handle)) {
entry            1761 drivers/vme/bridges/vme_tsi148.c 	if (entry->list.prev != &list->entries) {
entry            1762 drivers/vme/bridges/vme_tsi148.c 		reg_split((unsigned long long)entry->dma_handle, &address_high,
entry            1764 drivers/vme/bridges/vme_tsi148.c 		prev = list_entry(entry->list.prev, struct tsi148_dma_entry,
entry            1777 drivers/vme/bridges/vme_tsi148.c 		kfree(entry);
entry            1811 drivers/vme/bridges/vme_tsi148.c 	struct tsi148_dma_entry *entry;
entry            1841 drivers/vme/bridges/vme_tsi148.c 	entry = list_first_entry(&list->entries, struct tsi148_dma_entry,
entry            1846 drivers/vme/bridges/vme_tsi148.c 	reg_split(entry->dma_handle, &bus_addr_high, &bus_addr_low);
entry            1902 drivers/vme/bridges/vme_tsi148.c 	struct tsi148_dma_entry *entry;
entry            1909 drivers/vme/bridges/vme_tsi148.c 		entry = list_entry(pos, struct tsi148_dma_entry, list);
entry            1911 drivers/vme/bridges/vme_tsi148.c 		dma_unmap_single(tsi148_bridge->parent, entry->dma_handle,
entry            1913 drivers/vme/bridges/vme_tsi148.c 		kfree(entry);
entry              53 drivers/vme/vme.c 		return list_entry(resource->entry, struct vme_master_resource,
entry              57 drivers/vme/vme.c 		return list_entry(resource->entry, struct vme_slave_resource,
entry              61 drivers/vme/vme.c 		return list_entry(resource->entry, struct vme_dma_resource,
entry              65 drivers/vme/vme.c 		return list_entry(resource->entry, struct vme_lm_resource,
entry             338 drivers/vme/vme.c 	resource->entry = &allocated_image->list;
entry             383 drivers/vme/vme.c 	image = list_entry(resource->entry, struct vme_slave_resource, list);
entry             432 drivers/vme/vme.c 	image = list_entry(resource->entry, struct vme_slave_resource, list);
entry             459 drivers/vme/vme.c 	slave_image = list_entry(resource->entry, struct vme_slave_resource,
entry             542 drivers/vme/vme.c 	resource->entry = &allocated_image->list;
entry             587 drivers/vme/vme.c 	image = list_entry(resource->entry, struct vme_master_resource, list);
entry             637 drivers/vme/vme.c 	image = list_entry(resource->entry, struct vme_master_resource, list);
entry             681 drivers/vme/vme.c 	image = list_entry(resource->entry, struct vme_master_resource, list);
entry             730 drivers/vme/vme.c 	image = list_entry(resource->entry, struct vme_master_resource, list);
entry             781 drivers/vme/vme.c 	image = list_entry(resource->entry, struct vme_master_resource, list);
entry             809 drivers/vme/vme.c 	image = list_entry(resource->entry, struct vme_master_resource, list);
entry             839 drivers/vme/vme.c 	master_image = list_entry(resource->entry, struct vme_master_resource,
entry             917 drivers/vme/vme.c 	resource->entry = &allocated_ctrlr->list;
entry             956 drivers/vme/vme.c 	dma_list->parent = list_entry(resource->entry,
entry            1231 drivers/vme/vme.c 	ctrlr = list_entry(resource->entry, struct vme_dma_resource, list);
entry            1517 drivers/vme/vme.c 	resource->entry = &allocated_lm->list;
entry            1552 drivers/vme/vme.c 	lm = list_entry(resource->entry, struct vme_lm_resource, list);
entry            1583 drivers/vme/vme.c 	lm = list_entry(resource->entry, struct vme_lm_resource, list);
entry            1619 drivers/vme/vme.c 	lm = list_entry(resource->entry, struct vme_lm_resource, list);
entry            1656 drivers/vme/vme.c 	lm = list_entry(resource->entry, struct vme_lm_resource, list);
entry            1689 drivers/vme/vme.c 	lm = list_entry(resource->entry, struct vme_lm_resource, list);
entry            1721 drivers/vme/vme.c 	lm = list_entry(resource->entry, struct vme_lm_resource, list);
entry              95 drivers/w1/w1_int.c 	struct w1_master *dev, *entry;
entry             115 drivers/w1/w1_int.c 		list_for_each_entry(entry, &w1_masters, w1_master_entry) {
entry             116 drivers/w1/w1_int.c 			if (entry->id == id) {
entry              25 drivers/watchdog/watchdog_pretimeout.c 	struct list_head		entry;
entry              36 drivers/watchdog/watchdog_pretimeout.c 	struct list_head		entry;
entry              43 drivers/watchdog/watchdog_pretimeout.c 	list_for_each_entry(priv, &governor_list, entry)
entry              57 drivers/watchdog/watchdog_pretimeout.c 	list_for_each_entry(priv, &governor_list, entry)
entry             132 drivers/watchdog/watchdog_pretimeout.c 	list_add(&priv->entry, &governor_list);
entry             139 drivers/watchdog/watchdog_pretimeout.c 		list_for_each_entry(p, &pretimeout_list, entry)
entry             158 drivers/watchdog/watchdog_pretimeout.c 	list_for_each_entry_safe(priv, t, &governor_list, entry) {
entry             160 drivers/watchdog/watchdog_pretimeout.c 			list_del(&priv->entry);
entry             167 drivers/watchdog/watchdog_pretimeout.c 	list_for_each_entry(p, &pretimeout_list, entry)
entry             188 drivers/watchdog/watchdog_pretimeout.c 	list_add(&p->entry, &pretimeout_list);
entry             206 drivers/watchdog/watchdog_pretimeout.c 	list_for_each_entry_safe(p, t, &pretimeout_list, entry) {
entry             208 drivers/watchdog/watchdog_pretimeout.c 			list_del(&p->entry);
entry              25 drivers/watchdog/wdat_wdt.c 	struct acpi_wdat_entry entry;
entry              67 drivers/watchdog/wdat_wdt.c 	const struct acpi_generic_address *gas = &instr->entry.register_region;
entry              92 drivers/watchdog/wdat_wdt.c 	const struct acpi_generic_address *gas = &instr->entry.register_region;
entry             129 drivers/watchdog/wdat_wdt.c 		const struct acpi_wdat_entry *entry = &instr->entry;
entry             135 drivers/watchdog/wdat_wdt.c 		gas = &entry->register_region;
entry             137 drivers/watchdog/wdat_wdt.c 		preserve = entry->instruction & ACPI_WDAT_PRESERVE_REGISTER;
entry             138 drivers/watchdog/wdat_wdt.c 		flags = entry->instruction & ~ACPI_WDAT_PRESERVE_REGISTER;
entry             139 drivers/watchdog/wdat_wdt.c 		value = entry->value;
entry             140 drivers/watchdog/wdat_wdt.c 		mask = entry->mask;
entry             393 drivers/watchdog/wdat_wdt.c 		instr->entry = entries[i];
entry             163 drivers/xen/grant-table.c static inline grant_ref_t *__gnttab_entry(grant_ref_t entry)
entry             165 drivers/xen/grant-table.c 	return &gnttab_list[(entry) / RPP][(entry) % RPP];
entry             168 drivers/xen/grant-table.c #define gnttab_entry(entry) (*__gnttab_entry(entry))
entry             376 drivers/xen/grant-table.c 		struct deferred_entry *entry
entry             380 drivers/xen/grant-table.c 		if (entry == first)
entry             382 drivers/xen/grant-table.c 		list_del(&entry->list);
entry             384 drivers/xen/grant-table.c 		if (_gnttab_end_foreign_access_ref(entry->ref, entry->ro)) {
entry             385 drivers/xen/grant-table.c 			put_free_entry(entry->ref);
entry             386 drivers/xen/grant-table.c 			if (entry->page) {
entry             388 drivers/xen/grant-table.c 					 entry->ref, page_to_pfn(entry->page));
entry             389 drivers/xen/grant-table.c 				put_page(entry->page);
entry             391 drivers/xen/grant-table.c 				pr_info("freeing g.e. %#x\n", entry->ref);
entry             392 drivers/xen/grant-table.c 			kfree(entry);
entry             393 drivers/xen/grant-table.c 			entry = NULL;
entry             395 drivers/xen/grant-table.c 			if (!--entry->warn_delay)
entry             396 drivers/xen/grant-table.c 				pr_info("g.e. %#x still pending\n", entry->ref);
entry             398 drivers/xen/grant-table.c 				first = entry;
entry             401 drivers/xen/grant-table.c 		if (entry)
entry             402 drivers/xen/grant-table.c 			list_add_tail(&entry->list, &deferred_list);
entry             416 drivers/xen/grant-table.c 	struct deferred_entry *entry = kmalloc(sizeof(*entry), GFP_ATOMIC);
entry             419 drivers/xen/grant-table.c 	if (entry) {
entry             422 drivers/xen/grant-table.c 		entry->ref = ref;
entry             423 drivers/xen/grant-table.c 		entry->ro = readonly;
entry             424 drivers/xen/grant-table.c 		entry->page = page;
entry             425 drivers/xen/grant-table.c 		entry->warn_delay = 60;
entry             427 drivers/xen/grant-table.c 		list_add_tail(&entry->list, &deferred_list);
entry             124 drivers/xen/mcelog.c 		struct xen_mce *m = &xen_mcelog.entry[i];
entry             130 drivers/xen/mcelog.c 	memset(xen_mcelog.entry, 0, num * sizeof(struct xen_mce));
entry             199 drivers/xen/mcelog.c 	unsigned entry;
entry             201 drivers/xen/mcelog.c 	entry = xen_mcelog.next;
entry             208 drivers/xen/mcelog.c 	if (entry >= XEN_MCE_LOG_LEN) {
entry             214 drivers/xen/mcelog.c 	memcpy(xen_mcelog.entry + entry, mce, sizeof(struct xen_mce));
entry             276 drivers/xen/privcmd.c 			  mmapcmd.entry);
entry              41 drivers/xen/xen-pciback/conf_space.c 			   const struct config_field_entry *entry,
entry              45 drivers/xen/xen-pciback/conf_space.c 	const struct config_field *field = entry->field;
entry              53 drivers/xen/xen-pciback/conf_space.c 					      entry->data);
entry              58 drivers/xen/xen-pciback/conf_space.c 					      entry->data);
entry              62 drivers/xen/xen-pciback/conf_space.c 			ret = field->u.dw.read(dev, offset, value, entry->data);
entry              69 drivers/xen/xen-pciback/conf_space.c 			    const struct config_field_entry *entry,
entry              73 drivers/xen/xen-pciback/conf_space.c 	const struct config_field *field = entry->field;
entry              79 drivers/xen/xen-pciback/conf_space.c 					       entry->data);
entry              84 drivers/xen/xen-pciback/conf_space.c 					       entry->data);
entry              89 drivers/xen/xen-pciback/conf_space.c 						entry->data);
entry             240 drivers/xen/xen-pciback/pciback_ops.c 		entries[i].entry = op->msix_entries[i].entry;
entry             247 drivers/xen/xen-pciback/pciback_ops.c 			op->msix_entries[i].entry = entries[i].entry;
entry              35 drivers/xen/xen-pciback/vpci.c 	struct pci_dev_entry *entry;
entry              45 drivers/xen/xen-pciback/vpci.c 		list_for_each_entry(entry,
entry              48 drivers/xen/xen-pciback/vpci.c 			if (PCI_FUNC(entry->dev->devfn) == PCI_FUNC(devfn)) {
entry              49 drivers/xen/xen-pciback/vpci.c 				dev = entry->dev;
entry             235 drivers/xen/xen-pciback/vpci.c 	struct pci_dev_entry *entry;
entry             242 drivers/xen/xen-pciback/vpci.c 		list_for_each_entry(entry,
entry             245 drivers/xen/xen-pciback/vpci.c 			dev = entry->dev;
entry             312 drivers/xen/xen-scsiback.c 	struct v2p_entry *entry = container_of(kref, struct v2p_entry, kref);
entry             313 drivers/xen/xen-scsiback.c 	struct scsiback_tpg *tpg = entry->tpg;
entry             319 drivers/xen/xen-scsiback.c 	kfree(entry);
entry             630 drivers/xen/xen-scsiback.c 	struct v2p_entry *entry;
entry             635 drivers/xen/xen-scsiback.c 	list_for_each_entry(entry, head, l) {
entry             636 drivers/xen/xen-scsiback.c 		if ((entry->v.chn == v->chn) &&
entry             637 drivers/xen/xen-scsiback.c 		    (entry->v.tgt == v->tgt) &&
entry             638 drivers/xen/xen-scsiback.c 		    (entry->v.lun == v->lun)) {
entry             639 drivers/xen/xen-scsiback.c 			kref_get(&entry->kref);
entry             643 drivers/xen/xen-scsiback.c 	entry = NULL;
entry             647 drivers/xen/xen-scsiback.c 	return entry;
entry             878 drivers/xen/xen-scsiback.c 	struct v2p_entry *entry;
entry             880 drivers/xen/xen-scsiback.c 	list_for_each_entry(entry, head, l)
entry             881 drivers/xen/xen-scsiback.c 		if ((entry->v.chn == v->chn) &&
entry             882 drivers/xen/xen-scsiback.c 		    (entry->v.tgt == v->tgt) &&
entry             883 drivers/xen/xen-scsiback.c 		    (entry->v.lun == v->lun))
entry             884 drivers/xen/xen-scsiback.c 			return entry;
entry             983 drivers/xen/xen-scsiback.c static void __scsiback_del_translation_entry(struct v2p_entry *entry)
entry             985 drivers/xen/xen-scsiback.c 	list_del(&entry->l);
entry             986 drivers/xen/xen-scsiback.c 	kref_put(&entry->kref, scsiback_free_translation_entry);
entry             995 drivers/xen/xen-scsiback.c 	struct v2p_entry *entry;
entry            1001 drivers/xen/xen-scsiback.c 	entry = scsiback_chk_translation_entry(info, v);
entry            1002 drivers/xen/xen-scsiback.c 	if (entry)
entry            1003 drivers/xen/xen-scsiback.c 		__scsiback_del_translation_entry(entry);
entry            1014 drivers/xen/xen-scsiback.c 	struct v2p_entry *entry;
entry            1020 drivers/xen/xen-scsiback.c 		entry = scsiback_chk_translation_entry(info, vir);
entry            1022 drivers/xen/xen-scsiback.c 		if (entry)
entry            1210 drivers/xen/xen-scsiback.c 	struct v2p_entry *entry, *tmp;
entry            1216 drivers/xen/xen-scsiback.c 	list_for_each_entry_safe(entry, tmp, head, l)
entry            1217 drivers/xen/xen-scsiback.c 		__scsiback_del_translation_entry(entry);
entry             103 drivers/zorro/proc.c 	struct proc_dir_entry *entry;
entry             107 drivers/zorro/proc.c 	entry = proc_create_data(name, 0, proc_bus_zorro_dir,
entry             110 drivers/zorro/proc.c 	if (!entry)
entry             112 drivers/zorro/proc.c 	proc_set_size(entry, sizeof(struct zorro_dev));
entry              92 fs/afs/callback.c 	struct afs_server_entry *entry = &slist->servers[index];
entry              94 fs/afs/callback.c 	struct afs_server *server = entry->server;
entry              99 fs/afs/callback.c 	if (vcbi && likely(vcbi == entry->cb_interest))
entry             103 fs/afs/callback.c 	cbi = afs_get_cb_interest(entry->cb_interest);
entry             128 fs/afs/callback.c 			if (entry->cb_interest) {
entry             134 fs/afs/callback.c 			entry->cb_interest = cbi;
entry             146 fs/afs/callback.c 		if (!entry->cb_interest) {
entry             147 fs/afs/callback.c 			entry->cb_interest = afs_get_cb_interest(new);
entry             151 fs/afs/callback.c 			cbi = afs_get_cb_interest(entry->cb_interest);
entry             288 fs/afs/proc.c  	const struct afs_vlserver_entry *entry;
entry             300 fs/afs/proc.c  	entry = v;
entry             301 fs/afs/proc.c  	vlserver = entry->server;
entry             305 fs/afs/proc.c  		   vlserver->name, entry->priority, entry->weight,
entry             306 fs/afs/proc.c  		   dns_record_sources[alist ? alist->source : entry->source],
entry             307 fs/afs/proc.c  		   dns_lookup_statuses[alist ? alist->status : entry->status]);
entry              20 fs/afs/vlclient.c 	struct afs_vldb_entry *entry;
entry              33 fs/afs/vlclient.c 	entry = call->ret_vldb;
entry              40 fs/afs/vlclient.c 		entry->name[i] = (u8)ntohl(uvldb->name[i]);
entry              41 fs/afs/vlclient.c 	entry->name[i] = 0;
entry              42 fs/afs/vlclient.c 	entry->name_len = strlen(entry->name);
entry              59 fs/afs/vlclient.c 		int n = entry->nr_servers;
entry              66 fs/afs/vlclient.c 			entry->fs_mask[n] |= AFS_VOL_VTM_RW;
entry              68 fs/afs/vlclient.c 				entry->fs_mask[n] |= AFS_VOL_VTM_BAK;
entry              71 fs/afs/vlclient.c 			entry->fs_mask[n] |= AFS_VOL_VTM_RO;
entry              72 fs/afs/vlclient.c 		if (!entry->fs_mask[n])
entry              76 fs/afs/vlclient.c 		uuid = (struct afs_uuid *)&entry->fs_server[n];
entry              85 fs/afs/vlclient.c 		entry->nr_servers++;
entry              89 fs/afs/vlclient.c 		entry->vid[i] = ntohl(uvldb->volumeId[i]);
entry              92 fs/afs/vlclient.c 		__set_bit(AFS_VLDB_HAS_RW, &entry->flags);
entry              94 fs/afs/vlclient.c 		__set_bit(AFS_VLDB_HAS_RO, &entry->flags);
entry              96 fs/afs/vlclient.c 		__set_bit(AFS_VLDB_HAS_BAK, &entry->flags);
entry              99 fs/afs/vlclient.c 		entry->error = -ENOMEDIUM;
entry             100 fs/afs/vlclient.c 		__set_bit(AFS_VLDB_QUERY_ERROR, &entry->flags);
entry             103 fs/afs/vlclient.c 	__set_bit(AFS_VLDB_QUERY_VALID, &entry->flags);
entry             132 fs/afs/vlclient.c 	struct afs_vldb_entry *entry;
entry             143 fs/afs/vlclient.c 	entry = kzalloc(sizeof(struct afs_vldb_entry), GFP_KERNEL);
entry             144 fs/afs/vlclient.c 	if (!entry)
entry             150 fs/afs/vlclient.c 		kfree(entry);
entry             155 fs/afs/vlclient.c 	call->ret_vldb = entry;
entry            1669 fs/aio.c       	if (!list_empty(&req->wait.entry)) {
entry            1670 fs/aio.c       		list_del_init(&req->wait.entry);
entry            1690 fs/aio.c       	list_del_init(&req->wait.entry);
entry            1769 fs/aio.c       	INIT_LIST_HEAD(&req->wait.entry);
entry            1776 fs/aio.c       		if (unlikely(list_empty(&req->wait.entry))) {
entry            1783 fs/aio.c       			list_del_init(&req->wait.entry);
entry              83 fs/binfmt_flat.c 		unsigned long entry;			/* Start address for this module */
entry             739 fs/binfmt_flat.c 			textpos, 0x00ffffff&ntohl(hdr->entry), ntohl(hdr->data_start));
entry             752 fs/binfmt_flat.c 	libinfo->lib_list[id].entry = (0x00ffffff & ntohl(hdr->entry)) + textpos;
entry             992 fs/binfmt_flat.c 	start_addr = libinfo.lib_list[0].entry;
entry            1002 fs/binfmt_flat.c 			start_addr = libinfo.lib_list[i].entry;
entry             283 fs/btrfs/delayed-ref.c 	struct btrfs_delayed_ref_head *entry;
entry             292 fs/btrfs/delayed-ref.c 		entry = rb_entry(parent_node, struct btrfs_delayed_ref_head,
entry             295 fs/btrfs/delayed-ref.c 		if (bytenr < entry->bytenr) {
entry             297 fs/btrfs/delayed-ref.c 		} else if (bytenr > entry->bytenr) {
entry             301 fs/btrfs/delayed-ref.c 			return entry;
entry             316 fs/btrfs/delayed-ref.c 	struct btrfs_delayed_ref_node *entry;
entry             323 fs/btrfs/delayed-ref.c 		entry = rb_entry(parent_node, struct btrfs_delayed_ref_node,
entry             325 fs/btrfs/delayed-ref.c 		comp = comp_refs(ins, entry, true);
entry             332 fs/btrfs/delayed-ref.c 			return entry;
entry             345 fs/btrfs/delayed-ref.c 	struct btrfs_delayed_ref_head *entry;
entry             351 fs/btrfs/delayed-ref.c 	entry = rb_entry(n, struct btrfs_delayed_ref_head, href_node);
entry             353 fs/btrfs/delayed-ref.c 	return entry;
entry             367 fs/btrfs/delayed-ref.c 	struct btrfs_delayed_ref_head *entry;
entry             370 fs/btrfs/delayed-ref.c 	entry = NULL;
entry             372 fs/btrfs/delayed-ref.c 		entry = rb_entry(n, struct btrfs_delayed_ref_head, href_node);
entry             374 fs/btrfs/delayed-ref.c 		if (bytenr < entry->bytenr)
entry             376 fs/btrfs/delayed-ref.c 		else if (bytenr > entry->bytenr)
entry             379 fs/btrfs/delayed-ref.c 			return entry;
entry             381 fs/btrfs/delayed-ref.c 	if (entry && return_bigger) {
entry             382 fs/btrfs/delayed-ref.c 		if (bytenr > entry->bytenr) {
entry             383 fs/btrfs/delayed-ref.c 			n = rb_next(&entry->href_node);
entry             386 fs/btrfs/delayed-ref.c 			entry = rb_entry(n, struct btrfs_delayed_ref_head,
entry             389 fs/btrfs/delayed-ref.c 		return entry;
entry            2082 fs/btrfs/extent-tree.c 	struct btrfs_delayed_ref_node *entry;
entry            2089 fs/btrfs/extent-tree.c 		entry = rb_entry(n, struct btrfs_delayed_ref_node, rb_node);
entry            2090 fs/btrfs/extent-tree.c 		first = entry->bytenr;
entry            2094 fs/btrfs/extent-tree.c 		entry = rb_entry(n, struct btrfs_delayed_ref_node, rb_node);
entry            2095 fs/btrfs/extent-tree.c 		last = entry->bytenr;
entry            2100 fs/btrfs/extent-tree.c 		entry = rb_entry(n, struct btrfs_delayed_ref_node, rb_node);
entry            2101 fs/btrfs/extent-tree.c 		WARN_ON(!entry->in_tree);
entry            2103 fs/btrfs/extent-tree.c 		middle = entry->bytenr;
entry              53 fs/btrfs/extent_io.c void btrfs_leak_debug_del(struct list_head *entry)
entry              58 fs/btrfs/extent_io.c 	list_del(entry);
entry             107 fs/btrfs/extent_io.c #define btrfs_leak_debug_del(entry)	do {} while (0)
entry             335 fs/btrfs/extent_io.c 	struct tree_entry *entry;
entry             346 fs/btrfs/extent_io.c 		entry = rb_entry(parent, struct tree_entry, rb_node);
entry             348 fs/btrfs/extent_io.c 		if (offset < entry->start)
entry             350 fs/btrfs/extent_io.c 		else if (offset > entry->end)
entry             390 fs/btrfs/extent_io.c 	struct tree_entry *entry;
entry             395 fs/btrfs/extent_io.c 		entry = rb_entry(prev, struct tree_entry, rb_node);
entry             396 fs/btrfs/extent_io.c 		prev_entry = entry;
entry             398 fs/btrfs/extent_io.c 		if (offset < entry->start)
entry             400 fs/btrfs/extent_io.c 		else if (offset > entry->end)
entry              98 fs/btrfs/extent_map.c 	struct extent_map *entry = NULL;
entry             105 fs/btrfs/extent_map.c 		entry = rb_entry(parent, struct extent_map, rb_node);
entry             107 fs/btrfs/extent_map.c 		if (em->start < entry->start) {
entry             109 fs/btrfs/extent_map.c 		} else if (em->start >= extent_map_end(entry)) {
entry             118 fs/btrfs/extent_map.c 	while (parent && em->start >= extent_map_end(entry)) {
entry             120 fs/btrfs/extent_map.c 		entry = rb_entry(parent, struct extent_map, rb_node);
entry             123 fs/btrfs/extent_map.c 		if (end > entry->start && em->start < extent_map_end(entry))
entry             127 fs/btrfs/extent_map.c 	entry = rb_entry(parent, struct extent_map, rb_node);
entry             128 fs/btrfs/extent_map.c 	while (parent && em->start < entry->start) {
entry             130 fs/btrfs/extent_map.c 		entry = rb_entry(parent, struct extent_map, rb_node);
entry             133 fs/btrfs/extent_map.c 		if (end > entry->start && em->start < extent_map_end(entry))
entry             152 fs/btrfs/extent_map.c 	struct extent_map *entry;
entry             156 fs/btrfs/extent_map.c 		entry = rb_entry(n, struct extent_map, rb_node);
entry             158 fs/btrfs/extent_map.c 		prev_entry = entry;
entry             160 fs/btrfs/extent_map.c 		if (offset < entry->start)
entry             162 fs/btrfs/extent_map.c 		else if (offset >= extent_map_end(entry))
entry              85 fs/btrfs/file.c 	struct inode_defrag *entry;
entry              93 fs/btrfs/file.c 		entry = rb_entry(parent, struct inode_defrag, rb_node);
entry              95 fs/btrfs/file.c 		ret = __compare_inode_defrag(defrag, entry);
entry             105 fs/btrfs/file.c 			if (defrag->transid < entry->transid)
entry             106 fs/btrfs/file.c 				entry->transid = defrag->transid;
entry             107 fs/btrfs/file.c 			if (defrag->last_offset > entry->last_offset)
entry             108 fs/btrfs/file.c 				entry->last_offset = defrag->last_offset;
entry             213 fs/btrfs/file.c 	struct inode_defrag *entry = NULL;
entry             226 fs/btrfs/file.c 		entry = rb_entry(parent, struct inode_defrag, rb_node);
entry             228 fs/btrfs/file.c 		ret = __compare_inode_defrag(&tmp, entry);
entry             237 fs/btrfs/file.c 	if (parent && __compare_inode_defrag(&tmp, entry) > 0) {
entry             240 fs/btrfs/file.c 			entry = rb_entry(parent, struct inode_defrag, rb_node);
entry             242 fs/btrfs/file.c 			entry = NULL;
entry             245 fs/btrfs/file.c 	if (entry)
entry             248 fs/btrfs/file.c 	return entry;
entry             519 fs/btrfs/free-space-cache.c 	struct btrfs_free_space_entry *entry;
entry             524 fs/btrfs/free-space-cache.c 	entry = io_ctl->cur;
entry             525 fs/btrfs/free-space-cache.c 	entry->offset = cpu_to_le64(offset);
entry             526 fs/btrfs/free-space-cache.c 	entry->bytes = cpu_to_le64(bytes);
entry             527 fs/btrfs/free-space-cache.c 	entry->type = (bitmap) ? BTRFS_FREE_SPACE_BITMAP :
entry             587 fs/btrfs/free-space-cache.c 			    struct btrfs_free_space *entry, u8 *type)
entry             599 fs/btrfs/free-space-cache.c 	entry->offset = le64_to_cpu(e->offset);
entry             600 fs/btrfs/free-space-cache.c 	entry->bytes = le64_to_cpu(e->bytes);
entry             614 fs/btrfs/free-space-cache.c 			      struct btrfs_free_space *entry)
entry             622 fs/btrfs/free-space-cache.c 	copy_page(entry->bitmap, io_ctl->cur);
entry            1103 fs/btrfs/free-space-cache.c 	struct btrfs_free_space *entry, *next;
entry            1107 fs/btrfs/free-space-cache.c 	list_for_each_entry_safe(entry, next, bitmap_list, list) {
entry            1108 fs/btrfs/free-space-cache.c 		ret = io_ctl_add_bitmap(io_ctl, entry->bitmap);
entry            1111 fs/btrfs/free-space-cache.c 		list_del_init(&entry->list);
entry            1132 fs/btrfs/free-space-cache.c 	struct btrfs_free_space *entry, *next;
entry            1134 fs/btrfs/free-space-cache.c 	list_for_each_entry_safe(entry, next, bitmap_list, list)
entry            1135 fs/btrfs/free-space-cache.c 		list_del_init(&entry->list);
entry            1512 fs/btrfs/free-space-cache.c 	struct btrfs_free_space *entry, *prev = NULL;
entry            1517 fs/btrfs/free-space-cache.c 			entry = NULL;
entry            1521 fs/btrfs/free-space-cache.c 		entry = rb_entry(n, struct btrfs_free_space, offset_index);
entry            1522 fs/btrfs/free-space-cache.c 		prev = entry;
entry            1524 fs/btrfs/free-space-cache.c 		if (offset < entry->offset)
entry            1526 fs/btrfs/free-space-cache.c 		else if (offset > entry->offset)
entry            1533 fs/btrfs/free-space-cache.c 		if (!entry)
entry            1535 fs/btrfs/free-space-cache.c 		if (entry->bitmap)
entry            1536 fs/btrfs/free-space-cache.c 			return entry;
entry            1545 fs/btrfs/free-space-cache.c 		entry = rb_entry(n, struct btrfs_free_space, offset_index);
entry            1546 fs/btrfs/free-space-cache.c 		if (entry->offset != offset)
entry            1549 fs/btrfs/free-space-cache.c 		WARN_ON(!entry->bitmap);
entry            1550 fs/btrfs/free-space-cache.c 		return entry;
entry            1551 fs/btrfs/free-space-cache.c 	} else if (entry) {
entry            1552 fs/btrfs/free-space-cache.c 		if (entry->bitmap) {
entry            1557 fs/btrfs/free-space-cache.c 			n = rb_prev(&entry->offset_index);
entry            1563 fs/btrfs/free-space-cache.c 					entry = prev;
entry            1566 fs/btrfs/free-space-cache.c 		return entry;
entry            1573 fs/btrfs/free-space-cache.c 	entry = prev;
entry            1574 fs/btrfs/free-space-cache.c 	if (entry->offset > offset) {
entry            1575 fs/btrfs/free-space-cache.c 		n = rb_prev(&entry->offset_index);
entry            1577 fs/btrfs/free-space-cache.c 			entry = rb_entry(n, struct btrfs_free_space,
entry            1579 fs/btrfs/free-space-cache.c 			ASSERT(entry->offset <= offset);
entry            1582 fs/btrfs/free-space-cache.c 				return entry;
entry            1588 fs/btrfs/free-space-cache.c 	if (entry->bitmap) {
entry            1589 fs/btrfs/free-space-cache.c 		n = rb_prev(&entry->offset_index);
entry            1597 fs/btrfs/free-space-cache.c 		if (entry->offset + BITS_PER_BITMAP * ctl->unit > offset)
entry            1598 fs/btrfs/free-space-cache.c 			return entry;
entry            1599 fs/btrfs/free-space-cache.c 	} else if (entry->offset + entry->bytes > offset)
entry            1600 fs/btrfs/free-space-cache.c 		return entry;
entry            1606 fs/btrfs/free-space-cache.c 		if (entry->bitmap) {
entry            1607 fs/btrfs/free-space-cache.c 			if (entry->offset + BITS_PER_BITMAP *
entry            1611 fs/btrfs/free-space-cache.c 			if (entry->offset + entry->bytes > offset)
entry            1615 fs/btrfs/free-space-cache.c 		n = rb_next(&entry->offset_index);
entry            1618 fs/btrfs/free-space-cache.c 		entry = rb_entry(n, struct btrfs_free_space, offset_index);
entry            1620 fs/btrfs/free-space-cache.c 	return entry;
entry            1799 fs/btrfs/free-space-cache.c static inline u64 get_max_extent_size(struct btrfs_free_space *entry)
entry            1801 fs/btrfs/free-space-cache.c 	if (entry->bitmap)
entry            1802 fs/btrfs/free-space-cache.c 		return entry->max_extent_size;
entry            1803 fs/btrfs/free-space-cache.c 	return entry->bytes;
entry            1811 fs/btrfs/free-space-cache.c 	struct btrfs_free_space *entry;
entry            1820 fs/btrfs/free-space-cache.c 	entry = tree_search_offset(ctl, offset_to_bitmap(ctl, *offset), 0, 1);
entry            1821 fs/btrfs/free-space-cache.c 	if (!entry)
entry            1824 fs/btrfs/free-space-cache.c 	for (node = &entry->offset_index; node; node = rb_next(node)) {
entry            1825 fs/btrfs/free-space-cache.c 		entry = rb_entry(node, struct btrfs_free_space, offset_index);
entry            1826 fs/btrfs/free-space-cache.c 		if (entry->bytes < *bytes) {
entry            1827 fs/btrfs/free-space-cache.c 			*max_extent_size = max(get_max_extent_size(entry),
entry            1836 fs/btrfs/free-space-cache.c 			tmp = entry->offset - ctl->start + align - 1;
entry            1839 fs/btrfs/free-space-cache.c 			align_off = tmp - entry->offset;
entry            1842 fs/btrfs/free-space-cache.c 			tmp = entry->offset;
entry            1845 fs/btrfs/free-space-cache.c 		if (entry->bytes < *bytes + align_off) {
entry            1846 fs/btrfs/free-space-cache.c 			*max_extent_size = max(get_max_extent_size(entry),
entry            1851 fs/btrfs/free-space-cache.c 		if (entry->bitmap) {
entry            1854 fs/btrfs/free-space-cache.c 			ret = search_bitmap(ctl, entry, &tmp, &size, true);
entry            1858 fs/btrfs/free-space-cache.c 				return entry;
entry            1861 fs/btrfs/free-space-cache.c 					max(get_max_extent_size(entry),
entry            1868 fs/btrfs/free-space-cache.c 		*bytes = entry->bytes - align_off;
entry            1869 fs/btrfs/free-space-cache.c 		return entry;
entry            2074 fs/btrfs/free-space-cache.c 		struct btrfs_free_space *entry;
entry            2086 fs/btrfs/free-space-cache.c 		entry = rb_entry(node, struct btrfs_free_space, offset_index);
entry            2087 fs/btrfs/free-space-cache.c 		if (!entry->bitmap) {
entry            2092 fs/btrfs/free-space-cache.c 		if (entry->offset == offset_to_bitmap(ctl, offset)) {
entry            2093 fs/btrfs/free-space-cache.c 			bytes_added = add_bytes_to_bitmap(ctl, entry,
entry            2545 fs/btrfs/free-space-cache.c 	struct btrfs_free_space *entry;
entry            2560 fs/btrfs/free-space-cache.c 		entry = rb_entry(node, struct btrfs_free_space, offset_index);
entry            2561 fs/btrfs/free-space-cache.c 		node = rb_next(&entry->offset_index);
entry            2562 fs/btrfs/free-space-cache.c 		rb_erase(&entry->offset_index, &cluster->root);
entry            2563 fs/btrfs/free-space-cache.c 		RB_CLEAR_NODE(&entry->offset_index);
entry            2565 fs/btrfs/free-space-cache.c 		bitmap = (entry->bitmap != NULL);
entry            2567 fs/btrfs/free-space-cache.c 			try_merge_free_space(ctl, entry, false);
entry            2568 fs/btrfs/free-space-cache.c 			steal_from_bitmap(ctl, entry, false);
entry            2571 fs/btrfs/free-space-cache.c 				   entry->offset, &entry->offset_index, bitmap);
entry            2634 fs/btrfs/free-space-cache.c 	struct btrfs_free_space *entry = NULL;
entry            2641 fs/btrfs/free-space-cache.c 	entry = find_free_space(ctl, &offset, &bytes_search,
entry            2643 fs/btrfs/free-space-cache.c 	if (!entry)
entry            2647 fs/btrfs/free-space-cache.c 	if (entry->bitmap) {
entry            2648 fs/btrfs/free-space-cache.c 		bitmap_clear_bits(ctl, entry, offset, bytes);
entry            2649 fs/btrfs/free-space-cache.c 		if (!entry->bytes)
entry            2650 fs/btrfs/free-space-cache.c 			free_bitmap(ctl, entry);
entry            2652 fs/btrfs/free-space-cache.c 		unlink_free_space(ctl, entry);
entry            2653 fs/btrfs/free-space-cache.c 		align_gap_len = offset - entry->offset;
entry            2654 fs/btrfs/free-space-cache.c 		align_gap = entry->offset;
entry            2656 fs/btrfs/free-space-cache.c 		entry->offset = offset + bytes;
entry            2657 fs/btrfs/free-space-cache.c 		WARN_ON(entry->bytes < bytes + align_gap_len);
entry            2659 fs/btrfs/free-space-cache.c 		entry->bytes -= bytes + align_gap_len;
entry            2660 fs/btrfs/free-space-cache.c 		if (!entry->bytes)
entry            2661 fs/btrfs/free-space-cache.c 			kmem_cache_free(btrfs_free_space_cachep, entry);
entry            2663 fs/btrfs/free-space-cache.c 			link_free_space(ctl, entry);
entry            2719 fs/btrfs/free-space-cache.c 				   struct btrfs_free_space *entry,
entry            2732 fs/btrfs/free-space-cache.c 	err = search_bitmap(ctl, entry, &search_start, &search_bytes, true);
entry            2734 fs/btrfs/free-space-cache.c 		*max_extent_size = max(get_max_extent_size(entry),
entry            2740 fs/btrfs/free-space-cache.c 	__bitmap_clear_bits(ctl, entry, ret, bytes);
entry            2755 fs/btrfs/free-space-cache.c 	struct btrfs_free_space *entry = NULL;
entry            2770 fs/btrfs/free-space-cache.c 	entry = rb_entry(node, struct btrfs_free_space, offset_index);
entry            2772 fs/btrfs/free-space-cache.c 		if (entry->bytes < bytes)
entry            2773 fs/btrfs/free-space-cache.c 			*max_extent_size = max(get_max_extent_size(entry),
entry            2776 fs/btrfs/free-space-cache.c 		if (entry->bytes < bytes ||
entry            2777 fs/btrfs/free-space-cache.c 		    (!entry->bitmap && entry->offset < min_start)) {
entry            2778 fs/btrfs/free-space-cache.c 			node = rb_next(&entry->offset_index);
entry            2781 fs/btrfs/free-space-cache.c 			entry = rb_entry(node, struct btrfs_free_space,
entry            2786 fs/btrfs/free-space-cache.c 		if (entry->bitmap) {
entry            2788 fs/btrfs/free-space-cache.c 						      cluster, entry, bytes,
entry            2792 fs/btrfs/free-space-cache.c 				node = rb_next(&entry->offset_index);
entry            2795 fs/btrfs/free-space-cache.c 				entry = rb_entry(node, struct btrfs_free_space,
entry            2801 fs/btrfs/free-space-cache.c 			ret = entry->offset;
entry            2803 fs/btrfs/free-space-cache.c 			entry->offset += bytes;
entry            2804 fs/btrfs/free-space-cache.c 			entry->bytes -= bytes;
entry            2807 fs/btrfs/free-space-cache.c 		if (entry->bytes == 0)
entry            2808 fs/btrfs/free-space-cache.c 			rb_erase(&entry->offset_index, &cluster->root);
entry            2820 fs/btrfs/free-space-cache.c 	if (entry->bytes == 0) {
entry            2822 fs/btrfs/free-space-cache.c 		if (entry->bitmap) {
entry            2824 fs/btrfs/free-space-cache.c 					entry->bitmap);
entry            2828 fs/btrfs/free-space-cache.c 		kmem_cache_free(btrfs_free_space_cachep, entry);
entry            2837 fs/btrfs/free-space-cache.c 				struct btrfs_free_space *entry,
entry            2853 fs/btrfs/free-space-cache.c 	i = offset_to_bit(entry->offset, ctl->unit,
entry            2854 fs/btrfs/free-space-cache.c 			  max_t(u64, offset, entry->offset));
entry            2862 fs/btrfs/free-space-cache.c 	if (entry->max_extent_size &&
entry            2863 fs/btrfs/free-space-cache.c 	    entry->max_extent_size < cont1_bytes)
entry            2867 fs/btrfs/free-space-cache.c 	for_each_set_bit_from(i, entry->bitmap, BITS_PER_BITMAP) {
entry            2868 fs/btrfs/free-space-cache.c 		next_zero = find_next_zero_bit(entry->bitmap,
entry            2882 fs/btrfs/free-space-cache.c 		entry->max_extent_size = (u64)max_bits * ctl->unit;
entry            2901 fs/btrfs/free-space-cache.c 	cluster->window_start = start * ctl->unit + entry->offset;
entry            2902 fs/btrfs/free-space-cache.c 	rb_erase(&entry->offset_index, &ctl->free_space_offset);
entry            2903 fs/btrfs/free-space-cache.c 	ret = tree_insert_offset(&cluster->root, entry->offset,
entry            2904 fs/btrfs/free-space-cache.c 				 &entry->offset_index, 1);
entry            2925 fs/btrfs/free-space-cache.c 	struct btrfs_free_space *entry = NULL;
entry            2932 fs/btrfs/free-space-cache.c 	entry = tree_search_offset(ctl, offset, 0, 1);
entry            2933 fs/btrfs/free-space-cache.c 	if (!entry)
entry            2940 fs/btrfs/free-space-cache.c 	while (entry->bitmap || entry->bytes < min_bytes) {
entry            2941 fs/btrfs/free-space-cache.c 		if (entry->bitmap && list_empty(&entry->list))
entry            2942 fs/btrfs/free-space-cache.c 			list_add_tail(&entry->list, bitmaps);
entry            2943 fs/btrfs/free-space-cache.c 		node = rb_next(&entry->offset_index);
entry            2946 fs/btrfs/free-space-cache.c 		entry = rb_entry(node, struct btrfs_free_space, offset_index);
entry            2949 fs/btrfs/free-space-cache.c 	window_free = entry->bytes;
entry            2950 fs/btrfs/free-space-cache.c 	max_extent = entry->bytes;
entry            2951 fs/btrfs/free-space-cache.c 	first = entry;
entry            2952 fs/btrfs/free-space-cache.c 	last = entry;
entry            2954 fs/btrfs/free-space-cache.c 	for (node = rb_next(&entry->offset_index); node;
entry            2955 fs/btrfs/free-space-cache.c 	     node = rb_next(&entry->offset_index)) {
entry            2956 fs/btrfs/free-space-cache.c 		entry = rb_entry(node, struct btrfs_free_space, offset_index);
entry            2958 fs/btrfs/free-space-cache.c 		if (entry->bitmap) {
entry            2959 fs/btrfs/free-space-cache.c 			if (list_empty(&entry->list))
entry            2960 fs/btrfs/free-space-cache.c 				list_add_tail(&entry->list, bitmaps);
entry            2964 fs/btrfs/free-space-cache.c 		if (entry->bytes < min_bytes)
entry            2967 fs/btrfs/free-space-cache.c 		last = entry;
entry            2968 fs/btrfs/free-space-cache.c 		window_free += entry->bytes;
entry            2969 fs/btrfs/free-space-cache.c 		if (entry->bytes > max_extent)
entry            2970 fs/btrfs/free-space-cache.c 			max_extent = entry->bytes;
entry            2987 fs/btrfs/free-space-cache.c 		entry = rb_entry(node, struct btrfs_free_space, offset_index);
entry            2988 fs/btrfs/free-space-cache.c 		node = rb_next(&entry->offset_index);
entry            2989 fs/btrfs/free-space-cache.c 		if (entry->bitmap || entry->bytes < min_bytes)
entry            2992 fs/btrfs/free-space-cache.c 		rb_erase(&entry->offset_index, &ctl->free_space_offset);
entry            2993 fs/btrfs/free-space-cache.c 		ret = tree_insert_offset(&cluster->root, entry->offset,
entry            2994 fs/btrfs/free-space-cache.c 					 &entry->offset_index, 0);
entry            2995 fs/btrfs/free-space-cache.c 		total_size += entry->bytes;
entry            2997 fs/btrfs/free-space-cache.c 	} while (node && entry != last);
entry            3015 fs/btrfs/free-space-cache.c 	struct btrfs_free_space *entry = NULL;
entry            3027 fs/btrfs/free-space-cache.c 		entry = list_first_entry(bitmaps, struct btrfs_free_space, list);
entry            3029 fs/btrfs/free-space-cache.c 	if (!entry || entry->offset != bitmap_offset) {
entry            3030 fs/btrfs/free-space-cache.c 		entry = tree_search_offset(ctl, bitmap_offset, 1, 0);
entry            3031 fs/btrfs/free-space-cache.c 		if (entry && list_empty(&entry->list))
entry            3032 fs/btrfs/free-space-cache.c 			list_add(&entry->list, bitmaps);
entry            3035 fs/btrfs/free-space-cache.c 	list_for_each_entry(entry, bitmaps, list) {
entry            3036 fs/btrfs/free-space-cache.c 		if (entry->bytes < bytes)
entry            3038 fs/btrfs/free-space-cache.c 		ret = btrfs_bitmap_cluster(block_group, entry, cluster, offset,
entry            3065 fs/btrfs/free-space-cache.c 	struct btrfs_free_space *entry, *tmp;
entry            3118 fs/btrfs/free-space-cache.c 	list_for_each_entry_safe(entry, tmp, &bitmaps, list)
entry            3119 fs/btrfs/free-space-cache.c 		list_del_init(&entry->list);
entry            3199 fs/btrfs/free-space-cache.c 	struct btrfs_free_space *entry;
entry            3218 fs/btrfs/free-space-cache.c 		entry = tree_search_offset(ctl, start, 0, 1);
entry            3219 fs/btrfs/free-space-cache.c 		if (!entry) {
entry            3226 fs/btrfs/free-space-cache.c 		while (entry->bitmap) {
entry            3227 fs/btrfs/free-space-cache.c 			node = rb_next(&entry->offset_index);
entry            3233 fs/btrfs/free-space-cache.c 			entry = rb_entry(node, struct btrfs_free_space,
entry            3237 fs/btrfs/free-space-cache.c 		if (entry->offset >= end) {
entry            3243 fs/btrfs/free-space-cache.c 		extent_start = entry->offset;
entry            3244 fs/btrfs/free-space-cache.c 		extent_bytes = entry->bytes;
entry            3253 fs/btrfs/free-space-cache.c 		unlink_free_space(ctl, entry);
entry            3254 fs/btrfs/free-space-cache.c 		kmem_cache_free(btrfs_free_space_cachep, entry);
entry            3284 fs/btrfs/free-space-cache.c 	struct btrfs_free_space *entry;
entry            3303 fs/btrfs/free-space-cache.c 		entry = tree_search_offset(ctl, offset, 1, 0);
entry            3304 fs/btrfs/free-space-cache.c 		if (!entry) {
entry            3312 fs/btrfs/free-space-cache.c 		ret2 = search_bitmap(ctl, entry, &start, &bytes, false);
entry            3327 fs/btrfs/free-space-cache.c 		bitmap_clear_bits(ctl, entry, start, bytes);
entry            3328 fs/btrfs/free-space-cache.c 		if (entry->bytes == 0)
entry            3329 fs/btrfs/free-space-cache.c 			free_bitmap(ctl, entry);
entry            3436 fs/btrfs/free-space-cache.c 	struct btrfs_free_space *entry = NULL;
entry            3444 fs/btrfs/free-space-cache.c 	entry = rb_entry(rb_first(&ctl->free_space_offset),
entry            3447 fs/btrfs/free-space-cache.c 	if (!entry->bitmap) {
entry            3448 fs/btrfs/free-space-cache.c 		ino = entry->offset;
entry            3450 fs/btrfs/free-space-cache.c 		unlink_free_space(ctl, entry);
entry            3451 fs/btrfs/free-space-cache.c 		entry->offset++;
entry            3452 fs/btrfs/free-space-cache.c 		entry->bytes--;
entry            3453 fs/btrfs/free-space-cache.c 		if (!entry->bytes)
entry            3454 fs/btrfs/free-space-cache.c 			kmem_cache_free(btrfs_free_space_cachep, entry);
entry            3456 fs/btrfs/free-space-cache.c 			link_free_space(ctl, entry);
entry            3462 fs/btrfs/free-space-cache.c 		ret = search_bitmap(ctl, entry, &offset, &count, true);
entry            3467 fs/btrfs/free-space-cache.c 		bitmap_clear_bits(ctl, entry, offset, 1);
entry            3468 fs/btrfs/free-space-cache.c 		if (entry->bytes == 0)
entry            3469 fs/btrfs/free-space-cache.c 			free_bitmap(ctl, entry);
entry            2518 fs/btrfs/inode.c 	struct sa_defrag_extent_backref *entry;
entry            2523 fs/btrfs/inode.c 		entry = rb_entry(parent, struct sa_defrag_extent_backref, node);
entry            2525 fs/btrfs/inode.c 		ret = backref_comp(backref, entry);
entry            4447 fs/btrfs/inode.c 	struct btrfs_inode *entry;
entry            4460 fs/btrfs/inode.c 		entry = rb_entry(node, struct btrfs_inode, rb_node);
entry            4462 fs/btrfs/inode.c 		if (objectid < btrfs_ino(entry))
entry            4464 fs/btrfs/inode.c 		else if (objectid > btrfs_ino(entry))
entry            4471 fs/btrfs/inode.c 			entry = rb_entry(prev, struct btrfs_inode, rb_node);
entry            4472 fs/btrfs/inode.c 			if (objectid <= btrfs_ino(entry)) {
entry            4480 fs/btrfs/inode.c 		entry = rb_entry(node, struct btrfs_inode, rb_node);
entry            4481 fs/btrfs/inode.c 		objectid = btrfs_ino(entry) + 1;
entry            4482 fs/btrfs/inode.c 		inode = igrab(&entry->vfs_inode);
entry            5770 fs/btrfs/inode.c 	struct btrfs_inode *entry;
entry            5783 fs/btrfs/inode.c 		entry = rb_entry(parent, struct btrfs_inode, rb_node);
entry            5785 fs/btrfs/inode.c 		if (ino < btrfs_ino(entry))
entry            5787 fs/btrfs/inode.c 		else if (ino > btrfs_ino(entry))
entry            5790 fs/btrfs/inode.c 			WARN_ON(!(entry->vfs_inode.i_state &
entry            6074 fs/btrfs/inode.c 		struct dir_entry *entry = addr;
entry            6075 fs/btrfs/inode.c 		char *name = (char *)(entry + 1);
entry            6077 fs/btrfs/inode.c 		ctx->pos = get_unaligned(&entry->offset);
entry            6078 fs/btrfs/inode.c 		if (!dir_emit(ctx, name, get_unaligned(&entry->name_len),
entry            6079 fs/btrfs/inode.c 					 get_unaligned(&entry->ino),
entry            6080 fs/btrfs/inode.c 					 get_unaligned(&entry->type)))
entry            6083 fs/btrfs/inode.c 			get_unaligned(&entry->name_len);
entry            6135 fs/btrfs/inode.c 		struct dir_entry *entry;
entry            6172 fs/btrfs/inode.c 		entry = addr;
entry            6173 fs/btrfs/inode.c 		put_unaligned(name_len, &entry->name_len);
entry            6174 fs/btrfs/inode.c 		name_ptr = (char *)(entry + 1);
entry            6178 fs/btrfs/inode.c 				&entry->type);
entry            6180 fs/btrfs/inode.c 		put_unaligned(location.objectid, &entry->ino);
entry            6181 fs/btrfs/inode.c 		put_unaligned(found_key.offset, &entry->offset);
entry            10738 fs/btrfs/inode.c 	struct btrfs_swapfile_pin *sp, *entry;
entry            10753 fs/btrfs/inode.c 		entry = rb_entry(parent, struct btrfs_swapfile_pin, node);
entry            10754 fs/btrfs/inode.c 		if (sp->ptr < entry->ptr ||
entry            10755 fs/btrfs/inode.c 		    (sp->ptr == entry->ptr && sp->inode < entry->inode)) {
entry            10757 fs/btrfs/inode.c 		} else if (sp->ptr > entry->ptr ||
entry            10758 fs/btrfs/inode.c 			   (sp->ptr == entry->ptr && sp->inode > entry->inode)) {
entry              21 fs/btrfs/ordered-data.c static u64 entry_end(struct btrfs_ordered_extent *entry)
entry              23 fs/btrfs/ordered-data.c 	if (entry->file_offset + entry->len < entry->file_offset)
entry              25 fs/btrfs/ordered-data.c 	return entry->file_offset + entry->len;
entry              36 fs/btrfs/ordered-data.c 	struct btrfs_ordered_extent *entry;
entry              40 fs/btrfs/ordered-data.c 		entry = rb_entry(parent, struct btrfs_ordered_extent, rb_node);
entry              42 fs/btrfs/ordered-data.c 		if (file_offset < entry->file_offset)
entry              44 fs/btrfs/ordered-data.c 		else if (file_offset >= entry_end(entry))
entry              73 fs/btrfs/ordered-data.c 	struct btrfs_ordered_extent *entry;
entry              77 fs/btrfs/ordered-data.c 		entry = rb_entry(n, struct btrfs_ordered_extent, rb_node);
entry              79 fs/btrfs/ordered-data.c 		prev_entry = entry;
entry              81 fs/btrfs/ordered-data.c 		if (file_offset < entry->file_offset)
entry              83 fs/btrfs/ordered-data.c 		else if (file_offset >= entry_end(entry))
entry             120 fs/btrfs/ordered-data.c static int offset_in_entry(struct btrfs_ordered_extent *entry, u64 file_offset)
entry             122 fs/btrfs/ordered-data.c 	if (file_offset < entry->file_offset ||
entry             123 fs/btrfs/ordered-data.c 	    entry->file_offset + entry->len <= file_offset)
entry             128 fs/btrfs/ordered-data.c static int range_overlaps(struct btrfs_ordered_extent *entry, u64 file_offset,
entry             131 fs/btrfs/ordered-data.c 	if (file_offset + len <= entry->file_offset ||
entry             132 fs/btrfs/ordered-data.c 	    entry->file_offset + entry->len <= file_offset)
entry             147 fs/btrfs/ordered-data.c 	struct btrfs_ordered_extent *entry;
entry             150 fs/btrfs/ordered-data.c 		entry = rb_entry(tree->last, struct btrfs_ordered_extent,
entry             152 fs/btrfs/ordered-data.c 		if (offset_in_entry(entry, file_offset))
entry             182 fs/btrfs/ordered-data.c 	struct btrfs_ordered_extent *entry;
entry             185 fs/btrfs/ordered-data.c 	entry = kmem_cache_zalloc(btrfs_ordered_extent_cache, GFP_NOFS);
entry             186 fs/btrfs/ordered-data.c 	if (!entry)
entry             189 fs/btrfs/ordered-data.c 	entry->file_offset = file_offset;
entry             190 fs/btrfs/ordered-data.c 	entry->start = start;
entry             191 fs/btrfs/ordered-data.c 	entry->len = len;
entry             192 fs/btrfs/ordered-data.c 	entry->disk_len = disk_len;
entry             193 fs/btrfs/ordered-data.c 	entry->bytes_left = len;
entry             194 fs/btrfs/ordered-data.c 	entry->inode = igrab(inode);
entry             195 fs/btrfs/ordered-data.c 	entry->compress_type = compress_type;
entry             196 fs/btrfs/ordered-data.c 	entry->truncated_len = (u64)-1;
entry             198 fs/btrfs/ordered-data.c 		set_bit(type, &entry->flags);
entry             203 fs/btrfs/ordered-data.c 		set_bit(BTRFS_ORDERED_DIRECT, &entry->flags);
entry             207 fs/btrfs/ordered-data.c 	refcount_set(&entry->refs, 1);
entry             208 fs/btrfs/ordered-data.c 	init_waitqueue_head(&entry->wait);
entry             209 fs/btrfs/ordered-data.c 	INIT_LIST_HEAD(&entry->list);
entry             210 fs/btrfs/ordered-data.c 	INIT_LIST_HEAD(&entry->root_extent_list);
entry             211 fs/btrfs/ordered-data.c 	INIT_LIST_HEAD(&entry->work_list);
entry             212 fs/btrfs/ordered-data.c 	init_completion(&entry->completion);
entry             213 fs/btrfs/ordered-data.c 	INIT_LIST_HEAD(&entry->log_list);
entry             214 fs/btrfs/ordered-data.c 	INIT_LIST_HEAD(&entry->trans_list);
entry             216 fs/btrfs/ordered-data.c 	trace_btrfs_ordered_extent_add(inode, entry);
entry             220 fs/btrfs/ordered-data.c 			   &entry->rb_node);
entry             226 fs/btrfs/ordered-data.c 	list_add_tail(&entry->root_extent_list,
entry             279 fs/btrfs/ordered-data.c void btrfs_add_ordered_sum(struct btrfs_ordered_extent *entry,
entry             284 fs/btrfs/ordered-data.c 	tree = &BTRFS_I(entry->inode)->ordered_tree;
entry             286 fs/btrfs/ordered-data.c 	list_add_tail(&sum->list, &entry->list);
entry             309 fs/btrfs/ordered-data.c 	struct btrfs_ordered_extent *entry = NULL;
entry             324 fs/btrfs/ordered-data.c 	entry = rb_entry(node, struct btrfs_ordered_extent, rb_node);
entry             325 fs/btrfs/ordered-data.c 	if (!offset_in_entry(entry, *file_offset)) {
entry             330 fs/btrfs/ordered-data.c 	dec_start = max(*file_offset, entry->file_offset);
entry             331 fs/btrfs/ordered-data.c 	dec_end = min(*file_offset + io_size, entry->file_offset +
entry             332 fs/btrfs/ordered-data.c 		      entry->len);
entry             339 fs/btrfs/ordered-data.c 	if (to_dec > entry->bytes_left) {
entry             342 fs/btrfs/ordered-data.c 			   entry->bytes_left, to_dec);
entry             344 fs/btrfs/ordered-data.c 	entry->bytes_left -= to_dec;
entry             346 fs/btrfs/ordered-data.c 		set_bit(BTRFS_ORDERED_IOERR, &entry->flags);
entry             348 fs/btrfs/ordered-data.c 	if (entry->bytes_left == 0) {
entry             349 fs/btrfs/ordered-data.c 		ret = test_and_set_bit(BTRFS_ORDERED_IO_DONE, &entry->flags);
entry             351 fs/btrfs/ordered-data.c 		cond_wake_up_nomb(&entry->wait);
entry             356 fs/btrfs/ordered-data.c 	if (!ret && cached && entry) {
entry             357 fs/btrfs/ordered-data.c 		*cached = entry;
entry             358 fs/btrfs/ordered-data.c 		refcount_inc(&entry->refs);
entry             379 fs/btrfs/ordered-data.c 	struct btrfs_ordered_extent *entry = NULL;
entry             386 fs/btrfs/ordered-data.c 		entry = *cached;
entry             396 fs/btrfs/ordered-data.c 	entry = rb_entry(node, struct btrfs_ordered_extent, rb_node);
entry             398 fs/btrfs/ordered-data.c 	if (!offset_in_entry(entry, file_offset)) {
entry             403 fs/btrfs/ordered-data.c 	if (io_size > entry->bytes_left) {
entry             406 fs/btrfs/ordered-data.c 		       entry->bytes_left, io_size);
entry             408 fs/btrfs/ordered-data.c 	entry->bytes_left -= io_size;
entry             410 fs/btrfs/ordered-data.c 		set_bit(BTRFS_ORDERED_IOERR, &entry->flags);
entry             412 fs/btrfs/ordered-data.c 	if (entry->bytes_left == 0) {
entry             413 fs/btrfs/ordered-data.c 		ret = test_and_set_bit(BTRFS_ORDERED_IO_DONE, &entry->flags);
entry             415 fs/btrfs/ordered-data.c 		cond_wake_up_nomb(&entry->wait);
entry             420 fs/btrfs/ordered-data.c 	if (!ret && cached && entry) {
entry             421 fs/btrfs/ordered-data.c 		*cached = entry;
entry             422 fs/btrfs/ordered-data.c 		refcount_inc(&entry->refs);
entry             432 fs/btrfs/ordered-data.c void btrfs_put_ordered_extent(struct btrfs_ordered_extent *entry)
entry             437 fs/btrfs/ordered-data.c 	trace_btrfs_ordered_extent_put(entry->inode, entry);
entry             439 fs/btrfs/ordered-data.c 	if (refcount_dec_and_test(&entry->refs)) {
entry             440 fs/btrfs/ordered-data.c 		ASSERT(list_empty(&entry->log_list));
entry             441 fs/btrfs/ordered-data.c 		ASSERT(list_empty(&entry->trans_list));
entry             442 fs/btrfs/ordered-data.c 		ASSERT(list_empty(&entry->root_extent_list));
entry             443 fs/btrfs/ordered-data.c 		ASSERT(RB_EMPTY_NODE(&entry->rb_node));
entry             444 fs/btrfs/ordered-data.c 		if (entry->inode)
entry             445 fs/btrfs/ordered-data.c 			btrfs_add_delayed_iput(entry->inode);
entry             446 fs/btrfs/ordered-data.c 		while (!list_empty(&entry->list)) {
entry             447 fs/btrfs/ordered-data.c 			cur = entry->list.next;
entry             452 fs/btrfs/ordered-data.c 		kmem_cache_free(btrfs_ordered_extent_cache, entry);
entry             461 fs/btrfs/ordered-data.c 				 struct btrfs_ordered_extent *entry)
entry             474 fs/btrfs/ordered-data.c 		btrfs_delalloc_release_metadata(btrfs_inode, entry->len, false);
entry             476 fs/btrfs/ordered-data.c 	if (test_bit(BTRFS_ORDERED_DIRECT, &entry->flags))
entry             477 fs/btrfs/ordered-data.c 		percpu_counter_add_batch(&fs_info->dio_bytes, -entry->len,
entry             482 fs/btrfs/ordered-data.c 	node = &entry->rb_node;
entry             487 fs/btrfs/ordered-data.c 	set_bit(BTRFS_ORDERED_COMPLETE, &entry->flags);
entry             491 fs/btrfs/ordered-data.c 	list_del_init(&entry->root_extent_list);
entry             494 fs/btrfs/ordered-data.c 	trace_btrfs_ordered_extent_remove(inode, entry);
entry             503 fs/btrfs/ordered-data.c 	wake_up(&entry->wait);
entry             622 fs/btrfs/ordered-data.c 				       struct btrfs_ordered_extent *entry,
entry             625 fs/btrfs/ordered-data.c 	u64 start = entry->file_offset;
entry             626 fs/btrfs/ordered-data.c 	u64 end = start + entry->len - 1;
entry             628 fs/btrfs/ordered-data.c 	trace_btrfs_ordered_extent_start(inode, entry);
entry             635 fs/btrfs/ordered-data.c 	if (!test_bit(BTRFS_ORDERED_DIRECT, &entry->flags))
entry             638 fs/btrfs/ordered-data.c 		wait_event(entry->wait, test_bit(BTRFS_ORDERED_COMPLETE,
entry             639 fs/btrfs/ordered-data.c 						 &entry->flags));
entry             717 fs/btrfs/ordered-data.c 	struct btrfs_ordered_extent *entry = NULL;
entry             725 fs/btrfs/ordered-data.c 	entry = rb_entry(node, struct btrfs_ordered_extent, rb_node);
entry             726 fs/btrfs/ordered-data.c 	if (!offset_in_entry(entry, file_offset))
entry             727 fs/btrfs/ordered-data.c 		entry = NULL;
entry             728 fs/btrfs/ordered-data.c 	if (entry)
entry             729 fs/btrfs/ordered-data.c 		refcount_inc(&entry->refs);
entry             732 fs/btrfs/ordered-data.c 	return entry;
entry             743 fs/btrfs/ordered-data.c 	struct btrfs_ordered_extent *entry = NULL;
entry             755 fs/btrfs/ordered-data.c 		entry = rb_entry(node, struct btrfs_ordered_extent, rb_node);
entry             756 fs/btrfs/ordered-data.c 		if (range_overlaps(entry, file_offset, len))
entry             759 fs/btrfs/ordered-data.c 		if (entry->file_offset >= file_offset + len) {
entry             760 fs/btrfs/ordered-data.c 			entry = NULL;
entry             763 fs/btrfs/ordered-data.c 		entry = NULL;
entry             769 fs/btrfs/ordered-data.c 	if (entry)
entry             770 fs/btrfs/ordered-data.c 		refcount_inc(&entry->refs);
entry             772 fs/btrfs/ordered-data.c 	return entry;
entry             784 fs/btrfs/ordered-data.c 	struct btrfs_ordered_extent *entry = NULL;
entry             792 fs/btrfs/ordered-data.c 	entry = rb_entry(node, struct btrfs_ordered_extent, rb_node);
entry             793 fs/btrfs/ordered-data.c 	refcount_inc(&entry->refs);
entry             796 fs/btrfs/ordered-data.c 	return entry;
entry             153 fs/btrfs/ordered-data.h void btrfs_put_ordered_extent(struct btrfs_ordered_extent *entry);
entry             155 fs/btrfs/ordered-data.h 				struct btrfs_ordered_extent *entry);
entry             170 fs/btrfs/ordered-data.h void btrfs_add_ordered_sum(struct btrfs_ordered_extent *entry,
entry             175 fs/btrfs/ordered-data.h 				struct btrfs_ordered_extent *entry, int wait);
entry            1550 fs/btrfs/qgroup.c 	struct btrfs_qgroup_extent_record *entry;
entry            1558 fs/btrfs/qgroup.c 		entry = rb_entry(parent_node, struct btrfs_qgroup_extent_record,
entry            1560 fs/btrfs/qgroup.c 		if (bytenr < entry->bytenr) {
entry            1562 fs/btrfs/qgroup.c 		} else if (bytenr > entry->bytenr) {
entry            1565 fs/btrfs/qgroup.c 			if (record->data_rsv && !entry->data_rsv) {
entry            1566 fs/btrfs/qgroup.c 				entry->data_rsv = record->data_rsv;
entry            1567 fs/btrfs/qgroup.c 				entry->data_rsv_refroot =
entry            3814 fs/btrfs/qgroup.c 		struct btrfs_qgroup_swapped_block *entry;
entry            3817 fs/btrfs/qgroup.c 		rbtree_postorder_for_each_entry_safe(entry, next, cur_root,
entry            3819 fs/btrfs/qgroup.c 			kfree(entry);
entry            3899 fs/btrfs/qgroup.c 		struct btrfs_qgroup_swapped_block *entry;
entry            3902 fs/btrfs/qgroup.c 		entry = rb_entry(parent, struct btrfs_qgroup_swapped_block,
entry            3905 fs/btrfs/qgroup.c 		if (entry->subvol_bytenr < block->subvol_bytenr) {
entry            3907 fs/btrfs/qgroup.c 		} else if (entry->subvol_bytenr > block->subvol_bytenr) {
entry            3910 fs/btrfs/qgroup.c 			if (entry->subvol_generation !=
entry            3912 fs/btrfs/qgroup.c 			    entry->reloc_bytenr != block->reloc_bytenr ||
entry            3913 fs/btrfs/qgroup.c 			    entry->reloc_generation !=
entry            4031 fs/btrfs/qgroup.c 	struct btrfs_qgroup_extent_record *entry;
entry            4036 fs/btrfs/qgroup.c 	rbtree_postorder_for_each_entry_safe(entry, next, root, node) {
entry            4037 fs/btrfs/qgroup.c 		ulist_free(entry->old_roots);
entry            4038 fs/btrfs/qgroup.c 		kfree(entry);
entry              80 fs/btrfs/ref-verify.c 	struct block_entry *entry;
entry              84 fs/btrfs/ref-verify.c 		entry = rb_entry(parent_node, struct block_entry, node);
entry              85 fs/btrfs/ref-verify.c 		if (entry->bytenr > be->bytenr)
entry              87 fs/btrfs/ref-verify.c 		else if (entry->bytenr < be->bytenr)
entry              90 fs/btrfs/ref-verify.c 			return entry;
entry             101 fs/btrfs/ref-verify.c 	struct block_entry *entry = NULL;
entry             105 fs/btrfs/ref-verify.c 		entry = rb_entry(n, struct block_entry, node);
entry             106 fs/btrfs/ref-verify.c 		if (entry->bytenr < bytenr)
entry             108 fs/btrfs/ref-verify.c 		else if (entry->bytenr > bytenr)
entry             111 fs/btrfs/ref-verify.c 			return entry;
entry             121 fs/btrfs/ref-verify.c 	struct root_entry *entry;
entry             125 fs/btrfs/ref-verify.c 		entry = rb_entry(parent_node, struct root_entry, node);
entry             126 fs/btrfs/ref-verify.c 		if (entry->root_objectid > re->root_objectid)
entry             128 fs/btrfs/ref-verify.c 		else if (entry->root_objectid < re->root_objectid)
entry             131 fs/btrfs/ref-verify.c 			return entry;
entry             166 fs/btrfs/ref-verify.c 	struct ref_entry *entry;
entry             171 fs/btrfs/ref-verify.c 		entry = rb_entry(parent_node, struct ref_entry, node);
entry             172 fs/btrfs/ref-verify.c 		cmp = comp_refs(entry, ref);
entry             178 fs/btrfs/ref-verify.c 			return entry;
entry             190 fs/btrfs/ref-verify.c 	struct root_entry *entry = NULL;
entry             194 fs/btrfs/ref-verify.c 		entry = rb_entry(n, struct root_entry, node);
entry             195 fs/btrfs/ref-verify.c 		if (entry->root_objectid < objectid)
entry             197 fs/btrfs/ref-verify.c 		else if (entry->root_objectid > objectid)
entry             200 fs/btrfs/ref-verify.c 			return entry;
entry             915 fs/btrfs/ref-verify.c 	struct block_entry *be = NULL, *entry;
entry             924 fs/btrfs/ref-verify.c 		entry = rb_entry(n, struct block_entry, node);
entry             925 fs/btrfs/ref-verify.c 		if (entry->bytenr < start) {
entry             927 fs/btrfs/ref-verify.c 		} else if (entry->bytenr > start) {
entry             930 fs/btrfs/ref-verify.c 			be = entry;
entry             935 fs/btrfs/ref-verify.c 		    (entry->bytenr < start && be->bytenr > start) ||
entry             936 fs/btrfs/ref-verify.c 		    (entry->bytenr < start && entry->bytenr > be->bytenr))
entry             937 fs/btrfs/ref-verify.c 			be = entry;
entry             287 fs/btrfs/relocation.c 	struct tree_entry *entry;
entry             291 fs/btrfs/relocation.c 		entry = rb_entry(parent, struct tree_entry, rb_node);
entry             293 fs/btrfs/relocation.c 		if (bytenr < entry->bytenr)
entry             295 fs/btrfs/relocation.c 		else if (bytenr > entry->bytenr)
entry             309 fs/btrfs/relocation.c 	struct tree_entry *entry;
entry             312 fs/btrfs/relocation.c 		entry = rb_entry(n, struct tree_entry, rb_node);
entry             314 fs/btrfs/relocation.c 		if (bytenr < entry->bytenr)
entry             316 fs/btrfs/relocation.c 		else if (bytenr > entry->bytenr)
entry            1555 fs/btrfs/relocation.c 	struct btrfs_inode *entry;
entry            1564 fs/btrfs/relocation.c 		entry = rb_entry(node, struct btrfs_inode, rb_node);
entry            1566 fs/btrfs/relocation.c 		if (objectid < btrfs_ino(entry))
entry            1568 fs/btrfs/relocation.c 		else if (objectid > btrfs_ino(entry))
entry            1575 fs/btrfs/relocation.c 			entry = rb_entry(prev, struct btrfs_inode, rb_node);
entry            1576 fs/btrfs/relocation.c 			if (objectid <= btrfs_ino(entry)) {
entry            1584 fs/btrfs/relocation.c 		entry = rb_entry(node, struct btrfs_inode, rb_node);
entry            1585 fs/btrfs/relocation.c 		inode = igrab(&entry->vfs_inode);
entry            1591 fs/btrfs/relocation.c 		objectid = btrfs_ino(entry) + 1;
entry             325 fs/btrfs/scrub.c 	struct full_stripe_lock *entry;
entry             333 fs/btrfs/scrub.c 		entry = rb_entry(parent, struct full_stripe_lock, node);
entry             334 fs/btrfs/scrub.c 		if (fstripe_logical < entry->logical) {
entry             336 fs/btrfs/scrub.c 		} else if (fstripe_logical > entry->logical) {
entry             339 fs/btrfs/scrub.c 			entry->refs++;
entry             340 fs/btrfs/scrub.c 			return entry;
entry             370 fs/btrfs/scrub.c 	struct full_stripe_lock *entry;
entry             376 fs/btrfs/scrub.c 		entry = rb_entry(node, struct full_stripe_lock, node);
entry             377 fs/btrfs/scrub.c 		if (fstripe_logical < entry->logical)
entry             379 fs/btrfs/scrub.c 		else if (fstripe_logical > entry->logical)
entry             382 fs/btrfs/scrub.c 			return entry;
entry            2863 fs/btrfs/send.c 	struct orphan_dir_info *entry, *odi;
entry            2867 fs/btrfs/send.c 		entry = rb_entry(parent, struct orphan_dir_info, node);
entry            2868 fs/btrfs/send.c 		if (dir_ino < entry->ino) {
entry            2870 fs/btrfs/send.c 		} else if (dir_ino > entry->ino) {
entry            2873 fs/btrfs/send.c 			return entry;
entry            2893 fs/btrfs/send.c 	struct orphan_dir_info *entry;
entry            2896 fs/btrfs/send.c 		entry = rb_entry(n, struct orphan_dir_info, node);
entry            2897 fs/btrfs/send.c 		if (dir_ino < entry->ino)
entry            2899 fs/btrfs/send.c 		else if (dir_ino > entry->ino)
entry            2902 fs/btrfs/send.c 			return entry;
entry            3022 fs/btrfs/send.c 	struct waiting_dir_move *entry = get_waiting_dir_move(sctx, ino);
entry            3024 fs/btrfs/send.c 	return entry != NULL;
entry            3031 fs/btrfs/send.c 	struct waiting_dir_move *entry, *dm;
entry            3042 fs/btrfs/send.c 		entry = rb_entry(parent, struct waiting_dir_move, node);
entry            3043 fs/btrfs/send.c 		if (ino < entry->ino) {
entry            3045 fs/btrfs/send.c 		} else if (ino > entry->ino) {
entry            3062 fs/btrfs/send.c 	struct waiting_dir_move *entry;
entry            3065 fs/btrfs/send.c 		entry = rb_entry(n, struct waiting_dir_move, node);
entry            3066 fs/btrfs/send.c 		if (ino < entry->ino)
entry            3068 fs/btrfs/send.c 		else if (ino > entry->ino)
entry            3071 fs/btrfs/send.c 			return entry;
entry            3095 fs/btrfs/send.c 	struct pending_dir_move *entry = NULL, *pm;
entry            3112 fs/btrfs/send.c 		entry = rb_entry(parent, struct pending_dir_move, node);
entry            3113 fs/btrfs/send.c 		if (parent_ino < entry->parent_ino) {
entry            3115 fs/btrfs/send.c 		} else if (parent_ino > entry->parent_ino) {
entry            3139 fs/btrfs/send.c 		list_add_tail(&pm->list, &entry->list);
entry            3157 fs/btrfs/send.c 	struct pending_dir_move *entry;
entry            3160 fs/btrfs/send.c 		entry = rb_entry(n, struct pending_dir_move, node);
entry            3161 fs/btrfs/send.c 		if (parent_ino < entry->parent_ino)
entry            3163 fs/btrfs/send.c 		else if (parent_ino > entry->parent_ino)
entry            3166 fs/btrfs/send.c 			return entry;
entry              48 fs/cachefiles/rdwr.c 	list_del(&wait->entry);
entry              41 fs/coda/dir.c  static struct dentry *coda_lookup(struct inode *dir, struct dentry *entry, unsigned int flags)
entry              44 fs/coda/dir.c  	const char *name = entry->d_name.name;
entry              45 fs/coda/dir.c  	size_t length = entry->d_name.len;
entry              72 fs/coda/dir.c  	return d_splice_alias(inode, entry);
entry              24 fs/configfs/item.c static inline struct config_item *to_item(struct list_head *entry)
entry              26 fs/configfs/item.c 	return container_of(entry, struct config_item, ci_entry);
entry             179 fs/configfs/item.c 	struct list_head *entry;
entry             182 fs/configfs/item.c 	list_for_each(entry, &group->cg_children) {
entry             183 fs/configfs/item.c 		struct config_item *item = to_item(entry);
entry              83 fs/dax.c       static unsigned long dax_to_pfn(void *entry)
entry              85 fs/dax.c       	return xa_to_value(entry) >> DAX_SHIFT;
entry              93 fs/dax.c       static bool dax_is_locked(void *entry)
entry              95 fs/dax.c       	return xa_to_value(entry) & DAX_LOCKED;
entry              98 fs/dax.c       static unsigned int dax_entry_order(void *entry)
entry             100 fs/dax.c       	if (xa_to_value(entry) & DAX_PMD)
entry             105 fs/dax.c       static unsigned long dax_is_pmd_entry(void *entry)
entry             107 fs/dax.c       	return xa_to_value(entry) & DAX_PMD;
entry             110 fs/dax.c       static bool dax_is_pte_entry(void *entry)
entry             112 fs/dax.c       	return !(xa_to_value(entry) & DAX_PMD);
entry             115 fs/dax.c       static int dax_is_zero_entry(void *entry)
entry             117 fs/dax.c       	return xa_to_value(entry) & DAX_ZERO_PAGE;
entry             120 fs/dax.c       static int dax_is_empty_entry(void *entry)
entry             122 fs/dax.c       	return xa_to_value(entry) & DAX_EMPTY;
entry             129 fs/dax.c       static bool dax_is_conflict(void *entry)
entry             131 fs/dax.c       	return entry == XA_RETRY_ENTRY;
entry             148 fs/dax.c       		void *entry, struct exceptional_entry_key *key)
entry             158 fs/dax.c       	if (dax_is_pmd_entry(entry))
entry             185 fs/dax.c       static void dax_wake_entry(struct xa_state *xas, void *entry, bool wake_all)
entry             190 fs/dax.c       	wq = dax_entry_waitqueue(xas, entry, &key);
entry             214 fs/dax.c       	void *entry;
entry             222 fs/dax.c       		entry = xas_find_conflict(xas);
entry             223 fs/dax.c       		if (!entry || WARN_ON_ONCE(!xa_is_value(entry)))
entry             224 fs/dax.c       			return entry;
entry             225 fs/dax.c       		if (dax_entry_order(entry) < order)
entry             227 fs/dax.c       		if (!dax_is_locked(entry))
entry             228 fs/dax.c       			return entry;
entry             230 fs/dax.c       		wq = dax_entry_waitqueue(xas, entry, &ewait.key);
entry             246 fs/dax.c       static void wait_entry_unlocked(struct xa_state *xas, void *entry)
entry             254 fs/dax.c       	wq = dax_entry_waitqueue(xas, entry, &ewait.key);
entry             267 fs/dax.c       static void put_unlocked_entry(struct xa_state *xas, void *entry)
entry             270 fs/dax.c       	if (entry && !dax_is_conflict(entry))
entry             271 fs/dax.c       		dax_wake_entry(xas, entry, false);
entry             279 fs/dax.c       static void dax_unlock_entry(struct xa_state *xas, void *entry)
entry             283 fs/dax.c       	BUG_ON(dax_is_locked(entry));
entry             286 fs/dax.c       	old = xas_store(xas, entry);
entry             289 fs/dax.c       	dax_wake_entry(xas, entry, false);
entry             295 fs/dax.c       static void *dax_lock_entry(struct xa_state *xas, void *entry)
entry             297 fs/dax.c       	unsigned long v = xa_to_value(entry);
entry             301 fs/dax.c       static unsigned long dax_entry_size(void *entry)
entry             303 fs/dax.c       	if (dax_is_zero_entry(entry))
entry             305 fs/dax.c       	else if (dax_is_empty_entry(entry))
entry             307 fs/dax.c       	else if (dax_is_pmd_entry(entry))
entry             313 fs/dax.c       static unsigned long dax_end_pfn(void *entry)
entry             315 fs/dax.c       	return dax_to_pfn(entry) + dax_entry_size(entry) / PAGE_SIZE;
entry             322 fs/dax.c       #define for_each_mapped_pfn(entry, pfn) \
entry             323 fs/dax.c       	for (pfn = dax_to_pfn(entry); \
entry             324 fs/dax.c       			pfn < dax_end_pfn(entry); pfn++)
entry             331 fs/dax.c       static void dax_associate_entry(void *entry, struct address_space *mapping,
entry             334 fs/dax.c       	unsigned long size = dax_entry_size(entry), pfn, index;
entry             341 fs/dax.c       	for_each_mapped_pfn(entry, pfn) {
entry             350 fs/dax.c       static void dax_disassociate_entry(void *entry, struct address_space *mapping,
entry             358 fs/dax.c       	for_each_mapped_pfn(entry, pfn) {
entry             368 fs/dax.c       static struct page *dax_busy_page(void *entry)
entry             372 fs/dax.c       	for_each_mapped_pfn(entry, pfn) {
entry             392 fs/dax.c       	void *entry;
entry             399 fs/dax.c       		entry = NULL;
entry             410 fs/dax.c       		entry = (void *)~0UL;
entry             421 fs/dax.c       		entry = xas_load(&xas);
entry             422 fs/dax.c       		if (dax_is_locked(entry)) {
entry             424 fs/dax.c       			wait_entry_unlocked(&xas, entry);
entry             428 fs/dax.c       		dax_lock_entry(&xas, entry);
entry             433 fs/dax.c       	return (dax_entry_t)entry;
entry             481 fs/dax.c       	void *entry;
entry             485 fs/dax.c       	entry = get_unlocked_entry(xas, order);
entry             487 fs/dax.c       	if (entry) {
entry             488 fs/dax.c       		if (dax_is_conflict(entry))
entry             490 fs/dax.c       		if (!xa_is_value(entry)) {
entry             496 fs/dax.c       			if (dax_is_pmd_entry(entry) &&
entry             497 fs/dax.c       			    (dax_is_zero_entry(entry) ||
entry             498 fs/dax.c       			     dax_is_empty_entry(entry))) {
entry             509 fs/dax.c       		dax_lock_entry(xas, entry);
entry             516 fs/dax.c       		if (dax_is_zero_entry(entry)) {
entry             525 fs/dax.c       		dax_disassociate_entry(entry, mapping, false);
entry             527 fs/dax.c       		dax_wake_entry(xas, entry, true);
entry             529 fs/dax.c       		entry = NULL;
entry             533 fs/dax.c       	if (entry) {
entry             534 fs/dax.c       		dax_lock_entry(xas, entry);
entry             540 fs/dax.c       		entry = dax_make_entry(pfn_to_pfn_t(0), flags);
entry             541 fs/dax.c       		dax_lock_entry(xas, entry);
entry             555 fs/dax.c       	return entry;
entry             579 fs/dax.c       	void *entry;
entry             607 fs/dax.c       	xas_for_each(&xas, entry, ULONG_MAX) {
entry             608 fs/dax.c       		if (WARN_ON_ONCE(!xa_is_value(entry)))
entry             610 fs/dax.c       		if (unlikely(dax_is_locked(entry)))
entry             611 fs/dax.c       			entry = get_unlocked_entry(&xas, 0);
entry             612 fs/dax.c       		if (entry)
entry             613 fs/dax.c       			page = dax_busy_page(entry);
entry             614 fs/dax.c       		put_unlocked_entry(&xas, entry);
entry             635 fs/dax.c       	void *entry;
entry             638 fs/dax.c       	entry = get_unlocked_entry(&xas, 0);
entry             639 fs/dax.c       	if (!entry || WARN_ON_ONCE(!xa_is_value(entry)))
entry             645 fs/dax.c       	dax_disassociate_entry(entry, mapping, trunc);
entry             650 fs/dax.c       	put_unlocked_entry(&xas, entry);
entry             718 fs/dax.c       		void *entry, pfn_t pfn, unsigned long flags, bool dirty)
entry             725 fs/dax.c       	if (dax_is_zero_entry(entry) && !(flags & DAX_ZERO_PAGE)) {
entry             728 fs/dax.c       		if (dax_is_pmd_entry(entry))
entry             737 fs/dax.c       	if (dax_is_zero_entry(entry) || dax_is_empty_entry(entry)) {
entry             740 fs/dax.c       		dax_disassociate_entry(entry, mapping, false);
entry             751 fs/dax.c       		WARN_ON_ONCE(old != xa_mk_value(xa_to_value(entry) |
entry             753 fs/dax.c       		entry = new_entry;
entry             762 fs/dax.c       	return entry;
entry             850 fs/dax.c       		struct address_space *mapping, void *entry)
entry             859 fs/dax.c       	if (WARN_ON(!xa_is_value(entry)))
entry             862 fs/dax.c       	if (unlikely(dax_is_locked(entry))) {
entry             863 fs/dax.c       		void *old_entry = entry;
entry             865 fs/dax.c       		entry = get_unlocked_entry(xas, 0);
entry             868 fs/dax.c       		if (!entry || WARN_ON_ONCE(!xa_is_value(entry)))
entry             875 fs/dax.c       		if (dax_to_pfn(old_entry) != dax_to_pfn(entry))
entry             877 fs/dax.c       		if (WARN_ON_ONCE(dax_is_empty_entry(entry) ||
entry             878 fs/dax.c       					dax_is_zero_entry(entry))) {
entry             889 fs/dax.c       	dax_lock_entry(xas, entry);
entry             908 fs/dax.c       	pfn = dax_to_pfn(entry);
entry             909 fs/dax.c       	count = 1UL << dax_entry_order(entry);
entry             922 fs/dax.c       	xas_store(xas, entry);
entry             924 fs/dax.c       	dax_wake_entry(xas, entry, false);
entry             930 fs/dax.c       	put_unlocked_entry(xas, entry);
entry             946 fs/dax.c       	void *entry;
entry             965 fs/dax.c       	xas_for_each_marked(&xas, entry, end_index, PAGECACHE_TAG_TOWRITE) {
entry             966 fs/dax.c       		ret = dax_writeback_one(&xas, dax_dev, mapping, entry);
entry            1031 fs/dax.c       		struct address_space *mapping, void **entry,
entry            1039 fs/dax.c       	*entry = dax_insert_entry(xas, mapping, vmf, *entry, pfn,
entry            1260 fs/dax.c       	void *entry;
entry            1277 fs/dax.c       	entry = grab_mapping_entry(&xas, mapping, 0);
entry            1278 fs/dax.c       	if (xa_is_internal(entry)) {
entry            1279 fs/dax.c       		ret = xa_to_internal(entry);
entry            1352 fs/dax.c       		entry = dax_insert_entry(&xas, mapping, vmf, entry, pfn,
entry            1370 fs/dax.c       		trace_dax_insert_mapping(inode, vmf, entry);
entry            1380 fs/dax.c       			ret = dax_load_hole(&xas, mapping, &entry, vmf);
entry            1407 fs/dax.c       	dax_unlock_entry(&xas, entry);
entry            1415 fs/dax.c       		struct iomap *iomap, void **entry)
entry            1433 fs/dax.c       	*entry = dax_insert_entry(xas, mapping, vmf, *entry, pfn,
entry            1456 fs/dax.c       	trace_dax_pmd_load_hole(inode, vmf, zero_page, *entry);
entry            1462 fs/dax.c       	trace_dax_pmd_load_hole_fallback(inode, vmf, zero_page, *entry);
entry            1480 fs/dax.c       	void *entry;
entry            1529 fs/dax.c       	entry = grab_mapping_entry(&xas, mapping, PMD_ORDER);
entry            1530 fs/dax.c       	if (xa_is_internal(entry)) {
entry            1531 fs/dax.c       		result = xa_to_internal(entry);
entry            1568 fs/dax.c       		entry = dax_insert_entry(&xas, mapping, vmf, entry, pfn,
entry            1585 fs/dax.c       		trace_dax_pmd_insert_mapping(inode, vmf, PMD_SIZE, pfn, entry);
entry            1592 fs/dax.c       		result = dax_pmd_load_hole(&xas, vmf, &iomap, &entry);
entry            1615 fs/dax.c       	dax_unlock_entry(&xas, entry);
entry            1674 fs/dax.c       	void *entry;
entry            1678 fs/dax.c       	entry = get_unlocked_entry(&xas, order);
entry            1680 fs/dax.c       	if (!entry || dax_is_conflict(entry) ||
entry            1681 fs/dax.c       	    (order == 0 && !dax_is_pte_entry(entry))) {
entry            1682 fs/dax.c       		put_unlocked_entry(&xas, entry);
entry            1689 fs/dax.c       	dax_lock_entry(&xas, entry);
entry            1699 fs/dax.c       	dax_unlock_entry(&xas, entry);
entry            1948 fs/dcache.c    void d_instantiate(struct dentry *entry, struct inode * inode)
entry            1950 fs/dcache.c    	BUG_ON(!hlist_unhashed(&entry->d_u.d_alias));
entry            1952 fs/dcache.c    		security_d_instantiate(entry, inode);
entry            1954 fs/dcache.c    		__d_instantiate(entry, inode);
entry            1966 fs/dcache.c    void d_instantiate_new(struct dentry *entry, struct inode *inode)
entry            1968 fs/dcache.c    	BUG_ON(!hlist_unhashed(&entry->d_u.d_alias));
entry            1971 fs/dcache.c    	security_d_instantiate(entry, inode);
entry            1973 fs/dcache.c    	__d_instantiate(entry, inode);
entry            2458 fs/dcache.c    static void __d_rehash(struct dentry *entry)
entry            2460 fs/dcache.c    	struct hlist_bl_head *b = d_hash(entry->d_name.hash);
entry            2463 fs/dcache.c    	hlist_bl_add_head_rcu(&entry->d_hash, b);
entry            2474 fs/dcache.c    void d_rehash(struct dentry * entry)
entry            2476 fs/dcache.c    	spin_lock(&entry->d_lock);
entry            2477 fs/dcache.c    	__d_rehash(entry);
entry            2478 fs/dcache.c    	spin_unlock(&entry->d_lock);
entry            2674 fs/dcache.c    void d_add(struct dentry *entry, struct inode *inode)
entry            2677 fs/dcache.c    		security_d_instantiate(entry, inode);
entry            2680 fs/dcache.c    	__d_add(entry, inode);
entry            2695 fs/dcache.c    struct dentry *d_exact_alias(struct dentry *entry, struct inode *inode)
entry            2698 fs/dcache.c    	unsigned int hash = entry->d_name.hash;
entry            2709 fs/dcache.c    		if (alias->d_parent != entry->d_parent)
entry            2711 fs/dcache.c    		if (!d_same_name(alias, entry->d_parent, &entry->d_name))
entry            1159 fs/debugfs/file.c 	struct debugfs_devm_entry *entry = inode->i_private;
entry            1161 fs/debugfs/file.c 	return single_open(f, entry->read, entry->dev);
entry            1187 fs/debugfs/file.c 	struct debugfs_devm_entry *entry;
entry            1192 fs/debugfs/file.c 	entry = devm_kzalloc(dev, sizeof(*entry), GFP_KERNEL);
entry            1193 fs/debugfs/file.c 	if (!entry)
entry            1196 fs/debugfs/file.c 	entry->read = read_fn;
entry            1197 fs/debugfs/file.c 	entry->dev = dev;
entry            1199 fs/debugfs/file.c 	return debugfs_create_file(name, S_IRUGO, parent, entry,
entry             427 fs/dlm/debug_fs.c 	unsigned bucket, entry;
entry             431 fs/dlm/debug_fs.c 	entry = n & ((1LL << 32) - 1);
entry             456 fs/dlm/debug_fs.c 			if (!entry--) {
entry            1402 fs/dlm/lowcomms.c 	struct writequeue_entry *entry;
entry            1404 fs/dlm/lowcomms.c 	entry = kmalloc(sizeof(struct writequeue_entry), allocation);
entry            1405 fs/dlm/lowcomms.c 	if (!entry)
entry            1408 fs/dlm/lowcomms.c 	entry->page = alloc_page(allocation);
entry            1409 fs/dlm/lowcomms.c 	if (!entry->page) {
entry            1410 fs/dlm/lowcomms.c 		kfree(entry);
entry            1414 fs/dlm/lowcomms.c 	entry->offset = 0;
entry            1415 fs/dlm/lowcomms.c 	entry->len = 0;
entry            1416 fs/dlm/lowcomms.c 	entry->end = 0;
entry            1417 fs/dlm/lowcomms.c 	entry->users = 0;
entry            1418 fs/dlm/lowcomms.c 	entry->con = con;
entry            1420 fs/dlm/lowcomms.c 	return entry;
entry             110 fs/efivarfs/super.c 	struct efivar_entry *entry;
entry             119 fs/efivarfs/super.c 	entry = kzalloc(sizeof(*entry), GFP_KERNEL);
entry             120 fs/efivarfs/super.c 	if (!entry)
entry             123 fs/efivarfs/super.c 	memcpy(entry->var.VariableName, name16, name_size);
entry             124 fs/efivarfs/super.c 	memcpy(&(entry->var.VendorGuid), &vendor, sizeof(efi_guid_t));
entry             126 fs/efivarfs/super.c 	len = ucs2_utf8size(entry->var.VariableName);
entry             133 fs/efivarfs/super.c 	ucs2_as_utf8(name, entry->var.VariableName, len);
entry             135 fs/efivarfs/super.c 	if (efivar_variable_is_removable(entry->var.VendorGuid, name, len))
entry             140 fs/efivarfs/super.c 	efi_guid_to_str(&entry->var.VendorGuid, name + len + 1);
entry             155 fs/efivarfs/super.c 	efivar_entry_size(entry, &size);
entry             156 fs/efivarfs/super.c 	err = efivar_entry_add(entry, &efivarfs_list);
entry             164 fs/efivarfs/super.c 	inode->i_private = entry;
entry             165 fs/efivarfs/super.c 	i_size_write(inode, size + sizeof(entry->var.Attributes));
entry             176 fs/efivarfs/super.c 	kfree(entry);
entry             180 fs/efivarfs/super.c static int efivarfs_destroy(struct efivar_entry *entry, void *data)
entry             182 fs/efivarfs/super.c 	int err = efivar_entry_remove(entry);
entry             186 fs/efivarfs/super.c 	kfree(entry);
entry             155 fs/erofs/xattr.c 	int (*entry)(struct xattr_iter *_it, struct erofs_xattr_entry *entry);
entry             219 fs/erofs/xattr.c 	struct erofs_xattr_entry entry;
entry             233 fs/erofs/xattr.c 	entry = *(struct erofs_xattr_entry *)(it->kaddr + it->ofs);
entry             235 fs/erofs/xattr.c 		unsigned int entry_sz = erofs_xattr_entry_size(&entry);
entry             246 fs/erofs/xattr.c 	value_sz = le16_to_cpu(entry.e_value_size);
entry             249 fs/erofs/xattr.c 	err = op->entry(it, &entry);
entry             251 fs/erofs/xattr.c 		it->ofs += entry.e_name_len + value_sz;
entry             258 fs/erofs/xattr.c 	while (processed < entry.e_name_len) {
entry             269 fs/erofs/xattr.c 			      entry.e_name_len - processed);
entry             274 fs/erofs/xattr.c 			it->ofs += entry.e_name_len - processed + value_sz;
entry             325 fs/erofs/xattr.c 			    struct erofs_xattr_entry *entry)
entry             329 fs/erofs/xattr.c 	return (it->index != entry->e_name_index ||
entry             330 fs/erofs/xattr.c 		it->name.len != entry->e_name_len) ? -ENOATTR : 0;
entry             361 fs/erofs/xattr.c 	.entry = xattr_entrymatch,
entry             532 fs/erofs/xattr.c 			   struct erofs_xattr_entry *entry)
entry             540 fs/erofs/xattr.c 		erofs_xattr_handler(entry->e_name_index);
entry             549 fs/erofs/xattr.c 		it->buffer_ofs += prefix_len + entry->e_name_len + 1;
entry             554 fs/erofs/xattr.c 		+ entry->e_name_len + 1 > it->buffer_size)
entry             584 fs/erofs/xattr.c 	.entry = xattr_entrylist,
entry            1299 fs/eventpoll.c 		list_del_init(&wait->entry);
entry            1928 fs/eventpoll.c 	if (!list_empty_careful(&wait.entry)) {
entry              43 fs/ext2/acl.c  		ext2_acl_entry *entry =
entry              47 fs/ext2/acl.c  		acl->a_entries[n].e_tag  = le16_to_cpu(entry->e_tag);
entry              48 fs/ext2/acl.c  		acl->a_entries[n].e_perm = le16_to_cpu(entry->e_perm);
entry              64 fs/ext2/acl.c  						  le32_to_cpu(entry->e_id));
entry              72 fs/ext2/acl.c  						  le32_to_cpu(entry->e_id));
entry             107 fs/ext2/acl.c  		ext2_acl_entry *entry = (ext2_acl_entry *)e;
entry             108 fs/ext2/acl.c  		entry->e_tag  = cpu_to_le16(acl_e->e_tag);
entry             109 fs/ext2/acl.c  		entry->e_perm = cpu_to_le16(acl_e->e_perm);
entry             112 fs/ext2/acl.c  				entry->e_id = cpu_to_le32(
entry             117 fs/ext2/acl.c  				entry->e_id = cpu_to_le32(
entry              72 fs/ext2/xattr.c #define IS_LAST_ENTRY(entry) (*(__u32 *)(entry) == 0)
entry             149 fs/ext2/xattr.c ext2_xattr_entry_valid(struct ext2_xattr_entry *entry,
entry             155 fs/ext2/xattr.c 	next = EXT2_XATTR_NEXT(entry);
entry             159 fs/ext2/xattr.c 	if (entry->e_value_block != 0)
entry             162 fs/ext2/xattr.c 	size = le32_to_cpu(entry->e_value_size);
entry             164 fs/ext2/xattr.c 	    le16_to_cpu(entry->e_value_offs) + size > end_offs)
entry             172 fs/ext2/xattr.c 		     struct ext2_xattr_entry *entry)
entry             176 fs/ext2/xattr.c 	cmp = name_index - entry->e_name_index;
entry             178 fs/ext2/xattr.c 		cmp = name_len - entry->e_name_len;
entry             180 fs/ext2/xattr.c 		cmp = memcmp(name, entry->e_name, name_len);
entry             200 fs/ext2/xattr.c 	struct ext2_xattr_entry *entry;
entry             237 fs/ext2/xattr.c 	entry = FIRST_ENTRY(bh);
entry             238 fs/ext2/xattr.c 	while (!IS_LAST_ENTRY(entry)) {
entry             239 fs/ext2/xattr.c 		if (!ext2_xattr_entry_valid(entry, end,
entry             244 fs/ext2/xattr.c 						 entry);
entry             250 fs/ext2/xattr.c 		entry = EXT2_XATTR_NEXT(entry);
entry             257 fs/ext2/xattr.c 	size = le32_to_cpu(entry->e_value_size);
entry             265 fs/ext2/xattr.c 		memcpy(buffer, bh->b_data + le16_to_cpu(entry->e_value_offs),
entry             292 fs/ext2/xattr.c 	struct ext2_xattr_entry *entry;
entry             323 fs/ext2/xattr.c 	entry = FIRST_ENTRY(bh);
entry             324 fs/ext2/xattr.c 	while (!IS_LAST_ENTRY(entry)) {
entry             325 fs/ext2/xattr.c 		if (!ext2_xattr_entry_valid(entry, end,
entry             328 fs/ext2/xattr.c 		entry = EXT2_XATTR_NEXT(entry);
entry             334 fs/ext2/xattr.c 	for (entry = FIRST_ENTRY(bh); !IS_LAST_ENTRY(entry);
entry             335 fs/ext2/xattr.c 	     entry = EXT2_XATTR_NEXT(entry)) {
entry             337 fs/ext2/xattr.c 			ext2_xattr_handler(entry->e_name_index);
entry             342 fs/ext2/xattr.c 			size_t size = prefix_len + entry->e_name_len + 1;
entry             351 fs/ext2/xattr.c 				memcpy(buffer, entry->e_name, entry->e_name_len);
entry             352 fs/ext2/xattr.c 				buffer += entry->e_name_len;
entry             990 fs/ext2/xattr.c 					 struct ext2_xattr_entry *entry)
entry             993 fs/ext2/xattr.c 	char *name = entry->e_name;
entry             996 fs/ext2/xattr.c 	for (n=0; n < entry->e_name_len; n++) {
entry            1002 fs/ext2/xattr.c 	if (entry->e_value_block == 0 && entry->e_value_size != 0) {
entry            1004 fs/ext2/xattr.c 			le16_to_cpu(entry->e_value_offs));
entry            1005 fs/ext2/xattr.c 		for (n = (le32_to_cpu(entry->e_value_size) +
entry            1012 fs/ext2/xattr.c 	entry->e_hash = cpu_to_le32(hash);
entry            1026 fs/ext2/xattr.c 			      struct ext2_xattr_entry *entry)
entry            1031 fs/ext2/xattr.c 	ext2_xattr_hash_entry(header, entry);
entry              51 fs/ext2/xattr.h #define EXT2_XATTR_NEXT(entry) \
entry              53 fs/ext2/xattr.h 	  (char *)(entry) + EXT2_XATTR_LEN((entry)->e_name_len)) )
entry              41 fs/ext4/acl.c  		ext4_acl_entry *entry =
entry              45 fs/ext4/acl.c  		acl->a_entries[n].e_tag  = le16_to_cpu(entry->e_tag);
entry              46 fs/ext4/acl.c  		acl->a_entries[n].e_perm = le16_to_cpu(entry->e_perm);
entry              63 fs/ext4/acl.c  					  le32_to_cpu(entry->e_id));
entry              71 fs/ext4/acl.c  					  le32_to_cpu(entry->e_id));
entry             106 fs/ext4/acl.c  		ext4_acl_entry *entry = (ext4_acl_entry *)e;
entry             107 fs/ext4/acl.c  		entry->e_tag  = cpu_to_le16(acl_e->e_tag);
entry             108 fs/ext4/acl.c  		entry->e_perm = cpu_to_le16(acl_e->e_perm);
entry             111 fs/ext4/acl.c  			entry->e_id = cpu_to_le32(
entry             116 fs/ext4/acl.c  			entry->e_id = cpu_to_le32(
entry              55 fs/ext4/block_validity.c 	struct ext4_system_zone	*entry, *n;
entry              57 fs/ext4/block_validity.c 	rbtree_postorder_for_each_entry_safe(entry, n,
entry              59 fs/ext4/block_validity.c 		kmem_cache_free(ext4_system_zone_cachep, entry);
entry              71 fs/ext4/block_validity.c 	struct ext4_system_zone *new_entry = NULL, *entry;
entry              77 fs/ext4/block_validity.c 		entry = rb_entry(parent, struct ext4_system_zone, node);
entry              78 fs/ext4/block_validity.c 		if (start_blk < entry->start_blk)
entry              80 fs/ext4/block_validity.c 		else if (start_blk >= (entry->start_blk + entry->count))
entry              83 fs/ext4/block_validity.c 			if (start_blk + count > (entry->start_blk +
entry              84 fs/ext4/block_validity.c 						 entry->count))
entry              85 fs/ext4/block_validity.c 				entry->count = (start_blk + count -
entry              86 fs/ext4/block_validity.c 						entry->start_blk);
entry             110 fs/ext4/block_validity.c 		entry = rb_entry(node, struct ext4_system_zone, node);
entry             111 fs/ext4/block_validity.c 		if (can_merge(entry, new_entry)) {
entry             112 fs/ext4/block_validity.c 			new_entry->start_blk = entry->start_blk;
entry             113 fs/ext4/block_validity.c 			new_entry->count += entry->count;
entry             115 fs/ext4/block_validity.c 			kmem_cache_free(ext4_system_zone_cachep, entry);
entry             122 fs/ext4/block_validity.c 		entry = rb_entry(node, struct ext4_system_zone, node);
entry             123 fs/ext4/block_validity.c 		if (can_merge(new_entry, entry)) {
entry             124 fs/ext4/block_validity.c 			new_entry->count += entry->count;
entry             126 fs/ext4/block_validity.c 			kmem_cache_free(ext4_system_zone_cachep, entry);
entry             135 fs/ext4/block_validity.c 	struct ext4_system_zone *entry;
entry             141 fs/ext4/block_validity.c 		entry = rb_entry(node, struct ext4_system_zone, node);
entry             143 fs/ext4/block_validity.c 		       entry->start_blk, entry->start_blk + entry->count - 1);
entry             160 fs/ext4/block_validity.c 	struct ext4_system_zone *entry;
entry             175 fs/ext4/block_validity.c 		entry = rb_entry(n, struct ext4_system_zone, node);
entry             176 fs/ext4/block_validity.c 		if (start_blk + count - 1 < entry->start_blk)
entry             178 fs/ext4/block_validity.c 		else if (start_blk >= (entry->start_blk + entry->count))
entry            3218 fs/ext4/ext4.h 			   const struct qstr *entry, bool quick);
entry              33 fs/ext4/inline.c 	struct ext4_xattr_entry *entry;
entry              54 fs/ext4/inline.c 	entry = IFIRST(header);
entry              57 fs/ext4/inline.c 	for (; !IS_LAST_ENTRY(entry); entry = EXT4_XATTR_NEXT(entry)) {
entry              58 fs/ext4/inline.c 		if (!entry->e_value_inum && entry->e_value_size) {
entry              59 fs/ext4/inline.c 			size_t offs = le16_to_cpu(entry->e_value_offs);
entry              65 fs/ext4/inline.c 		((void *)entry - (void *)IFIRST(header)) - sizeof(__u32);
entry              68 fs/ext4/inline.c 		entry = (struct ext4_xattr_entry *)
entry              71 fs/ext4/inline.c 		free += EXT4_XATTR_SIZE(le32_to_cpu(entry->e_value_size));
entry             168 fs/ext4/inline.c 	struct ext4_xattr_entry *entry;
entry             191 fs/ext4/inline.c 	entry = (struct ext4_xattr_entry *)((void *)raw_inode +
entry             194 fs/ext4/inline.c 		    (unsigned int)le32_to_cpu(entry->e_value_size));
entry             197 fs/ext4/inline.c 	       (void *)IFIRST(header) + le16_to_cpu(entry->e_value_offs), len);
entry             213 fs/ext4/inline.c 	struct ext4_xattr_entry *entry;
entry             242 fs/ext4/inline.c 	entry = (struct ext4_xattr_entry *)((void *)raw_inode +
entry             245 fs/ext4/inline.c 	memcpy((void *)IFIRST(header) + le16_to_cpu(entry->e_value_offs) + pos,
entry            1058 fs/ext4/inline.c 	struct ext4_xattr_entry *entry;
entry            1064 fs/ext4/inline.c 	entry = (struct ext4_xattr_entry *)((void *)ext4_raw_inode(iloc) +
entry            1067 fs/ext4/inline.c 	return (void *)IFIRST(header) + le16_to_cpu(entry->e_value_offs);
entry            2799 fs/ext4/mballoc.c 				    struct ext4_free_data *entry)
entry            2806 fs/ext4/mballoc.c 		 entry->efd_count, entry->efd_group, entry);
entry            2808 fs/ext4/mballoc.c 	err = ext4_mb_load_buddy(sb, entry->efd_group, &e4b);
entry            2813 fs/ext4/mballoc.c 	EXT4_SB(sb)->s_mb_free_pending -= entry->efd_count;
entry            2818 fs/ext4/mballoc.c 	count += entry->efd_count;
entry            2820 fs/ext4/mballoc.c 	ext4_lock_group(sb, entry->efd_group);
entry            2822 fs/ext4/mballoc.c 	rb_erase(&entry->efd_node, &(db->bb_free_root));
entry            2823 fs/ext4/mballoc.c 	mb_free_blocks(NULL, &e4b, entry->efd_start_cluster, entry->efd_count);
entry            2841 fs/ext4/mballoc.c 	ext4_unlock_group(sb, entry->efd_group);
entry            2842 fs/ext4/mballoc.c 	kmem_cache_free(ext4_free_data_cachep, entry);
entry            2855 fs/ext4/mballoc.c 	struct ext4_free_data *entry, *tmp;
entry            2864 fs/ext4/mballoc.c 	list_for_each_entry(entry, &sbi->s_freed_data_list, efd_list) {
entry            2865 fs/ext4/mballoc.c 		if (entry->efd_tid != commit_tid)
entry            2867 fs/ext4/mballoc.c 		cut_pos = &entry->efd_list;
entry            2875 fs/ext4/mballoc.c 		list_for_each_entry(entry, &freed_data_list, efd_list) {
entry            2876 fs/ext4/mballoc.c 			err = ext4_issue_discard(sb, entry->efd_group,
entry            2877 fs/ext4/mballoc.c 						 entry->efd_start_cluster,
entry            2878 fs/ext4/mballoc.c 						 entry->efd_count,
entry            2883 fs/ext4/mballoc.c 					 " with %d", entry->efd_group,
entry            2884 fs/ext4/mballoc.c 					 entry->efd_start_cluster,
entry            2885 fs/ext4/mballoc.c 					 entry->efd_count, err);
entry            2896 fs/ext4/mballoc.c 	list_for_each_entry_safe(entry, tmp, &freed_data_list, efd_list)
entry            2897 fs/ext4/mballoc.c 		ext4_free_data_in_buddy(sb, entry);
entry            3518 fs/ext4/mballoc.c 	struct ext4_free_data *entry;
entry            3524 fs/ext4/mballoc.c 		entry = rb_entry(n, struct ext4_free_data, efd_node);
entry            3525 fs/ext4/mballoc.c 		ext4_set_bits(bitmap, entry->efd_start_cluster, entry->efd_count);
entry            4622 fs/ext4/mballoc.c 					struct ext4_free_data *entry,
entry            4626 fs/ext4/mballoc.c 	if ((entry->efd_tid != new_entry->efd_tid) ||
entry            4627 fs/ext4/mballoc.c 	    (entry->efd_group != new_entry->efd_group))
entry            4629 fs/ext4/mballoc.c 	if (entry->efd_start_cluster + entry->efd_count ==
entry            4631 fs/ext4/mballoc.c 		new_entry->efd_start_cluster = entry->efd_start_cluster;
entry            4632 fs/ext4/mballoc.c 		new_entry->efd_count += entry->efd_count;
entry            4634 fs/ext4/mballoc.c 		   entry->efd_start_cluster) {
entry            4635 fs/ext4/mballoc.c 		new_entry->efd_count += entry->efd_count;
entry            4639 fs/ext4/mballoc.c 	list_del(&entry->efd_list);
entry            4641 fs/ext4/mballoc.c 	rb_erase(&entry->efd_node, entry_rb_root);
entry            4642 fs/ext4/mballoc.c 	kmem_cache_free(ext4_free_data_cachep, entry);
entry            4652 fs/ext4/mballoc.c 	struct ext4_free_data *entry;
entry            4677 fs/ext4/mballoc.c 		entry = rb_entry(parent, struct ext4_free_data, efd_node);
entry            4678 fs/ext4/mballoc.c 		if (cluster < entry->efd_start_cluster)
entry            4680 fs/ext4/mballoc.c 		else if (cluster >= (entry->efd_start_cluster + entry->efd_count))
entry            4697 fs/ext4/mballoc.c 		entry = rb_entry(node, struct ext4_free_data, efd_node);
entry            4698 fs/ext4/mballoc.c 		ext4_try_merge_freed_extent(sbi, entry, new_entry,
entry            4704 fs/ext4/mballoc.c 		entry = rb_entry(node, struct ext4_free_data, efd_node);
entry            4705 fs/ext4/mballoc.c 		ext4_try_merge_freed_extent(sbi, entry, new_entry,
entry             261 fs/ext4/namei.c static inline ext4_lblk_t dx_get_block(struct dx_entry *entry);
entry             262 fs/ext4/namei.c static void dx_set_block(struct dx_entry *entry, ext4_lblk_t value);
entry             263 fs/ext4/namei.c static inline unsigned dx_get_hash(struct dx_entry *entry);
entry             264 fs/ext4/namei.c static void dx_set_hash(struct dx_entry *entry, unsigned value);
entry             532 fs/ext4/namei.c static inline ext4_lblk_t dx_get_block(struct dx_entry *entry)
entry             534 fs/ext4/namei.c 	return le32_to_cpu(entry->block) & 0x0fffffff;
entry             537 fs/ext4/namei.c static inline void dx_set_block(struct dx_entry *entry, ext4_lblk_t value)
entry             539 fs/ext4/namei.c 	entry->block = cpu_to_le32(value);
entry             542 fs/ext4/namei.c static inline unsigned dx_get_hash(struct dx_entry *entry)
entry             544 fs/ext4/namei.c 	return le32_to_cpu(entry->hash);
entry             547 fs/ext4/namei.c static inline void dx_set_hash(struct dx_entry *entry, unsigned value)
entry             549 fs/ext4/namei.c 	entry->hash = cpu_to_le32(value);
entry            1283 fs/ext4/namei.c 		    const struct qstr *entry, bool quick)
entry            1290 fs/ext4/namei.c 		ret = utf8_strncasecmp_folded(um, name, entry);
entry            1292 fs/ext4/namei.c 		ret = utf8_strncasecmp(um, name, entry);
entry            1301 fs/ext4/namei.c 		if (name->len != entry->len)
entry            1304 fs/ext4/namei.c 		return !!memcmp(name->name, entry->name, name->len);
entry            1348 fs/ext4/namei.c 	const struct qstr entry = {.name = de->name, .len = de->name_len};
entry            1365 fs/ext4/namei.c 			return !ext4_ci_compare(parent, &cf, &entry, true);
entry            1367 fs/ext4/namei.c 		return !ext4_ci_compare(parent, fname->usr_fname, &entry,
entry             183 fs/ext4/xattr.c ext4_xattr_check_entries(struct ext4_xattr_entry *entry, void *end,
entry             186 fs/ext4/xattr.c 	struct ext4_xattr_entry *e = entry;
entry             199 fs/ext4/xattr.c 	while (!IS_LAST_ENTRY(entry)) {
entry             200 fs/ext4/xattr.c 		u32 size = le32_to_cpu(entry->e_value_size);
entry             205 fs/ext4/xattr.c 		if (size != 0 && entry->e_value_inum == 0) {
entry             206 fs/ext4/xattr.c 			u16 offs = le16_to_cpu(entry->e_value_offs);
entry             223 fs/ext4/xattr.c 		entry = EXT4_XATTR_NEXT(entry);
entry             284 fs/ext4/xattr.c 	struct ext4_xattr_entry *entry, *next;
entry             291 fs/ext4/xattr.c 	for (entry = *pentry; !IS_LAST_ENTRY(entry); entry = next) {
entry             292 fs/ext4/xattr.c 		next = EXT4_XATTR_NEXT(entry);
entry             297 fs/ext4/xattr.c 		cmp = name_index - entry->e_name_index;
entry             299 fs/ext4/xattr.c 			cmp = name_len - entry->e_name_len;
entry             301 fs/ext4/xattr.c 			cmp = memcmp(name, entry->e_name, name_len);
entry             305 fs/ext4/xattr.c 	*pentry = entry;
entry             439 fs/ext4/xattr.c 			       struct ext4_xattr_entry *entry, void *buffer,
entry             449 fs/ext4/xattr.c 	if (entry) {
entry             454 fs/ext4/xattr.c 		e_hash = ext4_xattr_hash_entry(entry->e_name, entry->e_name_len,
entry             456 fs/ext4/xattr.c 		if (e_hash != entry->e_hash)
entry             466 fs/ext4/xattr.c ext4_xattr_inode_get(struct inode *inode, struct ext4_xattr_entry *entry,
entry             473 fs/ext4/xattr.c 	err = ext4_xattr_inode_iget(inode, le32_to_cpu(entry->e_value_inum),
entry             474 fs/ext4/xattr.c 				    le32_to_cpu(entry->e_hash), &ea_inode);
entry             493 fs/ext4/xattr.c 		err = ext4_xattr_inode_verify_hashes(ea_inode, entry, buffer,
entry             516 fs/ext4/xattr.c 	struct ext4_xattr_entry *entry;
entry             538 fs/ext4/xattr.c 	entry = BFIRST(bh);
entry             540 fs/ext4/xattr.c 	error = xattr_find_entry(inode, &entry, end, name_index, name, 1);
entry             543 fs/ext4/xattr.c 	size = le32_to_cpu(entry->e_value_size);
entry             550 fs/ext4/xattr.c 		if (entry->e_value_inum) {
entry             551 fs/ext4/xattr.c 			error = ext4_xattr_inode_get(inode, entry, buffer,
entry             556 fs/ext4/xattr.c 			u16 offset = le16_to_cpu(entry->e_value_offs);
entry             576 fs/ext4/xattr.c 	struct ext4_xattr_entry *entry;
entry             594 fs/ext4/xattr.c 	entry = IFIRST(header);
entry             595 fs/ext4/xattr.c 	error = xattr_find_entry(inode, &entry, end, name_index, name, 0);
entry             598 fs/ext4/xattr.c 	size = le32_to_cpu(entry->e_value_size);
entry             605 fs/ext4/xattr.c 		if (entry->e_value_inum) {
entry             606 fs/ext4/xattr.c 			error = ext4_xattr_inode_get(inode, entry, buffer,
entry             611 fs/ext4/xattr.c 			u16 offset = le16_to_cpu(entry->e_value_offs);
entry             659 fs/ext4/xattr.c ext4_xattr_list_entries(struct dentry *dentry, struct ext4_xattr_entry *entry,
entry             664 fs/ext4/xattr.c 	for (; !IS_LAST_ENTRY(entry); entry = EXT4_XATTR_NEXT(entry)) {
entry             666 fs/ext4/xattr.c 			ext4_xattr_handler(entry->e_name_index);
entry             671 fs/ext4/xattr.c 			size_t size = prefix_len + entry->e_name_len + 1;
entry             678 fs/ext4/xattr.c 				memcpy(buffer, entry->e_name, entry->e_name_len);
entry             679 fs/ext4/xattr.c 				buffer += entry->e_name_len;
entry             804 fs/ext4/xattr.c 	struct ext4_xattr_entry *entry;
entry             822 fs/ext4/xattr.c 		for (entry = IFIRST(header); !IS_LAST_ENTRY(entry);
entry             823 fs/ext4/xattr.c 		     entry = EXT4_XATTR_NEXT(entry))
entry             824 fs/ext4/xattr.c 			if (entry->e_value_inum)
entry             840 fs/ext4/xattr.c 		for (entry = BFIRST(bh); !IS_LAST_ENTRY(entry);
entry             841 fs/ext4/xattr.c 		     entry = EXT4_XATTR_NEXT(entry))
entry             842 fs/ext4/xattr.c 			if (entry->e_value_inum)
entry             960 fs/ext4/xattr.c 		struct ext4_xattr_entry *entry = BFIRST(block_bh);
entry             962 fs/ext4/xattr.c 		for (; !IS_LAST_ENTRY(entry); entry = EXT4_XATTR_NEXT(entry))
entry             963 fs/ext4/xattr.c 			if (entry->e_value_inum)
entry            1100 fs/ext4/xattr.c 	struct ext4_xattr_entry *entry;
entry            1105 fs/ext4/xattr.c 	for (entry = first; !IS_LAST_ENTRY(entry);
entry            1106 fs/ext4/xattr.c 	     entry = EXT4_XATTR_NEXT(entry)) {
entry            1107 fs/ext4/xattr.c 		if (!entry->e_value_inum)
entry            1109 fs/ext4/xattr.c 		ea_ino = le32_to_cpu(entry->e_value_inum);
entry            1111 fs/ext4/xattr.c 					    le32_to_cpu(entry->e_hash),
entry            1127 fs/ext4/xattr.c 	failed_entry = entry;
entry            1129 fs/ext4/xattr.c 	for (entry = first; entry != failed_entry;
entry            1130 fs/ext4/xattr.c 	     entry = EXT4_XATTR_NEXT(entry)) {
entry            1131 fs/ext4/xattr.c 		if (!entry->e_value_inum)
entry            1133 fs/ext4/xattr.c 		ea_ino = le32_to_cpu(entry->e_value_inum);
entry            1135 fs/ext4/xattr.c 					    le32_to_cpu(entry->e_hash),
entry            1160 fs/ext4/xattr.c 	struct ext4_xattr_entry *entry;
entry            1169 fs/ext4/xattr.c 	for (entry = first; !IS_LAST_ENTRY(entry);
entry            1170 fs/ext4/xattr.c 	     entry = EXT4_XATTR_NEXT(entry)) {
entry            1171 fs/ext4/xattr.c 		if (!entry->e_value_inum)
entry            1173 fs/ext4/xattr.c 		ea_ino = le32_to_cpu(entry->e_value_inum);
entry            1175 fs/ext4/xattr.c 					    le32_to_cpu(entry->e_hash),
entry            1205 fs/ext4/xattr.c 					      le32_to_cpu(entry->e_value_size));
entry            1213 fs/ext4/xattr.c 		entry->e_value_inum = 0;
entry            1214 fs/ext4/xattr.c 		entry->e_value_size = 0;
entry            2521 fs/ext4/xattr.c static void ext4_xattr_shift_entries(struct ext4_xattr_entry *entry,
entry            2525 fs/ext4/xattr.c 	struct ext4_xattr_entry *last = entry;
entry            2548 fs/ext4/xattr.c 				    struct ext4_xattr_entry *entry)
entry            2553 fs/ext4/xattr.c 	size_t value_size = le32_to_cpu(entry->e_value_size);
entry            2557 fs/ext4/xattr.c 		.name_index = entry->e_name_index,
entry            2558 fs/ext4/xattr.c 		.in_inode = !!entry->e_value_inum,
entry            2566 fs/ext4/xattr.c 	b_entry_name = kmalloc(entry->e_name_len + 1, GFP_NOFS);
entry            2578 fs/ext4/xattr.c 	if (entry->e_value_inum) {
entry            2579 fs/ext4/xattr.c 		error = ext4_xattr_inode_get(inode, entry, buffer, value_size);
entry            2583 fs/ext4/xattr.c 		size_t value_offs = le16_to_cpu(entry->e_value_offs);
entry            2587 fs/ext4/xattr.c 	memcpy(b_entry_name, entry->e_name, entry->e_name_len);
entry            2588 fs/ext4/xattr.c 	b_entry_name[entry->e_name_len] = '\0';
entry            2635 fs/ext4/xattr.c 	struct ext4_xattr_entry *entry;
entry            2643 fs/ext4/xattr.c 		entry = NULL;
entry            2663 fs/ext4/xattr.c 					entry = last;
entry            2669 fs/ext4/xattr.c 		if (entry == NULL) {
entry            2672 fs/ext4/xattr.c 			entry = small_entry;
entry            2675 fs/ext4/xattr.c 		entry_size = EXT4_XATTR_LEN(entry->e_name_len);
entry            2677 fs/ext4/xattr.c 		if (!entry->e_value_inum)
entry            2679 fs/ext4/xattr.c 					      le32_to_cpu(entry->e_value_size));
entry            2681 fs/ext4/xattr.c 						 entry);
entry            2861 fs/ext4/xattr.c 	struct ext4_xattr_entry *entry;
entry            2915 fs/ext4/xattr.c 			for (entry = BFIRST(bh); !IS_LAST_ENTRY(entry);
entry            2916 fs/ext4/xattr.c 			     entry = EXT4_XATTR_NEXT(entry)) {
entry            2917 fs/ext4/xattr.c 				if (!entry->e_value_inum)
entry            2920 fs/ext4/xattr.c 					      le32_to_cpu(entry->e_value_inum),
entry            2921 fs/ext4/xattr.c 					      le32_to_cpu(entry->e_hash),
entry            2926 fs/ext4/xattr.c 					      le32_to_cpu(entry->e_value_size));
entry              60 fs/ext4/xattr.h #define EXT4_XATTR_NEXT(entry) \
entry              62 fs/ext4/xattr.h 	 (char *)(entry) + EXT4_XATTR_LEN((entry)->e_name_len)))
entry              94 fs/ext4/xattr.h #define IS_LAST_ENTRY(entry) (*(__u32 *)(entry) == 0)
entry              50 fs/f2fs/acl.c  	struct f2fs_acl_entry *entry = (struct f2fs_acl_entry *)(hdr + 1);
entry              71 fs/f2fs/acl.c  		if ((char *)entry > end)
entry              74 fs/f2fs/acl.c  		acl->a_entries[i].e_tag  = le16_to_cpu(entry->e_tag);
entry              75 fs/f2fs/acl.c  		acl->a_entries[i].e_perm = le16_to_cpu(entry->e_perm);
entry              82 fs/f2fs/acl.c  			entry = (struct f2fs_acl_entry *)((char *)entry +
entry              89 fs/f2fs/acl.c  						le32_to_cpu(entry->e_id));
entry              90 fs/f2fs/acl.c  			entry = (struct f2fs_acl_entry *)((char *)entry +
entry              96 fs/f2fs/acl.c  						le32_to_cpu(entry->e_id));
entry              97 fs/f2fs/acl.c  			entry = (struct f2fs_acl_entry *)((char *)entry +
entry             104 fs/f2fs/acl.c  	if ((char *)entry != end)
entry             116 fs/f2fs/acl.c  	struct f2fs_acl_entry *entry;
entry             126 fs/f2fs/acl.c  	entry = (struct f2fs_acl_entry *)(f2fs_acl + 1);
entry             130 fs/f2fs/acl.c  		entry->e_tag  = cpu_to_le16(acl->a_entries[i].e_tag);
entry             131 fs/f2fs/acl.c  		entry->e_perm = cpu_to_le16(acl->a_entries[i].e_perm);
entry             135 fs/f2fs/acl.c  			entry->e_id = cpu_to_le32(
entry             138 fs/f2fs/acl.c  			entry = (struct f2fs_acl_entry *)((char *)entry +
entry             142 fs/f2fs/acl.c  			entry->e_id = cpu_to_le32(
entry             145 fs/f2fs/acl.c  			entry = (struct f2fs_acl_entry *)((char *)entry +
entry             152 fs/f2fs/acl.c  			entry = (struct f2fs_acl_entry *)((char *)entry +
entry             115 fs/f2fs/dir.c  				const struct qstr *entry, bool quick)
entry             122 fs/f2fs/dir.c  		ret = utf8_strncasecmp_folded(um, name, entry);
entry             124 fs/f2fs/dir.c  		ret = utf8_strncasecmp(um, name, entry);
entry             133 fs/f2fs/dir.c  		if (name->len != entry->len)
entry             136 fs/f2fs/dir.c  		return !!memcmp(name->name, entry->name, name->len);
entry             177 fs/f2fs/dir.c  	struct qstr entry;
entry             184 fs/f2fs/dir.c  	entry.name = d->filename[bit_pos];
entry             185 fs/f2fs/dir.c  	entry.len = de->name_len;
entry             191 fs/f2fs/dir.c  			return !f2fs_ci_compare(parent, &cf, &entry, true);
entry             193 fs/f2fs/dir.c  		return !f2fs_ci_compare(parent, fname->usr_fname, &entry,
entry            2207 fs/f2fs/f2fs.h 	void *entry;
entry            2209 fs/f2fs/f2fs.h 	entry = kmem_cache_alloc(cachep, flags);
entry            2210 fs/f2fs/f2fs.h 	if (!entry)
entry            2211 fs/f2fs/f2fs.h 		entry = kmem_cache_alloc(cachep, flags | __GFP_NOFAIL);
entry            2212 fs/f2fs/f2fs.h 	return entry;
entry            2964 fs/f2fs/f2fs.h 			   const struct qstr *entry,
entry             509 fs/f2fs/gc.c   	struct f2fs_summary *entry;
entry             519 fs/f2fs/gc.c   	entry = sum;
entry             524 fs/f2fs/gc.c   	for (off = 0; off < sbi->blocks_per_seg; off++, entry++) {
entry             525 fs/f2fs/gc.c   		nid_t nid = le32_to_cpu(entry->nid);
entry             996 fs/f2fs/gc.c   	struct f2fs_summary *entry;
entry            1005 fs/f2fs/gc.c   	entry = sum;
entry            1007 fs/f2fs/gc.c   	for (off = 0; off < sbi->blocks_per_seg; off++, entry++) {
entry            1013 fs/f2fs/gc.c   		nid_t nid = le32_to_cpu(entry->nid);
entry            1034 fs/f2fs/gc.c   		if (!is_alive(sbi, entry, &dni, start_addr + off, &nofs))
entry            1042 fs/f2fs/gc.c   		ofs_in_node = le16_to_cpu(entry->ofs_in_node);
entry              59 fs/f2fs/recovery.c 	struct fsync_inode_entry *entry;
entry              61 fs/f2fs/recovery.c 	list_for_each_entry(entry, head, list)
entry              62 fs/f2fs/recovery.c 		if (entry->inode->i_ino == ino)
entry              63 fs/f2fs/recovery.c 			return entry;
entry              72 fs/f2fs/recovery.c 	struct fsync_inode_entry *entry;
entry              89 fs/f2fs/recovery.c 	entry = f2fs_kmem_cache_alloc(fsync_entry_slab, GFP_F2FS_ZERO);
entry              90 fs/f2fs/recovery.c 	entry->inode = inode;
entry              91 fs/f2fs/recovery.c 	list_add_tail(&entry->list, head);
entry              93 fs/f2fs/recovery.c 	return entry;
entry              99 fs/f2fs/recovery.c static void del_fsync_inode(struct fsync_inode_entry *entry, int drop)
entry             103 fs/f2fs/recovery.c 		f2fs_inode_synced(entry->inode);
entry             105 fs/f2fs/recovery.c 	iput(entry->inode);
entry             106 fs/f2fs/recovery.c 	list_del(&entry->list);
entry             107 fs/f2fs/recovery.c 	kmem_cache_free(fsync_entry_slab, entry);
entry             119 fs/f2fs/recovery.c 	struct fsync_inode_entry *entry;
entry             123 fs/f2fs/recovery.c 	entry = get_fsync_inode(dir_list, pino);
entry             124 fs/f2fs/recovery.c 	if (!entry) {
entry             125 fs/f2fs/recovery.c 		entry = add_fsync_inode(F2FS_I_SB(inode), dir_list,
entry             127 fs/f2fs/recovery.c 		if (IS_ERR(entry)) {
entry             128 fs/f2fs/recovery.c 			dir = ERR_CAST(entry);
entry             129 fs/f2fs/recovery.c 			err = PTR_ERR(entry);
entry             134 fs/f2fs/recovery.c 	dir = entry->inode;
entry             315 fs/f2fs/recovery.c 		struct fsync_inode_entry *entry;
entry             334 fs/f2fs/recovery.c 		entry = get_fsync_inode(head, ino_of_node(page));
entry             335 fs/f2fs/recovery.c 		if (!entry) {
entry             352 fs/f2fs/recovery.c 			entry = add_fsync_inode(sbi, head, ino_of_node(page),
entry             354 fs/f2fs/recovery.c 			if (IS_ERR(entry)) {
entry             355 fs/f2fs/recovery.c 				err = PTR_ERR(entry);
entry             364 fs/f2fs/recovery.c 		entry->blkaddr = blkaddr;
entry             367 fs/f2fs/recovery.c 			entry->last_dentry = blkaddr;
entry             391 fs/f2fs/recovery.c 	struct fsync_inode_entry *entry, *tmp;
entry             393 fs/f2fs/recovery.c 	list_for_each_entry_safe(entry, tmp, head, list)
entry             394 fs/f2fs/recovery.c 		del_fsync_inode(entry, drop);
entry             660 fs/f2fs/recovery.c 		struct fsync_inode_entry *entry;
entry             678 fs/f2fs/recovery.c 		entry = get_fsync_inode(inode_list, ino_of_node(page));
entry             679 fs/f2fs/recovery.c 		if (!entry)
entry             687 fs/f2fs/recovery.c 			err = recover_inode(entry->inode, page);
entry             693 fs/f2fs/recovery.c 		if (entry->last_dentry == blkaddr) {
entry             694 fs/f2fs/recovery.c 			err = recover_dentry(entry->inode, page, dir_list);
entry             700 fs/f2fs/recovery.c 		err = do_recover_data(sbi, entry->inode, page);
entry             706 fs/f2fs/recovery.c 		if (entry->blkaddr == blkaddr)
entry             707 fs/f2fs/recovery.c 			list_move_tail(&entry->list, tmp_inode_list);
entry            1901 fs/f2fs/segment.c static void release_discard_addr(struct discard_entry *entry)
entry            1903 fs/f2fs/segment.c 	list_del(&entry->list);
entry            1904 fs/f2fs/segment.c 	kmem_cache_free(discard_entry_slab, entry);
entry            1910 fs/f2fs/segment.c 	struct discard_entry *entry, *this;
entry            1913 fs/f2fs/segment.c 	list_for_each_entry_safe(entry, this, head, list)
entry            1914 fs/f2fs/segment.c 		release_discard_addr(entry);
entry            1936 fs/f2fs/segment.c 	struct discard_entry *entry, *this;
entry            1996 fs/f2fs/segment.c 	list_for_each_entry_safe(entry, this, head, list) {
entry            1998 fs/f2fs/segment.c 		bool is_valid = test_bit_le(0, entry->discard_map);
entry            2002 fs/f2fs/segment.c 			next_pos = find_next_zero_bit_le(entry->discard_map,
entry            2010 fs/f2fs/segment.c 			f2fs_issue_discard(sbi, entry->start_blkaddr + cur_pos,
entry            2014 fs/f2fs/segment.c 			next_pos = find_next_bit_le(entry->discard_map,
entry            2024 fs/f2fs/segment.c 		release_discard_addr(entry);
entry             210 fs/f2fs/xattr.c 	struct f2fs_xattr_entry *entry;
entry             212 fs/f2fs/xattr.c 	list_for_each_xattr(entry, base_addr) {
entry             213 fs/f2fs/xattr.c 		if ((void *)(entry) + sizeof(__u32) > last_base_addr ||
entry             214 fs/f2fs/xattr.c 			(void *)XATTR_NEXT_ENTRY(entry) > last_base_addr)
entry             217 fs/f2fs/xattr.c 		if (entry->e_name_index != index)
entry             219 fs/f2fs/xattr.c 		if (entry->e_name_len != len)
entry             221 fs/f2fs/xattr.c 		if (!memcmp(entry->e_name, name, len))
entry             224 fs/f2fs/xattr.c 	return entry;
entry             231 fs/f2fs/xattr.c 	struct f2fs_xattr_entry *entry;
entry             235 fs/f2fs/xattr.c 	list_for_each_xattr(entry, base_addr) {
entry             236 fs/f2fs/xattr.c 		if ((void *)entry + sizeof(__u32) > max_addr ||
entry             237 fs/f2fs/xattr.c 			(void *)XATTR_NEXT_ENTRY(entry) > max_addr) {
entry             238 fs/f2fs/xattr.c 			*last_addr = entry;
entry             241 fs/f2fs/xattr.c 		if (entry->e_name_index != index)
entry             243 fs/f2fs/xattr.c 		if (entry->e_name_len != len)
entry             245 fs/f2fs/xattr.c 		if (!memcmp(entry->e_name, name, len))
entry             250 fs/f2fs/xattr.c 	if (IS_XATTR_LAST_ENTRY(entry) &&
entry             251 fs/f2fs/xattr.c 		(void *)entry + sizeof(__u32) > max_addr) {
entry             252 fs/f2fs/xattr.c 		*last_addr = entry;
entry             255 fs/f2fs/xattr.c 	return entry;
entry             497 fs/f2fs/xattr.c 	struct f2fs_xattr_entry *entry = NULL;
entry             512 fs/f2fs/xattr.c 				&entry, &base_addr, &base_size);
entry             517 fs/f2fs/xattr.c 	size = le16_to_cpu(entry->e_value_size);
entry             525 fs/f2fs/xattr.c 		char *pval = entry->e_name + entry->e_name_len;
entry             543 fs/f2fs/xattr.c 	struct f2fs_xattr_entry *entry;
entry             556 fs/f2fs/xattr.c 	list_for_each_xattr(entry, base_addr) {
entry             558 fs/f2fs/xattr.c 			f2fs_xattr_handler(entry->e_name_index);
entry             563 fs/f2fs/xattr.c 		if ((void *)(entry) + sizeof(__u32) > last_base_addr ||
entry             564 fs/f2fs/xattr.c 			(void *)XATTR_NEXT_ENTRY(entry) > last_base_addr) {
entry             577 fs/f2fs/xattr.c 		size = prefix_len + entry->e_name_len + 1;
entry             585 fs/f2fs/xattr.c 			memcpy(buffer, entry->e_name, entry->e_name_len);
entry             586 fs/f2fs/xattr.c 			buffer += entry->e_name_len;
entry             597 fs/f2fs/xattr.c static bool f2fs_xattr_value_same(struct f2fs_xattr_entry *entry,
entry             600 fs/f2fs/xattr.c 	void *pval = entry->e_name + entry->e_name_len;
entry             602 fs/f2fs/xattr.c 	return (le16_to_cpu(entry->e_value_size) == size) &&
entry              62 fs/f2fs/xattr.h #define ENTRY_SIZE(entry) (XATTR_ALIGN(sizeof(struct f2fs_xattr_entry) + \
entry              63 fs/f2fs/xattr.h 			(entry)->e_name_len + le16_to_cpu((entry)->e_value_size)))
entry              65 fs/f2fs/xattr.h #define XATTR_NEXT_ENTRY(entry)	((struct f2fs_xattr_entry *)((char *)(entry) +\
entry              66 fs/f2fs/xattr.h 			ENTRY_SIZE(entry)))
entry              68 fs/f2fs/xattr.h #define IS_XATTR_LAST_ENTRY(entry) (*(__u32 *)(entry) == 0)
entry              70 fs/f2fs/xattr.h #define list_for_each_xattr(entry, addr) \
entry              71 fs/f2fs/xattr.h 		for (entry = XATTR_FIRST_ENTRY(addr);\
entry              72 fs/f2fs/xattr.h 				!IS_XATTR_LAST_ENTRY(entry);\
entry              73 fs/f2fs/xattr.h 				entry = XATTR_NEXT_ENTRY(entry))
entry             342 fs/fat/fat.h   	int entry;
entry             356 fs/fat/fat.h   	fatent->entry = 0;
entry             362 fs/fat/fat.h   static inline void fatent_set_entry(struct fat_entry *fatent, int entry)
entry             364 fs/fat/fat.h   	fatent->entry = entry;
entry             379 fs/fat/fat.h   static inline bool fat_valid_entry(struct msdos_sb_info *sbi, int entry)
entry             381 fs/fat/fat.h   	return FAT_START_ENT <= entry && entry < sbi->max_cluster;
entry             386 fs/fat/fat.h   			int entry);
entry              22 fs/fat/fatent.c static void fat12_ent_blocknr(struct super_block *sb, int entry,
entry              26 fs/fat/fatent.c 	int bytes = entry + (entry >> 1);
entry              27 fs/fat/fatent.c 	WARN_ON(!fat_valid_entry(sbi, entry));
entry              32 fs/fat/fatent.c static void fat_ent_blocknr(struct super_block *sb, int entry,
entry              36 fs/fat/fatent.c 	int bytes = (entry << sbi->fatent_shift);
entry              37 fs/fat/fatent.c 	WARN_ON(!fat_valid_entry(sbi, entry));
entry             124 fs/fat/fatent.c 	if (fatent->entry & 1)
entry             162 fs/fat/fatent.c 	if (fatent->entry & 1) {
entry             197 fs/fat/fatent.c 	u8 *nextp = ent12_p[1] + 1 + (fatent->entry & 1);
entry             199 fs/fat/fatent.c 	fatent->entry++;
entry             229 fs/fat/fatent.c 	fatent->entry++;
entry             241 fs/fat/fatent.c 	fatent->entry++;
entry             347 fs/fat/fatent.c int fat_ent_read(struct inode *inode, struct fat_entry *fatent, int entry)
entry             355 fs/fat/fatent.c 	if (!fat_valid_entry(sbi, entry)) {
entry             357 fs/fat/fatent.c 		fat_fs_error(sb, "invalid access to FAT (entry 0x%08x)", entry);
entry             361 fs/fat/fatent.c 	fatent_set_entry(fatent, entry);
entry             362 fs/fat/fatent.c 	ops->ent_blocknr(sb, entry, &offset, &blocknr);
entry             428 fs/fat/fatent.c 		if (fatent->entry < sbi->max_cluster)
entry             442 fs/fat/fatent.c 	ops->ent_blocknr(sb, fatent->entry, &offset, &blocknr);
entry             488 fs/fat/fatent.c 		if (fatent.entry >= sbi->max_cluster)
entry             489 fs/fat/fatent.c 			fatent.entry = FAT_START_ENT;
entry             490 fs/fat/fatent.c 		fatent_set_entry(&fatent, fatent.entry);
entry             498 fs/fat/fatent.c 				int entry = fatent.entry;
entry             503 fs/fat/fatent.c 					ops->ent_put(&prev_ent, entry);
entry             507 fs/fat/fatent.c 				sbi->prev_free = entry;
entry             511 fs/fat/fatent.c 				cluster[idx_clus] = entry;
entry             583 fs/fat/fatent.c 			if (cluster != fatent.entry + 1) {
entry             584 fs/fat/fatent.c 				int nr_clus = fatent.entry - first_cl + 1;
entry             645 fs/fat/fatent.c 	ops->ent_blocknr(sb, fatent->entry, &offset, &blocknr);
entry             670 fs/fat/fatent.c 	while (fatent.entry < sbi->max_cluster) {
entry             736 fs/fat/fatent.c 	while (fatent.entry <= ent_end) {
entry             752 fs/fat/fatent.c 					u32 clus = fatent.entry - free;
entry             763 fs/fat/fatent.c 		} while (fat_ent_next(sbi, &fatent) && fatent.entry <= ent_end);
entry             779 fs/fat/fatent.c 		u32 clus = fatent.entry - free;
entry              59 fs/fs_pin.c    		if (likely(list_empty(&wait.entry)))
entry             334 fs/fuse/control.c 	list_for_each_entry(fc, &fuse_conn_list, entry) {
entry             368 fs/fuse/control.c 	list_for_each_entry(fc, &fuse_conn_list, entry)
entry            2105 fs/fuse/dev.c  		list_for_each_entry(fud, &fc->devices, entry) {
entry              28 fs/fuse/dir.c  static inline void __fuse_dentry_settime(struct dentry *entry, u64 time)
entry              30 fs/fuse/dir.c  	entry->d_fsdata = (void *) time;
entry              33 fs/fuse/dir.c  static inline u64 fuse_dentry_time(const struct dentry *entry)
entry              35 fs/fuse/dir.c  	return (u64)entry->d_fsdata;
entry              49 fs/fuse/dir.c  static inline u64 fuse_dentry_time(const struct dentry *entry)
entry              51 fs/fuse/dir.c  	return ((union fuse_dentry *) entry->d_fsdata)->time;
entry             102 fs/fuse/dir.c  void fuse_change_entry_timeout(struct dentry *entry, struct fuse_entry_out *o)
entry             104 fs/fuse/dir.c  	fuse_dentry_settime(entry,
entry             156 fs/fuse/dir.c  void fuse_invalidate_entry_cache(struct dentry *entry)
entry             158 fs/fuse/dir.c  	fuse_dentry_settime(entry, 0);
entry             165 fs/fuse/dir.c  static void fuse_invalidate_entry(struct dentry *entry)
entry             167 fs/fuse/dir.c  	d_invalidate(entry);
entry             168 fs/fuse/dir.c  	fuse_invalidate_entry_cache(entry);
entry             195 fs/fuse/dir.c  static int fuse_dentry_revalidate(struct dentry *entry, unsigned int flags)
entry             203 fs/fuse/dir.c  	inode = d_inode_rcu(entry);
entry             206 fs/fuse/dir.c  	else if (time_before64(fuse_dentry_time(entry), get_jiffies_64()) ||
entry             230 fs/fuse/dir.c  		parent = dget_parent(entry);
entry             232 fs/fuse/dir.c  				 &entry->d_name, &outarg);
entry             259 fs/fuse/dir.c  		fuse_change_entry_timeout(entry, &outarg);
entry             266 fs/fuse/dir.c  			parent = dget_parent(entry);
entry             379 fs/fuse/dir.c  static struct dentry *fuse_lookup(struct inode *dir, struct dentry *entry,
entry             390 fs/fuse/dir.c  	err = fuse_lookup_name(dir->i_sb, get_node_id(dir), &entry->d_name,
entry             404 fs/fuse/dir.c  	newent = d_splice_alias(inode, entry);
entry             409 fs/fuse/dir.c  	entry = newent ? newent : entry;
entry             411 fs/fuse/dir.c  		fuse_change_entry_timeout(entry, &outarg);
entry             413 fs/fuse/dir.c  		fuse_invalidate_entry_cache(entry);
entry             431 fs/fuse/dir.c  static int fuse_create_open(struct inode *dir, struct dentry *entry,
entry             473 fs/fuse/dir.c  	args.in_args[1].size = entry->d_name.len + 1;
entry             474 fs/fuse/dir.c  	args.in_args[1].value = entry->d_name.name;
entry             502 fs/fuse/dir.c  	d_instantiate(entry, inode);
entry             503 fs/fuse/dir.c  	fuse_change_entry_timeout(entry, &outentry);
entry             505 fs/fuse/dir.c  	err = finish_open(file, entry, generic_file_open);
entry             524 fs/fuse/dir.c  static int fuse_atomic_open(struct inode *dir, struct dentry *entry,
entry             532 fs/fuse/dir.c  	if (d_in_lookup(entry)) {
entry             533 fs/fuse/dir.c  		res = fuse_lookup(dir, entry, 0);
entry             538 fs/fuse/dir.c  			entry = res;
entry             541 fs/fuse/dir.c  	if (!(flags & O_CREAT) || d_really_is_positive(entry))
entry             550 fs/fuse/dir.c  	err = fuse_create_open(dir, entry, file, flags, mode);
entry             560 fs/fuse/dir.c  	err = fuse_mknod(dir, entry, mode, 0);
entry             571 fs/fuse/dir.c  			    struct inode *dir, struct dentry *entry,
entry             608 fs/fuse/dir.c  	d_drop(entry);
entry             609 fs/fuse/dir.c  	d = d_splice_alias(inode, entry);
entry             617 fs/fuse/dir.c  		fuse_change_entry_timeout(entry, &outarg);
entry             627 fs/fuse/dir.c  static int fuse_mknod(struct inode *dir, struct dentry *entry, umode_t mode,
entry             645 fs/fuse/dir.c  	args.in_args[1].size = entry->d_name.len + 1;
entry             646 fs/fuse/dir.c  	args.in_args[1].value = entry->d_name.name;
entry             647 fs/fuse/dir.c  	return create_new_entry(fc, &args, dir, entry, mode);
entry             650 fs/fuse/dir.c  static int fuse_create(struct inode *dir, struct dentry *entry, umode_t mode,
entry             653 fs/fuse/dir.c  	return fuse_mknod(dir, entry, mode, 0);
entry             656 fs/fuse/dir.c  static int fuse_mkdir(struct inode *dir, struct dentry *entry, umode_t mode)
entry             672 fs/fuse/dir.c  	args.in_args[1].size = entry->d_name.len + 1;
entry             673 fs/fuse/dir.c  	args.in_args[1].value = entry->d_name.name;
entry             674 fs/fuse/dir.c  	return create_new_entry(fc, &args, dir, entry, S_IFDIR);
entry             677 fs/fuse/dir.c  static int fuse_symlink(struct inode *dir, struct dentry *entry,
entry             686 fs/fuse/dir.c  	args.in_args[0].size = entry->d_name.len + 1;
entry             687 fs/fuse/dir.c  	args.in_args[0].value = entry->d_name.name;
entry             690 fs/fuse/dir.c  	return create_new_entry(fc, &args, dir, entry, S_IFLNK);
entry             701 fs/fuse/dir.c  static int fuse_unlink(struct inode *dir, struct dentry *entry)
entry             710 fs/fuse/dir.c  	args.in_args[0].size = entry->d_name.len + 1;
entry             711 fs/fuse/dir.c  	args.in_args[0].value = entry->d_name.name;
entry             714 fs/fuse/dir.c  		struct inode *inode = d_inode(entry);
entry             730 fs/fuse/dir.c  		fuse_invalidate_entry_cache(entry);
entry             733 fs/fuse/dir.c  		fuse_invalidate_entry(entry);
entry             737 fs/fuse/dir.c  static int fuse_rmdir(struct inode *dir, struct dentry *entry)
entry             746 fs/fuse/dir.c  	args.in_args[0].size = entry->d_name.len + 1;
entry             747 fs/fuse/dir.c  	args.in_args[0].value = entry->d_name.name;
entry             750 fs/fuse/dir.c  		clear_nlink(d_inode(entry));
entry             752 fs/fuse/dir.c  		fuse_invalidate_entry_cache(entry);
entry             754 fs/fuse/dir.c  		fuse_invalidate_entry(entry);
entry             844 fs/fuse/dir.c  static int fuse_link(struct dentry *entry, struct inode *newdir,
entry             849 fs/fuse/dir.c  	struct inode *inode = d_inode(entry);
entry            1011 fs/fuse/dir.c  	struct dentry *entry;
entry            1027 fs/fuse/dir.c  	entry = d_lookup(dir, name);
entry            1029 fs/fuse/dir.c  	if (!entry)
entry            1033 fs/fuse/dir.c  	fuse_invalidate_entry(entry);
entry            1035 fs/fuse/dir.c  	if (child_nodeid != 0 && d_really_is_positive(entry)) {
entry            1036 fs/fuse/dir.c  		inode_lock(d_inode(entry));
entry            1037 fs/fuse/dir.c  		if (get_node_id(d_inode(entry)) != child_nodeid) {
entry            1041 fs/fuse/dir.c  		if (d_mountpoint(entry)) {
entry            1045 fs/fuse/dir.c  		if (d_is_dir(entry)) {
entry            1046 fs/fuse/dir.c  			shrink_dcache_parent(entry);
entry            1047 fs/fuse/dir.c  			if (!simple_empty(entry)) {
entry            1051 fs/fuse/dir.c  			d_inode(entry)->i_flags |= S_DEAD;
entry            1053 fs/fuse/dir.c  		dont_mount(entry);
entry            1054 fs/fuse/dir.c  		clear_nlink(d_inode(entry));
entry            1057 fs/fuse/dir.c  		inode_unlock(d_inode(entry));
entry            1059 fs/fuse/dir.c  			d_delete(entry);
entry            1063 fs/fuse/dir.c  	dput(entry);
entry            1627 fs/fuse/dir.c  static int fuse_setattr(struct dentry *entry, struct iattr *attr)
entry            1629 fs/fuse/dir.c  	struct inode *inode = d_inode(entry);
entry            1670 fs/fuse/dir.c  	ret = fuse_do_setattr(entry, attr, file);
entry            1680 fs/fuse/dir.c  		if (d_is_dir(entry) && (attr->ia_valid & ATTR_MODE))
entry            1681 fs/fuse/dir.c  			fuse_invalidate_entry_cache(entry);
entry             466 fs/fuse/fuse_i.h 	struct list_head entry;
entry             729 fs/fuse/fuse_i.h 	struct list_head entry;
entry             936 fs/fuse/fuse_i.h void fuse_invalidate_entry_cache(struct dentry *entry);
entry             941 fs/fuse/fuse_i.h void fuse_change_entry_timeout(struct dentry *entry, struct fuse_entry_out *o);
entry            1070 fs/fuse/fuse_i.h ssize_t fuse_listxattr(struct dentry *entry, char *list, size_t size);
entry             390 fs/fuse/inode.c 	list_del(&fc->entry);
entry             615 fs/fuse/inode.c 	INIT_LIST_HEAD(&fc->entry);
entry             675 fs/fuse/inode.c 	struct dentry *entry;
entry             705 fs/fuse/inode.c 	entry = d_obtain_alias(inode);
entry             706 fs/fuse/inode.c 	if (!IS_ERR(entry) && get_node_id(inode) != FUSE_ROOT_ID)
entry             707 fs/fuse/inode.c 		fuse_invalidate_entry_cache(entry);
entry             709 fs/fuse/inode.c 	return entry;
entry            1085 fs/fuse/inode.c 	list_add_tail(&fud->entry, &fc->devices);
entry            1109 fs/fuse/inode.c 		list_del(&fud->entry);
entry            1206 fs/fuse/inode.c 	list_add_tail(&fc->entry, &fuse_conn_list);
entry             107 fs/fuse/xattr.c ssize_t fuse_listxattr(struct dentry *entry, char *list, size_t size)
entry             109 fs/fuse/xattr.c 	struct inode *inode = d_inode(entry);
entry              65 fs/hfs/brec.c  int hfs_brec_insert(struct hfs_find_data *fd, void *entry, int entry_len)
entry             133 fs/hfs/brec.c  	hfs_bnode_write(node, entry, data_off + key_len, entry_len);
entry             155 fs/hfs/brec.c  		entry = &cnid;
entry              86 fs/hfs/catalog.c 	union hfs_cat_rec entry;
entry             109 fs/hfs/catalog.c 	entry_size = hfs_cat_build_thread(sb, &entry, S_ISDIR(inode->i_mode) ?
entry             118 fs/hfs/catalog.c 	err = hfs_brec_insert(&fd, &entry, entry_size);
entry             123 fs/hfs/catalog.c 	entry_size = hfs_cat_build_record(&entry, cnid, inode);
entry             131 fs/hfs/catalog.c 	err = hfs_brec_insert(&fd, &entry, entry_size);
entry             293 fs/hfs/catalog.c 	union hfs_cat_rec entry;
entry             319 fs/hfs/catalog.c 	if (src_fd.entrylength > sizeof(entry) || src_fd.entrylength < 0) {
entry             324 fs/hfs/catalog.c 	hfs_bnode_read(src_fd.bnode, &entry, src_fd.entryoffset,
entry             336 fs/hfs/catalog.c 	err = hfs_brec_insert(&dst_fd, &entry, src_fd.entrylength);
entry             355 fs/hfs/catalog.c 	type = entry.type;
entry             356 fs/hfs/catalog.c 	if (type == HFS_CDR_FIL && !(entry.file.Flags & HFS_FIL_THD))
entry             370 fs/hfs/catalog.c 	entry_size = hfs_cat_build_thread(sb, &entry, type == HFS_CDR_FIL ? HFS_CDR_FTH : HFS_CDR_THD,
entry             378 fs/hfs/catalog.c 	err = hfs_brec_insert(&dst_fd, &entry, entry_size);
entry              54 fs/hfs/dir.c   	union hfs_cat_rec entry;
entry              77 fs/hfs/dir.c   		if (fd.entrylength > sizeof(entry) || fd.entrylength < 0) {
entry              82 fs/hfs/dir.c   		hfs_bnode_read(fd.bnode, &entry, fd.entryoffset, fd.entrylength);
entry              83 fs/hfs/dir.c   		if (entry.type != HFS_CDR_THD) {
entry              94 fs/hfs/dir.c   			    be32_to_cpu(entry.thread.ParID), DT_DIR))
entry             111 fs/hfs/dir.c   		if (fd.entrylength > sizeof(entry) || fd.entrylength < 0) {
entry             116 fs/hfs/dir.c   		hfs_bnode_read(fd.bnode, &entry, fd.entryoffset, fd.entrylength);
entry             117 fs/hfs/dir.c   		type = entry.type;
entry             126 fs/hfs/dir.c   				    be32_to_cpu(entry.dir.DirID), DT_DIR))
entry             135 fs/hfs/dir.c   				    be32_to_cpu(entry.file.FlNum), DT_REG))
entry              88 fs/hfsplus/attributes.c void hfsplus_destroy_attr_entry(hfsplus_attr_entry *entry)
entry              90 fs/hfsplus/attributes.c 	if (entry)
entry              91 fs/hfsplus/attributes.c 		kmem_cache_free(hfsplus_attr_tree_cachep, entry);
entry              96 fs/hfsplus/attributes.c static int hfsplus_attr_build_record(hfsplus_attr_entry *entry, int record_type,
entry             104 fs/hfsplus/attributes.c 		memset(entry, 0, sizeof(*entry));
entry             111 fs/hfsplus/attributes.c 		memset(entry, 0, sizeof(*entry));
entry             116 fs/hfsplus/attributes.c 		memset(entry, 0, sizeof(struct hfsplus_attr_inline_data));
entry             117 fs/hfsplus/attributes.c 		entry->inline_data.record_type = cpu_to_be32(record_type);
entry             122 fs/hfsplus/attributes.c 		entry->inline_data.length = cpu_to_be16(len);
entry             123 fs/hfsplus/attributes.c 		memcpy(entry->inline_data.raw_bytes, value, len);
entry             132 fs/hfsplus/attributes.c 		memset(entry, 0, sizeof(*entry));
entry              63 fs/hfsplus/brec.c int hfs_brec_insert(struct hfs_find_data *fd, void *entry, int entry_len)
entry             133 fs/hfsplus/brec.c 	hfs_bnode_write(node, entry, data_off + key_len, entry_len);
entry             155 fs/hfsplus/brec.c 		entry = &cnid;
entry             105 fs/hfsplus/catalog.c static int hfsplus_cat_build_record(hfsplus_cat_entry *entry,
entry             113 fs/hfsplus/catalog.c 		folder = &entry->folder;
entry             132 fs/hfsplus/catalog.c 		file = &entry->file;
entry             177 fs/hfsplus/catalog.c 				   hfsplus_cat_entry *entry, int type,
entry             182 fs/hfsplus/catalog.c 	entry->type = cpu_to_be16(type);
entry             183 fs/hfsplus/catalog.c 	entry->thread.reserved = 0;
entry             184 fs/hfsplus/catalog.c 	entry->thread.parentID = cpu_to_be32(parentid);
entry             185 fs/hfsplus/catalog.c 	err = hfsplus_asc2uni(sb, &entry->thread.nodeName, HFSPLUS_MAX_STRLEN,
entry             190 fs/hfsplus/catalog.c 	return 10 + be16_to_cpu(entry->thread.nodeName.length) * 2;
entry             258 fs/hfsplus/catalog.c 	hfsplus_cat_entry entry;
entry             277 fs/hfsplus/catalog.c 	entry_size = hfsplus_fill_cat_thread(sb, &entry,
entry             292 fs/hfsplus/catalog.c 	err = hfs_brec_insert(&fd, &entry, entry_size);
entry             300 fs/hfsplus/catalog.c 	entry_size = hfsplus_cat_build_record(&entry, cnid, inode);
entry             308 fs/hfsplus/catalog.c 	err = hfs_brec_insert(&fd, &entry, entry_size);
entry             440 fs/hfsplus/catalog.c 	hfsplus_cat_entry entry;
entry             469 fs/hfsplus/catalog.c 	if (src_fd.entrylength > sizeof(entry) || src_fd.entrylength < 0) {
entry             474 fs/hfsplus/catalog.c 	hfs_bnode_read(src_fd.bnode, &entry, src_fd.entryoffset,
entry             476 fs/hfsplus/catalog.c 	type = be16_to_cpu(entry.type);
entry             491 fs/hfsplus/catalog.c 	err = hfs_brec_insert(&dst_fd, &entry, src_fd.entrylength);
entry             528 fs/hfsplus/catalog.c 	entry_size = hfsplus_fill_cat_thread(sb, &entry, type,
entry             541 fs/hfsplus/catalog.c 	err = hfs_brec_insert(&dst_fd, &entry, entry_size);
entry              36 fs/hfsplus/dir.c 	hfsplus_cat_entry entry;
entry              52 fs/hfsplus/dir.c 	err = hfs_brec_read(&fd, &entry, sizeof(entry));
entry              62 fs/hfsplus/dir.c 	type = be16_to_cpu(entry.type);
entry              68 fs/hfsplus/dir.c 		cnid = be32_to_cpu(entry.folder.id);
entry              75 fs/hfsplus/dir.c 		cnid = be32_to_cpu(entry.file.id);
entry              76 fs/hfsplus/dir.c 		if (entry.file.user_info.fdType ==
entry              78 fs/hfsplus/dir.c 				entry.file.user_info.fdCreator ==
entry              81 fs/hfsplus/dir.c 				(entry.file.create_date ==
entry              84 fs/hfsplus/dir.c 				entry.file.create_date ==
entry             100 fs/hfsplus/dir.c 					be32_to_cpu(entry.file.permissions.dev);
entry             136 fs/hfsplus/dir.c 	hfsplus_cat_entry entry;
entry             164 fs/hfsplus/dir.c 		if (fd.entrylength > sizeof(entry) || fd.entrylength < 0) {
entry             169 fs/hfsplus/dir.c 		hfs_bnode_read(fd.bnode, &entry, fd.entryoffset,
entry             171 fs/hfsplus/dir.c 		if (be16_to_cpu(entry.type) != HFSPLUS_FOLDER_THREAD) {
entry             182 fs/hfsplus/dir.c 			    be32_to_cpu(entry.thread.parentID), DT_DIR))
entry             198 fs/hfsplus/dir.c 		if (fd.entrylength > sizeof(entry) || fd.entrylength < 0) {
entry             203 fs/hfsplus/dir.c 		hfs_bnode_read(fd.bnode, &entry, fd.entryoffset,
entry             205 fs/hfsplus/dir.c 		type = be16_to_cpu(entry.type);
entry             219 fs/hfsplus/dir.c 					be32_to_cpu(entry.folder.id))
entry             222 fs/hfsplus/dir.c 				    be32_to_cpu(entry.folder.id), DT_DIR))
entry             234 fs/hfsplus/dir.c 			mode = be16_to_cpu(entry.file.permissions.mode);
entry             249 fs/hfsplus/dir.c 				      be32_to_cpu(entry.file.id), type))
entry             379 fs/hfsplus/hfsplus_fs.h void hfsplus_destroy_attr_entry(hfsplus_attr_entry *entry);
entry             428 fs/hfsplus/hfsplus_fs.h int hfs_brec_insert(struct hfs_find_data *fd, void *entry, int entry_len);
entry             490 fs/hfsplus/inode.c 	hfsplus_cat_entry entry;
entry             498 fs/hfsplus/inode.c 		struct hfsplus_cat_folder *folder = &entry.folder;
entry             502 fs/hfsplus/inode.c 		hfs_bnode_read(fd->bnode, &entry, fd->entryoffset,
entry             519 fs/hfsplus/inode.c 		struct hfsplus_cat_file *file = &entry.file;
entry             523 fs/hfsplus/inode.c 		hfs_bnode_read(fd->bnode, &entry, fd->entryoffset,
entry             560 fs/hfsplus/inode.c 	hfsplus_cat_entry entry;
entry             577 fs/hfsplus/inode.c 		struct hfsplus_cat_folder *folder = &entry.folder;
entry             581 fs/hfsplus/inode.c 		hfs_bnode_read(fd.bnode, &entry, fd.entryoffset,
entry             593 fs/hfsplus/inode.c 		hfs_bnode_write(fd.bnode, &entry, fd.entryoffset,
entry             596 fs/hfsplus/inode.c 		struct hfsplus_cat_file *file = &entry.file;
entry             597 fs/hfsplus/inode.c 		hfs_bnode_read(fd.bnode, &entry, fd.entryoffset,
entry             600 fs/hfsplus/inode.c 		hfs_bnode_write(fd.bnode, &entry, fd.entryoffset,
entry             603 fs/hfsplus/inode.c 		struct hfsplus_cat_file *file = &entry.file;
entry             607 fs/hfsplus/inode.c 		hfs_bnode_read(fd.bnode, &entry, fd.entryoffset,
entry             620 fs/hfsplus/inode.c 		hfs_bnode_write(fd.bnode, &entry, fd.entryoffset,
entry             379 fs/hfsplus/super.c 	hfsplus_cat_entry entry;
entry             526 fs/hfsplus/super.c 	if (!hfs_brec_read(&fd, &entry, sizeof(entry))) {
entry             528 fs/hfsplus/super.c 		if (entry.type != cpu_to_be16(HFSPLUS_FOLDER)) {
entry             532 fs/hfsplus/super.c 		inode = hfsplus_iget(sb, be32_to_cpu(entry.folder.id));
entry             263 fs/hfsplus/xattr.c 	hfsplus_cat_entry entry;
entry             296 fs/hfsplus/xattr.c 		hfs_bnode_read(cat_fd.bnode, &entry, cat_fd.entryoffset,
entry             298 fs/hfsplus/xattr.c 		if (be16_to_cpu(entry.type) == HFSPLUS_FOLDER) {
entry             300 fs/hfsplus/xattr.c 				memcpy(&entry.folder.user_info, value,
entry             302 fs/hfsplus/xattr.c 				hfs_bnode_write(cat_fd.bnode, &entry,
entry             311 fs/hfsplus/xattr.c 		} else if (be16_to_cpu(entry.type) == HFSPLUS_FILE) {
entry             313 fs/hfsplus/xattr.c 				memcpy(&entry.file.user_info, value,
entry             315 fs/hfsplus/xattr.c 				hfs_bnode_write(cat_fd.bnode, &entry,
entry             494 fs/hfsplus/xattr.c 	hfsplus_attr_entry *entry;
entry             511 fs/hfsplus/xattr.c 	entry = hfsplus_alloc_attr_entry();
entry             512 fs/hfsplus/xattr.c 	if (!entry) {
entry             557 fs/hfsplus/xattr.c 		hfs_bnode_read(fd.bnode, entry, fd.entryoffset,
entry             563 fs/hfsplus/xattr.c 		memcpy(value, entry->inline_data.raw_bytes, record_length);
entry             572 fs/hfsplus/xattr.c 	hfsplus_destroy_attr_entry(entry);
entry            1729 fs/io_uring.c  	if (!list_empty(&poll->wait.entry)) {
entry            1730 fs/io_uring.c  		list_del_init(&poll->wait.entry);
entry            1840 fs/io_uring.c  	list_del_init(&poll->wait.entry);
entry            1908 fs/io_uring.c  	INIT_LIST_HEAD(&poll->wait.entry);
entry            1918 fs/io_uring.c  		if (unlikely(list_empty(&poll->wait.entry))) {
entry            1925 fs/io_uring.c  			list_del_init(&poll->wait.entry);
entry            1981 fs/io_uring.c  	struct list_head *entry;
entry            2005 fs/io_uring.c  		entry = ctx->timeout_list.prev;
entry            2018 fs/io_uring.c  	list_for_each_prev(entry, &ctx->timeout_list) {
entry            2019 fs/io_uring.c  		struct io_kiocb *nxt = list_entry(entry, struct io_kiocb, list);
entry            2053 fs/io_uring.c  	list_add(&req->list, entry);
entry            2977 fs/io_uring.c  			.entry		= LIST_HEAD_INIT(iowq.wq.entry),
entry              59 fs/jffs2/acl.c 	struct jffs2_acl_entry *entry;
entry              86 fs/jffs2/acl.c 		entry = value;
entry              89 fs/jffs2/acl.c 		acl->a_entries[i].e_tag = je16_to_cpu(entry->e_tag);
entry              90 fs/jffs2/acl.c 		acl->a_entries[i].e_perm = je16_to_cpu(entry->e_perm);
entry             105 fs/jffs2/acl.c 						  je32_to_cpu(entry->e_id));
entry             113 fs/jffs2/acl.c 						  je32_to_cpu(entry->e_id));
entry             131 fs/jffs2/acl.c 	struct jffs2_acl_entry *entry;
entry             144 fs/jffs2/acl.c 		entry = e;
entry             145 fs/jffs2/acl.c 		entry->e_tag = cpu_to_je16(acl_e->e_tag);
entry             146 fs/jffs2/acl.c 		entry->e_perm = cpu_to_je16(acl_e->e_perm);
entry             149 fs/jffs2/acl.c 				entry->e_id = cpu_to_je32(
entry             154 fs/jffs2/acl.c 				entry->e_id = cpu_to_je32(
entry            4521 fs/jfs/jfs_dtree.c 	struct ldtentry *entry;
entry            4553 fs/jfs/jfs_dtree.c 	entry = (struct ldtentry *) & p->slot[entry_si];
entry            4556 fs/jfs/jfs_dtree.c 	entry->inumber = cpu_to_le32(new_ino);
entry              73 fs/lockd/procfs.c 	struct proc_dir_entry *entry;
entry              75 fs/lockd/procfs.c 	entry = proc_mkdir("fs/lockd", NULL);
entry              76 fs/lockd/procfs.c 	if (!entry)
entry              78 fs/lockd/procfs.c 	entry = proc_create("nlm_end_grace", S_IRUGO|S_IWUSR, entry,
entry              80 fs/lockd/procfs.c 	if (!entry) {
entry              77 fs/mbcache.c   	struct mb_cache_entry *entry, *dup;
entry              88 fs/mbcache.c   	entry = kmem_cache_alloc(mb_entry_cache, mask);
entry              89 fs/mbcache.c   	if (!entry)
entry              92 fs/mbcache.c   	INIT_LIST_HEAD(&entry->e_list);
entry              94 fs/mbcache.c   	atomic_set(&entry->e_refcnt, 1);
entry              95 fs/mbcache.c   	entry->e_key = key;
entry              96 fs/mbcache.c   	entry->e_value = value;
entry              97 fs/mbcache.c   	entry->e_reusable = reusable;
entry              98 fs/mbcache.c   	entry->e_referenced = 0;
entry             104 fs/mbcache.c   			kmem_cache_free(mb_entry_cache, entry);
entry             108 fs/mbcache.c   	hlist_bl_add_head(&entry->e_hash_list, head);
entry             112 fs/mbcache.c   	list_add_tail(&entry->e_list, &cache->c_list);
entry             114 fs/mbcache.c   	atomic_inc(&entry->e_refcnt);
entry             122 fs/mbcache.c   void __mb_cache_entry_free(struct mb_cache_entry *entry)
entry             124 fs/mbcache.c   	kmem_cache_free(mb_entry_cache, entry);
entry             129 fs/mbcache.c   					   struct mb_cache_entry *entry,
entry             132 fs/mbcache.c   	struct mb_cache_entry *old_entry = entry;
entry             138 fs/mbcache.c   	if (entry && !hlist_bl_unhashed(&entry->e_hash_list))
entry             139 fs/mbcache.c   		node = entry->e_hash_list.next;
entry             143 fs/mbcache.c   		entry = hlist_bl_entry(node, struct mb_cache_entry,
entry             145 fs/mbcache.c   		if (entry->e_key == key && entry->e_reusable) {
entry             146 fs/mbcache.c   			atomic_inc(&entry->e_refcnt);
entry             151 fs/mbcache.c   	entry = NULL;
entry             157 fs/mbcache.c   	return entry;
entry             186 fs/mbcache.c   						struct mb_cache_entry *entry)
entry             188 fs/mbcache.c   	return __entry_find(cache, entry, entry->e_key);
entry             203 fs/mbcache.c   	struct mb_cache_entry *entry;
entry             207 fs/mbcache.c   	hlist_bl_for_each_entry(entry, node, head, e_hash_list) {
entry             208 fs/mbcache.c   		if (entry->e_key == key && entry->e_value == value) {
entry             209 fs/mbcache.c   			atomic_inc(&entry->e_refcnt);
entry             213 fs/mbcache.c   	entry = NULL;
entry             216 fs/mbcache.c   	return entry;
entry             231 fs/mbcache.c   	struct mb_cache_entry *entry;
entry             235 fs/mbcache.c   	hlist_bl_for_each_entry(entry, node, head, e_hash_list) {
entry             236 fs/mbcache.c   		if (entry->e_key == key && entry->e_value == value) {
entry             238 fs/mbcache.c   			hlist_bl_del_init(&entry->e_hash_list);
entry             241 fs/mbcache.c   			if (!list_empty(&entry->e_list)) {
entry             242 fs/mbcache.c   				list_del_init(&entry->e_list);
entry             246 fs/mbcache.c   				atomic_dec(&entry->e_refcnt);
entry             249 fs/mbcache.c   			mb_cache_entry_put(cache, entry);
entry             264 fs/mbcache.c   			  struct mb_cache_entry *entry)
entry             266 fs/mbcache.c   	entry->e_referenced = 1;
entry             283 fs/mbcache.c   	struct mb_cache_entry *entry;
entry             289 fs/mbcache.c   		entry = list_first_entry(&cache->c_list,
entry             291 fs/mbcache.c   		if (entry->e_referenced) {
entry             292 fs/mbcache.c   			entry->e_referenced = 0;
entry             293 fs/mbcache.c   			list_move_tail(&entry->e_list, &cache->c_list);
entry             296 fs/mbcache.c   		list_del_init(&entry->e_list);
entry             303 fs/mbcache.c   		head = mb_cache_entry_head(cache, entry->e_key);
entry             305 fs/mbcache.c   		if (!hlist_bl_unhashed(&entry->e_hash_list)) {
entry             306 fs/mbcache.c   			hlist_bl_del_init(&entry->e_hash_list);
entry             307 fs/mbcache.c   			atomic_dec(&entry->e_refcnt);
entry             310 fs/mbcache.c   		if (mb_cache_entry_put(cache, entry))
entry             394 fs/mbcache.c   	struct mb_cache_entry *entry, *next;
entry             402 fs/mbcache.c   	list_for_each_entry_safe(entry, next, &cache->c_list, e_list) {
entry             403 fs/mbcache.c   		if (!hlist_bl_unhashed(&entry->e_hash_list)) {
entry             404 fs/mbcache.c   			hlist_bl_del_init(&entry->e_hash_list);
entry             405 fs/mbcache.c   			atomic_dec(&entry->e_refcnt);
entry             408 fs/mbcache.c   		list_del(&entry->e_list);
entry             409 fs/mbcache.c   		WARN_ON(atomic_read(&entry->e_refcnt) != 1);
entry             410 fs/mbcache.c   		mb_cache_entry_put(cache, entry);
entry             214 fs/nfs/dir.c   int nfs_readdir_add_to_array(struct nfs_entry *entry, struct page *page)
entry             227 fs/nfs/dir.c   	cache_entry->cookie = entry->prev_cookie;
entry             228 fs/nfs/dir.c   	cache_entry->ino = entry->ino;
entry             229 fs/nfs/dir.c   	cache_entry->d_type = entry->d_type;
entry             230 fs/nfs/dir.c   	ret = nfs_readdir_make_qstr(&cache_entry->string, entry->name, entry->len);
entry             233 fs/nfs/dir.c   	array->last_cookie = entry->cookie;
entry             235 fs/nfs/dir.c   	if (entry->eof != 0)
entry             346 fs/nfs/dir.c   			struct nfs_entry *entry, struct file *file, struct inode *inode)
entry             356 fs/nfs/dir.c   	error = NFS_PROTO(inode)->readdir(file_dentry(file), cred, entry->cookie, pages,
entry             375 fs/nfs/dir.c   		      struct nfs_entry *entry, struct xdr_stream *xdr)
entry             379 fs/nfs/dir.c   	error = desc->decode(xdr, entry, desc->plus);
entry             382 fs/nfs/dir.c   	entry->fattr->time_start = desc->timestamp;
entry             383 fs/nfs/dir.c   	entry->fattr->gencount = desc->gencount;
entry             391 fs/nfs/dir.c   int nfs_same_file(struct dentry *dentry, struct nfs_entry *entry)
entry             404 fs/nfs/dir.c   	if (entry->fattr->fileid != nfsi->fileid)
entry             406 fs/nfs/dir.c   	if (entry->fh->size && nfs_compare_fh(entry->fh, &nfsi->fh) != 0)
entry             457 fs/nfs/dir.c   void nfs_prime_dcache(struct dentry *parent, struct nfs_entry *entry)
entry             459 fs/nfs/dir.c   	struct qstr filename = QSTR_INIT(entry->name, entry->len);
entry             467 fs/nfs/dir.c   	if (!(entry->fattr->valid & NFS_ATTR_FATTR_FILEID))
entry             469 fs/nfs/dir.c   	if (!(entry->fattr->valid & NFS_ATTR_FATTR_FSID))
entry             497 fs/nfs/dir.c   					&entry->fattr->fsid))
entry             499 fs/nfs/dir.c   		if (nfs_same_file(dentry, entry)) {
entry             500 fs/nfs/dir.c   			if (!entry->fh->size)
entry             503 fs/nfs/dir.c   			status = nfs_refresh_inode(d_inode(dentry), entry->fattr);
entry             505 fs/nfs/dir.c   				nfs_setsecurity(d_inode(dentry), entry->fattr, entry->label);
entry             514 fs/nfs/dir.c   	if (!entry->fh->size) {
entry             519 fs/nfs/dir.c   	inode = nfs_fhget(dentry->d_sb, entry->fh, entry->fattr, entry->label);
entry             535 fs/nfs/dir.c   int nfs_readdir_page_filler(nfs_readdir_descriptor_t *desc, struct nfs_entry *entry,
entry             556 fs/nfs/dir.c   		status = xdr_decode(desc, entry, &stream);
entry             566 fs/nfs/dir.c   			nfs_prime_dcache(file_dentry(desc->file), entry);
entry             568 fs/nfs/dir.c   		status = nfs_readdir_add_to_array(entry, page);
entry             571 fs/nfs/dir.c   	} while (!entry->eof);
entry             574 fs/nfs/dir.c   	if (count == 0 || (status == -EBADCOOKIE && entry->eof != 0)) {
entry             619 fs/nfs/dir.c   	struct nfs_entry entry;
entry             627 fs/nfs/dir.c   	entry.prev_cookie = 0;
entry             628 fs/nfs/dir.c   	entry.cookie = desc->last_cookie;
entry             629 fs/nfs/dir.c   	entry.eof = 0;
entry             630 fs/nfs/dir.c   	entry.fh = nfs_alloc_fhandle();
entry             631 fs/nfs/dir.c   	entry.fattr = nfs_alloc_fattr();
entry             632 fs/nfs/dir.c   	entry.server = NFS_SERVER(inode);
entry             633 fs/nfs/dir.c   	if (entry.fh == NULL || entry.fattr == NULL)
entry             636 fs/nfs/dir.c   	entry.label = nfs4_label_alloc(NFS_SERVER(inode), GFP_NOWAIT);
entry             637 fs/nfs/dir.c   	if (IS_ERR(entry.label)) {
entry             638 fs/nfs/dir.c   		status = PTR_ERR(entry.label);
entry             649 fs/nfs/dir.c   		status = nfs_readdir_xdr_filler(pages, desc, &entry, file, inode);
entry             654 fs/nfs/dir.c   		status = nfs_readdir_page_filler(desc, &entry, pages, page, pglen);
entry             665 fs/nfs/dir.c   	nfs4_label_free(entry.label);
entry             667 fs/nfs/dir.c   	nfs_free_fattr(entry.fattr);
entry             668 fs/nfs/dir.c   	nfs_free_fhandle(entry.fh);
entry            2171 fs/nfs/dir.c   static void nfs_access_free_entry(struct nfs_access_entry *entry)
entry            2173 fs/nfs/dir.c   	put_cred(entry->cred);
entry            2174 fs/nfs/dir.c   	kfree_rcu(entry, rcu_head);
entry            2269 fs/nfs/dir.c   	struct nfs_access_entry *entry;
entry            2273 fs/nfs/dir.c   		entry = rb_entry(n, struct nfs_access_entry, rb_node);
entry            2275 fs/nfs/dir.c   		list_move(&entry->lru, head);
entry            2304 fs/nfs/dir.c   		struct nfs_access_entry *entry =
entry            2306 fs/nfs/dir.c   		int cmp = cred_fscmp(cred, entry->cred);
entry            2313 fs/nfs/dir.c   			return entry;
entry            2397 fs/nfs/dir.c   	struct nfs_access_entry *entry;
entry            2403 fs/nfs/dir.c   		entry = rb_entry(parent, struct nfs_access_entry, rb_node);
entry            2404 fs/nfs/dir.c   		cmp = cred_fscmp(set->cred, entry->cred);
entry            2421 fs/nfs/dir.c   	list_del(&entry->lru);
entry            2423 fs/nfs/dir.c   	nfs_access_free_entry(entry);
entry            2490 fs/nfs/dir.c   void nfs_access_set_mask(struct nfs_access_entry *entry, u32 access_result)
entry            2492 fs/nfs/dir.c   	entry->mask = access_result;
entry             930 fs/nfs/nfs2xdr.c int nfs2_decode_dirent(struct xdr_stream *xdr, struct nfs_entry *entry,
entry             945 fs/nfs/nfs2xdr.c 		entry->eof = 1;
entry             952 fs/nfs/nfs2xdr.c 	entry->ino = be32_to_cpup(p);
entry             954 fs/nfs/nfs2xdr.c 	error = decode_filename_inline(xdr, &entry->name, &entry->len);
entry             962 fs/nfs/nfs2xdr.c 	entry->prev_cookie = entry->cookie;
entry             966 fs/nfs/nfs2xdr.c 	entry->cookie = be32_to_cpup(p);
entry             968 fs/nfs/nfs2xdr.c 	entry->d_type = DT_UNKNOWN;
entry             191 fs/nfs/nfs3proc.c static int nfs3_proc_access(struct inode *inode, struct nfs_access_entry *entry)
entry             195 fs/nfs/nfs3proc.c 		.access		= entry->mask,
entry             202 fs/nfs/nfs3proc.c 		.rpc_cred	= entry->cred,
entry             214 fs/nfs/nfs3proc.c 		nfs_access_set_mask(entry, res.access);
entry            1966 fs/nfs/nfs3xdr.c int nfs3_decode_dirent(struct xdr_stream *xdr, struct nfs_entry *entry,
entry            1969 fs/nfs/nfs3xdr.c 	struct user_namespace *userns = rpc_userns(entry->server->client);
entry            1970 fs/nfs/nfs3xdr.c 	struct nfs_entry old = *entry;
entry            1984 fs/nfs/nfs3xdr.c 		entry->eof = 1;
entry            1988 fs/nfs/nfs3xdr.c 	error = decode_fileid3(xdr, &entry->ino);
entry            1992 fs/nfs/nfs3xdr.c 	error = decode_inline_filename3(xdr, &entry->name, &entry->len);
entry            2000 fs/nfs/nfs3xdr.c 	entry->d_type = DT_UNKNOWN;
entry            2003 fs/nfs/nfs3xdr.c 		entry->fattr->valid = 0;
entry            2004 fs/nfs/nfs3xdr.c 		error = decode_post_op_attr(xdr, entry->fattr, userns);
entry            2007 fs/nfs/nfs3xdr.c 		if (entry->fattr->valid & NFS_ATTR_FATTR_V3)
entry            2008 fs/nfs/nfs3xdr.c 			entry->d_type = nfs_umode_to_dtype(entry->fattr->mode);
entry            2010 fs/nfs/nfs3xdr.c 		if (entry->fattr->fileid != entry->ino) {
entry            2011 fs/nfs/nfs3xdr.c 			entry->fattr->mounted_on_fileid = entry->ino;
entry            2012 fs/nfs/nfs3xdr.c 			entry->fattr->valid |= NFS_ATTR_FATTR_MOUNTED_ON_FILEID;
entry            2020 fs/nfs/nfs3xdr.c 			error = decode_nfs_fh3(xdr, entry->fh);
entry            2027 fs/nfs/nfs3xdr.c 			zero_nfs_fh3(entry->fh);
entry            2030 fs/nfs/nfs3xdr.c 	entry->prev_cookie = entry->cookie;
entry            2031 fs/nfs/nfs3xdr.c 	entry->cookie = new_cookie;
entry            2037 fs/nfs/nfs3xdr.c 	*entry = old;
entry            4341 fs/nfs/nfs4proc.c static int _nfs4_proc_access(struct inode *inode, struct nfs_access_entry *entry)
entry            4346 fs/nfs/nfs4proc.c 		.access = entry->mask,
entry            4355 fs/nfs/nfs4proc.c 		.rpc_cred = entry->cred,
entry            4367 fs/nfs/nfs4proc.c 		nfs_access_set_mask(entry, res.access);
entry            4375 fs/nfs/nfs4proc.c static int nfs4_proc_access(struct inode *inode, struct nfs_access_entry *entry)
entry            4382 fs/nfs/nfs4proc.c 		err = _nfs4_proc_access(inode, entry);
entry            7105 fs/nfs/nfs4proc.c 		list_del_init(&wait->entry);
entry            7362 fs/nfs/nfs4xdr.c int nfs4_decode_dirent(struct xdr_stream *xdr, struct nfs_entry *entry,
entry            7378 fs/nfs/nfs4xdr.c 		entry->eof = 1;
entry            7386 fs/nfs/nfs4xdr.c 	entry->len = be32_to_cpup(p);
entry            7388 fs/nfs/nfs4xdr.c 	p = xdr_inline_decode(xdr, entry->len);
entry            7391 fs/nfs/nfs4xdr.c 	entry->name = (const char *) p;
entry            7398 fs/nfs/nfs4xdr.c 	entry->ino = 1;
entry            7399 fs/nfs/nfs4xdr.c 	entry->fattr->valid = 0;
entry            7407 fs/nfs/nfs4xdr.c 	if (decode_getfattr_attrs(xdr, bitmap, entry->fattr, entry->fh,
entry            7408 fs/nfs/nfs4xdr.c 			NULL, entry->label, entry->server) < 0)
entry            7410 fs/nfs/nfs4xdr.c 	if (entry->fattr->valid & NFS_ATTR_FATTR_MOUNTED_ON_FILEID)
entry            7411 fs/nfs/nfs4xdr.c 		entry->ino = entry->fattr->mounted_on_fileid;
entry            7412 fs/nfs/nfs4xdr.c 	else if (entry->fattr->valid & NFS_ATTR_FATTR_FILEID)
entry            7413 fs/nfs/nfs4xdr.c 		entry->ino = entry->fattr->fileid;
entry            7415 fs/nfs/nfs4xdr.c 	entry->d_type = DT_UNKNOWN;
entry            7416 fs/nfs/nfs4xdr.c 	if (entry->fattr->valid & NFS_ATTR_FATTR_TYPE)
entry            7417 fs/nfs/nfs4xdr.c 		entry->d_type = nfs_umode_to_dtype(entry->fattr->mode);
entry            7419 fs/nfs/nfs4xdr.c 	entry->prev_cookie = entry->cookie;
entry            7420 fs/nfs/nfs4xdr.c 	entry->cookie = new_cookie;
entry              55 fs/nfs_common/nfsacl.c 	struct posix_acl_entry *entry =
entry              58 fs/nfs_common/nfsacl.c 	*p++ = htonl(entry->e_tag | nfsacl_desc->typeflag);
entry              59 fs/nfs_common/nfsacl.c 	switch(entry->e_tag) {
entry              67 fs/nfs_common/nfsacl.c 			*p++ = htonl(from_kuid(&init_user_ns, entry->e_uid));
entry              70 fs/nfs_common/nfsacl.c 			*p++ = htonl(from_kgid(&init_user_ns, entry->e_gid));
entry              76 fs/nfs_common/nfsacl.c 	*p++ = htonl(entry->e_perm & S_IRWXO);
entry             151 fs/nfs_common/nfsacl.c 	struct posix_acl_entry *entry;
entry             163 fs/nfs_common/nfsacl.c 	entry = &nfsacl_desc->acl->a_entries[nfsacl_desc->count++];
entry             164 fs/nfs_common/nfsacl.c 	entry->e_tag = ntohl(*p++) & ~NFS_ACL_DEFAULT;
entry             166 fs/nfs_common/nfsacl.c 	entry->e_perm = ntohl(*p++);
entry             168 fs/nfs_common/nfsacl.c 	switch(entry->e_tag) {
entry             170 fs/nfs_common/nfsacl.c 			entry->e_uid = make_kuid(&init_user_ns, id);
entry             171 fs/nfs_common/nfsacl.c 			if (!uid_valid(entry->e_uid))
entry             175 fs/nfs_common/nfsacl.c 			entry->e_gid = make_kgid(&init_user_ns, id);
entry             176 fs/nfs_common/nfsacl.c 			if (!gid_valid(entry->e_gid))
entry             182 fs/nfs_common/nfsacl.c 			if (entry->e_perm & ~S_IRWXO)
entry             187 fs/nfs_common/nfsacl.c 			entry->e_perm &= S_IRWXO;
entry             283 fs/nfsd/nfs4recover.c 	struct name_list *entry;
entry             287 fs/nfsd/nfs4recover.c 	entry = kmalloc(sizeof(struct name_list), GFP_KERNEL);
entry             288 fs/nfsd/nfs4recover.c 	if (entry == NULL)
entry             290 fs/nfsd/nfs4recover.c 	memcpy(entry->name, name, HEXDIR_LEN - 1);
entry             291 fs/nfsd/nfs4recover.c 	entry->name[HEXDIR_LEN - 1] = '\0';
entry             292 fs/nfsd/nfs4recover.c 	list_add(&entry->list, &ctx->names);
entry             305 fs/nfsd/nfs4recover.c 	struct name_list *entry, *tmp;
entry             321 fs/nfsd/nfs4recover.c 	list_for_each_entry_safe(entry, tmp, &ctx.names, list) {
entry             324 fs/nfsd/nfs4recover.c 			dentry = lookup_one_len(entry->name, dir, HEXDIR_LEN-1);
entry             332 fs/nfsd/nfs4recover.c 		list_del(&entry->list);
entry             333 fs/nfsd/nfs4recover.c 		kfree(entry);
entry             338 fs/nfsd/nfs4recover.c 	list_for_each_entry_safe(entry, tmp, &ctx.names, list) {
entry             339 fs/nfsd/nfs4recover.c 		dprintk("NFSD: %s. Left entry %s\n", __func__, entry->name);
entry             340 fs/nfsd/nfs4recover.c 		list_del(&entry->list);
entry             341 fs/nfsd/nfs4recover.c 		kfree(entry);
entry            1430 fs/nfsd/nfsctl.c 	struct proc_dir_entry *entry;
entry            1432 fs/nfsd/nfsctl.c 	entry = proc_mkdir("fs/nfs", NULL);
entry            1433 fs/nfsd/nfsctl.c 	if (!entry)
entry            1435 fs/nfsd/nfsctl.c 	entry = proc_create("exports", 0, entry,
entry            1437 fs/nfsd/nfsctl.c 	if (!entry) {
entry              78 fs/nilfs2/dat.c 	struct nilfs_dat_entry *entry;
entry              82 fs/nilfs2/dat.c 	entry = nilfs_palloc_block_get_entry(dat, req->pr_entry_nr,
entry              84 fs/nilfs2/dat.c 	entry->de_start = cpu_to_le64(NILFS_CNO_MIN);
entry              85 fs/nilfs2/dat.c 	entry->de_end = cpu_to_le64(NILFS_CNO_MAX);
entry              86 fs/nilfs2/dat.c 	entry->de_blocknr = cpu_to_le64(0);
entry             102 fs/nilfs2/dat.c 	struct nilfs_dat_entry *entry;
entry             106 fs/nilfs2/dat.c 	entry = nilfs_palloc_block_get_entry(dat, req->pr_entry_nr,
entry             108 fs/nilfs2/dat.c 	entry->de_start = cpu_to_le64(NILFS_CNO_MIN);
entry             109 fs/nilfs2/dat.c 	entry->de_end = cpu_to_le64(NILFS_CNO_MIN);
entry             110 fs/nilfs2/dat.c 	entry->de_blocknr = cpu_to_le64(0);
entry             129 fs/nilfs2/dat.c 	struct nilfs_dat_entry *entry;
entry             133 fs/nilfs2/dat.c 	entry = nilfs_palloc_block_get_entry(dat, req->pr_entry_nr,
entry             135 fs/nilfs2/dat.c 	entry->de_start = cpu_to_le64(nilfs_mdt_cno(dat));
entry             136 fs/nilfs2/dat.c 	entry->de_blocknr = cpu_to_le64(blocknr);
entry             144 fs/nilfs2/dat.c 	struct nilfs_dat_entry *entry;
entry             156 fs/nilfs2/dat.c 	entry = nilfs_palloc_block_get_entry(dat, req->pr_entry_nr,
entry             158 fs/nilfs2/dat.c 	blocknr = le64_to_cpu(entry->de_blocknr);
entry             175 fs/nilfs2/dat.c 	struct nilfs_dat_entry *entry;
entry             181 fs/nilfs2/dat.c 	entry = nilfs_palloc_block_get_entry(dat, req->pr_entry_nr,
entry             183 fs/nilfs2/dat.c 	end = start = le64_to_cpu(entry->de_start);
entry             188 fs/nilfs2/dat.c 	entry->de_end = cpu_to_le64(end);
entry             189 fs/nilfs2/dat.c 	blocknr = le64_to_cpu(entry->de_blocknr);
entry             200 fs/nilfs2/dat.c 	struct nilfs_dat_entry *entry;
entry             206 fs/nilfs2/dat.c 	entry = nilfs_palloc_block_get_entry(dat, req->pr_entry_nr,
entry             208 fs/nilfs2/dat.c 	start = le64_to_cpu(entry->de_start);
entry             209 fs/nilfs2/dat.c 	blocknr = le64_to_cpu(entry->de_blocknr);
entry             316 fs/nilfs2/dat.c 	struct nilfs_dat_entry *entry;
entry             341 fs/nilfs2/dat.c 	entry = nilfs_palloc_block_get_entry(dat, vblocknr, entry_bh, kaddr);
entry             342 fs/nilfs2/dat.c 	if (unlikely(entry->de_blocknr == cpu_to_le64(0))) {
entry             346 fs/nilfs2/dat.c 			  (unsigned long long)le64_to_cpu(entry->de_start),
entry             347 fs/nilfs2/dat.c 			  (unsigned long long)le64_to_cpu(entry->de_end));
entry             353 fs/nilfs2/dat.c 	entry->de_blocknr = cpu_to_le64(blocknr);
entry             386 fs/nilfs2/dat.c 	struct nilfs_dat_entry *entry;
entry             405 fs/nilfs2/dat.c 	entry = nilfs_palloc_block_get_entry(dat, vblocknr, entry_bh, kaddr);
entry             406 fs/nilfs2/dat.c 	blocknr = le64_to_cpu(entry->de_blocknr);
entry             423 fs/nilfs2/dat.c 	struct nilfs_dat_entry *entry;
entry             445 fs/nilfs2/dat.c 			entry = nilfs_palloc_block_get_entry(
entry             447 fs/nilfs2/dat.c 			vinfo->vi_start = le64_to_cpu(entry->de_start);
entry             448 fs/nilfs2/dat.c 			vinfo->vi_end = le64_to_cpu(entry->de_end);
entry             449 fs/nilfs2/dat.c 			vinfo->vi_blocknr = le64_to_cpu(entry->de_blocknr);
entry            2198 fs/nilfs2/segment.c 	list_for_each_entry_safe(wrq, n, &sci->sc_wait_request.head, wq.entry) {
entry              45 fs/ntfs/index.c 	if (ictx->entry) {
entry             202 fs/ntfs/index.c 			ictx->entry = ie;
entry              63 fs/ntfs/index.h 	INDEX_ENTRY *entry;
entry              51 fs/ocfs2/acl.c 		struct ocfs2_acl_entry *entry =
entry              54 fs/ocfs2/acl.c 		acl->a_entries[n].e_tag  = le16_to_cpu(entry->e_tag);
entry              55 fs/ocfs2/acl.c 		acl->a_entries[n].e_perm = le16_to_cpu(entry->e_perm);
entry              60 fs/ocfs2/acl.c 					  le32_to_cpu(entry->e_id));
entry              65 fs/ocfs2/acl.c 					  le32_to_cpu(entry->e_id));
entry              81 fs/ocfs2/acl.c 	struct ocfs2_acl_entry *entry = NULL;
entry              91 fs/ocfs2/acl.c 	entry = (struct ocfs2_acl_entry *)ocfs2_acl;
entry              92 fs/ocfs2/acl.c 	for (n = 0; n < acl->a_count; n++, entry++) {
entry              93 fs/ocfs2/acl.c 		entry->e_tag  = cpu_to_le16(acl->a_entries[n].e_tag);
entry              94 fs/ocfs2/acl.c 		entry->e_perm = cpu_to_le16(acl->a_entries[n].e_perm);
entry              97 fs/ocfs2/acl.c 			entry->e_id = cpu_to_le32(
entry             102 fs/ocfs2/acl.c 			entry->e_id = cpu_to_le32(
entry             107 fs/ocfs2/acl.c 			entry->e_id = cpu_to_le32(ACL_UNDEFINED_ID);
entry             102 fs/ocfs2/filecheck.c 	struct ocfs2_filecheck_sysfs_entry *entry = container_of(kobj,
entry             105 fs/ocfs2/filecheck.c 	complete(&entry->fs_kobj_unregister);
entry             149 fs/ocfs2/filecheck.c ocfs2_filecheck_sysfs_free(struct ocfs2_filecheck_sysfs_entry *entry)
entry             153 fs/ocfs2/filecheck.c 	spin_lock(&entry->fs_fcheck->fc_lock);
entry             154 fs/ocfs2/filecheck.c 	while (!list_empty(&entry->fs_fcheck->fc_head)) {
entry             155 fs/ocfs2/filecheck.c 		p = list_first_entry(&entry->fs_fcheck->fc_head,
entry             161 fs/ocfs2/filecheck.c 	spin_unlock(&entry->fs_fcheck->fc_lock);
entry             163 fs/ocfs2/filecheck.c 	kfree(entry->fs_fcheck);
entry             164 fs/ocfs2/filecheck.c 	entry->fs_fcheck = NULL;
entry             171 fs/ocfs2/filecheck.c 	struct ocfs2_filecheck_sysfs_entry *entry = &osb->osb_fc_ent;
entry             183 fs/ocfs2/filecheck.c 	entry->fs_kobj.kset = osb->osb_dev_kset;
entry             184 fs/ocfs2/filecheck.c 	init_completion(&entry->fs_kobj_unregister);
entry             185 fs/ocfs2/filecheck.c 	ret = kobject_init_and_add(&entry->fs_kobj, &ocfs2_ktype_filecheck,
entry             188 fs/ocfs2/filecheck.c 		kobject_put(&entry->fs_kobj);
entry             193 fs/ocfs2/filecheck.c 	entry->fs_fcheck = fcheck;
entry             402 fs/ocfs2/filecheck.c 			   struct ocfs2_filecheck_entry *entry)
entry             405 fs/ocfs2/filecheck.c 	entry->fe_done = 1;
entry             434 fs/ocfs2/filecheck.c 			     struct ocfs2_filecheck_entry *entry)
entry             439 fs/ocfs2/filecheck.c 	if (entry->fe_type == OCFS2_FILECHECK_TYPE_CHK)
entry             440 fs/ocfs2/filecheck.c 		entry->fe_status = ocfs2_filecheck_handle(osb,
entry             441 fs/ocfs2/filecheck.c 				entry->fe_ino, OCFS2_FI_FLAG_FILECHECK_CHK);
entry             442 fs/ocfs2/filecheck.c 	else if (entry->fe_type == OCFS2_FILECHECK_TYPE_FIX)
entry             443 fs/ocfs2/filecheck.c 		entry->fe_status = ocfs2_filecheck_handle(osb,
entry             444 fs/ocfs2/filecheck.c 				entry->fe_ino, OCFS2_FI_FLAG_FILECHECK_FIX);
entry             446 fs/ocfs2/filecheck.c 		entry->fe_status = OCFS2_FILECHECK_ERR_UNSUPPORTED;
entry             448 fs/ocfs2/filecheck.c 	ocfs2_filecheck_done_entry(ent, entry);
entry             457 fs/ocfs2/filecheck.c 	struct ocfs2_filecheck_entry *entry;
entry             472 fs/ocfs2/filecheck.c 	entry = kmalloc(sizeof(struct ocfs2_filecheck_entry), GFP_NOFS);
entry             473 fs/ocfs2/filecheck.c 	if (!entry) {
entry             481 fs/ocfs2/filecheck.c 		kfree(entry);
entry             489 fs/ocfs2/filecheck.c 		kfree(entry);
entry             500 fs/ocfs2/filecheck.c 		entry->fe_ino = args.fa_ino;
entry             501 fs/ocfs2/filecheck.c 		entry->fe_type = args.fa_type;
entry             502 fs/ocfs2/filecheck.c 		entry->fe_done = 0;
entry             503 fs/ocfs2/filecheck.c 		entry->fe_status = OCFS2_FILECHECK_ERR_INPROGRESS;
entry             504 fs/ocfs2/filecheck.c 		list_add_tail(&entry->fe_list, &ent->fs_fcheck->fc_head);
entry             510 fs/ocfs2/filecheck.c 		ocfs2_filecheck_handle_entry(ent, entry);
entry             939 fs/ocfs2/xattr.c 		struct ocfs2_xattr_entry *entry = &header->xh_entries[i];
entry             940 fs/ocfs2/xattr.c 		type = ocfs2_xattr_get_type(entry);
entry             942 fs/ocfs2/xattr.c 			le16_to_cpu(entry->xe_name_offset);
entry             947 fs/ocfs2/xattr.c 					     entry->xe_name_len);
entry            1075 fs/ocfs2/xattr.c 	struct ocfs2_xattr_entry *entry;
entry            1083 fs/ocfs2/xattr.c 	entry = xs->here;
entry            1085 fs/ocfs2/xattr.c 		cmp = name_index - ocfs2_xattr_get_type(entry);
entry            1087 fs/ocfs2/xattr.c 			cmp = name_len - entry->xe_name_len;
entry            1090 fs/ocfs2/xattr.c 				     le16_to_cpu(entry->xe_name_offset)),
entry            1094 fs/ocfs2/xattr.c 		entry += 1;
entry            1096 fs/ocfs2/xattr.c 	xs->here = entry;
entry            1630 fs/ocfs2/xattr.c 	struct ocfs2_xattr_entry *entry = loc->xl_entry;
entry            1634 fs/ocfs2/xattr.c 	namevalue_offset = le16_to_cpu(entry->xe_name_offset);
entry            1635 fs/ocfs2/xattr.c 	namevalue_size = namevalue_size_xe(entry);
entry            1936 fs/ocfs2/xattr.c 	struct ocfs2_xattr_entry *entry = loc->xl_entry;
entry            1950 fs/ocfs2/xattr.c 		index = ((char *)entry - (char *)&xh->xh_entries) /
entry            2270 fs/ocfs2/xattr.c 				     struct ocfs2_xattr_entry *entry)
entry            2279 fs/ocfs2/xattr.c 	loc->xl_entry = entry;
entry            2289 fs/ocfs2/xattr.c 					  struct ocfs2_xattr_entry *entry)
entry            2300 fs/ocfs2/xattr.c 	loc->xl_entry = entry;
entry            2307 fs/ocfs2/xattr.c 					   struct ocfs2_xattr_entry *entry)
entry            2313 fs/ocfs2/xattr.c 	loc->xl_entry = entry;
entry            2376 fs/ocfs2/xattr.c 		struct ocfs2_xattr_entry *entry = &header->xh_entries[i];
entry            2378 fs/ocfs2/xattr.c 		if (ocfs2_xattr_is_local(entry))
entry            2382 fs/ocfs2/xattr.c 			le16_to_cpu(entry->xe_name_offset);
entry            2384 fs/ocfs2/xattr.c 			(val + OCFS2_XATTR_SIZE(entry->xe_name_len));
entry            4055 fs/ocfs2/xattr.c 		struct ocfs2_xattr_entry *entry = &bucket_xh(bucket)->xh_entries[i];
entry            4056 fs/ocfs2/xattr.c 		type = ocfs2_xattr_get_type(entry);
entry            4073 fs/ocfs2/xattr.c 					     entry->xe_name_len);
entry             119 fs/omfs/dir.c  	__be64 *entry;
entry             127 fs/omfs/dir.c  	entry = (__be64 *) &bh->b_data[ofs];
entry             128 fs/omfs/dir.c  	block = be64_to_cpu(*entry);
entry             129 fs/omfs/dir.c  	*entry = cpu_to_be64(inode->i_ino);
entry             164 fs/omfs/dir.c  	__be64 *entry, next;
entry             174 fs/omfs/dir.c  	entry = (__be64 *) &bh->b_data[ofs];
entry             175 fs/omfs/dir.c  	block = be64_to_cpu(*entry);
entry             195 fs/omfs/dir.c  		entry = &oi->i_sibling;
entry             198 fs/omfs/dir.c  	*entry = next;
entry              35 fs/omfs/file.c 	struct omfs_extent_entry *entry;
entry              71 fs/omfs/file.c 		entry = &oe->e_entry;
entry              76 fs/omfs/file.c 			start = be64_to_cpu(entry->e_cluster);
entry              77 fs/omfs/file.c 			count = be64_to_cpu(entry->e_blocks);
entry              80 fs/omfs/file.c 			entry++;
entry             120 fs/omfs/file.c 	struct omfs_extent_entry *entry = &oe->e_entry;
entry             142 fs/omfs/file.c 	terminator = entry + extent_count - 1;
entry             144 fs/omfs/file.c 		entry = terminator-1;
entry             145 fs/omfs/file.c 		new_block = be64_to_cpu(entry->e_cluster) +
entry             146 fs/omfs/file.c 			be64_to_cpu(entry->e_blocks);
entry             149 fs/omfs/file.c 			be64_add_cpu(&entry->e_blocks, 1);
entry             168 fs/omfs/file.c 	entry = terminator;
entry             170 fs/omfs/file.c 	memcpy(terminator, entry, sizeof(struct omfs_extent_entry));
entry             172 fs/omfs/file.c 	entry->e_cluster = cpu_to_be64(new_block);
entry             173 fs/omfs/file.c 	entry->e_blocks = cpu_to_be64((u64) new_count);
entry             227 fs/omfs/file.c 	struct omfs_extent_entry *entry;
entry             248 fs/omfs/file.c 		entry = &oe->e_entry;
entry             253 fs/omfs/file.c 		offset = find_block(inode, entry, block, extent_count, &remain);
entry              50 fs/orangefs/orangefs-bufmap.c 			if (likely(list_empty(&wait.entry)))
entry              88 fs/orangefs/orangefs-bufmap.c 		if (likely(list_empty(&wait.entry)))
entry             112 fs/orangefs/orangefs-bufmap.c 	if (!list_empty(&wait.entry))
entry             113 fs/orangefs/orangefs-bufmap.c 		list_del(&wait.entry);
entry             671 fs/posix_acl.c 	struct posix_acl_xattr_entry *entry = (void *)(header + 1), *end;
entry             689 fs/posix_acl.c 	for (end = entry + count; entry != end; entry++) {
entry             690 fs/posix_acl.c 		switch(le16_to_cpu(entry->e_tag)) {
entry             692 fs/posix_acl.c 			uid = make_kuid(from, le32_to_cpu(entry->e_id));
entry             693 fs/posix_acl.c 			entry->e_id = cpu_to_le32(from_kuid(to, uid));
entry             696 fs/posix_acl.c 			gid = make_kgid(from, le32_to_cpu(entry->e_id));
entry             697 fs/posix_acl.c 			entry->e_id = cpu_to_le32(from_kgid(to, gid));
entry             729 fs/posix_acl.c 	const struct posix_acl_xattr_entry *entry = (const void *)(header + 1), *end;
entry             752 fs/posix_acl.c 	for (end = entry + count; entry != end; acl_e++, entry++) {
entry             753 fs/posix_acl.c 		acl_e->e_tag  = le16_to_cpu(entry->e_tag);
entry             754 fs/posix_acl.c 		acl_e->e_perm = le16_to_cpu(entry->e_perm);
entry             766 fs/posix_acl.c 						  le32_to_cpu(entry->e_id));
entry             773 fs/posix_acl.c 						  le32_to_cpu(entry->e_id));
entry             113 fs/proc/namespaces.c 	const struct proc_ns_operations **entry, **last;
entry             122 fs/proc/namespaces.c 	entry = ns_entries + (ctx->pos - 2);
entry             124 fs/proc/namespaces.c 	while (entry <= last) {
entry             125 fs/proc/namespaces.c 		const struct proc_ns_operations *ops = *entry;
entry             130 fs/proc/namespaces.c 		entry++;
entry             147 fs/proc/namespaces.c 	const struct proc_ns_operations **entry, **last;
entry             155 fs/proc/namespaces.c 	for (entry = ns_entries; entry < last; entry++) {
entry             156 fs/proc/namespaces.c 		if (strlen((*entry)->name) != len)
entry             158 fs/proc/namespaces.c 		if (!memcmp(dentry->d_name.name, (*entry)->name, len))
entry             161 fs/proc/namespaces.c 	if (entry == last)
entry             164 fs/proc/namespaces.c 	res = proc_ns_instantiate(dentry, task, *entry);
entry             113 fs/proc/proc_sysctl.c 	struct ctl_table *entry;
entry             124 fs/proc/proc_sysctl.c 		entry = &head->ctl_table[ctl_node - head->node];
entry             125 fs/proc/proc_sysctl.c 		procname = entry->procname;
entry             134 fs/proc/proc_sysctl.c 			return entry;
entry             140 fs/proc/proc_sysctl.c static int insert_entry(struct ctl_table_header *head, struct ctl_table *entry)
entry             142 fs/proc/proc_sysctl.c 	struct rb_node *node = &head->node[entry - head->ctl_table].node;
entry             145 fs/proc/proc_sysctl.c 	const char *name = entry->procname;
entry             169 fs/proc/proc_sysctl.c 			pr_cont("/%s\n", entry->procname);
entry             179 fs/proc/proc_sysctl.c static void erase_entry(struct ctl_table_header *head, struct ctl_table *entry)
entry             181 fs/proc/proc_sysctl.c 	struct rb_node *node = &head->node[entry - head->ctl_table].node;
entry             202 fs/proc/proc_sysctl.c 		struct ctl_table *entry;
entry             203 fs/proc/proc_sysctl.c 		for (entry = table; entry->procname; entry++, node++)
entry             210 fs/proc/proc_sysctl.c 	struct ctl_table *entry;
entry             211 fs/proc/proc_sysctl.c 	for (entry = head->ctl_table; entry->procname; entry++)
entry             212 fs/proc/proc_sysctl.c 		erase_entry(head, entry);
entry             217 fs/proc/proc_sysctl.c 	struct ctl_table *entry;
entry             236 fs/proc/proc_sysctl.c 	for (entry = header->ctl_table; entry->procname; entry++) {
entry             237 fs/proc/proc_sysctl.c 		err = insert_entry(header, entry);
entry             372 fs/proc/proc_sysctl.c 	struct ctl_table *entry;
entry             375 fs/proc/proc_sysctl.c 	entry = find_entry(&head, dir, name, namelen);
entry             376 fs/proc/proc_sysctl.c 	if (entry && use_table(head))
entry             379 fs/proc/proc_sysctl.c 		entry = NULL;
entry             381 fs/proc/proc_sysctl.c 	return entry;
entry             400 fs/proc/proc_sysctl.c 	struct ctl_table *entry = NULL;
entry             408 fs/proc/proc_sysctl.c 		entry = &head->ctl_table[ctl_node - head->node];
entry             411 fs/proc/proc_sysctl.c 	*pentry = entry;
entry             417 fs/proc/proc_sysctl.c 	struct ctl_table *entry = *pentry;
entry             418 fs/proc/proc_sysctl.c 	struct ctl_node *ctl_node = &head->node[entry - head->ctl_table];
entry             428 fs/proc/proc_sysctl.c 		entry = &head->ctl_table[ctl_node - head->node];
entry             431 fs/proc/proc_sysctl.c 	*pentry = entry;
entry             782 fs/proc/proc_sysctl.c 	struct ctl_table *entry;
entry             796 fs/proc/proc_sysctl.c 	for (first_entry(ctl_dir, &h, &entry); h; next_entry(&h, &entry)) {
entry             797 fs/proc/proc_sysctl.c 		if (!scan(h, entry, &pos, file, ctx)) {
entry             954 fs/proc/proc_sysctl.c 	struct ctl_table *entry;
entry             956 fs/proc/proc_sysctl.c 	entry = find_entry(&head, dir, name, namelen);
entry             957 fs/proc/proc_sysctl.c 	if (!entry)
entry             959 fs/proc/proc_sysctl.c 	if (!S_ISDIR(entry->mode))
entry            1071 fs/proc/proc_sysctl.c 	struct ctl_table *entry;
entry            1085 fs/proc/proc_sysctl.c 		entry = find_entry(&head, dir, procname, strlen(procname));
entry            1087 fs/proc/proc_sysctl.c 		if (entry && use_table(head)) {
entry            1090 fs/proc/proc_sysctl.c 			*pentry = entry;
entry            1165 fs/proc/proc_sysctl.c 	struct ctl_table *link_table, *entry, *link;
entry            1173 fs/proc/proc_sysctl.c 	for (entry = table; entry->procname; entry++) {
entry            1175 fs/proc/proc_sysctl.c 		name_bytes += strlen(entry->procname) + 1;
entry            1191 fs/proc/proc_sysctl.c 	for (link = link_table, entry = table; entry->procname; link++, entry++) {
entry            1192 fs/proc/proc_sysctl.c 		int len = strlen(entry->procname) + 1;
entry            1193 fs/proc/proc_sysctl.c 		memcpy(link_name, entry->procname, len);
entry            1209 fs/proc/proc_sysctl.c 	struct ctl_table *entry, *link;
entry            1212 fs/proc/proc_sysctl.c 	for (entry = table; entry->procname; entry++) {
entry            1213 fs/proc/proc_sysctl.c 		const char *procname = entry->procname;
entry            1217 fs/proc/proc_sysctl.c 		if (S_ISDIR(link->mode) && S_ISDIR(entry->mode))
entry            1225 fs/proc/proc_sysctl.c 	for (entry = table; entry->procname; entry++) {
entry            1226 fs/proc/proc_sysctl.c 		const char *procname = entry->procname;
entry            1324 fs/proc/proc_sysctl.c 	struct ctl_table *entry;
entry            1328 fs/proc/proc_sysctl.c 	for (entry = table; entry->procname; entry++)
entry            1417 fs/proc/proc_sysctl.c 	struct ctl_table *entry;
entry            1423 fs/proc/proc_sysctl.c 	for (entry = table; entry->procname; entry++) {
entry            1424 fs/proc/proc_sysctl.c 		if (entry->child)
entry            1425 fs/proc/proc_sysctl.c 			nr_subheaders += count_subheaders(entry->child);
entry            1437 fs/proc/proc_sysctl.c 	struct ctl_table *entry, *files;
entry            1442 fs/proc/proc_sysctl.c 	for (entry = table; entry->procname; entry++) {
entry            1443 fs/proc/proc_sysctl.c 		if (entry->child)
entry            1459 fs/proc/proc_sysctl.c 		for (new = files, entry = table; entry->procname; entry++) {
entry            1460 fs/proc/proc_sysctl.c 			if (entry->child)
entry            1462 fs/proc/proc_sysctl.c 			*new = *entry;
entry            1483 fs/proc/proc_sysctl.c 	for (entry = table; entry->procname; entry++) {
entry            1486 fs/proc/proc_sysctl.c 		if (!entry->child)
entry            1490 fs/proc/proc_sysctl.c 		child_pos = append_path(path, pos, entry->procname);
entry            1495 fs/proc/proc_sysctl.c 						  set, entry->child);
entry            1619 fs/proc/proc_sysctl.c 	struct ctl_table *entry;
entry            1628 fs/proc/proc_sysctl.c 	for (entry = header->ctl_table; entry->procname; entry++) {
entry            1631 fs/proc/proc_sysctl.c 		const char *name = entry->procname;
entry            1635 fs/proc/proc_sysctl.c 		    ((S_ISDIR(link->mode) && S_ISDIR(entry->mode)) ||
entry            1338 fs/proc/task_mmu.c 		swp_entry_t entry;
entry            1341 fs/proc/task_mmu.c 		entry = pte_to_swp_entry(pte);
entry            1343 fs/proc/task_mmu.c 			frame = swp_type(entry) |
entry            1344 fs/proc/task_mmu.c 				(swp_offset(entry) << MAX_SWAPFILES_SHIFT);
entry            1346 fs/proc/task_mmu.c 		if (is_migration_entry(entry))
entry            1347 fs/proc/task_mmu.c 			page = migration_entry_to_page(entry);
entry            1349 fs/proc/task_mmu.c 		if (is_device_private_entry(entry))
entry            1350 fs/proc/task_mmu.c 			page = device_private_entry_to_page(entry);
entry            1394 fs/proc/task_mmu.c 			swp_entry_t entry = pmd_to_swp_entry(pmd);
entry            1398 fs/proc/task_mmu.c 				offset = swp_offset(entry) +
entry            1400 fs/proc/task_mmu.c 				frame = swp_type(entry) |
entry            1407 fs/proc/task_mmu.c 			page = migration_entry_to_page(entry);
entry             161 fs/reiserfs/journal.c 	struct list_head *entry = journal->j_bitmap_nodes.next;
entry             166 fs/reiserfs/journal.c 	if (entry != &journal->j_bitmap_nodes) {
entry             167 fs/reiserfs/journal.c 		bn = list_entry(entry, struct reiserfs_bitmap_node, list);
entry             168 fs/reiserfs/journal.c 		list_del(entry);
entry             584 fs/reiserfs/journal.c 	struct list_head *entry = &journal->j_journal_list;
entry             587 fs/reiserfs/journal.c 	if (!list_empty(entry)) {
entry             588 fs/reiserfs/journal.c 		jl = JOURNAL_LIST_ENTRY(entry->next);
entry             891 fs/reiserfs/journal.c 	struct list_head *entry;
entry             900 fs/reiserfs/journal.c 	entry = jl->j_list.prev;
entry             902 fs/reiserfs/journal.c 		other_jl = JOURNAL_LIST_ENTRY(entry);
entry             903 fs/reiserfs/journal.c 		if (entry == &journal->j_journal_list ||
entry             908 fs/reiserfs/journal.c 		entry = other_jl->j_list.prev;
entry             916 fs/reiserfs/journal.c 	entry = &first_jl->j_list;
entry             918 fs/reiserfs/journal.c 		other_jl = JOURNAL_LIST_ENTRY(entry);
entry             940 fs/reiserfs/journal.c 			entry = entry->next;
entry             941 fs/reiserfs/journal.c 			if (entry == &journal->j_journal_list)
entry            1301 fs/reiserfs/journal.c 	struct list_head *entry;
entry            1311 fs/reiserfs/journal.c 	entry = journal->j_journal_list.next;
entry            1313 fs/reiserfs/journal.c 	if (entry == &journal->j_journal_list)
entry            1315 fs/reiserfs/journal.c 	other_jl = JOURNAL_LIST_ENTRY(entry);
entry            1727 fs/reiserfs/journal.c 	struct list_head *entry;
entry            1755 fs/reiserfs/journal.c 		entry = jl->j_list.next;
entry            1758 fs/reiserfs/journal.c 		if (entry == &journal->j_journal_list) {
entry            1761 fs/reiserfs/journal.c 		jl = JOURNAL_LIST_ENTRY(entry);
entry            3534 fs/reiserfs/journal.c 	struct list_head *entry;
entry            3539 fs/reiserfs/journal.c 		entry = journal->j_journal_list.prev;
entry            3540 fs/reiserfs/journal.c 		jl = JOURNAL_LIST_ENTRY(entry);
entry            3987 fs/reiserfs/journal.c 	struct list_head *entry, *safe;
entry            4307 fs/reiserfs/journal.c 	list_for_each_safe(entry, safe, &journal->j_journal_list) {
entry            4308 fs/reiserfs/journal.c 		temp_jl = JOURNAL_LIST_ENTRY(entry);
entry              88 fs/reiserfs/xattr_acl.c 		reiserfs_acl_entry *entry = (reiserfs_acl_entry *) value;
entry              91 fs/reiserfs/xattr_acl.c 		acl->a_entries[n].e_tag = le16_to_cpu(entry->e_tag);
entry              92 fs/reiserfs/xattr_acl.c 		acl->a_entries[n].e_perm = le16_to_cpu(entry->e_perm);
entry             108 fs/reiserfs/xattr_acl.c 					  le32_to_cpu(entry->e_id));
entry             116 fs/reiserfs/xattr_acl.c 					  le32_to_cpu(entry->e_id));
entry             152 fs/reiserfs/xattr_acl.c 		reiserfs_acl_entry *entry = (reiserfs_acl_entry *) e;
entry             153 fs/reiserfs/xattr_acl.c 		entry->e_tag = cpu_to_le16(acl->a_entries[n].e_tag);
entry             154 fs/reiserfs/xattr_acl.c 		entry->e_perm = cpu_to_le16(acl->a_entries[n].e_perm);
entry             157 fs/reiserfs/xattr_acl.c 			entry->e_id = cpu_to_le32(
entry             162 fs/reiserfs/xattr_acl.c 			entry->e_id = cpu_to_le32(
entry              99 fs/select.c    	struct poll_table_entry * entry;
entry             104 fs/select.c    	((unsigned long)((table)->entry+1) > PAGE_SIZE + (unsigned long)(table))
entry             132 fs/select.c    static void free_poll_entry(struct poll_table_entry *entry)
entry             134 fs/select.c    	remove_wait_queue(entry->wait_address, &entry->wait);
entry             135 fs/select.c    	fput(entry->filp);
entry             145 fs/select.c    		struct poll_table_entry * entry;
entry             148 fs/select.c    		entry = p->entry;
entry             150 fs/select.c    			entry--;
entry             151 fs/select.c    			free_poll_entry(entry);
entry             152 fs/select.c    		} while (entry > p->entries);
entry             175 fs/select.c    		new_table->entry = new_table->entries;
entry             181 fs/select.c    	return table->entry++;
entry             212 fs/select.c    	struct poll_table_entry *entry;
entry             214 fs/select.c    	entry = container_of(wait, struct poll_table_entry, wait);
entry             215 fs/select.c    	if (key && !(key_to_poll(key) & entry->key))
entry             225 fs/select.c    	struct poll_table_entry *entry = poll_get_entry(pwq);
entry             226 fs/select.c    	if (!entry)
entry             228 fs/select.c    	entry->filp = get_file(filp);
entry             229 fs/select.c    	entry->wait_address = wait_address;
entry             230 fs/select.c    	entry->key = p->_key;
entry             231 fs/select.c    	init_waitqueue_func_entry(&entry->wait, pollwake);
entry             232 fs/select.c    	entry->wait.private = pwq;
entry             233 fs/select.c    	add_wait_queue(wait_address, &entry->wait);
entry              56 fs/squashfs/cache.c 	struct squashfs_cache_entry *entry;
entry              62 fs/squashfs/cache.c 			if (cache->entry[i].block == block) {
entry              90 fs/squashfs/cache.c 				if (cache->entry[i].refcount == 0)
entry              96 fs/squashfs/cache.c 			entry = &cache->entry[i];
entry             103 fs/squashfs/cache.c 			entry->block = block;
entry             104 fs/squashfs/cache.c 			entry->refcount = 1;
entry             105 fs/squashfs/cache.c 			entry->pending = 1;
entry             106 fs/squashfs/cache.c 			entry->num_waiters = 0;
entry             107 fs/squashfs/cache.c 			entry->error = 0;
entry             110 fs/squashfs/cache.c 			entry->length = squashfs_read_data(sb, block, length,
entry             111 fs/squashfs/cache.c 				&entry->next_index, entry->actor);
entry             115 fs/squashfs/cache.c 			if (entry->length < 0)
entry             116 fs/squashfs/cache.c 				entry->error = entry->length;
entry             118 fs/squashfs/cache.c 			entry->pending = 0;
entry             125 fs/squashfs/cache.c 			if (entry->num_waiters) {
entry             127 fs/squashfs/cache.c 				wake_up_all(&entry->wait_queue);
entry             140 fs/squashfs/cache.c 		entry = &cache->entry[i];
entry             141 fs/squashfs/cache.c 		if (entry->refcount == 0)
entry             143 fs/squashfs/cache.c 		entry->refcount++;
entry             149 fs/squashfs/cache.c 		if (entry->pending) {
entry             150 fs/squashfs/cache.c 			entry->num_waiters++;
entry             152 fs/squashfs/cache.c 			wait_event(entry->wait_queue, !entry->pending);
entry             161 fs/squashfs/cache.c 		cache->name, i, entry->block, entry->refcount, entry->error);
entry             163 fs/squashfs/cache.c 	if (entry->error)
entry             166 fs/squashfs/cache.c 	return entry;
entry             173 fs/squashfs/cache.c void squashfs_cache_put(struct squashfs_cache_entry *entry)
entry             175 fs/squashfs/cache.c 	struct squashfs_cache *cache = entry->cache;
entry             178 fs/squashfs/cache.c 	entry->refcount--;
entry             179 fs/squashfs/cache.c 	if (entry->refcount == 0) {
entry             205 fs/squashfs/cache.c 		if (cache->entry[i].data) {
entry             207 fs/squashfs/cache.c 				kfree(cache->entry[i].data[j]);
entry             208 fs/squashfs/cache.c 			kfree(cache->entry[i].data);
entry             210 fs/squashfs/cache.c 		kfree(cache->entry[i].actor);
entry             213 fs/squashfs/cache.c 	kfree(cache->entry);
entry             234 fs/squashfs/cache.c 	cache->entry = kcalloc(entries, sizeof(*(cache->entry)), GFP_KERNEL);
entry             235 fs/squashfs/cache.c 	if (cache->entry == NULL) {
entry             253 fs/squashfs/cache.c 		struct squashfs_cache_entry *entry = &cache->entry[i];
entry             255 fs/squashfs/cache.c 		init_waitqueue_head(&cache->entry[i].wait_queue);
entry             256 fs/squashfs/cache.c 		entry->cache = cache;
entry             257 fs/squashfs/cache.c 		entry->block = SQUASHFS_INVALID_BLK;
entry             258 fs/squashfs/cache.c 		entry->data = kcalloc(cache->pages, sizeof(void *), GFP_KERNEL);
entry             259 fs/squashfs/cache.c 		if (entry->data == NULL) {
entry             265 fs/squashfs/cache.c 			entry->data[j] = kmalloc(PAGE_SIZE, GFP_KERNEL);
entry             266 fs/squashfs/cache.c 			if (entry->data[j] == NULL) {
entry             272 fs/squashfs/cache.c 		entry->actor = squashfs_page_actor_init(entry->data,
entry             274 fs/squashfs/cache.c 		if (entry->actor == NULL) {
entry             293 fs/squashfs/cache.c int squashfs_copy_data(void *buffer, struct squashfs_cache_entry *entry,
entry             301 fs/squashfs/cache.c 		return min(length, entry->length - offset);
entry             303 fs/squashfs/cache.c 	while (offset < entry->length) {
entry             304 fs/squashfs/cache.c 		void *buff = entry->data[offset / PAGE_SIZE]
entry             306 fs/squashfs/cache.c 		int bytes = min_t(int, entry->length - offset,
entry             336 fs/squashfs/cache.c 	struct squashfs_cache_entry *entry;
entry             344 fs/squashfs/cache.c 		entry = squashfs_cache_get(sb, msblk->block_cache, *block, 0);
entry             345 fs/squashfs/cache.c 		if (entry->error) {
entry             346 fs/squashfs/cache.c 			res = entry->error;
entry             348 fs/squashfs/cache.c 		} else if (*offset >= entry->length) {
entry             353 fs/squashfs/cache.c 		bytes = squashfs_copy_data(buffer, entry, *offset, length);
entry             359 fs/squashfs/cache.c 		if (*offset == entry->length) {
entry             360 fs/squashfs/cache.c 			*block = entry->next_index;
entry             364 fs/squashfs/cache.c 		squashfs_cache_put(entry);
entry             370 fs/squashfs/cache.c 	squashfs_cache_put(entry);
entry              26 fs/squashfs/squashfs_fs_sb.h 	struct squashfs_cache_entry *entry;
entry              44 fs/squashfs/symlink.c 	struct squashfs_cache_entry *entry;
entry              71 fs/squashfs/symlink.c 		entry = squashfs_cache_get(sb, msblk->block_cache, block, 0);
entry              72 fs/squashfs/symlink.c 		if (entry->error) {
entry              76 fs/squashfs/symlink.c 			squashfs_cache_put(entry);
entry              81 fs/squashfs/symlink.c 		copied = squashfs_copy_data(pageaddr + bytes, entry, offset,
entry              86 fs/squashfs/symlink.c 			block = entry->next_index;
entry              88 fs/squashfs/symlink.c 		squashfs_cache_put(entry);
entry              45 fs/squashfs/xattr.c 		struct squashfs_xattr_entry entry;
entry              50 fs/squashfs/xattr.c 		err = squashfs_read_metadata(sb, &entry, &start, &offset,
entry              51 fs/squashfs/xattr.c 							sizeof(entry));
entry              55 fs/squashfs/xattr.c 		name_size = le16_to_cpu(entry.size);
entry              56 fs/squashfs/xattr.c 		handler = squashfs_xattr_handler(le16_to_cpu(entry.type));
entry             123 fs/squashfs/xattr.c 		struct squashfs_xattr_entry entry;
entry             127 fs/squashfs/xattr.c 		err = squashfs_read_metadata(sb, &entry, &start, &offset,
entry             128 fs/squashfs/xattr.c 							sizeof(entry));
entry             132 fs/squashfs/xattr.c 		name_size = le16_to_cpu(entry.size);
entry             133 fs/squashfs/xattr.c 		type = le16_to_cpu(entry.type);
entry             429 fs/sysfs/group.c 	struct kernfs_node *entry;
entry             445 fs/sysfs/group.c 	entry = kernfs_find_and_get(target_kobj->sd, target_name);
entry             446 fs/sysfs/group.c 	if (!entry) {
entry             451 fs/sysfs/group.c 	link = kernfs_create_link(kobj->sd, target_name, entry);
entry             455 fs/sysfs/group.c 	kernfs_put(entry);
entry             136 fs/udf/partition.c 			struct sparingEntry *entry = &st->mapEntry[i];
entry             137 fs/udf/partition.c 			u32 origLoc = le32_to_cpu(entry->origLocation);
entry             141 fs/udf/partition.c 				return le32_to_cpu(entry->mappedLocation) +
entry             188 fs/udf/partition.c 				struct sparingEntry *entry = &st->mapEntry[k];
entry             189 fs/udf/partition.c 				u32 origLoc = le32_to_cpu(entry->origLocation);
entry             200 fs/udf/partition.c 						entry->origLocation =
entry             210 fs/udf/partition.c 							entry->mappedLocation) +
entry             218 fs/udf/partition.c 							entry->mappedLocation) +
entry             229 fs/udf/partition.c 				struct sparingEntry *entry = &st->mapEntry[l];
entry             230 fs/udf/partition.c 				u32 origLoc = le32_to_cpu(entry->origLocation);
entry             140 fs/userfaultfd.c 		list_del_init(&wq->entry);
entry             564 fs/userfaultfd.c 	if (!list_empty_careful(&uwq.wq.entry)) {
entry             570 fs/userfaultfd.c 		list_del(&uwq.wq.entry);
entry             958 fs/userfaultfd.c 	wq = list_last_entry(&wqh->head, typeof(*wq), entry);
entry            1089 fs/userfaultfd.c 			list_del(&uwq->wq.entry);
entry            1111 fs/userfaultfd.c 				list_move(&uwq->wq.entry, &fork_event);
entry            1158 fs/userfaultfd.c 					       wq.entry);
entry            1169 fs/userfaultfd.c 			list_del(&uwq->wq.entry);
entry            1902 fs/userfaultfd.c 	list_for_each_entry(wq, &ctx->fault_pending_wqh.head, entry) {
entry            1906 fs/userfaultfd.c 	list_for_each_entry(wq, &ctx->fault_wqh.head, entry) {
entry             883 fs/xfs/libxfs/xfs_attr_leaf.c 	struct xfs_attr_leaf_entry *entry;
entry             892 fs/xfs/libxfs/xfs_attr_leaf.c 	entry = xfs_attr3_leaf_entryp(leaf);
entry             895 fs/xfs/libxfs/xfs_attr_leaf.c 	for (i = 0; i < leafhdr.count; entry++, i++) {
entry             896 fs/xfs/libxfs/xfs_attr_leaf.c 		if (entry->flags & XFS_ATTR_INCOMPLETE)
entry             898 fs/xfs/libxfs/xfs_attr_leaf.c 		if (!(entry->flags & XFS_ATTR_LOCAL))
entry            1001 fs/xfs/libxfs/xfs_attr_leaf.c 	struct xfs_attr_leaf_entry *entry;
entry            1019 fs/xfs/libxfs/xfs_attr_leaf.c 	entry = xfs_attr3_leaf_entryp(leaf);
entry            1051 fs/xfs/libxfs/xfs_attr_leaf.c 	for (i = 0; i < ichdr.count; entry++, i++) {
entry            1052 fs/xfs/libxfs/xfs_attr_leaf.c 		if (entry->flags & XFS_ATTR_INCOMPLETE)
entry            1054 fs/xfs/libxfs/xfs_attr_leaf.c 		if (!entry->nameidx)
entry            1056 fs/xfs/libxfs/xfs_attr_leaf.c 		ASSERT(entry->flags & XFS_ATTR_LOCAL);
entry            1062 fs/xfs/libxfs/xfs_attr_leaf.c 		nargs.hashval = be32_to_cpu(entry->hashval);
entry            1063 fs/xfs/libxfs/xfs_attr_leaf.c 		nargs.flags = XFS_ATTR_NSP_ONDISK_TO_ARGS(entry->flags);
entry            1348 fs/xfs/libxfs/xfs_attr_leaf.c 	struct xfs_attr_leaf_entry *entry;
entry            1364 fs/xfs/libxfs/xfs_attr_leaf.c 	entry = &xfs_attr3_leaf_entryp(leaf)[args->index];
entry            1368 fs/xfs/libxfs/xfs_attr_leaf.c 		memmove(entry + 1, entry, tmp);
entry            1370 fs/xfs/libxfs/xfs_attr_leaf.c 		    XFS_DA_LOGRANGE(leaf, entry, tmp + sizeof(*entry)));
entry            1387 fs/xfs/libxfs/xfs_attr_leaf.c 	entry->nameidx = cpu_to_be16(ichdr->freemap[mapindex].base +
entry            1389 fs/xfs/libxfs/xfs_attr_leaf.c 	entry->hashval = cpu_to_be32(args->hashval);
entry            1390 fs/xfs/libxfs/xfs_attr_leaf.c 	entry->flags = tmp ? XFS_ATTR_LOCAL : 0;
entry            1391 fs/xfs/libxfs/xfs_attr_leaf.c 	entry->flags |= XFS_ATTR_NSP_ARGS_TO_ONDISK(args->flags);
entry            1393 fs/xfs/libxfs/xfs_attr_leaf.c 		entry->flags |= XFS_ATTR_INCOMPLETE;
entry            1400 fs/xfs/libxfs/xfs_attr_leaf.c 			  XFS_DA_LOGRANGE(leaf, entry, sizeof(*entry)));
entry            1402 fs/xfs/libxfs/xfs_attr_leaf.c 	       (be32_to_cpu(entry->hashval) >= be32_to_cpu((entry-1)->hashval)));
entry            1404 fs/xfs/libxfs/xfs_attr_leaf.c 	       (be32_to_cpu(entry->hashval) <= be32_to_cpu((entry+1)->hashval)));
entry            1413 fs/xfs/libxfs/xfs_attr_leaf.c 	if (entry->flags & XFS_ATTR_LOCAL) {
entry            1424 fs/xfs/libxfs/xfs_attr_leaf.c 		entry->flags |= XFS_ATTR_INCOMPLETE;
entry            1439 fs/xfs/libxfs/xfs_attr_leaf.c 	if (be16_to_cpu(entry->nameidx) < ichdr->firstused)
entry            1440 fs/xfs/libxfs/xfs_attr_leaf.c 		ichdr->firstused = be16_to_cpu(entry->nameidx);
entry            1769 fs/xfs/libxfs/xfs_attr_leaf.c 	struct xfs_attr_leaf_entry	*entry;
entry            1784 fs/xfs/libxfs/xfs_attr_leaf.c 	half = (max + 1) * sizeof(*entry);
entry            1789 fs/xfs/libxfs/xfs_attr_leaf.c 	entry = xfs_attr3_leaf_entryp(leaf1);
entry            1790 fs/xfs/libxfs/xfs_attr_leaf.c 	for (count = index = 0; count < max; entry++, index++, count++) {
entry            1797 fs/xfs/libxfs/xfs_attr_leaf.c 			tmp = totallen + sizeof(*entry) +
entry            1811 fs/xfs/libxfs/xfs_attr_leaf.c 			entry = xfs_attr3_leaf_entryp(leaf1);
entry            1818 fs/xfs/libxfs/xfs_attr_leaf.c 		tmp = totallen + sizeof(*entry) + xfs_attr_leaf_entsize(leaf1,
entry            1831 fs/xfs/libxfs/xfs_attr_leaf.c 	totallen -= count * sizeof(*entry);
entry            1833 fs/xfs/libxfs/xfs_attr_leaf.c 		totallen -= sizeof(*entry) +
entry            1991 fs/xfs/libxfs/xfs_attr_leaf.c 	struct xfs_attr_leaf_entry *entry;
entry            2007 fs/xfs/libxfs/xfs_attr_leaf.c 	ASSERT(ichdr.firstused >= ichdr.count * sizeof(*entry) +
entry            2010 fs/xfs/libxfs/xfs_attr_leaf.c 	entry = &xfs_attr3_leaf_entryp(leaf)[args->index];
entry            2012 fs/xfs/libxfs/xfs_attr_leaf.c 	ASSERT(be16_to_cpu(entry->nameidx) >= ichdr.firstused);
entry            2013 fs/xfs/libxfs/xfs_attr_leaf.c 	ASSERT(be16_to_cpu(entry->nameidx) < args->geo->blksize);
entry            2036 fs/xfs/libxfs/xfs_attr_leaf.c 				be16_to_cpu(entry->nameidx)) {
entry            2039 fs/xfs/libxfs/xfs_attr_leaf.c 				(be16_to_cpu(entry->nameidx) + entsize)) {
entry            2060 fs/xfs/libxfs/xfs_attr_leaf.c 			ichdr.freemap[after].base = be16_to_cpu(entry->nameidx);
entry            2068 fs/xfs/libxfs/xfs_attr_leaf.c 			ichdr.freemap[smallest].base = be16_to_cpu(entry->nameidx);
entry            2076 fs/xfs/libxfs/xfs_attr_leaf.c 	if (be16_to_cpu(entry->nameidx) == ichdr.firstused)
entry            2091 fs/xfs/libxfs/xfs_attr_leaf.c 	memmove(entry, entry + 1, tmp);
entry            2094 fs/xfs/libxfs/xfs_attr_leaf.c 	    XFS_DA_LOGRANGE(leaf, entry, tmp + sizeof(xfs_attr_leaf_entry_t)));
entry            2096 fs/xfs/libxfs/xfs_attr_leaf.c 	entry = &xfs_attr3_leaf_entryp(leaf)[ichdr.count];
entry            2097 fs/xfs/libxfs/xfs_attr_leaf.c 	memset(entry, 0, sizeof(xfs_attr_leaf_entry_t));
entry            2107 fs/xfs/libxfs/xfs_attr_leaf.c 		entry = xfs_attr3_leaf_entryp(leaf);
entry            2108 fs/xfs/libxfs/xfs_attr_leaf.c 		for (i = ichdr.count - 1; i >= 0; entry++, i--) {
entry            2109 fs/xfs/libxfs/xfs_attr_leaf.c 			ASSERT(be16_to_cpu(entry->nameidx) >= ichdr.firstused);
entry            2110 fs/xfs/libxfs/xfs_attr_leaf.c 			ASSERT(be16_to_cpu(entry->nameidx) < args->geo->blksize);
entry            2112 fs/xfs/libxfs/xfs_attr_leaf.c 			if (be16_to_cpu(entry->nameidx) < tmp)
entry            2113 fs/xfs/libxfs/xfs_attr_leaf.c 				tmp = be16_to_cpu(entry->nameidx);
entry            2148 fs/xfs/libxfs/xfs_attr_leaf.c 	struct xfs_attr_leaf_entry *entry;
entry            2156 fs/xfs/libxfs/xfs_attr_leaf.c 	entry = xfs_attr3_leaf_entryp(drop_leaf);
entry            2161 fs/xfs/libxfs/xfs_attr_leaf.c 	drop_blk->hashval = be32_to_cpu(entry[drophdr.count - 1].hashval);
entry            2243 fs/xfs/libxfs/xfs_attr_leaf.c 	entry = xfs_attr3_leaf_entryp(save_leaf);
entry            2244 fs/xfs/libxfs/xfs_attr_leaf.c 	save_blk->hashval = be32_to_cpu(entry[savehdr.count - 1].hashval);
entry            2271 fs/xfs/libxfs/xfs_attr_leaf.c 	struct xfs_attr_leaf_entry *entry;
entry            2292 fs/xfs/libxfs/xfs_attr_leaf.c 	for (entry = &entries[probe]; span > 4; entry = &entries[probe]) {
entry            2294 fs/xfs/libxfs/xfs_attr_leaf.c 		if (be32_to_cpu(entry->hashval) < hashval)
entry            2296 fs/xfs/libxfs/xfs_attr_leaf.c 		else if (be32_to_cpu(entry->hashval) > hashval)
entry            2303 fs/xfs/libxfs/xfs_attr_leaf.c 	if (!(span <= 4 || be32_to_cpu(entry->hashval) == hashval))
entry            2310 fs/xfs/libxfs/xfs_attr_leaf.c 	while (probe > 0 && be32_to_cpu(entry->hashval) >= hashval) {
entry            2311 fs/xfs/libxfs/xfs_attr_leaf.c 		entry--;
entry            2315 fs/xfs/libxfs/xfs_attr_leaf.c 	       be32_to_cpu(entry->hashval) < hashval) {
entry            2316 fs/xfs/libxfs/xfs_attr_leaf.c 		entry++;
entry            2319 fs/xfs/libxfs/xfs_attr_leaf.c 	if (probe == ichdr.count || be32_to_cpu(entry->hashval) != hashval) {
entry            2327 fs/xfs/libxfs/xfs_attr_leaf.c 	for (; probe < ichdr.count && (be32_to_cpu(entry->hashval) == hashval);
entry            2328 fs/xfs/libxfs/xfs_attr_leaf.c 			entry++, probe++) {
entry            2337 fs/xfs/libxfs/xfs_attr_leaf.c 		    (entry->flags & XFS_ATTR_INCOMPLETE)) {
entry            2340 fs/xfs/libxfs/xfs_attr_leaf.c 		if (entry->flags & XFS_ATTR_LOCAL) {
entry            2347 fs/xfs/libxfs/xfs_attr_leaf.c 			if (!xfs_attr_namesp_match(args->flags, entry->flags))
entry            2358 fs/xfs/libxfs/xfs_attr_leaf.c 			if (!xfs_attr_namesp_match(args->flags, entry->flags))
entry            2388 fs/xfs/libxfs/xfs_attr_leaf.c 	struct xfs_attr_leaf_entry *entry;
entry            2397 fs/xfs/libxfs/xfs_attr_leaf.c 	entry = &xfs_attr3_leaf_entryp(leaf)[args->index];
entry            2398 fs/xfs/libxfs/xfs_attr_leaf.c 	if (entry->flags & XFS_ATTR_LOCAL) {
entry            2649 fs/xfs/libxfs/xfs_attr_leaf.c 	struct xfs_attr_leaf_entry *entry;
entry            2669 fs/xfs/libxfs/xfs_attr_leaf.c 	entry = &xfs_attr3_leaf_entryp(leaf)[args->index];
entry            2670 fs/xfs/libxfs/xfs_attr_leaf.c 	ASSERT(entry->flags & XFS_ATTR_INCOMPLETE);
entry            2677 fs/xfs/libxfs/xfs_attr_leaf.c 	if (entry->flags & XFS_ATTR_LOCAL) {
entry            2686 fs/xfs/libxfs/xfs_attr_leaf.c 	ASSERT(be32_to_cpu(entry->hashval) == args->hashval);
entry            2691 fs/xfs/libxfs/xfs_attr_leaf.c 	entry->flags &= ~XFS_ATTR_INCOMPLETE;
entry            2693 fs/xfs/libxfs/xfs_attr_leaf.c 			 XFS_DA_LOGRANGE(leaf, entry, sizeof(*entry)));
entry            2696 fs/xfs/libxfs/xfs_attr_leaf.c 		ASSERT((entry->flags & XFS_ATTR_LOCAL) == 0);
entry            2718 fs/xfs/libxfs/xfs_attr_leaf.c 	struct xfs_attr_leaf_entry *entry;
entry            2741 fs/xfs/libxfs/xfs_attr_leaf.c 	entry = &xfs_attr3_leaf_entryp(leaf)[args->index];
entry            2743 fs/xfs/libxfs/xfs_attr_leaf.c 	ASSERT((entry->flags & XFS_ATTR_INCOMPLETE) == 0);
entry            2744 fs/xfs/libxfs/xfs_attr_leaf.c 	entry->flags |= XFS_ATTR_INCOMPLETE;
entry            2746 fs/xfs/libxfs/xfs_attr_leaf.c 			XFS_DA_LOGRANGE(leaf, entry, sizeof(*entry)));
entry            2747 fs/xfs/libxfs/xfs_attr_leaf.c 	if ((entry->flags & XFS_ATTR_LOCAL) == 0) {
entry             124 fs/xfs/scrub/dabtree.c 	struct xfs_da_node_entry	*entry;
entry             139 fs/xfs/scrub/dabtree.c 	entry = xchk_da_btree_entry(ds, level - 1, blks[level - 1].index);
entry             140 fs/xfs/scrub/dabtree.c 	parent_hash = be32_to_cpu(entry->hashval);
entry             110 fs/xfs/xfs_attr_inactive.c 	struct xfs_attr_leaf_entry *entry;
entry             128 fs/xfs/xfs_attr_inactive.c 	entry = xfs_attr3_leaf_entryp(leaf);
entry             129 fs/xfs/xfs_attr_inactive.c 	for (i = 0; i < ichdr.count; entry++, i++) {
entry             130 fs/xfs/xfs_attr_inactive.c 		if (be16_to_cpu(entry->nameidx) &&
entry             131 fs/xfs/xfs_attr_inactive.c 		    ((entry->flags & XFS_ATTR_LOCAL) == 0)) {
entry             156 fs/xfs/xfs_attr_inactive.c 	entry = xfs_attr3_leaf_entryp(leaf);
entry             157 fs/xfs/xfs_attr_inactive.c 	for (i = 0; i < ichdr.count; entry++, i++) {
entry             158 fs/xfs/xfs_attr_inactive.c 		if (be16_to_cpu(entry->nameidx) &&
entry             159 fs/xfs/xfs_attr_inactive.c 		    ((entry->flags & XFS_ATTR_LOCAL) == 0)) {
entry             387 fs/xfs/xfs_attr_list.c 	struct xfs_attr_leaf_entry	*entry;
entry             404 fs/xfs/xfs_attr_list.c 		entry = &entries[0];
entry             405 fs/xfs/xfs_attr_list.c 		for (i = 0; i < ichdr.count; entry++, i++) {
entry             406 fs/xfs/xfs_attr_list.c 			if (be32_to_cpu(entry->hashval) == cursor->hashval) {
entry             412 fs/xfs/xfs_attr_list.c 			} else if (be32_to_cpu(entry->hashval) >
entry             423 fs/xfs/xfs_attr_list.c 		entry = &entries[0];
entry             431 fs/xfs/xfs_attr_list.c 	for (; i < ichdr.count; entry++, i++) {
entry             435 fs/xfs/xfs_attr_list.c 		if (be32_to_cpu(entry->hashval) != cursor->hashval) {
entry             436 fs/xfs/xfs_attr_list.c 			cursor->hashval = be32_to_cpu(entry->hashval);
entry             440 fs/xfs/xfs_attr_list.c 		if ((entry->flags & XFS_ATTR_INCOMPLETE) &&
entry             444 fs/xfs/xfs_attr_list.c 		if (entry->flags & XFS_ATTR_LOCAL) {
entry             460 fs/xfs/xfs_attr_list.c 		context->put_listent(context, entry->flags,
entry             209 include/acpi/acpi_bus.h 	struct proc_dir_entry *entry;
entry             212 include/acpi/acpi_bus.h #define acpi_device_dir(d)	((d)->dir.entry)
entry              89 include/acpi/actbl3.h 	u8 entry[1];		/* Real size = localities^2 */
entry              32 include/asm-generic/pgtable.h 				 pte_t entry, int dirty);
entry              39 include/asm-generic/pgtable.h 				 pmd_t entry, int dirty);
entry              42 include/asm-generic/pgtable.h 				 pud_t entry, int dirty);
entry              46 include/asm-generic/pgtable.h 					pmd_t entry, int dirty)
entry              53 include/asm-generic/pgtable.h 					pud_t entry, int dirty)
entry             559 include/asm-generic/vmlinux.lds.h 		*(.entry.text)						\
entry             359 include/drm/drm_mm.h #define drm_mm_for_each_node(entry, mm) \
entry             360 include/drm/drm_mm.h 	list_for_each_entry(entry, drm_mm_nodes(mm), node_list)
entry             371 include/drm/drm_mm.h #define drm_mm_for_each_node_safe(entry, next, mm) \
entry             372 include/drm/drm_mm.h 	list_for_each_entry_safe(entry, next, drm_mm_nodes(mm), node_list)
entry              35 include/drm/drm_os_linux.h 	DECLARE_WAITQUEUE(entry, current);			\
entry              37 include/drm/drm_os_linux.h 	add_wait_queue(&(queue), &entry);			\
entry              54 include/drm/drm_os_linux.h 	remove_wait_queue(&(queue), &entry);			\
entry             211 include/linux/acpi.h #define BAD_MADT_ENTRY(entry, end) (					    \
entry             212 include/linux/acpi.h 		(!entry) || (unsigned long)entry + sizeof(*entry) > end ||  \
entry             213 include/linux/acpi.h 		((struct acpi_subtable_header *)entry)->length < sizeof(*entry))
entry             142 include/linux/atmdev.h 	struct list_head entry;		/* next address */
entry             114 include/linux/backlight.h 	struct list_head entry;
entry             640 include/linux/ccp.h 	struct list_head entry;
entry             675 include/linux/compat.h asmlinkage long compat_sys_kexec_load(compat_ulong_t entry,
entry             856 include/linux/device.h 			      const struct device_attribute *entry);
entry             955 include/linux/efi.h 	efi_memory_desc_t entry[0];
entry            1489 include/linux/efi.h int efivar_entry_add(struct efivar_entry *entry, struct list_head *head);
entry            1490 include/linux/efi.h int efivar_entry_remove(struct efivar_entry *entry);
entry            1492 include/linux/efi.h int __efivar_entry_delete(struct efivar_entry *entry);
entry            1493 include/linux/efi.h int efivar_entry_delete(struct efivar_entry *entry);
entry            1495 include/linux/efi.h int efivar_entry_size(struct efivar_entry *entry, unsigned long *size);
entry            1496 include/linux/efi.h int __efivar_entry_get(struct efivar_entry *entry, u32 *attributes,
entry            1498 include/linux/efi.h int efivar_entry_get(struct efivar_entry *entry, u32 *attributes,
entry            1500 include/linux/efi.h int efivar_entry_set(struct efivar_entry *entry, u32 attributes,
entry            1502 include/linux/efi.h int efivar_entry_set_get_size(struct efivar_entry *entry, u32 attributes,
entry            1787 include/linux/efi.h 	} entry[0];
entry            1791 include/linux/efi.h 	(count) * sizeof(((struct linux_efi_memreserve *)0)->entry[0]))
entry            1794 include/linux/efi.h 	/ sizeof(((struct linux_efi_memreserve *)0)->entry[0]))
entry              23 include/linux/flat.h 	__be32	entry;        /* Offset of first executable instruction
entry             208 include/linux/hmm.h 						    uint64_t entry)
entry             210 include/linux/hmm.h 	if (entry == range->values[HMM_PFN_NONE])
entry             212 include/linux/hmm.h 	if (entry == range->values[HMM_PFN_ERROR])
entry             214 include/linux/hmm.h 	if (entry == range->values[HMM_PFN_SPECIAL])
entry             216 include/linux/hmm.h 	if (!(entry & range->flags[HMM_PFN_VALID]))
entry             218 include/linux/hmm.h 	return pfn_to_page(entry >> range->pfn_shift);
entry             449 include/linux/hugetlb.h static inline pte_t arch_make_huge_pte(pte_t entry, struct vm_area_struct *vma,
entry             452 include/linux/hugetlb.h 	return entry;
entry             187 include/linux/idr.h #define idr_for_each_entry(idr, entry, id)			\
entry             188 include/linux/idr.h 	for (id = 0; ((entry) = idr_get_next(idr, &(id))) != NULL; id += 1U)
entry             201 include/linux/idr.h #define idr_for_each_entry_ul(idr, entry, tmp, id)			\
entry             203 include/linux/idr.h 	     tmp <= id && ((entry) = idr_get_next_ul(idr, &(id))) != NULL; \
entry             214 include/linux/idr.h #define idr_for_each_entry_continue(idr, entry, id)			\
entry             215 include/linux/idr.h 	for ((entry) = idr_get_next((idr), &(id));			\
entry             216 include/linux/idr.h 	     entry;							\
entry             217 include/linux/idr.h 	     ++id, (entry) = idr_get_next((idr), &(id)))
entry             228 include/linux/idr.h #define idr_for_each_entry_continue_ul(idr, entry, tmp, id)		\
entry             230 include/linux/idr.h 	     tmp <= id && ((entry) = idr_get_next_ul(idr, &(id))) != NULL; \
entry             122 include/linux/init.h static inline initcall_t initcall_from_entry(initcall_entry_t *entry)
entry             124 include/linux/init.h 	return offset_to_ptr(entry);
entry             129 include/linux/init.h static inline initcall_t initcall_from_entry(initcall_entry_t *entry)
entry             131 include/linux/init.h 	return *entry;
entry             762 include/linux/irq.h extern int irq_set_msi_desc(unsigned int irq, struct msi_desc *entry);
entry             764 include/linux/irq.h 				struct msi_desc *entry);
entry             128 include/linux/jump_label.h static inline unsigned long jump_entry_code(const struct jump_entry *entry)
entry             130 include/linux/jump_label.h 	return (unsigned long)&entry->code + entry->code;
entry             133 include/linux/jump_label.h static inline unsigned long jump_entry_target(const struct jump_entry *entry)
entry             135 include/linux/jump_label.h 	return (unsigned long)&entry->target + entry->target;
entry             138 include/linux/jump_label.h static inline struct static_key *jump_entry_key(const struct jump_entry *entry)
entry             140 include/linux/jump_label.h 	long offset = entry->key & ~3L;
entry             142 include/linux/jump_label.h 	return (struct static_key *)((unsigned long)&entry->key + offset);
entry             147 include/linux/jump_label.h static inline unsigned long jump_entry_code(const struct jump_entry *entry)
entry             149 include/linux/jump_label.h 	return entry->code;
entry             152 include/linux/jump_label.h static inline unsigned long jump_entry_target(const struct jump_entry *entry)
entry             154 include/linux/jump_label.h 	return entry->target;
entry             157 include/linux/jump_label.h static inline struct static_key *jump_entry_key(const struct jump_entry *entry)
entry             159 include/linux/jump_label.h 	return (struct static_key *)((unsigned long)entry->key & ~3UL);
entry             164 include/linux/jump_label.h static inline bool jump_entry_is_branch(const struct jump_entry *entry)
entry             166 include/linux/jump_label.h 	return (unsigned long)entry->key & 1UL;
entry             169 include/linux/jump_label.h static inline bool jump_entry_is_init(const struct jump_entry *entry)
entry             171 include/linux/jump_label.h 	return (unsigned long)entry->key & 2UL;
entry             174 include/linux/jump_label.h static inline void jump_entry_set_init(struct jump_entry *entry)
entry             176 include/linux/jump_label.h 	entry->key |= 2;
entry             214 include/linux/jump_label.h extern void arch_jump_label_transform(struct jump_entry *entry,
entry             216 include/linux/jump_label.h extern void arch_jump_label_transform_static(struct jump_entry *entry,
entry             218 include/linux/jump_label.h extern bool arch_jump_label_transform_queue(struct jump_entry *entry,
entry             246 include/linux/kexec.h 	kimage_entry_t *entry;
entry             380 include/linux/kexec.h static inline void *boot_phys_to_virt(unsigned long entry)
entry             382 include/linux/kexec.h 	return phys_to_virt(boot_phys_to_phys(entry));
entry              67 include/linux/kobject.h 	struct list_head	entry;
entry             238 include/linux/kprobes.h extern int kprobe_add_ksym_blacklist(unsigned long entry);
entry              36 include/linux/list.h extern bool __list_del_entry_valid(struct list_head *entry);
entry              44 include/linux/list.h static inline bool __list_del_entry_valid(struct list_head *entry)
entry             117 include/linux/list.h static inline void __list_del_clearprev(struct list_head *entry)
entry             119 include/linux/list.h 	__list_del(entry->prev, entry->next);
entry             120 include/linux/list.h 	entry->prev = NULL;
entry             129 include/linux/list.h static inline void __list_del_entry(struct list_head *entry)
entry             131 include/linux/list.h 	if (!__list_del_entry_valid(entry))
entry             134 include/linux/list.h 	__list_del(entry->prev, entry->next);
entry             137 include/linux/list.h static inline void list_del(struct list_head *entry)
entry             139 include/linux/list.h 	__list_del_entry(entry);
entry             140 include/linux/list.h 	entry->next = LIST_POISON1;
entry             141 include/linux/list.h 	entry->prev = LIST_POISON2;
entry             188 include/linux/list.h static inline void list_del_init(struct list_head *entry)
entry             190 include/linux/list.h 	__list_del_entry(entry);
entry             191 include/linux/list.h 	INIT_LIST_HEAD(entry);
entry             332 include/linux/list.h 		struct list_head *head, struct list_head *entry)
entry             334 include/linux/list.h 	struct list_head *new_first = entry->next;
entry             337 include/linux/list.h 	list->prev = entry;
entry             338 include/linux/list.h 	entry->next = list;
entry             358 include/linux/list.h 		struct list_head *head, struct list_head *entry)
entry             363 include/linux/list.h 		(head->next != entry && head != entry))
entry             365 include/linux/list.h 	if (entry == head)
entry             368 include/linux/list.h 		__list_cut_position(list, head, entry);
entry             387 include/linux/list.h 				   struct list_head *entry)
entry             389 include/linux/list.h 	if (head->next == entry) {
entry             395 include/linux/list.h 	list->prev = entry->prev;
entry             397 include/linux/list.h 	head->next = entry;
entry             398 include/linux/list.h 	entry->prev = head;
entry             190 include/linux/lockdep.h 	struct list_head		entry;
entry             218 include/linux/lockdep.h 	struct hlist_node		entry;
entry              32 include/linux/mbcache.h void __mb_cache_entry_free(struct mb_cache_entry *entry);
entry              34 include/linux/mbcache.h 				     struct mb_cache_entry *entry)
entry              36 include/linux/mbcache.h 	if (!atomic_dec_and_test(&entry->e_refcnt))
entry              38 include/linux/mbcache.h 	__mb_cache_entry_free(entry);
entry              48 include/linux/mbcache.h 						struct mb_cache_entry *entry);
entry              50 include/linux/mbcache.h 			  struct mb_cache_entry *entry);
entry             143 include/linux/migrate.h 			pmd_t *pmd, pmd_t entry,
entry             149 include/linux/migrate.h 			pmd_t *pmd, pmd_t entry,
entry            1446 include/linux/mlx4/device.h void mlx4_set_admin_guid(struct mlx4_dev *dev, __be64 guid, int entry,
entry            1448 include/linux/mlx4/device.h __be64 mlx4_get_admin_guid(struct mlx4_dev *dev, int entry, int port);
entry            1449 include/linux/mlx4/device.h void mlx4_set_random_admin_guid(struct mlx4_dev *dev, int entry, int port);
entry              20 include/linux/msi.h void __get_cached_msi_msg(struct msi_desc *entry, struct msi_msg *msg);
entry              97 include/linux/msi.h 	void (*write_msi_msg)(struct msi_desc *entry, void *data);
entry             186 include/linux/msi.h void free_msi_entry(struct msi_desc *entry);
entry             187 include/linux/msi.h void __pci_read_msi_msg(struct msi_desc *entry, struct msi_msg *msg);
entry             188 include/linux/msi.h void __pci_write_msi_msg(struct msi_desc *entry, struct msi_msg *msg);
entry             725 include/linux/mtd/nand.h 				 unsigned int entry);
entry             726 include/linux/mtd/nand.h int nanddev_bbt_set_block_status(struct nand_device *nand, unsigned int entry,
entry             132 include/linux/netfilter.h nf_hook_entry_hookfn(const struct nf_hook_entry *entry, struct sk_buff *skb,
entry             135 include/linux/netfilter.h 	return entry->hook(entry->priv, skb, state);
entry             351 include/linux/netfilter.h int nf_reroute(struct sk_buff *skb, struct nf_queue_entry *entry);
entry              20 include/linux/netfilter_arp/arp_tables.h 	struct arpt_entry entry;
entry              25 include/linux/netfilter_arp/arp_tables.h 	struct arpt_entry entry;
entry              37 include/linux/netfilter_arp/arp_tables.h 	.entry		= ARPT_ENTRY_INIT(sizeof(struct arpt_standard)),       \
entry              45 include/linux/netfilter_arp/arp_tables.h 	.entry		= ARPT_ENTRY_INIT(sizeof(struct arpt_error)),	       \
entry              27 include/linux/netfilter_bridge/ebtables.h 	bool (*checkentry)(const char *table, const void *entry,
entry              44 include/linux/netfilter_bridge/ebtables.h 	bool (*checkentry)(const char *table, const void *entry,
entry              62 include/linux/netfilter_bridge/ebtables.h 	bool (*checkentry)(const char *table, const void *entry,
entry              33 include/linux/netfilter_ipv4/ip_tables.h 	struct ipt_entry entry;
entry              38 include/linux/netfilter_ipv4/ip_tables.h 	struct ipt_entry entry;
entry              50 include/linux/netfilter_ipv4/ip_tables.h 	.entry		= IPT_ENTRY_INIT(sizeof(struct ipt_standard)),	       \
entry              58 include/linux/netfilter_ipv4/ip_tables.h 	.entry		= IPT_ENTRY_INIT(sizeof(struct ipt_error)),	       \
entry              59 include/linux/netfilter_ipv6.h 	int (*reroute)(struct sk_buff *skb, const struct nf_queue_entry *entry);
entry             189 include/linux/oprofile.h void oprofile_write_reserve(struct op_entry *entry,
entry             192 include/linux/oprofile.h int oprofile_add_data(struct op_entry *entry, unsigned long val);
entry             193 include/linux/oprofile.h int oprofile_add_data64(struct op_entry *entry, u64 val);
entry             194 include/linux/oprofile.h int oprofile_write_commit(struct op_entry *entry);
entry            1429 include/linux/pci.h 	u16	entry;	/* Driver uses to specify entry, OS writes */
entry              67 include/linux/perf_event.h 	struct perf_callchain_entry *entry;
entry             255 include/linux/perf_event.h 	struct list_head		entry;
entry             488 include/linux/perf_event.h 	struct list_head	entry;
entry            1190 include/linux/perf_event.h extern void perf_callchain_user(struct perf_callchain_entry_ctx *entry, struct pt_regs *regs);
entry            1191 include/linux/perf_event.h extern void perf_callchain_kernel(struct perf_callchain_entry_ctx *entry, struct pt_regs *regs);
entry            1205 include/linux/perf_event.h 		struct perf_callchain_entry *entry = ctx->entry;
entry            1206 include/linux/perf_event.h 		entry->ip[entry->nr++] = ip;
entry            1218 include/linux/perf_event.h 		struct perf_callchain_entry *entry = ctx->entry;
entry            1219 include/linux/perf_event.h 		entry->ip[entry->nr++] = ip;
entry             588 include/linux/pm.h 	struct list_head	entry;
entry              46 include/linux/pm_wakeup.h 	struct list_head	entry;
entry             229 include/linux/power/charger-manager.h 	struct list_head entry;
entry              38 include/linux/qcom_scm.h extern int qcom_scm_set_cold_boot_addr(void *entry, const cpumask_t *cpus);
entry              39 include/linux/qcom_scm.h extern int qcom_scm_set_warm_boot_addr(void *entry, const cpumask_t *cpus);
entry              68 include/linux/qcom_scm.h int qcom_scm_set_cold_boot_addr(void *entry, const cpumask_t *cpus)
entry              73 include/linux/qcom_scm.h int qcom_scm_set_warm_boot_addr(void *entry, const cpumask_t *cpus)
entry             215 include/linux/radix-tree.h 			  void __rcu **slot, void *entry);
entry             217 include/linux/radix-tree.h 		const struct radix_tree_iter *, void __rcu **slot, void *entry);
entry             219 include/linux/radix-tree.h 			     void __rcu **slot, void *entry);
entry             146 include/linux/rculist.h static inline void list_del_rcu(struct list_head *entry)
entry             148 include/linux/rculist.h 	__list_del_entry(entry);
entry             149 include/linux/rculist.h 	entry->prev = LIST_POISON2;
entry              34 include/linux/resource_ext.h static inline void resource_list_add(struct resource_entry *entry,
entry              37 include/linux/resource_ext.h 	list_add(&entry->node, head);
entry              40 include/linux/resource_ext.h static inline void resource_list_add_tail(struct resource_entry *entry,
entry              43 include/linux/resource_ext.h 	list_add_tail(&entry->node, head);
entry              46 include/linux/resource_ext.h static inline void resource_list_del(struct resource_entry *entry)
entry              48 include/linux/resource_ext.h 	list_del(&entry->node);
entry              51 include/linux/resource_ext.h static inline void resource_list_free_entry(struct resource_entry *entry)
entry              53 include/linux/resource_ext.h 	kfree(entry);
entry              57 include/linux/resource_ext.h resource_list_destroy_entry(struct resource_entry *entry)
entry              59 include/linux/resource_ext.h 	resource_list_del(entry);
entry              60 include/linux/resource_ext.h 	resource_list_free_entry(entry);
entry              63 include/linux/resource_ext.h #define resource_list_for_each_entry(entry, list)	\
entry              64 include/linux/resource_ext.h 	list_for_each_entry((entry), (list), node)
entry              66 include/linux/resource_ext.h #define resource_list_for_each_entry_safe(entry, tmp, list)	\
entry              67 include/linux/resource_ext.h 	list_for_each_entry_safe((entry), (tmp), (list), node)
entry             562 include/linux/sbitmap.h 			.entry		= LIST_HEAD_INIT((name).wait.entry),	\
entry             691 include/linux/spi/spi.h 	struct list_head        entry;
entry             405 include/linux/swap.h #define swap_address_space(entry)			    \
entry             406 include/linux/swap.h 	(&swapper_spaces[swp_type(entry)][swp_offset(entry) \
entry             412 include/linux/swap.h extern int __add_to_swap_cache(struct page *page, swp_entry_t entry);
entry             413 include/linux/swap.h extern void __delete_from_swap_cache(struct page *, swp_entry_t entry);
entry             417 include/linux/swap.h extern struct page *lookup_swap_cache(swp_entry_t entry,
entry             426 include/linux/swap.h extern struct page *swap_cluster_readahead(swp_entry_t entry, gfp_t flag,
entry             428 include/linux/swap.h extern struct page *swapin_readahead(swp_entry_t entry, gfp_t flag,
entry             450 include/linux/swap.h extern void put_swap_page(struct page *page, swp_entry_t entry);
entry             465 include/linux/swap.h extern int __swap_count(swp_entry_t entry);
entry             466 include/linux/swap.h extern int __swp_swapcount(swp_entry_t entry);
entry             467 include/linux/swap.h extern int swp_swapcount(swp_entry_t entry);
entry             469 include/linux/swap.h extern struct swap_info_struct *swp_swap_info(swp_entry_t entry);
entry             475 include/linux/swap.h extern struct swap_info_struct *get_swap_device(swp_entry_t entry);
entry             489 include/linux/swap.h static inline struct swap_info_struct *swp_swap_info(swp_entry_t entry)
entry             494 include/linux/swap.h #define swap_address_space(entry)		(NULL)
entry             538 include/linux/swap.h static inline struct page *swap_cluster_readahead(swp_entry_t entry,
entry             567 include/linux/swap.h static inline int add_to_swap_cache(struct page *page, swp_entry_t entry,
entry             574 include/linux/swap.h 							swp_entry_t entry)
entry             587 include/linux/swap.h static inline int __swap_count(swp_entry_t entry)
entry             592 include/linux/swap.h static inline int __swp_swapcount(swp_entry_t entry)
entry             597 include/linux/swap.h static inline int swp_swapcount(swp_entry_t entry)
entry             612 include/linux/swap.h 	swp_entry_t entry;
entry             613 include/linux/swap.h 	entry.val = 0;
entry             614 include/linux/swap.h 	return entry;
entry             620 include/linux/swap.h extern int split_swap_cluster(swp_entry_t entry);
entry             622 include/linux/swap.h static inline int split_swap_cluster(swp_entry_t entry)
entry             659 include/linux/swap.h extern void mem_cgroup_swapout(struct page *page, swp_entry_t entry);
entry             660 include/linux/swap.h extern int mem_cgroup_try_charge_swap(struct page *page, swp_entry_t entry);
entry             661 include/linux/swap.h extern void mem_cgroup_uncharge_swap(swp_entry_t entry, unsigned int nr_pages);
entry             665 include/linux/swap.h static inline void mem_cgroup_swapout(struct page *page, swp_entry_t entry)
entry             670 include/linux/swap.h 					     swp_entry_t entry)
entry             675 include/linux/swap.h static inline void mem_cgroup_uncharge_swap(swp_entry_t entry,
entry              27 include/linux/swap_slots.h int free_swap_slot(swp_entry_t entry);
entry              41 include/linux/swapops.h static inline unsigned swp_type(swp_entry_t entry)
entry              43 include/linux/swapops.h 	return (entry.val >> SWP_TYPE_SHIFT);
entry              50 include/linux/swapops.h static inline pgoff_t swp_offset(swp_entry_t entry)
entry              52 include/linux/swapops.h 	return entry.val & SWP_OFFSET_MASK;
entry              79 include/linux/swapops.h static inline pte_t swp_entry_to_pte(swp_entry_t entry)
entry              83 include/linux/swapops.h 	arch_entry = __swp_entry(swp_type(entry), swp_offset(entry));
entry              89 include/linux/swapops.h 	swp_entry_t entry;
entry              91 include/linux/swapops.h 	entry.val = xa_to_value(arg);
entry              92 include/linux/swapops.h 	return entry;
entry              95 include/linux/swapops.h static inline void *swp_to_radix_entry(swp_entry_t entry)
entry              97 include/linux/swapops.h 	return xa_mk_value(entry.val);
entry             107 include/linux/swapops.h static inline bool is_device_private_entry(swp_entry_t entry)
entry             109 include/linux/swapops.h 	int type = swp_type(entry);
entry             113 include/linux/swapops.h static inline void make_device_private_entry_read(swp_entry_t *entry)
entry             115 include/linux/swapops.h 	*entry = swp_entry(SWP_DEVICE_READ, swp_offset(*entry));
entry             118 include/linux/swapops.h static inline bool is_write_device_private_entry(swp_entry_t entry)
entry             120 include/linux/swapops.h 	return unlikely(swp_type(entry) == SWP_DEVICE_WRITE);
entry             123 include/linux/swapops.h static inline unsigned long device_private_entry_to_pfn(swp_entry_t entry)
entry             125 include/linux/swapops.h 	return swp_offset(entry);
entry             128 include/linux/swapops.h static inline struct page *device_private_entry_to_page(swp_entry_t entry)
entry             130 include/linux/swapops.h 	return pfn_to_page(swp_offset(entry));
entry             138 include/linux/swapops.h static inline void make_device_private_entry_read(swp_entry_t *entry)
entry             142 include/linux/swapops.h static inline bool is_device_private_entry(swp_entry_t entry)
entry             147 include/linux/swapops.h static inline bool is_write_device_private_entry(swp_entry_t entry)
entry             152 include/linux/swapops.h static inline unsigned long device_private_entry_to_pfn(swp_entry_t entry)
entry             157 include/linux/swapops.h static inline struct page *device_private_entry_to_page(swp_entry_t entry)
entry             172 include/linux/swapops.h static inline int is_migration_entry(swp_entry_t entry)
entry             174 include/linux/swapops.h 	return unlikely(swp_type(entry) == SWP_MIGRATION_READ ||
entry             175 include/linux/swapops.h 			swp_type(entry) == SWP_MIGRATION_WRITE);
entry             178 include/linux/swapops.h static inline int is_write_migration_entry(swp_entry_t entry)
entry             180 include/linux/swapops.h 	return unlikely(swp_type(entry) == SWP_MIGRATION_WRITE);
entry             183 include/linux/swapops.h static inline unsigned long migration_entry_to_pfn(swp_entry_t entry)
entry             185 include/linux/swapops.h 	return swp_offset(entry);
entry             188 include/linux/swapops.h static inline struct page *migration_entry_to_page(swp_entry_t entry)
entry             190 include/linux/swapops.h 	struct page *p = pfn_to_page(swp_offset(entry));
entry             199 include/linux/swapops.h static inline void make_migration_entry_read(swp_entry_t *entry)
entry             201 include/linux/swapops.h 	*entry = swp_entry(SWP_MIGRATION_READ, swp_offset(*entry));
entry             218 include/linux/swapops.h static inline unsigned long migration_entry_to_pfn(swp_entry_t entry)
entry             223 include/linux/swapops.h static inline struct page *migration_entry_to_page(swp_entry_t entry)
entry             235 include/linux/swapops.h static inline int is_write_migration_entry(swp_entry_t entry)
entry             263 include/linux/swapops.h static inline pmd_t swp_entry_to_pmd(swp_entry_t entry)
entry             267 include/linux/swapops.h 	arch_entry = __swp_entry(swp_type(entry), swp_offset(entry));
entry             295 include/linux/swapops.h static inline pmd_t swp_entry_to_pmd(swp_entry_t entry)
entry             319 include/linux/swapops.h static inline int is_hwpoison_entry(swp_entry_t entry)
entry             321 include/linux/swapops.h 	return swp_type(entry) == SWP_HWPOISON;
entry             353 include/linux/swapops.h static inline int non_swap_entry(swp_entry_t entry)
entry             355 include/linux/swapops.h 	return swp_type(entry) >= MAX_SWAPFILES;
entry             358 include/linux/swapops.h static inline int non_swap_entry(swp_entry_t entry)
entry             600 include/linux/syscalls.h asmlinkage long sys_kexec_load(unsigned long entry, unsigned long nr_segments,
entry              16 include/linux/timer.h 	struct hlist_node	entry;
entry              76 include/linux/timer.h 		.entry = { .next = TIMER_ENTRY_STATIC },	\
entry             167 include/linux/timer.h 	return timer->entry.pprev != NULL;
entry              74 include/linux/usb/composite.h 	struct list_head	entry;
entry             246 include/linux/usb/usbnet.h 	struct skb_data *entry = (struct skb_data *) skb->cb;
entry             248 include/linux/usb/usbnet.h 	entry->packets = packets;
entry             249 include/linux/usb/usbnet.h 	entry->length = bytes_delta;
entry              80 include/linux/vme.h 	struct list_head *entry;
entry              31 include/linux/wait.h 	struct list_head	entry;
entry              49 include/linux/wait.h 	.entry		= { NULL, NULL } }
entry             169 include/linux/wait.h 	list_add(&wq_entry->entry, &wq_head->head);
entry             184 include/linux/wait.h 	list_add_tail(&wq_entry->entry, &wq_head->head);
entry             197 include/linux/wait.h 	list_del(&wq_entry->entry);
entry            1135 include/linux/wait.h 		.entry		= LIST_HEAD_INIT((name).entry),			\
entry            1144 include/linux/wait.h 		INIT_LIST_HEAD(&(wait)->entry);					\
entry              44 include/linux/wait_bit.h 			.entry		=					\
entry              45 include/linux/wait_bit.h 				LIST_HEAD_INIT((name).wq_entry.entry),		\
entry             104 include/linux/workqueue.h 	struct list_head entry;
entry             186 include/linux/workqueue.h 	.entry	= { &(n).entry, &(n).entry },				\
entry             236 include/linux/workqueue.h 		INIT_LIST_HEAD(&(_work)->entry);			\
entry             244 include/linux/workqueue.h 		INIT_LIST_HEAD(&(_work)->entry);			\
entry              65 include/linux/xarray.h static inline unsigned long xa_to_value(const void *entry)
entry              67 include/linux/xarray.h 	return (unsigned long)entry >> 1;
entry              77 include/linux/xarray.h static inline bool xa_is_value(const void *entry)
entry              79 include/linux/xarray.h 	return (unsigned long)entry & 1;
entry             110 include/linux/xarray.h static inline void *xa_untag_pointer(void *entry)
entry             112 include/linux/xarray.h 	return (void *)((unsigned long)entry & ~3UL);
entry             125 include/linux/xarray.h static inline unsigned int xa_pointer_tag(void *entry)
entry             127 include/linux/xarray.h 	return (unsigned long)entry & 3UL;
entry             155 include/linux/xarray.h static inline unsigned long xa_to_internal(const void *entry)
entry             157 include/linux/xarray.h 	return (unsigned long)entry >> 2;
entry             167 include/linux/xarray.h static inline bool xa_is_internal(const void *entry)
entry             169 include/linux/xarray.h 	return ((unsigned long)entry & 3) == 2;
entry             183 include/linux/xarray.h static inline bool xa_is_zero(const void *entry)
entry             185 include/linux/xarray.h 	return unlikely(entry == XA_ZERO_ENTRY);
entry             199 include/linux/xarray.h static inline bool xa_is_err(const void *entry)
entry             201 include/linux/xarray.h 	return unlikely(xa_is_internal(entry) &&
entry             202 include/linux/xarray.h 			entry >= xa_mk_internal(-MAX_ERRNO));
entry             217 include/linux/xarray.h static inline int xa_err(void *entry)
entry             220 include/linux/xarray.h 	if (xa_is_err(entry))
entry             221 include/linux/xarray.h 		return (long)entry >> 2;
entry             348 include/linux/xarray.h void *xa_store(struct xarray *, unsigned long index, void *entry, gfp_t);
entry             351 include/linux/xarray.h 			void *entry, gfp_t);
entry             442 include/linux/xarray.h #define xa_for_each_start(xa, index, entry, start)			\
entry             444 include/linux/xarray.h 	     entry = xa_find(xa, &index, ULONG_MAX, XA_PRESENT);	\
entry             445 include/linux/xarray.h 	     entry;							\
entry             446 include/linux/xarray.h 	     entry = xa_find_after(xa, &index, ULONG_MAX, XA_PRESENT))
entry             469 include/linux/xarray.h #define xa_for_each(xa, index, entry) \
entry             470 include/linux/xarray.h 	xa_for_each_start(xa, index, entry, 0)
entry             496 include/linux/xarray.h #define xa_for_each_marked(xa, index, entry, filter) \
entry             497 include/linux/xarray.h 	for (index = 0, entry = xa_find(xa, &index, ULONG_MAX, filter); \
entry             498 include/linux/xarray.h 	     entry; entry = xa_find_after(xa, &index, ULONG_MAX, filter))
entry             520 include/linux/xarray.h void *__xa_store(struct xarray *, unsigned long index, void *entry, gfp_t);
entry             522 include/linux/xarray.h 		void *entry, gfp_t);
entry             524 include/linux/xarray.h 		void *entry, gfp_t);
entry             525 include/linux/xarray.h int __must_check __xa_alloc(struct xarray *, u32 *id, void *entry,
entry             527 include/linux/xarray.h int __must_check __xa_alloc_cyclic(struct xarray *, u32 *id, void *entry,
entry             547 include/linux/xarray.h 		void *entry, gfp_t gfp)
entry             552 include/linux/xarray.h 	curr = __xa_store(xa, index, entry, gfp);
entry             573 include/linux/xarray.h 		void *entry, gfp_t gfp)
entry             578 include/linux/xarray.h 	curr = __xa_store(xa, index, entry, gfp);
entry             599 include/linux/xarray.h 	void *entry;
entry             602 include/linux/xarray.h 	entry = __xa_erase(xa, index);
entry             605 include/linux/xarray.h 	return entry;
entry             623 include/linux/xarray.h 	void *entry;
entry             626 include/linux/xarray.h 	entry = __xa_erase(xa, index);
entry             629 include/linux/xarray.h 	return entry;
entry             648 include/linux/xarray.h 			void *old, void *entry, gfp_t gfp)
entry             653 include/linux/xarray.h 	curr = __xa_cmpxchg(xa, index, old, entry, gfp);
entry             675 include/linux/xarray.h 			void *old, void *entry, gfp_t gfp)
entry             680 include/linux/xarray.h 	curr = __xa_cmpxchg(xa, index, old, entry, gfp);
entry             702 include/linux/xarray.h 			void *old, void *entry, gfp_t gfp)
entry             707 include/linux/xarray.h 	curr = __xa_cmpxchg(xa, index, old, entry, gfp);
entry             731 include/linux/xarray.h 		unsigned long index, void *entry, gfp_t gfp)
entry             736 include/linux/xarray.h 	err = __xa_insert(xa, index, entry, gfp);
entry             760 include/linux/xarray.h 		unsigned long index, void *entry, gfp_t gfp)
entry             765 include/linux/xarray.h 	err = __xa_insert(xa, index, entry, gfp);
entry             789 include/linux/xarray.h 		unsigned long index, void *entry, gfp_t gfp)
entry             794 include/linux/xarray.h 	err = __xa_insert(xa, index, entry, gfp);
entry             818 include/linux/xarray.h 		void *entry, struct xa_limit limit, gfp_t gfp)
entry             823 include/linux/xarray.h 	err = __xa_alloc(xa, id, entry, limit, gfp);
entry             847 include/linux/xarray.h 		void *entry, struct xa_limit limit, gfp_t gfp)
entry             852 include/linux/xarray.h 	err = __xa_alloc(xa, id, entry, limit, gfp);
entry             876 include/linux/xarray.h 		void *entry, struct xa_limit limit, gfp_t gfp)
entry             881 include/linux/xarray.h 	err = __xa_alloc(xa, id, entry, limit, gfp);
entry             908 include/linux/xarray.h static inline int xa_alloc_cyclic(struct xarray *xa, u32 *id, void *entry,
entry             914 include/linux/xarray.h 	err = __xa_alloc_cyclic(xa, id, entry, limit, next, gfp);
entry             941 include/linux/xarray.h static inline int xa_alloc_cyclic_bh(struct xarray *xa, u32 *id, void *entry,
entry             947 include/linux/xarray.h 	err = __xa_alloc_cyclic(xa, id, entry, limit, next, gfp);
entry             974 include/linux/xarray.h static inline int xa_alloc_cyclic_irq(struct xarray *xa, u32 *id, void *entry,
entry             980 include/linux/xarray.h 	err = __xa_alloc_cyclic(xa, id, entry, limit, next, gfp);
entry            1181 include/linux/xarray.h static inline struct xa_node *xa_to_node(const void *entry)
entry            1183 include/linux/xarray.h 	return (struct xa_node *)((unsigned long)entry - 2);
entry            1187 include/linux/xarray.h static inline bool xa_is_node(const void *entry)
entry            1189 include/linux/xarray.h 	return xa_is_internal(entry) && (unsigned long)entry > 4096;
entry            1199 include/linux/xarray.h static inline unsigned long xa_to_sibling(const void *entry)
entry            1201 include/linux/xarray.h 	return xa_to_internal(entry);
entry            1210 include/linux/xarray.h static inline bool xa_is_sibling(const void *entry)
entry            1212 include/linux/xarray.h 	return IS_ENABLED(CONFIG_XARRAY_MULTI) && xa_is_internal(entry) &&
entry            1213 include/linux/xarray.h 		(entry < xa_mk_sibling(XA_CHUNK_SIZE - 1));
entry            1224 include/linux/xarray.h static inline bool xa_is_retry(const void *entry)
entry            1226 include/linux/xarray.h 	return unlikely(entry == XA_RETRY_ENTRY);
entry            1235 include/linux/xarray.h static inline bool xa_is_advanced(const void *entry)
entry            1237 include/linux/xarray.h 	return xa_is_internal(entry) && (entry <= XA_RETRY_ENTRY);
entry            1447 include/linux/xarray.h static inline bool xas_retry(struct xa_state *xas, const void *entry)
entry            1449 include/linux/xarray.h 	if (xa_is_zero(entry))
entry            1451 include/linux/xarray.h 	if (!xa_is_retry(entry))
entry            1458 include/linux/xarray.h void *xas_store(struct xa_state *, void *entry);
entry            1558 include/linux/xarray.h 	void *entry;
entry            1569 include/linux/xarray.h 		entry = xa_entry(xas->xa, node, xas->xa_offset + 1);
entry            1570 include/linux/xarray.h 		if (unlikely(xa_is_internal(entry)))
entry            1574 include/linux/xarray.h 	} while (!entry);
entry            1576 include/linux/xarray.h 	return entry;
entry            1616 include/linux/xarray.h 	void *entry;
entry            1628 include/linux/xarray.h 	entry = xa_entry(xas->xa, node, offset);
entry            1629 include/linux/xarray.h 	if (!entry)
entry            1631 include/linux/xarray.h 	return entry;
entry            1655 include/linux/xarray.h #define xas_for_each(xas, entry, max) \
entry            1656 include/linux/xarray.h 	for (entry = xas_find(xas, max); entry; \
entry            1657 include/linux/xarray.h 	     entry = xas_next_entry(xas, max))
entry            1673 include/linux/xarray.h #define xas_for_each_marked(xas, entry, max, mark) \
entry            1674 include/linux/xarray.h 	for (entry = xas_find_marked(xas, max, mark); entry; \
entry            1675 include/linux/xarray.h 	     entry = xas_next_marked(xas, max, mark))
entry            1690 include/linux/xarray.h #define xas_for_each_conflict(xas, entry) \
entry            1691 include/linux/xarray.h 	while ((entry = xas_find_conflict(xas)))
entry             349 include/media/v4l2-subdev.h 	struct v4l2_mbus_frame_desc_entry entry[V4L2_FRAME_DESC_ENTRY_MAX];
entry              24 include/net/atmclip.h 	struct atmarp_entry *entry;	/* ATMARP table entry, NULL if IP addr.
entry             529 include/net/bluetooth/mgmt.h 	} entry[0];
entry             821 include/net/devlink.h 				   struct devlink_dpipe_entry *entry);
entry             823 include/net/devlink.h void devlink_dpipe_entry_clear(struct devlink_dpipe_entry *entry);
entry             392 include/net/flow_offload.h void flow_indr_add_block_cb(struct flow_indr_block_entry *entry);
entry             394 include/net/flow_offload.h void flow_indr_del_block_cb(struct flow_indr_block_entry *entry);
entry             323 include/net/ip.h 					c, stats_list[i].entry, \
entry             335 include/net/ip.h 						c, stats_list[i].entry); \
entry              28 include/net/netfilter/nf_queue.h 	int		(*outfn)(struct nf_queue_entry *entry,
entry              35 include/net/netfilter/nf_queue.h void nf_reinject(struct nf_queue_entry *entry, unsigned int verdict);
entry              37 include/net/netfilter/nf_queue.h void nf_queue_entry_get_refs(struct nf_queue_entry *entry);
entry              38 include/net/netfilter/nf_queue.h void nf_queue_entry_release_refs(struct nf_queue_entry *entry);
entry              31 include/net/snmp.h 	int entry;
entry              36 include/net/snmp.h 	.entry = _entry,			\
entry              41 include/net/snmp.h 	.entry = 0,		\
entry            2514 include/rdma/ib_verbs.h 			      struct rdma_restrack_entry *entry);
entry             110 include/rdma/rdmavt_cq.h bool rvt_cq_enter(struct rvt_cq *cq, struct ib_wc *entry, bool solicited);
entry             210 include/scsi/fc_encode.h 	struct fc_fdmi_attr_entry *entry;
entry             245 include/scsi/fc_encode.h 		entry = (struct fc_fdmi_attr_entry *)hba_attrs->attr;
entry             250 include/scsi/fc_encode.h 				   &entry->type);
entry             251 include/scsi/fc_encode.h 		put_unaligned_be16(len, &entry->len);
entry             253 include/scsi/fc_encode.h 				   (__be64 *)&entry->value[0]);
entry             256 include/scsi/fc_encode.h 		entry = (struct fc_fdmi_attr_entry *)((char *)entry->value +
entry             261 include/scsi/fc_encode.h 				   &entry->type);
entry             262 include/scsi/fc_encode.h 		put_unaligned_be16(len, &entry->len);
entry             263 include/scsi/fc_encode.h 		strncpy((char *)&entry->value,
entry             268 include/scsi/fc_encode.h 		entry = (struct fc_fdmi_attr_entry *)((char *)entry->value +
entry             273 include/scsi/fc_encode.h 				   &entry->type);
entry             274 include/scsi/fc_encode.h 		put_unaligned_be16(len, &entry->len);
entry             275 include/scsi/fc_encode.h 		strncpy((char *)&entry->value,
entry             280 include/scsi/fc_encode.h 		entry = (struct fc_fdmi_attr_entry *)((char *)entry->value +
entry             285 include/scsi/fc_encode.h 				   &entry->type);
entry             286 include/scsi/fc_encode.h 		put_unaligned_be16(len, &entry->len);
entry             287 include/scsi/fc_encode.h 		strncpy((char *)&entry->value,
entry             292 include/scsi/fc_encode.h 		entry = (struct fc_fdmi_attr_entry *)((char *)entry->value +
entry             297 include/scsi/fc_encode.h 				   &entry->type);
entry             298 include/scsi/fc_encode.h 		put_unaligned_be16(len, &entry->len);
entry             299 include/scsi/fc_encode.h 		strncpy((char *)&entry->value,
entry             304 include/scsi/fc_encode.h 		entry = (struct fc_fdmi_attr_entry *)((char *)entry->value +
entry             309 include/scsi/fc_encode.h 				   &entry->type);
entry             310 include/scsi/fc_encode.h 		put_unaligned_be16(len, &entry->len);
entry             311 include/scsi/fc_encode.h 		strncpy((char *)&entry->value,
entry             316 include/scsi/fc_encode.h 		entry = (struct fc_fdmi_attr_entry *)((char *)entry->value +
entry             321 include/scsi/fc_encode.h 				   &entry->type);
entry             322 include/scsi/fc_encode.h 		put_unaligned_be16(len, &entry->len);
entry             323 include/scsi/fc_encode.h 		strncpy((char *)&entry->value,
entry             328 include/scsi/fc_encode.h 		entry = (struct fc_fdmi_attr_entry *)((char *)entry->value +
entry             333 include/scsi/fc_encode.h 				   &entry->type);
entry             334 include/scsi/fc_encode.h 		put_unaligned_be16(len, &entry->len);
entry             335 include/scsi/fc_encode.h 		strncpy((char *)&entry->value,
entry             340 include/scsi/fc_encode.h 		entry = (struct fc_fdmi_attr_entry *)((char *)entry->value +
entry             345 include/scsi/fc_encode.h 				   &entry->type);
entry             346 include/scsi/fc_encode.h 		put_unaligned_be16(len, &entry->len);
entry             347 include/scsi/fc_encode.h 		strncpy((char *)&entry->value,
entry             352 include/scsi/fc_encode.h 		entry = (struct fc_fdmi_attr_entry *)((char *)entry->value +
entry             357 include/scsi/fc_encode.h 				   &entry->type);
entry             358 include/scsi/fc_encode.h 		put_unaligned_be16(len, &entry->len);
entry             359 include/scsi/fc_encode.h 		snprintf((char *)&entry->value,
entry             388 include/scsi/fc_encode.h 		entry = (struct fc_fdmi_attr_entry *)hba_attrs->attr;
entry             394 include/scsi/fc_encode.h 				   &entry->type);
entry             395 include/scsi/fc_encode.h 		put_unaligned_be16(len, &entry->len);
entry             396 include/scsi/fc_encode.h 		memcpy(&entry->value, fc_host_supported_fc4s(lport->host),
entry             400 include/scsi/fc_encode.h 		entry = (struct fc_fdmi_attr_entry *)((char *)entry->value +
entry             405 include/scsi/fc_encode.h 				   &entry->type);
entry             406 include/scsi/fc_encode.h 		put_unaligned_be16(len, &entry->len);
entry             409 include/scsi/fc_encode.h 				   &entry->value);
entry             412 include/scsi/fc_encode.h 		entry = (struct fc_fdmi_attr_entry *)((char *)entry->value +
entry             417 include/scsi/fc_encode.h 				   &entry->type);
entry             418 include/scsi/fc_encode.h 		put_unaligned_be16(len, &entry->len);
entry             420 include/scsi/fc_encode.h 				   &entry->value);
entry             423 include/scsi/fc_encode.h 		entry = (struct fc_fdmi_attr_entry *)((char *)entry->value +
entry             428 include/scsi/fc_encode.h 				   &entry->type);
entry             429 include/scsi/fc_encode.h 		put_unaligned_be16(len, &entry->len);
entry             431 include/scsi/fc_encode.h 				   &entry->value);
entry             434 include/scsi/fc_encode.h 		entry = (struct fc_fdmi_attr_entry *)((char *)entry->value +
entry             439 include/scsi/fc_encode.h 				   &entry->type);
entry             440 include/scsi/fc_encode.h 		put_unaligned_be16(len, &entry->len);
entry             442 include/scsi/fc_encode.h 		strncpy((char *)&entry->value,
entry             448 include/scsi/fc_encode.h 		entry = (struct fc_fdmi_attr_entry *)((char *)entry->value +
entry             453 include/scsi/fc_encode.h 				   &entry->type);
entry             454 include/scsi/fc_encode.h 		put_unaligned_be16(len, &entry->len);
entry             456 include/scsi/fc_encode.h 			strncpy((char *)&entry->value,
entry             461 include/scsi/fc_encode.h 			strncpy((char *)&entry->value,
entry              30 include/sound/info.h 	void (*read)(struct snd_info_entry *entry,
entry              32 include/sound/info.h 	void (*write)(struct snd_info_entry *entry,
entry              37 include/sound/info.h 	int (*open)(struct snd_info_entry *entry,
entry              39 include/sound/info.h 	int (*release)(struct snd_info_entry *entry,
entry              41 include/sound/info.h 	ssize_t (*read)(struct snd_info_entry *entry, void *file_private_data,
entry              44 include/sound/info.h 	ssize_t (*write)(struct snd_info_entry *entry, void *file_private_data,
entry              47 include/sound/info.h 	loff_t (*llseek)(struct snd_info_entry *entry,
entry              50 include/sound/info.h 	__poll_t (*poll)(struct snd_info_entry *entry,
entry              53 include/sound/info.h 	int (*ioctl)(struct snd_info_entry *entry, void *file_private_data,
entry              55 include/sound/info.h 	int (*mmap)(struct snd_info_entry *entry, void *file_private_data,
entry              72 include/sound/info.h 	void (*private_free)(struct snd_info_entry *entry);
entry             120 include/sound/info.h void snd_info_free_entry(struct snd_info_entry *entry);
entry             121 include/sound/info.h int snd_info_store_text(struct snd_info_entry *entry);
entry             122 include/sound/info.h int snd_info_restore_text(struct snd_info_entry *entry);
entry             129 include/sound/info.h int snd_info_register(struct snd_info_entry *entry);
entry             139 include/sound/info.h static inline void snd_info_set_text_ops(struct snd_info_entry *entry, 
entry             143 include/sound/info.h 	entry->private_data = private_data;
entry             144 include/sound/info.h 	entry->c.text.read = read;
entry             151 include/sound/info.h 			 void (*write)(struct snd_info_entry *entry,
entry             169 include/sound/info.h static inline void snd_info_free_entry(struct snd_info_entry *entry) { ; }
entry             176 include/sound/info.h static inline int snd_info_register(struct snd_info_entry *entry) { return 0; }
entry             180 include/sound/info.h static inline void snd_info_set_text_ops(struct snd_info_entry *entry __attribute__((unused)),
entry             187 include/sound/info.h 				       void (*write)(struct snd_info_entry *entry,
entry               8 include/trace/bpf_probe.h #define __entry entry
entry              14 include/trace/events/erofs.h #define show_dev_nid(entry)	show_dev(entry->dev), entry->nid
entry              11 include/trace/events/f2fs.h #define show_dev_ino(entry)	show_dev(entry->dev), (unsigned long)entry->ino
entry              20 include/trace/events/ib_mad.h 			  struct trace_event_raw_ib_mad_send_template *entry);
entry              49 include/trace/events/task.h 		memcpy(entry->oldcomm, task->comm, TASK_COMM_LEN);
entry              50 include/trace/events/task.h 		strlcpy(entry->newcomm, comm, TASK_COMM_LEN);
entry               8 include/trace/perf.h #define __entry entry
entry              37 include/trace/perf.h 	struct trace_event_raw_##call *entry;				\
entry              54 include/trace/perf.h 	__entry_size = ALIGN(__data_size + sizeof(*entry) + sizeof(u32),\
entry              58 include/trace/perf.h 	entry = perf_trace_buf_alloc(__entry_size, &__regs, &rctx);	\
entry              59 include/trace/perf.h 	if (!entry)							\
entry              68 include/trace/perf.h 	perf_trace_run_bpf_submit(entry, __entry_size, rctx,		\
entry             375 include/trace/trace_events.h 	struct trace_entry *entry;					\
entry             378 include/trace/trace_events.h 	entry = iter->ent;						\
entry             380 include/trace/trace_events.h 	if (entry->type != event_##call.event.type) {			\
entry             385 include/trace/trace_events.h 	field = (typeof(field))entry;					\
entry             474 include/trace/trace_events.h #define __entry entry
entry             495 include/trace/trace_events.h 			       offsetof(typeof(*entry), __data);	\
entry             533 include/trace/trace_events.h 	struct trace_event_raw_##call __maybe_unused *entry;		\
entry             654 include/trace/trace_events.h #define __entry entry
entry             704 include/trace/trace_events.h 	struct trace_event_raw_##call *entry;				\
entry             712 include/trace/trace_events.h 	entry = trace_event_buffer_reserve(&fbuffer, trace_file,	\
entry             713 include/trace/trace_events.h 				 sizeof(*entry) + __data_size);		\
entry             715 include/trace/trace_events.h 	if (!entry)							\
entry             140 include/uapi/linux/coff.h   char	entry[4];		/* entry pt.				 */
entry            1590 include/uapi/linux/kvm.h 	__u16 entry; /* The index of entry in the MSI-X table */
entry              23 include/uapi/linux/mpls.h 	__be32 entry;
entry             180 include/uapi/linux/netfilter/x_tables.h #define xt_ematch_foreach(pos, entry) \
entry             181 include/uapi/linux/netfilter/x_tables.h 	for ((pos) = (struct xt_entry_match *)entry->elems; \
entry             182 include/uapi/linux/netfilter/x_tables.h 	     (pos) < (struct xt_entry_match *)((char *)(entry) + \
entry             183 include/uapi/linux/netfilter/x_tables.h 	             (entry)->target_offset); \
entry             133 include/uapi/linux/netfilter_ipv6/ip6_tables.h 	struct ip6t_entry entry;
entry             138 include/uapi/linux/netfilter_ipv6/ip6_tables.h 	struct ip6t_entry entry;
entry             150 include/uapi/linux/netfilter_ipv6/ip6_tables.h 	.entry		= IP6T_ENTRY_INIT(sizeof(struct ip6t_standard)),       \
entry             158 include/uapi/linux/netfilter_ipv6/ip6_tables.h 	.entry		= IP6T_ENTRY_INIT(sizeof(struct ip6t_error)),	       \
entry              91 include/uapi/linux/ptrace.h 		} entry;
entry            1951 include/uapi/linux/videodev2.h 	struct v4l2_enc_idx_entry entry[V4L2_ENC_IDX_ENTRIES];
entry              59 include/uapi/xen/privcmd.h 	struct privcmd_mmap_entry __user *entry;
entry              21 include/video/udlfb.h 	struct list_head entry;
entry              67 include/xen/interface/io/pciif.h 	uint16_t entry;
entry             373 include/xen/interface/xen-mca.h 	struct xen_mce entry[XEN_MCE_LOG_LEN];
entry             811 init/main.c    	struct blacklist_entry *entry;
entry             818 init/main.c    			entry = memblock_alloc(sizeof(*entry),
entry             820 init/main.c    			if (!entry)
entry             822 init/main.c    				      __func__, sizeof(*entry));
entry             823 init/main.c    			entry->buf = memblock_alloc(strlen(str_entry) + 1,
entry             825 init/main.c    			if (!entry->buf)
entry             828 init/main.c    			strcpy(entry->buf, str_entry);
entry             829 init/main.c    			list_add(&entry->next, &blacklisted_initcalls);
entry             838 init/main.c    	struct blacklist_entry *entry;
entry             854 init/main.c    	list_for_each_entry(entry, &blacklisted_initcalls, next) {
entry             855 init/main.c    		if (!strcmp(fn_name, entry->buf)) {
entry             111 kernel/async.c 	struct async_entry *entry =
entry             119 kernel/async.c 			(long long)entry->cookie,
entry             120 kernel/async.c 			entry->func, task_pid_nr(current));
entry             123 kernel/async.c 	entry->func(entry->data, entry->cookie);
entry             128 kernel/async.c 			(long long)entry->cookie,
entry             129 kernel/async.c 			entry->func,
entry             135 kernel/async.c 	list_del_init(&entry->domain_list);
entry             136 kernel/async.c 	list_del_init(&entry->global_list);
entry             139 kernel/async.c 	kfree(entry);
entry             168 kernel/async.c 	struct async_entry *entry;
entry             173 kernel/async.c 	entry = kzalloc(sizeof(struct async_entry), GFP_ATOMIC);
entry             179 kernel/async.c 	if (!entry || atomic_read(&entry_count) > MAX_WORK) {
entry             180 kernel/async.c 		kfree(entry);
entry             189 kernel/async.c 	INIT_LIST_HEAD(&entry->domain_list);
entry             190 kernel/async.c 	INIT_LIST_HEAD(&entry->global_list);
entry             191 kernel/async.c 	INIT_WORK(&entry->work, async_run_entry_fn);
entry             192 kernel/async.c 	entry->func = func;
entry             193 kernel/async.c 	entry->data = data;
entry             194 kernel/async.c 	entry->domain = domain;
entry             199 kernel/async.c 	newcookie = entry->cookie = next_cookie++;
entry             201 kernel/async.c 	list_add_tail(&entry->domain_list, &domain->pending);
entry             203 kernel/async.c 		list_add_tail(&entry->global_list, &async_global_pending);
entry             212 kernel/async.c 	queue_work_node(node, system_unbound_wq, &entry->work);
entry             237 kernel/audit.h extern int audit_del_rule(struct audit_entry *entry);
entry             148 kernel/audit_fsnotify.c 	struct audit_entry *entry = container_of(rule, struct audit_entry, rule);
entry             151 kernel/audit_fsnotify.c 	audit_del_rule(entry);
entry             547 kernel/audit_tree.c 	struct audit_entry *entry;
entry             550 kernel/audit_tree.c 		entry = container_of(rule, struct audit_entry, rule);
entry             556 kernel/audit_tree.c 			if (entry->rule.exe)
entry             557 kernel/audit_tree.c 				audit_remove_mark(entry->rule.exe);
entry             559 kernel/audit_tree.c 			list_del_rcu(&entry->list);
entry             560 kernel/audit_tree.c 			list_del(&entry->rule.list);
entry             561 kernel/audit_tree.c 			call_rcu(&entry->rcu, audit_free_rule_rcu);
entry              64 kernel/audit_watch.c static void audit_watch_free_mark(struct fsnotify_mark *entry)
entry              68 kernel/audit_watch.c 	parent = container_of(entry, struct audit_parent, mark);
entry              91 kernel/audit_watch.c 	struct fsnotify_mark *entry;
entry              93 kernel/audit_watch.c 	entry = fsnotify_find_mark(&inode->i_fsnotify_marks, audit_watch_group);
entry              94 kernel/audit_watch.c 	if (entry)
entry              95 kernel/audit_watch.c 		parent = container_of(entry, struct audit_parent, mark);
entry             106 kernel/auditfilter.c 	struct audit_entry *entry;
entry             109 kernel/auditfilter.c 	entry = kzalloc(sizeof(*entry), GFP_KERNEL);
entry             110 kernel/auditfilter.c 	if (unlikely(!entry))
entry             115 kernel/auditfilter.c 		kfree(entry);
entry             118 kernel/auditfilter.c 	entry->rule.fields = fields;
entry             120 kernel/auditfilter.c 	return entry;
entry             208 kernel/auditfilter.c static int audit_match_signal(struct audit_entry *entry)
entry             210 kernel/auditfilter.c 	struct audit_field *arch = entry->rule.arch_f;
entry             216 kernel/auditfilter.c 					       entry->rule.mask) &&
entry             218 kernel/auditfilter.c 					       entry->rule.mask));
entry             224 kernel/auditfilter.c 					       entry->rule.mask));
entry             227 kernel/auditfilter.c 					       entry->rule.mask));
entry             238 kernel/auditfilter.c 	struct audit_entry *entry;
entry             268 kernel/auditfilter.c 	entry = audit_init_entry(rule->field_count);
entry             269 kernel/auditfilter.c 	if (!entry)
entry             272 kernel/auditfilter.c 	entry->rule.flags = rule->flags & AUDIT_FILTER_PREPEND;
entry             273 kernel/auditfilter.c 	entry->rule.listnr = listnr;
entry             274 kernel/auditfilter.c 	entry->rule.action = rule->action;
entry             275 kernel/auditfilter.c 	entry->rule.field_count = rule->field_count;
entry             278 kernel/auditfilter.c 		entry->rule.mask[i] = rule->mask[i];
entry             282 kernel/auditfilter.c 		__u32 *p = &entry->rule.mask[AUDIT_WORD(bit)];
entry             292 kernel/auditfilter.c 				entry->rule.mask[j] |= class[j];
entry             296 kernel/auditfilter.c 	return entry;
entry             323 kernel/auditfilter.c static int audit_field_valid(struct audit_entry *entry, struct audit_field *f)
entry             327 kernel/auditfilter.c 		if (entry->rule.listnr != AUDIT_FILTER_EXCLUDE &&
entry             328 kernel/auditfilter.c 		    entry->rule.listnr != AUDIT_FILTER_USER)
entry             332 kernel/auditfilter.c 		if (entry->rule.listnr != AUDIT_FILTER_FS)
entry             337 kernel/auditfilter.c 	switch (entry->rule.listnr) {
entry             445 kernel/auditfilter.c 	struct audit_entry *entry;
entry             452 kernel/auditfilter.c 	entry = audit_to_entry_common(data);
entry             453 kernel/auditfilter.c 	if (IS_ERR(entry))
entry             458 kernel/auditfilter.c 		struct audit_field *f = &entry->rule.fields[i];
entry             474 kernel/auditfilter.c 			entry->rule.pflags |= AUDIT_LOGINUID_LEGACY;
entry             477 kernel/auditfilter.c 		err = audit_field_valid(entry, f);
entry             504 kernel/auditfilter.c 			entry->rule.arch_f = f;
entry             521 kernel/auditfilter.c 			entry->rule.buflen += f_val;
entry             540 kernel/auditfilter.c 			err = audit_to_watch(&entry->rule, str, f_val, f->op);
entry             545 kernel/auditfilter.c 			entry->rule.buflen += f_val;
entry             553 kernel/auditfilter.c 			err = audit_make_tree(&entry->rule, str, f->op);
entry             557 kernel/auditfilter.c 			entry->rule.buflen += f_val;
entry             561 kernel/auditfilter.c 			err = audit_to_inode(&entry->rule, f);
entry             566 kernel/auditfilter.c 			if (entry->rule.filterkey || f_val > AUDIT_MAX_KEY_LEN)
entry             573 kernel/auditfilter.c 			entry->rule.buflen += f_val;
entry             574 kernel/auditfilter.c 			entry->rule.filterkey = str;
entry             577 kernel/auditfilter.c 			if (entry->rule.exe || f_val > PATH_MAX)
entry             584 kernel/auditfilter.c 			audit_mark = audit_alloc_mark(&entry->rule, str, f_val);
entry             590 kernel/auditfilter.c 			entry->rule.buflen += f_val;
entry             591 kernel/auditfilter.c 			entry->rule.exe = audit_mark;
entry             599 kernel/auditfilter.c 	if (entry->rule.inode_f && entry->rule.inode_f->op == Audit_not_equal)
entry             600 kernel/auditfilter.c 		entry->rule.inode_f = NULL;
entry             603 kernel/auditfilter.c 	return entry;
entry             606 kernel/auditfilter.c 	if (entry->rule.tree)
entry             607 kernel/auditfilter.c 		audit_put_tree(entry->rule.tree); /* that's the temporary one */
entry             608 kernel/auditfilter.c 	if (entry->rule.exe)
entry             609 kernel/auditfilter.c 		audit_remove_mark(entry->rule.exe); /* that's the template one */
entry             610 kernel/auditfilter.c 	audit_free_rule(entry);
entry             814 kernel/auditfilter.c 	struct audit_entry *entry;
entry             819 kernel/auditfilter.c 	entry = audit_init_entry(fcount);
entry             820 kernel/auditfilter.c 	if (unlikely(!entry))
entry             823 kernel/auditfilter.c 	new = &entry->rule;
entry             876 kernel/auditfilter.c 			audit_free_rule(entry);
entry             886 kernel/auditfilter.c 	return entry;
entry             891 kernel/auditfilter.c static struct audit_entry *audit_find_rule(struct audit_entry *entry,
entry             898 kernel/auditfilter.c 	if (entry->rule.inode_f) {
entry             899 kernel/auditfilter.c 		h = audit_hash_ino(entry->rule.inode_f->val);
entry             901 kernel/auditfilter.c 	} else if (entry->rule.watch) {
entry             906 kernel/auditfilter.c 				if (!audit_compare_rule(&entry->rule, &e->rule)) {
entry             913 kernel/auditfilter.c 		*p = list = &audit_filter_list[entry->rule.listnr];
entry             917 kernel/auditfilter.c 		if (!audit_compare_rule(&entry->rule, &e->rule)) {
entry             930 kernel/auditfilter.c static inline int audit_add_rule(struct audit_entry *entry)
entry             933 kernel/auditfilter.c 	struct audit_watch *watch = entry->rule.watch;
entry             934 kernel/auditfilter.c 	struct audit_tree *tree = entry->rule.tree;
entry             941 kernel/auditfilter.c 	switch(entry->rule.listnr) {
entry             950 kernel/auditfilter.c 	e = audit_find_rule(entry, &list);
entry             962 kernel/auditfilter.c 		err = audit_add_watch(&entry->rule, &list);
entry             975 kernel/auditfilter.c 		err = audit_add_tree_rule(&entry->rule);
entry             982 kernel/auditfilter.c 	entry->rule.prio = ~0ULL;
entry             983 kernel/auditfilter.c 	if (entry->rule.listnr == AUDIT_FILTER_EXIT) {
entry             984 kernel/auditfilter.c 		if (entry->rule.flags & AUDIT_FILTER_PREPEND)
entry             985 kernel/auditfilter.c 			entry->rule.prio = ++prio_high;
entry             987 kernel/auditfilter.c 			entry->rule.prio = --prio_low;
entry             990 kernel/auditfilter.c 	if (entry->rule.flags & AUDIT_FILTER_PREPEND) {
entry             991 kernel/auditfilter.c 		list_add(&entry->rule.list,
entry             992 kernel/auditfilter.c 			 &audit_rules_list[entry->rule.listnr]);
entry             993 kernel/auditfilter.c 		list_add_rcu(&entry->list, list);
entry             994 kernel/auditfilter.c 		entry->rule.flags &= ~AUDIT_FILTER_PREPEND;
entry             996 kernel/auditfilter.c 		list_add_tail(&entry->rule.list,
entry             997 kernel/auditfilter.c 			      &audit_rules_list[entry->rule.listnr]);
entry             998 kernel/auditfilter.c 		list_add_tail_rcu(&entry->list, list);
entry            1004 kernel/auditfilter.c 	if (!audit_match_signal(entry))
entry            1013 kernel/auditfilter.c int audit_del_rule(struct audit_entry *entry)
entry            1016 kernel/auditfilter.c 	struct audit_tree *tree = entry->rule.tree;
entry            1023 kernel/auditfilter.c 	switch(entry->rule.listnr) {
entry            1032 kernel/auditfilter.c 	e = audit_find_rule(entry, &list);
entry            1051 kernel/auditfilter.c 	if (!audit_match_signal(entry))
entry            1126 kernel/auditfilter.c 	struct audit_entry *entry;
entry            1130 kernel/auditfilter.c 		entry = audit_data_to_entry(data, datasz);
entry            1131 kernel/auditfilter.c 		if (IS_ERR(entry))
entry            1132 kernel/auditfilter.c 			return PTR_ERR(entry);
entry            1133 kernel/auditfilter.c 		err = audit_add_rule(entry);
entry            1134 kernel/auditfilter.c 		audit_log_rule_change("add_rule", &entry->rule, !err);
entry            1137 kernel/auditfilter.c 		entry = audit_data_to_entry(data, datasz);
entry            1138 kernel/auditfilter.c 		if (IS_ERR(entry))
entry            1139 kernel/auditfilter.c 			return PTR_ERR(entry);
entry            1140 kernel/auditfilter.c 		err = audit_del_rule(entry);
entry            1141 kernel/auditfilter.c 		audit_log_rule_change("remove_rule", &entry->rule, !err);
entry            1149 kernel/auditfilter.c 		if (entry->rule.exe)
entry            1150 kernel/auditfilter.c 			audit_remove_mark(entry->rule.exe);
entry            1151 kernel/auditfilter.c 		audit_free_rule(entry);
entry            1395 kernel/auditfilter.c 	struct audit_entry *entry = container_of(r, struct audit_entry, rule);
entry            1403 kernel/auditfilter.c 	if (entry->rule.exe)
entry            1404 kernel/auditfilter.c 		audit_remove_mark(entry->rule.exe);
entry            1412 kernel/auditfilter.c 		list_del_rcu(&entry->list);
entry            1417 kernel/auditfilter.c 		list_replace_rcu(&entry->list, &nentry->list);
entry            1420 kernel/auditfilter.c 	call_rcu(&entry->rcu, audit_free_rule_rcu);
entry              39 kernel/bpf/stackmap.c static void do_up_read(struct irq_work *entry)
entry              43 kernel/bpf/stackmap.c 	work = container_of(entry, struct stack_map_irq_work, irq_work);
entry              58 kernel/configs.c 	struct proc_dir_entry *entry;
entry              61 kernel/configs.c 	entry = proc_create("config.gz", S_IFREG | S_IRUGO, NULL,
entry              63 kernel/configs.c 	if (!entry)
entry              66 kernel/configs.c 	proc_set_size(entry, &kernel_config_data_end - &kernel_config_data);
entry             163 kernel/dma/debug.c static inline void dump_entry_trace(struct dma_debug_entry *entry)
entry             166 kernel/dma/debug.c 	if (entry) {
entry             168 kernel/dma/debug.c 		stack_trace_print(entry->stack_entries, entry->stack_len, 0);
entry             214 kernel/dma/debug.c #define err_printk(dev, entry, format, arg...) do {			\
entry             221 kernel/dma/debug.c 			dump_entry_trace(entry);			\
entry             233 kernel/dma/debug.c static int hash_fn(struct dma_debug_entry *entry)
entry             239 kernel/dma/debug.c 	return (entry->dev_addr >> HASH_FN_SHIFT) & HASH_FN_MASK;
entry             245 kernel/dma/debug.c static struct hash_bucket *get_hash_bucket(struct dma_debug_entry *entry,
entry             249 kernel/dma/debug.c 	int idx = hash_fn(entry);
entry             295 kernel/dma/debug.c 	struct dma_debug_entry *entry, *ret = NULL;
entry             298 kernel/dma/debug.c 	list_for_each_entry(entry, &bucket->list, list) {
entry             299 kernel/dma/debug.c 		if (!match(ref, entry))
entry             314 kernel/dma/debug.c 		entry->size         == ref->size         ? ++match_lvl : 0;
entry             315 kernel/dma/debug.c 		entry->type         == ref->type         ? ++match_lvl : 0;
entry             316 kernel/dma/debug.c 		entry->direction    == ref->direction    ? ++match_lvl : 0;
entry             317 kernel/dma/debug.c 		entry->sg_call_ents == ref->sg_call_ents ? ++match_lvl : 0;
entry             321 kernel/dma/debug.c 			return entry;
entry             328 kernel/dma/debug.c 			ret      = entry;
entry             353 kernel/dma/debug.c 	struct dma_debug_entry *entry, index = *ref;
entry             357 kernel/dma/debug.c 		entry = __hash_bucket_find(*bucket, ref, containing_match);
entry             359 kernel/dma/debug.c 		if (entry)
entry             360 kernel/dma/debug.c 			return entry;
entry             378 kernel/dma/debug.c 			    struct dma_debug_entry *entry)
entry             380 kernel/dma/debug.c 	list_add_tail(&entry->list, &bucket->list);
entry             386 kernel/dma/debug.c static void hash_bucket_del(struct dma_debug_entry *entry)
entry             388 kernel/dma/debug.c 	list_del(&entry->list);
entry             391 kernel/dma/debug.c static unsigned long long phys_addr(struct dma_debug_entry *entry)
entry             393 kernel/dma/debug.c 	if (entry->type == dma_debug_resource)
entry             394 kernel/dma/debug.c 		return __pfn_to_phys(entry->pfn) + entry->offset;
entry             396 kernel/dma/debug.c 	return page_to_phys(pfn_to_page(entry->pfn)) + entry->offset;
entry             408 kernel/dma/debug.c 		struct dma_debug_entry *entry;
entry             413 kernel/dma/debug.c 		list_for_each_entry(entry, &bucket->list, list) {
entry             414 kernel/dma/debug.c 			if (!dev || dev == entry->dev) {
entry             415 kernel/dma/debug.c 				dev_info(entry->dev,
entry             417 kernel/dma/debug.c 					 type2name[entry->type], idx,
entry             418 kernel/dma/debug.c 					 phys_addr(entry), entry->pfn,
entry             419 kernel/dma/debug.c 					 entry->dev_addr, entry->size,
entry             420 kernel/dma/debug.c 					 dir2name[entry->direction],
entry             421 kernel/dma/debug.c 					 maperr2str[entry->map_err_type]);
entry             459 kernel/dma/debug.c static phys_addr_t to_cacheline_number(struct dma_debug_entry *entry)
entry             461 kernel/dma/debug.c 	return (entry->pfn << CACHELINE_PER_PAGE_SHIFT) +
entry             462 kernel/dma/debug.c 		(entry->offset >> L1_CACHE_SHIFT);
entry             515 kernel/dma/debug.c static int active_cacheline_insert(struct dma_debug_entry *entry)
entry             517 kernel/dma/debug.c 	phys_addr_t cln = to_cacheline_number(entry);
entry             525 kernel/dma/debug.c 	if (entry->direction == DMA_TO_DEVICE)
entry             529 kernel/dma/debug.c 	rc = radix_tree_insert(&dma_active_cacheline, cln, entry);
entry             537 kernel/dma/debug.c static void active_cacheline_remove(struct dma_debug_entry *entry)
entry             539 kernel/dma/debug.c 	phys_addr_t cln = to_cacheline_number(entry);
entry             543 kernel/dma/debug.c 	if (entry->direction == DMA_TO_DEVICE)
entry             567 kernel/dma/debug.c 	struct dma_debug_entry *entry = NULL;
entry             587 kernel/dma/debug.c 			entry = ents[i];
entry             594 kernel/dma/debug.c 	if (!entry)
entry             597 kernel/dma/debug.c 	cln = to_cacheline_number(entry);
entry             598 kernel/dma/debug.c 	err_printk(entry->dev, entry,
entry             607 kernel/dma/debug.c static void add_dma_entry(struct dma_debug_entry *entry)
entry             613 kernel/dma/debug.c 	bucket = get_hash_bucket(entry, &flags);
entry             614 kernel/dma/debug.c 	hash_bucket_add(bucket, entry);
entry             617 kernel/dma/debug.c 	rc = active_cacheline_insert(entry);
entry             630 kernel/dma/debug.c 	struct dma_debug_entry *entry;
entry             633 kernel/dma/debug.c 	entry = (void *)get_zeroed_page(gfp);
entry             634 kernel/dma/debug.c 	if (!entry)
entry             638 kernel/dma/debug.c 		list_add_tail(&entry[i].list, &free_entries);
entry             648 kernel/dma/debug.c 	struct dma_debug_entry *entry;
entry             650 kernel/dma/debug.c 	entry = list_entry(free_entries.next, struct dma_debug_entry, list);
entry             651 kernel/dma/debug.c 	list_del(&entry->list);
entry             652 kernel/dma/debug.c 	memset(entry, 0, sizeof(*entry));
entry             658 kernel/dma/debug.c 	return entry;
entry             680 kernel/dma/debug.c 	struct dma_debug_entry *entry;
entry             694 kernel/dma/debug.c 	entry = __dma_entry_alloc();
entry             699 kernel/dma/debug.c 	entry->stack_len = stack_trace_save(entry->stack_entries,
entry             700 kernel/dma/debug.c 					    ARRAY_SIZE(entry->stack_entries),
entry             703 kernel/dma/debug.c 	return entry;
entry             706 kernel/dma/debug.c static void dma_entry_free(struct dma_debug_entry *entry)
entry             710 kernel/dma/debug.c 	active_cacheline_remove(entry);
entry             717 kernel/dma/debug.c 	list_add(&entry->list, &free_entries);
entry             828 kernel/dma/debug.c 		struct dma_debug_entry *entry;
entry             832 kernel/dma/debug.c 		list_for_each_entry(entry, &bucket->list, list) {
entry             835 kernel/dma/debug.c 				   dev_name(entry->dev),
entry             836 kernel/dma/debug.c 				   dev_driver_string(entry->dev),
entry             837 kernel/dma/debug.c 				   type2name[entry->type], idx,
entry             838 kernel/dma/debug.c 				   phys_addr(entry), entry->pfn,
entry             839 kernel/dma/debug.c 				   entry->dev_addr, entry->size,
entry             840 kernel/dma/debug.c 				   dir2name[entry->direction],
entry             841 kernel/dma/debug.c 				   maperr2str[entry->map_err_type]);
entry             866 kernel/dma/debug.c 	struct dma_debug_entry *entry;
entry             872 kernel/dma/debug.c 		list_for_each_entry(entry, &dma_entry_hash[i].list, list) {
entry             873 kernel/dma/debug.c 			if (entry->dev == dev) {
entry             875 kernel/dma/debug.c 				*out_entry = entry;
entry             887 kernel/dma/debug.c 	struct dma_debug_entry *uninitialized_var(entry);
entry             895 kernel/dma/debug.c 		count = device_dma_allocations(dev, &entry);
entry             898 kernel/dma/debug.c 		err_printk(dev, entry, "device driver has pending "
entry             904 kernel/dma/debug.c 			count, entry->dev_addr, entry->size,
entry             905 kernel/dma/debug.c 			dir2name[entry->direction], type2name[entry->type]);
entry             999 kernel/dma/debug.c 	struct dma_debug_entry *entry;
entry            1004 kernel/dma/debug.c 	entry = bucket_find_exact(bucket, ref);
entry            1006 kernel/dma/debug.c 	if (!entry) {
entry            1024 kernel/dma/debug.c 	if (ref->size != entry->size) {
entry            1025 kernel/dma/debug.c 		err_printk(ref->dev, entry, "device driver frees "
entry            1029 kernel/dma/debug.c 			   ref->dev_addr, entry->size, ref->size);
entry            1032 kernel/dma/debug.c 	if (ref->type != entry->type) {
entry            1033 kernel/dma/debug.c 		err_printk(ref->dev, entry, "device driver frees "
entry            1038 kernel/dma/debug.c 			   type2name[entry->type], type2name[ref->type]);
entry            1039 kernel/dma/debug.c 	} else if ((entry->type == dma_debug_coherent) &&
entry            1040 kernel/dma/debug.c 		   (phys_addr(ref) != phys_addr(entry))) {
entry            1041 kernel/dma/debug.c 		err_printk(ref->dev, entry, "device driver frees "
entry            1047 kernel/dma/debug.c 			   phys_addr(entry),
entry            1052 kernel/dma/debug.c 	    ref->sg_call_ents != entry->sg_call_ents) {
entry            1053 kernel/dma/debug.c 		err_printk(ref->dev, entry, "device driver frees "
entry            1056 kernel/dma/debug.c 			   entry->sg_call_ents, ref->sg_call_ents);
entry            1063 kernel/dma/debug.c 	if (ref->direction != entry->direction) {
entry            1064 kernel/dma/debug.c 		err_printk(ref->dev, entry, "device driver frees "
entry            1069 kernel/dma/debug.c 			   dir2name[entry->direction],
entry            1078 kernel/dma/debug.c 	if (entry->map_err_type == MAP_ERR_NOT_CHECKED) {
entry            1079 kernel/dma/debug.c 		err_printk(ref->dev, entry,
entry            1084 kernel/dma/debug.c 			   type2name[entry->type]);
entry            1087 kernel/dma/debug.c 	hash_bucket_del(entry);
entry            1088 kernel/dma/debug.c 	dma_entry_free(entry);
entry            1142 kernel/dma/debug.c 	struct dma_debug_entry *entry;
entry            1148 kernel/dma/debug.c 	entry = bucket_find_contain(&bucket, ref, &flags);
entry            1150 kernel/dma/debug.c 	if (!entry) {
entry            1158 kernel/dma/debug.c 	if (ref->size > entry->size) {
entry            1159 kernel/dma/debug.c 		err_printk(dev, entry, "device driver syncs"
entry            1164 kernel/dma/debug.c 				entry->dev_addr, entry->size,
entry            1168 kernel/dma/debug.c 	if (entry->direction == DMA_BIDIRECTIONAL)
entry            1171 kernel/dma/debug.c 	if (ref->direction != entry->direction) {
entry            1172 kernel/dma/debug.c 		err_printk(dev, entry, "device driver syncs "
entry            1176 kernel/dma/debug.c 				(unsigned long long)ref->dev_addr, entry->size,
entry            1177 kernel/dma/debug.c 				dir2name[entry->direction],
entry            1181 kernel/dma/debug.c 	if (to_cpu && !(entry->direction == DMA_FROM_DEVICE) &&
entry            1183 kernel/dma/debug.c 		err_printk(dev, entry, "device driver syncs "
entry            1187 kernel/dma/debug.c 				(unsigned long long)ref->dev_addr, entry->size,
entry            1188 kernel/dma/debug.c 				dir2name[entry->direction],
entry            1191 kernel/dma/debug.c 	if (!to_cpu && !(entry->direction == DMA_TO_DEVICE) &&
entry            1193 kernel/dma/debug.c 		err_printk(dev, entry, "device driver syncs "
entry            1197 kernel/dma/debug.c 				(unsigned long long)ref->dev_addr, entry->size,
entry            1198 kernel/dma/debug.c 				dir2name[entry->direction],
entry            1202 kernel/dma/debug.c 	    ref->sg_call_ents != entry->sg_call_ents) {
entry            1203 kernel/dma/debug.c 		err_printk(ref->dev, entry, "device driver syncs "
entry            1206 kernel/dma/debug.c 			   entry->sg_call_ents, ref->sg_call_ents);
entry            1258 kernel/dma/debug.c 	struct dma_debug_entry *entry;
entry            1266 kernel/dma/debug.c 	entry = dma_entry_alloc();
entry            1267 kernel/dma/debug.c 	if (!entry)
entry            1270 kernel/dma/debug.c 	entry->dev       = dev;
entry            1271 kernel/dma/debug.c 	entry->type      = dma_debug_single;
entry            1272 kernel/dma/debug.c 	entry->pfn	 = page_to_pfn(page);
entry            1273 kernel/dma/debug.c 	entry->offset	 = offset,
entry            1274 kernel/dma/debug.c 	entry->dev_addr  = dma_addr;
entry            1275 kernel/dma/debug.c 	entry->size      = size;
entry            1276 kernel/dma/debug.c 	entry->direction = direction;
entry            1277 kernel/dma/debug.c 	entry->map_err_type = MAP_ERR_NOT_CHECKED;
entry            1287 kernel/dma/debug.c 	add_dma_entry(entry);
entry            1294 kernel/dma/debug.c 	struct dma_debug_entry *entry;
entry            1305 kernel/dma/debug.c 	list_for_each_entry(entry, &bucket->list, list) {
entry            1306 kernel/dma/debug.c 		if (!exact_match(&ref, entry))
entry            1319 kernel/dma/debug.c 		if (entry->map_err_type == MAP_ERR_NOT_CHECKED) {
entry            1320 kernel/dma/debug.c 			entry->map_err_type = MAP_ERR_CHECKED;
entry            1349 kernel/dma/debug.c 	struct dma_debug_entry *entry;
entry            1357 kernel/dma/debug.c 		entry = dma_entry_alloc();
entry            1358 kernel/dma/debug.c 		if (!entry)
entry            1361 kernel/dma/debug.c 		entry->type           = dma_debug_sg;
entry            1362 kernel/dma/debug.c 		entry->dev            = dev;
entry            1363 kernel/dma/debug.c 		entry->pfn	      = page_to_pfn(sg_page(s));
entry            1364 kernel/dma/debug.c 		entry->offset	      = s->offset,
entry            1365 kernel/dma/debug.c 		entry->size           = sg_dma_len(s);
entry            1366 kernel/dma/debug.c 		entry->dev_addr       = sg_dma_address(s);
entry            1367 kernel/dma/debug.c 		entry->direction      = direction;
entry            1368 kernel/dma/debug.c 		entry->sg_call_ents   = nents;
entry            1369 kernel/dma/debug.c 		entry->sg_mapped_ents = mapped_ents;
entry            1379 kernel/dma/debug.c 		add_dma_entry(entry);
entry            1387 kernel/dma/debug.c 	struct dma_debug_entry *entry;
entry            1393 kernel/dma/debug.c 	entry        = bucket_find_exact(bucket, ref);
entry            1396 kernel/dma/debug.c 	if (entry)
entry            1397 kernel/dma/debug.c 		mapped_ents = entry->sg_mapped_ents;
entry            1439 kernel/dma/debug.c 	struct dma_debug_entry *entry;
entry            1451 kernel/dma/debug.c 	entry = dma_entry_alloc();
entry            1452 kernel/dma/debug.c 	if (!entry)
entry            1455 kernel/dma/debug.c 	entry->type      = dma_debug_coherent;
entry            1456 kernel/dma/debug.c 	entry->dev       = dev;
entry            1457 kernel/dma/debug.c 	entry->offset	 = offset_in_page(virt);
entry            1458 kernel/dma/debug.c 	entry->size      = size;
entry            1459 kernel/dma/debug.c 	entry->dev_addr  = dma_addr;
entry            1460 kernel/dma/debug.c 	entry->direction = DMA_BIDIRECTIONAL;
entry            1463 kernel/dma/debug.c 		entry->pfn = vmalloc_to_pfn(virt);
entry            1465 kernel/dma/debug.c 		entry->pfn = page_to_pfn(virt_to_page(virt));
entry            1467 kernel/dma/debug.c 	add_dma_entry(entry);
entry            1500 kernel/dma/debug.c 	struct dma_debug_entry *entry;
entry            1505 kernel/dma/debug.c 	entry = dma_entry_alloc();
entry            1506 kernel/dma/debug.c 	if (!entry)
entry            1509 kernel/dma/debug.c 	entry->type		= dma_debug_resource;
entry            1510 kernel/dma/debug.c 	entry->dev		= dev;
entry            1511 kernel/dma/debug.c 	entry->pfn		= PHYS_PFN(addr);
entry            1512 kernel/dma/debug.c 	entry->offset		= offset_in_page(addr);
entry            1513 kernel/dma/debug.c 	entry->size		= size;
entry            1514 kernel/dma/debug.c 	entry->dev_addr		= dma_addr;
entry            1515 kernel/dma/debug.c 	entry->direction	= direction;
entry            1516 kernel/dma/debug.c 	entry->map_err_type	= MAP_ERR_NOT_CHECKED;
entry            1518 kernel/dma/debug.c 	add_dma_entry(entry);
entry              38 kernel/events/callchain.c __weak void perf_callchain_kernel(struct perf_callchain_entry_ctx *entry,
entry              43 kernel/events/callchain.c __weak void perf_callchain_user(struct perf_callchain_entry_ctx *entry,
entry             181 kernel/events/callchain.c 	struct perf_callchain_entry *entry;
entry             185 kernel/events/callchain.c 	entry = get_callchain_entry(&rctx);
entry             189 kernel/events/callchain.c 	if (!entry)
entry             192 kernel/events/callchain.c 	ctx.entry     = entry;
entry             194 kernel/events/callchain.c 	ctx.nr	      = entry->nr = init_nr;
entry             231 kernel/events/callchain.c 	return entry;
entry            1710 kernel/events/core.c 	int entry = sizeof(u64); /* value */
entry            1721 kernel/events/core.c 		entry += sizeof(u64);
entry            1728 kernel/events/core.c 	size += entry * nr;
entry            6009 kernel/events/core.c static void perf_pending_event(struct irq_work *entry)
entry            6011 kernel/events/core.c 	struct perf_event *event = container_of(entry, struct perf_event, pending);
entry            6892 kernel/events/core.c 	list_for_each_entry(filter, &ifh->list, entry) {
entry            7621 kernel/events/core.c 	list_for_each_entry(filter, &ifh->list, entry) {
entry            8752 kernel/events/core.c 		struct trace_entry *entry = record;
entry            8764 kernel/events/core.c 			if (event->attr.config != entry->type)
entry            9165 kernel/events/core.c 	INIT_LIST_HEAD(&filter->entry);
entry            9166 kernel/events/core.c 	list_add_tail(&filter->entry, filters);
entry            9175 kernel/events/core.c 	list_for_each_entry_safe(filter, iter, filters, entry) {
entry            9177 kernel/events/core.c 		list_del(&filter->entry);
entry            9258 kernel/events/core.c 	list_for_each_entry(filter, &ifh->list, entry) {
entry            9894 kernel/events/core.c 	list_for_each_entry(pmu, &pmus, entry) {
entry            10153 kernel/events/core.c 	list_add_rcu(&pmu->entry, &pmus);
entry            10178 kernel/events/core.c 	list_del_rcu(&pmu->entry);
entry            10282 kernel/events/core.c 	list_for_each_entry_rcu(pmu, &pmus, entry) {
entry            12165 kernel/events/core.c 	list_for_each_entry(pmu, &pmus, entry) {
entry            12193 kernel/events/core.c 	list_for_each_entry(pmu, &pmus, entry) {
entry            12282 kernel/events/core.c 	list_for_each_entry(pmu, &pmus, entry) {
entry            3673 kernel/futex.c static inline int fetch_robust_entry(struct robust_list __user **entry,
entry            3682 kernel/futex.c 	*entry = (void __user *)(uentry & ~1UL);
entry            3697 kernel/futex.c 	struct robust_list __user *entry, *next_entry, *pending;
entry            3710 kernel/futex.c 	if (fetch_robust_entry(&entry, &head->list.next, &pi))
entry            3725 kernel/futex.c 	while (entry != &head->list) {
entry            3730 kernel/futex.c 		rc = fetch_robust_entry(&next_entry, &entry->next, &next_pi);
entry            3735 kernel/futex.c 		if (entry != pending) {
entry            3736 kernel/futex.c 			if (handle_futex_death((void __user *)entry + futex_offset,
entry            3742 kernel/futex.c 		entry = next_entry;
entry            3967 kernel/futex.c compat_fetch_robust_entry(compat_uptr_t *uentry, struct robust_list __user **entry,
entry            3973 kernel/futex.c 	*entry = compat_ptr((*uentry) & ~1);
entry            3979 kernel/futex.c static void __user *futex_uaddr(struct robust_list __user *entry,
entry            3982 kernel/futex.c 	compat_uptr_t base = ptr_to_compat(entry);
entry            3997 kernel/futex.c 	struct robust_list __user *entry, *next_entry, *pending;
entry            4011 kernel/futex.c 	if (compat_fetch_robust_entry(&uentry, &entry, &head->list.next, &pi))
entry            4027 kernel/futex.c 	while (entry != (struct robust_list __user *) &head->list) {
entry            4033 kernel/futex.c 			(compat_uptr_t __user *)&entry->next, &next_pi);
entry            4038 kernel/futex.c 		if (entry != pending) {
entry            4039 kernel/futex.c 			void __user *uaddr = futex_uaddr(entry, futex_offset);
entry            4048 kernel/futex.c 		entry = next_entry;
entry             112 kernel/irq/chip.c 			 struct msi_desc *entry)
entry             119 kernel/irq/chip.c 	desc->irq_common_data.msi_desc = entry;
entry             120 kernel/irq/chip.c 	if (entry && !irq_offset)
entry             121 kernel/irq/chip.c 		entry->irq = irq_base;
entry             133 kernel/irq/chip.c int irq_set_msi_desc(unsigned int irq, struct msi_desc *entry)
entry             135 kernel/irq/chip.c 	return irq_set_msi_desc_off(irq, 0, entry);
entry              53 kernel/irq/msi.c void free_msi_entry(struct msi_desc *entry)
entry              55 kernel/irq/msi.c 	kfree(entry->affinity);
entry              56 kernel/irq/msi.c 	kfree(entry);
entry              59 kernel/irq/msi.c void __get_cached_msi_msg(struct msi_desc *entry, struct msi_msg *msg)
entry              61 kernel/irq/msi.c 	*msg = entry->msg;
entry              66 kernel/irq/msi.c 	struct msi_desc *entry = irq_get_msi_desc(irq);
entry              68 kernel/irq/msi.c 	__get_cached_msi_msg(entry, msg);
entry             309 kernel/jump_label.c static int addr_conflict(struct jump_entry *entry, void *start, void *end)
entry             311 kernel/jump_label.c 	if (jump_entry_code(entry) <= (unsigned long)end &&
entry             312 kernel/jump_label.c 	    jump_entry_code(entry) + JUMP_LABEL_NOP_SIZE > (unsigned long)start)
entry             339 kernel/jump_label.c void __weak __init_or_module arch_jump_label_transform_static(struct jump_entry *entry,
entry             342 kernel/jump_label.c 	arch_jump_label_transform(entry, type);
entry             391 kernel/jump_label.c static enum jump_label_type jump_label_type(struct jump_entry *entry)
entry             393 kernel/jump_label.c 	struct static_key *key = jump_entry_key(entry);
entry             395 kernel/jump_label.c 	bool branch = jump_entry_is_branch(entry);
entry             401 kernel/jump_label.c static bool jump_label_can_update(struct jump_entry *entry, bool init)
entry             406 kernel/jump_label.c 	if (!init && jump_entry_is_init(entry))
entry             409 kernel/jump_label.c 	if (!kernel_text_address(jump_entry_code(entry))) {
entry             410 kernel/jump_label.c 		WARN_ONCE(!jump_entry_is_init(entry),
entry             412 kernel/jump_label.c 			  (void *)jump_entry_code(entry));
entry             421 kernel/jump_label.c 				struct jump_entry *entry,
entry             425 kernel/jump_label.c 	for (; (entry < stop) && (jump_entry_key(entry) == key); entry++) {
entry             426 kernel/jump_label.c 		if (jump_label_can_update(entry, init))
entry             427 kernel/jump_label.c 			arch_jump_label_transform(entry, jump_label_type(entry));
entry             432 kernel/jump_label.c 				struct jump_entry *entry,
entry             436 kernel/jump_label.c 	for (; (entry < stop) && (jump_entry_key(entry) == key); entry++) {
entry             438 kernel/jump_label.c 		if (!jump_label_can_update(entry, init))
entry             441 kernel/jump_label.c 		if (!arch_jump_label_transform_queue(entry, jump_label_type(entry))) {
entry             446 kernel/jump_label.c 			BUG_ON(!arch_jump_label_transform_queue(entry, jump_label_type(entry)));
entry             500 kernel/jump_label.c static enum jump_label_type jump_label_init_type(struct jump_entry *entry)
entry             502 kernel/jump_label.c 	struct static_key *key = jump_entry_key(entry);
entry             504 kernel/jump_label.c 	bool branch = jump_entry_is_branch(entry);
entry             790 kernel/jump_label.c 	struct jump_entry *entry;
entry             805 kernel/jump_label.c 	entry = static_key_entries(key);
entry             807 kernel/jump_label.c 	if (entry)
entry             808 kernel/jump_label.c 		__jump_label_update(key, entry, stop,
entry              39 kernel/kexec.c static int kimage_alloc_init(struct kimage **rimage, unsigned long entry,
entry              50 kernel/kexec.c 		if ((entry < phys_to_boot_phys(crashk_res.start)) ||
entry              51 kernel/kexec.c 		    (entry > phys_to_boot_phys(crashk_res.end)))
entry              60 kernel/kexec.c 	image->start = entry;
entry             106 kernel/kexec.c static int do_kexec_load(unsigned long entry, unsigned long nr_segments,
entry             135 kernel/kexec.c 	ret = kimage_alloc_init(&image, entry, nr_segments, segments, flags);
entry             232 kernel/kexec.c SYSCALL_DEFINE4(kexec_load, unsigned long, entry, unsigned long, nr_segments,
entry             257 kernel/kexec.c 	result = do_kexec_load(entry, nr_segments, segments, flags);
entry             265 kernel/kexec.c COMPAT_SYSCALL_DEFINE4(kexec_load, compat_ulong_t, entry,
entry             311 kernel/kexec.c 	result = do_kexec_load(entry, nr_segments, ksegments, flags);
entry             264 kernel/kexec_core.c 	image->entry = &image->head;
entry             534 kernel/kexec_core.c static int kimage_add_entry(struct kimage *image, kimage_entry_t entry)
entry             536 kernel/kexec_core.c 	if (*image->entry != 0)
entry             537 kernel/kexec_core.c 		image->entry++;
entry             539 kernel/kexec_core.c 	if (image->entry == image->last_entry) {
entry             548 kernel/kexec_core.c 		*image->entry = virt_to_boot_phys(ind_page) | IND_INDIRECTION;
entry             549 kernel/kexec_core.c 		image->entry = ind_page;
entry             553 kernel/kexec_core.c 	*image->entry = entry;
entry             554 kernel/kexec_core.c 	image->entry++;
entry             555 kernel/kexec_core.c 	*image->entry = 0;
entry             594 kernel/kexec_core.c 	if (*image->entry != 0)
entry             595 kernel/kexec_core.c 		image->entry++;
entry             597 kernel/kexec_core.c 	*image->entry = IND_DONE;
entry             600 kernel/kexec_core.c #define for_each_kimage_entry(image, ptr, entry) \
entry             601 kernel/kexec_core.c 	for (ptr = &image->head; (entry = *ptr) && !(entry & IND_DONE); \
entry             602 kernel/kexec_core.c 		ptr = (entry & IND_INDIRECTION) ? \
entry             603 kernel/kexec_core.c 			boot_phys_to_virt((entry & PAGE_MASK)) : ptr + 1)
entry             605 kernel/kexec_core.c static void kimage_free_entry(kimage_entry_t entry)
entry             609 kernel/kexec_core.c 	page = boot_pfn_to_page(entry >> PAGE_SHIFT);
entry             615 kernel/kexec_core.c 	kimage_entry_t *ptr, entry;
entry             627 kernel/kexec_core.c 	for_each_kimage_entry(image, ptr, entry) {
entry             628 kernel/kexec_core.c 		if (entry & IND_INDIRECTION) {
entry             635 kernel/kexec_core.c 			ind = entry;
entry             636 kernel/kexec_core.c 		} else if (entry & IND_SOURCE)
entry             637 kernel/kexec_core.c 			kimage_free_entry(entry);
entry             662 kernel/kexec_core.c 	kimage_entry_t *ptr, entry;
entry             665 kernel/kexec_core.c 	for_each_kimage_entry(image, ptr, entry) {
entry             666 kernel/kexec_core.c 		if (entry & IND_DESTINATION)
entry             667 kernel/kexec_core.c 			destination = entry & PAGE_MASK;
entry             668 kernel/kexec_core.c 		else if (entry & IND_SOURCE) {
entry            1848 kernel/kprobes.c unsigned long __weak arch_deref_entry_point(void *entry)
entry            1850 kernel/kprobes.c 	return (unsigned long)entry;
entry            2146 kernel/kprobes.c int kprobe_add_ksym_blacklist(unsigned long entry)
entry            2151 kernel/kprobes.c 	if (!kernel_text_address(entry) ||
entry            2152 kernel/kprobes.c 	    !kallsyms_lookup_size_offset(entry, &size, &offset))
entry            2158 kernel/kprobes.c 	ent->start_addr = entry;
entry            2159 kernel/kprobes.c 	ent->end_addr = entry + size;
entry            2169 kernel/kprobes.c 	unsigned long entry;
entry            2172 kernel/kprobes.c 	for (entry = start; entry < end; entry += ret) {
entry            2173 kernel/kprobes.c 		ret = kprobe_add_ksym_blacklist(entry);
entry            2198 kernel/kprobes.c 	unsigned long entry;
entry            2203 kernel/kprobes.c 		entry = arch_deref_entry_point((void *)*iter);
entry            2204 kernel/kprobes.c 		ret = kprobe_add_ksym_blacklist(entry);
entry             924 kernel/locking/lockdep.c 	list_for_each_entry(e, h, entry) {
entry            1011 kernel/locking/lockdep.c 		hlist_for_each_entry_rcu(chain, head, entry) {
entry            1023 kernel/locking/lockdep.c 		if (!in_any_class_list(&e->entry)) {
entry            1038 kernel/locking/lockdep.c 		if (in_any_class_list(&e->entry)) {
entry            1307 kernel/locking/lockdep.c 	struct lock_list *entry;
entry            1312 kernel/locking/lockdep.c 	entry = alloc_list_entry();
entry            1313 kernel/locking/lockdep.c 	if (!entry)
entry            1316 kernel/locking/lockdep.c 	entry->class = this;
entry            1317 kernel/locking/lockdep.c 	entry->links_to = links_to;
entry            1318 kernel/locking/lockdep.c 	entry->distance = distance;
entry            1319 kernel/locking/lockdep.c 	entry->trace = trace;
entry            1325 kernel/locking/lockdep.c 	list_add_tail_rcu(&entry->entry, head);
entry            1462 kernel/locking/lockdep.c 		 int (*match)(struct lock_list *entry, void *data),
entry            1466 kernel/locking/lockdep.c 	struct lock_list *entry;
entry            1496 kernel/locking/lockdep.c 		list_for_each_entry_rcu(entry, head, entry) {
entry            1497 kernel/locking/lockdep.c 			if (!lock_accessed(entry)) {
entry            1499 kernel/locking/lockdep.c 				mark_lock_accessed(entry, lock);
entry            1500 kernel/locking/lockdep.c 				if (match(entry, data)) {
entry            1501 kernel/locking/lockdep.c 					*target_entry = entry;
entry            1506 kernel/locking/lockdep.c 				if (__cq_enqueue(cq, entry)) {
entry            1522 kernel/locking/lockdep.c 			int (*match)(struct lock_list *entry, void *data),
entry            1532 kernel/locking/lockdep.c 			int (*match)(struct lock_list *entry, void *data),
entry            1616 kernel/locking/lockdep.c print_circular_bug_header(struct lock_list *entry, unsigned int depth,
entry            1640 kernel/locking/lockdep.c 	print_circular_bug_entry(entry, depth);
entry            1643 kernel/locking/lockdep.c static inline int class_equal(struct lock_list *entry, void *data)
entry            1645 kernel/locking/lockdep.c 	return entry->class == data;
entry            1698 kernel/locking/lockdep.c static int noop_count(struct lock_list *entry, void *data)
entry            1852 kernel/locking/lockdep.c static inline int usage_accumulate(struct lock_list *entry, void *mask)
entry            1854 kernel/locking/lockdep.c 	*(unsigned long *)mask |= entry->class->usage_mask;
entry            1865 kernel/locking/lockdep.c static inline int usage_match(struct lock_list *entry, void *mask)
entry            1867 kernel/locking/lockdep.c 	return entry->class->usage_mask & *(unsigned long *)mask;
entry            1949 kernel/locking/lockdep.c 	struct lock_list *entry = leaf;
entry            1956 kernel/locking/lockdep.c 		print_lock_class_header(entry->class, depth);
entry            1958 kernel/locking/lockdep.c 		print_lock_trace(entry->trace, 2);
entry            1961 kernel/locking/lockdep.c 		if (depth == 0 && (entry != root)) {
entry            1966 kernel/locking/lockdep.c 		entry = get_lock_parent(entry);
entry            1968 kernel/locking/lockdep.c 	} while (entry && (depth >= 0));
entry            2448 kernel/locking/lockdep.c 	struct lock_list *entry;
entry            2504 kernel/locking/lockdep.c 	list_for_each_entry(entry, &hlock_class(prev)->locks_after, entry) {
entry            2505 kernel/locking/lockdep.c 		if (entry->class == hlock_class(next)) {
entry            2507 kernel/locking/lockdep.c 				entry->distance = 1;
entry            2848 kernel/locking/lockdep.c 	hlist_add_head_rcu(&chain->entry, hash_head);
entry            2864 kernel/locking/lockdep.c 	hlist_for_each_entry_rcu(chain, hash_head, entry) {
entry            3131 kernel/locking/lockdep.c 	struct lock_list *entry = other;
entry            3158 kernel/locking/lockdep.c 		if (depth == 0 && (entry != root)) {
entry            3162 kernel/locking/lockdep.c 		middle = entry;
entry            3163 kernel/locking/lockdep.c 		entry = get_lock_parent(entry);
entry            3165 kernel/locking/lockdep.c 	} while (entry && entry != root && (depth >= 0));
entry            4812 kernel/locking/lockdep.c 	hlist_del_rcu(&chain->entry);
entry            4828 kernel/locking/lockdep.c 	hlist_add_head_rcu(&new_chain->entry, chainhashentry(chain_key));
entry            4842 kernel/locking/lockdep.c 		hlist_for_each_entry_rcu(chain, head, entry) {
entry            4853 kernel/locking/lockdep.c 	struct lock_list *entry;
entry            4863 kernel/locking/lockdep.c 		entry = list_entries + i;
entry            4864 kernel/locking/lockdep.c 		if (entry->class != class && entry->links_to != class)
entry            4868 kernel/locking/lockdep.c 		list_del_rcu(&entry->entry);
entry              61 kernel/locking/lockdep_proc.c 	struct lock_list *entry;
entry              85 kernel/locking/lockdep_proc.c 	list_for_each_entry(entry, &class->locks_after, entry) {
entry              86 kernel/locking/lockdep_proc.c 		if (entry->distance == 1) {
entry              87 kernel/locking/lockdep_proc.c 			seq_printf(m, " -> [%p] ", entry->class->key);
entry              88 kernel/locking/lockdep_proc.c 			print_name(m, entry->class);
entry             277 kernel/locking/rtmutex.c 	struct rt_mutex_waiter *entry;
entry             282 kernel/locking/rtmutex.c 		entry = rb_entry(parent, struct rt_mutex_waiter, tree_entry);
entry             283 kernel/locking/rtmutex.c 		if (rt_mutex_waiter_less(waiter, entry)) {
entry             310 kernel/locking/rtmutex.c 	struct rt_mutex_waiter *entry;
entry             315 kernel/locking/rtmutex.c 		entry = rb_entry(parent, struct rt_mutex_waiter, pi_tree_entry);
entry             316 kernel/locking/rtmutex.c 		if (rt_mutex_waiter_less(waiter, entry)) {
entry              48 kernel/power/console.c 	struct pm_vt_switch *entry, *tmp;
entry              59 kernel/power/console.c 	entry = kmalloc(sizeof(*entry), GFP_KERNEL);
entry              60 kernel/power/console.c 	if (!entry)
entry              63 kernel/power/console.c 	entry->required = required;
entry              64 kernel/power/console.c 	entry->dev = dev;
entry              66 kernel/power/console.c 	list_add(&entry->head, &pm_vt_switch_list);
entry             109 kernel/power/console.c 	struct pm_vt_switch *entry;
entry             119 kernel/power/console.c 	list_for_each_entry(entry, &pm_vt_switch_list, head) {
entry             120 kernel/power/console.c 		if (entry->required)
entry             528 kernel/profile.c 	struct proc_dir_entry *entry;
entry             550 kernel/profile.c 	entry = proc_create("profile", S_IWUSR | S_IRUGO,
entry             552 kernel/profile.c 	if (!entry)
entry             554 kernel/profile.c 	proc_set_size(entry, (1 + prof_len) * sizeof(atomic_t));
entry             912 kernel/ptrace.c 	unsigned long args[ARRAY_SIZE(info->entry.args)];
entry             916 kernel/ptrace.c 	info->entry.nr = syscall_get_nr(child, regs);
entry             919 kernel/ptrace.c 		info->entry.args[i] = args[i];
entry             922 kernel/ptrace.c 	return offsetofend(struct ptrace_syscall_info, entry.args);
entry             969 kernel/ptrace.c 	unsigned long actual_size = offsetof(struct ptrace_syscall_info, entry);
entry              38 kernel/rcu/srcutiny.c 	INIT_LIST_HEAD(&ssp->srcu_work.entry);
entry             172 kernel/rcu/srcutiny.c 		else if (list_empty(&ssp->srcu_work.entry))
entry             173 kernel/rcu/srcutiny.c 			list_add(&ssp->srcu_work.entry, &srcu_boot_list);
entry             211 kernel/rcu/srcutiny.c 				      struct srcu_struct, srcu_work.entry);
entry             212 kernel/rcu/srcutiny.c 		list_del_init(&ssp->srcu_work.entry);
entry             687 kernel/rcu/srcutree.c 		else if (list_empty(&ssp->work.work.entry))
entry             688 kernel/rcu/srcutree.c 			list_add(&ssp->work.work.entry, &srcu_boot_list);
entry            1309 kernel/rcu/srcutree.c 				      work.work.entry);
entry            1311 kernel/rcu/srcutree.c 		list_del_init(&ssp->work.work.entry);
entry            1625 kernel/resource.c 	struct resource_entry *entry;
entry            1627 kernel/resource.c 	entry = kzalloc(sizeof(*entry) + extra_size, GFP_KERNEL);
entry            1628 kernel/resource.c 	if (entry) {
entry            1629 kernel/resource.c 		INIT_LIST_HEAD(&entry->node);
entry            1630 kernel/resource.c 		entry->res = res ? res : &entry->__res;
entry            1633 kernel/resource.c 	return entry;
entry            1639 kernel/resource.c 	struct resource_entry *entry, *tmp;
entry            1641 kernel/resource.c 	list_for_each_entry_safe(entry, tmp, head, node)
entry            1642 kernel/resource.c 		resource_list_destroy_entry(entry);
entry             450 kernel/sched/deadline.c 	struct task_struct *entry;
entry             457 kernel/sched/deadline.c 		entry = rb_entry(parent, struct task_struct,
entry             459 kernel/sched/deadline.c 		if (dl_entity_preempt(&p->dl, &entry->dl))
entry            1359 kernel/sched/deadline.c 		struct sched_dl_entity *entry;
entry            1361 kernel/sched/deadline.c 		entry = rb_entry(leftmost, struct sched_dl_entity, rb_node);
entry            1362 kernel/sched/deadline.c 		dl_rq->earliest_dl.curr = entry->deadline;
entry            1363 kernel/sched/deadline.c 		cpudl_set(&rq->rd->cpudl, rq->cpu, entry->deadline);
entry            1407 kernel/sched/deadline.c 	struct sched_dl_entity *entry;
entry            1414 kernel/sched/deadline.c 		entry = rb_entry(parent, struct sched_dl_entity, rb_node);
entry            1415 kernel/sched/deadline.c 		if (dl_time_before(dl_se->deadline, entry->deadline))
entry             209 kernel/sched/debug.c 	struct ctl_table *entry =
entry             212 kernel/sched/debug.c 	return entry;
entry             217 kernel/sched/debug.c 	struct ctl_table *entry;
entry             225 kernel/sched/debug.c 	for (entry = *tablep; entry->mode; entry++) {
entry             226 kernel/sched/debug.c 		if (entry->child)
entry             227 kernel/sched/debug.c 			sd_free_ctl_entry(&entry->child);
entry             228 kernel/sched/debug.c 		if (entry->proc_handler == NULL)
entry             229 kernel/sched/debug.c 			kfree(entry->procname);
entry             237 kernel/sched/debug.c set_table_entry(struct ctl_table *entry,
entry             241 kernel/sched/debug.c 	entry->procname = procname;
entry             242 kernel/sched/debug.c 	entry->data = data;
entry             243 kernel/sched/debug.c 	entry->maxlen = maxlen;
entry             244 kernel/sched/debug.c 	entry->mode = mode;
entry             245 kernel/sched/debug.c 	entry->proc_handler = proc_handler;
entry             271 kernel/sched/debug.c 	struct ctl_table *entry, *table;
entry             278 kernel/sched/debug.c 	entry = table = sd_alloc_ctl_entry(domain_num + 1);
entry             285 kernel/sched/debug.c 		entry->procname = kstrdup(buf, GFP_KERNEL);
entry             286 kernel/sched/debug.c 		entry->mode = 0555;
entry             287 kernel/sched/debug.c 		entry->child = sd_alloc_ctl_domain_table(sd);
entry             288 kernel/sched/debug.c 		entry++;
entry             569 kernel/sched/fair.c 	struct sched_entity *entry;
entry             577 kernel/sched/fair.c 		entry = rb_entry(parent, struct sched_entity, run_node);
entry             582 kernel/sched/fair.c 		if (entity_before(se, entry)) {
entry              76 kernel/sched/wait.c 		curr = list_next_entry(bookmark, entry);
entry              78 kernel/sched/wait.c 		list_del(&bookmark->entry);
entry              81 kernel/sched/wait.c 		curr = list_first_entry(&wq_head->head, wait_queue_entry_t, entry);
entry              83 kernel/sched/wait.c 	if (&curr->entry == &wq_head->head)
entry              86 kernel/sched/wait.c 	list_for_each_entry_safe_from(curr, next, &wq_head->head, entry) {
entry             100 kernel/sched/wait.c 				(&next->entry != &wq_head->head)) {
entry             102 kernel/sched/wait.c 			list_add_tail(&bookmark->entry, &next->entry);
entry             119 kernel/sched/wait.c 	INIT_LIST_HEAD(&bookmark.entry);
entry             228 kernel/sched/wait.c 	if (list_empty(&wq_entry->entry))
entry             242 kernel/sched/wait.c 	if (list_empty(&wq_entry->entry))
entry             254 kernel/sched/wait.c 	INIT_LIST_HEAD(&wq_entry->entry);
entry             277 kernel/sched/wait.c 		list_del_init(&wq_entry->entry);
entry             280 kernel/sched/wait.c 		if (list_empty(&wq_entry->entry)) {
entry             303 kernel/sched/wait.c 	if (likely(list_empty(&wait->entry)))
entry             320 kernel/sched/wait.c 	if (likely(list_empty(&wait->entry)))
entry             362 kernel/sched/wait.c 	if (!list_empty_careful(&wq_entry->entry)) {
entry             364 kernel/sched/wait.c 		list_del_init(&wq_entry->entry);
entry             375 kernel/sched/wait.c 		list_del_init(&wq_entry->entry);
entry             184 kernel/sched/wait_bit.c 			.entry	 = LIST_HEAD_INIT(wbq_entry->wq_entry.entry),
entry             213 kernel/smp.c   	struct llist_node *entry;
entry             220 kernel/smp.c   	entry = llist_del_all(head);
entry             221 kernel/smp.c   	entry = llist_reverse_order(entry);
entry             233 kernel/smp.c   		llist_for_each_entry(csd, entry, llist)
entry             238 kernel/smp.c   	llist_for_each_entry_safe(csd, csd_next, entry, llist) {
entry             834 kernel/time/clocksource.c 	struct list_head *entry = &clocksource_list;
entry             841 kernel/time/clocksource.c 		entry = &tmp->list;
entry             843 kernel/time/clocksource.c 	list_add(&cs->list, entry);
entry             540 kernel/time/timer.c 	hlist_add_head(&timer->entry, base->vectors + idx);
entry             612 kernel/time/timer.c 	return (timer->entry.pprev == NULL &&
entry             613 kernel/time/timer.c 		timer->entry.next == TIMER_ENTRY_STATIC);
entry             782 kernel/time/timer.c 	timer->entry.pprev = NULL;
entry             811 kernel/time/timer.c 	struct hlist_node *entry = &timer->entry;
entry             815 kernel/time/timer.c 	__hlist_del(entry);
entry             817 kernel/time/timer.c 		entry->pprev = NULL;
entry             818 kernel/time/timer.c 	entry->next = LIST_POISON2;
entry             829 kernel/time/timer.c 	if (hlist_is_singular_node(&timer->entry, base->vectors + idx))
entry            1435 kernel/time/timer.c 		timer = hlist_entry(head->first, struct timer_list, entry);
entry            1951 kernel/time/timer.c 		timer = hlist_entry(head->first, struct timer_list, entry);
entry             629 kernel/trace/bpf_trace.c static void do_bpf_send_signal(struct irq_work *entry)
entry             633 kernel/trace/bpf_trace.c 	work = container_of(entry, struct send_signal_irq_work, irq_work);
entry             964 kernel/trace/ftrace.c 	struct dentry *entry;
entry             995 kernel/trace/ftrace.c 	entry = tracefs_create_file("function_profile_enabled", 0644,
entry             997 kernel/trace/ftrace.c 	if (!entry)
entry            1129 kernel/trace/ftrace.c 	struct ftrace_func_entry *entry;
entry            1135 kernel/trace/ftrace.c 	hlist_for_each_entry_rcu_notrace(entry, hhd, hlist) {
entry            1136 kernel/trace/ftrace.c 		if (entry->ip == ip)
entry            1137 kernel/trace/ftrace.c 			return entry;
entry            1162 kernel/trace/ftrace.c 			     struct ftrace_func_entry *entry)
entry            1167 kernel/trace/ftrace.c 	key = ftrace_hash_key(hash, entry->ip);
entry            1169 kernel/trace/ftrace.c 	hlist_add_head(&entry->hlist, hhd);
entry            1175 kernel/trace/ftrace.c 	struct ftrace_func_entry *entry;
entry            1177 kernel/trace/ftrace.c 	entry = kmalloc(sizeof(*entry), GFP_KERNEL);
entry            1178 kernel/trace/ftrace.c 	if (!entry)
entry            1181 kernel/trace/ftrace.c 	entry->ip = ip;
entry            1182 kernel/trace/ftrace.c 	__add_hash_entry(hash, entry);
entry            1189 kernel/trace/ftrace.c 		  struct ftrace_func_entry *entry)
entry            1191 kernel/trace/ftrace.c 	hlist_del(&entry->hlist);
entry            1192 kernel/trace/ftrace.c 	kfree(entry);
entry            1198 kernel/trace/ftrace.c 		  struct ftrace_func_entry *entry)
entry            1200 kernel/trace/ftrace.c 	hlist_del_rcu(&entry->hlist);
entry            1208 kernel/trace/ftrace.c 	struct ftrace_func_entry *entry;
entry            1217 kernel/trace/ftrace.c 		hlist_for_each_entry_safe(entry, tn, hhd, hlist)
entry            1218 kernel/trace/ftrace.c 			free_hash_entry(hash, entry);
entry            1330 kernel/trace/ftrace.c 	struct ftrace_func_entry *entry;
entry            1349 kernel/trace/ftrace.c 		hlist_for_each_entry(entry, &hash->buckets[i], hlist) {
entry            1350 kernel/trace/ftrace.c 			ret = add_hash_entry(new_hash, entry->ip);
entry            1376 kernel/trace/ftrace.c 	struct ftrace_func_entry *entry;
entry            1409 kernel/trace/ftrace.c 		hlist_for_each_entry_safe(entry, tn, hhd, hlist) {
entry            1410 kernel/trace/ftrace.c 			remove_hash_entry(src, entry);
entry            1411 kernel/trace/ftrace.c 			__add_hash_entry(new_hash, entry);
entry            3714 kernel/trace/ftrace.c 	struct ftrace_func_entry *entry;
entry            3717 kernel/trace/ftrace.c 	entry = ftrace_lookup_ip(hash, rec->ip);
entry            3720 kernel/trace/ftrace.c 		if (!entry)
entry            3723 kernel/trace/ftrace.c 		free_hash_entry(hash, entry);
entry            3726 kernel/trace/ftrace.c 		if (entry)
entry            4132 kernel/trace/ftrace.c 	struct ftrace_func_entry	entry;
entry            4172 kernel/trace/ftrace.c 	struct ftrace_func_entry *entry;
entry            4175 kernel/trace/ftrace.c 	entry = ftrace_lookup_ip(&mapper->hash, ip);
entry            4176 kernel/trace/ftrace.c 	if (!entry)
entry            4179 kernel/trace/ftrace.c 	map = (struct ftrace_func_map *)entry;
entry            4194 kernel/trace/ftrace.c 	struct ftrace_func_entry *entry;
entry            4197 kernel/trace/ftrace.c 	entry = ftrace_lookup_ip(&mapper->hash, ip);
entry            4198 kernel/trace/ftrace.c 	if (entry)
entry            4205 kernel/trace/ftrace.c 	map->entry.ip = ip;
entry            4208 kernel/trace/ftrace.c 	__add_hash_entry(&mapper->hash, &map->entry);
entry            4226 kernel/trace/ftrace.c 	struct ftrace_func_entry *entry;
entry            4230 kernel/trace/ftrace.c 	entry = ftrace_lookup_ip(&mapper->hash, ip);
entry            4231 kernel/trace/ftrace.c 	if (!entry)
entry            4234 kernel/trace/ftrace.c 	map = (struct ftrace_func_map *)entry;
entry            4237 kernel/trace/ftrace.c 	remove_hash_entry(&mapper->hash, entry);
entry            4238 kernel/trace/ftrace.c 	kfree(entry);
entry            4254 kernel/trace/ftrace.c 	struct ftrace_func_entry *entry;
entry            4266 kernel/trace/ftrace.c 			hlist_for_each_entry(entry, hhd, hlist) {
entry            4267 kernel/trace/ftrace.c 				map = (struct ftrace_func_map *)entry;
entry            4314 kernel/trace/ftrace.c 	struct ftrace_func_entry *entry;
entry            4381 kernel/trace/ftrace.c 		hlist_for_each_entry(entry, &hash->buckets[i], hlist) {
entry            4382 kernel/trace/ftrace.c 			if (ftrace_lookup_ip(old_hash, entry->ip))
entry            4391 kernel/trace/ftrace.c 						      entry->ip, data,
entry            4443 kernel/trace/ftrace.c 		hlist_for_each_entry(entry, &hash->buckets[i], hlist) {
entry            4444 kernel/trace/ftrace.c 			if (ftrace_lookup_ip(old_hash, entry->ip))
entry            4446 kernel/trace/ftrace.c 			probe_ops->free(probe_ops, tr, entry->ip, probe->data);
entry            4457 kernel/trace/ftrace.c 	struct ftrace_func_entry *entry;
entry            4522 kernel/trace/ftrace.c 		hlist_for_each_entry_safe(entry, tmp, &hash->buckets[i], hlist) {
entry            4525 kernel/trace/ftrace.c 				kallsyms_lookup(entry->ip, NULL, NULL,
entry            4531 kernel/trace/ftrace.c 			remove_hash_entry(hash, entry);
entry            4532 kernel/trace/ftrace.c 			hlist_add_head(&entry->hlist, &hhd);
entry            4560 kernel/trace/ftrace.c 	hlist_for_each_entry_safe(entry, tmp, &hhd, hlist) {
entry            4561 kernel/trace/ftrace.c 		hlist_del(&entry->hlist);
entry            4563 kernel/trace/ftrace.c 			probe_ops->free(probe_ops, tr, entry->ip, probe->data);
entry            4564 kernel/trace/ftrace.c 		kfree(entry);
entry            4731 kernel/trace/ftrace.c 	struct ftrace_func_entry *entry;
entry            4737 kernel/trace/ftrace.c 		entry = ftrace_lookup_ip(hash, ip);
entry            4738 kernel/trace/ftrace.c 		if (!entry)
entry            4740 kernel/trace/ftrace.c 		free_hash_entry(hash, entry);
entry            5117 kernel/trace/ftrace.c 	struct ftrace_func_entry	*entry;
entry            5129 kernel/trace/ftrace.c 	struct ftrace_func_entry *entry = fgd->entry;
entry            5136 kernel/trace/ftrace.c 	if (entry) {
entry            5137 kernel/trace/ftrace.c 		hlist_for_each_entry_continue(entry, hlist) {
entry            5138 kernel/trace/ftrace.c 			fgd->entry = entry;
entry            5139 kernel/trace/ftrace.c 			return entry;
entry            5147 kernel/trace/ftrace.c 		hlist_for_each_entry(entry, head, hlist) {
entry            5148 kernel/trace/ftrace.c 			fgd->entry = entry;
entry            5150 kernel/trace/ftrace.c 			return entry;
entry            5181 kernel/trace/ftrace.c 	fgd->entry = NULL;
entry            5192 kernel/trace/ftrace.c 	struct ftrace_func_entry *entry = v;
entry            5194 kernel/trace/ftrace.c 	if (!entry)
entry            5197 kernel/trace/ftrace.c 	if (entry == FTRACE_GRAPH_EMPTY) {
entry            5207 kernel/trace/ftrace.c 	seq_printf(m, "%ps\n", (void *)entry->ip);
entry            5407 kernel/trace/ftrace.c 	struct ftrace_func_entry *entry;
entry            5430 kernel/trace/ftrace.c 			entry = ftrace_lookup_ip(hash, rec->ip);
entry            5435 kernel/trace/ftrace.c 				if (entry)
entry            5440 kernel/trace/ftrace.c 				if (entry) {
entry            5441 kernel/trace/ftrace.c 					free_hash_entry(hash, entry);
entry            5712 kernel/trace/ftrace.c 	struct ftrace_func_entry *entry;
entry            5721 kernel/trace/ftrace.c 		entry = __ftrace_lookup_ip(hash, rec->ip);
entry            5727 kernel/trace/ftrace.c 		if (entry)
entry            5728 kernel/trace/ftrace.c 			entry->ip = 0;
entry            6074 kernel/trace/ftrace.c 	struct ftrace_func_entry *entry;
entry            6076 kernel/trace/ftrace.c 	entry = ftrace_lookup_ip(hash, func->ip);
entry            6082 kernel/trace/ftrace.c 	if (entry)
entry            6083 kernel/trace/ftrace.c 		entry->ip = 0;
entry              88 kernel/trace/ring_buffer_benchmark.c 	int *entry;
entry              95 kernel/trace/ring_buffer_benchmark.c 	entry = ring_buffer_event_data(event);
entry              96 kernel/trace/ring_buffer_benchmark.c 	if (*entry != cpu) {
entry             111 kernel/trace/ring_buffer_benchmark.c 	int *entry;
entry             145 kernel/trace/ring_buffer_benchmark.c 				entry = ring_buffer_event_data(event);
entry             146 kernel/trace/ring_buffer_benchmark.c 				if (*entry != cpu) {
entry             158 kernel/trace/ring_buffer_benchmark.c 				entry = ring_buffer_event_data(event);
entry             159 kernel/trace/ring_buffer_benchmark.c 				if (*entry != cpu) {
entry             250 kernel/trace/ring_buffer_benchmark.c 		int *entry;
entry             259 kernel/trace/ring_buffer_benchmark.c 				entry = ring_buffer_event_data(event);
entry             260 kernel/trace/ring_buffer_benchmark.c 				*entry = smp_processor_id();
entry             837 kernel/trace/trace.c 	struct print_entry *entry;
entry             850 kernel/trace/trace.c 	alloc = sizeof(*entry) + size + 2; /* possible \n added */
entry             859 kernel/trace/trace.c 	entry = ring_buffer_event_data(event);
entry             860 kernel/trace/trace.c 	entry->ip = ip;
entry             862 kernel/trace/trace.c 	memcpy(&entry->buf, str, size);
entry             865 kernel/trace/trace.c 	if (entry->buf[size - 1] != '\n') {
entry             866 kernel/trace/trace.c 		entry->buf[size] = '\n';
entry             867 kernel/trace/trace.c 		entry->buf[size + 1] = '\0';
entry             869 kernel/trace/trace.c 		entry->buf[size] = '\0';
entry             887 kernel/trace/trace.c 	struct bputs_entry *entry;
entry             907 kernel/trace/trace.c 	entry = ring_buffer_event_data(event);
entry             908 kernel/trace/trace.c 	entry->ip			= ip;
entry             909 kernel/trace/trace.c 	entry->str			= str;
entry            2340 kernel/trace/trace.c tracing_generic_entry_update(struct trace_entry *entry, unsigned short type,
entry            2345 kernel/trace/trace.c 	entry->preempt_count		= pc & 0xff;
entry            2346 kernel/trace/trace.c 	entry->pid			= (tsk) ? tsk->pid : 0;
entry            2347 kernel/trace/trace.c 	entry->type			= type;
entry            2348 kernel/trace/trace.c 	entry->flags =
entry            2490 kernel/trace/trace.c 	struct ring_buffer_event *entry;
entry            2497 kernel/trace/trace.c 	    (entry = this_cpu_read(trace_buffered_event))) {
entry            2501 kernel/trace/trace.c 			trace_event_setup(entry, type, flags, pc);
entry            2502 kernel/trace/trace.c 			entry->array[0] = len;
entry            2503 kernel/trace/trace.c 			return entry;
entry            2508 kernel/trace/trace.c 	entry = __trace_buffer_lock_reserve(*current_rb,
entry            2516 kernel/trace/trace.c 	if (!entry && trace_file->flags & EVENT_FILE_FL_TRIGGER_COND) {
entry            2518 kernel/trace/trace.c 		entry = __trace_buffer_lock_reserve(*current_rb,
entry            2521 kernel/trace/trace.c 	return entry;
entry            2548 kernel/trace/trace.c 	iter->ent = fbuffer->entry;
entry            2595 kernel/trace/trace.c 				    fbuffer->event, fbuffer->entry,
entry            2641 kernel/trace/trace.c 	struct trace_entry *entry;
entry            2644 kernel/trace/trace.c 	entry = ring_buffer_event_data(event);
entry            2646 kernel/trace/trace.c 	export->write(export, entry, size);
entry            2768 kernel/trace/trace.c 	struct ftrace_entry *entry;
entry            2770 kernel/trace/trace.c 	event = __trace_buffer_lock_reserve(buffer, TRACE_FN, sizeof(*entry),
entry            2774 kernel/trace/trace.c 	entry	= ring_buffer_event_data(event);
entry            2775 kernel/trace/trace.c 	entry->ip			= ip;
entry            2776 kernel/trace/trace.c 	entry->parent_ip		= parent_ip;
entry            2778 kernel/trace/trace.c 	if (!call_filter_check_discard(call, entry, buffer, event)) {
entry            2812 kernel/trace/trace.c 	struct stack_entry *entry;
entry            2859 kernel/trace/trace.c 					    sizeof(*entry) + size, flags, pc);
entry            2862 kernel/trace/trace.c 	entry = ring_buffer_event_data(event);
entry            2864 kernel/trace/trace.c 	memcpy(&entry->caller, fstack->calls, size);
entry            2865 kernel/trace/trace.c 	entry->size = nr_entries;
entry            2867 kernel/trace/trace.c 	if (!call_filter_check_discard(call, entry, buffer, event))
entry            2943 kernel/trace/trace.c 	struct userstack_entry *entry;
entry            2966 kernel/trace/trace.c 					    sizeof(*entry), flags, pc);
entry            2969 kernel/trace/trace.c 	entry	= ring_buffer_event_data(event);
entry            2971 kernel/trace/trace.c 	entry->tgid		= current->tgid;
entry            2972 kernel/trace/trace.c 	memset(&entry->caller, 0, sizeof(entry->caller));
entry            2974 kernel/trace/trace.c 	stack_trace_save_user(entry->caller, FTRACE_STACK_ENTRIES);
entry            2975 kernel/trace/trace.c 	if (!call_filter_check_discard(call, entry, buffer, event))
entry            3111 kernel/trace/trace.c 	struct bprint_entry *entry;
entry            3137 kernel/trace/trace.c 	size = sizeof(*entry) + sizeof(u32) * len;
entry            3143 kernel/trace/trace.c 	entry = ring_buffer_event_data(event);
entry            3144 kernel/trace/trace.c 	entry->ip			= ip;
entry            3145 kernel/trace/trace.c 	entry->fmt			= fmt;
entry            3147 kernel/trace/trace.c 	memcpy(entry->buf, tbuffer, sizeof(u32) * len);
entry            3148 kernel/trace/trace.c 	if (!call_filter_check_discard(call, entry, buffer, event)) {
entry            3172 kernel/trace/trace.c 	struct print_entry *entry;
entry            3195 kernel/trace/trace.c 	size = sizeof(*entry) + len + 1;
entry            3200 kernel/trace/trace.c 	entry = ring_buffer_event_data(event);
entry            3201 kernel/trace/trace.c 	entry->ip = ip;
entry            3203 kernel/trace/trace.c 	memcpy(&entry->buf, tbuffer, len + 1);
entry            3204 kernel/trace/trace.c 	if (!call_filter_check_discard(call, entry, buffer, event)) {
entry            3728 kernel/trace/trace.c 	struct trace_entry *entry;
entry            3731 kernel/trace/trace.c 	entry = iter->ent;
entry            3735 kernel/trace/trace.c 	event = ftrace_find_event(entry->type);
entry            3750 kernel/trace/trace.c 	trace_seq_printf(s, "Unknown type %d\n", entry->type);
entry            3759 kernel/trace/trace.c 	struct trace_entry *entry;
entry            3762 kernel/trace/trace.c 	entry = iter->ent;
entry            3766 kernel/trace/trace.c 				 entry->pid, iter->cpu, iter->ts);
entry            3771 kernel/trace/trace.c 	event = ftrace_find_event(entry->type);
entry            3775 kernel/trace/trace.c 	trace_seq_printf(s, "%d ?\n", entry->type);
entry            3785 kernel/trace/trace.c 	struct trace_entry *entry;
entry            3788 kernel/trace/trace.c 	entry = iter->ent;
entry            3791 kernel/trace/trace.c 		SEQ_PUT_HEX_FIELD(s, entry->pid);
entry            3798 kernel/trace/trace.c 	event = ftrace_find_event(entry->type);
entry            3814 kernel/trace/trace.c 	struct trace_entry *entry;
entry            3817 kernel/trace/trace.c 	entry = iter->ent;
entry            3820 kernel/trace/trace.c 		SEQ_PUT_FIELD(s, entry->pid);
entry            3827 kernel/trace/trace.c 	event = ftrace_find_event(entry->type);
entry            6387 kernel/trace/trace.c 	struct print_entry *entry;
entry            6409 kernel/trace/trace.c 	size = sizeof(*entry) + cnt + 2; /* add '\0' and possible '\n' */
entry            6422 kernel/trace/trace.c 	entry = ring_buffer_event_data(event);
entry            6423 kernel/trace/trace.c 	entry->ip = _THIS_IP_;
entry            6425 kernel/trace/trace.c 	len = __copy_from_user_inatomic(&entry->buf, ubuf, cnt);
entry            6427 kernel/trace/trace.c 		memcpy(&entry->buf, FAULTED_STR, FAULTED_SIZE);
entry            6436 kernel/trace/trace.c 		entry->buf[cnt] = '\0';
entry            6437 kernel/trace/trace.c 		tt = event_triggers_call(tr->trace_marker_file, entry, event);
entry            6440 kernel/trace/trace.c 	if (entry->buf[cnt - 1] != '\n') {
entry            6441 kernel/trace/trace.c 		entry->buf[cnt] = '\n';
entry            6442 kernel/trace/trace.c 		entry->buf[cnt + 1] = '\0';
entry            6444 kernel/trace/trace.c 		entry->buf[cnt] = '\0';
entry            6467 kernel/trace/trace.c 	struct raw_data_entry *entry;
entry            6491 kernel/trace/trace.c 	size = sizeof(*entry) + cnt;
entry            6502 kernel/trace/trace.c 	entry = ring_buffer_event_data(event);
entry            6504 kernel/trace/trace.c 	len = __copy_from_user_inatomic(&entry->id, ubuf, cnt);
entry            6506 kernel/trace/trace.c 		entry->id = -1;
entry            6507 kernel/trace/trace.c 		memcpy(&entry->buf, FAULTED_STR, FAULTED_SIZE);
entry            8062 kernel/trace/trace.c 	topt->entry = trace_create_file(opt->name, 0644, t_options, topt,
entry            8122 kernel/trace/trace.c 		WARN_ONCE(topts[cnt].entry == NULL,
entry             369 kernel/trace/trace.h #define IF_ASSIGN(var, entry, etype, id)			\
entry             371 kernel/trace/trace.h 		var = (typeof(var))(entry);			\
entry             372 kernel/trace/trace.h 		WARN_ON(id != 0 && (entry)->type != id);	\
entry             443 kernel/trace/trace.h 	struct dentry			*entry;
entry            1365 kernel/trace/trace.h 	struct dentry			*entry;
entry            1422 kernel/trace/trace.h 			     void *entry,
entry            1428 kernel/trace/trace.h 		*tt = event_triggers_call(file, entry, event);
entry            1432 kernel/trace/trace.h 	     !filter_match_preds(file->filter, entry))) {
entry            1457 kernel/trace/trace.h 			    void *entry, unsigned long irq_flags, int pc)
entry            1461 kernel/trace/trace.h 	if (!__event_trigger_test_discard(file, buffer, event, entry, &tt))
entry            1488 kernel/trace/trace.h 				 void *entry, unsigned long irq_flags, int pc,
entry            1493 kernel/trace/trace.h 	if (!__event_trigger_test_discard(file, buffer, event, entry, &tt))
entry              37 kernel/trace/trace_branch.c 	struct trace_branch *entry;
entry              65 kernel/trace/trace_branch.c 					  sizeof(*entry), flags, pc);
entry              69 kernel/trace/trace_branch.c 	entry	= ring_buffer_event_data(event);
entry              77 kernel/trace/trace_branch.c 	strncpy(entry->func, f->data.func, TRACE_FUNC_SIZE);
entry              78 kernel/trace/trace_branch.c 	strncpy(entry->file, p, TRACE_FILE_SIZE);
entry              79 kernel/trace/trace_branch.c 	entry->func[TRACE_FUNC_SIZE] = 0;
entry              80 kernel/trace/trace_branch.c 	entry->file[TRACE_FILE_SIZE] = 0;
entry              81 kernel/trace/trace_branch.c 	entry->constant = f->constant;
entry              82 kernel/trace/trace_branch.c 	entry->line = f->data.line;
entry              83 kernel/trace/trace_branch.c 	entry->correct = val == expect;
entry              85 kernel/trace/trace_branch.c 	if (!call_filter_check_discard(call, entry, buffer, event))
entry             210 kernel/trace/trace_dynevent.c 	struct dentry *entry;
entry             216 kernel/trace/trace_dynevent.c 	entry = tracefs_create_file("dynamic_events", 0644, d_tracer,
entry             220 kernel/trace/trace_dynevent.c 	if (!entry)
entry             418 kernel/trace/trace_event_perf.c 	struct trace_entry *entry = record;
entry             423 kernel/trace/trace_event_perf.c 	tracing_generic_entry_update(entry, type, flags, pc);
entry             432 kernel/trace/trace_event_perf.c 	struct ftrace_entry *entry;
entry             459 kernel/trace/trace_event_perf.c 	entry = perf_trace_buf_alloc(ENTRY_SIZE, NULL, &rctx);
entry             460 kernel/trace/trace_event_perf.c 	if (!entry)
entry             463 kernel/trace/trace_event_perf.c 	entry->ip = ip;
entry             464 kernel/trace/trace_event_perf.c 	entry->parent_ip = parent_ip;
entry             465 kernel/trace/trace_event_perf.c 	perf_trace_buf_submit(entry, ENTRY_SIZE, rctx, TRACE_FN,
entry             275 kernel/trace/trace_events.c 	fbuffer->entry = ring_buffer_event_data(fbuffer->event);
entry             276 kernel/trace/trace_events.c 	return fbuffer->entry;
entry             700 kernel/trace/trace_events.c 		tracefs_remove_recursive(dir->entry);
entry            1879 kernel/trace/trace_events.c 	struct dentry *entry;
entry            1887 kernel/trace/trace_events.c 			return dir->entry;
entry            1911 kernel/trace/trace_events.c 	dir->entry = tracefs_create_dir(name, parent);
entry            1912 kernel/trace/trace_events.c 	if (!dir->entry) {
entry            1924 kernel/trace/trace_events.c 	entry = tracefs_create_file("filter", 0644, dir->entry, dir,
entry            1926 kernel/trace/trace_events.c 	if (!entry) {
entry            1932 kernel/trace/trace_events.c 	trace_create_file("enable", 0644, dir->entry, dir,
entry            1937 kernel/trace/trace_events.c 	return dir->entry;
entry            2912 kernel/trace/trace_events.c 	struct dentry *entry;
entry            2914 kernel/trace/trace_events.c 	entry = tracefs_create_file("set_event", 0644, parent,
entry            2916 kernel/trace/trace_events.c 	if (!entry) {
entry            2927 kernel/trace/trace_events.c 	entry = trace_create_file("enable", 0644, d_events,
entry            2929 kernel/trace/trace_events.c 	if (!entry) {
entry            2936 kernel/trace/trace_events.c 	entry = tracefs_create_file("set_event_pid", 0644, parent,
entry            2938 kernel/trace/trace_events.c 	if (!entry)
entry            2942 kernel/trace/trace_events.c 	entry = trace_create_file("header_page", 0444, d_events,
entry            2945 kernel/trace/trace_events.c 	if (!entry)
entry            2948 kernel/trace/trace_events.c 	entry = trace_create_file("header_event", 0444, d_events,
entry            2951 kernel/trace/trace_events.c 	if (!entry)
entry            3143 kernel/trace/trace_events.c 	struct dentry *entry;
entry            3154 kernel/trace/trace_events.c 	entry = tracefs_create_file("available_events", 0444, d_tracer,
entry            3156 kernel/trace/trace_events.c 	if (!entry)
entry            3363 kernel/trace/trace_events.c 	struct ftrace_entry *entry;
entry            3380 kernel/trace/trace_events.c 						TRACE_FN, sizeof(*entry),
entry            3384 kernel/trace/trace_events.c 	entry	= ring_buffer_event_data(event);
entry            3385 kernel/trace/trace_events.c 	entry->ip			= ip;
entry            3386 kernel/trace/trace_events.c 	entry->parent_ip		= parent_ip;
entry            3389 kernel/trace/trace_events.c 				    entry, flags, pc);
entry             843 kernel/trace/trace_events_hist.c 	struct synth_trace_event *entry;
entry             849 kernel/trace/trace_events_hist.c 	entry = (struct synth_trace_event *)iter->ent;
entry             869 kernel/trace/trace_events_hist.c 					 (char *)&entry->fields[n_u64],
entry             880 kernel/trace/trace_events_hist.c 						  entry->fields[n_u64],
entry             886 kernel/trace/trace_events_hist.c 						      entry->fields[n_u64],
entry             908 kernel/trace/trace_events_hist.c 	struct synth_trace_event *entry;
entry             929 kernel/trace/trace_events_hist.c 	entry = trace_event_buffer_reserve(&fbuffer, trace_file,
entry             930 kernel/trace/trace_events_hist.c 					   sizeof(*entry) + fields_size);
entry             931 kernel/trace/trace_events_hist.c 	if (!entry)
entry             938 kernel/trace/trace_events_hist.c 			char *str_field = (char *)&entry->fields[n_u64];
entry             948 kernel/trace/trace_events_hist.c 				*(u8 *)&entry->fields[n_u64] = (u8)val;
entry             952 kernel/trace/trace_events_hist.c 				*(u16 *)&entry->fields[n_u64] = (u16)val;
entry             956 kernel/trace/trace_events_hist.c 				*(u32 *)&entry->fields[n_u64] = (u32)val;
entry             960 kernel/trace/trace_events_hist.c 				entry->fields[n_u64] = val;
entry            6535 kernel/trace/trace_events_hist.c 	struct dentry *entry = NULL;
entry            6551 kernel/trace/trace_events_hist.c 	entry = tracefs_create_file("synthetic_events", 0644, d_tracer,
entry            6553 kernel/trace/trace_events_hist.c 	if (!entry) {
entry             105 kernel/trace/trace_functions_graph.c 	struct ftrace_graph_ent_entry *entry;
entry             108 kernel/trace/trace_functions_graph.c 					  sizeof(*entry), flags, pc);
entry             111 kernel/trace/trace_functions_graph.c 	entry	= ring_buffer_event_data(event);
entry             112 kernel/trace/trace_functions_graph.c 	entry->graph_ent			= *trace;
entry             113 kernel/trace/trace_functions_graph.c 	if (!call_filter_check_discard(call, entry, buffer, event))
entry             225 kernel/trace/trace_functions_graph.c 	struct ftrace_graph_ret_entry *entry;
entry             228 kernel/trace/trace_functions_graph.c 					  sizeof(*entry), flags, pc);
entry             231 kernel/trace/trace_functions_graph.c 	entry	= ring_buffer_event_data(event);
entry             232 kernel/trace/trace_functions_graph.c 	entry->ret				= *trace;
entry             233 kernel/trace/trace_functions_graph.c 	if (!call_filter_check_discard(call, entry, buffer, event))
entry             376 kernel/trace/trace_functions_graph.c static void print_graph_lat_fmt(struct trace_seq *s, struct trace_entry *entry)
entry             379 kernel/trace/trace_functions_graph.c 	trace_print_lat_fmt(s, entry);
entry             627 kernel/trace/trace_functions_graph.c 		struct ftrace_graph_ent_entry *entry,
entry             640 kernel/trace/trace_functions_graph.c 	call = &entry->graph_ent;
entry             678 kernel/trace/trace_functions_graph.c 			 struct ftrace_graph_ent_entry *entry,
entry             681 kernel/trace/trace_functions_graph.c 	struct ftrace_graph_ent *call = &entry->graph_ent;
entry            1043 kernel/trace/trace_functions_graph.c 	struct trace_entry *entry = iter->ent;
entry            1069 kernel/trace/trace_functions_graph.c 	switch (entry->type) {
entry            1078 kernel/trace/trace_functions_graph.c 		trace_assign_type(field, entry);
entry            1084 kernel/trace/trace_functions_graph.c 		trace_assign_type(field, entry);
entry            1085 kernel/trace/trace_functions_graph.c 		return print_graph_return(&field->ret, s, entry, iter, flags);
entry            1093 kernel/trace/trace_functions_graph.c 		return print_graph_comment(s, entry, iter, flags);
entry             109 kernel/trace/trace_hwlat.c 	struct hwlat_entry *entry;
entry             116 kernel/trace/trace_hwlat.c 	event = trace_buffer_lock_reserve(buffer, TRACE_HWLAT, sizeof(*entry),
entry             120 kernel/trace/trace_hwlat.c 	entry	= ring_buffer_event_data(event);
entry             121 kernel/trace/trace_hwlat.c 	entry->seqnum			= sample->seqnum;
entry             122 kernel/trace/trace_hwlat.c 	entry->duration			= sample->duration;
entry             123 kernel/trace/trace_hwlat.c 	entry->outer_duration		= sample->outer_duration;
entry             124 kernel/trace/trace_hwlat.c 	entry->timestamp		= sample->timestamp;
entry             125 kernel/trace/trace_hwlat.c 	entry->nmi_total_ts		= sample->nmi_total_ts;
entry             126 kernel/trace/trace_hwlat.c 	entry->nmi_count		= sample->nmi_count;
entry             128 kernel/trace/trace_hwlat.c 	if (!call_filter_check_discard(call, entry, buffer, event))
entry             411 kernel/trace/trace_hwlat.c 	u64 *entry = filp->private_data;
entry             415 kernel/trace/trace_hwlat.c 	if (!entry)
entry             421 kernel/trace/trace_hwlat.c 	val = *entry;
entry            1179 kernel/trace/trace_kprobe.c 	struct kprobe_trace_entry_head *entry;
entry            1195 kernel/trace/trace_kprobe.c 	size = sizeof(*entry) + tk->tp.size + dsize;
entry            1203 kernel/trace/trace_kprobe.c 	entry = ring_buffer_event_data(event);
entry            1204 kernel/trace/trace_kprobe.c 	entry->ip = (unsigned long)tk->rp.kp.addr;
entry            1205 kernel/trace/trace_kprobe.c 	store_trace_args(&entry[1], &tk->tp, regs, sizeof(*entry), dsize);
entry            1208 kernel/trace/trace_kprobe.c 					 entry, irq_flags, pc, regs);
entry            1227 kernel/trace/trace_kprobe.c 	struct kretprobe_trace_entry_head *entry;
entry            1243 kernel/trace/trace_kprobe.c 	size = sizeof(*entry) + tk->tp.size + dsize;
entry            1251 kernel/trace/trace_kprobe.c 	entry = ring_buffer_event_data(event);
entry            1252 kernel/trace/trace_kprobe.c 	entry->func = (unsigned long)tk->rp.kp.addr;
entry            1253 kernel/trace/trace_kprobe.c 	entry->ret_ip = (unsigned long)ri->ret_addr;
entry            1254 kernel/trace/trace_kprobe.c 	store_trace_args(&entry[1], &tk->tp, regs, sizeof(*entry), dsize);
entry            1257 kernel/trace/trace_kprobe.c 					 entry, irq_flags, pc, regs);
entry            1377 kernel/trace/trace_kprobe.c 	struct kprobe_trace_entry_head *entry;
entry            1404 kernel/trace/trace_kprobe.c 	__size = sizeof(*entry) + tk->tp.size + dsize;
entry            1408 kernel/trace/trace_kprobe.c 	entry = perf_trace_buf_alloc(size, NULL, &rctx);
entry            1409 kernel/trace/trace_kprobe.c 	if (!entry)
entry            1412 kernel/trace/trace_kprobe.c 	entry->ip = (unsigned long)tk->rp.kp.addr;
entry            1413 kernel/trace/trace_kprobe.c 	memset(&entry[1], 0, dsize);
entry            1414 kernel/trace/trace_kprobe.c 	store_trace_args(&entry[1], &tk->tp, regs, sizeof(*entry), dsize);
entry            1415 kernel/trace/trace_kprobe.c 	perf_trace_buf_submit(entry, size, rctx, call->event.type, 1, regs,
entry            1427 kernel/trace/trace_kprobe.c 	struct kretprobe_trace_entry_head *entry;
entry            1440 kernel/trace/trace_kprobe.c 	__size = sizeof(*entry) + tk->tp.size + dsize;
entry            1444 kernel/trace/trace_kprobe.c 	entry = perf_trace_buf_alloc(size, NULL, &rctx);
entry            1445 kernel/trace/trace_kprobe.c 	if (!entry)
entry            1448 kernel/trace/trace_kprobe.c 	entry->func = (unsigned long)tk->rp.kp.addr;
entry            1449 kernel/trace/trace_kprobe.c 	entry->ret_ip = (unsigned long)ri->ret_addr;
entry            1450 kernel/trace/trace_kprobe.c 	store_trace_args(&entry[1], &tk->tp, regs, sizeof(*entry), dsize);
entry            1451 kernel/trace/trace_kprobe.c 	perf_trace_buf_submit(entry, size, rctx, call->event.type, 1, regs,
entry            1695 kernel/trace/trace_kprobe.c 	struct dentry *entry;
entry            1709 kernel/trace/trace_kprobe.c 	entry = tracefs_create_file("kprobe_events", 0644, d_tracer,
entry            1713 kernel/trace/trace_kprobe.c 	if (!entry)
entry            1717 kernel/trace/trace_kprobe.c 	entry = tracefs_create_file("kprobe_profile", 0444, d_tracer,
entry            1720 kernel/trace/trace_kprobe.c 	if (!entry)
entry             169 kernel/trace/trace_mmiotrace.c 	struct trace_entry *entry = iter->ent;
entry             177 kernel/trace/trace_mmiotrace.c 	trace_assign_type(field, entry);
entry             214 kernel/trace/trace_mmiotrace.c 	struct trace_entry *entry = iter->ent;
entry             222 kernel/trace/trace_mmiotrace.c 	trace_assign_type(field, entry);
entry             248 kernel/trace/trace_mmiotrace.c 	struct trace_entry *entry = iter->ent;
entry             249 kernel/trace/trace_mmiotrace.c 	struct print_entry *print = (struct print_entry *)entry;
entry             302 kernel/trace/trace_mmiotrace.c 	struct trace_mmiotrace_rw *entry;
entry             306 kernel/trace/trace_mmiotrace.c 					  sizeof(*entry), 0, pc);
entry             311 kernel/trace/trace_mmiotrace.c 	entry	= ring_buffer_event_data(event);
entry             312 kernel/trace/trace_mmiotrace.c 	entry->rw			= *rw;
entry             314 kernel/trace/trace_mmiotrace.c 	if (!call_filter_check_discard(call, entry, buffer, event))
entry             332 kernel/trace/trace_mmiotrace.c 	struct trace_mmiotrace_map *entry;
entry             336 kernel/trace/trace_mmiotrace.c 					  sizeof(*entry), 0, pc);
entry             341 kernel/trace/trace_mmiotrace.c 	entry	= ring_buffer_event_data(event);
entry             342 kernel/trace/trace_mmiotrace.c 	entry->map			= *map;
entry             344 kernel/trace/trace_mmiotrace.c 	if (!call_filter_check_discard(call, entry, buffer, event))
entry              28 kernel/trace/trace_output.c 	struct trace_entry *entry = iter->ent;
entry              31 kernel/trace/trace_output.c 	trace_assign_type(field, entry);
entry              41 kernel/trace/trace_output.c 	struct trace_entry *entry = iter->ent;
entry              44 kernel/trace/trace_output.c 	trace_assign_type(field, entry);
entry              54 kernel/trace/trace_output.c 	struct trace_entry *entry = iter->ent;
entry              57 kernel/trace/trace_output.c 	trace_assign_type(field, entry);
entry             283 kernel/trace/trace_output.c 	struct trace_entry *entry;
entry             286 kernel/trace/trace_output.c 	entry = iter->ent;
entry             288 kernel/trace/trace_output.c 	if (entry->type != event->event.type) {
entry             425 kernel/trace/trace_output.c int trace_print_lat_fmt(struct trace_seq *s, struct trace_entry *entry)
entry             434 kernel/trace/trace_output.c 	nmi = entry->flags & TRACE_FLAG_NMI;
entry             435 kernel/trace/trace_output.c 	hardirq = entry->flags & TRACE_FLAG_HARDIRQ;
entry             436 kernel/trace/trace_output.c 	softirq = entry->flags & TRACE_FLAG_SOFTIRQ;
entry             439 kernel/trace/trace_output.c 		(entry->flags & TRACE_FLAG_IRQS_OFF) ? 'd' :
entry             440 kernel/trace/trace_output.c 		(entry->flags & TRACE_FLAG_IRQS_NOSUPPORT) ? 'X' :
entry             443 kernel/trace/trace_output.c 	switch (entry->flags & (TRACE_FLAG_NEED_RESCHED |
entry             470 kernel/trace/trace_output.c 	if (entry->preempt_count)
entry             471 kernel/trace/trace_output.c 		trace_seq_printf(s, "%x", entry->preempt_count);
entry             479 kernel/trace/trace_output.c lat_print_generic(struct trace_seq *s, struct trace_entry *entry, int cpu)
entry             483 kernel/trace/trace_output.c 	trace_find_cmdline(entry->pid, comm);
entry             486 kernel/trace/trace_output.c 			 comm, entry->pid, cpu);
entry             488 kernel/trace/trace_output.c 	return trace_print_lat_fmt(s, entry);
entry             569 kernel/trace/trace_output.c 	struct trace_entry *entry = iter->ent;
entry             574 kernel/trace/trace_output.c 	trace_find_cmdline(entry->pid, comm);
entry             576 kernel/trace/trace_output.c 	trace_seq_printf(s, "%16s-%-5d ", comm, entry->pid);
entry             579 kernel/trace/trace_output.c 		unsigned int tgid = trace_find_tgid(entry->pid);
entry             590 kernel/trace/trace_output.c 		trace_print_lat_fmt(s, entry);
entry             610 kernel/trace/trace_output.c 	struct trace_entry *entry = iter->ent,
entry             624 kernel/trace/trace_output.c 		trace_find_cmdline(entry->pid, comm);
entry             628 kernel/trace/trace_output.c 			comm, entry->pid, iter->cpu, entry->flags,
entry             629 kernel/trace/trace_output.c 			entry->preempt_count, iter->idx);
entry             631 kernel/trace/trace_output.c 		lat_print_generic(s, entry, iter->cpu);
entry            1140 kernel/trace/trace_output.c 	struct trace_entry *entry = iter->ent;
entry            1144 kernel/trace/trace_output.c 	trace_assign_type(field, entry);
entry            1205 kernel/trace/trace_output.c 	struct trace_entry *entry = iter->ent;
entry            1209 kernel/trace/trace_output.c 	trace_assign_type(field, entry);
entry            1249 kernel/trace/trace_output.c 	struct trace_entry *entry = iter->ent;
entry            1253 kernel/trace/trace_output.c 	trace_assign_type(field, entry);
entry              29 kernel/trace/trace_output.h trace_print_lat_fmt(struct trace_seq *s, struct trace_entry *entry);
entry             383 kernel/trace/trace_sched_wakeup.c 	struct ctx_switch_entry *entry;
entry             386 kernel/trace/trace_sched_wakeup.c 					  sizeof(*entry), flags, pc);
entry             389 kernel/trace/trace_sched_wakeup.c 	entry	= ring_buffer_event_data(event);
entry             390 kernel/trace/trace_sched_wakeup.c 	entry->prev_pid			= prev->pid;
entry             391 kernel/trace/trace_sched_wakeup.c 	entry->prev_prio		= prev->prio;
entry             392 kernel/trace/trace_sched_wakeup.c 	entry->prev_state		= task_state_index(prev);
entry             393 kernel/trace/trace_sched_wakeup.c 	entry->next_pid			= next->pid;
entry             394 kernel/trace/trace_sched_wakeup.c 	entry->next_prio		= next->prio;
entry             395 kernel/trace/trace_sched_wakeup.c 	entry->next_state		= task_state_index(next);
entry             396 kernel/trace/trace_sched_wakeup.c 	entry->next_cpu	= task_cpu(next);
entry             398 kernel/trace/trace_sched_wakeup.c 	if (!call_filter_check_discard(call, entry, buffer, event))
entry             410 kernel/trace/trace_sched_wakeup.c 	struct ctx_switch_entry *entry;
entry             414 kernel/trace/trace_sched_wakeup.c 					  sizeof(*entry), flags, pc);
entry             417 kernel/trace/trace_sched_wakeup.c 	entry	= ring_buffer_event_data(event);
entry             418 kernel/trace/trace_sched_wakeup.c 	entry->prev_pid			= curr->pid;
entry             419 kernel/trace/trace_sched_wakeup.c 	entry->prev_prio		= curr->prio;
entry             420 kernel/trace/trace_sched_wakeup.c 	entry->prev_state		= task_state_index(curr);
entry             421 kernel/trace/trace_sched_wakeup.c 	entry->next_pid			= wakee->pid;
entry             422 kernel/trace/trace_sched_wakeup.c 	entry->next_prio		= wakee->prio;
entry             423 kernel/trace/trace_sched_wakeup.c 	entry->next_state		= task_state_index(wakee);
entry             424 kernel/trace/trace_sched_wakeup.c 	entry->next_cpu			= task_cpu(wakee);
entry             426 kernel/trace/trace_sched_wakeup.c 	if (!call_filter_check_discard(call, entry, buffer, event))
entry              10 kernel/trace/trace_selftest.c static inline int trace_valid_entry(struct trace_entry *entry)
entry              12 kernel/trace/trace_selftest.c 	switch (entry->type) {
entry              29 kernel/trace/trace_selftest.c 	struct trace_entry *entry;
entry              33 kernel/trace/trace_selftest.c 		entry = ring_buffer_event_data(event);
entry              44 kernel/trace/trace_selftest.c 		if (!trace_valid_entry(entry)) {
entry              46 kernel/trace/trace_selftest.c 				entry->type);
entry              25 kernel/trace/trace_syscalls.c 	struct syscall_metadata *entry = call->data;
entry              27 kernel/trace/trace_syscalls.c 	return &entry->enter_fields;
entry             112 kernel/trace/trace_syscalls.c 	struct syscall_metadata *entry;
entry             114 kernel/trace/trace_syscalls.c 	entry = syscall_nr_to_meta(syscall);
entry             115 kernel/trace/trace_syscalls.c 	if (!entry)
entry             118 kernel/trace/trace_syscalls.c 	return entry->name;
entry             129 kernel/trace/trace_syscalls.c 	struct syscall_metadata *entry;
entry             134 kernel/trace/trace_syscalls.c 	entry = syscall_nr_to_meta(syscall);
entry             136 kernel/trace/trace_syscalls.c 	if (!entry)
entry             139 kernel/trace/trace_syscalls.c 	if (entry->enter_event->event.type != ent->type) {
entry             144 kernel/trace/trace_syscalls.c 	trace_seq_printf(s, "%s(", entry->name);
entry             146 kernel/trace/trace_syscalls.c 	for (i = 0; i < entry->nb_args; i++) {
entry             153 kernel/trace/trace_syscalls.c 			trace_seq_printf(s, "%s ", entry->types[i]);
entry             156 kernel/trace/trace_syscalls.c 		trace_seq_printf(s, "%s: %lx%s", entry->args[i],
entry             158 kernel/trace/trace_syscalls.c 				 i == entry->nb_args - 1 ? "" : ", ");
entry             176 kernel/trace/trace_syscalls.c 	struct syscall_metadata *entry;
entry             180 kernel/trace/trace_syscalls.c 	entry = syscall_nr_to_meta(syscall);
entry             182 kernel/trace/trace_syscalls.c 	if (!entry) {
entry             187 kernel/trace/trace_syscalls.c 	if (entry->exit_event->event.type != ent->type) {
entry             192 kernel/trace/trace_syscalls.c 	trace_seq_printf(s, "%s -> 0x%lx\n", entry->name,
entry             208 kernel/trace/trace_syscalls.c __set_enter_print_fmt(struct syscall_metadata *entry, char *buf, int len)
entry             217 kernel/trace/trace_syscalls.c 	for (i = 0; i < entry->nb_args; i++) {
entry             219 kernel/trace/trace_syscalls.c 				entry->args[i], sizeof(unsigned long),
entry             220 kernel/trace/trace_syscalls.c 				i == entry->nb_args - 1 ? "" : ", ");
entry             224 kernel/trace/trace_syscalls.c 	for (i = 0; i < entry->nb_args; i++) {
entry             226 kernel/trace/trace_syscalls.c 				", ((unsigned long)(REC->%s))", entry->args[i]);
entry             239 kernel/trace/trace_syscalls.c 	struct syscall_metadata *entry = call->data;
entry             241 kernel/trace/trace_syscalls.c 	if (entry->enter_event != call) {
entry             247 kernel/trace/trace_syscalls.c 	len = __set_enter_print_fmt(entry, NULL, 0);
entry             254 kernel/trace/trace_syscalls.c 	__set_enter_print_fmt(entry, print_fmt, len + 1);
entry             262 kernel/trace/trace_syscalls.c 	struct syscall_metadata *entry = call->data;
entry             264 kernel/trace/trace_syscalls.c 	if (entry->enter_event == call)
entry             312 kernel/trace/trace_syscalls.c 	struct syscall_trace_enter *entry;
entry             338 kernel/trace/trace_syscalls.c 	size = sizeof(*entry) + sizeof(unsigned long) * sys_data->nb_args;
entry             349 kernel/trace/trace_syscalls.c 	entry = ring_buffer_event_data(event);
entry             350 kernel/trace/trace_syscalls.c 	entry->nr = syscall_nr;
entry             352 kernel/trace/trace_syscalls.c 	memcpy(entry->args, args, sizeof(unsigned long) * sys_data->nb_args);
entry             354 kernel/trace/trace_syscalls.c 	event_trigger_unlock_commit(trace_file, buffer, event, entry,
entry             362 kernel/trace/trace_syscalls.c 	struct syscall_trace_exit *entry;
entry             391 kernel/trace/trace_syscalls.c 			sys_data->exit_event->event.type, sizeof(*entry),
entry             396 kernel/trace/trace_syscalls.c 	entry = ring_buffer_event_data(event);
entry             397 kernel/trace/trace_syscalls.c 	entry->nr = syscall_nr;
entry             398 kernel/trace/trace_syscalls.c 	entry->ret = syscall_get_return_value(current, regs);
entry             400 kernel/trace/trace_syscalls.c 	event_trigger_unlock_commit(trace_file, buffer, event, entry,
entry              34 kernel/trace/trace_uprobe.c #define DATAOF_TRACE_ENTRY(entry, is_return)		\
entry              35 kernel/trace/trace_uprobe.c 	((void*)(entry) + SIZEOF_TRACE_ENTRY(is_return))
entry             933 kernel/trace/trace_uprobe.c 	struct uprobe_trace_entry_head *entry;
entry             955 kernel/trace/trace_uprobe.c 	entry = ring_buffer_event_data(event);
entry             957 kernel/trace/trace_uprobe.c 		entry->vaddr[0] = func;
entry             958 kernel/trace/trace_uprobe.c 		entry->vaddr[1] = instruction_pointer(regs);
entry             959 kernel/trace/trace_uprobe.c 		data = DATAOF_TRACE_ENTRY(entry, true);
entry             961 kernel/trace/trace_uprobe.c 		entry->vaddr[0] = instruction_pointer(regs);
entry             962 kernel/trace/trace_uprobe.c 		data = DATAOF_TRACE_ENTRY(entry, false);
entry             967 kernel/trace/trace_uprobe.c 	event_trigger_unlock_commit(trace_file, buffer, event, entry, 0, 0);
entry            1003 kernel/trace/trace_uprobe.c 	struct uprobe_trace_entry_head *entry;
entry            1008 kernel/trace/trace_uprobe.c 	entry = (struct uprobe_trace_entry_head *)iter->ent;
entry            1017 kernel/trace/trace_uprobe.c 				 entry->vaddr[1], entry->vaddr[0]);
entry            1018 kernel/trace/trace_uprobe.c 		data = DATAOF_TRACE_ENTRY(entry, true);
entry            1022 kernel/trace/trace_uprobe.c 				 entry->vaddr[0]);
entry            1023 kernel/trace/trace_uprobe.c 		data = DATAOF_TRACE_ENTRY(entry, false);
entry            1026 kernel/trace/trace_uprobe.c 	if (print_probe_args(s, tu->tp.args, tu->tp.nr_args, data, entry) < 0)
entry            1330 kernel/trace/trace_uprobe.c 	struct uprobe_trace_entry_head *entry;
entry            1351 kernel/trace/trace_uprobe.c 	entry = perf_trace_buf_alloc(size, NULL, &rctx);
entry            1352 kernel/trace/trace_uprobe.c 	if (!entry)
entry            1356 kernel/trace/trace_uprobe.c 		entry->vaddr[0] = func;
entry            1357 kernel/trace/trace_uprobe.c 		entry->vaddr[1] = instruction_pointer(regs);
entry            1358 kernel/trace/trace_uprobe.c 		data = DATAOF_TRACE_ENTRY(entry, true);
entry            1360 kernel/trace/trace_uprobe.c 		entry->vaddr[0] = instruction_pointer(regs);
entry            1361 kernel/trace/trace_uprobe.c 		data = DATAOF_TRACE_ENTRY(entry, false);
entry            1372 kernel/trace/trace_uprobe.c 	perf_trace_buf_submit(entry, size, rctx, call->event.type, 1, regs,
entry             517 kernel/trace/tracing_map.c 	struct tracing_map_entry *entry;
entry             527 kernel/trace/tracing_map.c 		entry = TRACING_MAP_ENTRY(map->map, idx);
entry             528 kernel/trace/tracing_map.c 		test_key = entry->key;
entry             531 kernel/trace/tracing_map.c 			val = READ_ONCE(entry->val);
entry             563 kernel/trace/tracing_map.c 			if (!cmpxchg(&entry->key, 0, key_hash)) {
entry             569 kernel/trace/tracing_map.c 					entry->key = 0;
entry             574 kernel/trace/tracing_map.c 				entry->val = elt;
entry             577 kernel/trace/tracing_map.c 				return entry->val;
entry             905 kernel/trace/tracing_map.c static void destroy_sort_entry(struct tracing_map_sort_entry *entry)
entry             907 kernel/trace/tracing_map.c 	if (!entry)
entry             910 kernel/trace/tracing_map.c 	if (entry->elt_copied)
entry             911 kernel/trace/tracing_map.c 		tracing_map_elt_free(entry->elt);
entry             913 kernel/trace/tracing_map.c 	kfree(entry);
entry            1074 kernel/trace/tracing_map.c 		struct tracing_map_entry *entry;
entry            1076 kernel/trace/tracing_map.c 		entry = TRACING_MAP_ENTRY(map->map, i);
entry            1078 kernel/trace/tracing_map.c 		if (!entry->key || !entry->val)
entry            1081 kernel/trace/tracing_map.c 		entries[n_entries] = create_sort_entry(entry->val->key,
entry            1082 kernel/trace/tracing_map.c 						       entry->val);
entry             825 kernel/workqueue.c 	return list_first_entry(&pool->idle_list, struct worker, entry);
entry            1068 kernel/workqueue.c 	list_for_each_entry_safe_from(work, n, NULL, entry) {
entry            1069 kernel/workqueue.c 		list_move_tail(&work->entry, head);
entry            1156 kernel/workqueue.c 						    struct work_struct, entry);
entry            1292 kernel/workqueue.c 		list_del_init(&work->entry);
entry            1332 kernel/workqueue.c 	list_add_tail(&work->entry, head);
entry            1477 kernel/workqueue.c 	if (WARN_ON(!list_empty(&work->entry)))
entry            1632 kernel/workqueue.c 	WARN_ON_ONCE(!list_empty(&work->entry));
entry            1773 kernel/workqueue.c 	    WARN_ON_ONCE(!list_empty(&worker->entry) &&
entry            1783 kernel/workqueue.c 	list_add(&worker->entry, &pool->idle_list);
entry            1816 kernel/workqueue.c 	list_del_init(&worker->entry);
entry            1825 kernel/workqueue.c 		INIT_LIST_HEAD(&worker->entry);
entry            1984 kernel/workqueue.c 	list_del_init(&worker->entry);
entry            2000 kernel/workqueue.c 		worker = list_entry(pool->idle_list.prev, struct worker, entry);
entry            2052 kernel/workqueue.c 		list_for_each_entry(work, &pool->worklist, entry)
entry            2216 kernel/workqueue.c 	list_del_init(&work->entry);
entry            2333 kernel/workqueue.c 						struct work_struct, entry);
entry            2373 kernel/workqueue.c 		WARN_ON_ONCE(!list_empty(&worker->entry));
entry            2412 kernel/workqueue.c 					 struct work_struct, entry);
entry            2515 kernel/workqueue.c 		list_for_each_entry_safe(work, n, &pool->worklist, entry) {
entry            2682 kernel/workqueue.c 		head = target->entry.next;
entry            4687 kernel/workqueue.c 			list_for_each_entry(work, &worker->scheduled, entry)
entry            4694 kernel/workqueue.c 	list_for_each_entry(work, &pool->worklist, entry) {
entry            4704 kernel/workqueue.c 		list_for_each_entry(work, &pool->worklist, entry) {
entry            4718 kernel/workqueue.c 		list_for_each_entry(work, &pwq->delayed_works, entry) {
entry            4788 kernel/workqueue.c 		list_for_each_entry(worker, &pool->idle_list, entry) {
entry              27 kernel/workqueue_internal.h 		struct list_head	entry;	/* L: while idle */
entry              63 lib/error-inject.c 	unsigned long entry, offset = 0, size = 0;
entry              67 lib/error-inject.c 		entry = arch_deref_entry_point((void *)iter->addr);
entry              69 lib/error-inject.c 		if (!kernel_text_address(entry) ||
entry              70 lib/error-inject.c 		    !kallsyms_lookup_size_offset(entry, &size, &offset)) {
entry              72 lib/error-inject.c 				(void *)entry);
entry              79 lib/error-inject.c 		ent->start_addr = entry;
entry              80 lib/error-inject.c 		ent->end_addr = entry + size;
entry             231 lib/idr.c      	void *entry = NULL;
entry             237 lib/idr.c      		entry = rcu_dereference_raw(*slot);
entry             238 lib/idr.c      		if (!entry)
entry             240 lib/idr.c      		if (!xa_is_internal(entry))
entry             242 lib/idr.c      		if (slot != &idr->idr_rt.xa_head && !xa_is_retry(entry))
entry             250 lib/idr.c      	return entry;
entry             267 lib/idr.c      	void *entry = idr_get_next_ul(idr, &id);
entry             272 lib/idr.c      	return entry;
entry             294 lib/idr.c      	void *entry;
entry             298 lib/idr.c      	entry = __radix_tree_lookup(&idr->idr_rt, id, &node, &slot);
entry             304 lib/idr.c      	return entry;
entry             556 lib/idr.c      static void ida_dump_entry(void *entry, unsigned long index)
entry             560 lib/idr.c      	if (!entry)
entry             563 lib/idr.c      	if (xa_is_node(entry)) {
entry             564 lib/idr.c      		struct xa_node *node = xa_to_node(entry);
entry             573 lib/idr.c      	} else if (xa_is_value(entry)) {
entry             575 lib/idr.c      		pr_cont("value: data %lx [%px]\n", xa_to_value(entry), entry);
entry             577 lib/idr.c      		struct ida_bitmap *bitmap = entry;
entry             196 lib/kobject.c  	list_add_tail(&kobj->entry, &kobj->kset->list);
entry             207 lib/kobject.c  	list_del_init(&kobj->entry);
entry             217 lib/kobject.c  	INIT_LIST_HEAD(&kobj->entry);
entry             901 lib/kobject.c  	list_for_each_entry(k, &kset->list, entry) {
entry              38 lib/list_debug.c bool __list_del_entry_valid(struct list_head *entry)
entry              42 lib/list_debug.c 	prev = entry->prev;
entry              43 lib/list_debug.c 	next = entry->next;
entry              47 lib/list_debug.c 			entry, LIST_POISON1) ||
entry              50 lib/list_debug.c 			entry, LIST_POISON2) ||
entry              51 lib/list_debug.c 	    CHECK_DATA_CORRUPTION(prev->next != entry,
entry              53 lib/list_debug.c 			entry, prev->next) ||
entry              54 lib/list_debug.c 	    CHECK_DATA_CORRUPTION(next->prev != entry,
entry              56 lib/list_debug.c 			entry, next->prev))
entry              55 lib/llist.c    	struct llist_node *entry, *old_entry, *next;
entry              57 lib/llist.c    	entry = smp_load_acquire(&head->first);
entry              59 lib/llist.c    		if (entry == NULL)
entry              61 lib/llist.c    		old_entry = entry;
entry              62 lib/llist.c    		next = READ_ONCE(entry->next);
entry              63 lib/llist.c    		entry = cmpxchg(&head->first, old_entry, next);
entry              64 lib/llist.c    		if (entry == old_entry)
entry              68 lib/llist.c    	return entry;
entry             238 lib/logic_pio.c 		struct logic_pio_hwaddr *entry = find_io_range(addr);	\
entry             240 lib/logic_pio.c 		if (entry && entry->ops)				\
entry             241 lib/logic_pio.c 			ret = entry->ops->in(entry->hostdata,		\
entry             254 lib/logic_pio.c 		struct logic_pio_hwaddr *entry = find_io_range(addr);	\
entry             256 lib/logic_pio.c 		if (entry && entry->ops)				\
entry             257 lib/logic_pio.c 			entry->ops->out(entry->hostdata,		\
entry             270 lib/logic_pio.c 		struct logic_pio_hwaddr *entry = find_io_range(addr);	\
entry             272 lib/logic_pio.c 		if (entry && entry->ops)				\
entry             273 lib/logic_pio.c 			entry->ops->ins(entry->hostdata,		\
entry             287 lib/logic_pio.c 		struct logic_pio_hwaddr *entry = find_io_range(addr);	\
entry             289 lib/logic_pio.c 		if (entry && entry->ops)				\
entry             290 lib/logic_pio.c 			entry->ops->outs(entry->hostdata,		\
entry              75 lib/nlattr.c   	const struct nlattr *entry;
entry              78 lib/nlattr.c   	nla_for_each_attr(entry, head, len, rem) {
entry              81 lib/nlattr.c   		if (nla_len(entry) == 0)
entry              84 lib/nlattr.c   		if (nla_len(entry) < NLA_HDRLEN) {
entry              85 lib/nlattr.c   			NL_SET_ERR_MSG_ATTR(extack, entry,
entry              90 lib/nlattr.c   		ret = __nla_validate(nla_data(entry), nla_len(entry),
entry              98 lib/radix-tree.c 	void __rcu **entry = rcu_dereference_raw(parent->slots[offset]);
entry             100 lib/radix-tree.c 	*nodep = (void *)entry;
entry             420 lib/radix-tree.c 	void *entry;
entry             429 lib/radix-tree.c 	entry = rcu_dereference_raw(root->xa_head);
entry             430 lib/radix-tree.c 	if (!entry && (!is_idr(root) || root_tag_get(root, IDR_FREE)))
entry             454 lib/radix-tree.c 		if (radix_tree_is_internal_node(entry)) {
entry             455 lib/radix-tree.c 			entry_to_node(entry)->parent = node;
entry             456 lib/radix-tree.c 		} else if (xa_is_value(entry)) {
entry             464 lib/radix-tree.c 		node->slots[0] = (void __rcu *)entry;
entry             465 lib/radix-tree.c 		entry = node_to_entry(node);
entry             466 lib/radix-tree.c 		rcu_assign_pointer(root->xa_head, entry);
entry             671 lib/radix-tree.c 		void *entry = rcu_dereference_raw(child->slots[offset]);
entry             672 lib/radix-tree.c 		if (xa_is_node(entry) && child->shift) {
entry             673 lib/radix-tree.c 			child = entry_to_node(entry);
entry            1427 lib/radix-tree.c 	void *entry;
entry            1429 lib/radix-tree.c 	entry = __radix_tree_lookup(root, index, &node, &slot);
entry            1432 lib/radix-tree.c 	if (!entry && (!is_idr(root) || node_tag_get(root, node, IDR_FREE,
entry            1436 lib/radix-tree.c 	if (item && entry != item)
entry            1441 lib/radix-tree.c 	return entry;
entry             222 lib/rhashtable.c 	struct rhash_head *head, *next, *entry;
entry             231 lib/rhashtable.c 	rht_for_each_from(entry, rht_ptr(bkt, old_tbl, old_hash),
entry             234 lib/rhashtable.c 		next = rht_dereference_bucket(entry->next, old_tbl, old_hash);
entry             239 lib/rhashtable.c 		pprev = &entry->next;
entry             245 lib/rhashtable.c 	new_hash = head_hashfn(ht, new_tbl, entry);
entry             251 lib/rhashtable.c 	RCU_INIT_POINTER(entry->next, head);
entry             253 lib/rhashtable.c 	rht_assign_unlock(new_tbl, &new_tbl->buckets[new_hash], entry);
entry             677 lib/sbitmap.c  	list_del_init(&sbq_wait->wait.entry);
entry              72 lib/test_xarray.c 		unsigned order, void *entry, gfp_t gfp)
entry              79 lib/test_xarray.c 		curr = xas_store(&xas, entry);
entry             105 lib/test_xarray.c 	void *entry;
entry             138 lib/test_xarray.c 	xas_for_each(&xas, entry, ULONG_MAX) {
entry             153 lib/test_xarray.c 			void *entry = xa_load(xa, j);
entry             155 lib/test_xarray.c 				XA_BUG_ON(xa, xa_to_value(entry) != j);
entry             157 lib/test_xarray.c 				XA_BUG_ON(xa, entry);
entry             164 lib/test_xarray.c 			void *entry = xa_load(xa, j);
entry             166 lib/test_xarray.c 				XA_BUG_ON(xa, xa_to_value(entry) != j);
entry             168 lib/test_xarray.c 				XA_BUG_ON(xa, entry);
entry             222 lib/test_xarray.c 			void *entry;
entry             230 lib/test_xarray.c 			xas_for_each(&xas, entry, ULONG_MAX)
entry             239 lib/test_xarray.c 			xas_for_each_marked(&xas, entry, ULONG_MAX, XA_MARK_0)
entry             259 lib/test_xarray.c 	void *entry;
entry             276 lib/test_xarray.c 	xas_for_each_marked(&xas, entry, ULONG_MAX, XA_MARK_0)
entry             282 lib/test_xarray.c 	xas_for_each(&xas, entry, ULONG_MAX) {
entry             403 lib/test_xarray.c 	void *entry;
entry             444 lib/test_xarray.c 	xa_for_each(xa, index, entry) {
entry             470 lib/test_xarray.c 	void *entry;
entry             495 lib/test_xarray.c 		xas_for_each(&xas, entry, ULONG_MAX) {
entry             496 lib/test_xarray.c 			XA_BUG_ON(xa, entry != xa_mk_index(j));
entry             549 lib/test_xarray.c 	void *entry;
entry             555 lib/test_xarray.c 	xas_for_each(&xas, entry, ULONG_MAX) {
entry             556 lib/test_xarray.c 		XA_BUG_ON(xa, entry != xa_mk_index(index));
entry             561 lib/test_xarray.c 	xas_for_each(&xas, entry, ULONG_MAX) {
entry             562 lib/test_xarray.c 		XA_BUG_ON(xa, entry != xa_mk_index(index));
entry             626 lib/test_xarray.c 				void *entry = xa_load(xa, (1UL << k) - 1);
entry             628 lib/test_xarray.c 					XA_BUG_ON(xa, entry != NULL);
entry             630 lib/test_xarray.c 					XA_BUG_ON(xa, entry != xa_mk_index(j));
entry             731 lib/test_xarray.c 	void *entry;
entry             762 lib/test_xarray.c 	xa_for_each(xa, index, entry) {
entry             784 lib/test_xarray.c 	void *entry;
entry             800 lib/test_xarray.c 			entry = xa_mk_index(i);
entry             802 lib/test_xarray.c 			entry = xa_mk_index(i - 0x3fff);
entry             803 lib/test_xarray.c 		XA_BUG_ON(xa, xa_alloc_cyclic(xa, &id, entry, limit,
entry             805 lib/test_xarray.c 		XA_BUG_ON(xa, xa_mk_index(id) != entry);
entry             823 lib/test_xarray.c 	xa_for_each(xa, index, entry)
entry             846 lib/test_xarray.c 	void *entry;
entry             851 lib/test_xarray.c 	xas_for_each_conflict(&xas, entry) {
entry             852 lib/test_xarray.c 		XA_BUG_ON(xa, !xa_is_value(entry));
entry             853 lib/test_xarray.c 		XA_BUG_ON(xa, entry < xa_mk_index(start));
entry             854 lib/test_xarray.c 		XA_BUG_ON(xa, entry > xa_mk_index(start + (1UL << order) - 1));
entry             945 lib/test_xarray.c 	void *entry;
entry             955 lib/test_xarray.c 			xas_for_each(&xas, entry, ULONG_MAX) {
entry             998 lib/test_xarray.c 				void *entry = xa_find(xa, &index, ULONG_MAX,
entry            1005 lib/test_xarray.c 					XA_BUG_ON(xa, entry != NULL);
entry            1008 lib/test_xarray.c 				entry = xa_find(xa, &index, ULONG_MAX,
entry            1015 lib/test_xarray.c 					XA_BUG_ON(xa, entry != NULL);
entry            1029 lib/test_xarray.c 	void *entry;
entry            1032 lib/test_xarray.c 	xa_for_each(xa, index, entry) {
entry            1039 lib/test_xarray.c 		xa_for_each(xa, index, entry) {
entry            1040 lib/test_xarray.c 			XA_BUG_ON(xa, xa_mk_index(index) != entry);
entry            1052 lib/test_xarray.c 	void *entry;
entry            1059 lib/test_xarray.c 				xas_for_each_marked(&xas, entry, k, XA_MARK_0)
entry            1076 lib/test_xarray.c 	void *entry;
entry            1080 lib/test_xarray.c 	entry = xa_find_after(xa, &index, ULONG_MAX, XA_PRESENT);
entry            1081 lib/test_xarray.c 	XA_BUG_ON(xa, entry != xa_mk_index(ULONG_MAX));
entry            1083 lib/test_xarray.c 	entry = xa_find_after(xa, &index, ULONG_MAX, XA_PRESENT);
entry            1084 lib/test_xarray.c 	XA_BUG_ON(xa, entry);
entry            1109 lib/test_xarray.c 	void *entry;
entry            1112 lib/test_xarray.c 	xas_for_each(&xas, entry, ULONG_MAX) {
entry            1113 lib/test_xarray.c 		if (xas_retry(&xas, entry))
entry            1115 lib/test_xarray.c 		if (entry == item)
entry            1124 lib/test_xarray.c 	return entry ? xas.xa_index : -1;
entry            1162 lib/test_xarray.c 	void *entry;
entry            1174 lib/test_xarray.c 	xas_for_each(&xas, entry, ULONG_MAX) {
entry            1175 lib/test_xarray.c 		XA_BUG_ON(xa, entry != xa_mk_index(1UL << count));
entry            1184 lib/test_xarray.c 	xas_for_each(&xas, entry, ULONG_MAX) {
entry            1185 lib/test_xarray.c 		XA_BUG_ON(xa, entry != xa_mk_index(1UL << count));
entry            1248 lib/test_xarray.c 		void *entry = xas_next(&xas);
entry            1253 lib/test_xarray.c 			XA_BUG_ON(xa, entry != xa_mk_index(i));
entry            1255 lib/test_xarray.c 			XA_BUG_ON(xa, entry != NULL);
entry            1261 lib/test_xarray.c 		void *entry = xas_prev(&xas);
entry            1267 lib/test_xarray.c 			XA_BUG_ON(xa, entry != xa_mk_index(i));
entry            1269 lib/test_xarray.c 			XA_BUG_ON(xa, entry != NULL);
entry            1296 lib/test_xarray.c 		void *entry = xas_prev(&xas);
entry            1298 lib/test_xarray.c 		XA_BUG_ON(xa, entry != xa_mk_index(i));
entry            1306 lib/test_xarray.c 		void *entry = xas_next(&xas);
entry            1307 lib/test_xarray.c 		XA_BUG_ON(xa, entry != xa_mk_index(i));
entry            1320 lib/test_xarray.c 		void *entry = xas_prev(&xas);
entry            1323 lib/test_xarray.c 			XA_BUG_ON(xa, entry != xa_mk_index(i));
entry            1325 lib/test_xarray.c 			XA_BUG_ON(xa, entry != NULL);
entry            1333 lib/test_xarray.c 		void *entry = xas_next(&xas);
entry            1335 lib/test_xarray.c 			XA_BUG_ON(xa, entry != xa_mk_index(i));
entry            1337 lib/test_xarray.c 			XA_BUG_ON(xa, entry != NULL);
entry            1511 lib/test_xarray.c 	void *entry;
entry            1518 lib/test_xarray.c 	xa_for_each(xa, index, entry)
entry            1519 lib/test_xarray.c 		XA_BUG_ON(xa, xa_is_err(entry));
entry             181 lib/xarray.c   	void *entry;
entry             188 lib/xarray.c   	entry = xa_head(xas->xa);
entry             189 lib/xarray.c   	if (!xa_is_node(entry)) {
entry             193 lib/xarray.c   		if ((xas->xa_index >> xa_to_node(entry)->shift) > XA_CHUNK_MASK)
entry             198 lib/xarray.c   	return entry;
entry             204 lib/xarray.c   	void *entry = xa_entry(xas->xa, node, offset);
entry             207 lib/xarray.c   	if (xa_is_sibling(entry)) {
entry             208 lib/xarray.c   		offset = xa_to_sibling(entry);
entry             209 lib/xarray.c   		entry = xa_entry(xas->xa, node, offset);
entry             213 lib/xarray.c   	return entry;
entry             233 lib/xarray.c   	void *entry = xas_start(xas);
entry             235 lib/xarray.c   	while (xa_is_node(entry)) {
entry             236 lib/xarray.c   		struct xa_node *node = xa_to_node(entry);
entry             240 lib/xarray.c   		entry = xas_descend(xas, node);
entry             244 lib/xarray.c   	return entry;
entry             426 lib/xarray.c   static unsigned long max_index(void *entry)
entry             428 lib/xarray.c   	if (!xa_is_node(entry))
entry             430 lib/xarray.c   	return (XA_CHUNK_SIZE << xa_to_node(entry)->shift) - 1;
entry             439 lib/xarray.c   		void *entry;
entry             444 lib/xarray.c   		entry = xa_entry_locked(xa, node, 0);
entry             445 lib/xarray.c   		if (!entry)
entry             447 lib/xarray.c   		if (!xa_is_node(entry) && node->shift)
entry             449 lib/xarray.c   		if (xa_is_zero(entry) && xa_zero_busy(xa))
entry             450 lib/xarray.c   			entry = NULL;
entry             453 lib/xarray.c   		RCU_INIT_POINTER(xa->xa_head, entry);
entry             459 lib/xarray.c   		if (!xa_is_node(entry))
entry             463 lib/xarray.c   		if (!xa_is_node(entry))
entry             465 lib/xarray.c   		node = xa_to_node(entry);
entry             525 lib/xarray.c   		void *entry = xa_entry_locked(xas->xa, node, offset);
entry             527 lib/xarray.c   		if (node->shift && xa_is_node(entry)) {
entry             528 lib/xarray.c   			node = xa_to_node(entry);
entry             532 lib/xarray.c   		if (entry)
entry             638 lib/xarray.c   	void *entry;
entry             645 lib/xarray.c   		entry = xa_head_locked(xa);
entry             647 lib/xarray.c   		if (!entry && xa_zero_busy(xa))
entry             648 lib/xarray.c   			entry = XA_ZERO_ENTRY;
entry             649 lib/xarray.c   		shift = xas_expand(xas, entry);
entry             654 lib/xarray.c   		entry = xa_head_locked(xa);
entry             662 lib/xarray.c   		entry = xa_entry_locked(xa, node, offset);
entry             666 lib/xarray.c   		entry = xa_head_locked(xa);
entry             672 lib/xarray.c   		if (!entry) {
entry             679 lib/xarray.c   		} else if (xa_is_node(entry)) {
entry             680 lib/xarray.c   			node = xa_to_node(entry);
entry             684 lib/xarray.c   		entry = xas_descend(xas, node);
entry             688 lib/xarray.c   	return entry;
entry             769 lib/xarray.c   void *xas_store(struct xa_state *xas, void *entry)
entry             777 lib/xarray.c   	bool value = xa_is_value(entry);
entry             779 lib/xarray.c   	if (entry) {
entry             780 lib/xarray.c   		bool allow_root = !xa_is_node(entry) && !xa_is_zero(entry);
entry             791 lib/xarray.c   	if ((first == entry) && !xas->xa_sibs)
entry             802 lib/xarray.c   	if (!entry)
entry             813 lib/xarray.c   		rcu_assign_pointer(*slot, entry);
entry             818 lib/xarray.c   		count += !next - !entry;
entry             820 lib/xarray.c   		if (entry) {
entry             823 lib/xarray.c   			if (!xa_is_sibling(entry))
entry             824 lib/xarray.c   				entry = xa_mk_sibling(xas->xa_offset);
entry             831 lib/xarray.c   			if (!entry && (offset > max))
entry             996 lib/xarray.c   	void *entry;
entry            1016 lib/xarray.c   		entry = xa_entry(xas->xa, xas->xa_node, xas->xa_offset);
entry            1017 lib/xarray.c   		if (!xa_is_node(entry))
entry            1018 lib/xarray.c   			return entry;
entry            1020 lib/xarray.c   		xas->xa_node = xa_to_node(entry);
entry            1035 lib/xarray.c   	void *entry;
entry            1055 lib/xarray.c   		entry = xa_entry(xas->xa, xas->xa_node, xas->xa_offset);
entry            1056 lib/xarray.c   		if (!xa_is_node(entry))
entry            1057 lib/xarray.c   			return entry;
entry            1059 lib/xarray.c   		xas->xa_node = xa_to_node(entry);
entry            1083 lib/xarray.c   	void *entry;
entry            1094 lib/xarray.c   		entry = xas_load(xas);
entry            1095 lib/xarray.c   		if (entry || xas_not_node(xas->xa_node))
entry            1096 lib/xarray.c   			return entry;
entry            1111 lib/xarray.c   		entry = xa_entry(xas->xa, xas->xa_node, xas->xa_offset);
entry            1112 lib/xarray.c   		if (xa_is_node(entry)) {
entry            1113 lib/xarray.c   			xas->xa_node = xa_to_node(entry);
entry            1117 lib/xarray.c   		if (entry && !xa_is_sibling(entry))
entry            1118 lib/xarray.c   			return entry;
entry            1154 lib/xarray.c   	void *entry;
entry            1166 lib/xarray.c   		entry = xa_head(xas->xa);
entry            1168 lib/xarray.c   		if (xas->xa_index > max_index(entry))
entry            1170 lib/xarray.c   		if (!xa_is_node(entry)) {
entry            1172 lib/xarray.c   				return entry;
entry            1176 lib/xarray.c   		xas->xa_node = xa_to_node(entry);
entry            1191 lib/xarray.c   			entry = xa_entry(xas->xa, xas->xa_node, xas->xa_offset);
entry            1192 lib/xarray.c   			if (xa_is_sibling(entry)) {
entry            1193 lib/xarray.c   				xas->xa_offset = xa_to_sibling(entry);
entry            1210 lib/xarray.c   		entry = xa_entry(xas->xa, xas->xa_node, xas->xa_offset);
entry            1211 lib/xarray.c   		if (!entry && !(xa_track_free(xas->xa) && mark == XA_FREE_MARK))
entry            1213 lib/xarray.c   		if (!xa_is_node(entry))
entry            1214 lib/xarray.c   			return entry;
entry            1215 lib/xarray.c   		xas->xa_node = xa_to_node(entry);
entry            1301 lib/xarray.c   	void *entry;
entry            1305 lib/xarray.c   		entry = xas_load(&xas);
entry            1306 lib/xarray.c   		if (xa_is_zero(entry))
entry            1307 lib/xarray.c   			entry = NULL;
entry            1308 lib/xarray.c   	} while (xas_retry(&xas, entry));
entry            1311 lib/xarray.c   	return entry;
entry            1357 lib/xarray.c   	void *entry;
entry            1360 lib/xarray.c   	entry = __xa_erase(xa, index);
entry            1363 lib/xarray.c   	return entry;
entry            1382 lib/xarray.c   void *__xa_store(struct xarray *xa, unsigned long index, void *entry, gfp_t gfp)
entry            1387 lib/xarray.c   	if (WARN_ON_ONCE(xa_is_advanced(entry)))
entry            1389 lib/xarray.c   	if (xa_track_free(xa) && !entry)
entry            1390 lib/xarray.c   		entry = XA_ZERO_ENTRY;
entry            1393 lib/xarray.c   		curr = xas_store(&xas, entry);
entry            1419 lib/xarray.c   void *xa_store(struct xarray *xa, unsigned long index, void *entry, gfp_t gfp)
entry            1424 lib/xarray.c   	curr = __xa_store(xa, index, entry, gfp);
entry            1448 lib/xarray.c   			void *old, void *entry, gfp_t gfp)
entry            1453 lib/xarray.c   	if (WARN_ON_ONCE(xa_is_advanced(entry)))
entry            1459 lib/xarray.c   			xas_store(&xas, entry);
entry            1460 lib/xarray.c   			if (xa_track_free(xa) && entry && !curr)
entry            1485 lib/xarray.c   int __xa_insert(struct xarray *xa, unsigned long index, void *entry, gfp_t gfp)
entry            1490 lib/xarray.c   	if (WARN_ON_ONCE(xa_is_advanced(entry)))
entry            1492 lib/xarray.c   	if (!entry)
entry            1493 lib/xarray.c   		entry = XA_ZERO_ENTRY;
entry            1498 lib/xarray.c   			xas_store(&xas, entry);
entry            1561 lib/xarray.c   		unsigned long last, void *entry, gfp_t gfp)
entry            1565 lib/xarray.c   	if (WARN_ON_ONCE(xa_is_internal(entry)))
entry            1572 lib/xarray.c   		if (entry) {
entry            1583 lib/xarray.c   			xas_store(&xas, entry);
entry            1614 lib/xarray.c   int __xa_alloc(struct xarray *xa, u32 *id, void *entry,
entry            1619 lib/xarray.c   	if (WARN_ON_ONCE(xa_is_advanced(entry)))
entry            1624 lib/xarray.c   	if (!entry)
entry            1625 lib/xarray.c   		entry = XA_ZERO_ENTRY;
entry            1634 lib/xarray.c   		xas_store(&xas, entry);
entry            1663 lib/xarray.c   int __xa_alloc_cyclic(struct xarray *xa, u32 *id, void *entry,
entry            1670 lib/xarray.c   	ret = __xa_alloc(xa, id, entry, limit, gfp);
entry            1678 lib/xarray.c   		ret = __xa_alloc(xa, id, entry, limit, gfp);
entry            1705 lib/xarray.c   	void *entry = xas_load(&xas);
entry            1707 lib/xarray.c   	if (entry)
entry            1723 lib/xarray.c   	void *entry = xas_load(&xas);
entry            1725 lib/xarray.c   	if (entry)
entry            1745 lib/xarray.c   	void *entry;
entry            1748 lib/xarray.c   	entry = xas_start(&xas);
entry            1750 lib/xarray.c   		if (!xa_is_node(entry))
entry            1752 lib/xarray.c   		entry = xas_descend(&xas, xa_to_node(entry));
entry            1819 lib/xarray.c   	void *entry;
entry            1824 lib/xarray.c   			entry = xas_find_marked(&xas, max, filter);
entry            1826 lib/xarray.c   			entry = xas_find(&xas, max);
entry            1827 lib/xarray.c   	} while (xas_retry(&xas, entry));
entry            1830 lib/xarray.c   	if (entry)
entry            1832 lib/xarray.c   	return entry;
entry            1869 lib/xarray.c   	void *entry;
entry            1877 lib/xarray.c   			entry = xas_find_marked(&xas, max, filter);
entry            1879 lib/xarray.c   			entry = xas_find(&xas, max);
entry            1885 lib/xarray.c   		if (!xas_retry(&xas, entry))
entry            1890 lib/xarray.c   	if (entry)
entry            1892 lib/xarray.c   	return entry;
entry            1899 lib/xarray.c   	void *entry;
entry            1903 lib/xarray.c   	xas_for_each(xas, entry, max) {
entry            1904 lib/xarray.c   		if (xas_retry(xas, entry))
entry            1906 lib/xarray.c   		dst[i++] = entry;
entry            1918 lib/xarray.c   	void *entry;
entry            1922 lib/xarray.c   	xas_for_each_marked(xas, entry, max, mark) {
entry            1923 lib/xarray.c   		if (xas_retry(xas, entry))
entry            1925 lib/xarray.c   		dst[i++] = entry;
entry            1990 lib/xarray.c   	void *entry;
entry            1994 lib/xarray.c   	entry = xa_head_locked(xa);
entry            2000 lib/xarray.c   	if (xa_is_node(entry))
entry            2001 lib/xarray.c   		xas_free_nodes(&xas, xa_to_node(entry));
entry            2039 lib/xarray.c   void xa_dump_entry(const void *entry, unsigned long index, unsigned long shift)
entry            2041 lib/xarray.c   	if (!entry)
entry            2046 lib/xarray.c   	if (xa_is_node(entry)) {
entry            2048 lib/xarray.c   			pr_cont("%px\n", entry);
entry            2051 lib/xarray.c   			struct xa_node *node = xa_to_node(entry);
entry            2057 lib/xarray.c   	} else if (xa_is_value(entry))
entry            2058 lib/xarray.c   		pr_cont("value %ld (0x%lx) [%px]\n", xa_to_value(entry),
entry            2059 lib/xarray.c   						xa_to_value(entry), entry);
entry            2060 lib/xarray.c   	else if (!xa_is_internal(entry))
entry            2061 lib/xarray.c   		pr_cont("%px\n", entry);
entry            2062 lib/xarray.c   	else if (xa_is_retry(entry))
entry            2063 lib/xarray.c   		pr_cont("retry (%ld)\n", xa_to_internal(entry));
entry            2064 lib/xarray.c   	else if (xa_is_sibling(entry))
entry            2065 lib/xarray.c   		pr_cont("sibling (slot %ld)\n", xa_to_sibling(entry));
entry            2066 lib/xarray.c   	else if (xa_is_zero(entry))
entry            2067 lib/xarray.c   		pr_cont("zero (%ld)\n", xa_to_internal(entry));
entry            2069 lib/xarray.c   		pr_cont("UNKNOWN ENTRY (%px)\n", entry);
entry            2074 lib/xarray.c   	void *entry = xa->xa_head;
entry            2077 lib/xarray.c   	pr_info("xarray: %px head %px flags %x marks %d %d %d\n", xa, entry,
entry            2080 lib/xarray.c   	if (xa_is_node(entry))
entry            2081 lib/xarray.c   		shift = xa_to_node(entry)->shift + XA_CHUNK_SHIFT;
entry            2082 lib/xarray.c   	xa_dump_entry(entry, 0, shift);
entry            1069 mm/filemap.c   	INIT_LIST_HEAD(&bookmark.entry);
entry            1161 mm/filemap.c   		if (likely(list_empty(&wait->entry))) {
entry            1465 mm/filemap.c   		void *entry = xas_next(&xas);
entry            1466 mm/filemap.c   		if (!entry || xa_is_value(entry))
entry            1501 mm/filemap.c   		void *entry = xas_prev(&xas);
entry            1502 mm/filemap.c   		if (!entry || xa_is_value(entry))
entry             248 mm/frontswap.c 	swp_entry_t entry = { .val = page_private(page), };
entry             249 mm/frontswap.c 	int type = swp_type(entry);
entry             251 mm/frontswap.c 	pgoff_t offset = swp_offset(entry);
entry             297 mm/frontswap.c 	swp_entry_t entry = { .val = page_private(page), };
entry             298 mm/frontswap.c 	int type = swp_type(entry);
entry             300 mm/frontswap.c 	pgoff_t offset = swp_offset(entry);
entry             147 mm/gup.c       		pte_t entry = *pte;
entry             150 mm/gup.c       			entry = pte_mkdirty(entry);
entry             151 mm/gup.c       		entry = pte_mkyoung(entry);
entry             153 mm/gup.c       		if (!pte_same(*pte, entry)) {
entry             154 mm/gup.c       			set_pte_at(vma->vm_mm, address, pte, entry);
entry             198 mm/gup.c       		swp_entry_t entry;
entry             208 mm/gup.c       		entry = pte_to_swp_entry(pte);
entry             209 mm/gup.c       		if (!is_migration_entry(entry))
entry             477 mm/hmm.c       		swp_entry_t entry = pte_to_swp_entry(pte);
entry             479 mm/hmm.c       		if (!non_swap_entry(entry)) {
entry             492 mm/hmm.c       		if (is_device_private_entry(entry)) {
entry             495 mm/hmm.c       			cpu_flags |= is_write_device_private_entry(entry) ?
entry             502 mm/hmm.c       					    swp_offset(entry));
entry             507 mm/hmm.c       		if (is_migration_entry(entry)) {
entry             737 mm/hmm.c       	pte_t entry;
entry             741 mm/hmm.c       	entry = huge_ptep_get(pte);
entry             746 mm/hmm.c       	cpu_flags = pte_to_hmm_pfn_flags(range, entry);
entry             755 mm/hmm.c       	pfn = pte_pfn(entry) + ((start & ~hmask) >> PAGE_SHIFT);
entry             610 mm/huge_memory.c 		pmd_t entry;
entry             629 mm/huge_memory.c 		entry = mk_huge_pmd(page, vma->vm_page_prot);
entry             630 mm/huge_memory.c 		entry = maybe_pmd_mkwrite(pmd_mkdirty(entry), vma);
entry             635 mm/huge_memory.c 		set_pmd_at(vma->vm_mm, haddr, vmf->pmd, entry);
entry             695 mm/huge_memory.c 	pmd_t entry;
entry             698 mm/huge_memory.c 	entry = mk_pmd(zero_page, vma->vm_page_prot);
entry             699 mm/huge_memory.c 	entry = pmd_mkhuge(entry);
entry             702 mm/huge_memory.c 	set_pmd_at(mm, haddr, pmd, entry);
entry             774 mm/huge_memory.c 	pmd_t entry;
entry             784 mm/huge_memory.c 			entry = pmd_mkyoung(*pmd);
entry             785 mm/huge_memory.c 			entry = maybe_pmd_mkwrite(pmd_mkdirty(entry), vma);
entry             786 mm/huge_memory.c 			if (pmdp_set_access_flags(vma, addr, pmd, entry, 1))
entry             793 mm/huge_memory.c 	entry = pmd_mkhuge(pfn_t_pmd(pfn, prot));
entry             795 mm/huge_memory.c 		entry = pmd_mkdevmap(entry);
entry             797 mm/huge_memory.c 		entry = pmd_mkyoung(pmd_mkdirty(entry));
entry             798 mm/huge_memory.c 		entry = maybe_pmd_mkwrite(entry, vma);
entry             807 mm/huge_memory.c 	set_pmd_at(mm, addr, pmd, entry);
entry             862 mm/huge_memory.c 	pud_t entry;
entry             872 mm/huge_memory.c 			entry = pud_mkyoung(*pud);
entry             873 mm/huge_memory.c 			entry = maybe_pud_mkwrite(pud_mkdirty(entry), vma);
entry             874 mm/huge_memory.c 			if (pudp_set_access_flags(vma, addr, pud, entry, 1))
entry             880 mm/huge_memory.c 	entry = pud_mkhuge(pfn_t_pud(pfn, prot));
entry             882 mm/huge_memory.c 		entry = pud_mkdevmap(entry);
entry             884 mm/huge_memory.c 		entry = pud_mkyoung(pud_mkdirty(entry));
entry             885 mm/huge_memory.c 		entry = maybe_pud_mkwrite(entry, vma);
entry             887 mm/huge_memory.c 	set_pud_at(mm, addr, pud, entry);
entry            1005 mm/huge_memory.c 		swp_entry_t entry = pmd_to_swp_entry(pmd);
entry            1008 mm/huge_memory.c 		if (is_write_migration_entry(entry)) {
entry            1009 mm/huge_memory.c 			make_migration_entry_read(&entry);
entry            1010 mm/huge_memory.c 			pmd = swp_entry_to_pmd(entry);
entry            1157 mm/huge_memory.c 	pud_t entry;
entry            1165 mm/huge_memory.c 	entry = pud_mkyoung(orig_pud);
entry            1167 mm/huge_memory.c 		entry = pud_mkdirty(entry);
entry            1169 mm/huge_memory.c 	if (pudp_set_access_flags(vmf->vma, haddr, vmf->pud, entry, write))
entry            1179 mm/huge_memory.c 	pmd_t entry;
entry            1187 mm/huge_memory.c 	entry = pmd_mkyoung(orig_pmd);
entry            1189 mm/huge_memory.c 		entry = pmd_mkdirty(entry);
entry            1191 mm/huge_memory.c 	if (pmdp_set_access_flags(vmf->vma, haddr, vmf->pmd, entry, write))
entry            1270 mm/huge_memory.c 		pte_t entry;
entry            1271 mm/huge_memory.c 		entry = mk_pte(pages[i], vma->vm_page_prot);
entry            1272 mm/huge_memory.c 		entry = maybe_mkwrite(pte_mkdirty(entry), vma);
entry            1280 mm/huge_memory.c 		set_pte_at(vma->vm_mm, haddr, vmf->pte, entry);
entry            1352 mm/huge_memory.c 		pmd_t entry;
entry            1353 mm/huge_memory.c 		entry = pmd_mkyoung(orig_pmd);
entry            1354 mm/huge_memory.c 		entry = maybe_pmd_mkwrite(pmd_mkdirty(entry), vma);
entry            1355 mm/huge_memory.c 		if (pmdp_set_access_flags(vma, haddr, vmf->pmd, entry,  1))
entry            1424 mm/huge_memory.c 		pmd_t entry;
entry            1425 mm/huge_memory.c 		entry = mk_huge_pmd(new_page, vma->vm_page_prot);
entry            1426 mm/huge_memory.c 		entry = maybe_pmd_mkwrite(pmd_mkdirty(entry), vma);
entry            1431 mm/huge_memory.c 		set_pmd_at(vma->vm_mm, haddr, vmf->pmd, entry);
entry            1813 mm/huge_memory.c 			swp_entry_t entry;
entry            1816 mm/huge_memory.c 			entry = pmd_to_swp_entry(orig_pmd);
entry            1817 mm/huge_memory.c 			page = pfn_to_page(swp_offset(entry));
entry            1929 mm/huge_memory.c 	pmd_t entry;
entry            1942 mm/huge_memory.c 		swp_entry_t entry = pmd_to_swp_entry(*pmd);
entry            1945 mm/huge_memory.c 		if (is_write_migration_entry(entry)) {
entry            1951 mm/huge_memory.c 			make_migration_entry_read(&entry);
entry            1952 mm/huge_memory.c 			newpmd = swp_entry_to_pmd(entry);
entry            1993 mm/huge_memory.c 	entry = pmdp_invalidate(vma, addr, pmd);
entry            1995 mm/huge_memory.c 	entry = pmd_modify(entry, newprot);
entry            1997 mm/huge_memory.c 		entry = pmd_mk_savedwrite(entry);
entry            1999 mm/huge_memory.c 	set_pmd_at(mm, addr, pmd, entry);
entry            2000 mm/huge_memory.c 	BUG_ON(vma_is_anonymous(vma) && !preserve_write && pmd_write(entry));
entry            2127 mm/huge_memory.c 		pte_t *pte, entry;
entry            2128 mm/huge_memory.c 		entry = pfn_pte(my_zero_pfn(haddr), vma->vm_page_prot);
entry            2129 mm/huge_memory.c 		entry = pte_mkspecial(entry);
entry            2132 mm/huge_memory.c 		set_pte_at(mm, haddr, pte, entry);
entry            2214 mm/huge_memory.c 		swp_entry_t entry;
entry            2216 mm/huge_memory.c 		entry = pmd_to_swp_entry(old_pmd);
entry            2217 mm/huge_memory.c 		page = pfn_to_page(swp_offset(entry));
entry            2218 mm/huge_memory.c 		write = is_write_migration_entry(entry);
entry            2240 mm/huge_memory.c 		pte_t entry, *pte;
entry            2249 mm/huge_memory.c 			entry = swp_entry_to_pte(swp_entry);
entry            2251 mm/huge_memory.c 				entry = pte_swp_mksoft_dirty(entry);
entry            2253 mm/huge_memory.c 			entry = mk_pte(page + i, READ_ONCE(vma->vm_page_prot));
entry            2254 mm/huge_memory.c 			entry = maybe_mkwrite(entry, vma);
entry            2256 mm/huge_memory.c 				entry = pte_wrprotect(entry);
entry            2258 mm/huge_memory.c 				entry = pte_mkold(entry);
entry            2260 mm/huge_memory.c 				entry = pte_mksoft_dirty(entry);
entry            2264 mm/huge_memory.c 		set_pte_at(mm, addr, pte, entry);
entry            2514 mm/huge_memory.c 		swp_entry_t entry = { .val = page_private(head) };
entry            2516 mm/huge_memory.c 		offset = swp_offset(entry);
entry            2517 mm/huge_memory.c 		swap_cache = swap_address_space(entry);
entry            2803 mm/huge_memory.c 			swp_entry_t entry = { .val = page_private(head) };
entry            2805 mm/huge_memory.c 			ret = split_swap_cluster(entry);
entry            3027 mm/huge_memory.c 	swp_entry_t entry;
entry            3037 mm/huge_memory.c 	entry = make_migration_entry(page, pmd_write(pmdval));
entry            3038 mm/huge_memory.c 	pmdswp = swp_entry_to_pmd(entry);
entry            3053 mm/huge_memory.c 	swp_entry_t entry;
entry            3058 mm/huge_memory.c 	entry = pmd_to_swp_entry(*pvmw->pmd);
entry            3063 mm/huge_memory.c 	if (is_write_migration_entry(entry))
entry            3367 mm/hugetlb.c   	pte_t entry;
entry            3370 mm/hugetlb.c   		entry = huge_pte_mkwrite(huge_pte_mkdirty(mk_huge_pte(page,
entry            3373 mm/hugetlb.c   		entry = huge_pte_wrprotect(mk_huge_pte(page,
entry            3376 mm/hugetlb.c   	entry = pte_mkyoung(entry);
entry            3377 mm/hugetlb.c   	entry = pte_mkhuge(entry);
entry            3378 mm/hugetlb.c   	entry = arch_make_huge_pte(entry, vma, page, writable);
entry            3380 mm/hugetlb.c   	return entry;
entry            3386 mm/hugetlb.c   	pte_t entry;
entry            3388 mm/hugetlb.c   	entry = huge_pte_mkwrite(huge_pte_mkdirty(huge_ptep_get(ptep)));
entry            3389 mm/hugetlb.c   	if (huge_ptep_set_access_flags(vma, address, ptep, entry, 1))
entry            3422 mm/hugetlb.c   	pte_t *src_pte, *dst_pte, entry, dst_entry;
entry            3467 mm/hugetlb.c   		entry = huge_ptep_get(src_pte);
entry            3469 mm/hugetlb.c   		if (huge_pte_none(entry) || !huge_pte_none(dst_entry)) {
entry            3476 mm/hugetlb.c   		} else if (unlikely(is_hugetlb_entry_migration(entry) ||
entry            3477 mm/hugetlb.c   				    is_hugetlb_entry_hwpoisoned(entry))) {
entry            3478 mm/hugetlb.c   			swp_entry_t swp_entry = pte_to_swp_entry(entry);
entry            3486 mm/hugetlb.c   				entry = swp_entry_to_pte(swp_entry);
entry            3488 mm/hugetlb.c   						     entry, sz);
entry            3490 mm/hugetlb.c   			set_huge_swap_pte_at(dst, addr, dst_pte, entry, sz);
entry            3502 mm/hugetlb.c   			entry = huge_ptep_get(src_pte);
entry            3503 mm/hugetlb.c   			ptepage = pte_page(entry);
entry            3506 mm/hugetlb.c   			set_huge_pte_at(dst, addr, dst_pte, entry);
entry            4122 mm/hugetlb.c   	pte_t *ptep, entry;
entry            4136 mm/hugetlb.c   		entry = huge_ptep_get(ptep);
entry            4137 mm/hugetlb.c   		if (unlikely(is_hugetlb_entry_migration(entry))) {
entry            4140 mm/hugetlb.c   		} else if (unlikely(is_hugetlb_entry_hwpoisoned(entry)))
entry            4160 mm/hugetlb.c   	entry = huge_ptep_get(ptep);
entry            4161 mm/hugetlb.c   	if (huge_pte_none(entry)) {
entry            4175 mm/hugetlb.c   	if (!pte_present(entry))
entry            4186 mm/hugetlb.c   	if ((flags & FAULT_FLAG_WRITE) && !huge_pte_write(entry)) {
entry            4202 mm/hugetlb.c   	if (unlikely(!pte_same(entry, huge_ptep_get(ptep))))
entry            4210 mm/hugetlb.c   	page = pte_page(entry);
entry            4220 mm/hugetlb.c   		if (!huge_pte_write(entry)) {
entry            4225 mm/hugetlb.c   		entry = huge_pte_mkdirty(entry);
entry            4227 mm/hugetlb.c   	entry = pte_mkyoung(entry);
entry            4228 mm/hugetlb.c   	if (huge_ptep_set_access_flags(vma, haddr, ptep, entry,
entry            4597 mm/hugetlb.c   			swp_entry_t entry = pte_to_swp_entry(pte);
entry            4599 mm/hugetlb.c   			if (is_write_migration_entry(entry)) {
entry            4602 mm/hugetlb.c   				make_migration_entry_read(&entry);
entry            4603 mm/hugetlb.c   				newpte = swp_entry_to_pte(entry);
entry            1063 mm/ksm.c       		pte_t entry;
entry            1081 mm/ksm.c       		entry = ptep_clear_flush(vma, pvmw.address, pvmw.pte);
entry            1087 mm/ksm.c       			set_pte_at(mm, pvmw.address, pvmw.pte, entry);
entry            1090 mm/ksm.c       		if (pte_dirty(entry))
entry            1093 mm/ksm.c       		if (pte_protnone(entry))
entry            1094 mm/ksm.c       			entry = pte_mkclean(pte_clear_savedwrite(entry));
entry            1096 mm/ksm.c       			entry = pte_mkclean(pte_wrprotect(entry));
entry            1097 mm/ksm.c       		set_pte_at_notify(mm, pvmw.address, pvmw.pte, entry);
entry             195 mm/madvise.c   		swp_entry_t entry;
entry             205 mm/madvise.c   		entry = pte_to_swp_entry(pte);
entry             206 mm/madvise.c   		if (unlikely(non_swap_entry(entry)))
entry             209 mm/madvise.c   		page = read_swap_cache_async(entry, GFP_HIGHUSER_MOVABLE,
entry             598 mm/madvise.c   			swp_entry_t entry;
entry             600 mm/madvise.c   			entry = pte_to_swp_entry(ptent);
entry             601 mm/madvise.c   			if (non_swap_entry(entry))
entry             604 mm/madvise.c   			free_swap_and_cache(entry);
entry            1983 mm/memcontrol.c 	INIT_LIST_HEAD(&owait.wait.entry);
entry            3145 mm/memcontrol.c static int mem_cgroup_move_swap_account(swp_entry_t entry,
entry            3153 mm/memcontrol.c 	if (swap_cgroup_cmpxchg(entry, old_id, new_id) == old_id) {
entry            3161 mm/memcontrol.c static inline int mem_cgroup_move_swap_account(swp_entry_t entry,
entry            5394 mm/memcontrol.c 			pte_t ptent, swp_entry_t *entry)
entry            5424 mm/memcontrol.c 		entry->val = ent.val;
entry            5430 mm/memcontrol.c 			pte_t ptent, swp_entry_t *entry)
entry            5437 mm/memcontrol.c 			unsigned long addr, pte_t ptent, swp_entry_t *entry)
entry            5459 mm/memcontrol.c 				*entry = swp;
entry            6615 mm/memcontrol.c 		swp_entry_t entry = { .val = page_private(page) };
entry            6621 mm/memcontrol.c 		mem_cgroup_uncharge_swap(entry, nr_pages);
entry            7032 mm/memcontrol.c void mem_cgroup_swapout(struct page *page, swp_entry_t entry)
entry            7060 mm/memcontrol.c 	oldid = swap_cgroup_record(entry, mem_cgroup_id(swap_memcg),
entry            7100 mm/memcontrol.c int mem_cgroup_try_charge_swap(struct page *page, swp_entry_t entry)
entry            7116 mm/memcontrol.c 	if (!entry.val) {
entry            7134 mm/memcontrol.c 	oldid = swap_cgroup_record(entry, mem_cgroup_id(memcg), nr_pages);
entry            7146 mm/memcontrol.c void mem_cgroup_uncharge_swap(swp_entry_t entry, unsigned int nr_pages)
entry            7154 mm/memcontrol.c 	id = swap_cgroup_record(entry, 0, nr_pages);
entry            1461 mm/memory-failure.c 	struct memory_failure_entry entry = {
entry            1468 mm/memory-failure.c 	if (kfifo_put(&mf_cpu->fifo, entry))
entry            1481 mm/memory-failure.c 	struct memory_failure_entry entry = { 0, };
entry            1488 mm/memory-failure.c 		gotten = kfifo_get(&mf_cpu->fifo, &entry);
entry            1492 mm/memory-failure.c 		if (entry.flags & MF_SOFT_OFFLINE)
entry            1493 mm/memory-failure.c 			soft_offline_page(pfn_to_page(entry.pfn), entry.flags);
entry            1495 mm/memory-failure.c 			memory_failure(entry.pfn, entry.flags);
entry             688 mm/memory.c    		swp_entry_t entry = pte_to_swp_entry(pte);
entry             690 mm/memory.c    		if (likely(!non_swap_entry(entry))) {
entry             691 mm/memory.c    			if (swap_duplicate(entry) < 0)
entry             692 mm/memory.c    				return entry.val;
entry             703 mm/memory.c    		} else if (is_migration_entry(entry)) {
entry             704 mm/memory.c    			page = migration_entry_to_page(entry);
entry             708 mm/memory.c    			if (is_write_migration_entry(entry) &&
entry             714 mm/memory.c    				make_migration_entry_read(&entry);
entry             715 mm/memory.c    				pte = swp_entry_to_pte(entry);
entry             720 mm/memory.c    		} else if (is_device_private_entry(entry)) {
entry             721 mm/memory.c    			page = device_private_entry_to_page(entry);
entry             743 mm/memory.c    			if (is_write_device_private_entry(entry) &&
entry             745 mm/memory.c    				make_device_private_entry_read(&entry);
entry             746 mm/memory.c    				pte = swp_entry_to_pte(entry);
entry             793 mm/memory.c    	swp_entry_t entry = (swp_entry_t){0};
entry             823 mm/memory.c    		entry.val = copy_one_pte(dst_mm, src_mm, dst_pte, src_pte,
entry             825 mm/memory.c    		if (entry.val)
entry             837 mm/memory.c    	if (entry.val) {
entry             838 mm/memory.c    		if (add_swap_count_continuation(entry, GFP_KERNEL) < 0)
entry            1015 mm/memory.c    	swp_entry_t entry;
entry            1073 mm/memory.c    		entry = pte_to_swp_entry(ptent);
entry            1074 mm/memory.c    		if (non_swap_entry(entry) && is_device_private_entry(entry)) {
entry            1075 mm/memory.c    			struct page *page = device_private_entry_to_page(entry);
entry            1099 mm/memory.c    		if (!non_swap_entry(entry))
entry            1101 mm/memory.c    		else if (is_migration_entry(entry)) {
entry            1104 mm/memory.c    			page = migration_entry_to_page(entry);
entry            1107 mm/memory.c    		if (unlikely(!free_swap_and_cache(entry)))
entry            1585 mm/memory.c    	pte_t *pte, entry;
entry            1607 mm/memory.c    			entry = pte_mkyoung(*pte);
entry            1608 mm/memory.c    			entry = maybe_mkwrite(pte_mkdirty(entry), vma);
entry            1609 mm/memory.c    			if (ptep_set_access_flags(vma, addr, pte, entry, 1))
entry            1617 mm/memory.c    		entry = pte_mkdevmap(pfn_t_pte(pfn, prot));
entry            1619 mm/memory.c    		entry = pte_mkspecial(pfn_t_pte(pfn, prot));
entry            1622 mm/memory.c    		entry = pte_mkyoung(entry);
entry            1623 mm/memory.c    		entry = maybe_mkwrite(pte_mkdirty(entry), vma);
entry            1626 mm/memory.c    	set_pte_at(mm, addr, pte, entry);
entry            2288 mm/memory.c    	pte_t entry;
entry            2298 mm/memory.c    	entry = pte_mkyoung(vmf->orig_pte);
entry            2299 mm/memory.c    	entry = maybe_mkwrite(pte_mkdirty(entry), vma);
entry            2300 mm/memory.c    	if (ptep_set_access_flags(vma, vmf->address, vmf->pte, entry, 1))
entry            2327 mm/memory.c    	pte_t entry;
entry            2373 mm/memory.c    		entry = mk_pte(new_page, vma->vm_page_prot);
entry            2374 mm/memory.c    		entry = maybe_mkwrite(pte_mkdirty(entry), vma);
entry            2390 mm/memory.c    		set_pte_at_notify(mm, vmf->address, vmf->pte, entry);
entry            2756 mm/memory.c    	swp_entry_t entry;
entry            2765 mm/memory.c    	entry = pte_to_swp_entry(vmf->orig_pte);
entry            2766 mm/memory.c    	if (unlikely(non_swap_entry(entry))) {
entry            2767 mm/memory.c    		if (is_migration_entry(entry)) {
entry            2770 mm/memory.c    		} else if (is_device_private_entry(entry)) {
entry            2771 mm/memory.c    			vmf->page = device_private_entry_to_page(entry);
entry            2773 mm/memory.c    		} else if (is_hwpoison_entry(entry)) {
entry            2784 mm/memory.c    	page = lookup_swap_cache(entry, vma, vmf->address);
entry            2788 mm/memory.c    		struct swap_info_struct *si = swp_swap_info(entry);
entry            2791 mm/memory.c    				__swap_count(entry) == 1) {
entry            2798 mm/memory.c    				set_page_private(page, entry.val);
entry            2803 mm/memory.c    			page = swapin_readahead(entry, GFP_HIGHUSER_MOVABLE,
entry            2850 mm/memory.c    			page_private(page) != entry.val)) && swapcache)
entry            2916 mm/memory.c    	swap_free(entry);
entry            2972 mm/memory.c    	pte_t entry;
entry            2998 mm/memory.c    		entry = pte_mkspecial(pfn_pte(my_zero_pfn(vmf->address),
entry            3033 mm/memory.c    	entry = mk_pte(page, vma->vm_page_prot);
entry            3035 mm/memory.c    		entry = pte_mkwrite(pte_mkdirty(entry));
entry            3059 mm/memory.c    	set_pte_at(vma->vm_mm, vmf->address, vmf->pte, entry);
entry            3208 mm/memory.c    	pmd_t entry;
entry            3236 mm/memory.c    	entry = mk_huge_pmd(page, vma->vm_page_prot);
entry            3238 mm/memory.c    		entry = maybe_pmd_mkwrite(pmd_mkdirty(entry), vma);
entry            3248 mm/memory.c    	set_pmd_at(vma->vm_mm, haddr, vmf->pmd, entry);
entry            3288 mm/memory.c    	pte_t entry;
entry            3312 mm/memory.c    	entry = mk_pte(page, vma->vm_page_prot);
entry            3314 mm/memory.c    		entry = maybe_mkwrite(pte_mkdirty(entry), vma);
entry            3325 mm/memory.c    	set_pte_at(vma->vm_mm, vmf->address, vmf->pte, entry);
entry            3823 mm/memory.c    	pte_t entry;
entry            3876 mm/memory.c    	entry = vmf->orig_pte;
entry            3877 mm/memory.c    	if (unlikely(!pte_same(*vmf->pte, entry)))
entry            3880 mm/memory.c    		if (!pte_write(entry))
entry            3882 mm/memory.c    		entry = pte_mkdirty(entry);
entry            3884 mm/memory.c    	entry = pte_mkyoung(entry);
entry            3885 mm/memory.c    	if (ptep_set_access_flags(vmf->vma, vmf->address, vmf->pte, entry,
entry             563 mm/mempolicy.c 	pte_t entry;
entry             566 mm/mempolicy.c 	entry = huge_ptep_get(pte);
entry             567 mm/mempolicy.c 	if (!pte_present(entry))
entry             569 mm/mempolicy.c 	page = pte_page(entry);
entry             215 mm/migrate.c   	swp_entry_t entry;
entry             242 mm/migrate.c   		entry = pte_to_swp_entry(*pvmw.pte);
entry             243 mm/migrate.c   		if (is_write_migration_entry(entry))
entry             248 mm/migrate.c   				entry = make_device_private_entry(new, pte_write(pte));
entry             249 mm/migrate.c   				pte = swp_entry_to_pte(entry);
entry             311 mm/migrate.c   	swp_entry_t entry;
entry             319 mm/migrate.c   	entry = pte_to_swp_entry(pte);
entry             320 mm/migrate.c   	if (!is_migration_entry(entry))
entry             323 mm/migrate.c   	page = migration_entry_to_page(entry);
entry            2026 mm/migrate.c   				pmd_t *pmd, pmd_t entry,
entry            2065 mm/migrate.c   	if (unlikely(!pmd_same(*pmd, entry) || !page_ref_freeze(page, 2))) {
entry            2086 mm/migrate.c   	entry = mk_huge_pmd(new_page, vma->vm_page_prot);
entry            2087 mm/migrate.c   	entry = maybe_pmd_mkwrite(pmd_mkdirty(entry), vma);
entry            2109 mm/migrate.c   	set_pmd_at(mm, start, pmd, entry);
entry            2110 mm/migrate.c   	update_mmu_cache_pmd(vma, address, &entry);
entry            2139 mm/migrate.c   	if (pmd_same(*pmd, entry)) {
entry            2140 mm/migrate.c   		entry = pmd_modify(entry, vma->vm_page_prot);
entry            2141 mm/migrate.c   		set_pmd_at(mm, start, pmd, entry);
entry            2142 mm/migrate.c   		update_mmu_cache_pmd(vma, address, &entry);
entry            2249 mm/migrate.c   		swp_entry_t entry;
entry            2268 mm/migrate.c   			entry = pte_to_swp_entry(pte);
entry            2269 mm/migrate.c   			if (!is_device_private_entry(entry))
entry            2272 mm/migrate.c   			page = device_private_entry_to_page(entry);
entry            2275 mm/migrate.c   			if (is_write_device_private_entry(entry))
entry            2319 mm/migrate.c   			entry = make_migration_entry(page, mpfn &
entry            2321 mm/migrate.c   			swp_pte = swp_entry_to_pte(entry);
entry            2714 mm/migrate.c   	pte_t entry;
entry            2773 mm/migrate.c   			entry = swp_entry_to_pte(swp_entry);
entry            2776 mm/migrate.c   		entry = mk_pte(page, vma->vm_page_prot);
entry            2778 mm/migrate.c   			entry = pte_mkwrite(pte_mkdirty(entry));
entry            2818 mm/migrate.c   		set_pte_at_notify(mm, addr, ptep, entry);
entry            2822 mm/migrate.c   		set_pte_at(mm, addr, ptep, entry);
entry             153 mm/mincore.c   			swp_entry_t entry = pte_to_swp_entry(pte);
entry             155 mm/mincore.c   			if (non_swap_entry(entry)) {
entry             163 mm/mincore.c   				*vec = mincore_page(swap_address_space(entry),
entry             164 mm/mincore.c   						    swp_offset(entry));
entry             126 mm/mprotect.c  			swp_entry_t entry = pte_to_swp_entry(oldpte);
entry             128 mm/mprotect.c  			if (is_write_migration_entry(entry)) {
entry             134 mm/mprotect.c  				make_migration_entry_read(&entry);
entry             135 mm/mprotect.c  				newpte = swp_entry_to_pte(entry);
entry             143 mm/mprotect.c  			if (is_write_device_private_entry(entry)) {
entry             150 mm/mprotect.c  				make_device_private_entry_read(&entry);
entry             151 mm/mprotect.c  				newpte = swp_entry_to_pte(entry);
entry              76 mm/page_io.c   	swp_entry_t entry;
entry             108 mm/page_io.c   	entry.val = page_private(page);
entry             109 mm/page_io.c   	if (disk->fops->swap_slot_free_notify && __swap_count(entry) == 1) {
entry             112 mm/page_io.c   		offset = swp_offset(entry);
entry              40 mm/page_vma_mapped.c 				swp_entry_t entry;
entry              43 mm/page_vma_mapped.c 				entry = pte_to_swp_entry(*pvmw->pte);
entry              44 mm/page_vma_mapped.c 				if (!is_device_private_entry(entry))
entry              86 mm/page_vma_mapped.c 		swp_entry_t entry;
entry              89 mm/page_vma_mapped.c 		entry = pte_to_swp_entry(*pvmw->pte);
entry              91 mm/page_vma_mapped.c 		if (!is_migration_entry(entry))
entry              94 mm/page_vma_mapped.c 		pfn = migration_entry_to_pfn(entry);
entry              96 mm/page_vma_mapped.c 		swp_entry_t entry;
entry              99 mm/page_vma_mapped.c 		entry = pte_to_swp_entry(*pvmw->pte);
entry             100 mm/page_vma_mapped.c 		if (!is_device_private_entry(entry))
entry             103 mm/page_vma_mapped.c 		pfn = device_private_entry_to_pfn(entry);
entry             196 mm/page_vma_mapped.c 					swp_entry_t entry = pmd_to_swp_entry(*pvmw->pmd);
entry             198 mm/page_vma_mapped.c 					if (migration_entry_to_page(entry) != page)
entry              57 mm/pgtable-generic.c 			  pte_t entry, int dirty)
entry              59 mm/pgtable-generic.c 	int changed = !pte_same(*ptep, entry);
entry              61 mm/pgtable-generic.c 		set_pte_at(vma->vm_mm, address, ptep, entry);
entry              98 mm/pgtable-generic.c 			  pmd_t entry, int dirty)
entry             100 mm/pgtable-generic.c 	int changed = !pmd_same(*pmdp, entry);
entry             103 mm/pgtable-generic.c 		set_pmd_at(vma->vm_mm, address, pmdp, entry);
entry             910 mm/rmap.c      			pte_t entry;
entry             917 mm/rmap.c      			entry = ptep_clear_flush(vma, address, pte);
entry             918 mm/rmap.c      			entry = pte_wrprotect(entry);
entry             919 mm/rmap.c      			entry = pte_mkclean(entry);
entry             920 mm/rmap.c      			set_pte_at(vma->vm_mm, address, pte, entry);
entry             925 mm/rmap.c      			pmd_t entry;
entry             931 mm/rmap.c      			entry = pmdp_invalidate(vma, address, pmd);
entry             932 mm/rmap.c      			entry = pmd_wrprotect(entry);
entry             933 mm/rmap.c      			entry = pmd_mkclean(entry);
entry             934 mm/rmap.c      			set_pmd_at(vma->vm_mm, address, pmd, entry);
entry            1461 mm/rmap.c      			swp_entry_t entry;
entry            1471 mm/rmap.c      			entry = make_migration_entry(page, 0);
entry            1472 mm/rmap.c      			swp_pte = swp_entry_to_pte(entry);
entry            1554 mm/rmap.c      			swp_entry_t entry;
entry            1569 mm/rmap.c      			entry = make_migration_entry(subpage,
entry            1571 mm/rmap.c      			swp_pte = swp_entry_to_pte(entry);
entry            1580 mm/rmap.c      			swp_entry_t entry = { .val = page_private(subpage) };
entry            1617 mm/rmap.c      			if (swap_duplicate(entry) < 0) {
entry            1637 mm/rmap.c      			swp_pte = swp_entry_to_pte(entry);
entry             625 mm/shmem.c     		void *entry;
entry             627 mm/shmem.c     		entry = xas_find_conflict(&xas);
entry             628 mm/shmem.c     		if (entry != expected)
entry            1128 mm/shmem.c     	swp_entry_t entry;
entry            1142 mm/shmem.c     		entry = radix_to_swp_entry(page);
entry            1143 mm/shmem.c     		if (swp_type(entry) != type)
entry            1146 mm/shmem.c     		    !frontswap_test(swap_info[type], swp_offset(entry)))
entry            1558 mm/shmem.c     	swp_entry_t entry;
entry            1563 mm/shmem.c     	entry.val = page_private(oldpage);
entry            1564 mm/shmem.c     	swap_index = swp_offset(entry);
entry            1583 mm/shmem.c     	set_page_private(newpage, entry.val);
entry            1986 mm/shmem.c     	list_del_init(&wait->entry);
entry             189 mm/slab.c      	void *entry[];	/*
entry             583 mm/slab.c      	memcpy(to->entry + to->avail, from->entry + from->avail -nr,
entry             700 mm/slab.c      		free_block(cachep, ac->entry, ac->avail, node, list);
entry             770 mm/slab.c      		ac->entry[ac->avail++] = objp;
entry             902 mm/slab.c      		free_block(cachep, n->shared->entry,
entry             964 mm/slab.c      		free_block(cachep, nc->entry, nc->avail, node, &list);
entry             974 mm/slab.c      			free_block(cachep, shared->entry,
entry            2131 mm/slab.c      	free_block(cachep, ac->entry, tofree, node, list);
entry            2133 mm/slab.c      	memmove(ac->entry, &(ac->entry[tofree]), sizeof(void *) * ac->avail);
entry            2148 mm/slab.c      	free_block(cachep, ac->entry, ac->avail, node, &list);
entry            2899 mm/slab.c      		ac->entry[ac->avail++] = slab_get_obj(cachep, page);
entry            2987 mm/slab.c      	return ac->entry[--ac->avail];
entry            3047 mm/slab.c      		objp = ac->entry[--ac->avail];
entry            3068 mm/slab.c      		kmemleak_erase(&ac->entry[ac->avail]);
entry            3389 mm/slab.c      			memcpy(&(shared_array->entry[shared_array->avail]),
entry            3390 mm/slab.c      			       ac->entry, sizeof(void *) * batchcount);
entry            3396 mm/slab.c      	free_block(cachep, ac->entry, batchcount, node, &list);
entry            3414 mm/slab.c      	memmove(ac->entry, &(ac->entry[batchcount]), sizeof(void *)*ac->avail);
entry            3468 mm/slab.c      	ac->entry[ac->avail++] = objp;
entry            3834 mm/slab.c      		free_block(cachep, ac->entry, ac->avail, node, &list);
entry             147 mm/sparse-vmemmap.c 		pte_t entry;
entry             151 mm/sparse-vmemmap.c 		entry = pfn_pte(__pa(p) >> PAGE_SHIFT, PAGE_KERNEL);
entry             152 mm/sparse-vmemmap.c 		set_pte_at(&init_mm, addr, pte, entry);
entry             278 mm/swap_slots.c int free_swap_slot(swp_entry_t entry)
entry             300 mm/swap_slots.c 		cache->slots_ret[cache->n_ret++] = entry;
entry             304 mm/swap_slots.c 		swapcache_free_entries(&entry, 1);
entry             312 mm/swap_slots.c 	swp_entry_t entry, *pentry;
entry             315 mm/swap_slots.c 	entry.val = 0;
entry             319 mm/swap_slots.c 			get_swap_pages(1, &entry, HPAGE_PMD_NR);
entry             340 mm/swap_slots.c 				entry = *pentry;
entry             349 mm/swap_slots.c 		if (entry.val)
entry             353 mm/swap_slots.c 	get_swap_pages(1, &entry, 1);
entry             355 mm/swap_slots.c 	if (mem_cgroup_try_charge_swap(page, entry)) {
entry             356 mm/swap_slots.c 		put_swap_page(page, entry);
entry             357 mm/swap_slots.c 		entry.val = 0;
entry             359 mm/swap_slots.c 	return entry;
entry              79 mm/swap_state.c 		swp_entry_t entry = swp_entry(i, 1);
entry              82 mm/swap_state.c 		if (!swp_swap_info(entry))
entry              85 mm/swap_state.c 		si = get_swap_device(entry);
entry             114 mm/swap_state.c int add_to_swap_cache(struct page *page, swp_entry_t entry, gfp_t gfp)
entry             116 mm/swap_state.c 	struct address_space *address_space = swap_address_space(entry);
entry             117 mm/swap_state.c 	pgoff_t idx = swp_offset(entry);
entry             135 mm/swap_state.c 			set_page_private(page + i, entry.val + i);
entry             158 mm/swap_state.c void __delete_from_swap_cache(struct page *page, swp_entry_t entry)
entry             160 mm/swap_state.c 	struct address_space *address_space = swap_address_space(entry);
entry             162 mm/swap_state.c 	pgoff_t idx = swp_offset(entry);
entry             170 mm/swap_state.c 		void *entry = xas_store(&xas, NULL);
entry             171 mm/swap_state.c 		VM_BUG_ON_PAGE(entry != page, entry);
entry             190 mm/swap_state.c 	swp_entry_t entry;
entry             196 mm/swap_state.c 	entry = get_swap_page(page);
entry             197 mm/swap_state.c 	if (!entry.val)
entry             211 mm/swap_state.c 	err = add_to_swap_cache(page, entry,
entry             234 mm/swap_state.c 	put_swap_page(page, entry);
entry             246 mm/swap_state.c 	swp_entry_t entry = { .val = page_private(page) };
entry             247 mm/swap_state.c 	struct address_space *address_space = swap_address_space(entry);
entry             250 mm/swap_state.c 	__delete_from_swap_cache(page, entry);
entry             253 mm/swap_state.c 	put_swap_page(page, entry);
entry             310 mm/swap_state.c struct page *lookup_swap_cache(swp_entry_t entry, struct vm_area_struct *vma,
entry             316 mm/swap_state.c 	si = get_swap_device(entry);
entry             319 mm/swap_state.c 	page = find_get_page(swap_address_space(entry), swp_offset(entry));
entry             359 mm/swap_state.c struct page *__read_swap_cache_async(swp_entry_t entry, gfp_t gfp_mask,
entry             374 mm/swap_state.c 		si = get_swap_device(entry);
entry             377 mm/swap_state.c 		found_page = find_get_page(swap_address_space(entry),
entry             378 mm/swap_state.c 					   swp_offset(entry));
entry             391 mm/swap_state.c 		if (!__swp_swapcount(entry) && swap_slot_cache_enabled)
entry             406 mm/swap_state.c 		err = swapcache_prepare(entry);
entry             421 mm/swap_state.c 		err = add_to_swap_cache(new_page, entry, gfp_mask & GFP_KERNEL);
entry             434 mm/swap_state.c 		put_swap_page(new_page, entry);
entry             448 mm/swap_state.c struct page *read_swap_cache_async(swp_entry_t entry, gfp_t gfp_mask,
entry             452 mm/swap_state.c 	struct page *retpage = __read_swap_cache_async(entry, gfp_mask,
entry             539 mm/swap_state.c struct page *swap_cluster_readahead(swp_entry_t entry, gfp_t gfp_mask,
entry             543 mm/swap_state.c 	unsigned long entry_offset = swp_offset(entry);
entry             547 mm/swap_state.c 	struct swap_info_struct *si = swp_swap_info(entry);
entry             577 mm/swap_state.c 			swp_entry(swp_type(entry), offset),
entry             594 mm/swap_state.c 	return read_swap_cache_async(entry, gfp_mask, vma, addr, do_poll);
entry             645 mm/swap_state.c 	swp_entry_t entry;
entry             663 mm/swap_state.c 	entry = pte_to_swp_entry(*pte);
entry             664 mm/swap_state.c 	if ((unlikely(non_swap_entry(entry)))) {
entry             729 mm/swap_state.c 	swp_entry_t entry;
entry             746 mm/swap_state.c 		entry = pte_to_swp_entry(pentry);
entry             747 mm/swap_state.c 		if (unlikely(non_swap_entry(entry)))
entry             749 mm/swap_state.c 		page = __read_swap_cache_async(entry, gfp_mask, vma,
entry             781 mm/swap_state.c struct page *swapin_readahead(swp_entry_t entry, gfp_t gfp_mask,
entry             785 mm/swap_state.c 			swap_vma_readahead(entry, gfp_mask, vmf) :
entry             786 mm/swap_state.c 			swap_cluster_readahead(entry, gfp_mask, vmf);
entry             130 mm/swapfile.c  	swp_entry_t entry = swp_entry(si->type, offset);
entry             134 mm/swapfile.c  	page = find_get_page(swap_address_space(entry), offset);
entry             980 mm/swapfile.c  	swp_entry_t entry;
entry             983 mm/swapfile.c  	n_ret = scan_swap_map_slots(si, usage, 1, &entry);
entry             986 mm/swapfile.c  		return swp_offset(entry);
entry            1104 mm/swapfile.c  static struct swap_info_struct *__swap_info_get(swp_entry_t entry)
entry            1109 mm/swapfile.c  	if (!entry.val)
entry            1111 mm/swapfile.c  	p = swp_swap_info(entry);
entry            1116 mm/swapfile.c  	offset = swp_offset(entry);
entry            1122 mm/swapfile.c  	pr_err("swap_info_get: %s%08lx\n", Bad_offset, entry.val);
entry            1125 mm/swapfile.c  	pr_err("swap_info_get: %s%08lx\n", Unused_file, entry.val);
entry            1128 mm/swapfile.c  	pr_err("swap_info_get: %s%08lx\n", Bad_file, entry.val);
entry            1133 mm/swapfile.c  static struct swap_info_struct *_swap_info_get(swp_entry_t entry)
entry            1137 mm/swapfile.c  	p = __swap_info_get(entry);
entry            1140 mm/swapfile.c  	if (!p->swap_map[swp_offset(entry)])
entry            1145 mm/swapfile.c  	pr_err("swap_info_get: %s%08lx\n", Unused_offset, entry.val);
entry            1151 mm/swapfile.c  static struct swap_info_struct *swap_info_get(swp_entry_t entry)
entry            1155 mm/swapfile.c  	p = _swap_info_get(entry);
entry            1161 mm/swapfile.c  static struct swap_info_struct *swap_info_get_cont(swp_entry_t entry,
entry            1166 mm/swapfile.c  	p = _swap_info_get(entry);
entry            1249 mm/swapfile.c  struct swap_info_struct *get_swap_device(swp_entry_t entry)
entry            1254 mm/swapfile.c  	if (!entry.val)
entry            1256 mm/swapfile.c  	si = swp_swap_info(entry);
entry            1263 mm/swapfile.c  	offset = swp_offset(entry);
entry            1269 mm/swapfile.c  	pr_err("%s: %s%08lx\n", __func__, Bad_file, entry.val);
entry            1278 mm/swapfile.c  				       swp_entry_t entry, unsigned char usage)
entry            1281 mm/swapfile.c  	unsigned long offset = swp_offset(entry);
entry            1287 mm/swapfile.c  		free_swap_slot(entry);
entry            1292 mm/swapfile.c  static void swap_entry_free(struct swap_info_struct *p, swp_entry_t entry)
entry            1295 mm/swapfile.c  	unsigned long offset = swp_offset(entry);
entry            1305 mm/swapfile.c  	mem_cgroup_uncharge_swap(entry, 1);
entry            1313 mm/swapfile.c  void swap_free(swp_entry_t entry)
entry            1317 mm/swapfile.c  	p = _swap_info_get(entry);
entry            1319 mm/swapfile.c  		__swap_entry_free(p, entry, 1);
entry            1325 mm/swapfile.c  void put_swap_page(struct page *page, swp_entry_t entry)
entry            1327 mm/swapfile.c  	unsigned long offset = swp_offset(entry);
entry            1336 mm/swapfile.c  	si = _swap_info_get(entry);
entry            1354 mm/swapfile.c  			mem_cgroup_uncharge_swap(entry, SWAPFILE_CLUSTER);
entry            1360 mm/swapfile.c  	for (i = 0; i < size; i++, entry.val++) {
entry            1363 mm/swapfile.c  			free_swap_slot(entry);
entry            1373 mm/swapfile.c  int split_swap_cluster(swp_entry_t entry)
entry            1377 mm/swapfile.c  	unsigned long offset = swp_offset(entry);
entry            1379 mm/swapfile.c  	si = _swap_info_get(entry);
entry            1434 mm/swapfile.c  	swp_entry_t entry;
entry            1437 mm/swapfile.c  	entry.val = page_private(page);
entry            1438 mm/swapfile.c  	p = _swap_info_get(entry);
entry            1440 mm/swapfile.c  		offset = swp_offset(entry);
entry            1448 mm/swapfile.c  int __swap_count(swp_entry_t entry)
entry            1451 mm/swapfile.c  	pgoff_t offset = swp_offset(entry);
entry            1454 mm/swapfile.c  	si = get_swap_device(entry);
entry            1462 mm/swapfile.c  static int swap_swapcount(struct swap_info_struct *si, swp_entry_t entry)
entry            1465 mm/swapfile.c  	pgoff_t offset = swp_offset(entry);
entry            1479 mm/swapfile.c  int __swp_swapcount(swp_entry_t entry)
entry            1484 mm/swapfile.c  	si = get_swap_device(entry);
entry            1486 mm/swapfile.c  		count = swap_swapcount(si, entry);
entry            1496 mm/swapfile.c  int swp_swapcount(swp_entry_t entry)
entry            1505 mm/swapfile.c  	p = _swap_info_get(entry);
entry            1509 mm/swapfile.c  	offset = swp_offset(entry);
entry            1539 mm/swapfile.c  					 swp_entry_t entry)
entry            1543 mm/swapfile.c  	unsigned long roffset = swp_offset(entry);
entry            1567 mm/swapfile.c  	swp_entry_t entry;
entry            1574 mm/swapfile.c  	entry.val = page_private(page);
entry            1575 mm/swapfile.c  	si = _swap_info_get(entry);
entry            1577 mm/swapfile.c  		return swap_page_trans_huge_swapped(si, entry);
entry            1607 mm/swapfile.c  		swp_entry_t entry;
entry            1609 mm/swapfile.c  		entry.val = page_private(page);
entry            1610 mm/swapfile.c  		si = _swap_info_get(entry);
entry            1613 mm/swapfile.c  			offset = swp_offset(entry);
entry            1673 mm/swapfile.c  			swp_entry_t entry;
entry            1676 mm/swapfile.c  			entry.val = page_private(page);
entry            1677 mm/swapfile.c  			p = swap_info_get(entry);
entry            1732 mm/swapfile.c  int free_swap_and_cache(swp_entry_t entry)
entry            1737 mm/swapfile.c  	if (non_swap_entry(entry))
entry            1740 mm/swapfile.c  	p = _swap_info_get(entry);
entry            1742 mm/swapfile.c  		count = __swap_entry_free(p, entry, 1);
entry            1744 mm/swapfile.c  		    !swap_page_trans_huge_swapped(p, entry))
entry            1745 mm/swapfile.c  			__try_to_reclaim_swap(p, swp_offset(entry),
entry            1854 mm/swapfile.c  		unsigned long addr, swp_entry_t entry, struct page *page)
entry            1874 mm/swapfile.c  	if (unlikely(!pte_same_as_swp(*pte, swp_entry_to_pte(entry)))) {
entry            1893 mm/swapfile.c  	swap_free(entry);
entry            1915 mm/swapfile.c  	swp_entry_t entry;
entry            1930 mm/swapfile.c  		entry = pte_to_swp_entry(*pte);
entry            1931 mm/swapfile.c  		if (swp_type(entry) != type)
entry            1934 mm/swapfile.c  		offset = swp_offset(entry);
entry            1943 mm/swapfile.c  		page = swapin_readahead(entry, GFP_HIGHUSER_MOVABLE, &vmf);
entry            1952 mm/swapfile.c  		ret = unuse_pte(vma, pmd, addr, entry, page);
entry            2132 mm/swapfile.c  	swp_entry_t entry;
entry            2184 mm/swapfile.c  		entry = swp_entry(type, i);
entry            2185 mm/swapfile.c  		page = find_get_page(swap_address_space(entry), i);
entry            2257 mm/swapfile.c  static sector_t map_swap_entry(swp_entry_t entry, struct block_device **bdev)
entry            2263 mm/swapfile.c  	sis = swp_swap_info(entry);
entry            2266 mm/swapfile.c  	offset = swp_offset(entry);
entry            2276 mm/swapfile.c  	swp_entry_t entry;
entry            2277 mm/swapfile.c  	entry.val = page_private(page);
entry            2278 mm/swapfile.c  	return map_swap_entry(entry, bdev);
entry            3377 mm/swapfile.c  static int __swap_duplicate(swp_entry_t entry, unsigned char usage)
entry            3386 mm/swapfile.c  	p = get_swap_device(entry);
entry            3390 mm/swapfile.c  	offset = swp_offset(entry);
entry            3445 mm/swapfile.c  void swap_shmem_alloc(swp_entry_t entry)
entry            3447 mm/swapfile.c  	__swap_duplicate(entry, SWAP_MAP_SHMEM);
entry            3457 mm/swapfile.c  int swap_duplicate(swp_entry_t entry)
entry            3461 mm/swapfile.c  	while (!err && __swap_duplicate(entry, 1) == -ENOMEM)
entry            3462 mm/swapfile.c  		err = add_swap_count_continuation(entry, GFP_ATOMIC);
entry            3474 mm/swapfile.c  int swapcache_prepare(swp_entry_t entry)
entry            3476 mm/swapfile.c  	return __swap_duplicate(entry, SWAP_HAS_CACHE);
entry            3479 mm/swapfile.c  struct swap_info_struct *swp_swap_info(swp_entry_t entry)
entry            3481 mm/swapfile.c  	return swap_type_to_swap_info(swp_type(entry));
entry            3486 mm/swapfile.c  	swp_entry_t entry = { .val = page_private(page) };
entry            3487 mm/swapfile.c  	return swp_swap_info(entry);
entry            3521 mm/swapfile.c  int add_swap_count_continuation(swp_entry_t entry, gfp_t gfp_mask)
entry            3538 mm/swapfile.c  	si = get_swap_device(entry);
entry            3548 mm/swapfile.c  	offset = swp_offset(entry);
entry              35 mm/truncate.c  				pgoff_t index, void *entry)
entry              40 mm/truncate.c  	if (xas_load(&xas) != entry)
entry              47 mm/truncate.c  			       void *entry)
entry              50 mm/truncate.c  	__clear_shadow_entry(mapping, index, entry);
entry             112 mm/truncate.c  					pgoff_t index, void *entry)
entry             117 mm/truncate.c  	clear_shadow_entry(mapping, index, entry);
entry             126 mm/truncate.c  					 pgoff_t index, void *entry)
entry             133 mm/truncate.c  	clear_shadow_entry(mapping, index, entry);
entry             677 mm/util.c      		swp_entry_t entry;
entry             679 mm/util.c      		entry.val = page_private(page);
entry             680 mm/util.c      		return swap_address_space(entry);
entry             199 mm/workingset.c 	unsigned long entry = xa_to_value(shadow);
entry             203 mm/workingset.c 	workingset = entry & 1;
entry             204 mm/workingset.c 	entry >>= 1;
entry             205 mm/workingset.c 	nid = entry & ((1UL << NODES_SHIFT) - 1);
entry             206 mm/workingset.c 	entry >>= NODES_SHIFT;
entry             207 mm/workingset.c 	memcgid = entry & ((1UL << MEM_CGROUP_ID_SHIFT) - 1);
entry             208 mm/workingset.c 	entry >>= MEM_CGROUP_ID_SHIFT;
entry             212 mm/workingset.c 	*evictionp = entry << bucket_order;
entry             250 mm/zswap.c     	struct zswap_entry *entry;
entry             251 mm/zswap.c     	entry = kmem_cache_alloc(zswap_entry_cache, gfp);
entry             252 mm/zswap.c     	if (!entry)
entry             254 mm/zswap.c     	entry->refcount = 1;
entry             255 mm/zswap.c     	RB_CLEAR_NODE(&entry->rbnode);
entry             256 mm/zswap.c     	return entry;
entry             259 mm/zswap.c     static void zswap_entry_cache_free(struct zswap_entry *entry)
entry             261 mm/zswap.c     	kmem_cache_free(zswap_entry_cache, entry);
entry             270 mm/zswap.c     	struct zswap_entry *entry;
entry             273 mm/zswap.c     		entry = rb_entry(node, struct zswap_entry, rbnode);
entry             274 mm/zswap.c     		if (entry->offset > offset)
entry             276 mm/zswap.c     		else if (entry->offset < offset)
entry             279 mm/zswap.c     			return entry;
entry             288 mm/zswap.c     static int zswap_rb_insert(struct rb_root *root, struct zswap_entry *entry,
entry             297 mm/zswap.c     		if (myentry->offset > entry->offset)
entry             299 mm/zswap.c     		else if (myentry->offset < entry->offset)
entry             306 mm/zswap.c     	rb_link_node(&entry->rbnode, parent, link);
entry             307 mm/zswap.c     	rb_insert_color(&entry->rbnode, root);
entry             311 mm/zswap.c     static void zswap_rb_erase(struct rb_root *root, struct zswap_entry *entry)
entry             313 mm/zswap.c     	if (!RB_EMPTY_NODE(&entry->rbnode)) {
entry             314 mm/zswap.c     		rb_erase(&entry->rbnode, root);
entry             315 mm/zswap.c     		RB_CLEAR_NODE(&entry->rbnode);
entry             323 mm/zswap.c     static void zswap_free_entry(struct zswap_entry *entry)
entry             325 mm/zswap.c     	if (!entry->length)
entry             328 mm/zswap.c     		zpool_free(entry->pool->zpool, entry->handle);
entry             329 mm/zswap.c     		zswap_pool_put(entry->pool);
entry             331 mm/zswap.c     	zswap_entry_cache_free(entry);
entry             337 mm/zswap.c     static void zswap_entry_get(struct zswap_entry *entry)
entry             339 mm/zswap.c     	entry->refcount++;
entry             346 mm/zswap.c     			struct zswap_entry *entry)
entry             348 mm/zswap.c     	int refcount = --entry->refcount;
entry             352 mm/zswap.c     		zswap_rb_erase(&tree->rbroot, entry);
entry             353 mm/zswap.c     		zswap_free_entry(entry);
entry             361 mm/zswap.c     	struct zswap_entry *entry;
entry             363 mm/zswap.c     	entry = zswap_rb_search(root, offset);
entry             364 mm/zswap.c     	if (entry)
entry             365 mm/zswap.c     		zswap_entry_get(entry);
entry             367 mm/zswap.c     	return entry;
entry             814 mm/zswap.c     static int zswap_get_swap_cache_page(swp_entry_t entry,
entry             819 mm/zswap.c     	*retpage = __read_swap_cache_async(entry, GFP_KERNEL,
entry             846 mm/zswap.c     	struct zswap_entry *entry;
entry             864 mm/zswap.c     	entry = zswap_entry_find_get(&tree->rbroot, offset);
entry             865 mm/zswap.c     	if (!entry) {
entry             872 mm/zswap.c     	BUG_ON(offset != entry->offset);
entry             891 mm/zswap.c     		tfm = *get_cpu_ptr(entry->pool->tfm);
entry             892 mm/zswap.c     		ret = crypto_comp_decompress(tfm, src, entry->length,
entry             894 mm/zswap.c     		put_cpu_ptr(entry->pool->tfm);
entry             913 mm/zswap.c     	zswap_entry_put(tree, entry);
entry             922 mm/zswap.c     	if (entry == zswap_rb_search(&tree->rbroot, offset))
entry             923 mm/zswap.c     		zswap_entry_put(tree, entry);
entry             937 mm/zswap.c     	zswap_entry_put(tree, entry);
entry             991 mm/zswap.c     	struct zswap_entry *entry, *dupentry;
entry            1032 mm/zswap.c     	entry = zswap_entry_cache_alloc(GFP_KERNEL);
entry            1033 mm/zswap.c     	if (!entry) {
entry            1043 mm/zswap.c     			entry->offset = offset;
entry            1044 mm/zswap.c     			entry->length = 0;
entry            1045 mm/zswap.c     			entry->value = value;
entry            1053 mm/zswap.c     	entry->pool = zswap_pool_current_get();
entry            1054 mm/zswap.c     	if (!entry->pool) {
entry            1061 mm/zswap.c     	tfm = *get_cpu_ptr(entry->pool->tfm);
entry            1065 mm/zswap.c     	put_cpu_ptr(entry->pool->tfm);
entry            1072 mm/zswap.c     	hlen = zpool_evictable(entry->pool->zpool) ? sizeof(zhdr) : 0;
entry            1074 mm/zswap.c     	if (zpool_malloc_support_movable(entry->pool->zpool))
entry            1076 mm/zswap.c     	ret = zpool_malloc(entry->pool->zpool, hlen + dlen, gfp, &handle);
entry            1085 mm/zswap.c     	buf = zpool_map_handle(entry->pool->zpool, handle, ZPOOL_MM_RW);
entry            1088 mm/zswap.c     	zpool_unmap_handle(entry->pool->zpool, handle);
entry            1092 mm/zswap.c     	entry->offset = offset;
entry            1093 mm/zswap.c     	entry->handle = handle;
entry            1094 mm/zswap.c     	entry->length = dlen;
entry            1100 mm/zswap.c     		ret = zswap_rb_insert(&tree->rbroot, entry, &dupentry);
entry            1118 mm/zswap.c     	zswap_pool_put(entry->pool);
entry            1120 mm/zswap.c     	zswap_entry_cache_free(entry);
entry            1133 mm/zswap.c     	struct zswap_entry *entry;
entry            1141 mm/zswap.c     	entry = zswap_entry_find_get(&tree->rbroot, offset);
entry            1142 mm/zswap.c     	if (!entry) {
entry            1149 mm/zswap.c     	if (!entry->length) {
entry            1151 mm/zswap.c     		zswap_fill_page(dst, entry->value);
entry            1158 mm/zswap.c     	src = zpool_map_handle(entry->pool->zpool, entry->handle, ZPOOL_MM_RO);
entry            1159 mm/zswap.c     	if (zpool_evictable(entry->pool->zpool))
entry            1162 mm/zswap.c     	tfm = *get_cpu_ptr(entry->pool->tfm);
entry            1163 mm/zswap.c     	ret = crypto_comp_decompress(tfm, src, entry->length, dst, &dlen);
entry            1164 mm/zswap.c     	put_cpu_ptr(entry->pool->tfm);
entry            1166 mm/zswap.c     	zpool_unmap_handle(entry->pool->zpool, entry->handle);
entry            1171 mm/zswap.c     	zswap_entry_put(tree, entry);
entry            1181 mm/zswap.c     	struct zswap_entry *entry;
entry            1185 mm/zswap.c     	entry = zswap_rb_search(&tree->rbroot, offset);
entry            1186 mm/zswap.c     	if (!entry) {
entry            1193 mm/zswap.c     	zswap_rb_erase(&tree->rbroot, entry);
entry            1196 mm/zswap.c     	zswap_entry_put(tree, entry);
entry            1205 mm/zswap.c     	struct zswap_entry *entry, *n;
entry            1212 mm/zswap.c     	rbtree_postorder_for_each_entry_safe(entry, n, &tree->rbroot, rbnode)
entry            1213 mm/zswap.c     		zswap_free_entry(entry);
entry             484 net/appletalk/aarp.c 	struct aarp_entry *entry;
entry             499 net/appletalk/aarp.c 	entry = aarp_alloc();
entry             501 net/appletalk/aarp.c 	if (!entry)
entry             504 net/appletalk/aarp.c 	entry->expires_at = -1;
entry             505 net/appletalk/aarp.c 	entry->status = ATIF_PROBE;
entry             506 net/appletalk/aarp.c 	entry->target_addr.s_node = sa->s_node;
entry             507 net/appletalk/aarp.c 	entry->target_addr.s_net = sa->s_net;
entry             508 net/appletalk/aarp.c 	entry->dev = atif->dev;
entry             513 net/appletalk/aarp.c 	entry->next = proxies[hash];
entry             514 net/appletalk/aarp.c 	proxies[hash] = entry;
entry             524 net/appletalk/aarp.c 		if (entry->status & ATIF_PROBE_FAIL)
entry             528 net/appletalk/aarp.c 	if (entry->status & ATIF_PROBE_FAIL) {
entry             529 net/appletalk/aarp.c 		entry->expires_at = jiffies - 1; /* free the entry */
entry             532 net/appletalk/aarp.c 		entry->status &= ~ATIF_PROBE;
entry             924 net/appletalk/aarp.c 	struct aarp_entry *entry;
entry             928 net/appletalk/aarp.c 		for (entry = table[ct]; entry; entry = entry->next) {
entry             932 net/appletalk/aarp.c 				return entry;
entry             965 net/appletalk/aarp.c 	struct aarp_entry *entry = v;
entry             972 net/appletalk/aarp.c 		entry = iter_next(iter, NULL);
entry             975 net/appletalk/aarp.c 	else if (entry->next)
entry             976 net/appletalk/aarp.c 		entry = entry->next;
entry             981 net/appletalk/aarp.c 		entry = iter_next(iter, NULL);
entry             983 net/appletalk/aarp.c 	return entry;
entry            1004 net/appletalk/aarp.c 	struct aarp_entry *entry = v;
entry            1013 net/appletalk/aarp.c 			   ntohs(entry->target_addr.s_net),
entry            1014 net/appletalk/aarp.c 			   (unsigned int) entry->target_addr.s_node,
entry            1015 net/appletalk/aarp.c 			   entry->dev ? entry->dev->name : "????");
entry            1016 net/appletalk/aarp.c 		seq_printf(seq, "%pM", entry->hwaddr);
entry            1018 net/appletalk/aarp.c 			   dt2str((long)entry->expires_at - (long)now));
entry            1021 net/appletalk/aarp.c 				   dt2str(now - entry->last_sent),
entry            1022 net/appletalk/aarp.c 				   entry->xmit_count);
entry              59 net/atm/addr.c 	list_for_each_entry_safe(this, p, head, entry) {
entry              60 net/atm/addr.c 		list_del(&this->entry);
entry              84 net/atm/addr.c 	list_for_each_entry(this, head, entry) {
entry              96 net/atm/addr.c 	list_add(&this->entry, head);
entry             119 net/atm/addr.c 	list_for_each_entry(this, head, entry) {
entry             121 net/atm/addr.c 			list_del(&this->entry);
entry             147 net/atm/addr.c 	list_for_each_entry(this, head, entry)
entry             154 net/atm/addr.c 	list_for_each_entry(this, head, entry)
entry              40 net/atm/atm_sysfs.c 	list_for_each_entry(aaddr, &adev->local, entry) {
entry              76 net/atm/clip.c static void link_vcc(struct clip_vcc *clip_vcc, struct atmarp_entry *entry)
entry              78 net/atm/clip.c 	pr_debug("%p to entry %p (neigh %p)\n", clip_vcc, entry, entry->neigh);
entry              79 net/atm/clip.c 	clip_vcc->entry = entry;
entry              81 net/atm/clip.c 	clip_vcc->next = entry->vccs;
entry              82 net/atm/clip.c 	entry->vccs = clip_vcc;
entry              83 net/atm/clip.c 	entry->neigh->used = jiffies;
entry              88 net/atm/clip.c 	struct atmarp_entry *entry = clip_vcc->entry;
entry              91 net/atm/clip.c 	if (!entry) {
entry              95 net/atm/clip.c 	netif_tx_lock_bh(entry->neigh->dev);	/* block clip_start_xmit() */
entry              96 net/atm/clip.c 	entry->neigh->used = jiffies;
entry              97 net/atm/clip.c 	for (walk = &entry->vccs; *walk; walk = &(*walk)->next)
entry             102 net/atm/clip.c 			clip_vcc->entry = NULL;
entry             104 net/atm/clip.c 				netif_wake_queue(entry->neigh->dev);
entry             105 net/atm/clip.c 			if (entry->vccs)
entry             107 net/atm/clip.c 			entry->expires = jiffies - 1;
entry             109 net/atm/clip.c 			error = neigh_update(entry->neigh, NULL, NUD_NONE,
entry             115 net/atm/clip.c 	pr_err("ATMARP: failed (entry %p, vcc 0x%p)\n", entry, clip_vcc);
entry             117 net/atm/clip.c 	netif_tx_unlock_bh(entry->neigh->dev);
entry             123 net/atm/clip.c 	struct atmarp_entry *entry = neighbour_priv(n);
entry             128 net/atm/clip.c 	for (cv = entry->vccs; cv; cv = cv->next) {
entry             133 net/atm/clip.c 				 cv, cv->vcc, entry);
entry             138 net/atm/clip.c 	if (entry->vccs || time_before(jiffies, entry->expires))
entry             204 net/atm/clip.c 		if (clip_vcc->entry)
entry             211 net/atm/clip.c 	skb->dev = clip_vcc->entry ? clip_vcc->entry->neigh->dev : clip_devs;
entry             292 net/atm/clip.c 	struct atmarp_entry *entry = neighbour_priv(neigh);
entry             303 net/atm/clip.c 	entry->neigh = neigh;
entry             304 net/atm/clip.c 	entry->vccs = NULL;
entry             305 net/atm/clip.c 	entry->expires = jiffies - 1;
entry             333 net/atm/clip.c 	struct atmarp_entry *entry;
entry             360 net/atm/clip.c 	entry = neighbour_priv(n);
entry             361 net/atm/clip.c 	if (!entry->vccs) {
entry             362 net/atm/clip.c 		if (time_after(jiffies, entry->expires)) {
entry             364 net/atm/clip.c 			entry->expires = jiffies + ATMARP_RETRY_DELAY * HZ;
entry             367 net/atm/clip.c 		if (entry->neigh->arp_queue.qlen < ATMARP_MAX_UNRES_PACKETS)
entry             368 net/atm/clip.c 			skb_queue_tail(&entry->neigh->arp_queue, skb);
entry             375 net/atm/clip.c 	pr_debug("neigh %p, vccs %p\n", entry, entry->vccs);
entry             376 net/atm/clip.c 	ATM_SKB(skb)->vcc = vcc = entry->vccs->vcc;
entry             378 net/atm/clip.c 	if (entry->vccs->encap) {
entry             386 net/atm/clip.c 	entry->vccs->last_use = jiffies;
entry             388 net/atm/clip.c 	old = xchg(&entry->vccs->xoff, 1);	/* assume XOFF ... */
entry             397 net/atm/clip.c 		entry->vccs->xoff = 0;
entry             403 net/atm/clip.c 	if (!entry->vccs->xoff)
entry             428 net/atm/clip.c 	clip_vcc->entry = NULL;
entry             447 net/atm/clip.c 	struct atmarp_entry *entry;
entry             458 net/atm/clip.c 		if (!clip_vcc->entry) {
entry             473 net/atm/clip.c 	entry = neighbour_priv(neigh);
entry             474 net/atm/clip.c 	if (entry != clip_vcc->entry) {
entry             475 net/atm/clip.c 		if (!clip_vcc->entry)
entry             481 net/atm/clip.c 		link_vcc(clip_vcc, entry);
entry             737 net/atm/clip.c 			struct atmarp_entry *entry, struct clip_vcc *clip_vcc)
entry             750 net/atm/clip.c 		exp = entry->neigh->used;
entry             766 net/atm/clip.c 		if (time_before(jiffies, entry->expires))
entry             770 net/atm/clip.c 				   refcount_read(&entry->neigh->refcnt));
entry             106 net/atm/lec.c  static inline void lec_arp_hold(struct lec_arp_table *entry)
entry             108 net/atm/lec.c  	refcount_inc(&entry->usage);
entry             111 net/atm/lec.c  static inline void lec_arp_put(struct lec_arp_table *entry)
entry             113 net/atm/lec.c  	if (refcount_dec_and_test(&entry->usage))
entry             114 net/atm/lec.c  		kfree(entry);
entry             211 net/atm/lec.c  	struct lec_arp_table *entry;
entry             282 net/atm/lec.c  	entry = NULL;
entry             283 net/atm/lec.c  	vcc = lec_arp_resolve(priv, dst, is_rdesc, &entry);
entry             285 net/atm/lec.c  		 dev->name, vcc, vcc ? vcc->flags : 0, entry);
entry             287 net/atm/lec.c  		if (entry && (entry->tx_wait.qlen < LEC_UNRES_QUE_LEN)) {
entry             290 net/atm/lec.c  			skb_queue_tail(&entry->tx_wait, skb);
entry             304 net/atm/lec.c  	while (entry && (skb2 = skb_dequeue(&entry->tx_wait))) {
entry             328 net/atm/lec.c  	if (entry)
entry             329 net/atm/lec.c  		lec_arp_put(entry);
entry             347 net/atm/lec.c  	struct lec_arp_table *entry;
entry             377 net/atm/lec.c  		entry = lec_arp_find(priv, mesg->content.normal.mac_addr);
entry             378 net/atm/lec.c  		lec_arp_remove(priv, entry);
entry             616 net/atm/lec.c  		struct lec_arp_table *entry;
entry             639 net/atm/lec.c  			entry = lec_arp_find(priv, src);
entry             640 net/atm/lec.c  			if (entry && entry->vcc != vcc) {
entry             641 net/atm/lec.c  				lec_arp_remove(priv, entry);
entry             642 net/atm/lec.c  				lec_arp_put(entry);
entry             800 net/atm/lec.c  static void lec_info(struct seq_file *seq, struct lec_arp_table *entry)
entry             805 net/atm/lec.c  		seq_printf(seq, "%2.2x", entry->mac_addr[i] & 0xff);
entry             808 net/atm/lec.c  		seq_printf(seq, "%2.2x", entry->atm_addr[i] & 0xff);
entry             809 net/atm/lec.c  	seq_printf(seq, " %s %4.4x", lec_arp_get_status_string(entry->status),
entry             810 net/atm/lec.c  		   entry->flags & 0xffff);
entry             811 net/atm/lec.c  	if (entry->vcc)
entry             812 net/atm/lec.c  		seq_printf(seq, "%3d %3d ", entry->vcc->vpi, entry->vcc->vci);
entry             815 net/atm/lec.c  	if (entry->recv_vcc) {
entry             816 net/atm/lec.c  		seq_printf(seq, "     %3d %3d", entry->recv_vcc->vpi,
entry             817 net/atm/lec.c  			   entry->recv_vcc->vci);
entry             979 net/atm/lec.c  		struct lec_arp_table *entry = hlist_entry(state->node,
entry             984 net/atm/lec.c  		lec_info(seq, entry);
entry            1177 net/atm/lec.c  	struct lec_arp_table *entry = lec_arp_find(priv, mac_addr);
entry            1179 net/atm/lec.c  	if (entry == NULL)
entry            1182 net/atm/lec.c  	kfree(entry->tlvs);
entry            1184 net/atm/lec.c  	entry->tlvs = kmemdup(tlvs, sizeoftlvs, GFP_KERNEL);
entry            1185 net/atm/lec.c  	if (entry->tlvs == NULL)
entry            1187 net/atm/lec.c  	entry->sizeoftlvs = sizeoftlvs;
entry            1255 net/atm/lec.c  static void lec_arp_clear_vccs(struct lec_arp_table *entry)
entry            1257 net/atm/lec.c  	if (entry->vcc) {
entry            1258 net/atm/lec.c  		struct atm_vcc *vcc = entry->vcc;
entry            1267 net/atm/lec.c  		vcc->push = entry->old_push;
entry            1269 net/atm/lec.c  		entry->vcc = NULL;
entry            1271 net/atm/lec.c  	if (entry->recv_vcc) {
entry            1272 net/atm/lec.c  		entry->recv_vcc->push = entry->old_recv_push;
entry            1273 net/atm/lec.c  		vcc_release_async(entry->recv_vcc, -EPIPE);
entry            1274 net/atm/lec.c  		entry->recv_vcc = NULL;
entry            1283 net/atm/lec.c  lec_arp_add(struct lec_priv *priv, struct lec_arp_table *entry)
entry            1287 net/atm/lec.c  	tmp = &priv->lec_arp_tables[HASH(entry->mac_addr[ETH_ALEN - 1])];
entry            1288 net/atm/lec.c  	hlist_add_head(&entry->next, tmp);
entry            1290 net/atm/lec.c  	pr_debug("Added entry:%pM\n", entry->mac_addr);
entry            1299 net/atm/lec.c  	struct lec_arp_table *entry;
entry            1317 net/atm/lec.c  			hlist_for_each_entry(entry,
entry            1320 net/atm/lec.c  					   entry->atm_addr, ATM_ESA_LEN) == 0) {
entry            1478 net/atm/lec.c  	struct lec_arp_table *entry;
entry            1489 net/atm/lec.c  		hlist_for_each_entry_safe(entry, next,
entry            1491 net/atm/lec.c  			lec_arp_remove(priv, entry);
entry            1492 net/atm/lec.c  			lec_arp_put(entry);
entry            1497 net/atm/lec.c  	hlist_for_each_entry_safe(entry, next,
entry            1499 net/atm/lec.c  		del_timer_sync(&entry->timer);
entry            1500 net/atm/lec.c  		lec_arp_clear_vccs(entry);
entry            1501 net/atm/lec.c  		hlist_del(&entry->next);
entry            1502 net/atm/lec.c  		lec_arp_put(entry);
entry            1506 net/atm/lec.c  	hlist_for_each_entry_safe(entry, next,
entry            1508 net/atm/lec.c  		del_timer_sync(&entry->timer);
entry            1509 net/atm/lec.c  		lec_arp_clear_vccs(entry);
entry            1510 net/atm/lec.c  		hlist_del(&entry->next);
entry            1511 net/atm/lec.c  		lec_arp_put(entry);
entry            1515 net/atm/lec.c  	hlist_for_each_entry_safe(entry, next, &priv->mcast_fwds, next) {
entry            1517 net/atm/lec.c  		lec_arp_clear_vccs(entry);
entry            1518 net/atm/lec.c  		hlist_del(&entry->next);
entry            1519 net/atm/lec.c  		lec_arp_put(entry);
entry            1533 net/atm/lec.c  	struct lec_arp_table *entry;
entry            1538 net/atm/lec.c  	hlist_for_each_entry(entry, head, next) {
entry            1539 net/atm/lec.c  		if (ether_addr_equal(mac_addr, entry->mac_addr))
entry            1540 net/atm/lec.c  			return entry;
entry            1568 net/atm/lec.c  	struct lec_arp_table *entry;
entry            1570 net/atm/lec.c  	entry = from_timer(entry, t, timer);
entry            1573 net/atm/lec.c  	if (entry->status == ESI_ARP_PENDING) {
entry            1574 net/atm/lec.c  		if (entry->no_tries <= entry->priv->max_retry_count) {
entry            1575 net/atm/lec.c  			if (entry->is_rdesc)
entry            1576 net/atm/lec.c  				send_to_lecd(entry->priv, l_rdesc_arp_xmt,
entry            1577 net/atm/lec.c  					     entry->mac_addr, NULL, NULL);
entry            1579 net/atm/lec.c  				send_to_lecd(entry->priv, l_arp_xmt,
entry            1580 net/atm/lec.c  					     entry->mac_addr, NULL, NULL);
entry            1581 net/atm/lec.c  			entry->no_tries++;
entry            1583 net/atm/lec.c  		mod_timer(&entry->timer, jiffies + (1 * HZ));
entry            1609 net/atm/lec.c  static bool __lec_arp_check_expire(struct lec_arp_table *entry,
entry            1615 net/atm/lec.c  	if ((entry->flags) & LEC_REMOTE_FLAG && priv->topology_change)
entry            1621 net/atm/lec.c  		 now, entry->last_used, time_to_check);
entry            1622 net/atm/lec.c  	if (time_after(now, entry->last_used + time_to_check) &&
entry            1623 net/atm/lec.c  	    !(entry->flags & LEC_PERMANENT_FLAG) &&
entry            1624 net/atm/lec.c  	    !(entry->mac_addr[0] & 0x01)) {	/* LANE2: 7.1.20 */
entry            1627 net/atm/lec.c  		lec_arp_remove(priv, entry);
entry            1628 net/atm/lec.c  		lec_arp_put(entry);
entry            1631 net/atm/lec.c  		if ((entry->status == ESI_VC_PENDING ||
entry            1632 net/atm/lec.c  		     entry->status == ESI_ARP_PENDING) &&
entry            1633 net/atm/lec.c  		    time_after_eq(now, entry->timestamp +
entry            1635 net/atm/lec.c  			entry->timestamp = jiffies;
entry            1636 net/atm/lec.c  			entry->packets_flooded = 0;
entry            1637 net/atm/lec.c  			if (entry->status == ESI_VC_PENDING)
entry            1639 net/atm/lec.c  					     entry->mac_addr,
entry            1640 net/atm/lec.c  					     entry->atm_addr,
entry            1643 net/atm/lec.c  		if (entry->status == ESI_FLUSH_PENDING &&
entry            1644 net/atm/lec.c  		    time_after_eq(now, entry->timestamp +
entry            1646 net/atm/lec.c  			lec_arp_hold(entry);
entry            1675 net/atm/lec.c  	struct lec_arp_table *entry;
entry            1684 net/atm/lec.c  		hlist_for_each_entry_safe(entry, next,
entry            1686 net/atm/lec.c  			if (__lec_arp_check_expire(entry, now, priv)) {
entry            1688 net/atm/lec.c  				struct atm_vcc *vcc = entry->vcc;
entry            1692 net/atm/lec.c  				while ((skb = skb_dequeue(&entry->tx_wait)))
entry            1694 net/atm/lec.c  				entry->last_used = jiffies;
entry            1695 net/atm/lec.c  				entry->status = ESI_FORWARD_DIRECT;
entry            1696 net/atm/lec.c  				lec_arp_put(entry);
entry            1717 net/atm/lec.c  	struct lec_arp_table *entry;
entry            1734 net/atm/lec.c  	entry = lec_arp_find(priv, mac_to_find);
entry            1736 net/atm/lec.c  	if (entry) {
entry            1737 net/atm/lec.c  		if (entry->status == ESI_FORWARD_DIRECT) {
entry            1739 net/atm/lec.c  			entry->last_used = jiffies;
entry            1740 net/atm/lec.c  			lec_arp_hold(entry);
entry            1741 net/atm/lec.c  			*ret_entry = entry;
entry            1742 net/atm/lec.c  			found = entry->vcc;
entry            1749 net/atm/lec.c  		if (entry->status == ESI_ARP_PENDING)
entry            1750 net/atm/lec.c  			entry->no_tries = 0;
entry            1757 net/atm/lec.c  		if (entry->status != ESI_FLUSH_PENDING &&
entry            1758 net/atm/lec.c  		    entry->packets_flooded <
entry            1760 net/atm/lec.c  			entry->packets_flooded++;
entry            1770 net/atm/lec.c  		lec_arp_hold(entry);
entry            1771 net/atm/lec.c  		*ret_entry = entry;
entry            1772 net/atm/lec.c  		pr_debug("entry->status %d entry->vcc %p\n", entry->status,
entry            1773 net/atm/lec.c  			 entry->vcc);
entry            1777 net/atm/lec.c  		entry = make_entry(priv, mac_to_find);
entry            1779 net/atm/lec.c  		if (!entry) {
entry            1783 net/atm/lec.c  		lec_arp_add(priv, entry);
entry            1785 net/atm/lec.c  		entry->packets_flooded = 1;
entry            1786 net/atm/lec.c  		entry->status = ESI_ARP_PENDING;
entry            1787 net/atm/lec.c  		entry->no_tries = 1;
entry            1788 net/atm/lec.c  		entry->last_used = entry->timestamp = jiffies;
entry            1789 net/atm/lec.c  		entry->is_rdesc = is_rdesc;
entry            1790 net/atm/lec.c  		if (entry->is_rdesc)
entry            1795 net/atm/lec.c  		entry->timer.expires = jiffies + (1 * HZ);
entry            1796 net/atm/lec.c  		entry->timer.function = lec_arp_expire_arp;
entry            1797 net/atm/lec.c  		add_timer(&entry->timer);
entry            1812 net/atm/lec.c  	struct lec_arp_table *entry;
entry            1818 net/atm/lec.c  		hlist_for_each_entry_safe(entry, next,
entry            1820 net/atm/lec.c  			if (!memcmp(atm_addr, entry->atm_addr, ATM_ESA_LEN) &&
entry            1822 net/atm/lec.c  			     !(entry->flags & LEC_PERMANENT_FLAG))) {
entry            1823 net/atm/lec.c  				lec_arp_remove(priv, entry);
entry            1824 net/atm/lec.c  				lec_arp_put(entry);
entry            1844 net/atm/lec.c  	struct lec_arp_table *entry, *tmp;
entry            1851 net/atm/lec.c  	entry = lec_arp_find(priv, mac_addr);
entry            1852 net/atm/lec.c  	if (entry == NULL && targetless_le_arp)
entry            1858 net/atm/lec.c  		hlist_for_each_entry_safe(entry, next,
entry            1860 net/atm/lec.c  			if (memcmp(entry->atm_addr, atm_addr, ATM_ESA_LEN) == 0) {
entry            1861 net/atm/lec.c  				hlist_del(&entry->next);
entry            1862 net/atm/lec.c  				del_timer(&entry->timer);
entry            1868 net/atm/lec.c  					tmp->vcc = entry->vcc;
entry            1869 net/atm/lec.c  					tmp->old_push = entry->old_push;
entry            1871 net/atm/lec.c  					del_timer(&entry->timer);
entry            1872 net/atm/lec.c  					lec_arp_put(entry);
entry            1873 net/atm/lec.c  					entry = tmp;
entry            1875 net/atm/lec.c  					entry->status = ESI_FORWARD_DIRECT;
entry            1876 net/atm/lec.c  					ether_addr_copy(entry->mac_addr,
entry            1878 net/atm/lec.c  					entry->last_used = jiffies;
entry            1879 net/atm/lec.c  					lec_arp_add(priv, entry);
entry            1882 net/atm/lec.c  					entry->flags |= LEC_REMOTE_FLAG;
entry            1884 net/atm/lec.c  					entry->flags &= ~LEC_REMOTE_FLAG;
entry            1892 net/atm/lec.c  	entry = lec_arp_find(priv, mac_addr);
entry            1893 net/atm/lec.c  	if (!entry) {
entry            1894 net/atm/lec.c  		entry = make_entry(priv, mac_addr);
entry            1895 net/atm/lec.c  		if (!entry)
entry            1897 net/atm/lec.c  		entry->status = ESI_UNKNOWN;
entry            1898 net/atm/lec.c  		lec_arp_add(priv, entry);
entry            1901 net/atm/lec.c  	memcpy(entry->atm_addr, atm_addr, ATM_ESA_LEN);
entry            1902 net/atm/lec.c  	del_timer(&entry->timer);
entry            1906 net/atm/lec.c  			if (entry != tmp &&
entry            1914 net/atm/lec.c  					entry->vcc = tmp->vcc;
entry            1915 net/atm/lec.c  					entry->old_push = tmp->old_push;
entry            1917 net/atm/lec.c  				entry->status = tmp->status;
entry            1923 net/atm/lec.c  		entry->flags |= LEC_REMOTE_FLAG;
entry            1925 net/atm/lec.c  		entry->flags &= ~LEC_REMOTE_FLAG;
entry            1926 net/atm/lec.c  	if (entry->status == ESI_ARP_PENDING || entry->status == ESI_UNKNOWN) {
entry            1927 net/atm/lec.c  		entry->status = ESI_VC_PENDING;
entry            1928 net/atm/lec.c  		send_to_lecd(priv, l_svc_setup, entry->mac_addr, atm_addr, NULL);
entry            1945 net/atm/lec.c  	struct lec_arp_table *entry;
entry            1953 net/atm/lec.c  		entry = lec_arp_find(priv, bus_mac);
entry            1954 net/atm/lec.c  		if (!entry) {
entry            1958 net/atm/lec.c  		memcpy(entry->atm_addr, ioc_data->atm_addr, ATM_ESA_LEN);
entry            1959 net/atm/lec.c  		entry->recv_vcc = vcc;
entry            1960 net/atm/lec.c  		entry->old_recv_push = old_push;
entry            1962 net/atm/lec.c  		entry = make_entry(priv, bus_mac);
entry            1963 net/atm/lec.c  		if (entry == NULL)
entry            1965 net/atm/lec.c  		del_timer(&entry->timer);
entry            1966 net/atm/lec.c  		memcpy(entry->atm_addr, ioc_data->atm_addr, ATM_ESA_LEN);
entry            1967 net/atm/lec.c  		entry->recv_vcc = vcc;
entry            1968 net/atm/lec.c  		entry->old_recv_push = old_push;
entry            1969 net/atm/lec.c  		hlist_add_head(&entry->next, &priv->mcast_fwds);
entry            1987 net/atm/lec.c  		entry = make_entry(priv, bus_mac);
entry            1988 net/atm/lec.c  		if (entry == NULL)
entry            1990 net/atm/lec.c  		memcpy(entry->atm_addr, ioc_data->atm_addr, ATM_ESA_LEN);
entry            1991 net/atm/lec.c  		eth_zero_addr(entry->mac_addr);
entry            1992 net/atm/lec.c  		entry->recv_vcc = vcc;
entry            1993 net/atm/lec.c  		entry->old_recv_push = old_push;
entry            1994 net/atm/lec.c  		entry->status = ESI_UNKNOWN;
entry            1995 net/atm/lec.c  		entry->timer.expires = jiffies + priv->vcc_timeout_period;
entry            1996 net/atm/lec.c  		entry->timer.function = lec_arp_expire_vcc;
entry            1997 net/atm/lec.c  		hlist_add_head(&entry->next, &priv->lec_no_forward);
entry            1998 net/atm/lec.c  		add_timer(&entry->timer);
entry            2014 net/atm/lec.c  		hlist_for_each_entry(entry,
entry            2017 net/atm/lec.c  			    (ioc_data->atm_addr, entry->atm_addr,
entry            2021 net/atm/lec.c  					 entry->vcc ? entry->vcc->vci : 0,
entry            2022 net/atm/lec.c  					 entry->recv_vcc ? entry->recv_vcc->
entry            2025 net/atm/lec.c  				del_timer(&entry->timer);
entry            2026 net/atm/lec.c  				entry->vcc = vcc;
entry            2027 net/atm/lec.c  				entry->old_push = old_push;
entry            2028 net/atm/lec.c  				if (entry->status == ESI_VC_PENDING) {
entry            2031 net/atm/lec.c  						entry->status =
entry            2034 net/atm/lec.c  						entry->timestamp = jiffies;
entry            2035 net/atm/lec.c  						entry->status =
entry            2040 net/atm/lec.c  							     entry->atm_addr,
entry            2070 net/atm/lec.c  	entry = make_entry(priv, bus_mac);
entry            2071 net/atm/lec.c  	if (!entry)
entry            2073 net/atm/lec.c  	entry->vcc = vcc;
entry            2074 net/atm/lec.c  	entry->old_push = old_push;
entry            2075 net/atm/lec.c  	memcpy(entry->atm_addr, ioc_data->atm_addr, ATM_ESA_LEN);
entry            2076 net/atm/lec.c  	eth_zero_addr(entry->mac_addr);
entry            2077 net/atm/lec.c  	entry->status = ESI_UNKNOWN;
entry            2078 net/atm/lec.c  	hlist_add_head(&entry->next, &priv->lec_arp_empty_ones);
entry            2079 net/atm/lec.c  	entry->timer.expires = jiffies + priv->vcc_timeout_period;
entry            2080 net/atm/lec.c  	entry->timer.function = lec_arp_expire_vcc;
entry            2081 net/atm/lec.c  	add_timer(&entry->timer);
entry            2091 net/atm/lec.c  	struct lec_arp_table *entry;
entry            2098 net/atm/lec.c  		hlist_for_each_entry(entry,
entry            2100 net/atm/lec.c  			if (entry->flush_tran_id == tran_id &&
entry            2101 net/atm/lec.c  			    entry->status == ESI_FLUSH_PENDING) {
entry            2103 net/atm/lec.c  				struct atm_vcc *vcc = entry->vcc;
entry            2105 net/atm/lec.c  				lec_arp_hold(entry);
entry            2108 net/atm/lec.c  				while ((skb = skb_dequeue(&entry->tx_wait)))
entry            2110 net/atm/lec.c  				entry->last_used = jiffies;
entry            2111 net/atm/lec.c  				entry->status = ESI_FORWARD_DIRECT;
entry            2112 net/atm/lec.c  				lec_arp_put(entry);
entry            2127 net/atm/lec.c  	struct lec_arp_table *entry;
entry            2132 net/atm/lec.c  		hlist_for_each_entry(entry,
entry            2134 net/atm/lec.c  			if (!memcmp(atm_addr, entry->atm_addr, ATM_ESA_LEN)) {
entry            2135 net/atm/lec.c  				entry->flush_tran_id = tran_id;
entry            2137 net/atm/lec.c  					 tran_id, entry);
entry            2185 net/atm/lec.c  	struct lec_arp_table *entry;
entry            2194 net/atm/lec.c  		hlist_for_each_entry_safe(entry, next,
entry            2196 net/atm/lec.c  			if (vcc == entry->vcc) {
entry            2197 net/atm/lec.c  				lec_arp_remove(priv, entry);
entry            2198 net/atm/lec.c  				lec_arp_put(entry);
entry            2205 net/atm/lec.c  	hlist_for_each_entry_safe(entry, next,
entry            2207 net/atm/lec.c  		if (entry->vcc == vcc) {
entry            2208 net/atm/lec.c  			lec_arp_clear_vccs(entry);
entry            2209 net/atm/lec.c  			del_timer(&entry->timer);
entry            2210 net/atm/lec.c  			hlist_del(&entry->next);
entry            2211 net/atm/lec.c  			lec_arp_put(entry);
entry            2215 net/atm/lec.c  	hlist_for_each_entry_safe(entry, next,
entry            2217 net/atm/lec.c  		if (entry->recv_vcc == vcc) {
entry            2218 net/atm/lec.c  			lec_arp_clear_vccs(entry);
entry            2219 net/atm/lec.c  			del_timer(&entry->timer);
entry            2220 net/atm/lec.c  			hlist_del(&entry->next);
entry            2221 net/atm/lec.c  			lec_arp_put(entry);
entry            2225 net/atm/lec.c  	hlist_for_each_entry_safe(entry, next, &priv->mcast_fwds, next) {
entry            2226 net/atm/lec.c  		if (entry->recv_vcc == vcc) {
entry            2227 net/atm/lec.c  			lec_arp_clear_vccs(entry);
entry            2229 net/atm/lec.c  			hlist_del(&entry->next);
entry            2230 net/atm/lec.c  			lec_arp_put(entry);
entry            2244 net/atm/lec.c  	struct lec_arp_table *entry, *tmp;
entry            2249 net/atm/lec.c  	hlist_for_each_entry_safe(entry, next,
entry            2251 net/atm/lec.c  		if (vcc == entry->vcc) {
entry            2252 net/atm/lec.c  			del_timer(&entry->timer);
entry            2253 net/atm/lec.c  			ether_addr_copy(entry->mac_addr, src);
entry            2254 net/atm/lec.c  			entry->status = ESI_FORWARD_DIRECT;
entry            2255 net/atm/lec.c  			entry->last_used = jiffies;
entry            2262 net/atm/lec.c  			hlist_del(&entry->next);
entry            2263 net/atm/lec.c  			lec_arp_add(priv, entry);
entry              87 net/atm/mpc.c  static void purge_egress_shortcut(struct atm_vcc *vcc, eg_cache_entry *entry);
entry             179 net/atm/mpc.c  	struct atm_mpoa_qos *entry;
entry             181 net/atm/mpc.c  	entry = atm_mpoa_search_qos(dst_ip);
entry             182 net/atm/mpc.c  	if (entry != NULL) {
entry             183 net/atm/mpc.c  		entry->qos = *qos;
entry             184 net/atm/mpc.c  		return entry;
entry             187 net/atm/mpc.c  	entry = kmalloc(sizeof(struct atm_mpoa_qos), GFP_KERNEL);
entry             188 net/atm/mpc.c  	if (entry == NULL) {
entry             190 net/atm/mpc.c  		return entry;
entry             193 net/atm/mpc.c  	entry->ipaddr = dst_ip;
entry             194 net/atm/mpc.c  	entry->qos = *qos;
entry             196 net/atm/mpc.c  	entry->next = qos_head;
entry             197 net/atm/mpc.c  	qos_head = entry;
entry             199 net/atm/mpc.c  	return entry;
entry             219 net/atm/mpc.c  int atm_mpoa_delete_qos(struct atm_mpoa_qos *entry)
entry             223 net/atm/mpc.c  	if (entry == NULL)
entry             225 net/atm/mpc.c  	if (entry == qos_head) {
entry             227 net/atm/mpc.c  		kfree(entry);
entry             233 net/atm/mpc.c  		if (curr->next == entry) {
entry             234 net/atm/mpc.c  			curr->next = entry->next;
entry             235 net/atm/mpc.c  			kfree(entry);
entry             494 net/atm/mpc.c  	in_cache_entry *entry;
entry             514 net/atm/mpc.c  	entry = mpc->in_ops->get(ipaddr, mpc);
entry             515 net/atm/mpc.c  	if (entry == NULL) {
entry             516 net/atm/mpc.c  		entry = mpc->in_ops->add_entry(ipaddr, mpc);
entry             517 net/atm/mpc.c  		if (entry != NULL)
entry             518 net/atm/mpc.c  			mpc->in_ops->put(entry);
entry             522 net/atm/mpc.c  	if (mpc->in_ops->cache_hit(entry, mpc) != OPEN) {
entry             525 net/atm/mpc.c  		mpc->in_ops->put(entry);
entry             535 net/atm/mpc.c  		mpc->in_ops->put(entry);
entry             542 net/atm/mpc.c  	if (entry->ctrl_info.tag != 0) {
entry             544 net/atm/mpc.c  			 mpc->dev->name, entry->ctrl_info.tag);
entry             545 net/atm/mpc.c  		tagged_llc_snap_hdr.tag = entry->ctrl_info.tag;
entry             559 net/atm/mpc.c  	atm_account_tx(entry->shortcut, skb);
entry             560 net/atm/mpc.c  	entry->shortcut->send(entry->shortcut, skb);
entry             561 net/atm/mpc.c  	entry->packets_fwded++;
entry             562 net/atm/mpc.c  	mpc->in_ops->put(entry);
entry            1083 net/atm/mpc.c  	in_cache_entry *entry;
entry            1085 net/atm/mpc.c  	entry = mpc->in_ops->get(dst_ip, mpc);
entry            1086 net/atm/mpc.c  	if (entry == NULL) {
entry            1087 net/atm/mpc.c  		entry = mpc->in_ops->add_entry(dst_ip, mpc);
entry            1088 net/atm/mpc.c  		entry->entry_state = INGRESS_RESOLVING;
entry            1090 net/atm/mpc.c  		msg->content.in_info = entry->ctrl_info;
entry            1092 net/atm/mpc.c  		entry->reply_wait = ktime_get_seconds();
entry            1093 net/atm/mpc.c  		mpc->in_ops->put(entry);
entry            1097 net/atm/mpc.c  	if (entry->entry_state == INGRESS_INVALID) {
entry            1098 net/atm/mpc.c  		entry->entry_state = INGRESS_RESOLVING;
entry            1100 net/atm/mpc.c  		msg->content.in_info = entry->ctrl_info;
entry            1102 net/atm/mpc.c  		entry->reply_wait = ktime_get_seconds();
entry            1103 net/atm/mpc.c  		mpc->in_ops->put(entry);
entry            1109 net/atm/mpc.c  	mpc->in_ops->put(entry);
entry            1118 net/atm/mpc.c  					in_cache_entry *entry)
entry            1129 net/atm/mpc.c  				entry->shortcut = eg_entry->shortcut;
entry            1131 net/atm/mpc.c  				entry->shortcut = eg_entry->shortcut;
entry            1133 net/atm/mpc.c  		if (entry->shortcut) {
entry            1158 net/atm/mpc.c  	in_cache_entry *entry = mpc->in_ops->get(dst_ip, mpc);
entry            1163 net/atm/mpc.c  		 mpc->dev->name, entry);
entry            1164 net/atm/mpc.c  	if (entry == NULL) {
entry            1169 net/atm/mpc.c  	ddprintk_cont(" entry_state = %d ", entry->entry_state);
entry            1171 net/atm/mpc.c  	if (entry->entry_state == INGRESS_RESOLVED) {
entry            1173 net/atm/mpc.c  		mpc->in_ops->put(entry);
entry            1177 net/atm/mpc.c  	entry->ctrl_info = msg->content.in_info;
entry            1178 net/atm/mpc.c  	entry->time = ktime_get_seconds();
entry            1180 net/atm/mpc.c  	entry->reply_wait = ktime_get_seconds();
entry            1181 net/atm/mpc.c  	entry->refresh_time = 0;
entry            1182 net/atm/mpc.c  	ddprintk_cont("entry->shortcut = %p\n", entry->shortcut);
entry            1184 net/atm/mpc.c  	if (entry->entry_state == INGRESS_RESOLVING &&
entry            1185 net/atm/mpc.c  	    entry->shortcut != NULL) {
entry            1186 net/atm/mpc.c  		entry->entry_state = INGRESS_RESOLVED;
entry            1187 net/atm/mpc.c  		mpc->in_ops->put(entry);
entry            1191 net/atm/mpc.c  	if (entry->shortcut != NULL) {
entry            1194 net/atm/mpc.c  		mpc->in_ops->put(entry);
entry            1198 net/atm/mpc.c  	check_qos_and_open_shortcut(msg, mpc, entry);
entry            1199 net/atm/mpc.c  	entry->entry_state = INGRESS_RESOLVED;
entry            1200 net/atm/mpc.c  	mpc->in_ops->put(entry);
entry            1210 net/atm/mpc.c  	in_cache_entry *entry = mpc->in_ops->get_with_mask(dst_ip, mpc, mask);
entry            1212 net/atm/mpc.c  	if (entry == NULL) {
entry            1222 net/atm/mpc.c  		mpc->in_ops->remove_entry(entry, mpc);
entry            1224 net/atm/mpc.c  		mpc->in_ops->put(entry);
entry            1225 net/atm/mpc.c  		entry = mpc->in_ops->get_with_mask(dst_ip, mpc, mask);
entry            1226 net/atm/mpc.c  	} while (entry != NULL);
entry            1232 net/atm/mpc.c  	eg_cache_entry *entry = mpc->eg_ops->get_by_cache_id(cache_id, mpc);
entry            1234 net/atm/mpc.c  	if (entry == NULL) {
entry            1241 net/atm/mpc.c  	mpc->eg_ops->remove_entry(entry, mpc);
entry            1244 net/atm/mpc.c  	mpc->eg_ops->put(entry);
entry            1247 net/atm/mpc.c  static void purge_egress_shortcut(struct atm_vcc *vcc, eg_cache_entry *entry)
entry            1269 net/atm/mpc.c  	if (entry != NULL)
entry            1270 net/atm/mpc.c  		purge_msg->content.eg_info = entry->ctrl_info;
entry            1286 net/atm/mpc.c  	eg_cache_entry *entry;
entry            1297 net/atm/mpc.c  	entry = mpc->eg_cache;
entry            1298 net/atm/mpc.c  	while (entry != NULL) {
entry            1299 net/atm/mpc.c  		purge_egress_shortcut(entry->shortcut, entry);
entry            1300 net/atm/mpc.c  		entry = entry->next;
entry            1312 net/atm/mpc.c  	eg_cache_entry *entry = mpc->eg_ops->get_by_cache_id(msg->content.eg_info.cache_id, mpc);
entry            1316 net/atm/mpc.c  		mpc->dev->name, entry, holding_time);
entry            1317 net/atm/mpc.c  	if (entry == NULL && holding_time) {
entry            1318 net/atm/mpc.c  		entry = mpc->eg_ops->add_entry(msg, mpc);
entry            1319 net/atm/mpc.c  		mpc->eg_ops->put(entry);
entry            1323 net/atm/mpc.c  		mpc->eg_ops->update(entry, holding_time);
entry            1328 net/atm/mpc.c  	mpc->eg_ops->remove_entry(entry, mpc);
entry            1331 net/atm/mpc.c  	mpc->eg_ops->put(entry);
entry            1392 net/atm/mpc.c  	eg_cache_entry *entry;
entry            1398 net/atm/mpc.c  	entry = mpc->eg_cache;
entry            1399 net/atm/mpc.c  	while (entry != NULL) {
entry            1400 net/atm/mpc.c  		msg->content.eg_info = entry->ctrl_info;
entry            1401 net/atm/mpc.c  		dprintk("cache_id %u\n", entry->ctrl_info.cache_id);
entry            1403 net/atm/mpc.c  		entry = entry->next;
entry              38 net/atm/mpoa_caches.c 	in_cache_entry *entry;
entry              41 net/atm/mpoa_caches.c 	entry = client->in_cache;
entry              42 net/atm/mpoa_caches.c 	while (entry != NULL) {
entry              43 net/atm/mpoa_caches.c 		if (entry->ctrl_info.in_dst_ip == dst_ip) {
entry              44 net/atm/mpoa_caches.c 			refcount_inc(&entry->use);
entry              46 net/atm/mpoa_caches.c 			return entry;
entry              48 net/atm/mpoa_caches.c 		entry = entry->next;
entry              59 net/atm/mpoa_caches.c 	in_cache_entry *entry;
entry              62 net/atm/mpoa_caches.c 	entry = client->in_cache;
entry              63 net/atm/mpoa_caches.c 	while (entry != NULL) {
entry              64 net/atm/mpoa_caches.c 		if ((entry->ctrl_info.in_dst_ip & mask) == (dst_ip & mask)) {
entry              65 net/atm/mpoa_caches.c 			refcount_inc(&entry->use);
entry              67 net/atm/mpoa_caches.c 			return entry;
entry              69 net/atm/mpoa_caches.c 		entry = entry->next;
entry              80 net/atm/mpoa_caches.c 	in_cache_entry *entry;
entry              83 net/atm/mpoa_caches.c 	entry = client->in_cache;
entry              84 net/atm/mpoa_caches.c 	while (entry != NULL) {
entry              85 net/atm/mpoa_caches.c 		if (entry->shortcut == vcc) {
entry              86 net/atm/mpoa_caches.c 			refcount_inc(&entry->use);
entry              88 net/atm/mpoa_caches.c 			return entry;
entry              90 net/atm/mpoa_caches.c 		entry = entry->next;
entry             100 net/atm/mpoa_caches.c 	in_cache_entry *entry = kzalloc(sizeof(in_cache_entry), GFP_KERNEL);
entry             102 net/atm/mpoa_caches.c 	if (entry == NULL) {
entry             109 net/atm/mpoa_caches.c 	refcount_set(&entry->use, 1);
entry             112 net/atm/mpoa_caches.c 	entry->next = client->in_cache;
entry             113 net/atm/mpoa_caches.c 	entry->prev = NULL;
entry             115 net/atm/mpoa_caches.c 		client->in_cache->prev = entry;
entry             116 net/atm/mpoa_caches.c 	client->in_cache = entry;
entry             118 net/atm/mpoa_caches.c 	memcpy(entry->MPS_ctrl_ATM_addr, client->mps_ctrl_addr, ATM_ESA_LEN);
entry             119 net/atm/mpoa_caches.c 	entry->ctrl_info.in_dst_ip = dst_ip;
entry             120 net/atm/mpoa_caches.c 	entry->time = ktime_get_seconds();
entry             121 net/atm/mpoa_caches.c 	entry->retry_time = client->parameters.mpc_p4;
entry             122 net/atm/mpoa_caches.c 	entry->count = 1;
entry             123 net/atm/mpoa_caches.c 	entry->entry_state = INGRESS_INVALID;
entry             124 net/atm/mpoa_caches.c 	entry->ctrl_info.holding_time = HOLDING_TIME_DEFAULT;
entry             125 net/atm/mpoa_caches.c 	refcount_inc(&entry->use);
entry             130 net/atm/mpoa_caches.c 	return entry;
entry             133 net/atm/mpoa_caches.c static int cache_hit(in_cache_entry *entry, struct mpoa_client *mpc)
entry             138 net/atm/mpoa_caches.c 	entry->count++;
entry             139 net/atm/mpoa_caches.c 	if (entry->entry_state == INGRESS_RESOLVED && entry->shortcut != NULL)
entry             142 net/atm/mpoa_caches.c 	if (entry->entry_state == INGRESS_REFRESHING) {
entry             143 net/atm/mpoa_caches.c 		if (entry->count > mpc->parameters.mpc_p1) {
entry             145 net/atm/mpoa_caches.c 			msg.content.in_info = entry->ctrl_info;
entry             147 net/atm/mpoa_caches.c 			qos = atm_mpoa_search_qos(entry->ctrl_info.in_dst_ip);
entry             151 net/atm/mpoa_caches.c 			entry->reply_wait = ktime_get_seconds();
entry             152 net/atm/mpoa_caches.c 			entry->entry_state = INGRESS_RESOLVING;
entry             154 net/atm/mpoa_caches.c 		if (entry->shortcut != NULL)
entry             159 net/atm/mpoa_caches.c 	if (entry->entry_state == INGRESS_RESOLVING && entry->shortcut != NULL)
entry             162 net/atm/mpoa_caches.c 	if (entry->count > mpc->parameters.mpc_p1 &&
entry             163 net/atm/mpoa_caches.c 	    entry->entry_state == INGRESS_INVALID) {
entry             165 net/atm/mpoa_caches.c 			mpc->dev->name, &entry->ctrl_info.in_dst_ip);
entry             166 net/atm/mpoa_caches.c 		entry->entry_state = INGRESS_RESOLVING;
entry             169 net/atm/mpoa_caches.c 		msg.content.in_info = entry->ctrl_info;
entry             170 net/atm/mpoa_caches.c 		qos = atm_mpoa_search_qos(entry->ctrl_info.in_dst_ip);
entry             174 net/atm/mpoa_caches.c 		entry->reply_wait = ktime_get_seconds();
entry             180 net/atm/mpoa_caches.c static void in_cache_put(in_cache_entry *entry)
entry             182 net/atm/mpoa_caches.c 	if (refcount_dec_and_test(&entry->use)) {
entry             183 net/atm/mpoa_caches.c 		kzfree(entry);
entry             190 net/atm/mpoa_caches.c static void in_cache_remove_entry(in_cache_entry *entry,
entry             196 net/atm/mpoa_caches.c 	vcc = entry->shortcut;
entry             198 net/atm/mpoa_caches.c 		&entry->ctrl_info.in_dst_ip);
entry             200 net/atm/mpoa_caches.c 	if (entry->prev != NULL)
entry             201 net/atm/mpoa_caches.c 		entry->prev->next = entry->next;
entry             203 net/atm/mpoa_caches.c 		client->in_cache = entry->next;
entry             204 net/atm/mpoa_caches.c 	if (entry->next != NULL)
entry             205 net/atm/mpoa_caches.c 		entry->next->prev = entry->prev;
entry             206 net/atm/mpoa_caches.c 	client->in_ops->put(entry);
entry             228 net/atm/mpoa_caches.c 	in_cache_entry *entry, *next_entry;
entry             234 net/atm/mpoa_caches.c 	entry = client->in_cache;
entry             235 net/atm/mpoa_caches.c 	while (entry != NULL) {
entry             236 net/atm/mpoa_caches.c 		entry->count = 0;
entry             237 net/atm/mpoa_caches.c 		next_entry = entry->next;
entry             238 net/atm/mpoa_caches.c 		if ((now - entry->time) > entry->ctrl_info.holding_time) {
entry             240 net/atm/mpoa_caches.c 				&entry->ctrl_info.in_dst_ip);
entry             241 net/atm/mpoa_caches.c 			client->in_ops->remove_entry(entry, client);
entry             243 net/atm/mpoa_caches.c 		entry = next_entry;
entry             253 net/atm/mpoa_caches.c 	in_cache_entry *entry;
entry             260 net/atm/mpoa_caches.c 	entry = client->in_cache;
entry             261 net/atm/mpoa_caches.c 	while (entry != NULL) {
entry             262 net/atm/mpoa_caches.c 		if (entry->entry_state == INGRESS_RESOLVING) {
entry             264 net/atm/mpoa_caches.c 			if ((now - entry->hold_down)
entry             266 net/atm/mpoa_caches.c 				entry = entry->next;	/* Entry in hold down */
entry             269 net/atm/mpoa_caches.c 			if ((now - entry->reply_wait) > entry->retry_time) {
entry             270 net/atm/mpoa_caches.c 				entry->retry_time = MPC_C1 * (entry->retry_time);
entry             275 net/atm/mpoa_caches.c 				if (entry->retry_time > client->parameters.mpc_p5) {
entry             276 net/atm/mpoa_caches.c 					entry->hold_down = ktime_get_seconds();
entry             277 net/atm/mpoa_caches.c 					entry->retry_time = client->parameters.mpc_p4;
entry             278 net/atm/mpoa_caches.c 					entry = entry->next;
entry             282 net/atm/mpoa_caches.c 				memset(&entry->hold_down, 0, sizeof(time64_t));
entry             285 net/atm/mpoa_caches.c 				msg.content.in_info = entry->ctrl_info;
entry             286 net/atm/mpoa_caches.c 				qos = atm_mpoa_search_qos(entry->ctrl_info.in_dst_ip);
entry             290 net/atm/mpoa_caches.c 				entry->reply_wait = ktime_get_seconds();
entry             293 net/atm/mpoa_caches.c 		entry = entry->next;
entry             302 net/atm/mpoa_caches.c 	struct in_cache_entry *entry = client->in_cache;
entry             308 net/atm/mpoa_caches.c 	while (entry != NULL) {
entry             309 net/atm/mpoa_caches.c 		if (entry->entry_state == INGRESS_RESOLVED) {
entry             310 net/atm/mpoa_caches.c 			if (!(entry->refresh_time))
entry             311 net/atm/mpoa_caches.c 				entry->refresh_time = (2 * (entry->ctrl_info.holding_time))/3;
entry             312 net/atm/mpoa_caches.c 			if ((now - entry->reply_wait) >
entry             313 net/atm/mpoa_caches.c 			    entry->refresh_time) {
entry             315 net/atm/mpoa_caches.c 				entry->entry_state = INGRESS_REFRESHING;
entry             319 net/atm/mpoa_caches.c 		entry = entry->next;
entry             335 net/atm/mpoa_caches.c 	eg_cache_entry *entry;
entry             338 net/atm/mpoa_caches.c 	entry = mpc->eg_cache;
entry             339 net/atm/mpoa_caches.c 	while (entry != NULL) {
entry             340 net/atm/mpoa_caches.c 		if (entry->ctrl_info.cache_id == cache_id) {
entry             341 net/atm/mpoa_caches.c 			refcount_inc(&entry->use);
entry             343 net/atm/mpoa_caches.c 			return entry;
entry             345 net/atm/mpoa_caches.c 		entry = entry->next;
entry             356 net/atm/mpoa_caches.c 	eg_cache_entry *entry;
entry             359 net/atm/mpoa_caches.c 	entry = mpc->eg_cache;
entry             360 net/atm/mpoa_caches.c 	while (entry != NULL) {
entry             361 net/atm/mpoa_caches.c 		if (entry->ctrl_info.tag == tag) {
entry             362 net/atm/mpoa_caches.c 			refcount_inc(&entry->use);
entry             364 net/atm/mpoa_caches.c 			return entry;
entry             366 net/atm/mpoa_caches.c 		entry = entry->next;
entry             378 net/atm/mpoa_caches.c 	eg_cache_entry *entry;
entry             381 net/atm/mpoa_caches.c 	entry = mpc->eg_cache;
entry             382 net/atm/mpoa_caches.c 	while (entry != NULL) {
entry             383 net/atm/mpoa_caches.c 		if (entry->shortcut == vcc) {
entry             384 net/atm/mpoa_caches.c 			refcount_inc(&entry->use);
entry             386 net/atm/mpoa_caches.c 			return entry;
entry             388 net/atm/mpoa_caches.c 		entry = entry->next;
entry             398 net/atm/mpoa_caches.c 	eg_cache_entry *entry;
entry             401 net/atm/mpoa_caches.c 	entry = mpc->eg_cache;
entry             402 net/atm/mpoa_caches.c 	while (entry != NULL) {
entry             403 net/atm/mpoa_caches.c 		if (entry->latest_ip_addr == ipaddr) {
entry             404 net/atm/mpoa_caches.c 			refcount_inc(&entry->use);
entry             406 net/atm/mpoa_caches.c 			return entry;
entry             408 net/atm/mpoa_caches.c 		entry = entry->next;
entry             415 net/atm/mpoa_caches.c static void eg_cache_put(eg_cache_entry *entry)
entry             417 net/atm/mpoa_caches.c 	if (refcount_dec_and_test(&entry->use)) {
entry             418 net/atm/mpoa_caches.c 		kzfree(entry);
entry             425 net/atm/mpoa_caches.c static void eg_cache_remove_entry(eg_cache_entry *entry,
entry             431 net/atm/mpoa_caches.c 	vcc = entry->shortcut;
entry             433 net/atm/mpoa_caches.c 	if (entry->prev != NULL)
entry             434 net/atm/mpoa_caches.c 		entry->prev->next = entry->next;
entry             436 net/atm/mpoa_caches.c 		client->eg_cache = entry->next;
entry             437 net/atm/mpoa_caches.c 	if (entry->next != NULL)
entry             438 net/atm/mpoa_caches.c 		entry->next->prev = entry->prev;
entry             439 net/atm/mpoa_caches.c 	client->eg_ops->put(entry);
entry             459 net/atm/mpoa_caches.c 	eg_cache_entry *entry = kzalloc(sizeof(eg_cache_entry), GFP_KERNEL);
entry             461 net/atm/mpoa_caches.c 	if (entry == NULL) {
entry             469 net/atm/mpoa_caches.c 	refcount_set(&entry->use, 1);
entry             472 net/atm/mpoa_caches.c 	entry->next = client->eg_cache;
entry             473 net/atm/mpoa_caches.c 	entry->prev = NULL;
entry             475 net/atm/mpoa_caches.c 		client->eg_cache->prev = entry;
entry             476 net/atm/mpoa_caches.c 	client->eg_cache = entry;
entry             478 net/atm/mpoa_caches.c 	memcpy(entry->MPS_ctrl_ATM_addr, client->mps_ctrl_addr, ATM_ESA_LEN);
entry             479 net/atm/mpoa_caches.c 	entry->ctrl_info = msg->content.eg_info;
entry             480 net/atm/mpoa_caches.c 	entry->time = ktime_get_seconds();
entry             481 net/atm/mpoa_caches.c 	entry->entry_state = EGRESS_RESOLVED;
entry             483 net/atm/mpoa_caches.c 		ntohl(entry->ctrl_info.cache_id));
entry             484 net/atm/mpoa_caches.c 	dprintk("mps_ip = %pI4\n", &entry->ctrl_info.mps_ip);
entry             485 net/atm/mpoa_caches.c 	refcount_inc(&entry->use);
entry             490 net/atm/mpoa_caches.c 	return entry;
entry             493 net/atm/mpoa_caches.c static void update_eg_cache_entry(eg_cache_entry *entry, uint16_t holding_time)
entry             495 net/atm/mpoa_caches.c 	entry->time = ktime_get_seconds();
entry             496 net/atm/mpoa_caches.c 	entry->entry_state = EGRESS_RESOLVED;
entry             497 net/atm/mpoa_caches.c 	entry->ctrl_info.holding_time = holding_time;
entry             502 net/atm/mpoa_caches.c 	eg_cache_entry *entry, *next_entry;
entry             509 net/atm/mpoa_caches.c 	entry = client->eg_cache;
entry             510 net/atm/mpoa_caches.c 	while (entry != NULL) {
entry             511 net/atm/mpoa_caches.c 		next_entry = entry->next;
entry             512 net/atm/mpoa_caches.c 		if ((now - entry->time) > entry->ctrl_info.holding_time) {
entry             514 net/atm/mpoa_caches.c 			msg.content.eg_info = entry->ctrl_info;
entry             516 net/atm/mpoa_caches.c 				ntohl(entry->ctrl_info.cache_id));
entry             518 net/atm/mpoa_caches.c 			client->eg_ops->remove_entry(entry, client);
entry             520 net/atm/mpoa_caches.c 		entry = next_entry;
entry              43 net/atm/mpoa_caches.h 	void            (*put)(in_cache_entry *entry);
entry              46 net/atm/mpoa_caches.h 	int             (*cache_hit)(in_cache_entry *entry,
entry              73 net/atm/mpoa_caches.h 	void            (*put)(eg_cache_entry *entry);
entry              74 net/atm/mpoa_caches.h 	void            (*remove_entry)(eg_cache_entry *entry, struct mpoa_client *client);
entry              75 net/atm/mpoa_caches.h 	void            (*update)(eg_cache_entry *entry, uint16_t holding_time);
entry             163 net/atm/proc.c 		dev = clip_vcc->entry ? clip_vcc->entry->neigh->dev : NULL;
entry             489 net/batman-adv/bridge_loop_avoidance.c 	struct batadv_bla_backbone_gw *entry;
entry             493 net/batman-adv/bridge_loop_avoidance.c 	entry = batadv_backbone_hash_find(bat_priv, orig, vid);
entry             495 net/batman-adv/bridge_loop_avoidance.c 	if (entry)
entry             496 net/batman-adv/bridge_loop_avoidance.c 		return entry;
entry             502 net/batman-adv/bridge_loop_avoidance.c 	entry = kzalloc(sizeof(*entry), GFP_ATOMIC);
entry             503 net/batman-adv/bridge_loop_avoidance.c 	if (!entry)
entry             506 net/batman-adv/bridge_loop_avoidance.c 	entry->vid = vid;
entry             507 net/batman-adv/bridge_loop_avoidance.c 	entry->lasttime = jiffies;
entry             508 net/batman-adv/bridge_loop_avoidance.c 	entry->crc = BATADV_BLA_CRC_INIT;
entry             509 net/batman-adv/bridge_loop_avoidance.c 	entry->bat_priv = bat_priv;
entry             510 net/batman-adv/bridge_loop_avoidance.c 	spin_lock_init(&entry->crc_lock);
entry             511 net/batman-adv/bridge_loop_avoidance.c 	atomic_set(&entry->request_sent, 0);
entry             512 net/batman-adv/bridge_loop_avoidance.c 	atomic_set(&entry->wait_periods, 0);
entry             513 net/batman-adv/bridge_loop_avoidance.c 	ether_addr_copy(entry->orig, orig);
entry             514 net/batman-adv/bridge_loop_avoidance.c 	INIT_WORK(&entry->report_work, batadv_bla_loopdetect_report);
entry             515 net/batman-adv/bridge_loop_avoidance.c 	kref_init(&entry->refcount);
entry             517 net/batman-adv/bridge_loop_avoidance.c 	kref_get(&entry->refcount);
entry             520 net/batman-adv/bridge_loop_avoidance.c 				     batadv_choose_backbone_gw, entry,
entry             521 net/batman-adv/bridge_loop_avoidance.c 				     &entry->hash_entry);
entry             525 net/batman-adv/bridge_loop_avoidance.c 		kfree(entry);
entry             538 net/batman-adv/bridge_loop_avoidance.c 		batadv_bla_send_announce(bat_priv, entry);
entry             541 net/batman-adv/bridge_loop_avoidance.c 		atomic_inc(&entry->request_sent);
entry             542 net/batman-adv/bridge_loop_avoidance.c 		atomic_set(&entry->wait_periods, BATADV_BLA_WAIT_PERIODS);
entry             546 net/batman-adv/bridge_loop_avoidance.c 	return entry;
entry            1600 net/batman-adv/bridge_loop_avoidance.c 	struct batadv_bcast_duplist_entry *entry;
entry            1613 net/batman-adv/bridge_loop_avoidance.c 		entry = &bat_priv->bla.bcast_duplist[curr];
entry            1618 net/batman-adv/bridge_loop_avoidance.c 		if (batadv_has_timed_out(entry->entrytime,
entry            1622 net/batman-adv/bridge_loop_avoidance.c 		if (entry->crc != crc)
entry            1625 net/batman-adv/bridge_loop_avoidance.c 		if (batadv_compare_eth(entry->orig, bcast_packet->orig))
entry            1639 net/batman-adv/bridge_loop_avoidance.c 	entry = &bat_priv->bla.bcast_duplist[curr];
entry            1640 net/batman-adv/bridge_loop_avoidance.c 	entry->crc = crc;
entry            1641 net/batman-adv/bridge_loop_avoidance.c 	entry->entrytime = jiffies;
entry            1642 net/batman-adv/bridge_loop_avoidance.c 	ether_addr_copy(entry->orig, bcast_packet->orig);
entry              41 net/batman-adv/fragmentation.c 	struct batadv_frag_list_entry *entry;
entry              44 net/batman-adv/fragmentation.c 	hlist_for_each_entry_safe(entry, node, head, list) {
entry              45 net/batman-adv/fragmentation.c 		hlist_del(&entry->list);
entry              48 net/batman-adv/fragmentation.c 			kfree_skb(entry->skb);
entry              50 net/batman-adv/fragmentation.c 			consume_skb(entry->skb);
entry              52 net/batman-adv/fragmentation.c 		kfree(entry);
entry             252 net/batman-adv/fragmentation.c 	struct batadv_frag_list_entry *entry;
entry             260 net/batman-adv/fragmentation.c 	entry = hlist_entry(chain->first, struct batadv_frag_list_entry, list);
entry             261 net/batman-adv/fragmentation.c 	hlist_del(&entry->list);
entry             262 net/batman-adv/fragmentation.c 	skb_out = entry->skb;
entry             263 net/batman-adv/fragmentation.c 	kfree(entry);
entry             287 net/batman-adv/fragmentation.c 	hlist_for_each_entry(entry, chain, list) {
entry             288 net/batman-adv/fragmentation.c 		size = entry->skb->len - hdr_size;
entry             289 net/batman-adv/fragmentation.c 		skb_put_data(skb_out, entry->skb->data + hdr_size, size);
entry             469 net/batman-adv/translation-table.c 	struct batadv_tt_change_node *tt_change_node, *entry, *safe;
entry             489 net/batman-adv/translation-table.c 	list_for_each_entry_safe(entry, safe, &bat_priv->tt.changes_list,
entry             491 net/batman-adv/translation-table.c 		if (!batadv_compare_eth(entry->change.addr, common->addr))
entry             501 net/batman-adv/translation-table.c 		del_op_entry = entry->change.flags & BATADV_TT_CLIENT_DEL;
entry             511 net/batman-adv/translation-table.c 			entry->change.flags = flags;
entry             515 net/batman-adv/translation-table.c 		list_del(&entry->list);
entry             516 net/batman-adv/translation-table.c 		kmem_cache_free(batadv_tt_change_cache, entry);
entry             996 net/batman-adv/translation-table.c 	struct batadv_tt_change_node *entry, *safe;
entry            1026 net/batman-adv/translation-table.c 	list_for_each_entry_safe(entry, safe, &bat_priv->tt.changes_list,
entry            1030 net/batman-adv/translation-table.c 			       &entry->change,
entry            1034 net/batman-adv/translation-table.c 		list_del(&entry->list);
entry            1035 net/batman-adv/translation-table.c 		kmem_cache_free(batadv_tt_change_cache, entry);
entry            1505 net/batman-adv/translation-table.c 	struct batadv_tt_change_node *entry, *safe;
entry            1509 net/batman-adv/translation-table.c 	list_for_each_entry_safe(entry, safe, &bat_priv->tt.changes_list,
entry            1511 net/batman-adv/translation-table.c 		list_del(&entry->list);
entry            1512 net/batman-adv/translation-table.c 		kmem_cache_free(batadv_tt_change_cache, entry);
entry            1531 net/batman-adv/translation-table.c batadv_tt_global_orig_entry_find(const struct batadv_tt_global_entry *entry,
entry            1538 net/batman-adv/translation-table.c 	head = &entry->orig_list;
entry            1566 net/batman-adv/translation-table.c batadv_tt_global_entry_has_orig(const struct batadv_tt_global_entry *entry,
entry            1573 net/batman-adv/translation-table.c 	orig_entry = batadv_tt_global_orig_entry_find(entry, orig_node);
entry             228 net/bluetooth/6lowpan.c 	struct lowpan_btle_dev *entry;
entry             233 net/bluetooth/6lowpan.c 	list_for_each_entry_rcu(entry, &bt_6lowpan_devices, list) {
entry             234 net/bluetooth/6lowpan.c 		peer = __peer_lookup_conn(entry, conn);
entry             246 net/bluetooth/6lowpan.c 	struct lowpan_btle_dev *entry;
entry             251 net/bluetooth/6lowpan.c 	list_for_each_entry_rcu(entry, &bt_6lowpan_devices, list) {
entry             252 net/bluetooth/6lowpan.c 		if (conn->hcon->hdev == entry->hdev) {
entry             253 net/bluetooth/6lowpan.c 			dev = entry;
entry             490 net/bluetooth/6lowpan.c 	struct lowpan_btle_dev *entry;
entry             495 net/bluetooth/6lowpan.c 	list_for_each_entry_rcu(entry, &bt_6lowpan_devices, list) {
entry             499 net/bluetooth/6lowpan.c 		if (entry->netdev != netdev)
entry             502 net/bluetooth/6lowpan.c 		dev = lowpan_btle_dev(entry->netdev);
entry             773 net/bluetooth/6lowpan.c 	struct lowpan_btle_dev *entry = container_of(work,
entry             777 net/bluetooth/6lowpan.c 	lowpan_unregister_netdev(entry->netdev);
entry             784 net/bluetooth/6lowpan.c 	struct lowpan_btle_dev *entry;
entry             804 net/bluetooth/6lowpan.c 	list_for_each_entry_rcu(entry, &bt_6lowpan_devices, list) {
entry             805 net/bluetooth/6lowpan.c 		dev = lowpan_btle_dev(entry->netdev);
entry             829 net/bluetooth/6lowpan.c 			INIT_WORK(&entry->delete_netdev, delete_netdev);
entry             830 net/bluetooth/6lowpan.c 			schedule_work(&entry->delete_netdev);
entry            1017 net/bluetooth/6lowpan.c 	struct lowpan_btle_dev *entry;
entry            1030 net/bluetooth/6lowpan.c 	list_for_each_entry_rcu(entry, &bt_6lowpan_devices, list) {
entry            1031 net/bluetooth/6lowpan.c 		list_for_each_entry_rcu(peer, &entry->peers, list) {
entry            1177 net/bluetooth/6lowpan.c 	struct lowpan_btle_dev *entry;
entry            1182 net/bluetooth/6lowpan.c 	list_for_each_entry(entry, &bt_6lowpan_devices, list) {
entry            1183 net/bluetooth/6lowpan.c 		list_for_each_entry(peer, &entry->peers, list)
entry            1208 net/bluetooth/6lowpan.c 	struct lowpan_btle_dev *entry, *tmp, *new_dev;
entry            1220 net/bluetooth/6lowpan.c 	list_for_each_entry_rcu(entry, &bt_6lowpan_devices, list) {
entry            1225 net/bluetooth/6lowpan.c 		new_dev->netdev = entry->netdev;
entry            1233 net/bluetooth/6lowpan.c 	list_for_each_entry_safe(entry, tmp, &devices, list) {
entry            1234 net/bluetooth/6lowpan.c 		ifdown(entry->netdev);
entry            1236 net/bluetooth/6lowpan.c 		       entry->netdev->name, entry->netdev);
entry            1237 net/bluetooth/6lowpan.c 		lowpan_unregister_netdev(entry->netdev);
entry            1238 net/bluetooth/6lowpan.c 		kfree(entry);
entry            1246 net/bluetooth/6lowpan.c 	struct lowpan_btle_dev *entry;
entry            1254 net/bluetooth/6lowpan.c 		list_for_each_entry(entry, &bt_6lowpan_devices, list) {
entry            1255 net/bluetooth/6lowpan.c 			if (entry->netdev == netdev) {
entry            1258 net/bluetooth/6lowpan.c 				list_del(&entry->list);
entry            2934 net/bluetooth/hci_core.c 	struct bdaddr_list *entry;
entry            2942 net/bluetooth/hci_core.c 	entry = kzalloc(sizeof(*entry), GFP_KERNEL);
entry            2943 net/bluetooth/hci_core.c 	if (!entry)
entry            2946 net/bluetooth/hci_core.c 	bacpy(&entry->bdaddr, bdaddr);
entry            2947 net/bluetooth/hci_core.c 	entry->bdaddr_type = type;
entry            2949 net/bluetooth/hci_core.c 	list_add(&entry->list, list);
entry            2957 net/bluetooth/hci_core.c 	struct bdaddr_list_with_irk *entry;
entry            2965 net/bluetooth/hci_core.c 	entry = kzalloc(sizeof(*entry), GFP_KERNEL);
entry            2966 net/bluetooth/hci_core.c 	if (!entry)
entry            2969 net/bluetooth/hci_core.c 	bacpy(&entry->bdaddr, bdaddr);
entry            2970 net/bluetooth/hci_core.c 	entry->bdaddr_type = type;
entry            2973 net/bluetooth/hci_core.c 		memcpy(entry->peer_irk, peer_irk, 16);
entry            2976 net/bluetooth/hci_core.c 		memcpy(entry->local_irk, local_irk, 16);
entry            2978 net/bluetooth/hci_core.c 	list_add(&entry->list, list);
entry            2985 net/bluetooth/hci_core.c 	struct bdaddr_list *entry;
entry            2992 net/bluetooth/hci_core.c 	entry = hci_bdaddr_list_lookup(list, bdaddr, type);
entry            2993 net/bluetooth/hci_core.c 	if (!entry)
entry            2996 net/bluetooth/hci_core.c 	list_del(&entry->list);
entry            2997 net/bluetooth/hci_core.c 	kfree(entry);
entry            3005 net/bluetooth/hci_core.c 	struct bdaddr_list_with_irk *entry;
entry            3012 net/bluetooth/hci_core.c 	entry = hci_bdaddr_list_lookup_with_irk(list, bdaddr, type);
entry            3013 net/bluetooth/hci_core.c 	if (!entry)
entry            3016 net/bluetooth/hci_core.c 	list_del(&entry->list);
entry            3017 net/bluetooth/hci_core.c 	kfree(entry);
entry             490 net/bluetooth/mgmt.c 	rp = kmalloc(struct_size(rp, entry, count), GFP_ATOMIC);
entry             511 net/bluetooth/mgmt.c 				rp->entry[count].type = 0x01;
entry             513 net/bluetooth/mgmt.c 				rp->entry[count].type = 0x00;
entry             515 net/bluetooth/mgmt.c 			rp->entry[count].type = 0x02;
entry             520 net/bluetooth/mgmt.c 		rp->entry[count].bus = d->bus;
entry             521 net/bluetooth/mgmt.c 		rp->entry[count++].index = cpu_to_le16(d->id);
entry             539 net/bluetooth/mgmt.c 				struct_size(rp, entry, count));
entry             219 net/bluetooth/rfcomm/tty.c 	struct rfcomm_dev *dev, *entry;
entry             232 net/bluetooth/rfcomm/tty.c 		list_for_each_entry(entry, &rfcomm_dev_list, list) {
entry             233 net/bluetooth/rfcomm/tty.c 			if (entry->id != dev->id)
entry             237 net/bluetooth/rfcomm/tty.c 			head = &entry->list;
entry             242 net/bluetooth/rfcomm/tty.c 		list_for_each_entry(entry, &rfcomm_dev_list, list) {
entry             243 net/bluetooth/rfcomm/tty.c 			if (entry->id == dev->id) {
entry             248 net/bluetooth/rfcomm/tty.c 			if (entry->id > dev->id - 1)
entry             251 net/bluetooth/rfcomm/tty.c 			head = &entry->list;
entry              67 net/bridge/br_mdb.c static void __mdb_entry_to_br_ip(struct br_mdb_entry *entry, struct br_ip *ip)
entry              70 net/bridge/br_mdb.c 	ip->vid = entry->vid;
entry              71 net/bridge/br_mdb.c 	ip->proto = entry->addr.proto;
entry              73 net/bridge/br_mdb.c 		ip->u.ip4 = entry->addr.u.ip4;
entry              76 net/bridge/br_mdb.c 		ip->u.ip6 = entry->addr.u.ip6;
entry             266 net/bridge/br_mdb.c 				   struct br_mdb_entry *entry, u32 pid,
entry             288 net/bridge/br_mdb.c 	if (nla_put(skb, MDBA_MDB_ENTRY_INFO, sizeof(*entry), entry))
entry             344 net/bridge/br_mdb.c 				       struct br_mdb_entry *entry, int type)
entry             351 net/bridge/br_mdb.c 		.vid = entry->vid,
entry             354 net/bridge/br_mdb.c 	if (entry->addr.proto == htons(ETH_P_IP))
entry             355 net/bridge/br_mdb.c 		ip_eth_mc_map(entry->addr.u.ip4, mdb.addr);
entry             358 net/bridge/br_mdb.c 		ipv6_eth_mc_map(&entry->addr.u.ip6, mdb.addr);
entry             373 net/bridge/br_mdb.c 				  struct br_mdb_entry *entry, int type)
entry             379 net/bridge/br_mdb.c 		br_mdb_switchdev_host_port(dev, lower_dev, entry, type);
entry             383 net/bridge/br_mdb.c 			    struct br_mdb_entry *entry, int type)
entry             391 net/bridge/br_mdb.c 		.vid = entry->vid,
entry             398 net/bridge/br_mdb.c 	port_dev = __dev_get_by_index(net, entry->ifindex);
entry             399 net/bridge/br_mdb.c 	if (entry->addr.proto == htons(ETH_P_IP))
entry             400 net/bridge/br_mdb.c 		ip_eth_mc_map(entry->addr.u.ip4, mdb.addr);
entry             403 net/bridge/br_mdb.c 		ipv6_eth_mc_map(&entry->addr.u.ip6, mdb.addr);
entry             411 net/bridge/br_mdb.c 			__mdb_entry_to_br_ip(entry, &complete_info->ip);
entry             422 net/bridge/br_mdb.c 		br_mdb_switchdev_host(dev, entry, type);
entry             428 net/bridge/br_mdb.c 	err = nlmsg_populate_mdb_fill(skb, dev, entry, 0, 0, type, NTF_SELF);
entry             443 net/bridge/br_mdb.c 	struct br_mdb_entry entry;
entry             445 net/bridge/br_mdb.c 	memset(&entry, 0, sizeof(entry));
entry             447 net/bridge/br_mdb.c 		entry.ifindex = port->dev->ifindex;
entry             449 net/bridge/br_mdb.c 		entry.ifindex = dev->ifindex;
entry             450 net/bridge/br_mdb.c 	entry.addr.proto = group->proto;
entry             451 net/bridge/br_mdb.c 	entry.addr.u.ip4 = group->u.ip4;
entry             453 net/bridge/br_mdb.c 	entry.addr.u.ip6 = group->u.ip6;
entry             455 net/bridge/br_mdb.c 	entry.vid = group->vid;
entry             456 net/bridge/br_mdb.c 	__mdb_entry_fill_flags(&entry, flags);
entry             457 net/bridge/br_mdb.c 	__br_mdb_notify(dev, port, &entry, type);
entry             527 net/bridge/br_mdb.c static bool is_valid_mdb_entry(struct br_mdb_entry *entry)
entry             529 net/bridge/br_mdb.c 	if (entry->ifindex == 0)
entry             532 net/bridge/br_mdb.c 	if (entry->addr.proto == htons(ETH_P_IP)) {
entry             533 net/bridge/br_mdb.c 		if (!ipv4_is_multicast(entry->addr.u.ip4))
entry             535 net/bridge/br_mdb.c 		if (ipv4_is_local_multicast(entry->addr.u.ip4))
entry             538 net/bridge/br_mdb.c 	} else if (entry->addr.proto == htons(ETH_P_IPV6)) {
entry             539 net/bridge/br_mdb.c 		if (ipv6_addr_is_ll_all_nodes(&entry->addr.u.ip6))
entry             544 net/bridge/br_mdb.c 	if (entry->state != MDB_PERMANENT && entry->state != MDB_TEMPORARY)
entry             546 net/bridge/br_mdb.c 	if (entry->vid >= VLAN_VID_MASK)
entry             556 net/bridge/br_mdb.c 	struct br_mdb_entry *entry;
entry             592 net/bridge/br_mdb.c 	entry = nla_data(tb[MDBA_SET_ENTRY]);
entry             593 net/bridge/br_mdb.c 	if (!is_valid_mdb_entry(entry)) {
entry             598 net/bridge/br_mdb.c 	*pentry = entry;
entry             652 net/bridge/br_mdb.c 			struct br_mdb_entry *entry)
entry             662 net/bridge/br_mdb.c 	if (entry->ifindex != br->dev->ifindex) {
entry             663 net/bridge/br_mdb.c 		dev = __dev_get_by_index(net, entry->ifindex);
entry             672 net/bridge/br_mdb.c 	__mdb_entry_to_br_ip(entry, &ip);
entry             675 net/bridge/br_mdb.c 	ret = br_mdb_add_group(br, p, &ip, entry->state);
entry             687 net/bridge/br_mdb.c 	struct br_mdb_entry *entry;
entry             692 net/bridge/br_mdb.c 	err = br_mdb_parse(skb, nlh, &dev, &entry);
entry             698 net/bridge/br_mdb.c 	if (entry->ifindex != br->dev->ifindex) {
entry             699 net/bridge/br_mdb.c 		pdev = __dev_get_by_index(net, entry->ifindex);
entry             714 net/bridge/br_mdb.c 	if (br_vlan_enabled(br->dev) && vg && entry->vid == 0) {
entry             716 net/bridge/br_mdb.c 			entry->vid = v->vid;
entry             717 net/bridge/br_mdb.c 			err = __br_mdb_add(net, br, entry);
entry             720 net/bridge/br_mdb.c 			__br_mdb_notify(dev, p, entry, RTM_NEWMDB);
entry             723 net/bridge/br_mdb.c 		err = __br_mdb_add(net, br, entry);
entry             725 net/bridge/br_mdb.c 			__br_mdb_notify(dev, p, entry, RTM_NEWMDB);
entry             731 net/bridge/br_mdb.c static int __br_mdb_del(struct net_bridge *br, struct br_mdb_entry *entry)
entry             742 net/bridge/br_mdb.c 	__mdb_entry_to_br_ip(entry, &ip);
entry             750 net/bridge/br_mdb.c 	if (entry->ifindex == mp->br->dev->ifindex && mp->host_joined) {
entry             761 net/bridge/br_mdb.c 		if (!p->port || p->port->dev->ifindex != entry->ifindex)
entry             767 net/bridge/br_mdb.c 		__mdb_entry_fill_flags(entry, p->flags);
entry             792 net/bridge/br_mdb.c 	struct br_mdb_entry *entry;
entry             797 net/bridge/br_mdb.c 	err = br_mdb_parse(skb, nlh, &dev, &entry);
entry             803 net/bridge/br_mdb.c 	if (entry->ifindex != br->dev->ifindex) {
entry             804 net/bridge/br_mdb.c 		pdev = __dev_get_by_index(net, entry->ifindex);
entry             819 net/bridge/br_mdb.c 	if (br_vlan_enabled(br->dev) && vg && entry->vid == 0) {
entry             821 net/bridge/br_mdb.c 			entry->vid = v->vid;
entry             822 net/bridge/br_mdb.c 			err = __br_mdb_del(br, entry);
entry             824 net/bridge/br_mdb.c 				__br_mdb_notify(dev, p, entry, RTM_DELMDB);
entry             827 net/bridge/br_mdb.c 		err = __br_mdb_del(br, entry);
entry             829 net/bridge/br_mdb.c 			__br_mdb_notify(dev, p, entry, RTM_DELMDB);
entry            2192 net/bridge/br_multicast.c 	struct br_ip_list *entry;
entry            2210 net/bridge/br_multicast.c 			entry = kmalloc(sizeof(*entry), GFP_ATOMIC);
entry            2211 net/bridge/br_multicast.c 			if (!entry)
entry            2214 net/bridge/br_multicast.c 			entry->addr = group->addr;
entry            2215 net/bridge/br_multicast.c 			list_add(&entry->list, br_ip_list);
entry             100 net/bridge/netfilter/ebtables.c ebt_dev_check(const char *entry, const struct net_device *device)
entry             105 net/bridge/netfilter/ebtables.c 	if (*entry == '\0')
entry             111 net/bridge/netfilter/ebtables.c 	while (entry[i] != '\0' && entry[i] != 1 && entry[i] == devname[i])
entry             113 net/bridge/netfilter/ebtables.c 	return devname[i] != entry[i] && entry[i] != 1;
entry             167 net/bridge/netfilter/ebtables.c struct ebt_entry *ebt_next_entry(const struct ebt_entry *entry)
entry             169 net/bridge/netfilter/ebtables.c 	return (void *)entry + entry->next_offset;
entry            2042 net/bridge/netfilter/ebtables.c static int size_entry_mwt(const struct ebt_entry *entry, const unsigned char *base,
entry            2056 net/bridge/netfilter/ebtables.c 	if (!entry->bitmask) {
entry            2058 net/bridge/netfilter/ebtables.c 		return ebt_buf_add(state, entry, sizeof(struct ebt_entries));
entry            2060 net/bridge/netfilter/ebtables.c 	if (*total < sizeof(*entry) || entry->next_offset < sizeof(*entry))
entry            2065 net/bridge/netfilter/ebtables.c 	ret = ebt_buf_add(state, entry,
entry            2071 net/bridge/netfilter/ebtables.c 	memcpy(&offsets[1], &entry->watchers_offset,
entry            2082 net/bridge/netfilter/ebtables.c 	buf_start = (char *) entry;
entry            2133 net/bridge/netfilter/ebtables.c 	if (next_expected_off != entry->next_offset)
entry            2136 net/bridge/netfilter/ebtables.c 	if (*total < entry->next_offset)
entry            2138 net/bridge/netfilter/ebtables.c 	*total -= entry->next_offset;
entry            2024 net/core/devlink.c 				   struct devlink_dpipe_entry *entry)
entry            2033 net/core/devlink.c 	if (nla_put_u64_64bit(skb, DEVLINK_ATTR_DPIPE_ENTRY_INDEX, entry->index,
entry            2036 net/core/devlink.c 	if (entry->counter_valid)
entry            2038 net/core/devlink.c 				      entry->counter, DEVLINK_ATTR_PAD))
entry            2046 net/core/devlink.c 	err = devlink_dpipe_match_values_put(skb, entry->match_values,
entry            2047 net/core/devlink.c 					     entry->match_values_count);
entry            2059 net/core/devlink.c 	err = devlink_dpipe_action_values_put(skb, entry->action_values,
entry            2060 net/core/devlink.c 					      entry->action_values_count);
entry            2125 net/core/devlink.c 				   struct devlink_dpipe_entry *entry)
entry            2127 net/core/devlink.c 	return devlink_dpipe_entry_put(dump_ctx->skb, entry);
entry            2139 net/core/devlink.c void devlink_dpipe_entry_clear(struct devlink_dpipe_entry *entry)
entry            2145 net/core/devlink.c 	value = entry->action_values;
entry            2146 net/core/devlink.c 	value_count = entry->action_values_count;
entry            2152 net/core/devlink.c 	value = entry->match_values;
entry            2153 net/core/devlink.c 	value_count = entry->match_values_count;
entry             411 net/core/flow_dissector.c 	u32 entry, label;
entry             423 net/core/flow_dissector.c 	entry = ntohl(hdr[0].entry);
entry             424 net/core/flow_dissector.c 	label = (entry & MPLS_LS_LABEL_MASK) >> MPLS_LS_LABEL_SHIFT;
entry             433 net/core/flow_dissector.c 		key_mpls->mpls_ttl = (entry & MPLS_LS_TTL_MASK)
entry             435 net/core/flow_dissector.c 		key_mpls->mpls_tc = (entry & MPLS_LS_TC_MASK)
entry             437 net/core/flow_dissector.c 		key_mpls->mpls_bos = (entry & MPLS_LS_S_MASK)
entry             445 net/core/flow_dissector.c 		key_keyid->keyid = hdr[1].entry & htonl(MPLS_LS_LABEL_MASK);
entry             400 net/core/flow_offload.c 	struct flow_indr_block_entry *entry;
entry             403 net/core/flow_offload.c 	list_for_each_entry(entry, &block_cb_list, list) {
entry             404 net/core/flow_offload.c 		entry->cb(dev, cb, cb_priv, command);
entry             501 net/core/flow_offload.c void flow_indr_add_block_cb(struct flow_indr_block_entry *entry)
entry             504 net/core/flow_offload.c 	list_add_tail(&entry->list, &block_cb_list);
entry             509 net/core/flow_offload.c void flow_indr_del_block_cb(struct flow_indr_block_entry *entry)
entry             512 net/core/flow_offload.c 	list_del(&entry->list);
entry             261 net/core/pktgen.c 	struct proc_dir_entry *entry;	/* proc file */
entry            1925 net/core/pktgen.c 			proc_remove(pkt_dev->entry);
entry            1927 net/core/pktgen.c 			pkt_dev->entry = proc_create_data(dev->name, 0600,
entry            1931 net/core/pktgen.c 			if (!pkt_dev->entry)
entry            3641 net/core/pktgen.c 	pkt_dev->entry = proc_create_data(ifname, 0600, t->net->proc_dir,
entry            3643 net/core/pktgen.c 	if (!pkt_dev->entry) {
entry            3768 net/core/pktgen.c 	proc_remove(pkt_dev->entry);
entry            1807 net/dcb/dcbnl.c 	struct dcb_app_type *entry;
entry            1809 net/dcb/dcbnl.c 	entry = kmalloc(sizeof(*entry), GFP_ATOMIC);
entry            1810 net/dcb/dcbnl.c 	if (!entry)
entry            1813 net/dcb/dcbnl.c 	memcpy(&entry->app, app, sizeof(*app));
entry            1814 net/dcb/dcbnl.c 	entry->ifindex = ifindex;
entry            1815 net/dcb/dcbnl.c 	list_add(&entry->list, &dcb_app_list);
entry              51 net/dccp/ccids/lib/packet_history.c 	struct tfrc_tx_hist_entry *entry = kmem_cache_alloc(tfrc_tx_hist_slab, gfp_any());
entry              53 net/dccp/ccids/lib/packet_history.c 	if (entry == NULL)
entry              55 net/dccp/ccids/lib/packet_history.c 	entry->seqno = seqno;
entry              56 net/dccp/ccids/lib/packet_history.c 	entry->stamp = ktime_get_real();
entry              57 net/dccp/ccids/lib/packet_history.c 	entry->next  = *headp;
entry              58 net/dccp/ccids/lib/packet_history.c 	*headp	     = entry;
entry              97 net/dccp/ccids/lib/packet_history.c static inline void tfrc_rx_hist_entry_from_skb(struct tfrc_rx_hist_entry *entry,
entry             103 net/dccp/ccids/lib/packet_history.c 	entry->tfrchrx_seqno = DCCP_SKB_CB(skb)->dccpd_seq;
entry             104 net/dccp/ccids/lib/packet_history.c 	entry->tfrchrx_ccval = dh->dccph_ccval;
entry             105 net/dccp/ccids/lib/packet_history.c 	entry->tfrchrx_type  = dh->dccph_type;
entry             106 net/dccp/ccids/lib/packet_history.c 	entry->tfrchrx_ndp   = ndp;
entry             107 net/dccp/ccids/lib/packet_history.c 	entry->tfrchrx_tstamp = ktime_get_real();
entry             114 net/dccp/ccids/lib/packet_history.c 	struct tfrc_rx_hist_entry *entry = tfrc_rx_hist_last_rcv(h);
entry             116 net/dccp/ccids/lib/packet_history.c 	tfrc_rx_hist_entry_from_skb(entry, skb, ndp);
entry             277 net/dccp/feat.c static void dccp_feat_print_entry(struct dccp_feat_entry const *entry)
entry             279 net/dccp/feat.c 	dccp_debug("   * %s %s = ", entry->is_local ? "local" : "remote",
entry             280 net/dccp/feat.c 				    dccp_feat_fname(entry->feat_num));
entry             281 net/dccp/feat.c 	dccp_feat_printval(entry->feat_num, &entry->val);
entry             282 net/dccp/feat.c 	dccp_pr_debug_cat(", state=%s %s\n", dccp_feat_sname[entry->state],
entry             283 net/dccp/feat.c 			  entry->needs_confirm ? "(Confirm pending)" : "");
entry             409 net/dccp/feat.c static void dccp_feat_entry_destructor(struct dccp_feat_entry *entry)
entry             411 net/dccp/feat.c 	if (entry != NULL) {
entry             412 net/dccp/feat.c 		dccp_feat_val_destructor(entry->feat_num, &entry->val);
entry             413 net/dccp/feat.c 		kfree(entry);
entry             429 net/dccp/feat.c 	struct dccp_feat_entry *entry;
entry             431 net/dccp/feat.c 	list_for_each_entry(entry, fn_list, node) {
entry             432 net/dccp/feat.c 		if (entry->feat_num == feat_num && entry->is_local == is_local)
entry             433 net/dccp/feat.c 			return entry;
entry             434 net/dccp/feat.c 		else if (entry->feat_num > feat_num)
entry             451 net/dccp/feat.c 	struct dccp_feat_entry *entry;
entry             453 net/dccp/feat.c 	list_for_each_entry(entry, head, node)
entry             454 net/dccp/feat.c 		if (entry->feat_num == feat && entry->is_local == local) {
entry             455 net/dccp/feat.c 			dccp_feat_val_destructor(entry->feat_num, &entry->val);
entry             456 net/dccp/feat.c 			return entry;
entry             457 net/dccp/feat.c 		} else if (entry->feat_num > feat) {
entry             458 net/dccp/feat.c 			head = &entry->node;
entry             462 net/dccp/feat.c 	entry = kmalloc(sizeof(*entry), gfp_any());
entry             463 net/dccp/feat.c 	if (entry != NULL) {
entry             464 net/dccp/feat.c 		entry->feat_num = feat;
entry             465 net/dccp/feat.c 		entry->is_local = local;
entry             466 net/dccp/feat.c 		list_add_tail(&entry->node, head);
entry             468 net/dccp/feat.c 	return entry;
entry             533 net/dccp/feat.c static inline void dccp_feat_list_pop(struct dccp_feat_entry *entry)
entry             535 net/dccp/feat.c 	list_del(&entry->node);
entry             536 net/dccp/feat.c 	dccp_feat_entry_destructor(entry);
entry             541 net/dccp/feat.c 	struct dccp_feat_entry *entry, *next;
entry             543 net/dccp/feat.c 	list_for_each_entry_safe(entry, next, fn_list, node)
entry             544 net/dccp/feat.c 		dccp_feat_entry_destructor(entry);
entry             552 net/dccp/feat.c 	struct dccp_feat_entry *entry, *new;
entry             555 net/dccp/feat.c 	list_for_each_entry(entry, from, node) {
entry             556 net/dccp/feat.c 		new = dccp_feat_clone_entry(entry);
entry             775 net/dccp/feat.c 		struct dccp_feat_entry *entry;
entry             777 net/dccp/feat.c 		entry = dccp_feat_list_lookup(&dp->dccps_featneg, feat, 1);
entry             778 net/dccp/feat.c 		if (entry != NULL)
entry             779 net/dccp/feat.c 			return entry->val.nn;
entry             805 net/dccp/feat.c 	struct dccp_feat_entry *entry;
entry             817 net/dccp/feat.c 	entry = dccp_feat_list_lookup(fn, feat, 1);
entry             818 net/dccp/feat.c 	if (entry != NULL) {
entry             820 net/dccp/feat.c 			      (unsigned long long)entry->val.nn,
entry             822 net/dccp/feat.c 		dccp_feat_list_pop(entry);
entry             971 net/dccp/feat.c 	struct dccp_feat_entry *entry;
entry             983 net/dccp/feat.c 	list_for_each_entry(entry, fn, node)
entry             984 net/dccp/feat.c 		if (entry->feat_num == DCCPF_CCID && entry->val.sp.len == 1)
entry             985 net/dccp/feat.c 			ccids[entry->is_local] = entry->val.sp.vec[0];
entry            1001 net/dccp/feat.c 	struct dccp_feat_entry *entry;
entry            1005 net/dccp/feat.c 		entry = dccp_feat_list_lookup(fn, DCCPF_CCID, is_local);
entry            1007 net/dccp/feat.c 		if (entry != NULL && !entry->empty_confirm)
entry            1008 net/dccp/feat.c 			ccid = entry->val.sp.vec[0];
entry            1102 net/dccp/feat.c 	struct dccp_feat_entry *entry;
entry            1129 net/dccp/feat.c 	entry = dccp_feat_list_lookup(fn, feat, local);
entry            1130 net/dccp/feat.c 	if (entry == NULL) {
entry            1161 net/dccp/feat.c 	} else if (entry->state == FEAT_UNSTABLE) {	/* 6.6.2 */
entry            1165 net/dccp/feat.c 	if (dccp_feat_reconcile(&entry->val, val, len, server, true)) {
entry            1166 net/dccp/feat.c 		entry->empty_confirm = false;
entry            1169 net/dccp/feat.c 	} else if (entry->state == FEAT_INITIALISING) {
entry            1180 net/dccp/feat.c 		if (!dccp_feat_reconcile(&entry->val, &defval, 1, server, true))
entry            1182 net/dccp/feat.c 		entry->empty_confirm = true;
entry            1184 net/dccp/feat.c 	entry->needs_confirm   = true;
entry            1185 net/dccp/feat.c 	entry->needs_mandatory = false;
entry            1186 net/dccp/feat.c 	entry->state	       = FEAT_STABLE;
entry            1213 net/dccp/feat.c 	struct dccp_feat_entry *entry = dccp_feat_list_lookup(fn, feat, local);
entry            1217 net/dccp/feat.c 	if (entry == NULL) {	/* nothing queued: ignore or handle error */
entry            1226 net/dccp/feat.c 	if (entry->state != FEAT_CHANGING)		/* 6.6.2 */
entry            1239 net/dccp/feat.c 		dccp_feat_list_pop(entry);
entry            1244 net/dccp/feat.c 		if (len > sizeof(entry->val.nn))
entry            1247 net/dccp/feat.c 		if (entry->val.nn == dccp_decode_value_var(val, len))
entry            1271 net/dccp/feat.c 	if (dccp_feat_reconcile(&entry->val, plist, plen, server, 0) != *val) {
entry            1275 net/dccp/feat.c 	entry->val.sp.vec[0] = *val;
entry            1278 net/dccp/feat.c 	entry->state = FEAT_STABLE;
entry            1310 net/dccp/feat.c 	struct dccp_feat_entry *entry;
entry            1347 net/dccp/feat.c 		entry = dccp_feat_list_lookup(fn, feat, local);
entry            1348 net/dccp/feat.c 		if (entry == NULL || entry->state != FEAT_CHANGING)
entry            1358 net/dccp/feat.c 		if (fval.nn != entry->val.nn)
entry            1365 net/dccp/feat.c 		dccp_feat_list_pop(entry);
entry              79 net/dccp/feat.h static inline u8 dccp_feat_genopt(struct dccp_feat_entry *entry)
entry              81 net/dccp/feat.h 	if (entry->needs_confirm)
entry              82 net/dccp/feat.h 		return entry->is_local ? DCCPO_CONFIRM_L : DCCPO_CONFIRM_R;
entry              83 net/dccp/feat.h 	return entry->is_local ? DCCPO_CHANGE_L : DCCPO_CHANGE_R;
entry             132 net/ipv4/cipso_ipv4.c static void cipso_v4_cache_entry_free(struct cipso_v4_map_cache_entry *entry)
entry             134 net/ipv4/cipso_ipv4.c 	if (entry->lsm_data)
entry             135 net/ipv4/cipso_ipv4.c 		netlbl_secattr_cache_free(entry->lsm_data);
entry             136 net/ipv4/cipso_ipv4.c 	kfree(entry->key);
entry             137 net/ipv4/cipso_ipv4.c 	kfree(entry);
entry             196 net/ipv4/cipso_ipv4.c 	struct cipso_v4_map_cache_entry *entry, *tmp_entry;
entry             201 net/ipv4/cipso_ipv4.c 		list_for_each_entry_safe(entry,
entry             204 net/ipv4/cipso_ipv4.c 			list_del(&entry->list);
entry             205 net/ipv4/cipso_ipv4.c 			cipso_v4_cache_entry_free(entry);
entry             239 net/ipv4/cipso_ipv4.c 	struct cipso_v4_map_cache_entry *entry;
entry             249 net/ipv4/cipso_ipv4.c 	list_for_each_entry(entry, &cipso_v4_cache[bkt].list, list) {
entry             250 net/ipv4/cipso_ipv4.c 		if (entry->hash == hash &&
entry             251 net/ipv4/cipso_ipv4.c 		    entry->key_len == key_len &&
entry             252 net/ipv4/cipso_ipv4.c 		    memcmp(entry->key, key, key_len) == 0) {
entry             253 net/ipv4/cipso_ipv4.c 			entry->activity += 1;
entry             254 net/ipv4/cipso_ipv4.c 			refcount_inc(&entry->lsm_data->refcount);
entry             255 net/ipv4/cipso_ipv4.c 			secattr->cache = entry->lsm_data;
entry             265 net/ipv4/cipso_ipv4.c 			if (entry->activity > prev_entry->activity &&
entry             266 net/ipv4/cipso_ipv4.c 			    entry->activity - prev_entry->activity >
entry             268 net/ipv4/cipso_ipv4.c 				__list_del(entry->list.prev, entry->list.next);
entry             269 net/ipv4/cipso_ipv4.c 				__list_add(&entry->list,
entry             277 net/ipv4/cipso_ipv4.c 		prev_entry = entry;
entry             302 net/ipv4/cipso_ipv4.c 	struct cipso_v4_map_cache_entry *entry = NULL;
entry             311 net/ipv4/cipso_ipv4.c 	entry = kzalloc(sizeof(*entry), GFP_ATOMIC);
entry             312 net/ipv4/cipso_ipv4.c 	if (!entry)
entry             314 net/ipv4/cipso_ipv4.c 	entry->key = kmemdup(cipso_ptr, cipso_ptr_len, GFP_ATOMIC);
entry             315 net/ipv4/cipso_ipv4.c 	if (!entry->key) {
entry             319 net/ipv4/cipso_ipv4.c 	entry->key_len = cipso_ptr_len;
entry             320 net/ipv4/cipso_ipv4.c 	entry->hash = cipso_v4_map_cache_hash(cipso_ptr, cipso_ptr_len);
entry             322 net/ipv4/cipso_ipv4.c 	entry->lsm_data = secattr->cache;
entry             324 net/ipv4/cipso_ipv4.c 	bkt = entry->hash & (CIPSO_V4_CACHE_BUCKETS - 1);
entry             327 net/ipv4/cipso_ipv4.c 		list_add(&entry->list, &cipso_v4_cache[bkt].list);
entry             333 net/ipv4/cipso_ipv4.c 		list_add(&entry->list, &cipso_v4_cache[bkt].list);
entry             341 net/ipv4/cipso_ipv4.c 	if (entry)
entry             342 net/ipv4/cipso_ipv4.c 		cipso_v4_cache_entry_free(entry);
entry             490 net/ipv4/cipso_ipv4.c static void cipso_v4_doi_free_rcu(struct rcu_head *entry)
entry             494 net/ipv4/cipso_ipv4.c 	doi_def = container_of(entry, struct cipso_v4_doi, rcu);
entry             551 net/ipv4/inet_diag.c 			    const struct inet_diag_entry *entry)
entry             567 net/ipv4/inet_diag.c 			yes = entry->sport == op[1].no;
entry             570 net/ipv4/inet_diag.c 			yes = entry->sport >= op[1].no;
entry             573 net/ipv4/inet_diag.c 			yes = entry->sport <= op[1].no;
entry             576 net/ipv4/inet_diag.c 			yes = entry->dport == op[1].no;
entry             579 net/ipv4/inet_diag.c 			yes = entry->dport >= op[1].no;
entry             582 net/ipv4/inet_diag.c 			yes = entry->dport <= op[1].no;
entry             585 net/ipv4/inet_diag.c 			yes = !(entry->userlocks & SOCK_BINDPORT_LOCK);
entry             595 net/ipv4/inet_diag.c 					     entry->sport : entry->dport)) {
entry             601 net/ipv4/inet_diag.c 				addr = entry->saddr;
entry             603 net/ipv4/inet_diag.c 				addr = entry->daddr;
entry             606 net/ipv4/inet_diag.c 			    cond->family != entry->family) {
entry             607 net/ipv4/inet_diag.c 				if (entry->family == AF_INET6 &&
entry             632 net/ipv4/inet_diag.c 			if (ifindex != entry->ifindex)
entry             640 net/ipv4/inet_diag.c 			if ((entry->mark & cond->mask) != cond->mark)
entry             659 net/ipv4/inet_diag.c static void entry_fill_addrs(struct inet_diag_entry *entry,
entry             664 net/ipv4/inet_diag.c 		entry->saddr = sk->sk_v6_rcv_saddr.s6_addr32;
entry             665 net/ipv4/inet_diag.c 		entry->daddr = sk->sk_v6_daddr.s6_addr32;
entry             669 net/ipv4/inet_diag.c 		entry->saddr = &sk->sk_rcv_saddr;
entry             670 net/ipv4/inet_diag.c 		entry->daddr = &sk->sk_daddr;
entry             677 net/ipv4/inet_diag.c 	struct inet_diag_entry entry;
entry             682 net/ipv4/inet_diag.c 	entry.family = sk->sk_family;
entry             683 net/ipv4/inet_diag.c 	entry_fill_addrs(&entry, sk);
entry             684 net/ipv4/inet_diag.c 	entry.sport = inet->inet_num;
entry             685 net/ipv4/inet_diag.c 	entry.dport = ntohs(inet->inet_dport);
entry             686 net/ipv4/inet_diag.c 	entry.ifindex = sk->sk_bound_dev_if;
entry             687 net/ipv4/inet_diag.c 	entry.userlocks = sk_fullsock(sk) ? sk->sk_userlocks : 0;
entry             689 net/ipv4/inet_diag.c 		entry.mark = sk->sk_mark;
entry             691 net/ipv4/inet_diag.c 		entry.mark = inet_rsk(inet_reqsk(sk))->ir_mark;
entry             693 net/ipv4/inet_diag.c 		entry.mark = 0;
entry             695 net/ipv4/inet_diag.c 	return inet_diag_bc_run(bc, &entry);
entry             177 net/ipv4/netfilter/arp_tables.c struct arpt_entry *arpt_next_entry(const struct arpt_entry *entry)
entry             179 net/ipv4/netfilter/arp_tables.c 	return (void *)entry + entry->next_offset;
entry             218 net/ipv4/netfilter/ip_tables.c struct ipt_entry *ipt_next_entry(const struct ipt_entry *entry)
entry             220 net/ipv4/netfilter/ip_tables.c 	return (void *)entry + entry->next_offset;
entry             151 net/ipv4/netfilter/ipt_CLUSTERIP.c clusterip_config_find_get(struct net *net, __be32 clusterip, int entry)
entry             165 net/ipv4/netfilter/ipt_CLUSTERIP.c 		else if (entry) {
entry            1102 net/ipv4/nexthop.c 	struct nexthop_grp *entry = nla_data(grps_attr);
entry            1103 net/ipv4/nexthop.c 	u16 num_nh = nla_len(grps_attr) / sizeof(*entry);
entry            1133 net/ipv4/nexthop.c 		nhe = nexthop_find_by_id(net, entry[i].id);
entry            1142 net/ipv4/nexthop.c 		nhg->nh_entries[i].weight = entry[i].weight + 1;
entry             411 net/ipv4/proc.c 		if (snmp4_tcp_list[i].entry == TCP_MIB_MAXCONN)
entry             472 net/ipv4/proc.c 					   snmp4_net_list[i].entry));
entry             482 net/ipv4/proc.c 					     snmp4_ipextstats_list[i].entry,
entry              98 net/ipv6/calipso.c static void calipso_cache_entry_free(struct calipso_map_cache_entry *entry)
entry             100 net/ipv6/calipso.c 	if (entry->lsm_data)
entry             101 net/ipv6/calipso.c 		netlbl_secattr_cache_free(entry->lsm_data);
entry             102 net/ipv6/calipso.c 	kfree(entry->key);
entry             103 net/ipv6/calipso.c 	kfree(entry);
entry             158 net/ipv6/calipso.c 	struct calipso_map_cache_entry *entry, *tmp_entry;
entry             163 net/ipv6/calipso.c 		list_for_each_entry_safe(entry,
entry             166 net/ipv6/calipso.c 			list_del(&entry->list);
entry             167 net/ipv6/calipso.c 			calipso_cache_entry_free(entry);
entry             201 net/ipv6/calipso.c 	struct calipso_map_cache_entry *entry;
entry             211 net/ipv6/calipso.c 	list_for_each_entry(entry, &calipso_cache[bkt].list, list) {
entry             212 net/ipv6/calipso.c 		if (entry->hash == hash &&
entry             213 net/ipv6/calipso.c 		    entry->key_len == key_len &&
entry             214 net/ipv6/calipso.c 		    memcmp(entry->key, key, key_len) == 0) {
entry             215 net/ipv6/calipso.c 			entry->activity += 1;
entry             216 net/ipv6/calipso.c 			refcount_inc(&entry->lsm_data->refcount);
entry             217 net/ipv6/calipso.c 			secattr->cache = entry->lsm_data;
entry             227 net/ipv6/calipso.c 			if (entry->activity > prev_entry->activity &&
entry             228 net/ipv6/calipso.c 			    entry->activity - prev_entry->activity >
entry             230 net/ipv6/calipso.c 				__list_del(entry->list.prev, entry->list.next);
entry             231 net/ipv6/calipso.c 				__list_add(&entry->list,
entry             239 net/ipv6/calipso.c 		prev_entry = entry;
entry             266 net/ipv6/calipso.c 	struct calipso_map_cache_entry *entry = NULL;
entry             275 net/ipv6/calipso.c 	entry = kzalloc(sizeof(*entry), GFP_ATOMIC);
entry             276 net/ipv6/calipso.c 	if (!entry)
entry             278 net/ipv6/calipso.c 	entry->key = kmemdup(calipso_ptr + 2, calipso_ptr_len, GFP_ATOMIC);
entry             279 net/ipv6/calipso.c 	if (!entry->key) {
entry             283 net/ipv6/calipso.c 	entry->key_len = calipso_ptr_len;
entry             284 net/ipv6/calipso.c 	entry->hash = calipso_map_cache_hash(calipso_ptr, calipso_ptr_len);
entry             286 net/ipv6/calipso.c 	entry->lsm_data = secattr->cache;
entry             288 net/ipv6/calipso.c 	bkt = entry->hash & (CALIPSO_CACHE_BUCKETS - 1);
entry             291 net/ipv6/calipso.c 		list_add(&entry->list, &calipso_cache[bkt].list);
entry             297 net/ipv6/calipso.c 		list_add(&entry->list, &calipso_cache[bkt].list);
entry             305 net/ipv6/calipso.c 	if (entry)
entry             306 net/ipv6/calipso.c 		calipso_cache_entry_free(entry);
entry             415 net/ipv6/calipso.c static void calipso_doi_free_rcu(struct rcu_head *entry)
entry             419 net/ipv6/calipso.c 	doi_def = container_of(entry, struct calipso_doi, rcu);
entry              78 net/ipv6/netfilter.c 			  const struct nf_queue_entry *entry)
entry              80 net/ipv6/netfilter.c 	struct ip6_rt_info *rt_info = nf_queue_entry_reroute(entry);
entry              82 net/ipv6/netfilter.c 	if (entry->state.hook == NF_INET_LOCAL_OUT) {
entry              87 net/ipv6/netfilter.c 			return ip6_route_me_harder(entry->state.net, skb);
entry             243 net/ipv6/netfilter/ip6_tables.c ip6t_next_entry(const struct ip6t_entry *entry)
entry             245 net/ipv6/netfilter/ip6_tables.c 	return (void *)entry + entry->next_offset;
entry             196 net/ipv6/proc.c 				   atomic_long_read(smib + itemlist[i].entry));
entry            2027 net/ipv6/route.c 			struct rt6_info *entry = rt6_ex->rt6i;
entry            2033 net/ipv6/route.c 			if (dst_metric_raw(&entry->dst, RTAX_MTU) &&
entry            2034 net/ipv6/route.c 			    rt6_mtu_change_route_allowed(idev, entry, mtu))
entry            2035 net/ipv6/route.c 				dst_metric_set(&entry->dst, RTAX_MTU, mtu);
entry            2060 net/ipv6/route.c 				struct rt6_info *entry = rt6_ex->rt6i;
entry            2062 net/ipv6/route.c 				if ((entry->rt6i_flags & RTF_CACHE_GATEWAY) ==
entry            2065 net/ipv6/route.c 						    &entry->rt6i_gateway)) {
entry              83 net/lapb/lapb_iface.c 	struct list_head *entry;
entry              86 net/lapb/lapb_iface.c 	list_for_each(entry, &lapb_list) {
entry              87 net/lapb/lapb_iface.c 		lapb = list_entry(entry, struct lapb_cb, node);
entry            2090 net/mac80211/rx.c 	struct ieee80211_fragment_entry *entry;
entry            2092 net/mac80211/rx.c 	entry = &sdata->fragments[sdata->fragment_next++];
entry            2096 net/mac80211/rx.c 	if (!skb_queue_empty(&entry->skb_list))
entry            2097 net/mac80211/rx.c 		__skb_queue_purge(&entry->skb_list);
entry            2099 net/mac80211/rx.c 	__skb_queue_tail(&entry->skb_list, *skb); /* no need for locking */
entry            2101 net/mac80211/rx.c 	entry->first_frag_time = jiffies;
entry            2102 net/mac80211/rx.c 	entry->seq = seq;
entry            2103 net/mac80211/rx.c 	entry->rx_queue = rx_queue;
entry            2104 net/mac80211/rx.c 	entry->last_frag = frag;
entry            2105 net/mac80211/rx.c 	entry->check_sequential_pn = false;
entry            2106 net/mac80211/rx.c 	entry->extra_len = 0;
entry            2108 net/mac80211/rx.c 	return entry;
entry            2116 net/mac80211/rx.c 	struct ieee80211_fragment_entry *entry;
entry            2128 net/mac80211/rx.c 		entry = &sdata->fragments[idx];
entry            2129 net/mac80211/rx.c 		if (skb_queue_empty(&entry->skb_list) || entry->seq != seq ||
entry            2130 net/mac80211/rx.c 		    entry->rx_queue != rx_queue ||
entry            2131 net/mac80211/rx.c 		    entry->last_frag + 1 != frag)
entry            2134 net/mac80211/rx.c 		f_skb = __skb_peek(&entry->skb_list);
entry            2146 net/mac80211/rx.c 		if (time_after(jiffies, entry->first_frag_time + 2 * HZ)) {
entry            2147 net/mac80211/rx.c 			__skb_queue_purge(&entry->skb_list);
entry            2150 net/mac80211/rx.c 		return entry;
entry            2163 net/mac80211/rx.c 	struct ieee80211_fragment_entry *entry;
entry            2198 net/mac80211/rx.c 		entry = ieee80211_reassemble_add(rx->sdata, frag, seq,
entry            2211 net/mac80211/rx.c 			entry->check_sequential_pn = true;
entry            2212 net/mac80211/rx.c 			memcpy(entry->last_pn,
entry            2230 net/mac80211/rx.c 	entry = ieee80211_reassemble_find(rx->sdata, frag, seq,
entry            2232 net/mac80211/rx.c 	if (!entry) {
entry            2242 net/mac80211/rx.c 	if (entry->check_sequential_pn) {
entry            2253 net/mac80211/rx.c 		memcpy(pn, entry->last_pn, IEEE80211_CCMP_PN_LEN);
entry            2263 net/mac80211/rx.c 		memcpy(entry->last_pn, pn, IEEE80211_CCMP_PN_LEN);
entry            2267 net/mac80211/rx.c 	__skb_queue_tail(&entry->skb_list, rx->skb);
entry            2268 net/mac80211/rx.c 	entry->last_frag = frag;
entry            2269 net/mac80211/rx.c 	entry->extra_len += rx->skb->len;
entry            2275 net/mac80211/rx.c 	rx->skb = __skb_dequeue(&entry->skb_list);
entry            2276 net/mac80211/rx.c 	if (skb_tailroom(rx->skb) < entry->extra_len) {
entry            2278 net/mac80211/rx.c 		if (unlikely(pskb_expand_head(rx->skb, 0, entry->extra_len,
entry            2281 net/mac80211/rx.c 			__skb_queue_purge(&entry->skb_list);
entry            2285 net/mac80211/rx.c 	while ((skb = __skb_dequeue(&entry->skb_list))) {
entry             353 net/mac802154/llsec.c 	struct mac802154_llsec_device *entry;
entry             364 net/mac802154/llsec.c 	entry = kmalloc(sizeof(*entry), GFP_KERNEL);
entry             365 net/mac802154/llsec.c 	if (!entry)
entry             368 net/mac802154/llsec.c 	entry->dev = *dev;
entry             369 net/mac802154/llsec.c 	spin_lock_init(&entry->lock);
entry             370 net/mac802154/llsec.c 	INIT_LIST_HEAD(&entry->dev.keys);
entry             373 net/mac802154/llsec.c 		hash_add_rcu(sec->devices_short, &entry->bucket_s, skey);
entry             375 net/mac802154/llsec.c 		INIT_HLIST_NODE(&entry->bucket_s);
entry             377 net/mac802154/llsec.c 	hash_add_rcu(sec->devices_hw, &entry->bucket_hw, hwkey);
entry             378 net/mac802154/llsec.c 	list_add_tail_rcu(&entry->dev.list, &sec->table.devices);
entry             490 net/mac802154/llsec.c 	struct mac802154_llsec_seclevel *entry;
entry             495 net/mac802154/llsec.c 	entry = kmalloc(sizeof(*entry), GFP_KERNEL);
entry             496 net/mac802154/llsec.c 	if (!entry)
entry             499 net/mac802154/llsec.c 	entry->level = *sl;
entry             501 net/mac802154/llsec.c 	list_add_tail_rcu(&entry->level.list, &sec->table.security_levels);
entry             189 net/mpls/internal.h 	unsigned entry = be32_to_cpu(hdr->label_stack_entry);
entry             191 net/mpls/internal.h 	result.label = (entry & MPLS_LS_LABEL_MASK) >> MPLS_LS_LABEL_SHIFT;
entry             192 net/mpls/internal.h 	result.ttl = (entry & MPLS_LS_TTL_MASK) >> MPLS_LS_TTL_SHIFT;
entry             193 net/mpls/internal.h 	result.tc =  (entry & MPLS_LS_TC_MASK) >> MPLS_LS_TC_SHIFT;
entry             194 net/mpls/internal.h 	result.bos = (entry & MPLS_LS_S_MASK) >> MPLS_LS_S_SHIFT;
entry            2584 net/netfilter/ipvs/ip_vs_ctl.c 	struct ip_vs_service_entry entry;
entry            2595 net/netfilter/ipvs/ip_vs_ctl.c 			memset(&entry, 0, sizeof(entry));
entry            2596 net/netfilter/ipvs/ip_vs_ctl.c 			ip_vs_copy_service(&entry, svc);
entry            2598 net/netfilter/ipvs/ip_vs_ctl.c 					 &entry, sizeof(entry))) {
entry            2614 net/netfilter/ipvs/ip_vs_ctl.c 			memset(&entry, 0, sizeof(entry));
entry            2615 net/netfilter/ipvs/ip_vs_ctl.c 			ip_vs_copy_service(&entry, svc);
entry            2617 net/netfilter/ipvs/ip_vs_ctl.c 					 &entry, sizeof(entry))) {
entry            2647 net/netfilter/ipvs/ip_vs_ctl.c 		struct ip_vs_dest_entry entry;
entry            2650 net/netfilter/ipvs/ip_vs_ctl.c 		memset(&entry, 0, sizeof(entry));
entry            2661 net/netfilter/ipvs/ip_vs_ctl.c 			entry.addr = dest->addr.ip;
entry            2662 net/netfilter/ipvs/ip_vs_ctl.c 			entry.port = dest->port;
entry            2663 net/netfilter/ipvs/ip_vs_ctl.c 			entry.conn_flags = atomic_read(&dest->conn_flags);
entry            2664 net/netfilter/ipvs/ip_vs_ctl.c 			entry.weight = atomic_read(&dest->weight);
entry            2665 net/netfilter/ipvs/ip_vs_ctl.c 			entry.u_threshold = dest->u_threshold;
entry            2666 net/netfilter/ipvs/ip_vs_ctl.c 			entry.l_threshold = dest->l_threshold;
entry            2667 net/netfilter/ipvs/ip_vs_ctl.c 			entry.activeconns = atomic_read(&dest->activeconns);
entry            2668 net/netfilter/ipvs/ip_vs_ctl.c 			entry.inactconns = atomic_read(&dest->inactconns);
entry            2669 net/netfilter/ipvs/ip_vs_ctl.c 			entry.persistconns = atomic_read(&dest->persistconns);
entry            2671 net/netfilter/ipvs/ip_vs_ctl.c 			ip_vs_export_stats_user(&entry.stats, &kstats);
entry            2673 net/netfilter/ipvs/ip_vs_ctl.c 					 &entry, sizeof(entry))) {
entry            2823 net/netfilter/ipvs/ip_vs_ctl.c 		struct ip_vs_service_entry *entry;
entry            2827 net/netfilter/ipvs/ip_vs_ctl.c 		entry = (struct ip_vs_service_entry *)arg;
entry            2828 net/netfilter/ipvs/ip_vs_ctl.c 		addr.ip = entry->addr;
entry            2830 net/netfilter/ipvs/ip_vs_ctl.c 		if (entry->fwmark)
entry            2831 net/netfilter/ipvs/ip_vs_ctl.c 			svc = __ip_vs_svc_fwm_find(ipvs, AF_INET, entry->fwmark);
entry            2834 net/netfilter/ipvs/ip_vs_ctl.c 						   entry->protocol, &addr,
entry            2835 net/netfilter/ipvs/ip_vs_ctl.c 						   entry->port);
entry            2838 net/netfilter/ipvs/ip_vs_ctl.c 			ip_vs_copy_service(entry, svc);
entry            2839 net/netfilter/ipvs/ip_vs_ctl.c 			if (copy_to_user(user, entry, sizeof(*entry)) != 0)
entry              58 net/netfilter/nf_dup_netdev.c 	struct flow_action_entry *entry;
entry              66 net/netfilter/nf_dup_netdev.c 	entry = &flow->rule->action.entries[ctx->num_actions++];
entry              67 net/netfilter/nf_dup_netdev.c 	entry->id = id;
entry              68 net/netfilter/nf_dup_netdev.c 	entry->dev = dev;
entry              63 net/netfilter/nf_flow_table_core.c 	struct flow_offload_entry *entry;
entry              70 net/netfilter/nf_flow_table_core.c 	entry = kzalloc(sizeof(*entry), GFP_ATOMIC);
entry              71 net/netfilter/nf_flow_table_core.c 	if (!entry)
entry              74 net/netfilter/nf_flow_table_core.c 	flow = &entry->flow;
entry              82 net/netfilter/nf_flow_table_core.c 	entry->ct = ct;
entry              97 net/netfilter/nf_flow_table_core.c 	kfree(entry);
entry              67 net/netfilter/nf_queue.c void nf_queue_entry_release_refs(struct nf_queue_entry *entry)
entry              69 net/netfilter/nf_queue.c 	struct nf_hook_state *state = &entry->state;
entry              79 net/netfilter/nf_queue.c 	nf_queue_entry_release_br_nf_refs(entry->skb);
entry             102 net/netfilter/nf_queue.c void nf_queue_entry_get_refs(struct nf_queue_entry *entry)
entry             104 net/netfilter/nf_queue.c 	struct nf_hook_state *state = &entry->state;
entry             113 net/netfilter/nf_queue.c 	nf_queue_entry_get_br_nf_refs(entry->skb);
entry             130 net/netfilter/nf_queue.c 			    struct nf_queue_entry *entry)
entry             132 net/netfilter/nf_queue.c 	struct ip_rt_info *rt_info = nf_queue_entry_reroute(entry);
entry             134 net/netfilter/nf_queue.c 	if (entry->state.hook == NF_INET_LOCAL_OUT) {
entry             145 net/netfilter/nf_queue.c 			     struct nf_queue_entry *entry)
entry             147 net/netfilter/nf_queue.c 	struct ip6_rt_info *rt_info = nf_queue_entry_reroute(entry);
entry             149 net/netfilter/nf_queue.c 	if (entry->state.hook == NF_INET_LOCAL_OUT) {
entry             162 net/netfilter/nf_queue.c 	struct nf_queue_entry *entry = NULL;
entry             186 net/netfilter/nf_queue.c 	entry = kmalloc(sizeof(*entry) + route_key_size, GFP_ATOMIC);
entry             187 net/netfilter/nf_queue.c 	if (!entry) {
entry             197 net/netfilter/nf_queue.c 	*entry = (struct nf_queue_entry) {
entry             201 net/netfilter/nf_queue.c 		.size	= sizeof(*entry) + route_key_size,
entry             204 net/netfilter/nf_queue.c 	nf_queue_entry_get_refs(entry);
entry             206 net/netfilter/nf_queue.c 	switch (entry->state.pf) {
entry             208 net/netfilter/nf_queue.c 		nf_ip_saveroute(skb, entry);
entry             211 net/netfilter/nf_queue.c 		nf_ip6_saveroute(skb, entry);
entry             215 net/netfilter/nf_queue.c 	status = qh->outfn(entry, queuenum);
entry             218 net/netfilter/nf_queue.c 		nf_queue_entry_release_refs(entry);
entry             225 net/netfilter/nf_queue.c 	kfree(entry);
entry             292 net/netfilter/nf_queue.c void nf_reinject(struct nf_queue_entry *entry, unsigned int verdict)
entry             296 net/netfilter/nf_queue.c 	struct sk_buff *skb = entry->skb;
entry             302 net/netfilter/nf_queue.c 	net = entry->state.net;
entry             303 net/netfilter/nf_queue.c 	pf = entry->state.pf;
entry             305 net/netfilter/nf_queue.c 	hooks = nf_hook_entries_head(net, pf, entry->state.hook);
entry             307 net/netfilter/nf_queue.c 	nf_queue_entry_release_refs(entry);
entry             309 net/netfilter/nf_queue.c 	i = entry->hook_index;
entry             312 net/netfilter/nf_queue.c 		kfree(entry);
entry             320 net/netfilter/nf_queue.c 		verdict = nf_hook_entry_hookfn(hook_entry, skb, &entry->state);
entry             323 net/netfilter/nf_queue.c 		if (nf_reroute(skb, entry) < 0)
entry             330 net/netfilter/nf_queue.c 		verdict = nf_iterate(skb, &entry->state, hooks, &i);
entry             337 net/netfilter/nf_queue.c 		entry->state.okfn(entry->state.net, entry->state.sk, skb);
entry             341 net/netfilter/nf_queue.c 		err = nf_queue(skb, &entry->state, i, verdict);
entry             351 net/netfilter/nf_queue.c 	kfree(entry);
entry              88 net/netfilter/nf_tables_offload.c 	struct flow_action_entry *entry;
entry              91 net/netfilter/nf_tables_offload.c 	flow_action_for_each(i, entry, &flow->rule->action) {
entry              92 net/netfilter/nf_tables_offload.c 		switch (entry->id) {
entry              95 net/netfilter/nf_tables_offload.c 			dev_put(entry->dev);
entry             191 net/netfilter/nfnetlink_queue.c __enqueue_entry(struct nfqnl_instance *queue, struct nf_queue_entry *entry)
entry             193 net/netfilter/nfnetlink_queue.c        list_add_tail(&entry->list, &queue->queue_list);
entry             198 net/netfilter/nfnetlink_queue.c __dequeue_entry(struct nfqnl_instance *queue, struct nf_queue_entry *entry)
entry             200 net/netfilter/nfnetlink_queue.c 	list_del(&entry->list);
entry             207 net/netfilter/nfnetlink_queue.c 	struct nf_queue_entry *entry = NULL, *i;
entry             213 net/netfilter/nfnetlink_queue.c 			entry = i;
entry             218 net/netfilter/nfnetlink_queue.c 	if (entry)
entry             219 net/netfilter/nfnetlink_queue.c 		__dequeue_entry(queue, entry);
entry             223 net/netfilter/nfnetlink_queue.c 	return entry;
entry             226 net/netfilter/nfnetlink_queue.c static void nfqnl_reinject(struct nf_queue_entry *entry, unsigned int verdict)
entry             237 net/netfilter/nfnetlink_queue.c 			err = ct_hook->update(entry->state.net, entry->skb);
entry             243 net/netfilter/nfnetlink_queue.c 	nf_reinject(entry, verdict);
entry             249 net/netfilter/nfnetlink_queue.c 	struct nf_queue_entry *entry, *next;
entry             252 net/netfilter/nfnetlink_queue.c 	list_for_each_entry_safe(entry, next, &queue->queue_list, list) {
entry             253 net/netfilter/nfnetlink_queue.c 		if (!cmpfn || cmpfn(entry, data)) {
entry             254 net/netfilter/nfnetlink_queue.c 			list_del(&entry->list);
entry             256 net/netfilter/nfnetlink_queue.c 			nfqnl_reinject(entry, NF_DROP);
entry             321 net/netfilter/nfnetlink_queue.c static u32 nfqnl_get_bridge_size(struct nf_queue_entry *entry)
entry             323 net/netfilter/nfnetlink_queue.c 	struct sk_buff *entskb = entry->skb;
entry             326 net/netfilter/nfnetlink_queue.c 	if (entry->state.pf != PF_BRIDGE || !skb_mac_header_was_set(entskb))
entry             340 net/netfilter/nfnetlink_queue.c static int nfqnl_put_bridge(struct nf_queue_entry *entry, struct sk_buff *skb)
entry             342 net/netfilter/nfnetlink_queue.c 	struct sk_buff *entskb = entry->skb;
entry             344 net/netfilter/nfnetlink_queue.c 	if (entry->state.pf != PF_BRIDGE || !skb_mac_header_was_set(entskb))
entry             376 net/netfilter/nfnetlink_queue.c 			   struct nf_queue_entry *entry,
entry             387 net/netfilter/nfnetlink_queue.c 	struct sk_buff *entskb = entry->skb;
entry             413 net/netfilter/nfnetlink_queue.c 	size += nfqnl_get_bridge_size(entry);
entry             415 net/netfilter/nfnetlink_queue.c 	if (entry->state.hook <= NF_INET_FORWARD ||
entry             416 net/netfilter/nfnetlink_queue.c 	   (entry->state.hook == NF_INET_POST_ROUTING && entskb->sk == NULL))
entry             421 net/netfilter/nfnetlink_queue.c 	outdev = entry->state.out;
entry             481 net/netfilter/nfnetlink_queue.c 	nfmsg->nfgen_family = entry->state.pf;
entry             488 net/netfilter/nfnetlink_queue.c 	pmsg->hook		= entry->state.hook;
entry             491 net/netfilter/nfnetlink_queue.c 	indev = entry->state.in;
entry             497 net/netfilter/nfnetlink_queue.c 		if (entry->state.pf == PF_BRIDGE) {
entry             531 net/netfilter/nfnetlink_queue.c 		if (entry->state.pf == PF_BRIDGE) {
entry             578 net/netfilter/nfnetlink_queue.c 	if (nfqnl_put_bridge(entry, skb) < 0)
entry             581 net/netfilter/nfnetlink_queue.c 	if (entry->state.hook <= NF_INET_FORWARD && entskb->tstamp) {
entry             638 net/netfilter/nfnetlink_queue.c static bool nf_ct_drop_unconfirmed(const struct nf_queue_entry *entry)
entry             642 net/netfilter/nfnetlink_queue.c 	const struct nf_conn *ct = (void *)skb_nfct(entry->skb);
entry             652 net/netfilter/nfnetlink_queue.c 			struct nf_queue_entry *entry)
entry             659 net/netfilter/nfnetlink_queue.c 	nskb = nfqnl_build_packet_message(net, queue, entry, &packet_id_ptr);
entry             666 net/netfilter/nfnetlink_queue.c 	if (nf_ct_drop_unconfirmed(entry))
entry             680 net/netfilter/nfnetlink_queue.c 	entry->id = ++queue->id_sequence;
entry             681 net/netfilter/nfnetlink_queue.c 	*packet_id_ptr = htonl(entry->id);
entry             695 net/netfilter/nfnetlink_queue.c 	__enqueue_entry(queue, entry);
entry             705 net/netfilter/nfnetlink_queue.c 		nfqnl_reinject(entry, NF_ACCEPT);
entry             713 net/netfilter/nfnetlink_queue.c 	struct nf_queue_entry *entry = kmemdup(e, e->size, GFP_ATOMIC);
entry             714 net/netfilter/nfnetlink_queue.c 	if (entry)
entry             715 net/netfilter/nfnetlink_queue.c 		nf_queue_entry_get_refs(entry);
entry             716 net/netfilter/nfnetlink_queue.c 	return entry;
entry             740 net/netfilter/nfnetlink_queue.c static void free_entry(struct nf_queue_entry *entry)
entry             742 net/netfilter/nfnetlink_queue.c 	nf_queue_entry_release_refs(entry);
entry             743 net/netfilter/nfnetlink_queue.c 	kfree(entry);
entry             748 net/netfilter/nfnetlink_queue.c 			   struct sk_buff *skb, struct nf_queue_entry *entry)
entry             756 net/netfilter/nfnetlink_queue.c 		struct sk_buff *gso_skb = entry->skb;
entry             757 net/netfilter/nfnetlink_queue.c 		entry->skb = skb;
entry             758 net/netfilter/nfnetlink_queue.c 		ret = __nfqnl_enqueue_packet(net, queue, entry);
entry             760 net/netfilter/nfnetlink_queue.c 			entry->skb = gso_skb;
entry             766 net/netfilter/nfnetlink_queue.c 	entry_seg = nf_queue_entry_dup(entry);
entry             777 net/netfilter/nfnetlink_queue.c nfqnl_enqueue_packet(struct nf_queue_entry *entry, unsigned int queuenum)
entry             783 net/netfilter/nfnetlink_queue.c 	struct net *net = entry->state.net;
entry             794 net/netfilter/nfnetlink_queue.c 	skb = entry->skb;
entry             796 net/netfilter/nfnetlink_queue.c 	switch (entry->state.pf) {
entry             806 net/netfilter/nfnetlink_queue.c 		return __nfqnl_enqueue_packet(net, queue, entry);
entry             822 net/netfilter/nfnetlink_queue.c 							segs, entry);
entry             832 net/netfilter/nfnetlink_queue.c 			free_entry(entry);
entry             901 net/netfilter/nfnetlink_queue.c dev_cmp(struct nf_queue_entry *entry, unsigned long ifindex)
entry             906 net/netfilter/nfnetlink_queue.c 	physinif = nf_bridge_get_physinif(entry->skb);
entry             907 net/netfilter/nfnetlink_queue.c 	physoutif = nf_bridge_get_physoutif(entry->skb);
entry             912 net/netfilter/nfnetlink_queue.c 	if (entry->state.in)
entry             913 net/netfilter/nfnetlink_queue.c 		if (entry->state.in->ifindex == ifindex)
entry             915 net/netfilter/nfnetlink_queue.c 	if (entry->state.out)
entry             916 net/netfilter/nfnetlink_queue.c 		if (entry->state.out->ifindex == ifindex)
entry            1066 net/netfilter/nfnetlink_queue.c 	struct nf_queue_entry *entry, *tmp;
entry            1088 net/netfilter/nfnetlink_queue.c 	list_for_each_entry_safe(entry, tmp, &queue->queue_list, list) {
entry            1089 net/netfilter/nfnetlink_queue.c 		if (nfq_id_after(entry->id, maxid))
entry            1091 net/netfilter/nfnetlink_queue.c 		__dequeue_entry(queue, entry);
entry            1092 net/netfilter/nfnetlink_queue.c 		list_add_tail(&entry->list, &batch_list);
entry            1100 net/netfilter/nfnetlink_queue.c 	list_for_each_entry_safe(entry, tmp, &batch_list, list) {
entry            1102 net/netfilter/nfnetlink_queue.c 			entry->skb->mark = ntohl(nla_get_be32(nfqa[NFQA_MARK]));
entry            1104 net/netfilter/nfnetlink_queue.c 		nfqnl_reinject(entry, verdict);
entry            1112 net/netfilter/nfnetlink_queue.c 				      struct nf_queue_entry *entry,
entry            1117 net/netfilter/nfnetlink_queue.c 	ct = nfnl_ct->get_ct(entry->skb, ctinfo);
entry            1126 net/netfilter/nfnetlink_queue.c 				      NETLINK_CB(entry->skb).portid,
entry            1131 net/netfilter/nfnetlink_queue.c static int nfqa_parse_bridge(struct nf_queue_entry *entry,
entry            1147 net/netfilter/nfnetlink_queue.c 		__vlan_hwaccel_put_tag(entry->skb,
entry            1153 net/netfilter/nfnetlink_queue.c 		int mac_header_len = entry->skb->network_header -
entry            1154 net/netfilter/nfnetlink_queue.c 			entry->skb->mac_header;
entry            1159 net/netfilter/nfnetlink_queue.c 			memcpy(skb_mac_header(entry->skb),
entry            1178 net/netfilter/nfnetlink_queue.c 	struct nf_queue_entry *entry;
entry            1196 net/netfilter/nfnetlink_queue.c 	entry = find_dequeue_entry(queue, ntohl(vhdr->id));
entry            1197 net/netfilter/nfnetlink_queue.c 	if (entry == NULL)
entry            1205 net/netfilter/nfnetlink_queue.c 			ct = nfqnl_ct_parse(nfnl_ct, nlh, nfqa, entry, &ctinfo);
entry            1208 net/netfilter/nfnetlink_queue.c 	if (entry->state.pf == PF_BRIDGE) {
entry            1209 net/netfilter/nfnetlink_queue.c 		err = nfqa_parse_bridge(entry, nfqa);
entry            1216 net/netfilter/nfnetlink_queue.c 		int diff = payload_len - entry->skb->len;
entry            1219 net/netfilter/nfnetlink_queue.c 				 payload_len, entry, diff) < 0)
entry            1223 net/netfilter/nfnetlink_queue.c 			nfnl_ct->seq_adjust(entry->skb, ct, ctinfo, diff);
entry            1227 net/netfilter/nfnetlink_queue.c 		entry->skb->mark = ntohl(nla_get_be32(nfqa[NFQA_MARK]));
entry            1229 net/netfilter/nfnetlink_queue.c 	nfqnl_reinject(entry, verdict);
entry             138 net/netfilter/nft_compat.c 			   union nft_entry *entry, u16 proto, bool inv)
entry             144 net/netfilter/nft_compat.c 		entry->e4.ip.proto = proto;
entry             145 net/netfilter/nft_compat.c 		entry->e4.ip.invflags = inv ? IPT_INV_PROTO : 0;
entry             149 net/netfilter/nft_compat.c 			entry->e6.ipv6.flags |= IP6T_F_PROTO;
entry             151 net/netfilter/nft_compat.c 		entry->e6.ipv6.proto = proto;
entry             152 net/netfilter/nft_compat.c 		entry->e6.ipv6.invflags = inv ? IP6T_INV_PROTO : 0;
entry             155 net/netfilter/nft_compat.c 		entry->ebt.ethproto = (__force __be16)proto;
entry             156 net/netfilter/nft_compat.c 		entry->ebt.invflags = inv ? EBT_IPROTO : 0;
entry             161 net/netfilter/nft_compat.c 	par->entryinfo	= entry;
entry             382 net/netfilter/nft_compat.c 			  union nft_entry *entry, u16 proto, bool inv)
entry             388 net/netfilter/nft_compat.c 		entry->e4.ip.proto = proto;
entry             389 net/netfilter/nft_compat.c 		entry->e4.ip.invflags = inv ? IPT_INV_PROTO : 0;
entry             393 net/netfilter/nft_compat.c 			entry->e6.ipv6.flags |= IP6T_F_PROTO;
entry             395 net/netfilter/nft_compat.c 		entry->e6.ipv6.proto = proto;
entry             396 net/netfilter/nft_compat.c 		entry->e6.ipv6.invflags = inv ? IP6T_INV_PROTO : 0;
entry             399 net/netfilter/nft_compat.c 		entry->ebt.ethproto = (__force __be16)proto;
entry             400 net/netfilter/nft_compat.c 		entry->ebt.invflags = inv ? EBT_IPROTO : 0;
entry             405 net/netfilter/nft_compat.c 	par->entryinfo	= entry;
entry             132 net/netfilter/nft_immediate.c 	struct flow_action_entry *entry;
entry             135 net/netfilter/nft_immediate.c 	entry = &flow->rule->action.entries[ctx->num_actions++];
entry             140 net/netfilter/nft_immediate.c 		entry->id = FLOW_ACTION_ACCEPT;
entry             143 net/netfilter/nft_immediate.c 		entry->id = FLOW_ACTION_DROP;
entry             182 net/netfilter/utils.c static int nf_ip_reroute(struct sk_buff *skb, const struct nf_queue_entry *entry)
entry             185 net/netfilter/utils.c 	const struct ip_rt_info *rt_info = nf_queue_entry_reroute(entry);
entry             187 net/netfilter/utils.c 	if (entry->state.hook == NF_INET_LOCAL_OUT) {
entry             194 net/netfilter/utils.c 			return ip_route_me_harder(entry->state.net, skb,
entry             201 net/netfilter/utils.c int nf_reroute(struct sk_buff *skb, struct nf_queue_entry *entry)
entry             206 net/netfilter/utils.c 	switch (entry->state.pf) {
entry             208 net/netfilter/utils.c 		ret = nf_ip_reroute(skb, entry);
entry             213 net/netfilter/utils.c 			ret = v6ops->reroute(skb, entry);
entry              32 net/netfilter/xt_IDLETIMER.c 	struct list_head entry;
entry              50 net/netfilter/xt_IDLETIMER.c 	struct idletimer_tg *entry;
entry              52 net/netfilter/xt_IDLETIMER.c 	list_for_each_entry(entry, &idletimer_tg_list, entry) {
entry              53 net/netfilter/xt_IDLETIMER.c 		if (!strcmp(label, entry->attr.attr.name))
entry              54 net/netfilter/xt_IDLETIMER.c 			return entry;
entry             143 net/netfilter/xt_IDLETIMER.c 	list_add(&info->timer->entry, &idletimer_tg_list);
entry             236 net/netfilter/xt_IDLETIMER.c 		list_del(&info->timer->entry);
entry             148 net/netlabel/netlabel_addrlist.c int netlbl_af4list_add(struct netlbl_af4list *entry, struct list_head *head)
entry             152 net/netlabel/netlabel_addrlist.c 	iter = netlbl_af4list_search(entry->addr, head);
entry             154 net/netlabel/netlabel_addrlist.c 	    iter->addr == entry->addr && iter->mask == entry->mask)
entry             163 net/netlabel/netlabel_addrlist.c 		    ntohl(entry->mask) > ntohl(iter->mask)) {
entry             164 net/netlabel/netlabel_addrlist.c 			__list_add_rcu(&entry->list,
entry             169 net/netlabel/netlabel_addrlist.c 	list_add_tail_rcu(&entry->list, head);
entry             185 net/netlabel/netlabel_addrlist.c int netlbl_af6list_add(struct netlbl_af6list *entry, struct list_head *head)
entry             189 net/netlabel/netlabel_addrlist.c 	iter = netlbl_af6list_search(&entry->addr, head);
entry             191 net/netlabel/netlabel_addrlist.c 	    ipv6_addr_equal(&iter->addr, &entry->addr) &&
entry             192 net/netlabel/netlabel_addrlist.c 	    ipv6_addr_equal(&iter->mask, &entry->mask))
entry             201 net/netlabel/netlabel_addrlist.c 		    ipv6_addr_cmp(&entry->mask, &iter->mask) > 0) {
entry             202 net/netlabel/netlabel_addrlist.c 			__list_add_rcu(&entry->list,
entry             207 net/netlabel/netlabel_addrlist.c 	list_add_tail_rcu(&entry->list, head);
entry             221 net/netlabel/netlabel_addrlist.c void netlbl_af4list_remove_entry(struct netlbl_af4list *entry)
entry             223 net/netlabel/netlabel_addrlist.c 	entry->valid = 0;
entry             224 net/netlabel/netlabel_addrlist.c 	list_del_rcu(&entry->list);
entry             242 net/netlabel/netlabel_addrlist.c 	struct netlbl_af4list *entry;
entry             244 net/netlabel/netlabel_addrlist.c 	entry = netlbl_af4list_search_exact(addr, mask, head);
entry             245 net/netlabel/netlabel_addrlist.c 	if (entry == NULL)
entry             247 net/netlabel/netlabel_addrlist.c 	netlbl_af4list_remove_entry(entry);
entry             248 net/netlabel/netlabel_addrlist.c 	return entry;
entry             261 net/netlabel/netlabel_addrlist.c void netlbl_af6list_remove_entry(struct netlbl_af6list *entry)
entry             263 net/netlabel/netlabel_addrlist.c 	entry->valid = 0;
entry             264 net/netlabel/netlabel_addrlist.c 	list_del_rcu(&entry->list);
entry             283 net/netlabel/netlabel_addrlist.c 	struct netlbl_af6list *entry;
entry             285 net/netlabel/netlabel_addrlist.c 	entry = netlbl_af6list_search_exact(addr, mask, head);
entry             286 net/netlabel/netlabel_addrlist.c 	if (entry == NULL)
entry             288 net/netlabel/netlabel_addrlist.c 	netlbl_af6list_remove_entry(entry);
entry             289 net/netlabel/netlabel_addrlist.c 	return entry;
entry              98 net/netlabel/netlabel_addrlist.h int netlbl_af4list_add(struct netlbl_af4list *entry,
entry             102 net/netlabel/netlabel_addrlist.h void netlbl_af4list_remove_entry(struct netlbl_af4list *entry);
entry             165 net/netlabel/netlabel_addrlist.h int netlbl_af6list_add(struct netlbl_af6list *entry,
entry             170 net/netlabel/netlabel_addrlist.h void netlbl_af6list_remove_entry(struct netlbl_af6list *entry);
entry             258 net/netlabel/netlabel_calipso.c static int netlbl_calipso_remove_cb(struct netlbl_dom_map *entry, void *arg)
entry             262 net/netlabel/netlabel_calipso.c 	if (entry->def.type == NETLBL_NLTYPE_CALIPSO &&
entry             263 net/netlabel/netlabel_calipso.c 	    entry->def.calipso->doi == cb_arg->doi)
entry             264 net/netlabel/netlabel_calipso.c 		return netlbl_domhsh_remove_entry(entry, cb_arg->audit_info);
entry             677 net/netlabel/netlabel_cipso_v4.c static int netlbl_cipsov4_remove_cb(struct netlbl_dom_map *entry, void *arg)
entry             681 net/netlabel/netlabel_cipso_v4.c 	if (entry->def.type == NETLBL_NLTYPE_CIPSOV4 &&
entry             682 net/netlabel/netlabel_cipso_v4.c 	    entry->def.cipso->doi == cb_arg->doi)
entry             683 net/netlabel/netlabel_cipso_v4.c 		return netlbl_domhsh_remove_entry(entry, cb_arg->audit_info);
entry             153 net/netlabel/netlabel_cipso_v4.h void netlbl_cipsov4_doi_free(struct rcu_head *entry);
entry              64 net/netlabel/netlabel_domainhash.c static void netlbl_domhsh_free_entry(struct rcu_head *entry)
entry              74 net/netlabel/netlabel_domainhash.c 	ptr = container_of(entry, struct netlbl_dom_map, rcu);
entry             173 net/netlabel/netlabel_domainhash.c 	struct netlbl_dom_map *entry;
entry             175 net/netlabel/netlabel_domainhash.c 	entry = netlbl_domhsh_search(domain, family);
entry             176 net/netlabel/netlabel_domainhash.c 	if (entry != NULL)
entry             177 net/netlabel/netlabel_domainhash.c 		return entry;
entry             179 net/netlabel/netlabel_domainhash.c 		entry = netlbl_domhsh_rcu_deref(netlbl_domhsh_def_ipv4);
entry             180 net/netlabel/netlabel_domainhash.c 		if (entry != NULL && entry->valid)
entry             181 net/netlabel/netlabel_domainhash.c 			return entry;
entry             184 net/netlabel/netlabel_domainhash.c 		entry = netlbl_domhsh_rcu_deref(netlbl_domhsh_def_ipv6);
entry             185 net/netlabel/netlabel_domainhash.c 		if (entry != NULL && entry->valid)
entry             186 net/netlabel/netlabel_domainhash.c 			return entry;
entry             206 net/netlabel/netlabel_domainhash.c static void netlbl_domhsh_audit_add(struct netlbl_dom_map *entry,
entry             220 net/netlabel/netlabel_domainhash.c 				 entry->domain ? entry->domain : "(default)");
entry             238 net/netlabel/netlabel_domainhash.c 			type = entry->def.type;
entry             239 net/netlabel/netlabel_domainhash.c 			cipsov4 = entry->def.cipso;
entry             240 net/netlabel/netlabel_domainhash.c 			calipso = entry->def.calipso;
entry             272 net/netlabel/netlabel_domainhash.c static int netlbl_domhsh_validate(const struct netlbl_dom_map *entry)
entry             281 net/netlabel/netlabel_domainhash.c 	if (entry == NULL)
entry             284 net/netlabel/netlabel_domainhash.c 	if (entry->family != AF_INET && entry->family != AF_INET6 &&
entry             285 net/netlabel/netlabel_domainhash.c 	    (entry->family != AF_UNSPEC ||
entry             286 net/netlabel/netlabel_domainhash.c 	     entry->def.type != NETLBL_NLTYPE_UNLABELED))
entry             289 net/netlabel/netlabel_domainhash.c 	switch (entry->def.type) {
entry             291 net/netlabel/netlabel_domainhash.c 		if (entry->def.cipso != NULL || entry->def.calipso != NULL ||
entry             292 net/netlabel/netlabel_domainhash.c 		    entry->def.addrsel != NULL)
entry             296 net/netlabel/netlabel_domainhash.c 		if (entry->family != AF_INET ||
entry             297 net/netlabel/netlabel_domainhash.c 		    entry->def.cipso == NULL)
entry             301 net/netlabel/netlabel_domainhash.c 		if (entry->family != AF_INET6 ||
entry             302 net/netlabel/netlabel_domainhash.c 		    entry->def.calipso == NULL)
entry             306 net/netlabel/netlabel_domainhash.c 		netlbl_af4list_foreach(iter4, &entry->def.addrsel->list4) {
entry             322 net/netlabel/netlabel_domainhash.c 		netlbl_af6list_foreach(iter6, &entry->def.addrsel->list6) {
entry             403 net/netlabel/netlabel_domainhash.c int netlbl_domhsh_add(struct netlbl_dom_map *entry,
entry             415 net/netlabel/netlabel_domainhash.c 	ret_val = netlbl_domhsh_validate(entry);
entry             425 net/netlabel/netlabel_domainhash.c 	if (entry->domain != NULL)
entry             426 net/netlabel/netlabel_domainhash.c 		entry_old = netlbl_domhsh_search(entry->domain, entry->family);
entry             428 net/netlabel/netlabel_domainhash.c 		entry_old = netlbl_domhsh_search_def(entry->domain,
entry             429 net/netlabel/netlabel_domainhash.c 						     entry->family);
entry             431 net/netlabel/netlabel_domainhash.c 		entry->valid = 1;
entry             433 net/netlabel/netlabel_domainhash.c 		if (entry->domain != NULL) {
entry             434 net/netlabel/netlabel_domainhash.c 			u32 bkt = netlbl_domhsh_hash(entry->domain);
entry             435 net/netlabel/netlabel_domainhash.c 			list_add_tail_rcu(&entry->list,
entry             438 net/netlabel/netlabel_domainhash.c 			INIT_LIST_HEAD(&entry->list);
entry             439 net/netlabel/netlabel_domainhash.c 			switch (entry->family) {
entry             442 net/netlabel/netlabel_domainhash.c 						   entry);
entry             446 net/netlabel/netlabel_domainhash.c 						   entry);
entry             449 net/netlabel/netlabel_domainhash.c 				if (entry->def.type !=
entry             462 net/netlabel/netlabel_domainhash.c 				entry->family = AF_INET;
entry             464 net/netlabel/netlabel_domainhash.c 						   entry);
entry             476 net/netlabel/netlabel_domainhash.c 		if (entry->def.type == NETLBL_NLTYPE_ADDRSELECT) {
entry             478 net/netlabel/netlabel_domainhash.c 						   &entry->def.addrsel->list4)
entry             479 net/netlabel/netlabel_domainhash.c 				netlbl_domhsh_audit_add(entry, iter4, NULL,
entry             483 net/netlabel/netlabel_domainhash.c 						   &entry->def.addrsel->list6)
entry             484 net/netlabel/netlabel_domainhash.c 				netlbl_domhsh_audit_add(entry, NULL, iter6,
entry             488 net/netlabel/netlabel_domainhash.c 			netlbl_domhsh_audit_add(entry, NULL, NULL,
entry             491 net/netlabel/netlabel_domainhash.c 		   entry->def.type == NETLBL_NLTYPE_ADDRSELECT) {
entry             500 net/netlabel/netlabel_domainhash.c 		netlbl_af4list_foreach_rcu(iter4, &entry->def.addrsel->list4)
entry             508 net/netlabel/netlabel_domainhash.c 		netlbl_af6list_foreach_rcu(iter6, &entry->def.addrsel->list6)
entry             518 net/netlabel/netlabel_domainhash.c 					    &entry->def.addrsel->list4) {
entry             529 net/netlabel/netlabel_domainhash.c 					    &entry->def.addrsel->list6) {
entry             559 net/netlabel/netlabel_domainhash.c int netlbl_domhsh_add_default(struct netlbl_dom_map *entry,
entry             562 net/netlabel/netlabel_domainhash.c 	return netlbl_domhsh_add(entry, audit_info);
entry             577 net/netlabel/netlabel_domainhash.c int netlbl_domhsh_remove_entry(struct netlbl_dom_map *entry,
entry             583 net/netlabel/netlabel_domainhash.c 	if (entry == NULL)
entry             587 net/netlabel/netlabel_domainhash.c 	if (entry->valid) {
entry             588 net/netlabel/netlabel_domainhash.c 		entry->valid = 0;
entry             589 net/netlabel/netlabel_domainhash.c 		if (entry == rcu_dereference(netlbl_domhsh_def_ipv4))
entry             591 net/netlabel/netlabel_domainhash.c 		else if (entry == rcu_dereference(netlbl_domhsh_def_ipv6))
entry             594 net/netlabel/netlabel_domainhash.c 			list_del_rcu(&entry->list);
entry             603 net/netlabel/netlabel_domainhash.c 				 entry->domain ? entry->domain : "(default)",
entry             616 net/netlabel/netlabel_domainhash.c 		switch (entry->def.type) {
entry             619 net/netlabel/netlabel_domainhash.c 					     &entry->def.addrsel->list4) {
entry             625 net/netlabel/netlabel_domainhash.c 					     &entry->def.addrsel->list6) {
entry             632 net/netlabel/netlabel_domainhash.c 			cipso_v4_doi_putdef(entry->def.cipso);
entry             636 net/netlabel/netlabel_domainhash.c 			calipso_doi_putdef(entry->def.calipso);
entry             640 net/netlabel/netlabel_domainhash.c 		call_rcu(&entry->rcu, netlbl_domhsh_free_entry);
entry             670 net/netlabel/netlabel_domainhash.c 	struct netlbl_domaddr4_map *entry;
entry             704 net/netlabel/netlabel_domainhash.c 	entry = netlbl_domhsh_addr4_entry(entry_addr);
entry             705 net/netlabel/netlabel_domainhash.c 	cipso_v4_doi_putdef(entry->def.cipso);
entry             706 net/netlabel/netlabel_domainhash.c 	kfree(entry);
entry             737 net/netlabel/netlabel_domainhash.c 	struct netlbl_domaddr6_map *entry;
entry             769 net/netlabel/netlabel_domainhash.c 	entry = netlbl_domhsh_addr6_entry(entry_addr);
entry             770 net/netlabel/netlabel_domainhash.c 	calipso_doi_putdef(entry->def.calipso);
entry             771 net/netlabel/netlabel_domainhash.c 	kfree(entry);
entry             797 net/netlabel/netlabel_domainhash.c 	struct netlbl_dom_map *entry;
entry             803 net/netlabel/netlabel_domainhash.c 			entry = netlbl_domhsh_search(domain, AF_INET);
entry             805 net/netlabel/netlabel_domainhash.c 			entry = netlbl_domhsh_search_def(domain, AF_INET);
entry             806 net/netlabel/netlabel_domainhash.c 		ret_val = netlbl_domhsh_remove_entry(entry, audit_info);
entry             814 net/netlabel/netlabel_domainhash.c 			entry = netlbl_domhsh_search(domain, AF_INET6);
entry             816 net/netlabel/netlabel_domainhash.c 			entry = netlbl_domhsh_search_def(domain, AF_INET6);
entry             817 net/netlabel/netlabel_domainhash.c 		ret_val2 = netlbl_domhsh_remove_entry(entry, audit_info);
entry             940 net/netlabel/netlabel_domainhash.c 		     int (*callback) (struct netlbl_dom_map *entry, void *arg),
entry              72 net/netlabel/netlabel_domainhash.h int netlbl_domhsh_add(struct netlbl_dom_map *entry,
entry              74 net/netlabel/netlabel_domainhash.h int netlbl_domhsh_add_default(struct netlbl_dom_map *entry,
entry              76 net/netlabel/netlabel_domainhash.h int netlbl_domhsh_remove_entry(struct netlbl_dom_map *entry,
entry             103 net/netlabel/netlabel_domainhash.h 		     int (*callback) (struct netlbl_dom_map *entry, void *arg),
entry             102 net/netlabel/netlabel_kapi.c 	struct netlbl_dom_map *entry;
entry             107 net/netlabel/netlabel_kapi.c 	entry = kzalloc(sizeof(*entry), GFP_ATOMIC);
entry             108 net/netlabel/netlabel_kapi.c 	if (entry == NULL)
entry             111 net/netlabel/netlabel_kapi.c 		entry->domain = kstrdup(domain, GFP_ATOMIC);
entry             112 net/netlabel/netlabel_kapi.c 		if (entry->domain == NULL)
entry             115 net/netlabel/netlabel_kapi.c 	entry->family = family;
entry             118 net/netlabel/netlabel_kapi.c 		entry->def.type = NETLBL_NLTYPE_UNLABELED;
entry             169 net/netlabel/netlabel_kapi.c 		entry->def.addrsel = addrmap;
entry             170 net/netlabel/netlabel_kapi.c 		entry->def.type = NETLBL_NLTYPE_ADDRSELECT;
entry             176 net/netlabel/netlabel_kapi.c 	ret_val = netlbl_domhsh_add(entry, audit_info);
entry             183 net/netlabel/netlabel_kapi.c 	kfree(entry->domain);
entry             184 net/netlabel/netlabel_kapi.c 	kfree(entry);
entry             331 net/netlabel/netlabel_kapi.c 	struct netlbl_dom_map *entry;
entry             339 net/netlabel/netlabel_kapi.c 	entry = kzalloc(sizeof(*entry), GFP_ATOMIC);
entry             340 net/netlabel/netlabel_kapi.c 	if (entry == NULL)
entry             342 net/netlabel/netlabel_kapi.c 	entry->family = AF_INET;
entry             344 net/netlabel/netlabel_kapi.c 		entry->domain = kstrdup(domain, GFP_ATOMIC);
entry             345 net/netlabel/netlabel_kapi.c 		if (entry->domain == NULL)
entry             350 net/netlabel/netlabel_kapi.c 		entry->def.cipso = doi_def;
entry             351 net/netlabel/netlabel_kapi.c 		entry->def.type = NETLBL_NLTYPE_CIPSOV4;
entry             371 net/netlabel/netlabel_kapi.c 		entry->def.addrsel = addrmap;
entry             372 net/netlabel/netlabel_kapi.c 		entry->def.type = NETLBL_NLTYPE_ADDRSELECT;
entry             378 net/netlabel/netlabel_kapi.c 	ret_val = netlbl_domhsh_add(entry, audit_info);
entry             389 net/netlabel/netlabel_kapi.c 	kfree(entry->domain);
entry             391 net/netlabel/netlabel_kapi.c 	kfree(entry);
entry             457 net/netlabel/netlabel_kapi.c 	struct netlbl_dom_map *entry;
entry             465 net/netlabel/netlabel_kapi.c 	entry = kzalloc(sizeof(*entry), GFP_ATOMIC);
entry             466 net/netlabel/netlabel_kapi.c 	if (entry == NULL)
entry             468 net/netlabel/netlabel_kapi.c 	entry->family = AF_INET6;
entry             470 net/netlabel/netlabel_kapi.c 		entry->domain = kstrdup(domain, GFP_ATOMIC);
entry             471 net/netlabel/netlabel_kapi.c 		if (entry->domain == NULL)
entry             476 net/netlabel/netlabel_kapi.c 		entry->def.calipso = doi_def;
entry             477 net/netlabel/netlabel_kapi.c 		entry->def.type = NETLBL_NLTYPE_CALIPSO;
entry             501 net/netlabel/netlabel_kapi.c 		entry->def.addrsel = addrmap;
entry             502 net/netlabel/netlabel_kapi.c 		entry->def.type = NETLBL_NLTYPE_ADDRSELECT;
entry             508 net/netlabel/netlabel_kapi.c 	ret_val = netlbl_domhsh_add(entry, audit_info);
entry             519 net/netlabel/netlabel_kapi.c 	kfree(entry->domain);
entry             521 net/netlabel/netlabel_kapi.c 	kfree(entry);
entry            1111 net/netlabel/netlabel_kapi.c 	struct netlbl_dommap_def *entry;
entry            1117 net/netlabel/netlabel_kapi.c 		entry = netlbl_domhsh_getentry_af4(secattr->domain,
entry            1119 net/netlabel/netlabel_kapi.c 		if (entry == NULL) {
entry            1123 net/netlabel/netlabel_kapi.c 		switch (entry->type) {
entry            1126 net/netlabel/netlabel_kapi.c 							entry->cipso, secattr);
entry            1141 net/netlabel/netlabel_kapi.c 		entry = netlbl_domhsh_getentry_af6(secattr->domain,
entry            1143 net/netlabel/netlabel_kapi.c 		if (entry == NULL) {
entry            1147 net/netlabel/netlabel_kapi.c 		switch (entry->type) {
entry            1150 net/netlabel/netlabel_kapi.c 						       entry->calipso, secattr);
entry            1186 net/netlabel/netlabel_kapi.c 	struct netlbl_dommap_def *entry;
entry            1192 net/netlabel/netlabel_kapi.c 		entry = netlbl_domhsh_getentry_af4(secattr->domain,
entry            1194 net/netlabel/netlabel_kapi.c 		if (entry == NULL) {
entry            1198 net/netlabel/netlabel_kapi.c 		switch (entry->type) {
entry            1201 net/netlabel/netlabel_kapi.c 						       entry->cipso, secattr);
entry            1213 net/netlabel/netlabel_kapi.c 		entry = netlbl_domhsh_getentry_af6(secattr->domain,
entry            1215 net/netlabel/netlabel_kapi.c 		if (entry == NULL) {
entry            1219 net/netlabel/netlabel_kapi.c 		switch (entry->type) {
entry            1222 net/netlabel/netlabel_kapi.c 						      entry->calipso, secattr);
entry            1284 net/netlabel/netlabel_kapi.c 	struct netlbl_dommap_def *entry;
entry            1290 net/netlabel/netlabel_kapi.c 		entry = netlbl_domhsh_getentry_af4(secattr->domain,
entry            1292 net/netlabel/netlabel_kapi.c 		if (entry == NULL) {
entry            1296 net/netlabel/netlabel_kapi.c 		switch (entry->type) {
entry            1298 net/netlabel/netlabel_kapi.c 			ret_val = cipso_v4_skbuff_setattr(skb, entry->cipso,
entry            1313 net/netlabel/netlabel_kapi.c 		entry = netlbl_domhsh_getentry_af6(secattr->domain,
entry            1315 net/netlabel/netlabel_kapi.c 		if (entry == NULL) {
entry            1319 net/netlabel/netlabel_kapi.c 		switch (entry->type) {
entry            1321 net/netlabel/netlabel_kapi.c 			ret_val = calipso_skbuff_setattr(skb, entry->calipso,
entry              86 net/netlabel/netlabel_mgmt.c 	struct netlbl_dom_map *entry = kzalloc(sizeof(*entry), GFP_KERNEL);
entry              88 net/netlabel/netlabel_mgmt.c 	if (!entry)
entry              90 net/netlabel/netlabel_mgmt.c 	entry->def.type = nla_get_u32(info->attrs[NLBL_MGMT_A_PROTOCOL]);
entry              93 net/netlabel/netlabel_mgmt.c 		entry->domain = kmalloc(tmp_size, GFP_KERNEL);
entry              94 net/netlabel/netlabel_mgmt.c 		if (entry->domain == NULL) {
entry              98 net/netlabel/netlabel_mgmt.c 		nla_strlcpy(entry->domain,
entry             107 net/netlabel/netlabel_mgmt.c 	switch (entry->def.type) {
entry             110 net/netlabel/netlabel_mgmt.c 			entry->family =
entry             113 net/netlabel/netlabel_mgmt.c 			entry->family = AF_UNSPEC;
entry             123 net/netlabel/netlabel_mgmt.c 		entry->family = AF_INET;
entry             124 net/netlabel/netlabel_mgmt.c 		entry->def.cipso = cipsov4;
entry             135 net/netlabel/netlabel_mgmt.c 		entry->family = AF_INET6;
entry             136 net/netlabel/netlabel_mgmt.c 		entry->def.calipso = calipso;
entry             143 net/netlabel/netlabel_mgmt.c 	if ((entry->family == AF_INET && info->attrs[NLBL_MGMT_A_IPV6ADDR]) ||
entry             144 net/netlabel/netlabel_mgmt.c 	    (entry->family == AF_INET6 && info->attrs[NLBL_MGMT_A_IPV4ADDR]))
entry             181 net/netlabel/netlabel_mgmt.c 		map->def.type = entry->def.type;
entry             191 net/netlabel/netlabel_mgmt.c 		entry->family = AF_INET;
entry             192 net/netlabel/netlabel_mgmt.c 		entry->def.type = NETLBL_NLTYPE_ADDRSELECT;
entry             193 net/netlabel/netlabel_mgmt.c 		entry->def.addrsel = addrmap;
entry             233 net/netlabel/netlabel_mgmt.c 		map->def.type = entry->def.type;
entry             243 net/netlabel/netlabel_mgmt.c 		entry->family = AF_INET6;
entry             244 net/netlabel/netlabel_mgmt.c 		entry->def.type = NETLBL_NLTYPE_ADDRSELECT;
entry             245 net/netlabel/netlabel_mgmt.c 		entry->def.addrsel = addrmap;
entry             249 net/netlabel/netlabel_mgmt.c 	ret_val = netlbl_domhsh_add(entry, audit_info);
entry             263 net/netlabel/netlabel_mgmt.c 	kfree(entry->domain);
entry             265 net/netlabel/netlabel_mgmt.c 	kfree(entry);
entry             281 net/netlabel/netlabel_mgmt.c 				 struct netlbl_dom_map *entry)
entry             291 net/netlabel/netlabel_mgmt.c 	if (entry->domain != NULL) {
entry             293 net/netlabel/netlabel_mgmt.c 					 NLBL_MGMT_A_DOMAIN, entry->domain);
entry             298 net/netlabel/netlabel_mgmt.c 	ret_val = nla_put_u16(skb, NLBL_MGMT_A_FAMILY, entry->family);
entry             302 net/netlabel/netlabel_mgmt.c 	switch (entry->def.type) {
entry             308 net/netlabel/netlabel_mgmt.c 		netlbl_af4list_foreach_rcu(iter4, &entry->def.addrsel->list4) {
entry             344 net/netlabel/netlabel_mgmt.c 		netlbl_af6list_foreach_rcu(iter6, &entry->def.addrsel->list6) {
entry             383 net/netlabel/netlabel_mgmt.c 				      entry->def.type);
entry             387 net/netlabel/netlabel_mgmt.c 				      entry->def.type);
entry             391 net/netlabel/netlabel_mgmt.c 				      entry->def.cipso->doi);
entry             395 net/netlabel/netlabel_mgmt.c 				      entry->def.type);
entry             399 net/netlabel/netlabel_mgmt.c 				      entry->def.calipso->doi);
entry             478 net/netlabel/netlabel_mgmt.c static int netlbl_mgmt_listall_cb(struct netlbl_dom_map *entry, void *arg)
entry             490 net/netlabel/netlabel_mgmt.c 	ret_val = netlbl_mgmt_listentry(cb_arg->skb, entry);
entry             600 net/netlabel/netlabel_mgmt.c 	struct netlbl_dom_map *entry;
entry             617 net/netlabel/netlabel_mgmt.c 	entry = netlbl_domhsh_getentry(NULL, family);
entry             618 net/netlabel/netlabel_mgmt.c 	if (entry == NULL) {
entry             622 net/netlabel/netlabel_mgmt.c 	ret_val = netlbl_mgmt_listentry(ans_skb, entry);
entry             147 net/netlabel/netlabel_unlabeled.c static void netlbl_unlhsh_free_iface(struct rcu_head *entry)
entry             157 net/netlabel/netlabel_unlabeled.c 	iface = container_of(entry, struct netlbl_unlhsh_iface, rcu);
entry             236 net/netlabel/netlabel_unlabeled.c 	struct netlbl_unlhsh_addr4 *entry;
entry             238 net/netlabel/netlabel_unlabeled.c 	entry = kzalloc(sizeof(*entry), GFP_ATOMIC);
entry             239 net/netlabel/netlabel_unlabeled.c 	if (entry == NULL)
entry             242 net/netlabel/netlabel_unlabeled.c 	entry->list.addr = addr->s_addr & mask->s_addr;
entry             243 net/netlabel/netlabel_unlabeled.c 	entry->list.mask = mask->s_addr;
entry             244 net/netlabel/netlabel_unlabeled.c 	entry->list.valid = 1;
entry             245 net/netlabel/netlabel_unlabeled.c 	entry->secid = secid;
entry             248 net/netlabel/netlabel_unlabeled.c 	ret_val = netlbl_af4list_add(&entry->list, &iface->addr4_list);
entry             252 net/netlabel/netlabel_unlabeled.c 		kfree(entry);
entry             276 net/netlabel/netlabel_unlabeled.c 	struct netlbl_unlhsh_addr6 *entry;
entry             278 net/netlabel/netlabel_unlabeled.c 	entry = kzalloc(sizeof(*entry), GFP_ATOMIC);
entry             279 net/netlabel/netlabel_unlabeled.c 	if (entry == NULL)
entry             282 net/netlabel/netlabel_unlabeled.c 	entry->list.addr = *addr;
entry             283 net/netlabel/netlabel_unlabeled.c 	entry->list.addr.s6_addr32[0] &= mask->s6_addr32[0];
entry             284 net/netlabel/netlabel_unlabeled.c 	entry->list.addr.s6_addr32[1] &= mask->s6_addr32[1];
entry             285 net/netlabel/netlabel_unlabeled.c 	entry->list.addr.s6_addr32[2] &= mask->s6_addr32[2];
entry             286 net/netlabel/netlabel_unlabeled.c 	entry->list.addr.s6_addr32[3] &= mask->s6_addr32[3];
entry             287 net/netlabel/netlabel_unlabeled.c 	entry->list.mask = *mask;
entry             288 net/netlabel/netlabel_unlabeled.c 	entry->list.valid = 1;
entry             289 net/netlabel/netlabel_unlabeled.c 	entry->secid = secid;
entry             292 net/netlabel/netlabel_unlabeled.c 	ret_val = netlbl_af6list_add(&entry->list, &iface->addr6_list);
entry             296 net/netlabel/netlabel_unlabeled.c 		kfree(entry);
entry             472 net/netlabel/netlabel_unlabeled.c 	struct netlbl_unlhsh_addr4 *entry;
entry             483 net/netlabel/netlabel_unlabeled.c 		entry = netlbl_unlhsh_addr4_entry(list_entry);
entry             485 net/netlabel/netlabel_unlabeled.c 		entry = NULL;
entry             496 net/netlabel/netlabel_unlabeled.c 		if (entry != NULL &&
entry             497 net/netlabel/netlabel_unlabeled.c 		    security_secid_to_secctx(entry->secid,
entry             502 net/netlabel/netlabel_unlabeled.c 		audit_log_format(audit_buf, " res=%u", entry != NULL ? 1 : 0);
entry             506 net/netlabel/netlabel_unlabeled.c 	if (entry == NULL)
entry             509 net/netlabel/netlabel_unlabeled.c 	kfree_rcu(entry, rcu);
entry             534 net/netlabel/netlabel_unlabeled.c 	struct netlbl_unlhsh_addr6 *entry;
entry             544 net/netlabel/netlabel_unlabeled.c 		entry = netlbl_unlhsh_addr6_entry(list_entry);
entry             546 net/netlabel/netlabel_unlabeled.c 		entry = NULL;
entry             557 net/netlabel/netlabel_unlabeled.c 		if (entry != NULL &&
entry             558 net/netlabel/netlabel_unlabeled.c 		    security_secid_to_secctx(entry->secid,
entry             563 net/netlabel/netlabel_unlabeled.c 		audit_log_format(audit_buf, " res=%u", entry != NULL ? 1 : 0);
entry             567 net/netlabel/netlabel_unlabeled.c 	if (entry == NULL)
entry             570 net/netlabel/netlabel_unlabeled.c 	kfree_rcu(entry, rcu);
entry            1528 net/netlabel/netlabel_unlabeled.c 	struct netlbl_dom_map *entry;
entry            1538 net/netlabel/netlabel_unlabeled.c 	entry = kzalloc(sizeof(*entry), GFP_KERNEL);
entry            1539 net/netlabel/netlabel_unlabeled.c 	if (entry == NULL)
entry            1541 net/netlabel/netlabel_unlabeled.c 	entry->family = AF_UNSPEC;
entry            1542 net/netlabel/netlabel_unlabeled.c 	entry->def.type = NETLBL_NLTYPE_UNLABELED;
entry            1543 net/netlabel/netlabel_unlabeled.c 	ret_val = netlbl_domhsh_add_default(entry, &audit_info);
entry              37 net/nfc/hci/llc.c 	list_for_each_entry_safe(llc_engine, n, &llc_engines, entry) {
entry              38 net/nfc/hci/llc.c 		list_del(&llc_engine->entry);
entry              59 net/nfc/hci/llc.c 	INIT_LIST_HEAD(&llc_engine->entry);
entry              60 net/nfc/hci/llc.c 	list_add_tail(&llc_engine->entry, &llc_engines);
entry              69 net/nfc/hci/llc.c 	list_for_each_entry(llc_engine, &llc_engines, entry) {
entry              85 net/nfc/hci/llc.c 	list_del(&llc_engine->entry);
entry              30 net/nfc/hci/llc.h 	struct list_head entry;
entry            3387 net/sched/cls_api.c 	struct flow_action_entry *entry;
entry            3390 net/sched/cls_api.c 	flow_action_for_each(i, entry, flow_action)
entry            3391 net/sched/cls_api.c 		if (entry->destructor)
entry            3392 net/sched/cls_api.c 			entry->destructor(entry->destructor_priv);
entry            3396 net/sched/cls_api.c static void tcf_mirred_get_dev(struct flow_action_entry *entry,
entry            3400 net/sched/cls_api.c 	entry->dev = act->ops->get_dev(act, &entry->destructor);
entry            3401 net/sched/cls_api.c 	if (!entry->dev)
entry            3403 net/sched/cls_api.c 	entry->destructor_priv = entry->dev;
entry            3414 net/sched/cls_api.c static int tcf_tunnel_encap_get_tunnel(struct flow_action_entry *entry,
entry            3417 net/sched/cls_api.c 	entry->tunnel = tcf_tunnel_info_copy(act);
entry            3418 net/sched/cls_api.c 	if (!entry->tunnel)
entry            3420 net/sched/cls_api.c 	entry->destructor = tcf_tunnel_encap_put_tunnel;
entry            3421 net/sched/cls_api.c 	entry->destructor_priv = entry->tunnel;
entry            3425 net/sched/cls_api.c static void tcf_sample_get_group(struct flow_action_entry *entry,
entry            3429 net/sched/cls_api.c 	entry->sample.psample_group =
entry            3430 net/sched/cls_api.c 		act->ops->get_psample_group(act, &entry->destructor);
entry            3431 net/sched/cls_api.c 	entry->destructor_priv = entry->sample.psample_group;
entry            3449 net/sched/cls_api.c 		struct flow_action_entry *entry;
entry            3451 net/sched/cls_api.c 		entry = &flow_action->entries[j];
entry            3453 net/sched/cls_api.c 			entry->id = FLOW_ACTION_ACCEPT;
entry            3455 net/sched/cls_api.c 			entry->id = FLOW_ACTION_DROP;
entry            3457 net/sched/cls_api.c 			entry->id = FLOW_ACTION_TRAP;
entry            3459 net/sched/cls_api.c 			entry->id = FLOW_ACTION_GOTO;
entry            3460 net/sched/cls_api.c 			entry->chain_index = tcf_gact_goto_chain_index(act);
entry            3462 net/sched/cls_api.c 			entry->id = FLOW_ACTION_REDIRECT;
entry            3463 net/sched/cls_api.c 			tcf_mirred_get_dev(entry, act);
entry            3465 net/sched/cls_api.c 			entry->id = FLOW_ACTION_MIRRED;
entry            3466 net/sched/cls_api.c 			tcf_mirred_get_dev(entry, act);
entry            3468 net/sched/cls_api.c 			entry->id = FLOW_ACTION_REDIRECT_INGRESS;
entry            3469 net/sched/cls_api.c 			tcf_mirred_get_dev(entry, act);
entry            3471 net/sched/cls_api.c 			entry->id = FLOW_ACTION_MIRRED_INGRESS;
entry            3472 net/sched/cls_api.c 			tcf_mirred_get_dev(entry, act);
entry            3476 net/sched/cls_api.c 				entry->id = FLOW_ACTION_VLAN_PUSH;
entry            3477 net/sched/cls_api.c 				entry->vlan.vid = tcf_vlan_push_vid(act);
entry            3478 net/sched/cls_api.c 				entry->vlan.proto = tcf_vlan_push_proto(act);
entry            3479 net/sched/cls_api.c 				entry->vlan.prio = tcf_vlan_push_prio(act);
entry            3482 net/sched/cls_api.c 				entry->id = FLOW_ACTION_VLAN_POP;
entry            3485 net/sched/cls_api.c 				entry->id = FLOW_ACTION_VLAN_MANGLE;
entry            3486 net/sched/cls_api.c 				entry->vlan.vid = tcf_vlan_push_vid(act);
entry            3487 net/sched/cls_api.c 				entry->vlan.proto = tcf_vlan_push_proto(act);
entry            3488 net/sched/cls_api.c 				entry->vlan.prio = tcf_vlan_push_prio(act);
entry            3495 net/sched/cls_api.c 			entry->id = FLOW_ACTION_TUNNEL_ENCAP;
entry            3496 net/sched/cls_api.c 			err = tcf_tunnel_encap_get_tunnel(entry, act);
entry            3500 net/sched/cls_api.c 			entry->id = FLOW_ACTION_TUNNEL_DECAP;
entry            3505 net/sched/cls_api.c 					entry->id = FLOW_ACTION_MANGLE;
entry            3508 net/sched/cls_api.c 					entry->id = FLOW_ACTION_ADD;
entry            3514 net/sched/cls_api.c 				entry->mangle.htype = tcf_pedit_htype(act, k);
entry            3515 net/sched/cls_api.c 				entry->mangle.mask = tcf_pedit_mask(act, k);
entry            3516 net/sched/cls_api.c 				entry->mangle.val = tcf_pedit_val(act, k);
entry            3517 net/sched/cls_api.c 				entry->mangle.offset = tcf_pedit_offset(act, k);
entry            3518 net/sched/cls_api.c 				entry = &flow_action->entries[++j];
entry            3521 net/sched/cls_api.c 			entry->id = FLOW_ACTION_CSUM;
entry            3522 net/sched/cls_api.c 			entry->csum_flags = tcf_csum_update_flags(act);
entry            3524 net/sched/cls_api.c 			entry->id = FLOW_ACTION_MARK;
entry            3525 net/sched/cls_api.c 			entry->mark = tcf_skbedit_mark(act);
entry            3527 net/sched/cls_api.c 			entry->id = FLOW_ACTION_SAMPLE;
entry            3528 net/sched/cls_api.c 			entry->sample.trunc_size = tcf_sample_trunc_size(act);
entry            3529 net/sched/cls_api.c 			entry->sample.truncate = tcf_sample_truncate(act);
entry            3530 net/sched/cls_api.c 			entry->sample.rate = tcf_sample_rate(act);
entry            3531 net/sched/cls_api.c 			tcf_sample_get_group(entry, act);
entry            3533 net/sched/cls_api.c 			entry->id = FLOW_ACTION_POLICE;
entry            3534 net/sched/cls_api.c 			entry->police.burst = tcf_police_tcfp_burst(act);
entry            3535 net/sched/cls_api.c 			entry->police.rate_bytes_ps =
entry            3538 net/sched/cls_api.c 			entry->id = FLOW_ACTION_CT;
entry            3539 net/sched/cls_api.c 			entry->ct.action = tcf_ct_action(act);
entry            3540 net/sched/cls_api.c 			entry->ct.zone = tcf_ct_zone(act);
entry            3544 net/sched/cls_api.c 				entry->id = FLOW_ACTION_MPLS_PUSH;
entry            3545 net/sched/cls_api.c 				entry->mpls_push.proto = tcf_mpls_proto(act);
entry            3546 net/sched/cls_api.c 				entry->mpls_push.label = tcf_mpls_label(act);
entry            3547 net/sched/cls_api.c 				entry->mpls_push.tc = tcf_mpls_tc(act);
entry            3548 net/sched/cls_api.c 				entry->mpls_push.bos = tcf_mpls_bos(act);
entry            3549 net/sched/cls_api.c 				entry->mpls_push.ttl = tcf_mpls_ttl(act);
entry            3552 net/sched/cls_api.c 				entry->id = FLOW_ACTION_MPLS_POP;
entry            3553 net/sched/cls_api.c 				entry->mpls_pop.proto = tcf_mpls_proto(act);
entry            3556 net/sched/cls_api.c 				entry->id = FLOW_ACTION_MPLS_MANGLE;
entry            3557 net/sched/cls_api.c 				entry->mpls_mangle.label = tcf_mpls_label(act);
entry            3558 net/sched/cls_api.c 				entry->mpls_mangle.tc = tcf_mpls_tc(act);
entry            3559 net/sched/cls_api.c 				entry->mpls_mangle.bos = tcf_mpls_bos(act);
entry            3560 net/sched/cls_api.c 				entry->mpls_mangle.ttl = tcf_mpls_ttl(act);
entry            3566 net/sched/cls_api.c 			entry->id = FLOW_ACTION_PTYPE;
entry            3567 net/sched/cls_api.c 			entry->ptype = tcf_skbedit_ptype(act);
entry             531 net/sched/sch_gred.c static void gred_vq_apply(struct gred_sched *table, const struct nlattr *entry)
entry             536 net/sched/sch_gred.c 	nla_parse_nested_deprecated(tb, TCA_GRED_VQ_MAX, entry,
entry             560 net/sched/sch_gred.c 			    const struct nlattr *entry,
entry             567 net/sched/sch_gred.c 	err = nla_parse_nested_deprecated(tb, TCA_GRED_VQ_MAX, entry,
entry             114 net/sched/sch_taprio.c 	struct sched_entry *entry, *n;
entry             119 net/sched/sch_taprio.c 	list_for_each_entry_safe(entry, n, &sched->entries, list) {
entry             120 net/sched/sch_taprio.c 		list_del(&entry->list);
entry             121 net/sched/sch_taprio.c 		kfree(entry);
entry             155 net/sched/sch_taprio.c 				     struct sched_entry *entry,
entry             162 net/sched/sch_taprio.c 	intv_end = ktime_add_ns(intv_start, entry->interval);
entry             195 net/sched/sch_taprio.c 	struct sched_entry *entry = NULL, *entry_found = NULL;
entry             216 net/sched/sch_taprio.c 	list_for_each_entry(entry, &sched->entries, list) {
entry             218 net/sched/sch_taprio.c 		curr_intv_end = get_interval_end_time(sched, admin, entry,
entry             224 net/sched/sch_taprio.c 		if (!(entry->gate_mask & BIT(tc)) ||
entry             225 net/sched/sch_taprio.c 		    packet_transmit_time > entry->interval)
entry             228 net/sched/sch_taprio.c 		txtime = entry->next_txtime;
entry             235 net/sched/sch_taprio.c 				entry_found = entry;
entry             244 net/sched/sch_taprio.c 				entry_found = entry;
entry             251 net/sched/sch_taprio.c 			entry_found = entry;
entry             266 net/sched/sch_taprio.c 	struct sched_entry *entry;
entry             272 net/sched/sch_taprio.c 	entry = find_entry_to_transmit(skb, sch, sched, admin, skb->tstamp,
entry             276 net/sched/sch_taprio.c 	return entry;
entry             349 net/sched/sch_taprio.c 	struct sched_entry *entry;
entry             376 net/sched/sch_taprio.c 		entry = find_entry_to_transmit(skb, sch, sched, admin,
entry             380 net/sched/sch_taprio.c 		if (!entry) {
entry             385 net/sched/sch_taprio.c 		txtime = entry->next_txtime;
entry             403 net/sched/sch_taprio.c 			entry->next_txtime = ktime_add(interval_start, sched->cycle_time);
entry             406 net/sched/sch_taprio.c 	entry->next_txtime = transmit_end_time;
entry             445 net/sched/sch_taprio.c 	struct sched_entry *entry;
entry             451 net/sched/sch_taprio.c 	entry = rcu_dereference(q->current_entry);
entry             452 net/sched/sch_taprio.c 	gate_mask = entry ? entry->gate_mask : TAPRIO_ALL_GATES_OPEN;
entry             515 net/sched/sch_taprio.c static void taprio_set_budget(struct taprio_sched *q, struct sched_entry *entry)
entry             517 net/sched/sch_taprio.c 	atomic_set(&entry->budget,
entry             518 net/sched/sch_taprio.c 		   div64_u64((u64)entry->interval * 1000,
entry             527 net/sched/sch_taprio.c 	struct sched_entry *entry;
entry             532 net/sched/sch_taprio.c 	entry = rcu_dereference(q->current_entry);
entry             538 net/sched/sch_taprio.c 	gate_mask = entry ? entry->gate_mask : TAPRIO_ALL_GATES_OPEN;
entry             580 net/sched/sch_taprio.c 		    ktime_after(guard, entry->close_time)) {
entry             587 net/sched/sch_taprio.c 		    atomic_sub_return(len, &entry->budget) < 0) {
entry             645 net/sched/sch_taprio.c 				 const struct sched_entry *entry)
entry             647 net/sched/sch_taprio.c 	if (list_is_last(&entry->list, &oper->entries))
entry             650 net/sched/sch_taprio.c 	if (ktime_compare(entry->close_time, oper->cycle_close_time) == 0)
entry             695 net/sched/sch_taprio.c 	struct sched_entry *entry, *next;
entry             700 net/sched/sch_taprio.c 	entry = rcu_dereference_protected(q->current_entry,
entry             716 net/sched/sch_taprio.c 	if (unlikely(!entry || entry->close_time == oper->base_time)) {
entry             723 net/sched/sch_taprio.c 	if (should_restart_cycle(oper, entry)) {
entry             729 net/sched/sch_taprio.c 		next = list_next_entry(entry, list);
entry             732 net/sched/sch_taprio.c 	close_time = ktime_add_ns(entry->close_time, next->interval);
entry             780 net/sched/sch_taprio.c static int fill_sched_entry(struct nlattr **tb, struct sched_entry *entry,
entry             786 net/sched/sch_taprio.c 		entry->command = nla_get_u8(
entry             790 net/sched/sch_taprio.c 		entry->gate_mask = nla_get_u32(
entry             802 net/sched/sch_taprio.c 	entry->interval = interval;
entry             807 net/sched/sch_taprio.c static int parse_sched_entry(struct nlattr *n, struct sched_entry *entry,
entry             820 net/sched/sch_taprio.c 	entry->index = index;
entry             822 net/sched/sch_taprio.c 	return fill_sched_entry(tb, entry, extack);
entry             837 net/sched/sch_taprio.c 		struct sched_entry *entry;
entry             844 net/sched/sch_taprio.c 		entry = kzalloc(sizeof(*entry), GFP_KERNEL);
entry             845 net/sched/sch_taprio.c 		if (!entry) {
entry             850 net/sched/sch_taprio.c 		err = parse_sched_entry(n, entry, i, extack);
entry             852 net/sched/sch_taprio.c 			kfree(entry);
entry             856 net/sched/sch_taprio.c 		list_add_tail(&entry->list, &sched->entries);
entry             892 net/sched/sch_taprio.c 		struct sched_entry *entry;
entry             895 net/sched/sch_taprio.c 		list_for_each_entry(entry, &new->entries, list)
entry             896 net/sched/sch_taprio.c 			cycle = ktime_add_ns(cycle, entry->interval);
entry            1100 net/sched/sch_taprio.c 	struct sched_entry *entry;
entry            1103 net/sched/sch_taprio.c 	list_for_each_entry(entry, &sched->entries, list) {
entry            1104 net/sched/sch_taprio.c 		entry->next_txtime = ktime_add_ns(base, interval);
entry            1105 net/sched/sch_taprio.c 		interval += entry->interval;
entry            1185 net/sched/sch_taprio.c 	struct sched_entry *entry;
entry            1192 net/sched/sch_taprio.c 	list_for_each_entry(entry, &sched->entries, list) {
entry            1195 net/sched/sch_taprio.c 		e->command = entry->command;
entry            1196 net/sched/sch_taprio.c 		e->interval = entry->interval;
entry            1197 net/sched/sch_taprio.c 		e->gate_mask = entry->gate_mask;
entry            1712 net/sched/sch_taprio.c 		      const struct sched_entry *entry)
entry            1720 net/sched/sch_taprio.c 	if (nla_put_u32(msg, TCA_TAPRIO_SCHED_ENTRY_INDEX, entry->index))
entry            1723 net/sched/sch_taprio.c 	if (nla_put_u8(msg, TCA_TAPRIO_SCHED_ENTRY_CMD, entry->command))
entry            1727 net/sched/sch_taprio.c 			entry->gate_mask))
entry            1731 net/sched/sch_taprio.c 			entry->interval))
entry            1745 net/sched/sch_taprio.c 	struct sched_entry *entry;
entry            1764 net/sched/sch_taprio.c 	list_for_each_entry(entry, &root->entries, list) {
entry            1765 net/sched/sch_taprio.c 		if (dump_entry(msg, entry) < 0)
entry             148 net/sctp/inqueue.c 		struct list_head *entry;
entry             152 net/sctp/inqueue.c 		entry = sctp_list_dequeue(&queue->in_chunk_list);
entry             153 net/sctp/inqueue.c 		if (!entry)
entry             156 net/sctp/inqueue.c 		chunk = list_entry(entry, struct sctp_chunk, list);
entry            3835 net/sctp/sm_statefuns.c 	struct list_head *entry;
entry            3840 net/sctp/sm_statefuns.c 	entry = asoc->addip_chunk_list.next;
entry            3841 net/sctp/sm_statefuns.c 	asconf = list_entry(entry, struct sctp_chunk, list);
entry            3843 net/sctp/sm_statefuns.c 	list_del_init(entry);
entry              62 net/sctp/stream_sched.c 	struct list_head *entry;
entry              71 net/sctp/stream_sched.c 		entry = q->out_chunk_list.next;
entry              72 net/sctp/stream_sched.c 		ch = list_entry(entry, struct sctp_chunk, list);
entry            2418 net/socket.c   	struct mmsghdr __user *entry;
entry            2437 net/socket.c   	entry = mmsg;
entry            2455 net/socket.c   					     (struct user_msghdr __user *)entry,
entry            2459 net/socket.c   			err = put_user(err, &entry->msg_len);
entry            2460 net/socket.c   			++entry;
entry            2640 net/socket.c   	struct mmsghdr __user *entry;
entry            2665 net/socket.c   	entry = mmsg;
entry            2682 net/socket.c   					     (struct user_msghdr __user *)entry,
entry            2687 net/socket.c   			err = put_user(err, &entry->msg_len);
entry            2688 net/socket.c   			++entry;
entry             572 net/sunrpc/auth.c 			*entry, *new;
entry             578 net/sunrpc/auth.c 	hlist_for_each_entry_rcu(entry, &cache->hashtable[nr], cr_hash) {
entry             579 net/sunrpc/auth.c 		if (!entry->cr_ops->crmatch(acred, entry, flags))
entry             581 net/sunrpc/auth.c 		cred = get_rpccred(entry);
entry             597 net/sunrpc/auth.c 	hlist_for_each_entry(entry, &cache->hashtable[nr], cr_hash) {
entry             598 net/sunrpc/auth.c 		if (!entry->cr_ops->crmatch(acred, entry, flags))
entry             600 net/sunrpc/auth.c 		cred = get_rpccred(entry);
entry            1320 net/sunrpc/cache.c 	unsigned int hash, entry;
entry            1327 net/sunrpc/cache.c 	entry = n & ((1LL<<32) - 1);
entry            1330 net/sunrpc/cache.c 		if (!entry--)
entry             966 net/sunrpc/xprt.c 	struct rpc_rqst *entry;
entry             968 net/sunrpc/xprt.c 	entry = xprt_request_rb_find(xprt, xid);
entry             969 net/sunrpc/xprt.c 	if (entry != NULL) {
entry             971 net/sunrpc/xprt.c 		entry->rq_rtt = ktime_sub(ktime_get(), entry->rq_xtime);
entry             972 net/sunrpc/xprt.c 		return entry;
entry              91 net/wireless/lib80211.c 	struct lib80211_crypt_data *entry, *next;
entry              95 net/wireless/lib80211.c 	list_for_each_entry_safe(entry, next, &info->crypt_deinit_list, list) {
entry              96 net/wireless/lib80211.c 		if (atomic_read(&entry->refcnt) != 0 && !force)
entry              99 net/wireless/lib80211.c 		list_del(&entry->list);
entry             101 net/wireless/lib80211.c 		if (entry->ops) {
entry             102 net/wireless/lib80211.c 			entry->ops->deinit(entry->priv);
entry             103 net/wireless/lib80211.c 			module_put(entry->ops->owner);
entry             105 net/wireless/lib80211.c 		kfree(entry);
entry             779 net/wireless/util.c 		ret = (ntohl(mpls->entry) & MPLS_LS_TC_MASK)
entry              22 net/x25/x25_forward.c 	struct list_head *entry;
entry              49 net/x25/x25_forward.c 	list_for_each(entry, &x25_forward_list) {
entry              50 net/x25/x25_forward.c 		x25_frwd = list_entry(entry, struct x25_forward, node);
entry              95 net/x25/x25_forward.c 	struct list_head *entry;
entry             102 net/x25/x25_forward.c 	list_for_each(entry, &x25_forward_list) {
entry             103 net/x25/x25_forward.c 		frwd = list_entry(entry, struct x25_forward, node);
entry             135 net/x25/x25_forward.c 	struct list_head *entry, *tmp;
entry             139 net/x25/x25_forward.c 	list_for_each_safe(entry, tmp, &x25_forward_list) {
entry             140 net/x25/x25_forward.c 		fwd = list_entry(entry, struct x25_forward, node);
entry             153 net/x25/x25_forward.c 	struct list_head *entry, *tmp;
entry             157 net/x25/x25_forward.c 	list_for_each_safe(entry, tmp, &x25_forward_list) {
entry             158 net/x25/x25_forward.c 		fwd = list_entry(entry, struct x25_forward, node);
entry             295 net/x25/x25_link.c 	struct list_head *entry, *tmp;
entry             299 net/x25/x25_link.c 	list_for_each_safe(entry, tmp, &x25_neigh_list) {
entry             300 net/x25/x25_link.c 		nb = list_entry(entry, struct x25_neigh, node);
entry             317 net/x25/x25_link.c 	struct list_head *entry;
entry             320 net/x25/x25_link.c 	list_for_each(entry, &x25_neigh_list) {
entry             321 net/x25/x25_link.c 		nb = list_entry(entry, struct x25_neigh, node);
entry             393 net/x25/x25_link.c 	struct list_head *entry, *tmp;
entry             397 net/x25/x25_link.c 	list_for_each_safe(entry, tmp, &x25_neigh_list) {
entry             400 net/x25/x25_link.c 		nb = list_entry(entry, struct x25_neigh, node);
entry              30 net/x25/x25_route.c 	struct list_head *entry;
entry              35 net/x25/x25_route.c 	list_for_each(entry, &x25_route_list) {
entry              36 net/x25/x25_route.c 		rt = list_entry(entry, struct x25_route, node);
entry              81 net/x25/x25_route.c 	struct list_head *entry;
entry              86 net/x25/x25_route.c 	list_for_each(entry, &x25_route_list) {
entry              87 net/x25/x25_route.c 		rt = list_entry(entry, struct x25_route, node);
entry             107 net/x25/x25_route.c 	struct list_head *entry, *tmp;
entry             111 net/x25/x25_route.c 	list_for_each_safe(entry, tmp, &x25_route_list) {
entry             112 net/x25/x25_route.c 		rt = list_entry(entry, struct x25_route, node);
entry             152 net/x25/x25_route.c 	struct list_head *entry;
entry             156 net/x25/x25_route.c 	list_for_each(entry, &x25_route_list) {
entry             157 net/x25/x25_route.c 		rt = list_entry(entry, struct x25_route, node);
entry             213 net/x25/x25_route.c 	struct list_head *entry, *tmp;
entry             216 net/x25/x25_route.c 	list_for_each_safe(entry, tmp, &x25_route_list) {
entry             217 net/x25/x25_route.c 		rt = list_entry(entry, struct x25_route, node);
entry             642 net/xfrm/xfrm_algo.c 	int match(const struct xfrm_algo_desc *entry, const void *data),
entry             669 net/xfrm/xfrm_algo.c static int xfrm_alg_id_match(const struct xfrm_algo_desc *entry,
entry             672 net/xfrm/xfrm_algo.c 	return entry->desc.sadb_alg_id == (unsigned long)data;
entry             696 net/xfrm/xfrm_algo.c static int xfrm_alg_name_match(const struct xfrm_algo_desc *entry,
entry             701 net/xfrm/xfrm_algo.c 	return name && (!strcmp(name, entry->name) ||
entry             702 net/xfrm/xfrm_algo.c 			(entry->compat && !strcmp(name, entry->compat)));
entry             731 net/xfrm/xfrm_algo.c static int xfrm_aead_name_match(const struct xfrm_algo_desc *entry,
entry             737 net/xfrm/xfrm_algo.c 	return aead->icvbits == entry->uinfo.aead.icv_truncbits && name &&
entry             738 net/xfrm/xfrm_algo.c 	       !strcmp(name, entry->name);
entry              72 samples/bpf/test_lru_dist.c static inline void __list_del_entry(struct list_head *entry)
entry              74 samples/bpf/test_lru_dist.c 	__list_del(entry->prev, entry->next);
entry              87 samples/mic/mpssd/mpssd.h char *readsysfs(char *dir, char *entry);
entry              88 samples/mic/mpssd/mpssd.h int setsysfs(char *dir, char *entry, char *value);
entry              15 samples/mic/mpssd/sysfs.c readsysfs(char *dir, char *entry)
entry              24 samples/mic/mpssd/sysfs.c 		snprintf(filename, PATH_MAX, "%s/%s", MICSYSFSDIR, entry);
entry              27 samples/mic/mpssd/sysfs.c 			 "%s/%s/%s", MICSYSFSDIR, dir, entry);
entry              57 samples/mic/mpssd/sysfs.c setsysfs(char *dir, char *entry, char *value)
entry              64 samples/mic/mpssd/sysfs.c 		snprintf(filename, PATH_MAX, "%s/%s", MICSYSFSDIR, entry);
entry              67 samples/mic/mpssd/sysfs.c 			 MICSYSFSDIR, dir, entry);
entry              69 samples/mic/mpssd/sysfs.c 	oldvalue = readsysfs(dir, entry);
entry            1396 scripts/asn1_compiler.c 	int entry;
entry            1404 scripts/asn1_compiler.c 		e->entry_index = entry = nr_entries;
entry            1417 scripts/asn1_compiler.c 			render_opcode(out, "_jump_target(%u),\n", entry);
entry            1424 scripts/asn1_compiler.c 			render_opcode(out, "_jump_target(%u),\n", entry);
entry            1443 scripts/asn1_compiler.c 	int entry, skippable = 0, outofline = 0;
entry            1574 scripts/asn1_compiler.c 			entry = nr_entries;
entry            1582 scripts/asn1_compiler.c 			render_opcode(out, "_jump_target(%u),\n", entry);
entry             891 scripts/dtc/livetree.c 	char *entry;
entry             900 scripts/dtc/livetree.c 	xasprintf(&entry, "%s:%s:%u",
entry             902 scripts/dtc/livetree.c 	append_to_property(fn, m->ref, entry, strlen(entry) + 1, TYPE_STRING);
entry             904 scripts/dtc/livetree.c 	free(entry);
entry             802 scripts/gcc-plugins/gcc-common.h static inline void cgraph_remove_function_insertion_hook(struct cgraph_node_hook_list *entry)
entry             804 scripts/gcc-plugins/gcc-common.h 	symtab->remove_cgraph_insertion_hook(entry);
entry             812 scripts/gcc-plugins/gcc-common.h static inline void cgraph_remove_node_removal_hook(struct cgraph_node_hook_list *entry)
entry             814 scripts/gcc-plugins/gcc-common.h 	symtab->remove_cgraph_removal_hook(entry);
entry             822 scripts/gcc-plugins/gcc-common.h static inline void cgraph_remove_node_duplication_hook(struct cgraph_2node_hook_list *entry)
entry             824 scripts/gcc-plugins/gcc-common.h 	symtab->remove_cgraph_duplication_hook(entry);
entry             789 scripts/gcc-plugins/randomize_layout_plugin.c 	const struct whitelist_entry *entry;
entry             792 scripts/gcc-plugins/randomize_layout_plugin.c 	for (entry = whitelist; entry->pathname; entry++) {
entry             793 scripts/gcc-plugins/randomize_layout_plugin.c 		if (!strstr(xloc.file, entry->pathname))
entry             796 scripts/gcc-plugins/randomize_layout_plugin.c 		if (type_name_eq(stmt, lhs_tree, entry->lhs) && type_name_eq(stmt, rhs_tree, entry->rhs))
entry             126 scripts/kconfig/list.h static inline void list_del(struct list_head *entry)
entry             128 scripts/kconfig/list.h 	__list_del(entry->prev, entry->next);
entry             129 scripts/kconfig/list.h 	entry->next = (struct list_head*)LIST_POISON1;
entry             130 scripts/kconfig/list.h 	entry->prev = (struct list_head*)LIST_POISON2;
entry            1801 scripts/kconfig/qconf.cc 	QString entry;
entry            1804 scripts/kconfig/qconf.cc 		entry = "single";
entry            1808 scripts/kconfig/qconf.cc 		entry = "split";
entry            1812 scripts/kconfig/qconf.cc 		entry = "full";
entry            1818 scripts/kconfig/qconf.cc 	configSettings->setValue("/listMode", entry);
entry             104 security/integrity/ima/ima.h 	struct ima_template_entry *entry;
entry             134 security/integrity/ima/ima.h int ima_add_template_entry(struct ima_template_entry *entry, int violation,
entry             156 security/integrity/ima/ima.h int ima_restore_measurement_entry(struct ima_template_entry *entry);
entry             223 security/integrity/ima/ima.h 			    struct ima_template_entry **entry,
entry             225 security/integrity/ima/ima.h int ima_store_template(struct ima_template_entry *entry, int violation,
entry             228 security/integrity/ima/ima.h void ima_free_template_entry(struct ima_template_entry *entry);
entry              23 security/integrity/ima/ima_api.c void ima_free_template_entry(struct ima_template_entry *entry)
entry              27 security/integrity/ima/ima_api.c 	for (i = 0; i < entry->template_desc->num_fields; i++)
entry              28 security/integrity/ima/ima_api.c 		kfree(entry->template_data[i].data);
entry              30 security/integrity/ima/ima_api.c 	kfree(entry);
entry              37 security/integrity/ima/ima_api.c 			    struct ima_template_entry **entry,
entry              48 security/integrity/ima/ima_api.c 	*entry = kzalloc(struct_size(*entry, template_data,
entry              50 security/integrity/ima/ima_api.c 	if (!*entry)
entry              53 security/integrity/ima/ima_api.c 	(*entry)->template_desc = template_desc;
entry              60 security/integrity/ima/ima_api.c 					   &((*entry)->template_data[i]));
entry              64 security/integrity/ima/ima_api.c 		len = (*entry)->template_data[i].len;
entry              65 security/integrity/ima/ima_api.c 		(*entry)->template_data_len += sizeof(len);
entry              66 security/integrity/ima/ima_api.c 		(*entry)->template_data_len += len;
entry              70 security/integrity/ima/ima_api.c 	ima_free_template_entry(*entry);
entry              71 security/integrity/ima/ima_api.c 	*entry = NULL;
entry              91 security/integrity/ima/ima_api.c int ima_store_template(struct ima_template_entry *entry,
entry              97 security/integrity/ima/ima_api.c 	char *template_name = entry->template_desc->name;
entry             105 security/integrity/ima/ima_api.c 		int num_fields = entry->template_desc->num_fields;
entry             109 security/integrity/ima/ima_api.c 		result = ima_calc_field_array_hash(&entry->template_data[0],
entry             110 security/integrity/ima/ima_api.c 						   entry->template_desc,
entry             118 security/integrity/ima/ima_api.c 		memcpy(entry->digest, hash.hdr.digest, hash.hdr.length);
entry             120 security/integrity/ima/ima_api.c 	entry->pcr = pcr;
entry             121 security/integrity/ima/ima_api.c 	result = ima_add_template_entry(entry, violation, op, inode, filename);
entry             136 security/integrity/ima/ima_api.c 	struct ima_template_entry *entry;
entry             148 security/integrity/ima/ima_api.c 	result = ima_alloc_init_template(&event_data, &entry, NULL);
entry             153 security/integrity/ima/ima_api.c 	result = ima_store_template(entry, violation, inode,
entry             156 security/integrity/ima/ima_api.c 		ima_free_template_entry(entry);
entry             303 security/integrity/ima/ima_api.c 	struct ima_template_entry *entry;
entry             321 security/integrity/ima/ima_api.c 	result = ima_alloc_init_template(&event_data, &entry, template_desc);
entry             328 security/integrity/ima/ima_api.c 	result = ima_store_template(entry, violation, inode, filename, pcr);
entry             334 security/integrity/ima/ima_api.c 		ima_free_template_entry(entry);
entry             139 security/integrity/ima/ima_fs.c 	e = qe->entry;
entry             227 security/integrity/ima/ima_fs.c 	e = qe->entry;
entry              46 security/integrity/ima/ima_init.c 	struct ima_template_entry *entry;
entry              71 security/integrity/ima/ima_init.c 	result = ima_alloc_init_template(&event_data, &entry, NULL);
entry              77 security/integrity/ima/ima_init.c 	result = ima_store_template(entry, violation, NULL,
entry              81 security/integrity/ima/ima_init.c 		ima_free_template_entry(entry);
entry             639 security/integrity/ima/ima_main.c 	struct ima_template_entry *entry = NULL;
entry             667 security/integrity/ima/ima_main.c 	ret = ima_alloc_init_template(&event_data, &entry, template_desc);
entry             671 security/integrity/ima/ima_main.c 	ret = ima_store_template(entry, violation, NULL, buf, pcr);
entry             674 security/integrity/ima/ima_main.c 		ima_free_template_entry(entry);
entry             252 security/integrity/ima/ima_policy.c static void ima_lsm_free_rule(struct ima_rule_entry *entry)
entry             257 security/integrity/ima/ima_policy.c 		kfree(entry->lsm[i].rule);
entry             258 security/integrity/ima/ima_policy.c 		kfree(entry->lsm[i].args_p);
entry             260 security/integrity/ima/ima_policy.c 	kfree(entry);
entry             263 security/integrity/ima/ima_policy.c static struct ima_rule_entry *ima_lsm_copy_rule(struct ima_rule_entry *entry)
entry             276 security/integrity/ima/ima_policy.c 	memcpy(nentry, entry, sizeof(*nentry));
entry             280 security/integrity/ima/ima_policy.c 		if (!entry->lsm[i].args_p)
entry             283 security/integrity/ima/ima_policy.c 		nentry->lsm[i].type = entry->lsm[i].type;
entry             284 security/integrity/ima/ima_policy.c 		nentry->lsm[i].args_p = kstrdup(entry->lsm[i].args_p,
entry             295 security/integrity/ima/ima_policy.c 				(char *)entry->lsm[i].args_p);
entry             304 security/integrity/ima/ima_policy.c static int ima_lsm_update_rule(struct ima_rule_entry *entry)
entry             308 security/integrity/ima/ima_policy.c 	nentry = ima_lsm_copy_rule(entry);
entry             312 security/integrity/ima/ima_policy.c 	list_replace_rcu(&entry->list, &nentry->list);
entry             314 security/integrity/ima/ima_policy.c 	ima_lsm_free_rule(entry);
entry             326 security/integrity/ima/ima_policy.c 	struct ima_rule_entry *entry, *e;
entry             329 security/integrity/ima/ima_policy.c 	list_for_each_entry_safe(entry, e, &ima_policy_rules, list) {
entry             332 security/integrity/ima/ima_policy.c 			if (entry->lsm[i].args_p) {
entry             340 security/integrity/ima/ima_policy.c 		result = ima_lsm_update_rule(entry);
entry             496 security/integrity/ima/ima_policy.c 	struct ima_rule_entry *entry;
entry             503 security/integrity/ima/ima_policy.c 	list_for_each_entry_rcu(entry, ima_rules, list) {
entry             505 security/integrity/ima/ima_policy.c 		if (!(entry->action & actmask))
entry             508 security/integrity/ima/ima_policy.c 		if (!ima_match_rules(entry, inode, cred, secid, func, mask))
entry             511 security/integrity/ima/ima_policy.c 		action |= entry->flags & IMA_ACTION_FLAGS;
entry             513 security/integrity/ima/ima_policy.c 		action |= entry->action & IMA_DO_MASK;
entry             514 security/integrity/ima/ima_policy.c 		if (entry->action & IMA_APPRAISE) {
entry             515 security/integrity/ima/ima_policy.c 			action |= get_subaction(entry, func);
entry             522 security/integrity/ima/ima_policy.c 		if (entry->action & IMA_DO_MASK)
entry             523 security/integrity/ima/ima_policy.c 			actmask &= ~(entry->action | entry->action << 1);
entry             525 security/integrity/ima/ima_policy.c 			actmask &= ~(entry->action | entry->action >> 1);
entry             527 security/integrity/ima/ima_policy.c 		if ((pcr) && (entry->flags & IMA_PCR))
entry             528 security/integrity/ima/ima_policy.c 			*pcr = entry->pcr;
entry             530 security/integrity/ima/ima_policy.c 		if (template_desc && entry->template)
entry             531 security/integrity/ima/ima_policy.c 			*template_desc = entry->template;
entry             549 security/integrity/ima/ima_policy.c 	struct ima_rule_entry *entry;
entry             551 security/integrity/ima/ima_policy.c 	list_for_each_entry(entry, ima_rules, list) {
entry             552 security/integrity/ima/ima_policy.c 		if (entry->action & IMA_DO_MASK)
entry             553 security/integrity/ima/ima_policy.c 			ima_policy_flag |= entry->action;
entry             580 security/integrity/ima/ima_policy.c 		struct ima_rule_entry *entry;
entry             586 security/integrity/ima/ima_policy.c 			entry = kmemdup(&entries[i], sizeof(*entry),
entry             588 security/integrity/ima/ima_policy.c 			if (!entry)
entry             591 security/integrity/ima/ima_policy.c 			list_add_tail(&entry->list, &ima_policy_rules);
entry             601 security/integrity/ima/ima_policy.c static int ima_parse_rule(char *rule, struct ima_rule_entry *entry);
entry             809 security/integrity/ima/ima_policy.c static int ima_lsm_rule_init(struct ima_rule_entry *entry,
entry             814 security/integrity/ima/ima_policy.c 	if (entry->lsm[lsm_rule].rule)
entry             817 security/integrity/ima/ima_policy.c 	entry->lsm[lsm_rule].args_p = match_strdup(args);
entry             818 security/integrity/ima/ima_policy.c 	if (!entry->lsm[lsm_rule].args_p)
entry             821 security/integrity/ima/ima_policy.c 	entry->lsm[lsm_rule].type = audit_type;
entry             822 security/integrity/ima/ima_policy.c 	result = security_filter_rule_init(entry->lsm[lsm_rule].type,
entry             824 security/integrity/ima/ima_policy.c 					   entry->lsm[lsm_rule].args_p,
entry             825 security/integrity/ima/ima_policy.c 					   &entry->lsm[lsm_rule].rule);
entry             826 security/integrity/ima/ima_policy.c 	if (!entry->lsm[lsm_rule].rule) {
entry             828 security/integrity/ima/ima_policy.c 			(char *)entry->lsm[lsm_rule].args_p);
entry             831 security/integrity/ima/ima_policy.c 			kfree(entry->lsm[lsm_rule].args_p);
entry             891 security/integrity/ima/ima_policy.c static int ima_parse_rule(char *rule, struct ima_rule_entry *entry)
entry             903 security/integrity/ima/ima_policy.c 	entry->uid = INVALID_UID;
entry             904 security/integrity/ima/ima_policy.c 	entry->fowner = INVALID_UID;
entry             905 security/integrity/ima/ima_policy.c 	entry->uid_op = &uid_eq;
entry             906 security/integrity/ima/ima_policy.c 	entry->fowner_op = &uid_eq;
entry             907 security/integrity/ima/ima_policy.c 	entry->action = UNKNOWN;
entry             922 security/integrity/ima/ima_policy.c 			if (entry->action != UNKNOWN)
entry             925 security/integrity/ima/ima_policy.c 			entry->action = MEASURE;
entry             930 security/integrity/ima/ima_policy.c 			if (entry->action != UNKNOWN)
entry             933 security/integrity/ima/ima_policy.c 			entry->action = DONT_MEASURE;
entry             938 security/integrity/ima/ima_policy.c 			if (entry->action != UNKNOWN)
entry             941 security/integrity/ima/ima_policy.c 			entry->action = APPRAISE;
entry             946 security/integrity/ima/ima_policy.c 			if (entry->action != UNKNOWN)
entry             949 security/integrity/ima/ima_policy.c 			entry->action = DONT_APPRAISE;
entry             954 security/integrity/ima/ima_policy.c 			if (entry->action != UNKNOWN)
entry             957 security/integrity/ima/ima_policy.c 			entry->action = AUDIT;
entry             962 security/integrity/ima/ima_policy.c 			if (entry->action != UNKNOWN)
entry             965 security/integrity/ima/ima_policy.c 			entry->action = HASH;
entry             970 security/integrity/ima/ima_policy.c 			if (entry->action != UNKNOWN)
entry             973 security/integrity/ima/ima_policy.c 			entry->action = DONT_HASH;
entry             978 security/integrity/ima/ima_policy.c 			if (entry->func)
entry             982 security/integrity/ima/ima_policy.c 				entry->func = FILE_CHECK;
entry             985 security/integrity/ima/ima_policy.c 				entry->func = FILE_CHECK;
entry             987 security/integrity/ima/ima_policy.c 				entry->func = MODULE_CHECK;
entry             989 security/integrity/ima/ima_policy.c 				entry->func = FIRMWARE_CHECK;
entry             992 security/integrity/ima/ima_policy.c 				entry->func = MMAP_CHECK;
entry             994 security/integrity/ima/ima_policy.c 				entry->func = BPRM_CHECK;
entry             996 security/integrity/ima/ima_policy.c 				entry->func = CREDS_CHECK;
entry             999 security/integrity/ima/ima_policy.c 				entry->func = KEXEC_KERNEL_CHECK;
entry            1002 security/integrity/ima/ima_policy.c 				entry->func = KEXEC_INITRAMFS_CHECK;
entry            1004 security/integrity/ima/ima_policy.c 				entry->func = POLICY_CHECK;
entry            1006 security/integrity/ima/ima_policy.c 				entry->func = KEXEC_CMDLINE;
entry            1010 security/integrity/ima/ima_policy.c 				entry->flags |= IMA_FUNC;
entry            1015 security/integrity/ima/ima_policy.c 			if (entry->mask)
entry            1023 security/integrity/ima/ima_policy.c 				entry->mask = MAY_EXEC;
entry            1025 security/integrity/ima/ima_policy.c 				entry->mask = MAY_WRITE;
entry            1027 security/integrity/ima/ima_policy.c 				entry->mask = MAY_READ;
entry            1029 security/integrity/ima/ima_policy.c 				entry->mask = MAY_APPEND;
entry            1033 security/integrity/ima/ima_policy.c 				entry->flags |= (*args[0].from == '^')
entry            1039 security/integrity/ima/ima_policy.c 			if (entry->fsmagic) {
entry            1044 security/integrity/ima/ima_policy.c 			result = kstrtoul(args[0].from, 16, &entry->fsmagic);
entry            1046 security/integrity/ima/ima_policy.c 				entry->flags |= IMA_FSMAGIC;
entry            1051 security/integrity/ima/ima_policy.c 			entry->fsname = kstrdup(args[0].from, GFP_KERNEL);
entry            1052 security/integrity/ima/ima_policy.c 			if (!entry->fsname) {
entry            1057 security/integrity/ima/ima_policy.c 			entry->flags |= IMA_FSNAME;
entry            1062 security/integrity/ima/ima_policy.c 			if (!uuid_is_null(&entry->fsuuid)) {
entry            1067 security/integrity/ima/ima_policy.c 			result = uuid_parse(args[0].from, &entry->fsuuid);
entry            1069 security/integrity/ima/ima_policy.c 				entry->flags |= IMA_FSUUID;
entry            1073 security/integrity/ima/ima_policy.c 			entry->uid_op = &uid_gt;
entry            1078 security/integrity/ima/ima_policy.c 				entry->uid_op = &uid_lt;
entry            1087 security/integrity/ima/ima_policy.c 					  args[0].from, entry->uid_op);
entry            1089 security/integrity/ima/ima_policy.c 			if (uid_valid(entry->uid)) {
entry            1096 security/integrity/ima/ima_policy.c 				entry->uid = make_kuid(current_user_ns(),
entry            1098 security/integrity/ima/ima_policy.c 				if (!uid_valid(entry->uid) ||
entry            1102 security/integrity/ima/ima_policy.c 					entry->flags |= uid_token
entry            1107 security/integrity/ima/ima_policy.c 			entry->fowner_op = &uid_gt;
entry            1111 security/integrity/ima/ima_policy.c 				entry->fowner_op = &uid_lt;
entry            1115 security/integrity/ima/ima_policy.c 					  entry->fowner_op);
entry            1117 security/integrity/ima/ima_policy.c 			if (uid_valid(entry->fowner)) {
entry            1124 security/integrity/ima/ima_policy.c 				entry->fowner = make_kuid(current_user_ns(), (uid_t)lnum);
entry            1125 security/integrity/ima/ima_policy.c 				if (!uid_valid(entry->fowner) || (((uid_t)lnum) != lnum))
entry            1128 security/integrity/ima/ima_policy.c 					entry->flags |= IMA_FOWNER;
entry            1133 security/integrity/ima/ima_policy.c 			result = ima_lsm_rule_init(entry, args,
entry            1139 security/integrity/ima/ima_policy.c 			result = ima_lsm_rule_init(entry, args,
entry            1145 security/integrity/ima/ima_policy.c 			result = ima_lsm_rule_init(entry, args,
entry            1151 security/integrity/ima/ima_policy.c 			result = ima_lsm_rule_init(entry, args,
entry            1157 security/integrity/ima/ima_policy.c 			result = ima_lsm_rule_init(entry, args,
entry            1163 security/integrity/ima/ima_policy.c 			result = ima_lsm_rule_init(entry, args,
entry            1168 security/integrity/ima/ima_policy.c 			if (entry->action != APPRAISE) {
entry            1175 security/integrity/ima/ima_policy.c 				entry->flags |= IMA_DIGSIG_REQUIRED;
entry            1176 security/integrity/ima/ima_policy.c 			else if (ima_hook_supports_modsig(entry->func) &&
entry            1178 security/integrity/ima/ima_policy.c 				entry->flags |= IMA_DIGSIG_REQUIRED |
entry            1184 security/integrity/ima/ima_policy.c 			entry->flags |= IMA_PERMIT_DIRECTIO;
entry            1187 security/integrity/ima/ima_policy.c 			if (entry->action != MEASURE) {
entry            1193 security/integrity/ima/ima_policy.c 			result = kstrtoint(args[0].from, 10, &entry->pcr);
entry            1194 security/integrity/ima/ima_policy.c 			if (result || INVALID_PCR(entry->pcr))
entry            1197 security/integrity/ima/ima_policy.c 				entry->flags |= IMA_PCR;
entry            1202 security/integrity/ima/ima_policy.c 			if (entry->action != MEASURE) {
entry            1207 security/integrity/ima/ima_policy.c 			if (!template_desc || entry->template) {
entry            1220 security/integrity/ima/ima_policy.c 			entry->template = template_desc;
entry            1228 security/integrity/ima/ima_policy.c 	if (!result && (entry->action == UNKNOWN))
entry            1230 security/integrity/ima/ima_policy.c 	else if (entry->action == APPRAISE)
entry            1231 security/integrity/ima/ima_policy.c 		temp_ima_appraise |= ima_appraise_flag(entry->func);
entry            1233 security/integrity/ima/ima_policy.c 	if (!result && entry->flags & IMA_MODSIG_ALLOWED) {
entry            1234 security/integrity/ima/ima_policy.c 		template_desc = entry->template ? entry->template :
entry            1255 security/integrity/ima/ima_policy.c 	struct ima_rule_entry *entry;
entry            1266 security/integrity/ima/ima_policy.c 	entry = kzalloc(sizeof(*entry), GFP_KERNEL);
entry            1267 security/integrity/ima/ima_policy.c 	if (!entry) {
entry            1273 security/integrity/ima/ima_policy.c 	INIT_LIST_HEAD(&entry->list);
entry            1275 security/integrity/ima/ima_policy.c 	result = ima_parse_rule(p, entry);
entry            1277 security/integrity/ima/ima_policy.c 		kfree(entry);
entry            1284 security/integrity/ima/ima_policy.c 	list_add_tail(&entry->list, &ima_temp_rules);
entry            1297 security/integrity/ima/ima_policy.c 	struct ima_rule_entry *entry, *tmp;
entry            1301 security/integrity/ima/ima_policy.c 	list_for_each_entry_safe(entry, tmp, &ima_temp_rules, list) {
entry            1303 security/integrity/ima/ima_policy.c 			kfree(entry->lsm[i].args_p);
entry            1305 security/integrity/ima/ima_policy.c 		list_del(&entry->list);
entry            1306 security/integrity/ima/ima_policy.c 		kfree(entry);
entry            1331 security/integrity/ima/ima_policy.c 	struct ima_rule_entry *entry;
entry            1334 security/integrity/ima/ima_policy.c 	list_for_each_entry_rcu(entry, ima_rules, list) {
entry            1337 security/integrity/ima/ima_policy.c 			return entry;
entry            1346 security/integrity/ima/ima_policy.c 	struct ima_rule_entry *entry = v;
entry            1349 security/integrity/ima/ima_policy.c 	entry = list_entry_rcu(entry->list.next, struct ima_rule_entry, list);
entry            1353 security/integrity/ima/ima_policy.c 	return (&entry->list == ima_rules) ? NULL : entry;
entry            1376 security/integrity/ima/ima_policy.c 	struct ima_rule_entry *entry = v;
entry            1383 security/integrity/ima/ima_policy.c 	if (entry->action & MEASURE)
entry            1385 security/integrity/ima/ima_policy.c 	if (entry->action & DONT_MEASURE)
entry            1387 security/integrity/ima/ima_policy.c 	if (entry->action & APPRAISE)
entry            1389 security/integrity/ima/ima_policy.c 	if (entry->action & DONT_APPRAISE)
entry            1391 security/integrity/ima/ima_policy.c 	if (entry->action & AUDIT)
entry            1393 security/integrity/ima/ima_policy.c 	if (entry->action & HASH)
entry            1395 security/integrity/ima/ima_policy.c 	if (entry->action & DONT_HASH)
entry            1400 security/integrity/ima/ima_policy.c 	if (entry->flags & IMA_FUNC)
entry            1401 security/integrity/ima/ima_policy.c 		policy_func_show(m, entry->func);
entry            1403 security/integrity/ima/ima_policy.c 	if ((entry->flags & IMA_MASK) || (entry->flags & IMA_INMASK)) {
entry            1404 security/integrity/ima/ima_policy.c 		if (entry->flags & IMA_MASK)
entry            1406 security/integrity/ima/ima_policy.c 		if (entry->mask & MAY_EXEC)
entry            1408 security/integrity/ima/ima_policy.c 		if (entry->mask & MAY_WRITE)
entry            1410 security/integrity/ima/ima_policy.c 		if (entry->mask & MAY_READ)
entry            1412 security/integrity/ima/ima_policy.c 		if (entry->mask & MAY_APPEND)
entry            1417 security/integrity/ima/ima_policy.c 	if (entry->flags & IMA_FSMAGIC) {
entry            1418 security/integrity/ima/ima_policy.c 		snprintf(tbuf, sizeof(tbuf), "0x%lx", entry->fsmagic);
entry            1423 security/integrity/ima/ima_policy.c 	if (entry->flags & IMA_FSNAME) {
entry            1424 security/integrity/ima/ima_policy.c 		snprintf(tbuf, sizeof(tbuf), "%s", entry->fsname);
entry            1429 security/integrity/ima/ima_policy.c 	if (entry->flags & IMA_PCR) {
entry            1430 security/integrity/ima/ima_policy.c 		snprintf(tbuf, sizeof(tbuf), "%d", entry->pcr);
entry            1435 security/integrity/ima/ima_policy.c 	if (entry->flags & IMA_FSUUID) {
entry            1436 security/integrity/ima/ima_policy.c 		seq_printf(m, "fsuuid=%pU", &entry->fsuuid);
entry            1440 security/integrity/ima/ima_policy.c 	if (entry->flags & IMA_UID) {
entry            1441 security/integrity/ima/ima_policy.c 		snprintf(tbuf, sizeof(tbuf), "%d", __kuid_val(entry->uid));
entry            1442 security/integrity/ima/ima_policy.c 		if (entry->uid_op == &uid_gt)
entry            1444 security/integrity/ima/ima_policy.c 		else if (entry->uid_op == &uid_lt)
entry            1451 security/integrity/ima/ima_policy.c 	if (entry->flags & IMA_EUID) {
entry            1452 security/integrity/ima/ima_policy.c 		snprintf(tbuf, sizeof(tbuf), "%d", __kuid_val(entry->uid));
entry            1453 security/integrity/ima/ima_policy.c 		if (entry->uid_op == &uid_gt)
entry            1455 security/integrity/ima/ima_policy.c 		else if (entry->uid_op == &uid_lt)
entry            1462 security/integrity/ima/ima_policy.c 	if (entry->flags & IMA_FOWNER) {
entry            1463 security/integrity/ima/ima_policy.c 		snprintf(tbuf, sizeof(tbuf), "%d", __kuid_val(entry->fowner));
entry            1464 security/integrity/ima/ima_policy.c 		if (entry->fowner_op == &uid_gt)
entry            1466 security/integrity/ima/ima_policy.c 		else if (entry->fowner_op == &uid_lt)
entry            1474 security/integrity/ima/ima_policy.c 		if (entry->lsm[i].rule) {
entry            1478 security/integrity/ima/ima_policy.c 					   (char *)entry->lsm[i].args_p);
entry            1482 security/integrity/ima/ima_policy.c 					   (char *)entry->lsm[i].args_p);
entry            1486 security/integrity/ima/ima_policy.c 					   (char *)entry->lsm[i].args_p);
entry            1490 security/integrity/ima/ima_policy.c 					   (char *)entry->lsm[i].args_p);
entry            1494 security/integrity/ima/ima_policy.c 					   (char *)entry->lsm[i].args_p);
entry            1498 security/integrity/ima/ima_policy.c 					   (char *)entry->lsm[i].args_p);
entry            1503 security/integrity/ima/ima_policy.c 	if (entry->template)
entry            1504 security/integrity/ima/ima_policy.c 		seq_printf(m, "template=%s ", entry->template->name);
entry            1505 security/integrity/ima/ima_policy.c 	if (entry->flags & IMA_DIGSIG_REQUIRED) {
entry            1506 security/integrity/ima/ima_policy.c 		if (entry->flags & IMA_MODSIG_ALLOWED)
entry            1511 security/integrity/ima/ima_policy.c 	if (entry->flags & IMA_PERMIT_DIRECTIO)
entry            1528 security/integrity/ima/ima_policy.c 	struct ima_rule_entry *entry;
entry            1538 security/integrity/ima/ima_policy.c 	list_for_each_entry_rcu(entry, ima_rules, list) {
entry            1539 security/integrity/ima/ima_policy.c 		if (entry->action != APPRAISE)
entry            1546 security/integrity/ima/ima_policy.c 		if (entry->func && entry->func != func)
entry            1553 security/integrity/ima/ima_policy.c 		if (entry->flags & IMA_DIGSIG_REQUIRED)
entry              60 security/integrity/ima/ima_queue.c 		rc = memcmp(qe->entry->digest, digest_value, TPM_DIGEST_SIZE);
entry              61 security/integrity/ima/ima_queue.c 		if ((rc == 0) && (qe->entry->pcr == pcr)) {
entry              75 security/integrity/ima/ima_queue.c static int get_binary_runtime_size(struct ima_template_entry *entry)
entry              80 security/integrity/ima/ima_queue.c 	size += sizeof(entry->digest);
entry              82 security/integrity/ima/ima_queue.c 	size += strlen(entry->template_desc->name);
entry              83 security/integrity/ima/ima_queue.c 	size += sizeof(entry->template_data_len);
entry              84 security/integrity/ima/ima_queue.c 	size += entry->template_data_len;
entry              94 security/integrity/ima/ima_queue.c static int ima_add_digest_entry(struct ima_template_entry *entry,
entry             105 security/integrity/ima/ima_queue.c 	qe->entry = entry;
entry             112 security/integrity/ima/ima_queue.c 		key = ima_hash_key(entry->digest);
entry             119 security/integrity/ima/ima_queue.c 		size = get_binary_runtime_size(entry);
entry             164 security/integrity/ima/ima_queue.c int ima_add_template_entry(struct ima_template_entry *entry, int violation,
entry             176 security/integrity/ima/ima_queue.c 		memcpy(digest, entry->digest, sizeof(digest));
entry             177 security/integrity/ima/ima_queue.c 		if (ima_lookup_digest_entry(digest, entry->pcr)) {
entry             184 security/integrity/ima/ima_queue.c 	result = ima_add_digest_entry(entry, 1);
entry             194 security/integrity/ima/ima_queue.c 	tpmresult = ima_pcr_extend(digest, entry->pcr);
entry             208 security/integrity/ima/ima_queue.c int ima_restore_measurement_entry(struct ima_template_entry *entry)
entry             213 security/integrity/ima/ima_queue.c 	result = ima_add_digest_entry(entry, 0);
entry             304 security/integrity/ima/ima_template.c 				     struct ima_template_entry **entry)
entry             309 security/integrity/ima/ima_template.c 	*entry = kzalloc(struct_size(*entry, template_data,
entry             311 security/integrity/ima/ima_template.c 	if (!*entry)
entry             316 security/integrity/ima/ima_template.c 			    (*entry)->template_data, NULL, NULL,
entry             319 security/integrity/ima/ima_template.c 		kfree(*entry);
entry             323 security/integrity/ima/ima_template.c 	(*entry)->template_desc = template_desc;
entry             325 security/integrity/ima/ima_template.c 		struct ima_field_data *field_data = &(*entry)->template_data[i];
entry             328 security/integrity/ima/ima_template.c 		(*entry)->template_data[i].data =
entry             330 security/integrity/ima/ima_template.c 		if (!(*entry)->template_data[i].data) {
entry             334 security/integrity/ima/ima_template.c 		memcpy((*entry)->template_data[i].data, data, field_data->len);
entry             335 security/integrity/ima/ima_template.c 		(*entry)->template_data_len += sizeof(field_data->len);
entry             336 security/integrity/ima/ima_template.c 		(*entry)->template_data_len += field_data->len;
entry             340 security/integrity/ima/ima_template.c 		ima_free_template_entry(*entry);
entry             341 security/integrity/ima/ima_template.c 		*entry = NULL;
entry             360 security/integrity/ima/ima_template.c 	struct ima_template_entry *entry;
entry             446 security/integrity/ima/ima_template.c 						&entry);
entry             450 security/integrity/ima/ima_template.c 		memcpy(entry->digest, hdr[HDR_DIGEST].data,
entry             452 security/integrity/ima/ima_template.c 		entry->pcr = !ima_canonical_fmt ? *(hdr[HDR_PCR].data) :
entry             454 security/integrity/ima/ima_template.c 		ret = ima_restore_measurement_entry(entry);
entry              42 security/selinux/ss/sidtab.c 	struct sidtab_isid_entry *entry;
entry              48 security/selinux/ss/sidtab.c 	entry = &s->isids[sid - 1];
entry              50 security/selinux/ss/sidtab.c 	rc = context_cpy(&entry->context, context);
entry              54 security/selinux/ss/sidtab.c 	entry->set = 1;
entry              93 security/selinux/ss/sidtab.c 	union sidtab_entry_inner *entry;
entry             105 security/selinux/ss/sidtab.c 	entry = &s->roots[level];
entry             110 security/selinux/ss/sidtab.c 		entry = &entry->ptr_inner->entries[leaf_index >> capacity_shift];
entry             113 security/selinux/ss/sidtab.c 		if (!entry->ptr_inner) {
entry             115 security/selinux/ss/sidtab.c 				entry->ptr_inner = kzalloc(SIDTAB_NODE_ALLOC_SIZE,
entry             117 security/selinux/ss/sidtab.c 			if (!entry->ptr_inner)
entry             121 security/selinux/ss/sidtab.c 	if (!entry->ptr_leaf) {
entry             123 security/selinux/ss/sidtab.c 			entry->ptr_leaf = kzalloc(SIDTAB_NODE_ALLOC_SIZE,
entry             125 security/selinux/ss/sidtab.c 		if (!entry->ptr_leaf)
entry             128 security/selinux/ss/sidtab.c 	return &entry->ptr_leaf->entries[index % SIDTAB_LEAF_ENTRIES].context;
entry             173 security/selinux/ss/sidtab.c static int sidtab_find_context(union sidtab_entry_inner entry,
entry             181 security/selinux/ss/sidtab.c 		struct sidtab_node_inner *node = entry.ptr_inner;
entry             193 security/selinux/ss/sidtab.c 		struct sidtab_node_leaf *node = entry.ptr_leaf;
entry             344 security/selinux/ss/sidtab.c 		struct sidtab_isid_entry *entry = &s->isids[i];
entry             346 security/selinux/ss/sidtab.c 		if (entry->set && context_cmp(context, &entry->context)) {
entry             456 security/selinux/ss/sidtab.c static void sidtab_destroy_tree(union sidtab_entry_inner entry, u32 level)
entry             461 security/selinux/ss/sidtab.c 		struct sidtab_node_inner *node = entry.ptr_inner;
entry             470 security/selinux/ss/sidtab.c 		struct sidtab_node_leaf *node = entry.ptr_leaf;
entry             369 security/tomoyo/audit.c 	struct tomoyo_log *entry;
entry             378 security/tomoyo/audit.c 	entry = kzalloc(sizeof(*entry), GFP_NOFS);
entry             379 security/tomoyo/audit.c 	if (!entry) {
entry             383 security/tomoyo/audit.c 	entry->log = buf;
entry             389 security/tomoyo/audit.c 	entry->size = len + tomoyo_round2(sizeof(*entry));
entry             392 security/tomoyo/audit.c 	    tomoyo_memory_used[TOMOYO_MEMORY_AUDIT] + entry->size >=
entry             396 security/tomoyo/audit.c 		tomoyo_memory_used[TOMOYO_MEMORY_AUDIT] += entry->size;
entry             397 security/tomoyo/audit.c 		list_add_tail(&entry->list, &tomoyo_log);
entry             403 security/tomoyo/audit.c 		kfree(entry);
entry             494 security/tomoyo/common.c 	struct tomoyo_profile *entry;
entry             501 security/tomoyo/common.c 	entry = kzalloc(sizeof(*entry), GFP_NOFS);
entry             505 security/tomoyo/common.c 	if (!ptr && tomoyo_memory_ok(entry)) {
entry             506 security/tomoyo/common.c 		ptr = entry;
entry             518 security/tomoyo/common.c 		entry = NULL;
entry             522 security/tomoyo/common.c 	kfree(entry);
entry            2065 security/tomoyo/common.c 	struct tomoyo_query entry = { };
entry            2097 security/tomoyo/common.c 	entry.query = tomoyo_init_log(r, len, fmt, args);
entry            2099 security/tomoyo/common.c 	if (!entry.query)
entry            2101 security/tomoyo/common.c 	entry.query_len = strlen(entry.query) + 1;
entry            2103 security/tomoyo/common.c 		tomoyo_add_entry(r->domain, entry.query);
entry            2106 security/tomoyo/common.c 	len = tomoyo_round2(entry.query_len);
entry            2107 security/tomoyo/common.c 	entry.domain = r->domain;
entry            2114 security/tomoyo/common.c 		entry.serial = tomoyo_serial++;
entry            2115 security/tomoyo/common.c 		entry.retry = r->retry;
entry            2117 security/tomoyo/common.c 		list_add_tail(&entry.list, &tomoyo_query_list);
entry            2123 security/tomoyo/common.c 	while (entry.timer < 10) {
entry            2126 security/tomoyo/common.c 		    (tomoyo_answer_wait, entry.answer ||
entry            2129 security/tomoyo/common.c 		entry.timer++;
entry            2132 security/tomoyo/common.c 	list_del(&entry.list);
entry            2135 security/tomoyo/common.c 	switch (entry.answer) {
entry            2149 security/tomoyo/common.c 	kfree(entry.query);
entry             515 security/tomoyo/common.h 	struct tomoyo_path_info entry;
entry            1182 security/tomoyo/common.h 			container_of(name, typeof(*ptr), entry);
entry             407 security/tomoyo/condition.c (struct tomoyo_condition *entry)
entry             419 security/tomoyo/condition.c 		if (!tomoyo_same_condition(ptr, entry) ||
entry             428 security/tomoyo/condition.c 		if (tomoyo_memory_ok(entry)) {
entry             429 security/tomoyo/condition.c 			atomic_set(&entry->head.users, 1);
entry             430 security/tomoyo/condition.c 			list_add(&entry->head.list, &tomoyo_condition_list);
entry             439 security/tomoyo/condition.c 		tomoyo_del_condition(&entry->head.list);
entry             440 security/tomoyo/condition.c 		kfree(entry);
entry             441 security/tomoyo/condition.c 		entry = ptr;
entry             443 security/tomoyo/condition.c 	return entry;
entry             497 security/tomoyo/condition.c 	struct tomoyo_condition *entry = NULL;
entry             551 security/tomoyo/condition.c 			if (entry) {
entry             553 security/tomoyo/condition.c 				    entry->grant_log != TOMOYO_GRANTLOG_AUTO)
entry             556 security/tomoyo/condition.c 					entry->grant_log = TOMOYO_GRANTLOG_YES;
entry             558 security/tomoyo/condition.c 					entry->grant_log = TOMOYO_GRANTLOG_NO;
entry             658 security/tomoyo/condition.c 	if (entry) {
entry             661 security/tomoyo/condition.c 		return tomoyo_commit_condition(entry);
entry             663 security/tomoyo/condition.c 	e.size = sizeof(*entry)
entry             669 security/tomoyo/condition.c 	entry = kzalloc(e.size, GFP_NOFS);
entry             670 security/tomoyo/condition.c 	if (!entry)
entry             672 security/tomoyo/condition.c 	*entry = e;
entry             674 security/tomoyo/condition.c 	condp = (struct tomoyo_condition_element *) (entry + 1);
entry             697 security/tomoyo/condition.c 	if (entry) {
entry             698 security/tomoyo/condition.c 		tomoyo_del_condition(&entry->head.list);
entry             699 security/tomoyo/condition.c 		kfree(entry);
entry              39 security/tomoyo/domain.c 	struct tomoyo_acl_head *entry;
entry              44 security/tomoyo/domain.c 	list_for_each_entry_rcu(entry, list, list,
entry              46 security/tomoyo/domain.c 		if (entry->is_deleted == TOMOYO_GC_IN_PROGRESS)
entry              48 security/tomoyo/domain.c 		if (!check_duplicate(entry, new_entry))
entry              50 security/tomoyo/domain.c 		entry->is_deleted = param->is_delete;
entry              55 security/tomoyo/domain.c 		entry = tomoyo_commit_ok(new_entry, size);
entry              56 security/tomoyo/domain.c 		if (entry) {
entry              57 security/tomoyo/domain.c 			list_add_tail_rcu(&entry->list, list);
entry             104 security/tomoyo/domain.c 	struct tomoyo_acl_info *entry;
entry             123 security/tomoyo/domain.c 	list_for_each_entry_rcu(entry, list, list,
entry             125 security/tomoyo/domain.c 		if (entry->is_deleted == TOMOYO_GC_IN_PROGRESS)
entry             127 security/tomoyo/domain.c 		if (!tomoyo_same_acl_head(entry, new_entry) ||
entry             128 security/tomoyo/domain.c 		    !check_duplicate(entry, new_entry))
entry             131 security/tomoyo/domain.c 			entry->is_deleted = merge_duplicate(entry, new_entry,
entry             134 security/tomoyo/domain.c 			entry->is_deleted = is_delete;
entry             139 security/tomoyo/domain.c 		entry = tomoyo_commit_ok(new_entry, size);
entry             140 security/tomoyo/domain.c 		if (entry) {
entry             141 security/tomoyo/domain.c 			list_add_tail_rcu(&entry->list, list);
entry             465 security/tomoyo/domain.c 	struct tomoyo_policy_namespace *entry;
entry             476 security/tomoyo/domain.c 	entry = kzalloc(sizeof(*entry) + len + 1, GFP_NOFS);
entry             477 security/tomoyo/domain.c 	if (!entry)
entry             482 security/tomoyo/domain.c 	if (!ptr && tomoyo_memory_ok(entry)) {
entry             483 security/tomoyo/domain.c 		char *name = (char *) (entry + 1);
entry             485 security/tomoyo/domain.c 		ptr = entry;
entry             488 security/tomoyo/domain.c 		entry->name = name;
entry             489 security/tomoyo/domain.c 		tomoyo_init_policy_namespace(entry);
entry             490 security/tomoyo/domain.c 		entry = NULL;
entry             494 security/tomoyo/domain.c 	kfree(entry);
entry             528 security/tomoyo/domain.c 	struct tomoyo_domain_info *entry = tomoyo_find_domain(domainname);
entry             531 security/tomoyo/domain.c 	if (entry) {
entry             540 security/tomoyo/domain.c 			    !entry->ns->profile_ptr[entry->profile])
entry             543 security/tomoyo/domain.c 		return entry;
entry             576 security/tomoyo/domain.c 	entry = tomoyo_find_domain(domainname);
entry             577 security/tomoyo/domain.c 	if (!entry) {
entry             578 security/tomoyo/domain.c 		entry = tomoyo_commit_ok(&e, sizeof(e));
entry             579 security/tomoyo/domain.c 		if (entry) {
entry             580 security/tomoyo/domain.c 			INIT_LIST_HEAD(&entry->acl_info_list);
entry             581 security/tomoyo/domain.c 			list_add_tail_rcu(&entry->list, &tomoyo_domain_list);
entry             588 security/tomoyo/domain.c 	if (entry && transit) {
entry             593 security/tomoyo/domain.c 			tomoyo_init_request_info(&r, entry,
entry             597 security/tomoyo/domain.c 					 entry->profile);
entry             599 security/tomoyo/domain.c 				if (test_bit(i, entry->group))
entry             605 security/tomoyo/domain.c 	return entry;
entry             165 security/tomoyo/gc.c 			struct tomoyo_path_acl *entry
entry             166 security/tomoyo/gc.c 				= container_of(acl, typeof(*entry), head);
entry             167 security/tomoyo/gc.c 			tomoyo_put_name_union(&entry->name);
entry             172 security/tomoyo/gc.c 			struct tomoyo_path2_acl *entry
entry             173 security/tomoyo/gc.c 				= container_of(acl, typeof(*entry), head);
entry             174 security/tomoyo/gc.c 			tomoyo_put_name_union(&entry->name1);
entry             175 security/tomoyo/gc.c 			tomoyo_put_name_union(&entry->name2);
entry             180 security/tomoyo/gc.c 			struct tomoyo_path_number_acl *entry
entry             181 security/tomoyo/gc.c 				= container_of(acl, typeof(*entry), head);
entry             182 security/tomoyo/gc.c 			tomoyo_put_name_union(&entry->name);
entry             183 security/tomoyo/gc.c 			tomoyo_put_number_union(&entry->number);
entry             188 security/tomoyo/gc.c 			struct tomoyo_mkdev_acl *entry
entry             189 security/tomoyo/gc.c 				= container_of(acl, typeof(*entry), head);
entry             190 security/tomoyo/gc.c 			tomoyo_put_name_union(&entry->name);
entry             191 security/tomoyo/gc.c 			tomoyo_put_number_union(&entry->mode);
entry             192 security/tomoyo/gc.c 			tomoyo_put_number_union(&entry->major);
entry             193 security/tomoyo/gc.c 			tomoyo_put_number_union(&entry->minor);
entry             198 security/tomoyo/gc.c 			struct tomoyo_mount_acl *entry
entry             199 security/tomoyo/gc.c 				= container_of(acl, typeof(*entry), head);
entry             200 security/tomoyo/gc.c 			tomoyo_put_name_union(&entry->dev_name);
entry             201 security/tomoyo/gc.c 			tomoyo_put_name_union(&entry->dir_name);
entry             202 security/tomoyo/gc.c 			tomoyo_put_name_union(&entry->fs_type);
entry             203 security/tomoyo/gc.c 			tomoyo_put_number_union(&entry->flags);
entry             208 security/tomoyo/gc.c 			struct tomoyo_env_acl *entry =
entry             209 security/tomoyo/gc.c 				container_of(acl, typeof(*entry), head);
entry             211 security/tomoyo/gc.c 			tomoyo_put_name(entry->env);
entry             216 security/tomoyo/gc.c 			struct tomoyo_inet_acl *entry =
entry             217 security/tomoyo/gc.c 				container_of(acl, typeof(*entry), head);
entry             219 security/tomoyo/gc.c 			tomoyo_put_group(entry->address.group);
entry             220 security/tomoyo/gc.c 			tomoyo_put_number_union(&entry->port);
entry             225 security/tomoyo/gc.c 			struct tomoyo_unix_acl *entry =
entry             226 security/tomoyo/gc.c 				container_of(acl, typeof(*entry), head);
entry             228 security/tomoyo/gc.c 			tomoyo_put_name_union(&entry->name);
entry             233 security/tomoyo/gc.c 			struct tomoyo_task_acl *entry =
entry             234 security/tomoyo/gc.c 				container_of(acl, typeof(*entry), head);
entry             236 security/tomoyo/gc.c 			tomoyo_put_name(entry->domainname);
entry             439 security/tomoyo/gc.c 				  head.list)->entry.name))
entry             121 security/tomoyo/memory.c 		struct tomoyo_group *entry = tomoyo_commit_ok(&e, sizeof(e));
entry             123 security/tomoyo/memory.c 		if (entry) {
entry             124 security/tomoyo/memory.c 			INIT_LIST_HEAD(&entry->member_list);
entry             125 security/tomoyo/memory.c 			atomic_set(&entry->head.users, 1);
entry             126 security/tomoyo/memory.c 			list_add_tail_rcu(&entry->head.list, list);
entry             127 security/tomoyo/memory.c 			group = entry;
entry             167 security/tomoyo/memory.c 		if (hash != ptr->entry.hash || strcmp(name, ptr->entry.name) ||
entry             175 security/tomoyo/memory.c 		ptr->entry.name = ((char *) ptr) + sizeof(*ptr);
entry             176 security/tomoyo/memory.c 		memmove((char *) ptr->entry.name, name, len);
entry             178 security/tomoyo/memory.c 		tomoyo_fill_path_info(&ptr->entry);
entry             186 security/tomoyo/memory.c 	return ptr ? &ptr->entry : NULL;
entry            1019 sound/core/compress_offload.c static void snd_compress_proc_info_read(struct snd_info_entry *entry,
entry            1022 sound/core/compress_offload.c 	struct snd_compr *compr = (struct snd_compr *)entry->private_data;
entry            1034 sound/core/compress_offload.c 	struct snd_info_entry *entry;
entry            1038 sound/core/compress_offload.c 	entry = snd_info_create_card_entry(compr->card, name,
entry            1040 sound/core/compress_offload.c 	if (!entry)
entry            1042 sound/core/compress_offload.c 	entry->mode = S_IFDIR | 0555;
entry            1043 sound/core/compress_offload.c 	compr->proc_root = entry;
entry            1045 sound/core/compress_offload.c 	entry = snd_info_create_card_entry(compr->card, "info",
entry            1047 sound/core/compress_offload.c 	if (entry)
entry            1048 sound/core/compress_offload.c 		snd_info_set_text_ops(entry, compr,
entry            1050 sound/core/compress_offload.c 	compr->proc_info_entry = entry;
entry             478 sound/core/hwdep.c static void snd_hwdep_proc_read(struct snd_info_entry *entry,
entry             494 sound/core/hwdep.c 	struct snd_info_entry *entry;
entry             496 sound/core/hwdep.c 	if ((entry = snd_info_create_module_entry(THIS_MODULE, "hwdep", NULL)) != NULL) {
entry             497 sound/core/hwdep.c 		entry->c.text.read = snd_hwdep_proc_read;
entry             498 sound/core/hwdep.c 		if (snd_info_register(entry) < 0) {
entry             499 sound/core/hwdep.c 			snd_info_free_entry(entry);
entry             500 sound/core/hwdep.c 			entry = NULL;
entry             503 sound/core/hwdep.c 	snd_hwdep_proc_entry = entry;
entry              55 sound/core/info.c 	struct snd_info_entry *entry;
entry              60 sound/core/info.c static void snd_info_disconnect(struct snd_info_entry *entry);
entry              74 sound/core/info.c static int alloc_info_private(struct snd_info_entry *entry,
entry              79 sound/core/info.c 	if (!entry || !entry->p)
entry              81 sound/core/info.c 	if (!try_module_get(entry->module))
entry              85 sound/core/info.c 		module_put(entry->module);
entry              88 sound/core/info.c 	data->entry = entry;
entry             108 sound/core/info.c 	struct snd_info_entry *entry;
entry             112 sound/core/info.c 	entry = data->entry;
entry             113 sound/core/info.c 	mutex_lock(&entry->access);
entry             114 sound/core/info.c 	if (entry->c.ops->llseek) {
entry             115 sound/core/info.c 		offset = entry->c.ops->llseek(entry,
entry             121 sound/core/info.c 	size = entry->size;
entry             143 sound/core/info.c 	mutex_unlock(&entry->access);
entry             151 sound/core/info.c 	struct snd_info_entry *entry = data->entry;
entry             158 sound/core/info.c 	if (pos >= entry->size)
entry             160 sound/core/info.c 	size = entry->size - pos;
entry             162 sound/core/info.c 	size = entry->c.ops->read(entry, data->file_private_data,
entry             173 sound/core/info.c 	struct snd_info_entry *entry = data->entry;
entry             181 sound/core/info.c 		size_t maxsize = entry->size - pos;
entry             183 sound/core/info.c 		size = entry->c.ops->write(entry, data->file_private_data,
entry             194 sound/core/info.c 	struct snd_info_entry *entry = data->entry;
entry             197 sound/core/info.c 	if (entry->c.ops->poll)
entry             198 sound/core/info.c 		return entry->c.ops->poll(entry,
entry             201 sound/core/info.c 	if (entry->c.ops->read)
entry             203 sound/core/info.c 	if (entry->c.ops->write)
entry             212 sound/core/info.c 	struct snd_info_entry *entry = data->entry;
entry             214 sound/core/info.c 	if (!entry->c.ops->ioctl)
entry             216 sound/core/info.c 	return entry->c.ops->ioctl(entry, data->file_private_data,
entry             224 sound/core/info.c 	struct snd_info_entry *entry;
entry             229 sound/core/info.c 	entry = data->entry;
entry             230 sound/core/info.c 	if (!entry->c.ops->mmap)
entry             232 sound/core/info.c 	return entry->c.ops->mmap(entry, data->file_private_data,
entry             238 sound/core/info.c 	struct snd_info_entry *entry = PDE_DATA(inode);
entry             243 sound/core/info.c 	err = alloc_info_private(entry, &data);
entry             248 sound/core/info.c 	if (((mode == O_RDONLY || mode == O_RDWR) && !entry->c.ops->read) ||
entry             249 sound/core/info.c 	    ((mode == O_WRONLY || mode == O_RDWR) && !entry->c.ops->write)) {
entry             254 sound/core/info.c 	if (entry->c.ops->open) {
entry             255 sound/core/info.c 		err = entry->c.ops->open(entry, mode, &data->file_private_data);
entry             266 sound/core/info.c 	module_put(entry->module);
entry             275 sound/core/info.c 	struct snd_info_entry *entry = data->entry;
entry             277 sound/core/info.c 	if (entry->c.ops->release)
entry             278 sound/core/info.c 		entry->c.ops->release(entry, file->f_flags & O_ACCMODE,
entry             280 sound/core/info.c 	module_put(entry->module);
entry             307 sound/core/info.c 	struct snd_info_entry *entry = data->entry;
entry             313 sound/core/info.c 	if (!entry->c.text.write)
entry             322 sound/core/info.c 	mutex_lock(&entry->access);
entry             347 sound/core/info.c 	mutex_unlock(&entry->access);
entry             357 sound/core/info.c 	struct snd_info_entry *entry = data->entry;
entry             359 sound/core/info.c 	if (!entry->c.text.read) {
entry             363 sound/core/info.c 		entry->c.text.read(entry, data->rbuffer);
entry             370 sound/core/info.c 	struct snd_info_entry *entry = PDE_DATA(inode);
entry             375 sound/core/info.c 	err = alloc_info_private(entry, &data);
entry             384 sound/core/info.c 	if (entry->size)
entry             386 sound/core/info.c 				       entry->size);
entry             397 sound/core/info.c 	module_put(entry->module);
entry             407 sound/core/info.c 	struct snd_info_entry *entry = data->entry;
entry             409 sound/core/info.c 	if (data->wbuffer && entry->c.text.write)
entry             410 sound/core/info.c 		entry->c.text.write(entry, data->wbuffer);
entry             419 sound/core/info.c 	module_put(entry->module);
entry             437 sound/core/info.c 	struct snd_info_entry *entry;
entry             439 sound/core/info.c 	entry = snd_info_create_module_entry(mod, name, NULL);
entry             440 sound/core/info.c 	if (!entry)
entry             442 sound/core/info.c 	entry->mode = S_IFDIR | 0555;
entry             443 sound/core/info.c 	if (snd_info_register(entry) < 0) {
entry             444 sound/core/info.c 		snd_info_free_entry(entry);
entry             447 sound/core/info.c 	return entry;
entry             492 sound/core/info.c static void snd_card_id_read(struct snd_info_entry *entry,
entry             495 sound/core/info.c 	struct snd_card *card = entry->private_data;
entry             507 sound/core/info.c 	struct snd_info_entry *entry;
entry             513 sound/core/info.c 	entry = create_subdir(card->module, str);
entry             514 sound/core/info.c 	if (!entry)
entry             516 sound/core/info.c 	card->proc_root = entry;
entry             685 sound/core/info.c 	struct snd_info_entry *entry;
entry             686 sound/core/info.c 	entry = kzalloc(sizeof(*entry), GFP_KERNEL);
entry             687 sound/core/info.c 	if (entry == NULL)
entry             689 sound/core/info.c 	entry->name = kstrdup(name, GFP_KERNEL);
entry             690 sound/core/info.c 	if (entry->name == NULL) {
entry             691 sound/core/info.c 		kfree(entry);
entry             694 sound/core/info.c 	entry->mode = S_IFREG | 0444;
entry             695 sound/core/info.c 	entry->content = SNDRV_INFO_CONTENT_TEXT;
entry             696 sound/core/info.c 	mutex_init(&entry->access);
entry             697 sound/core/info.c 	INIT_LIST_HEAD(&entry->children);
entry             698 sound/core/info.c 	INIT_LIST_HEAD(&entry->list);
entry             699 sound/core/info.c 	entry->parent = parent;
entry             700 sound/core/info.c 	entry->module = module;
entry             703 sound/core/info.c 		list_add_tail(&entry->list, &parent->children);
entry             706 sound/core/info.c 	return entry;
entry             749 sound/core/info.c static void snd_info_disconnect(struct snd_info_entry *entry)
entry             753 sound/core/info.c 	if (!entry->p)
entry             755 sound/core/info.c 	list_for_each_entry(p, &entry->children, list)
entry             757 sound/core/info.c 	proc_remove(entry->p);
entry             758 sound/core/info.c 	entry->p = NULL;
entry             767 sound/core/info.c void snd_info_free_entry(struct snd_info_entry * entry)
entry             771 sound/core/info.c 	if (!entry)
entry             773 sound/core/info.c 	if (entry->p) {
entry             775 sound/core/info.c 		snd_info_disconnect(entry);
entry             780 sound/core/info.c 	list_for_each_entry_safe(p, n, &entry->children, list)
entry             783 sound/core/info.c 	p = entry->parent;
entry             786 sound/core/info.c 		list_del(&entry->list);
entry             789 sound/core/info.c 	kfree(entry->name);
entry             790 sound/core/info.c 	if (entry->private_free)
entry             791 sound/core/info.c 		entry->private_free(entry);
entry             792 sound/core/info.c 	kfree(entry);
entry             796 sound/core/info.c static int __snd_info_register(struct snd_info_entry *entry)
entry             800 sound/core/info.c 	if (snd_BUG_ON(!entry))
entry             802 sound/core/info.c 	root = entry->parent == NULL ? snd_proc_root->p : entry->parent->p;
entry             804 sound/core/info.c 	if (entry->p || !root)
entry             806 sound/core/info.c 	if (S_ISDIR(entry->mode)) {
entry             807 sound/core/info.c 		p = proc_mkdir_mode(entry->name, entry->mode, root);
entry             814 sound/core/info.c 		if (entry->content == SNDRV_INFO_CONTENT_DATA)
entry             818 sound/core/info.c 		p = proc_create_data(entry->name, entry->mode, root,
entry             819 sound/core/info.c 				     ops, entry);
entry             824 sound/core/info.c 		proc_set_size(p, entry->size);
entry             826 sound/core/info.c 	entry->p = p;
entry             841 sound/core/info.c int snd_info_register(struct snd_info_entry *entry)
entry             846 sound/core/info.c 	if (!entry->p) {
entry             847 sound/core/info.c 		err = __snd_info_register(entry);
entry             852 sound/core/info.c 	list_for_each_entry(p, &entry->children, list) {
entry             877 sound/core/info.c 			 void (*write)(struct snd_info_entry *entry,
entry             880 sound/core/info.c 	struct snd_info_entry *entry;
entry             882 sound/core/info.c 	entry = snd_info_create_card_entry(card, name, card->proc_root);
entry             883 sound/core/info.c 	if (!entry)
entry             885 sound/core/info.c 	snd_info_set_text_ops(entry, private_data, read);
entry             887 sound/core/info.c 		entry->mode |= 0200;
entry             888 sound/core/info.c 		entry->c.text.write = write;
entry             898 sound/core/info.c static void snd_info_version_read(struct snd_info_entry *entry, struct snd_info_buffer *buffer)
entry             907 sound/core/info.c 	struct snd_info_entry *entry;
entry             909 sound/core/info.c 	entry = snd_info_create_module_entry(THIS_MODULE, "version", NULL);
entry             910 sound/core/info.c 	if (entry == NULL)
entry             912 sound/core/info.c 	entry->c.text.read = snd_info_version_read;
entry             913 sound/core/info.c 	return snd_info_register(entry); /* freed in error path */
entry              74 sound/core/info_oss.c static void snd_sndstat_proc_read(struct snd_info_entry *entry,
entry              98 sound/core/info_oss.c 	struct snd_info_entry *entry;
entry             101 sound/core/info_oss.c 	entry = snd_info_create_module_entry(THIS_MODULE, "sndstat",
entry             103 sound/core/info_oss.c 	if (!entry)
entry             105 sound/core/info_oss.c 	entry->c.text.read = snd_sndstat_proc_read;
entry             106 sound/core/info_oss.c 	return snd_info_register(entry); /* freed in error path */
entry             789 sound/core/init.c static void snd_card_info_read(struct snd_info_entry *entry,
entry             835 sound/core/init.c static void snd_card_module_info_read(struct snd_info_entry *entry,
entry             853 sound/core/init.c 	struct snd_info_entry *entry;
entry             855 sound/core/init.c 	entry = snd_info_create_module_entry(THIS_MODULE, "cards", NULL);
entry             856 sound/core/init.c 	if (! entry)
entry             858 sound/core/init.c 	entry->c.text.read = snd_card_info_read;
entry             859 sound/core/init.c 	if (snd_info_register(entry) < 0)
entry             863 sound/core/init.c 	entry = snd_info_create_module_entry(THIS_MODULE, "modules", NULL);
entry             864 sound/core/init.c 	if (!entry)
entry             866 sound/core/init.c 	entry->c.text.read = snd_card_module_info_read;
entry             867 sound/core/init.c 	if (snd_info_register(entry) < 0)
entry            1142 sound/core/oss/mixer_oss.c static void snd_mixer_oss_proc_read(struct snd_info_entry *entry,
entry            1145 sound/core/oss/mixer_oss.c 	struct snd_mixer_oss *mixer = entry->private_data;
entry            1166 sound/core/oss/mixer_oss.c static void snd_mixer_oss_proc_write(struct snd_info_entry *entry,
entry            1169 sound/core/oss/mixer_oss.c 	struct snd_mixer_oss *mixer = entry->private_data;
entry            1228 sound/core/oss/mixer_oss.c 	struct snd_info_entry *entry;
entry            1230 sound/core/oss/mixer_oss.c 	entry = snd_info_create_card_entry(mixer->card, "oss_mixer",
entry            1232 sound/core/oss/mixer_oss.c 	if (! entry)
entry            1234 sound/core/oss/mixer_oss.c 	entry->content = SNDRV_INFO_CONTENT_TEXT;
entry            1235 sound/core/oss/mixer_oss.c 	entry->mode = S_IFREG | 0644;
entry            1236 sound/core/oss/mixer_oss.c 	entry->c.text.read = snd_mixer_oss_proc_read;
entry            1237 sound/core/oss/mixer_oss.c 	entry->c.text.write = snd_mixer_oss_proc_write;
entry            1238 sound/core/oss/mixer_oss.c 	entry->private_data = mixer;
entry            1239 sound/core/oss/mixer_oss.c 	if (snd_info_register(entry) < 0) {
entry            1240 sound/core/oss/mixer_oss.c 		snd_info_free_entry(entry);
entry            1241 sound/core/oss/mixer_oss.c 		entry = NULL;
entry            1243 sound/core/oss/mixer_oss.c 	mixer->proc_entry = entry;
entry            2910 sound/core/oss/pcm_oss.c static void snd_pcm_oss_proc_read(struct snd_info_entry *entry,
entry            2913 sound/core/oss/pcm_oss.c 	struct snd_pcm_str *pstr = entry->private_data;
entry            2945 sound/core/oss/pcm_oss.c static void snd_pcm_oss_proc_write(struct snd_info_entry *entry,
entry            2948 sound/core/oss/pcm_oss.c 	struct snd_pcm_str *pstr = entry->private_data;
entry            3027 sound/core/oss/pcm_oss.c 		struct snd_info_entry *entry;
entry            3031 sound/core/oss/pcm_oss.c 		if ((entry = snd_info_create_card_entry(pcm->card, "oss", pstr->proc_root)) != NULL) {
entry            3032 sound/core/oss/pcm_oss.c 			entry->content = SNDRV_INFO_CONTENT_TEXT;
entry            3033 sound/core/oss/pcm_oss.c 			entry->mode = S_IFREG | 0644;
entry            3034 sound/core/oss/pcm_oss.c 			entry->c.text.read = snd_pcm_oss_proc_read;
entry            3035 sound/core/oss/pcm_oss.c 			entry->c.text.write = snd_pcm_oss_proc_write;
entry            3036 sound/core/oss/pcm_oss.c 			entry->private_data = pstr;
entry            3037 sound/core/oss/pcm_oss.c 			if (snd_info_register(entry) < 0) {
entry            3038 sound/core/oss/pcm_oss.c 				snd_info_free_entry(entry);
entry            3039 sound/core/oss/pcm_oss.c 				entry = NULL;
entry            3042 sound/core/oss/pcm_oss.c 		pstr->oss.proc_entry = entry;
entry             363 sound/core/pcm.c static void snd_pcm_stream_proc_info_read(struct snd_info_entry *entry,
entry             366 sound/core/pcm.c 	snd_pcm_proc_info_read(((struct snd_pcm_str *)entry->private_data)->substream,
entry             370 sound/core/pcm.c static void snd_pcm_substream_proc_info_read(struct snd_info_entry *entry,
entry             373 sound/core/pcm.c 	snd_pcm_proc_info_read(entry->private_data, buffer);
entry             376 sound/core/pcm.c static void snd_pcm_substream_proc_hw_params_read(struct snd_info_entry *entry,
entry             379 sound/core/pcm.c 	struct snd_pcm_substream *substream = entry->private_data;
entry             413 sound/core/pcm.c static void snd_pcm_substream_proc_sw_params_read(struct snd_info_entry *entry,
entry             416 sound/core/pcm.c 	struct snd_pcm_substream *substream = entry->private_data;
entry             441 sound/core/pcm.c static void snd_pcm_substream_proc_status_read(struct snd_info_entry *entry,
entry             444 sound/core/pcm.c 	struct snd_pcm_substream *substream = entry->private_data;
entry             478 sound/core/pcm.c static void snd_pcm_xrun_injection_write(struct snd_info_entry *entry,
entry             481 sound/core/pcm.c 	struct snd_pcm_substream *substream = entry->private_data;
entry             486 sound/core/pcm.c static void snd_pcm_xrun_debug_read(struct snd_info_entry *entry,
entry             489 sound/core/pcm.c 	struct snd_pcm_str *pstr = entry->private_data;
entry             493 sound/core/pcm.c static void snd_pcm_xrun_debug_write(struct snd_info_entry *entry,
entry             496 sound/core/pcm.c 	struct snd_pcm_str *pstr = entry->private_data;
entry             506 sound/core/pcm.c 	struct snd_info_entry *entry;
entry             511 sound/core/pcm.c 	entry = snd_info_create_card_entry(pcm->card, name,
entry             513 sound/core/pcm.c 	if (!entry)
entry             515 sound/core/pcm.c 	entry->mode = S_IFDIR | 0555;
entry             516 sound/core/pcm.c 	pstr->proc_root = entry;
entry             517 sound/core/pcm.c 	entry = snd_info_create_card_entry(pcm->card, "info", pstr->proc_root);
entry             518 sound/core/pcm.c 	if (entry)
entry             519 sound/core/pcm.c 		snd_info_set_text_ops(entry, pstr, snd_pcm_stream_proc_info_read);
entry             521 sound/core/pcm.c 	entry = snd_info_create_card_entry(pcm->card, "xrun_debug",
entry             523 sound/core/pcm.c 	if (entry) {
entry             524 sound/core/pcm.c 		snd_info_set_text_ops(entry, pstr, snd_pcm_xrun_debug_read);
entry             525 sound/core/pcm.c 		entry->c.text.write = snd_pcm_xrun_debug_write;
entry             526 sound/core/pcm.c 		entry->mode |= 0200;
entry             545 sound/core/pcm.c 	struct snd_info_entry *entry;
entry             547 sound/core/pcm.c 	entry = snd_info_create_card_entry(substream->pcm->card, name,
entry             549 sound/core/pcm.c 	if (entry)
entry             550 sound/core/pcm.c 		snd_info_set_text_ops(entry, substream, read);
entry             551 sound/core/pcm.c 	return entry;
entry             556 sound/core/pcm.c 	struct snd_info_entry *entry;
entry             563 sound/core/pcm.c 	entry = snd_info_create_card_entry(card, name,
entry             565 sound/core/pcm.c 	if (!entry)
entry             567 sound/core/pcm.c 	entry->mode = S_IFDIR | 0555;
entry             568 sound/core/pcm.c 	substream->proc_root = entry;
entry             580 sound/core/pcm.c 	entry = create_substream_info_entry(substream, "xrun_injection", NULL);
entry             581 sound/core/pcm.c 	if (entry) {
entry             582 sound/core/pcm.c 		entry->c.text.write = snd_pcm_xrun_injection_write;
entry             583 sound/core/pcm.c 		entry->mode = S_IFREG | 0200;
entry            1162 sound/core/pcm.c static void snd_pcm_proc_read(struct snd_info_entry *entry,
entry            1186 sound/core/pcm.c 	struct snd_info_entry *entry;
entry            1188 sound/core/pcm.c 	entry = snd_info_create_module_entry(THIS_MODULE, "pcm", NULL);
entry            1189 sound/core/pcm.c 	if (entry) {
entry            1190 sound/core/pcm.c 		snd_info_set_text_ops(entry, NULL, snd_pcm_proc_read);
entry            1191 sound/core/pcm.c 		if (snd_info_register(entry) < 0) {
entry            1192 sound/core/pcm.c 			snd_info_free_entry(entry);
entry            1193 sound/core/pcm.c 			entry = NULL;
entry            1196 sound/core/pcm.c 	snd_pcm_proc_entry = entry;
entry             104 sound/core/pcm_memory.c static void snd_pcm_lib_preallocate_proc_read(struct snd_info_entry *entry,
entry             107 sound/core/pcm_memory.c 	struct snd_pcm_substream *substream = entry->private_data;
entry             116 sound/core/pcm_memory.c static void snd_pcm_lib_preallocate_max_proc_read(struct snd_info_entry *entry,
entry             119 sound/core/pcm_memory.c 	struct snd_pcm_substream *substream = entry->private_data;
entry             128 sound/core/pcm_memory.c static void snd_pcm_lib_preallocate_proc_write(struct snd_info_entry *entry,
entry             131 sound/core/pcm_memory.c 	struct snd_pcm_substream *substream = entry->private_data;
entry             172 sound/core/pcm_memory.c 	struct snd_info_entry *entry;
entry             174 sound/core/pcm_memory.c 	entry = snd_info_create_card_entry(substream->pcm->card, "prealloc",
entry             176 sound/core/pcm_memory.c 	if (entry) {
entry             177 sound/core/pcm_memory.c 		snd_info_set_text_ops(entry, substream,
entry             179 sound/core/pcm_memory.c 		entry->c.text.write = snd_pcm_lib_preallocate_proc_write;
entry             180 sound/core/pcm_memory.c 		entry->mode |= 0200;
entry             182 sound/core/pcm_memory.c 	entry = snd_info_create_card_entry(substream->pcm->card, "prealloc_max",
entry             184 sound/core/pcm_memory.c 	if (entry)
entry             185 sound/core/pcm_memory.c 		snd_info_set_text_ops(entry, substream,
entry            1449 sound/core/rawmidi.c static void snd_rawmidi_proc_info_read(struct snd_info_entry *entry,
entry            1456 sound/core/rawmidi.c 	rmidi = entry->private_data;
entry            1676 sound/core/rawmidi.c 	struct snd_info_entry *entry;
entry            1733 sound/core/rawmidi.c 	entry = snd_info_create_card_entry(rmidi->card, name, rmidi->card->proc_root);
entry            1734 sound/core/rawmidi.c 	if (entry) {
entry            1735 sound/core/rawmidi.c 		entry->private_data = rmidi;
entry            1736 sound/core/rawmidi.c 		entry->c.text.read = snd_rawmidi_proc_info_read;
entry            1737 sound/core/rawmidi.c 		if (snd_info_register(entry) < 0) {
entry            1738 sound/core/rawmidi.c 			snd_info_free_entry(entry);
entry            1739 sound/core/rawmidi.c 			entry = NULL;
entry            1742 sound/core/rawmidi.c 	rmidi->proc_entry = entry;
entry             259 sound/core/seq/oss/seq_oss.c info_read(struct snd_info_entry *entry, struct snd_info_buffer *buf)
entry             273 sound/core/seq/oss/seq_oss.c 	struct snd_info_entry *entry;
entry             275 sound/core/seq/oss/seq_oss.c 	entry = snd_info_create_module_entry(THIS_MODULE, SNDRV_SEQ_OSS_PROCNAME, snd_seq_root);
entry             276 sound/core/seq/oss/seq_oss.c 	if (entry == NULL)
entry             279 sound/core/seq/oss/seq_oss.c 	entry->content = SNDRV_INFO_CONTENT_TEXT;
entry             280 sound/core/seq/oss/seq_oss.c 	entry->private_data = NULL;
entry             281 sound/core/seq/oss/seq_oss.c 	entry->c.text.read = info_read;
entry             282 sound/core/seq/oss/seq_oss.c 	if (snd_info_register(entry) < 0) {
entry             283 sound/core/seq/oss/seq_oss.c 		snd_info_free_entry(entry);
entry             286 sound/core/seq/oss/seq_oss.c 	info_entry = entry;
entry            2445 sound/core/seq/seq_clientmgr.c void snd_seq_info_clients_read(struct snd_info_entry *entry, 
entry              24 sound/core/seq/seq_info.c 	struct snd_info_entry *entry;
entry              26 sound/core/seq/seq_info.c 	entry = snd_info_create_module_entry(THIS_MODULE, name, snd_seq_root);
entry              27 sound/core/seq/seq_info.c 	if (entry == NULL)
entry              29 sound/core/seq/seq_info.c 	entry->content = SNDRV_INFO_CONTENT_TEXT;
entry              30 sound/core/seq/seq_info.c 	entry->c.text.read = read;
entry              31 sound/core/seq/seq_info.c 	if (snd_info_register(entry) < 0) {
entry              32 sound/core/seq/seq_info.c 		snd_info_free_entry(entry);
entry              35 sound/core/seq/seq_info.c 	return entry;
entry              12 sound/core/seq/seq_info.h void snd_seq_info_clients_read(struct snd_info_entry *entry, struct snd_info_buffer *buffer);
entry              13 sound/core/seq/seq_info.h void snd_seq_info_timer_read(struct snd_info_entry *entry, struct snd_info_buffer *buffer);
entry              14 sound/core/seq/seq_info.h void snd_seq_info_queues_read(struct snd_info_entry *entry, struct snd_info_buffer *buffer);
entry             744 sound/core/seq/seq_queue.c void snd_seq_info_queues_read(struct snd_info_entry *entry, 
entry             462 sound/core/seq/seq_timer.c void snd_seq_info_timer_read(struct snd_info_entry *entry,
entry              74 sound/core/seq_device.c static void snd_seq_device_info(struct snd_info_entry *entry,
entry             345 sound/core/sound.c static void snd_minor_info_read(struct snd_info_entry *entry, struct snd_info_buffer *buffer)
entry             372 sound/core/sound.c 	struct snd_info_entry *entry;
entry             374 sound/core/sound.c 	entry = snd_info_create_module_entry(THIS_MODULE, "devices", NULL);
entry             375 sound/core/sound.c 	if (!entry)
entry             377 sound/core/sound.c 	entry->c.text.read = snd_minor_info_read;
entry             378 sound/core/sound.c 	return snd_info_register(entry); /* freed in error path */
entry             212 sound/core/sound_oss.c static void snd_minor_info_oss_read(struct snd_info_entry *entry,
entry             236 sound/core/sound_oss.c 	struct snd_info_entry *entry;
entry             238 sound/core/sound_oss.c 	entry = snd_info_create_module_entry(THIS_MODULE, "devices", snd_oss_root);
entry             239 sound/core/sound_oss.c 	if (!entry)
entry             241 sound/core/sound_oss.c 	entry->c.text.read = snd_minor_info_oss_read;
entry             242 sound/core/sound_oss.c 	return snd_info_register(entry); /* freed in error path */
entry            1216 sound/core/timer.c static void snd_timer_proc_read(struct snd_info_entry *entry,
entry            1266 sound/core/timer.c 	struct snd_info_entry *entry;
entry            1268 sound/core/timer.c 	entry = snd_info_create_module_entry(THIS_MODULE, "timers", NULL);
entry            1269 sound/core/timer.c 	if (entry != NULL) {
entry            1270 sound/core/timer.c 		entry->c.text.read = snd_timer_proc_read;
entry            1271 sound/core/timer.c 		if (snd_info_register(entry) < 0) {
entry            1272 sound/core/timer.c 			snd_info_free_entry(entry);
entry            1273 sound/core/timer.c 			entry = NULL;
entry            1276 sound/core/timer.c 	snd_timer_proc_entry = entry;
entry            1107 sound/drivers/aloop.c static void print_cable_info(struct snd_info_entry *entry,
entry            1110 sound/drivers/aloop.c 	struct loopback *loopback = entry->private_data;
entry            1114 sound/drivers/aloop.c 	num = entry->name[strlen(entry->name)-1];
entry             973 sound/drivers/dummy.c static void dummy_proc_read(struct snd_info_entry *entry,
entry             976 sound/drivers/dummy.c 	struct snd_dummy *dummy = entry->private_data;
entry             995 sound/drivers/dummy.c static void dummy_proc_write(struct snd_info_entry *entry,
entry             998 sound/drivers/dummy.c 	struct snd_dummy *dummy = entry->private_data;
entry              12 sound/drivers/opl4/opl4_proc.c static int snd_opl4_mem_proc_open(struct snd_info_entry *entry,
entry              15 sound/drivers/opl4/opl4_proc.c 	struct snd_opl4 *opl4 = entry->private_data;
entry              27 sound/drivers/opl4/opl4_proc.c static int snd_opl4_mem_proc_release(struct snd_info_entry *entry,
entry              30 sound/drivers/opl4/opl4_proc.c 	struct snd_opl4 *opl4 = entry->private_data;
entry              38 sound/drivers/opl4/opl4_proc.c static ssize_t snd_opl4_mem_proc_read(struct snd_info_entry *entry,
entry              43 sound/drivers/opl4/opl4_proc.c 	struct snd_opl4 *opl4 = entry->private_data;
entry              58 sound/drivers/opl4/opl4_proc.c static ssize_t snd_opl4_mem_proc_write(struct snd_info_entry *entry,
entry              64 sound/drivers/opl4/opl4_proc.c 	struct snd_opl4 *opl4 = entry->private_data;
entry              88 sound/drivers/opl4/opl4_proc.c 	struct snd_info_entry *entry;
entry              90 sound/drivers/opl4/opl4_proc.c 	entry = snd_info_create_card_entry(opl4->card, "opl4-mem", opl4->card->proc_root);
entry              91 sound/drivers/opl4/opl4_proc.c 	if (entry) {
entry              94 sound/drivers/opl4/opl4_proc.c 			entry->mode |= 0200;
entry              95 sound/drivers/opl4/opl4_proc.c 			entry->size = 4 * 1024 * 1024;
entry              98 sound/drivers/opl4/opl4_proc.c 			entry->size = 1 * 1024 * 1024;
entry             100 sound/drivers/opl4/opl4_proc.c 		entry->content = SNDRV_INFO_CONTENT_DATA;
entry             101 sound/drivers/opl4/opl4_proc.c 		entry->c.ops = &snd_opl4_mem_proc_ops;
entry             102 sound/drivers/opl4/opl4_proc.c 		entry->module = THIS_MODULE;
entry             103 sound/drivers/opl4/opl4_proc.c 		entry->private_data = opl4;
entry             105 sound/drivers/opl4/opl4_proc.c 	opl4->proc_entry = entry;
entry             588 sound/drivers/vx/vx_core.c static void vx_proc_read(struct snd_info_entry *entry, struct snd_info_buffer *buffer)
entry             590 sound/drivers/vx/vx_core.c 	struct vx_core *chip = entry->private_data;
entry             902 sound/firewire/amdtp-stream.c 	} *entry, initial_state[] = {
entry             927 sound/firewire/amdtp-stream.c 		entry = &initial_state[s->sfc];
entry             930 sound/firewire/amdtp-stream.c 		s->ctx_data.rx.data_block_state = entry->data_block;
entry             931 sound/firewire/amdtp-stream.c 		s->ctx_data.rx.syt_offset_state = entry->syt_offset;
entry             203 sound/firewire/bebob/bebob.c 		if (bebob->entry->model_id == MODEL_MAUDIO_FW1814)
entry             247 sound/firewire/bebob/bebob.c bebob_probe(struct fw_unit *unit, const struct ieee1394_device_id *entry)
entry             252 sound/firewire/bebob/bebob.c 	if (entry->vendor_id == VEN_FOCUSRITE &&
entry             253 sound/firewire/bebob/bebob.c 	    entry->model_id == MODEL_FOCUSRITE_SAFFIRE_BOTH)
entry             255 sound/firewire/bebob/bebob.c 	else if (entry->vendor_id == VEN_MAUDIO1 &&
entry             256 sound/firewire/bebob/bebob.c 		 entry->model_id == MODEL_MAUDIO_AUDIOPHILE_BOTH &&
entry             260 sound/firewire/bebob/bebob.c 		spec = (const struct snd_bebob_spec *)entry->driver_data;
entry             263 sound/firewire/bebob/bebob.c 		if (entry->vendor_id == VEN_MAUDIO1 ||
entry             264 sound/firewire/bebob/bebob.c 		    entry->vendor_id == VEN_MAUDIO2)
entry             278 sound/firewire/bebob/bebob.c 	bebob->entry = entry;
entry             287 sound/firewire/bebob/bebob.c 	if (entry->vendor_id != VEN_MAUDIO1 ||
entry             288 sound/firewire/bebob/bebob.c 	    (entry->model_id != MODEL_MAUDIO_FW1814 &&
entry             289 sound/firewire/bebob/bebob.c 	     entry->model_id != MODEL_MAUDIO_PROJECTMIX)) {
entry              89 sound/firewire/bebob/bebob.h 	const struct ieee1394_device_id *entry;
entry              35 sound/firewire/bebob/bebob_proc.c proc_read_hw_info(struct snd_info_entry *entry,
entry              38 sound/firewire/bebob/bebob_proc.c 	struct snd_bebob *bebob = entry->private_data;
entry              71 sound/firewire/bebob/bebob_proc.c proc_read_meters(struct snd_info_entry *entry,
entry              74 sound/firewire/bebob/bebob_proc.c 	struct snd_bebob *bebob = entry->private_data;
entry             104 sound/firewire/bebob/bebob_proc.c proc_read_formation(struct snd_info_entry *entry,
entry             107 sound/firewire/bebob/bebob_proc.c 	struct snd_bebob *bebob = entry->private_data;
entry             131 sound/firewire/bebob/bebob_proc.c proc_read_clock(struct snd_info_entry *entry,
entry             139 sound/firewire/bebob/bebob_proc.c 	struct snd_bebob *bebob = entry->private_data;
entry             162 sound/firewire/bebob/bebob_proc.c 	struct snd_info_entry *entry;
entry             164 sound/firewire/bebob/bebob_proc.c 	entry = snd_info_create_card_entry(bebob->card, name, root);
entry             165 sound/firewire/bebob/bebob_proc.c 	if (entry)
entry             166 sound/firewire/bebob/bebob_proc.c 		snd_info_set_text_ops(entry, bebob, op);
entry              27 sound/firewire/dice/dice-presonus.c 	} *entry, entries[] = {
entry              44 sound/firewire/dice/dice-presonus.c 		entry = entries + i;
entry              45 sound/firewire/dice/dice-presonus.c 		if (entry->model_id == model_id)
entry              51 sound/firewire/dice/dice-presonus.c 	memcpy(dice->tx_pcm_chs, entry->spec->tx_pcm_chs,
entry              53 sound/firewire/dice/dice-presonus.c 	memcpy(dice->rx_pcm_chs, entry->spec->rx_pcm_chs,
entry              56 sound/firewire/dice/dice-presonus.c 	if (entry->spec->has_midi) {
entry              56 sound/firewire/dice/dice-proc.c static void dice_proc_read(struct snd_info_entry *entry,
entry              70 sound/firewire/dice/dice-proc.c 	struct snd_dice *dice = entry->private_data;
entry             245 sound/firewire/dice/dice-proc.c static void dice_proc_read_formation(struct snd_info_entry *entry,
entry             253 sound/firewire/dice/dice-proc.c 	struct snd_dice *dice = entry->private_data;
entry             281 sound/firewire/dice/dice-proc.c 		     void (*op)(struct snd_info_entry *entry,
entry             284 sound/firewire/dice/dice-proc.c 	struct snd_info_entry *entry;
entry             286 sound/firewire/dice/dice-proc.c 	entry = snd_info_create_card_entry(dice->card, name, root);
entry             287 sound/firewire/dice/dice-proc.c 	if (entry)
entry             288 sound/firewire/dice/dice-proc.c 		snd_info_set_text_ops(entry, dice, op);
entry              63 sound/firewire/dice/dice-tcelectronic.c 	} *entry, entries[] = {
entry              86 sound/firewire/dice/dice-tcelectronic.c 		entry = entries + i;
entry              87 sound/firewire/dice/dice-tcelectronic.c 		if (entry->model_id == model_id)
entry              93 sound/firewire/dice/dice-tcelectronic.c 	memcpy(dice->tx_pcm_chs, entry->spec->tx_pcm_chs,
entry              95 sound/firewire/dice/dice-tcelectronic.c 	memcpy(dice->rx_pcm_chs, entry->spec->rx_pcm_chs,
entry              98 sound/firewire/dice/dice-tcelectronic.c 	if (entry->spec->has_midi) {
entry             196 sound/firewire/dice/dice.c 		      const struct ieee1394_device_id *entry)
entry             201 sound/firewire/dice/dice.c 	if (!entry->driver_data && entry->vendor_id != OUI_SSL) {
entry             214 sound/firewire/dice/dice.c 	if (!entry->driver_data) {
entry             218 sound/firewire/dice/dice.c 				(snd_dice_detect_formats_t)entry->driver_data;
entry              25 sound/firewire/digi00x/digi00x-proc.c static void proc_read_clock(struct snd_info_entry *entry,
entry              38 sound/firewire/digi00x/digi00x-proc.c 	struct snd_dg00x *dg00x = entry->private_data;
entry              70 sound/firewire/digi00x/digi00x-proc.c 	struct snd_info_entry *root, *entry;
entry              83 sound/firewire/digi00x/digi00x-proc.c 	entry = snd_info_create_card_entry(dg00x->card, "clock", root);
entry              84 sound/firewire/digi00x/digi00x-proc.c 	if (entry)
entry              85 sound/firewire/digi00x/digi00x-proc.c 		snd_info_set_text_ops(entry, dg00x, proc_read_clock);
entry             107 sound/firewire/digi00x/digi00x.c 			   const struct ieee1394_device_id *entry)
entry             124 sound/firewire/digi00x/digi00x.c 	dg00x->is_console = entry->model_id == MODEL_CONSOLE;
entry              27 sound/firewire/fireface/ff-proc.c static void proc_dump_status(struct snd_info_entry *entry,
entry              30 sound/firewire/fireface/ff-proc.c 	struct snd_ff *ff = entry->private_data;
entry              40 sound/firewire/fireface/ff-proc.c 	struct snd_info_entry *entry;
entry              42 sound/firewire/fireface/ff-proc.c 	entry = snd_info_create_card_entry(ff->card, name, root);
entry              43 sound/firewire/fireface/ff-proc.c 	if (entry)
entry              44 sound/firewire/fireface/ff-proc.c 		snd_info_set_text_ops(entry, ff, op);
entry              90 sound/firewire/fireface/ff.c 			   const struct ieee1394_device_id *entry)
entry             104 sound/firewire/fireface/ff.c 	ff->spec = (const struct snd_ff_spec *)entry->driver_data;
entry             282 sound/firewire/fireworks/fireworks.c efw_probe(struct fw_unit *unit, const struct ieee1394_device_id *entry)
entry              28 sound/firewire/fireworks/fireworks_proc.c proc_read_hwinfo(struct snd_info_entry *entry, struct snd_info_buffer *buffer)
entry              30 sound/firewire/fireworks/fireworks_proc.c 	struct snd_efw *efw = entry->private_data;
entry             105 sound/firewire/fireworks/fireworks_proc.c proc_read_clock(struct snd_info_entry *entry, struct snd_info_buffer *buffer)
entry             107 sound/firewire/fireworks/fireworks_proc.c 	struct snd_efw *efw = entry->private_data;
entry             127 sound/firewire/fireworks/fireworks_proc.c proc_read_phys_meters(struct snd_info_entry *entry,
entry             130 sound/firewire/fireworks/fireworks_proc.c 	struct snd_efw *efw = entry->private_data;
entry             178 sound/firewire/fireworks/fireworks_proc.c proc_read_queues_state(struct snd_info_entry *entry,
entry             181 sound/firewire/fireworks/fireworks_proc.c 	struct snd_efw *efw = entry->private_data;
entry             198 sound/firewire/fireworks/fireworks_proc.c 	struct snd_info_entry *entry;
entry             200 sound/firewire/fireworks/fireworks_proc.c 	entry = snd_info_create_card_entry(efw->card, name, root);
entry             201 sound/firewire/fireworks/fireworks_proc.c 	if (entry)
entry             202 sound/firewire/fireworks/fireworks_proc.c 		snd_info_set_text_ops(entry, efw, op);
entry              24 sound/firewire/motu/motu-proc.c static void proc_read_clock(struct snd_info_entry *entry,
entry              28 sound/firewire/motu/motu-proc.c 	struct snd_motu *motu = entry->private_data;
entry              42 sound/firewire/motu/motu-proc.c static void proc_read_format(struct snd_info_entry *entry,
entry              45 sound/firewire/motu/motu-proc.c 	struct snd_motu *motu = entry->private_data;
entry              86 sound/firewire/motu/motu-proc.c 	struct snd_info_entry *entry;
entry              88 sound/firewire/motu/motu-proc.c 	entry = snd_info_create_card_entry(motu->card, name, root);
entry              89 sound/firewire/motu/motu-proc.c 	if (entry)
entry              90 sound/firewire/motu/motu-proc.c 		snd_info_set_text_ops(entry, motu, op);
entry             120 sound/firewire/motu/motu.c 		      const struct ieee1394_device_id *entry)
entry             131 sound/firewire/motu/motu.c 	motu->spec = (const struct snd_motu_spec *)entry->driver_data;
entry              10 sound/firewire/oxfw/oxfw-proc.c static void proc_read_formation(struct snd_info_entry *entry,
entry              13 sound/firewire/oxfw/oxfw-proc.c 	struct snd_oxfw *oxfw = entry->private_data;
entry              82 sound/firewire/oxfw/oxfw-proc.c 	struct snd_info_entry *entry;
entry              84 sound/firewire/oxfw/oxfw-proc.c 	entry = snd_info_create_card_entry(oxfw->card, name, root);
entry              85 sound/firewire/oxfw/oxfw-proc.c 	if (entry)
entry              86 sound/firewire/oxfw/oxfw-proc.c 		snd_info_set_text_ops(entry, oxfw, op);
entry              92 sound/firewire/oxfw/oxfw.c 	if (oxfw->entry->vendor_id == VENDOR_GRIFFIN ||
entry              93 sound/firewire/oxfw/oxfw.c 	    oxfw->entry->vendor_id == VENDOR_LACIE) {
entry              94 sound/firewire/oxfw/oxfw.c 		info = (const struct compat_info *)oxfw->entry->driver_data;
entry             135 sound/firewire/oxfw/oxfw.c 	if (oxfw->entry->vendor_id == VENDOR_GRIFFIN)
entry             137 sound/firewire/oxfw/oxfw.c 	if (oxfw->entry->vendor_id == VENDOR_LACIE)
entry             144 sound/firewire/oxfw/oxfw.c 	if (oxfw->entry->vendor_id == OUI_STANTON) {
entry             156 sound/firewire/oxfw/oxfw.c 	if (oxfw->entry->vendor_id == VENDOR_TASCAM) {
entry             241 sound/firewire/oxfw/oxfw.c 		      const struct ieee1394_device_id *entry)
entry             245 sound/firewire/oxfw/oxfw.c 	if (entry->vendor_id == VENDOR_LOUD && !detect_loud_models(unit))
entry             255 sound/firewire/oxfw/oxfw.c 	oxfw->entry = entry;
entry             281 sound/firewire/oxfw/oxfw.c 		if (oxfw->entry->vendor_id == OUI_STANTON)
entry              64 sound/firewire/oxfw/oxfw.h 	const struct ieee1394_device_id *entry;
entry             158 sound/firewire/tascam/amdtp-tascam.c 				struct snd_firewire_tascam_change *entry =
entry             163 sound/firewire/tascam/amdtp-tascam.c 				entry->index = index;
entry             164 sound/firewire/tascam/amdtp-tascam.c 				entry->before = before;
entry             165 sound/firewire/tascam/amdtp-tascam.c 				entry->after = after;
entry              10 sound/firewire/tascam/tascam-proc.c static void proc_read_firmware(struct snd_info_entry *entry,
entry              13 sound/firewire/tascam/tascam-proc.c 	struct snd_tscm *tscm = entry->private_data;
entry              57 sound/firewire/tascam/tascam-proc.c 	struct snd_info_entry *entry;
entry              59 sound/firewire/tascam/tascam-proc.c 	entry = snd_info_create_card_entry(tscm->card, name, root);
entry              60 sound/firewire/tascam/tascam-proc.c 	if (entry)
entry              61 sound/firewire/tascam/tascam-proc.c 		snd_info_set_text_ops(entry, tscm, op);
entry             150 sound/firewire/tascam/tascam.c 			   const struct ieee1394_device_id *entry)
entry             466 sound/i2c/other/ak4113.c static void snd_ak4113_proc_regs_read(struct snd_info_entry *entry,
entry             469 sound/i2c/other/ak4113.c 	struct ak4113 *ak4113 = entry->private_data;
entry             439 sound/i2c/other/ak4114.c static void snd_ak4114_proc_regs_read(struct snd_info_entry *entry,
entry             442 sound/i2c/other/ak4114.c 	struct ak4114 *ak4114 = entry->private_data;
entry             849 sound/i2c/other/ak4xxx-adda.c static void proc_regs_read(struct snd_info_entry *entry,
entry             852 sound/i2c/other/ak4xxx-adda.c 	struct snd_akm4xxx *ak = entry->private_data;
entry             102 sound/isa/gus/gus_irq.c static void snd_gus_irq_info_read(struct snd_info_entry *entry, 
entry             109 sound/isa/gus/gus_irq.c 	gus = entry->private_data;
entry              14 sound/isa/gus/gus_mem.c static void snd_gf1_mem_info_read(struct snd_info_entry *entry, 
entry             269 sound/isa/gus/gus_mem.c static void snd_gf1_mem_info_read(struct snd_info_entry *entry, 
entry             278 sound/isa/gus/gus_mem.c 	gus = entry->private_data;
entry              19 sound/isa/gus/gus_mem_proc.c static ssize_t snd_gf1_mem_proc_dump(struct snd_info_entry *entry,
entry              24 sound/isa/gus/gus_mem_proc.c 	struct gus_proc_private *priv = entry->private_data;
entry              34 sound/isa/gus/gus_mem_proc.c static void snd_gf1_mem_proc_free(struct snd_info_entry *entry)
entry              36 sound/isa/gus/gus_mem_proc.c 	struct gus_proc_private *priv = entry->private_data;
entry              49 sound/isa/gus/gus_mem_proc.c 	struct snd_info_entry *entry;
entry              58 sound/isa/gus/gus_mem_proc.c 			if (! snd_card_proc_new(gus->card, name, &entry)) {
entry              59 sound/isa/gus/gus_mem_proc.c 				entry->content = SNDRV_INFO_CONTENT_DATA;
entry              60 sound/isa/gus/gus_mem_proc.c 				entry->private_data = priv;
entry              61 sound/isa/gus/gus_mem_proc.c 				entry->private_free = snd_gf1_mem_proc_free;
entry              62 sound/isa/gus/gus_mem_proc.c 				entry->c.ops = &snd_gf1_mem_proc_ops;
entry              64 sound/isa/gus/gus_mem_proc.c 				priv->size = entry->size = gus->gf1.mem_alloc.banks_8[idx].size;
entry              76 sound/isa/gus/gus_mem_proc.c 			if (! snd_card_proc_new(gus->card, name, &entry)) {
entry              77 sound/isa/gus/gus_mem_proc.c 				entry->content = SNDRV_INFO_CONTENT_DATA;
entry              78 sound/isa/gus/gus_mem_proc.c 				entry->private_data = priv;
entry              79 sound/isa/gus/gus_mem_proc.c 				entry->private_free = snd_gf1_mem_proc_free;
entry              80 sound/isa/gus/gus_mem_proc.c 				entry->c.ops = &snd_gf1_mem_proc_ops;
entry              82 sound/isa/gus/gus_mem_proc.c 				priv->size = entry->size = gus->gf1.rom_memory;
entry             882 sound/isa/opti9xx/miro.c static void snd_miro_proc_read(struct snd_info_entry * entry, 
entry             885 sound/isa/opti9xx/miro.c 	struct snd_miro *miro = (struct snd_miro *) entry->private_data;
entry              97 sound/isa/sb/sb16_csp.c static void info_read(struct snd_info_entry *entry, struct snd_info_buffer *buffer);
entry            1121 sound/isa/sb/sb16_csp.c static void info_read(struct snd_info_entry *entry, struct snd_info_buffer *buffer)
entry            1123 sound/isa/sb/sb16_csp.c 	struct snd_sb_csp *p = entry->private_data;
entry             328 sound/pci/ac97/ac97_proc.c static void snd_ac97_proc_read(struct snd_info_entry *entry, struct snd_info_buffer *buffer)
entry             330 sound/pci/ac97/ac97_proc.c 	struct snd_ac97 *ac97 = entry->private_data;
entry             363 sound/pci/ac97/ac97_proc.c static void snd_ac97_proc_regs_write(struct snd_info_entry *entry, struct snd_info_buffer *buffer)
entry             365 sound/pci/ac97/ac97_proc.c 	struct snd_ac97 *ac97 = entry->private_data;
entry             390 sound/pci/ac97/ac97_proc.c static void snd_ac97_proc_regs_read(struct snd_info_entry *entry, 
entry             393 sound/pci/ac97/ac97_proc.c 	struct snd_ac97 *ac97 = entry->private_data;
entry             416 sound/pci/ac97/ac97_proc.c 	struct snd_info_entry *entry;
entry             424 sound/pci/ac97/ac97_proc.c 	entry = snd_info_create_card_entry(ac97->bus->card, name,
entry             426 sound/pci/ac97/ac97_proc.c 	if (entry)
entry             427 sound/pci/ac97/ac97_proc.c 		snd_info_set_text_ops(entry, ac97, snd_ac97_proc_read);
entry             428 sound/pci/ac97/ac97_proc.c 	ac97->proc = entry;
entry             430 sound/pci/ac97/ac97_proc.c 	entry = snd_info_create_card_entry(ac97->bus->card, name,
entry             432 sound/pci/ac97/ac97_proc.c 	if (entry) {
entry             433 sound/pci/ac97/ac97_proc.c 		snd_info_set_text_ops(entry, ac97, snd_ac97_proc_regs_read);
entry             435 sound/pci/ac97/ac97_proc.c 		entry->mode |= 0200;
entry             436 sound/pci/ac97/ac97_proc.c 		entry->c.text.write = snd_ac97_proc_regs_write;
entry             439 sound/pci/ac97/ac97_proc.c 	ac97->proc_regs = entry;
entry             452 sound/pci/ac97/ac97_proc.c 	struct snd_info_entry *entry;
entry             456 sound/pci/ac97/ac97_proc.c 	entry = snd_info_create_card_entry(bus->card, name,
entry             458 sound/pci/ac97/ac97_proc.c 	if (entry)
entry             459 sound/pci/ac97/ac97_proc.c 		entry->mode = S_IFDIR | 0555;
entry             460 sound/pci/ac97/ac97_proc.c 	bus->proc = entry;
entry             644 sound/pci/ad1889.c snd_ad1889_proc_read(struct snd_info_entry *entry, struct snd_info_buffer *buffer)
entry             646 sound/pci/ad1889.c 	struct snd_ad1889 *chip = entry->private_data;
entry             454 sound/pci/ak4531_codec.c static void snd_ak4531_proc_read(struct snd_info_entry *entry, 
entry             457 sound/pci/ak4531_codec.c 	struct snd_ak4531 *ak4531 = entry->private_data;
entry            2027 sound/pci/ali5451/ali5451.c static void snd_ali_proc_read(struct snd_info_entry *entry,
entry            2030 sound/pci/ali5451/ali5451.c 	struct snd_ali *codec = entry->private_data;
entry            2720 sound/pci/asihpi/asihpi.c snd_asihpi_proc_read(struct snd_info_entry *entry,
entry            2723 sound/pci/asihpi/asihpi.c 	struct snd_card_asihpi *asihpi = entry->private_data;
entry            1520 sound/pci/atiixp.c static void snd_atiixp_proc_read(struct snd_info_entry *entry,
entry            1523 sound/pci/atiixp.c 	struct atiixp *chip = entry->private_data;
entry            1147 sound/pci/atiixp_modem.c static void snd_atiixp_proc_read(struct snd_info_entry *entry,
entry            1150 sound/pci/atiixp_modem.c 	struct atiixp_modem *chip = entry->private_data;
entry             258 sound/pci/ca0106/ca0106_proc.c static void snd_ca0106_proc_iec958(struct snd_info_entry *entry, 
entry             261 sound/pci/ca0106/ca0106_proc.c 	struct snd_ca0106 *emu = entry->private_data;
entry             280 sound/pci/ca0106/ca0106_proc.c static void snd_ca0106_proc_reg_write32(struct snd_info_entry *entry, 
entry             283 sound/pci/ca0106/ca0106_proc.c 	struct snd_ca0106 *emu = entry->private_data;
entry             298 sound/pci/ca0106/ca0106_proc.c static void snd_ca0106_proc_reg_read32(struct snd_info_entry *entry, 
entry             301 sound/pci/ca0106/ca0106_proc.c 	struct snd_ca0106 *emu = entry->private_data;
entry             314 sound/pci/ca0106/ca0106_proc.c static void snd_ca0106_proc_reg_read16(struct snd_info_entry *entry, 
entry             317 sound/pci/ca0106/ca0106_proc.c 	struct snd_ca0106 *emu = entry->private_data;
entry             330 sound/pci/ca0106/ca0106_proc.c static void snd_ca0106_proc_reg_read8(struct snd_info_entry *entry, 
entry             333 sound/pci/ca0106/ca0106_proc.c 	struct snd_ca0106 *emu = entry->private_data;
entry             346 sound/pci/ca0106/ca0106_proc.c static void snd_ca0106_proc_reg_read1(struct snd_info_entry *entry, 
entry             349 sound/pci/ca0106/ca0106_proc.c 	struct snd_ca0106 *emu = entry->private_data;
entry             364 sound/pci/ca0106/ca0106_proc.c static void snd_ca0106_proc_reg_read2(struct snd_info_entry *entry, 
entry             367 sound/pci/ca0106/ca0106_proc.c 	struct snd_ca0106 *emu = entry->private_data;
entry             382 sound/pci/ca0106/ca0106_proc.c static void snd_ca0106_proc_reg_write(struct snd_info_entry *entry, 
entry             385 sound/pci/ca0106/ca0106_proc.c 	struct snd_ca0106 *emu = entry->private_data;
entry             396 sound/pci/ca0106/ca0106_proc.c static void snd_ca0106_proc_i2c_write(struct snd_info_entry *entry, 
entry             399 sound/pci/ca0106/ca0106_proc.c 	struct snd_ca0106 *emu = entry->private_data;
entry            2762 sound/pci/cmipci.c static void snd_cmipci_proc_read(struct snd_info_entry *entry, 
entry            2765 sound/pci/cmipci.c 	struct cmipci *cm = entry->private_data;
entry            1115 sound/pci/cs4281.c static void snd_cs4281_proc_read(struct snd_info_entry *entry, 
entry            1118 sound/pci/cs4281.c 	struct cs4281 *chip = entry->private_data;
entry            1125 sound/pci/cs4281.c static ssize_t snd_cs4281_BA0_read(struct snd_info_entry *entry,
entry            1130 sound/pci/cs4281.c 	struct cs4281 *chip = entry->private_data;
entry            1137 sound/pci/cs4281.c static ssize_t snd_cs4281_BA1_read(struct snd_info_entry *entry,
entry            1142 sound/pci/cs4281.c 	struct cs4281 *chip = entry->private_data;
entry            1159 sound/pci/cs4281.c 	struct snd_info_entry *entry;
entry            1162 sound/pci/cs4281.c 	if (! snd_card_proc_new(chip->card, "cs4281_BA0", &entry)) {
entry            1163 sound/pci/cs4281.c 		entry->content = SNDRV_INFO_CONTENT_DATA;
entry            1164 sound/pci/cs4281.c 		entry->private_data = chip;
entry            1165 sound/pci/cs4281.c 		entry->c.ops = &snd_cs4281_proc_ops_BA0;
entry            1166 sound/pci/cs4281.c 		entry->size = CS4281_BA0_SIZE;
entry            1168 sound/pci/cs4281.c 	if (! snd_card_proc_new(chip->card, "cs4281_BA1", &entry)) {
entry            1169 sound/pci/cs4281.c 		entry->content = SNDRV_INFO_CONTENT_DATA;
entry            1170 sound/pci/cs4281.c 		entry->private_data = chip;
entry            1171 sound/pci/cs4281.c 		entry->c.ops = &snd_cs4281_proc_ops_BA1;
entry            1172 sound/pci/cs4281.c 		entry->size = CS4281_BA1_SIZE;
entry             421 sound/pci/cs46xx/cs46xx_lib.c 		struct dsp_symbol_entry *entry =
entry             425 sound/pci/cs46xx/cs46xx_lib.c 		entry->address = le32_to_cpu(fwdat[fwlen++]);
entry             426 sound/pci/cs46xx/cs46xx_lib.c 		memcpy(entry->symbol_name, &fwdat[fwlen], DSP_MAX_SYMBOL_NAME - 1);
entry             428 sound/pci/cs46xx/cs46xx_lib.c 		entry->symbol_type = le32_to_cpu(fwdat[fwlen++]);
entry             441 sound/pci/cs46xx/cs46xx_lib.c 		struct dsp_segment_desc *entry = &module->segments[i];
entry             444 sound/pci/cs46xx/cs46xx_lib.c 		entry->segment_type = le32_to_cpu(fwdat[fwlen++]);
entry             445 sound/pci/cs46xx/cs46xx_lib.c 		entry->offset = le32_to_cpu(fwdat[fwlen++]);
entry             446 sound/pci/cs46xx/cs46xx_lib.c 		entry->size = le32_to_cpu(fwdat[fwlen++]);
entry             447 sound/pci/cs46xx/cs46xx_lib.c 		if (fwlen + entry->size > fwsize)
entry             449 sound/pci/cs46xx/cs46xx_lib.c 		entry->data = kmalloc_array(entry->size, 4, GFP_KERNEL);
entry             450 sound/pci/cs46xx/cs46xx_lib.c 		if (!entry->data)
entry             452 sound/pci/cs46xx/cs46xx_lib.c 		memcpy_le32(entry->data, &fwdat[fwlen], entry->size * 4);
entry             453 sound/pci/cs46xx/cs46xx_lib.c 		fwlen += entry->size;
entry            2812 sound/pci/cs46xx/cs46xx_lib.c static ssize_t snd_cs46xx_io_read(struct snd_info_entry *entry,
entry            2817 sound/pci/cs46xx/cs46xx_lib.c 	struct snd_cs46xx_region *region = entry->private_data;
entry            2830 sound/pci/cs46xx/cs46xx_lib.c 	struct snd_info_entry *entry;
entry            2835 sound/pci/cs46xx/cs46xx_lib.c 		if (! snd_card_proc_new(card, region->name, &entry)) {
entry            2836 sound/pci/cs46xx/cs46xx_lib.c 			entry->content = SNDRV_INFO_CONTENT_DATA;
entry            2837 sound/pci/cs46xx/cs46xx_lib.c 			entry->private_data = chip;
entry            2838 sound/pci/cs46xx/cs46xx_lib.c 			entry->c.ops = &snd_cs46xx_proc_io_ops;
entry            2839 sound/pci/cs46xx/cs46xx_lib.c 			entry->size = region->size;
entry            2840 sound/pci/cs46xx/cs46xx_lib.c 			entry->mode = S_IFREG | 0400;
entry             493 sound/pci/cs46xx/dsp_spos.c static void cs46xx_dsp_proc_symbol_table_read (struct snd_info_entry *entry,
entry             496 sound/pci/cs46xx/dsp_spos.c 	struct snd_cs46xx *chip = entry->private_data;
entry             521 sound/pci/cs46xx/dsp_spos.c static void cs46xx_dsp_proc_modules_read (struct snd_info_entry *entry,
entry             524 sound/pci/cs46xx/dsp_spos.c 	struct snd_cs46xx *chip = entry->private_data;
entry             544 sound/pci/cs46xx/dsp_spos.c static void cs46xx_dsp_proc_task_tree_read (struct snd_info_entry *entry,
entry             547 sound/pci/cs46xx/dsp_spos.c 	struct snd_cs46xx *chip = entry->private_data;
entry             572 sound/pci/cs46xx/dsp_spos.c static void cs46xx_dsp_proc_scb_read (struct snd_info_entry *entry,
entry             575 sound/pci/cs46xx/dsp_spos.c 	struct snd_cs46xx *chip = entry->private_data;
entry             605 sound/pci/cs46xx/dsp_spos.c static void cs46xx_dsp_proc_parameter_dump_read (struct snd_info_entry *entry,
entry             608 sound/pci/cs46xx/dsp_spos.c 	struct snd_cs46xx *chip = entry->private_data;
entry             633 sound/pci/cs46xx/dsp_spos.c static void cs46xx_dsp_proc_sample_dump_read (struct snd_info_entry *entry,
entry             636 sound/pci/cs46xx/dsp_spos.c 	struct snd_cs46xx *chip = entry->private_data;
entry             783 sound/pci/cs46xx/dsp_spos.c 	struct snd_info_entry *entry;
entry             789 sound/pci/cs46xx/dsp_spos.c 	entry = snd_info_create_card_entry(card, "dsp", card->proc_root);
entry             790 sound/pci/cs46xx/dsp_spos.c 	if (entry)
entry             791 sound/pci/cs46xx/dsp_spos.c 		entry->mode = S_IFDIR | 0555;
entry             792 sound/pci/cs46xx/dsp_spos.c 	ins->proc_dsp_dir = entry;
entry             797 sound/pci/cs46xx/dsp_spos.c 	entry = snd_info_create_card_entry(card, "spos_symbols",
entry             799 sound/pci/cs46xx/dsp_spos.c 	if (entry)
entry             800 sound/pci/cs46xx/dsp_spos.c 		snd_info_set_text_ops(entry, chip,
entry             803 sound/pci/cs46xx/dsp_spos.c 	entry = snd_info_create_card_entry(card, "spos_modules",
entry             805 sound/pci/cs46xx/dsp_spos.c 	if (entry)
entry             806 sound/pci/cs46xx/dsp_spos.c 		snd_info_set_text_ops(entry, chip,
entry             809 sound/pci/cs46xx/dsp_spos.c 	entry = snd_info_create_card_entry(card, "parameter",
entry             811 sound/pci/cs46xx/dsp_spos.c 	if (entry)
entry             812 sound/pci/cs46xx/dsp_spos.c 		snd_info_set_text_ops(entry, chip,
entry             815 sound/pci/cs46xx/dsp_spos.c 	entry = snd_info_create_card_entry(card, "sample",
entry             817 sound/pci/cs46xx/dsp_spos.c 	if (entry)
entry             818 sound/pci/cs46xx/dsp_spos.c 		snd_info_set_text_ops(entry, chip,
entry             821 sound/pci/cs46xx/dsp_spos.c 	entry = snd_info_create_card_entry(card, "task_tree",
entry             823 sound/pci/cs46xx/dsp_spos.c 	if (entry)
entry             824 sound/pci/cs46xx/dsp_spos.c 		snd_info_set_text_ops(entry, chip,
entry             827 sound/pci/cs46xx/dsp_spos.c 	entry = snd_info_create_card_entry(card, "scb_info",
entry             829 sound/pci/cs46xx/dsp_spos.c 	if (entry)
entry             830 sound/pci/cs46xx/dsp_spos.c 		snd_info_set_text_ops(entry, chip,
entry              57 sound/pci/cs46xx/dsp_spos_scb_lib.c static void cs46xx_dsp_proc_scb_info_read (struct snd_info_entry *entry,
entry              60 sound/pci/cs46xx/dsp_spos_scb_lib.c 	struct proc_scb_info * scb_info  = entry->private_data;
entry             236 sound/pci/cs46xx/dsp_spos_scb_lib.c 	struct snd_info_entry * entry;
entry             243 sound/pci/cs46xx/dsp_spos_scb_lib.c 		entry = snd_info_create_card_entry(ins->snd_card, scb->scb_name,
entry             245 sound/pci/cs46xx/dsp_spos_scb_lib.c 		if (entry) {
entry             248 sound/pci/cs46xx/dsp_spos_scb_lib.c 				snd_info_free_entry(entry);
entry             249 sound/pci/cs46xx/dsp_spos_scb_lib.c 				entry = NULL;
entry             255 sound/pci/cs46xx/dsp_spos_scb_lib.c 			snd_info_set_text_ops(entry, scb_info,
entry             259 sound/pci/cs46xx/dsp_spos_scb_lib.c 		scb->proc_info = entry;
entry             158 sound/pci/ctxfi/ctdaio.c 	struct imapper *entry;
entry             162 sound/pci/ctxfi/ctdaio.c 	entry = kzalloc((sizeof(*entry) * daio->rscl.msr), GFP_KERNEL);
entry             163 sound/pci/ctxfi/ctdaio.c 	if (!entry)
entry             170 sound/pci/ctxfi/ctdaio.c 	for (i = 0; i < daio->rscl.msr; i++, entry++) {
entry             171 sound/pci/ctxfi/ctdaio.c 		entry->slot = input->ops->output_slot(input);
entry             172 sound/pci/ctxfi/ctdaio.c 		entry->user = entry->addr = daio->rscl.ops->index(&daio->rscl);
entry             173 sound/pci/ctxfi/ctdaio.c 		dao->mgr->imap_add(dao->mgr, entry);
entry             174 sound/pci/ctxfi/ctdaio.c 		dao->imappers[i] = entry;
entry             187 sound/pci/ctxfi/ctdaio.c 	struct imapper *entry;
entry             191 sound/pci/ctxfi/ctdaio.c 	entry = kzalloc((sizeof(*entry) * daio->rscr.msr), GFP_KERNEL);
entry             192 sound/pci/ctxfi/ctdaio.c 	if (!entry)
entry             199 sound/pci/ctxfi/ctdaio.c 	for (i = 0; i < daio->rscr.msr; i++, entry++) {
entry             200 sound/pci/ctxfi/ctdaio.c 		entry->slot = input->ops->output_slot(input);
entry             201 sound/pci/ctxfi/ctdaio.c 		entry->user = entry->addr = daio->rscr.ops->index(&daio->rscr);
entry             202 sound/pci/ctxfi/ctdaio.c 		dao->mgr->imap_add(dao->mgr, entry);
entry             203 sound/pci/ctxfi/ctdaio.c 		dao->imappers[daio->rscl.msr + i] = entry;
entry             216 sound/pci/ctxfi/ctdaio.c 	struct imapper *entry;
entry             223 sound/pci/ctxfi/ctdaio.c 	entry = dao->imappers[0];
entry             224 sound/pci/ctxfi/ctdaio.c 	dao->mgr->imap_delete(dao->mgr, entry);
entry             227 sound/pci/ctxfi/ctdaio.c 		entry = dao->imappers[i];
entry             228 sound/pci/ctxfi/ctdaio.c 		dao->mgr->imap_delete(dao->mgr, entry);
entry             240 sound/pci/ctxfi/ctdaio.c 	struct imapper *entry;
entry             247 sound/pci/ctxfi/ctdaio.c 	entry = dao->imappers[daio->rscl.msr];
entry             248 sound/pci/ctxfi/ctdaio.c 	dao->mgr->imap_delete(dao->mgr, entry);
entry             251 sound/pci/ctxfi/ctdaio.c 		entry = dao->imappers[daio->rscl.msr + i];
entry             252 sound/pci/ctxfi/ctdaio.c 		dao->mgr->imap_delete(dao->mgr, entry);
entry             632 sound/pci/ctxfi/ctdaio.c static int daio_map_op(void *data, struct imapper *entry)
entry             637 sound/pci/ctxfi/ctdaio.c 	hw->daio_mgr_set_imaparc(mgr->ctrl_blk, entry->slot);
entry             638 sound/pci/ctxfi/ctdaio.c 	hw->daio_mgr_set_imapnxt(mgr->ctrl_blk, entry->next);
entry             639 sound/pci/ctxfi/ctdaio.c 	hw->daio_mgr_set_imapaddr(mgr->ctrl_blk, entry->addr);
entry             645 sound/pci/ctxfi/ctdaio.c static int daio_imap_add(struct daio_mgr *mgr, struct imapper *entry)
entry             651 sound/pci/ctxfi/ctdaio.c 	if (!entry->addr && mgr->init_imap_added) {
entry             656 sound/pci/ctxfi/ctdaio.c 	err = input_mapper_add(&mgr->imappers, entry, daio_map_op, mgr);
entry             662 sound/pci/ctxfi/ctdaio.c static int daio_imap_delete(struct daio_mgr *mgr, struct imapper *entry)
entry             668 sound/pci/ctxfi/ctdaio.c 	err = input_mapper_delete(&mgr->imappers, entry, daio_map_op, mgr);
entry             691 sound/pci/ctxfi/ctdaio.c 	struct imapper *entry;
entry             705 sound/pci/ctxfi/ctdaio.c 	entry = kzalloc(sizeof(*entry), GFP_KERNEL);
entry             706 sound/pci/ctxfi/ctdaio.c 	if (!entry) {
entry             710 sound/pci/ctxfi/ctdaio.c 	entry->slot = entry->addr = entry->next = entry->user = 0;
entry             711 sound/pci/ctxfi/ctdaio.c 	list_add(&entry->list, &daio_mgr->imappers);
entry             712 sound/pci/ctxfi/ctdaio.c 	daio_mgr->init_imap = entry;
entry             112 sound/pci/ctxfi/ctdaio.h 	int (*imap_add)(struct daio_mgr *mgr, struct imapper *entry);
entry             113 sound/pci/ctxfi/ctdaio.h 	int (*imap_delete)(struct daio_mgr *mgr, struct imapper *entry);
entry              18 sound/pci/ctxfi/ctimap.c int input_mapper_add(struct list_head *mappers, struct imapper *entry,
entry              27 sound/pci/ctxfi/ctimap.c 		entry->next = entry->addr;
entry              28 sound/pci/ctxfi/ctimap.c 		map_op(data, entry);
entry              29 sound/pci/ctxfi/ctimap.c 		list_add(&entry->list, head);
entry              35 sound/pci/ctxfi/ctimap.c 		if (pos_ent->slot > entry->slot) {
entry              46 sound/pci/ctxfi/ctimap.c 		__list_add(&entry->list, pos->prev, pos);
entry              50 sound/pci/ctxfi/ctimap.c 		list_add_tail(&entry->list, head);
entry              56 sound/pci/ctxfi/ctimap.c 	entry->next = pos_ent->addr;
entry              57 sound/pci/ctxfi/ctimap.c 	map_op(data, entry);
entry              58 sound/pci/ctxfi/ctimap.c 	pre_ent->next = entry->addr;
entry              64 sound/pci/ctxfi/ctimap.c int input_mapper_delete(struct list_head *mappers, struct imapper *entry,
entry              75 sound/pci/ctxfi/ctimap.c 	pre = (entry->list.prev == head) ? head->prev : entry->list.prev;
entry              76 sound/pci/ctxfi/ctimap.c 	next = (entry->list.next == head) ? head->next : entry->list.next;
entry              78 sound/pci/ctxfi/ctimap.c 	if (pre == &entry->list) {
entry              80 sound/pci/ctxfi/ctimap.c 		entry->next = entry->addr = entry->user = entry->slot = 0;
entry              81 sound/pci/ctxfi/ctimap.c 		map_op(data, entry);
entry              82 sound/pci/ctxfi/ctimap.c 		list_del(&entry->list);
entry              91 sound/pci/ctxfi/ctimap.c 	list_del(&entry->list);
entry              98 sound/pci/ctxfi/ctimap.c 	struct imapper *entry;
entry             104 sound/pci/ctxfi/ctimap.c 		entry = list_entry(pos, struct imapper, list);
entry             105 sound/pci/ctxfi/ctimap.c 		kfree(entry);
entry              28 sound/pci/ctxfi/ctimap.h int input_mapper_add(struct list_head *mappers, struct imapper *entry,
entry              31 sound/pci/ctxfi/ctimap.h int input_mapper_delete(struct list_head *mappers, struct imapper *entry,
entry             619 sound/pci/ctxfi/ctsrc.c 	struct imapper *entry;
entry             628 sound/pci/ctxfi/ctsrc.c 		entry = &srcimp->imappers[i];
entry             629 sound/pci/ctxfi/ctsrc.c 		entry->slot = input->ops->output_slot(input);
entry             630 sound/pci/ctxfi/ctsrc.c 		entry->user = src->rsc.ops->index(&src->rsc);
entry             631 sound/pci/ctxfi/ctsrc.c 		entry->addr = srcimp->rsc.ops->index(&srcimp->rsc);
entry             632 sound/pci/ctxfi/ctsrc.c 		srcimp->mgr->imap_add(srcimp->mgr, entry);
entry             777 sound/pci/ctxfi/ctsrc.c static int srcimp_map_op(void *data, struct imapper *entry)
entry             782 sound/pci/ctxfi/ctsrc.c 	hw->srcimp_mgr_set_imaparc(mgr->ctrl_blk, entry->slot);
entry             783 sound/pci/ctxfi/ctsrc.c 	hw->srcimp_mgr_set_imapuser(mgr->ctrl_blk, entry->user);
entry             784 sound/pci/ctxfi/ctsrc.c 	hw->srcimp_mgr_set_imapnxt(mgr->ctrl_blk, entry->next);
entry             785 sound/pci/ctxfi/ctsrc.c 	hw->srcimp_mgr_set_imapaddr(mgr->ctrl_blk, entry->addr);
entry             791 sound/pci/ctxfi/ctsrc.c static int srcimp_imap_add(struct srcimp_mgr *mgr, struct imapper *entry)
entry             797 sound/pci/ctxfi/ctsrc.c 	if ((0 == entry->addr) && (mgr->init_imap_added)) {
entry             802 sound/pci/ctxfi/ctsrc.c 	err = input_mapper_add(&mgr->imappers, entry, srcimp_map_op, mgr);
entry             808 sound/pci/ctxfi/ctsrc.c static int srcimp_imap_delete(struct srcimp_mgr *mgr, struct imapper *entry)
entry             814 sound/pci/ctxfi/ctsrc.c 	err = input_mapper_delete(&mgr->imappers, entry, srcimp_map_op, mgr);
entry             829 sound/pci/ctxfi/ctsrc.c 	struct imapper *entry;
entry             843 sound/pci/ctxfi/ctsrc.c 	entry = kzalloc(sizeof(*entry), GFP_KERNEL);
entry             844 sound/pci/ctxfi/ctsrc.c 	if (!entry) {
entry             848 sound/pci/ctxfi/ctsrc.c 	entry->slot = entry->addr = entry->next = entry->user = 0;
entry             849 sound/pci/ctxfi/ctsrc.c 	list_add(&entry->list, &srcimp_mgr->imappers);
entry             850 sound/pci/ctxfi/ctsrc.c 	srcimp_mgr->init_imap = entry;
entry             137 sound/pci/ctxfi/ctsrc.h 	int (*imap_add)(struct srcimp_mgr *mgr, struct imapper *entry);
entry             138 sound/pci/ctxfi/ctsrc.h 	int (*imap_delete)(struct srcimp_mgr *mgr, struct imapper *entry);
entry              32 sound/pci/ctxfi/ctvmem.c 	struct ct_vm_block *block = NULL, *entry;
entry              44 sound/pci/ctxfi/ctvmem.c 		entry = list_entry(pos, struct ct_vm_block, list);
entry              45 sound/pci/ctxfi/ctvmem.c 		if (entry->size >= size)
entry              51 sound/pci/ctxfi/ctvmem.c 	if (entry->size == size) {
entry              53 sound/pci/ctxfi/ctvmem.c 		list_move(&entry->list, &vm->used);
entry              55 sound/pci/ctxfi/ctvmem.c 		block = entry;
entry              63 sound/pci/ctxfi/ctvmem.c 	block->addr = entry->addr;
entry              66 sound/pci/ctxfi/ctvmem.c 	entry->addr += size;
entry              67 sound/pci/ctxfi/ctvmem.c 	entry->size -= size;
entry              77 sound/pci/ctxfi/ctvmem.c 	struct ct_vm_block *entry, *pre_ent;
entry              87 sound/pci/ctxfi/ctvmem.c 		entry = list_entry(pos, struct ct_vm_block, list);
entry              88 sound/pci/ctxfi/ctvmem.c 		if (entry->addr >= (block->addr + block->size))
entry              93 sound/pci/ctxfi/ctvmem.c 		entry = block;
entry              95 sound/pci/ctxfi/ctvmem.c 		if ((block->addr + block->size) == entry->addr) {
entry              96 sound/pci/ctxfi/ctvmem.c 			entry->addr = block->addr;
entry              97 sound/pci/ctxfi/ctvmem.c 			entry->size += block->size;
entry             101 sound/pci/ctxfi/ctvmem.c 			entry = block;
entry             105 sound/pci/ctxfi/ctvmem.c 	pos = &entry->list;
entry             108 sound/pci/ctxfi/ctvmem.c 		entry = list_entry(pos, struct ct_vm_block, list);
entry             110 sound/pci/ctxfi/ctvmem.c 		if ((pre_ent->addr + pre_ent->size) > entry->addr)
entry             113 sound/pci/ctxfi/ctvmem.c 		pre_ent->size += entry->size;
entry             115 sound/pci/ctxfi/ctvmem.c 		kfree(entry);
entry             219 sound/pci/ctxfi/ctvmem.c 	struct ct_vm_block *entry;
entry             225 sound/pci/ctxfi/ctvmem.c 		entry = list_entry(pos, struct ct_vm_block, list);
entry             226 sound/pci/ctxfi/ctvmem.c 		kfree(entry);
entry             231 sound/pci/ctxfi/ctvmem.c 		entry = list_entry(pos, struct ct_vm_block, list);
entry             232 sound/pci/ctxfi/ctvmem.c 		kfree(entry);
entry            1008 sound/pci/emu10k1/emu10k1x.c static void snd_emu10k1x_proc_reg_read(struct snd_info_entry *entry, 
entry            1011 sound/pci/emu10k1/emu10k1x.c 	struct emu10k1x *emu = entry->private_data;
entry            1036 sound/pci/emu10k1/emu10k1x.c static void snd_emu10k1x_proc_reg_write(struct snd_info_entry *entry, 
entry            1039 sound/pci/emu10k1/emu10k1x.c 	struct emu10k1x *emu = entry->private_data;
entry              66 sound/pci/emu10k1/emuproc.c static void snd_emu10k1_proc_read(struct snd_info_entry *entry, 
entry             172 sound/pci/emu10k1/emuproc.c 	struct snd_emu10k1 *emu = entry->private_data;
entry             223 sound/pci/emu10k1/emuproc.c static void snd_emu10k1_proc_spdif_read(struct snd_info_entry *entry, 
entry             226 sound/pci/emu10k1/emuproc.c 	struct snd_emu10k1 *emu = entry->private_data;
entry             262 sound/pci/emu10k1/emuproc.c static void snd_emu10k1_proc_rates_read(struct snd_info_entry *entry, 
entry             266 sound/pci/emu10k1/emuproc.c 	struct snd_emu10k1 *emu = entry->private_data;
entry             276 sound/pci/emu10k1/emuproc.c static void snd_emu10k1_proc_acode_read(struct snd_info_entry *entry, 
entry             280 sound/pci/emu10k1/emuproc.c 	struct snd_emu10k1 *emu = entry->private_data;
entry             319 sound/pci/emu10k1/emuproc.c static ssize_t snd_emu10k1_fx8010_read(struct snd_info_entry *entry,
entry             324 sound/pci/emu10k1/emuproc.c 	struct snd_emu10k1 *emu = entry->private_data;
entry             331 sound/pci/emu10k1/emuproc.c 	if (!strcmp(entry->name, "fx8010_tram_addr")) {
entry             334 sound/pci/emu10k1/emuproc.c 	} else if (!strcmp(entry->name, "fx8010_tram_data")) {
entry             336 sound/pci/emu10k1/emuproc.c 	} else if (!strcmp(entry->name, "fx8010_code")) {
entry             362 sound/pci/emu10k1/emuproc.c static void snd_emu10k1_proc_voices_read(struct snd_info_entry *entry, 
entry             365 sound/pci/emu10k1/emuproc.c 	struct snd_emu10k1 *emu = entry->private_data;
entry             383 sound/pci/emu10k1/emuproc.c static void snd_emu_proc_emu1010_reg_read(struct snd_info_entry *entry,
entry             386 sound/pci/emu10k1/emuproc.c 	struct snd_emu10k1 *emu = entry->private_data;
entry             397 sound/pci/emu10k1/emuproc.c static void snd_emu_proc_io_reg_read(struct snd_info_entry *entry,
entry             400 sound/pci/emu10k1/emuproc.c 	struct snd_emu10k1 *emu = entry->private_data;
entry             413 sound/pci/emu10k1/emuproc.c static void snd_emu_proc_io_reg_write(struct snd_info_entry *entry,
entry             416 sound/pci/emu10k1/emuproc.c 	struct snd_emu10k1 *emu = entry->private_data;
entry             466 sound/pci/emu10k1/emuproc.c static void snd_emu_proc_ptr_reg_read(struct snd_info_entry *entry,
entry             469 sound/pci/emu10k1/emuproc.c 	struct snd_emu10k1 *emu = entry->private_data;
entry             490 sound/pci/emu10k1/emuproc.c static void snd_emu_proc_ptr_reg_write(struct snd_info_entry *entry,
entry             493 sound/pci/emu10k1/emuproc.c 	struct snd_emu10k1 *emu = entry->private_data;
entry             504 sound/pci/emu10k1/emuproc.c static void snd_emu_proc_ptr_reg_write00(struct snd_info_entry *entry,
entry             507 sound/pci/emu10k1/emuproc.c 	snd_emu_proc_ptr_reg_write(entry, buffer, 0);
entry             510 sound/pci/emu10k1/emuproc.c static void snd_emu_proc_ptr_reg_write20(struct snd_info_entry *entry,
entry             513 sound/pci/emu10k1/emuproc.c 	snd_emu_proc_ptr_reg_write(entry, buffer, 0x20);
entry             517 sound/pci/emu10k1/emuproc.c static void snd_emu_proc_ptr_reg_read00a(struct snd_info_entry *entry,
entry             520 sound/pci/emu10k1/emuproc.c 	snd_emu_proc_ptr_reg_read(entry, buffer, 0, 0, 0x40, 64);
entry             523 sound/pci/emu10k1/emuproc.c static void snd_emu_proc_ptr_reg_read00b(struct snd_info_entry *entry,
entry             526 sound/pci/emu10k1/emuproc.c 	snd_emu_proc_ptr_reg_read(entry, buffer, 0, 0x40, 0x40, 64);
entry             529 sound/pci/emu10k1/emuproc.c static void snd_emu_proc_ptr_reg_read20a(struct snd_info_entry *entry,
entry             532 sound/pci/emu10k1/emuproc.c 	snd_emu_proc_ptr_reg_read(entry, buffer, 0x20, 0, 0x40, 4);
entry             535 sound/pci/emu10k1/emuproc.c static void snd_emu_proc_ptr_reg_read20b(struct snd_info_entry *entry,
entry             538 sound/pci/emu10k1/emuproc.c 	snd_emu_proc_ptr_reg_read(entry, buffer, 0x20, 0x40, 0x40, 4);
entry             541 sound/pci/emu10k1/emuproc.c static void snd_emu_proc_ptr_reg_read20c(struct snd_info_entry *entry,
entry             544 sound/pci/emu10k1/emuproc.c 	snd_emu_proc_ptr_reg_read(entry, buffer, 0x20, 0x80, 0x20, 4);
entry             554 sound/pci/emu10k1/emuproc.c 	struct snd_info_entry *entry;
entry             592 sound/pci/emu10k1/emuproc.c 	if (! snd_card_proc_new(emu->card, "fx8010_gpr", &entry)) {
entry             593 sound/pci/emu10k1/emuproc.c 		entry->content = SNDRV_INFO_CONTENT_DATA;
entry             594 sound/pci/emu10k1/emuproc.c 		entry->private_data = emu;
entry             595 sound/pci/emu10k1/emuproc.c 		entry->mode = S_IFREG | 0444 /*| S_IWUSR*/;
entry             596 sound/pci/emu10k1/emuproc.c 		entry->size = emu->audigy ? A_TOTAL_SIZE_GPR : TOTAL_SIZE_GPR;
entry             597 sound/pci/emu10k1/emuproc.c 		entry->c.ops = &snd_emu10k1_proc_ops_fx8010;
entry             599 sound/pci/emu10k1/emuproc.c 	if (! snd_card_proc_new(emu->card, "fx8010_tram_data", &entry)) {
entry             600 sound/pci/emu10k1/emuproc.c 		entry->content = SNDRV_INFO_CONTENT_DATA;
entry             601 sound/pci/emu10k1/emuproc.c 		entry->private_data = emu;
entry             602 sound/pci/emu10k1/emuproc.c 		entry->mode = S_IFREG | 0444 /*| S_IWUSR*/;
entry             603 sound/pci/emu10k1/emuproc.c 		entry->size = emu->audigy ? A_TOTAL_SIZE_TANKMEM_DATA : TOTAL_SIZE_TANKMEM_DATA ;
entry             604 sound/pci/emu10k1/emuproc.c 		entry->c.ops = &snd_emu10k1_proc_ops_fx8010;
entry             606 sound/pci/emu10k1/emuproc.c 	if (! snd_card_proc_new(emu->card, "fx8010_tram_addr", &entry)) {
entry             607 sound/pci/emu10k1/emuproc.c 		entry->content = SNDRV_INFO_CONTENT_DATA;
entry             608 sound/pci/emu10k1/emuproc.c 		entry->private_data = emu;
entry             609 sound/pci/emu10k1/emuproc.c 		entry->mode = S_IFREG | 0444 /*| S_IWUSR*/;
entry             610 sound/pci/emu10k1/emuproc.c 		entry->size = emu->audigy ? A_TOTAL_SIZE_TANKMEM_ADDR : TOTAL_SIZE_TANKMEM_ADDR ;
entry             611 sound/pci/emu10k1/emuproc.c 		entry->c.ops = &snd_emu10k1_proc_ops_fx8010;
entry             613 sound/pci/emu10k1/emuproc.c 	if (! snd_card_proc_new(emu->card, "fx8010_code", &entry)) {
entry             614 sound/pci/emu10k1/emuproc.c 		entry->content = SNDRV_INFO_CONTENT_DATA;
entry             615 sound/pci/emu10k1/emuproc.c 		entry->private_data = emu;
entry             616 sound/pci/emu10k1/emuproc.c 		entry->mode = S_IFREG | 0444 /*| S_IWUSR*/;
entry             617 sound/pci/emu10k1/emuproc.c 		entry->size = emu->audigy ? A_TOTAL_SIZE_CODE : TOTAL_SIZE_CODE;
entry             618 sound/pci/emu10k1/emuproc.c 		entry->c.ops = &snd_emu10k1_proc_ops_fx8010;
entry            1870 sound/pci/ens1370.c static void snd_ensoniq_proc_read(struct snd_info_entry *entry, 
entry            1873 sound/pci/ens1370.c 	struct ensoniq *ensoniq = entry->private_data;
entry             738 sound/pci/hda/hda_proc.c static void print_codec_info(struct snd_info_entry *entry,
entry             741 sound/pci/hda/hda_proc.c 	struct hda_codec *codec = entry->private_data;
entry             483 sound/pci/hda/patch_hdmi.c static void print_eld_info(struct snd_info_entry *entry,
entry             486 sound/pci/hda/patch_hdmi.c 	struct hdmi_spec_per_pin *per_pin = entry->private_data;
entry             493 sound/pci/hda/patch_hdmi.c static void write_eld_info(struct snd_info_entry *entry,
entry             496 sound/pci/hda/patch_hdmi.c 	struct hdmi_spec_per_pin *per_pin = entry->private_data;
entry             507 sound/pci/hda/patch_hdmi.c 	struct snd_info_entry *entry;
entry             511 sound/pci/hda/patch_hdmi.c 	err = snd_card_proc_new(codec->card, name, &entry);
entry             515 sound/pci/hda/patch_hdmi.c 	snd_info_set_text_ops(entry, per_pin, print_eld_info);
entry             516 sound/pci/hda/patch_hdmi.c 	entry->c.text.write = write_eld_info;
entry             517 sound/pci/hda/patch_hdmi.c 	entry->mode |= 0200;
entry             518 sound/pci/hda/patch_hdmi.c 	per_pin->proc_entry = entry;
entry            1550 sound/pci/ice1712/ice1712.c static void snd_ice1712_proc_read(struct snd_info_entry *entry,
entry            1553 sound/pci/ice1712/ice1712.c 	struct snd_ice1712 *ice = entry->private_data;
entry            1519 sound/pci/ice1712/ice1724.c static void snd_vt1724_proc_read(struct snd_info_entry *entry,
entry            1522 sound/pci/ice1712/ice1724.c 	struct snd_ice1712 *ice = entry->private_data;
entry             618 sound/pci/ice1712/pontis.c static void wm_proc_regs_write(struct snd_info_entry *entry, struct snd_info_buffer *buffer)
entry             620 sound/pci/ice1712/pontis.c 	struct snd_ice1712 *ice = entry->private_data;
entry             633 sound/pci/ice1712/pontis.c static void wm_proc_regs_read(struct snd_info_entry *entry, struct snd_info_buffer *buffer)
entry             635 sound/pci/ice1712/pontis.c 	struct snd_ice1712 *ice = entry->private_data;
entry             652 sound/pci/ice1712/pontis.c static void cs_proc_regs_read(struct snd_info_entry *entry, struct snd_info_buffer *buffer)
entry             654 sound/pci/ice1712/pontis.c 	struct snd_ice1712 *ice = entry->private_data;
entry             625 sound/pci/ice1712/prodigy192.c static void stac9460_proc_regs_read(struct snd_info_entry *entry,
entry             628 sound/pci/ice1712/prodigy192.c 	struct snd_ice1712 *ice = entry->private_data;
entry             861 sound/pci/ice1712/prodigy_hifi.c static void wm_proc_regs_write(struct snd_info_entry *entry,
entry             864 sound/pci/ice1712/prodigy_hifi.c 	struct snd_ice1712 *ice = entry->private_data;
entry             877 sound/pci/ice1712/prodigy_hifi.c static void wm_proc_regs_read(struct snd_info_entry *entry,
entry             880 sound/pci/ice1712/prodigy_hifi.c 	struct snd_ice1712 *ice = entry->private_data;
entry             474 sound/pci/ice1712/quartet.c static void proc_regs_read(struct snd_info_entry *entry,
entry             477 sound/pci/ice1712/quartet.c 	struct snd_ice1712 *ice = entry->private_data;
entry            2815 sound/pci/intel8x0.c static void snd_intel8x0_proc_read(struct snd_info_entry * entry,
entry            2818 sound/pci/intel8x0.c 	struct intel8x0 *chip = entry->private_data;
entry            1050 sound/pci/intel8x0m.c static void snd_intel8x0m_proc_read(struct snd_info_entry * entry,
entry            1053 sound/pci/intel8x0m.c 	struct intel8x0m *chip = entry->private_data;
entry            2049 sound/pci/korg1212/korg1212.c static void snd_korg1212_proc_read(struct snd_info_entry *entry,
entry            2053 sound/pci/korg1212/korg1212.c 	struct snd_korg1212 *korg1212 = entry->private_data;
entry              97 sound/pci/lola/lola_proc.c static void lola_proc_codec_read(struct snd_info_entry *entry,
entry             100 sound/pci/lola/lola_proc.c 	struct lola *chip = entry->private_data;
entry             134 sound/pci/lola/lola_proc.c static void lola_proc_codec_rw_write(struct snd_info_entry *entry,
entry             137 sound/pci/lola/lola_proc.c 	struct lola *chip = entry->private_data;
entry             149 sound/pci/lola/lola_proc.c static void lola_proc_codec_rw_read(struct snd_info_entry *entry,
entry             152 sound/pci/lola/lola_proc.c 	struct lola *chip = entry->private_data;
entry             159 sound/pci/lola/lola_proc.c static void lola_proc_regs_read(struct snd_info_entry *entry,
entry             162 sound/pci/lola/lola_proc.c 	struct lola *chip = entry->private_data;
entry             904 sound/pci/lx6464es/lx6464es.c static void lx_proc_levels_read(struct snd_info_entry *entry,
entry             910 sound/pci/lx6464es/lx6464es.c 	struct lx6464es *chip = entry->private_data;
entry            1131 sound/pci/mixart/mixart.c static ssize_t snd_mixart_BA0_read(struct snd_info_entry *entry,
entry            1136 sound/pci/mixart/mixart.c 	struct mixart_mgr *mgr = entry->private_data;
entry            1147 sound/pci/mixart/mixart.c static ssize_t snd_mixart_BA1_read(struct snd_info_entry *entry,
entry            1152 sound/pci/mixart/mixart.c 	struct mixart_mgr *mgr = entry->private_data;
entry            1169 sound/pci/mixart/mixart.c static void snd_mixart_proc_read(struct snd_info_entry *entry, 
entry            1172 sound/pci/mixart/mixart.c 	struct snd_mixart *chip = entry->private_data;        
entry            1207 sound/pci/mixart/mixart.c 	struct snd_info_entry *entry;
entry            1213 sound/pci/mixart/mixart.c 	if (! snd_card_proc_new(chip->card, "mixart_BA0", &entry)) {
entry            1214 sound/pci/mixart/mixart.c 		entry->content = SNDRV_INFO_CONTENT_DATA;
entry            1215 sound/pci/mixart/mixart.c 		entry->private_data = chip->mgr;	
entry            1216 sound/pci/mixart/mixart.c 		entry->c.ops = &snd_mixart_proc_ops_BA0;
entry            1217 sound/pci/mixart/mixart.c 		entry->size = MIXART_BA0_SIZE;
entry            1219 sound/pci/mixart/mixart.c 	if (! snd_card_proc_new(chip->card, "mixart_BA1", &entry)) {
entry            1220 sound/pci/mixart/mixart.c 		entry->content = SNDRV_INFO_CONTENT_DATA;
entry            1221 sound/pci/mixart/mixart.c 		entry->private_data = chip->mgr;
entry            1222 sound/pci/mixart/mixart.c 		entry->c.ops = &snd_mixart_proc_ops_BA1;
entry            1223 sound/pci/mixart/mixart.c 		entry->size = MIXART_BA1_SIZE;
entry             187 sound/pci/oxygen/oxygen_lib.c static void oxygen_proc_read(struct snd_info_entry *entry,
entry             190 sound/pci/oxygen/oxygen_lib.c 	struct oxygen *chip = entry->private_data;
entry            1234 sound/pci/pcxhr/pcxhr.c static void pcxhr_proc_info(struct snd_info_entry *entry,
entry            1237 sound/pci/pcxhr/pcxhr.c 	struct snd_pcxhr *chip = entry->private_data;
entry            1305 sound/pci/pcxhr/pcxhr.c static void pcxhr_proc_sync(struct snd_info_entry *entry,
entry            1308 sound/pci/pcxhr/pcxhr.c 	struct snd_pcxhr *chip = entry->private_data;
entry            1347 sound/pci/pcxhr/pcxhr.c static void pcxhr_proc_gpio_read(struct snd_info_entry *entry,
entry            1350 sound/pci/pcxhr/pcxhr.c 	struct snd_pcxhr *chip = entry->private_data;
entry            1364 sound/pci/pcxhr/pcxhr.c static void pcxhr_proc_gpo_write(struct snd_info_entry *entry,
entry            1367 sound/pci/pcxhr/pcxhr.c 	struct snd_pcxhr *chip = entry->private_data;
entry            1392 sound/pci/pcxhr/pcxhr.c static void pcxhr_proc_ltc(struct snd_info_entry *entry,
entry            1395 sound/pci/pcxhr/pcxhr.c 	struct snd_pcxhr *chip = entry->private_data;
entry            1890 sound/pci/riptide/riptide.c snd_riptide_proc_read(struct snd_info_entry *entry,
entry            1893 sound/pci/riptide/riptide.c 	struct snd_riptide *chip = entry->private_data;
entry            1461 sound/pci/rme32.c snd_rme32_proc_read(struct snd_info_entry * entry, struct snd_info_buffer *buffer)
entry            1464 sound/pci/rme32.c 	struct rme32 *rme32 = (struct rme32 *) entry->private_data;
entry            1727 sound/pci/rme96.c snd_rme96_proc_read(struct snd_info_entry *entry, struct snd_info_buffer *buffer)
entry            1730 sound/pci/rme96.c 	struct rme96 *rme96 = entry->private_data;
entry            3320 sound/pci/rme9652/hdsp.c snd_hdsp_proc_read(struct snd_info_entry *entry, struct snd_info_buffer *buffer)
entry            3322 sound/pci/rme9652/hdsp.c 	struct hdsp *hdsp = entry->private_data;
entry            4748 sound/pci/rme9652/hdspm.c snd_hdspm_proc_read_tco(struct snd_info_entry *entry,
entry            4751 sound/pci/rme9652/hdspm.c 	struct hdspm *hdspm = entry->private_data;
entry            4860 sound/pci/rme9652/hdspm.c snd_hdspm_proc_read_madi(struct snd_info_entry *entry,
entry            4863 sound/pci/rme9652/hdspm.c 	struct hdspm *hdspm = entry->private_data;
entry            5016 sound/pci/rme9652/hdspm.c 	snd_hdspm_proc_read_tco(entry, buffer);
entry            5022 sound/pci/rme9652/hdspm.c snd_hdspm_proc_read_aes32(struct snd_info_entry * entry,
entry            5025 sound/pci/rme9652/hdspm.c 	struct hdspm *hdspm = entry->private_data;
entry            5169 sound/pci/rme9652/hdspm.c 	snd_hdspm_proc_read_tco(entry, buffer);
entry            5175 sound/pci/rme9652/hdspm.c snd_hdspm_proc_read_raydat(struct snd_info_entry *entry,
entry            5178 sound/pci/rme9652/hdspm.c 	struct hdspm *hdspm = entry->private_data;
entry            5233 sound/pci/rme9652/hdspm.c snd_hdspm_proc_read_debug(struct snd_info_entry *entry,
entry            5236 sound/pci/rme9652/hdspm.c 	struct hdspm *hdspm = entry->private_data;
entry            5250 sound/pci/rme9652/hdspm.c static void snd_hdspm_proc_ports_in(struct snd_info_entry *entry,
entry            5253 sound/pci/rme9652/hdspm.c 	struct hdspm *hdspm = entry->private_data;
entry            5263 sound/pci/rme9652/hdspm.c static void snd_hdspm_proc_ports_out(struct snd_info_entry *entry,
entry            5266 sound/pci/rme9652/hdspm.c 	struct hdspm *hdspm = entry->private_data;
entry            1552 sound/pci/rme9652/rme9652.c snd_rme9652_proc_read(struct snd_info_entry *entry, struct snd_info_buffer *buffer)
entry            1554 sound/pci/rme9652/rme9652.c 	struct snd_rme9652 *rme9652 = (struct snd_rme9652 *) entry->private_data;
entry            1128 sound/pci/sonicvibes.c static void snd_sonicvibes_proc_read(struct snd_info_entry *entry, 
entry            1131 sound/pci/sonicvibes.c 	struct sonicvibes *sonic = entry->private_data;
entry            3272 sound/pci/trident/trident_main.c static void snd_trident_proc_read(struct snd_info_entry *entry, 
entry            3275 sound/pci/trident/trident_main.c 	struct snd_trident *trident = entry->private_data;
entry            2119 sound/pci/via82xx.c static void snd_via82xx_proc_read(struct snd_info_entry *entry,
entry            2122 sound/pci/via82xx.c 	struct via82xx *chip = entry->private_data;
entry             911 sound/pci/via82xx_modem.c static void snd_via82xx_proc_read(struct snd_info_entry *entry, struct snd_info_buffer *buffer)
entry             913 sound/pci/via82xx_modem.c 	struct via82xx_modem *chip = entry->private_data;
entry            1961 sound/pci/ymfpci/ymfpci_main.c static void snd_ymfpci_proc_read(struct snd_info_entry *entry, 
entry            1964 sound/pci/ymfpci/ymfpci_main.c 	struct snd_ymfpci *chip = entry->private_data;
entry             124 sound/pcmcia/pdaudiocf/pdaudiocf_core.c static void pdacf_proc_read(struct snd_info_entry * entry,
entry             127 sound/pcmcia/pdaudiocf/pdaudiocf_core.c 	struct snd_pdacf *chip = entry->private_data;
entry             149 sound/soc/codecs/wm_hubs.c 				  struct wm_hubs_dcs_cache **entry)
entry             165 sound/soc/codecs/wm_hubs.c 		*entry = cache;
entry              67 sound/soc/intel/baytrail/sst-baytrail-dsp.c 	template.entry = module->entry_point;
entry             131 sound/soc/intel/common/sst-dsp-priv.h 	u32 entry;			/* entry point */
entry             195 sound/soc/intel/common/sst-dsp-priv.h 	u32 entry;			/* module entry point */
entry             495 sound/soc/intel/common/sst-firmware.c 	sst_module->entry = template->entry;
entry             109 sound/soc/intel/haswell/sst-haswell-dsp.c 	template.entry = entry_point - 4;
entry            1160 sound/soc/intel/haswell/sst-haswell-ipc.c 	map->module_entries[0].entry_point = module->entry;
entry            1927 sound/soc/intel/haswell/sst-haswell-ipc.c 	config.map.module_entries[0].entry_point = module->entry;
entry             350 sound/soc/sh/rcar/dma.c 	const u8 *entry = NULL;
entry             358 sound/soc/sh/rcar/dma.c 		entry = gen2_id_table_ssiu;
entry             362 sound/soc/sh/rcar/dma.c 		entry = gen2_id_table_scu;
entry             366 sound/soc/sh/rcar/dma.c 		entry = gen2_id_table_cmd;
entry             371 sound/soc/sh/rcar/dma.c 	if ((!entry) || (size <= id)) {
entry             380 sound/soc/sh/rcar/dma.c 	return entry[id];
entry            2414 sound/sparc/dbri.c #define CS4215_SINGLE(xname, entry, shift, mask, invert)	\
entry            2418 sound/sparc/dbri.c   .private_value = (entry) | ((shift) << 8) | ((mask) << 16) |	\
entry            2477 sound/sparc/dbri.c static void dbri_regs_read(struct snd_info_entry *entry,
entry            2480 sound/sparc/dbri.c 	struct snd_dbri *dbri = entry->private_data;
entry            2489 sound/sparc/dbri.c static void dbri_debug_read(struct snd_info_entry *entry,
entry            2492 sound/sparc/dbri.c 	struct snd_dbri *dbri = entry->private_data;
entry              15 sound/synth/emux/emux_proc.c snd_emux_proc_info_read(struct snd_info_entry *entry, 
entry              21 sound/synth/emux/emux_proc.c 	emu = entry->private_data;
entry              94 sound/synth/emux/emux_proc.c 	struct snd_info_entry *entry;
entry              98 sound/synth/emux/emux_proc.c 	entry = snd_info_create_card_entry(card, name, card->proc_root);
entry              99 sound/synth/emux/emux_proc.c 	if (entry == NULL)
entry             102 sound/synth/emux/emux_proc.c 	entry->content = SNDRV_INFO_CONTENT_TEXT;
entry             103 sound/synth/emux/emux_proc.c 	entry->private_data = emu;
entry             104 sound/synth/emux/emux_proc.c 	entry->c.text.read = snd_emux_proc_info_read;
entry            3239 sound/usb/mixer.c static void snd_usb_mixer_proc_read(struct snd_info_entry *entry,
entry            3242 sound/usb/mixer.c 	struct snd_usb_audio *chip = entry->private_data;
entry             398 sound/usb/mixer_quirks.c static void snd_audigy2nx_proc_read(struct snd_info_entry *entry,
entry             417 sound/usb/mixer_quirks.c 	struct usb_mixer_interface *mixer = entry->private_data;
entry              33 sound/usb/proc.c static void proc_audio_usbbus_read(struct snd_info_entry *entry, struct snd_info_buffer *buffer)
entry              35 sound/usb/proc.c 	struct snd_usb_audio *chip = entry->private_data;
entry              40 sound/usb/proc.c static void proc_audio_usbid_read(struct snd_info_entry *entry, struct snd_info_buffer *buffer)
entry              42 sound/usb/proc.c 	struct snd_usb_audio *chip = entry->private_data;
entry             137 sound/usb/proc.c static void proc_pcm_format_read(struct snd_info_entry *entry, struct snd_info_buffer *buffer)
entry             139 sound/usb/proc.c 	struct snd_usb_stream *stream = entry->private_data;
entry             558 tools/hv/hv_kvp_daemon.c 	struct dirent *entry;
entry             569 tools/hv/hv_kvp_daemon.c 	while ((entry = readdir(dir)) != NULL) {
entry             574 tools/hv/hv_kvp_daemon.c 			 KVP_NET_DIR, entry->d_name);
entry             591 tools/hv/hv_kvp_daemon.c 				if_name = strdup(entry->d_name);
entry             949 tools/hv/hv_kvp_daemon.c 	struct dirent *entry;
entry             962 tools/hv/hv_kvp_daemon.c 	while ((entry = readdir(dir)) != NULL) {
entry             967 tools/hv/hv_kvp_daemon.c 			 entry->d_name);
entry             992 tools/hv/hv_kvp_daemon.c 		if_name = entry->d_name;
entry              66 tools/hv/hv_vss_daemon.c 	struct dirent *entry;
entry              93 tools/hv/hv_vss_daemon.c 	while ((entry = readdir(dir)) != NULL) {
entry              94 tools/hv/hv_vss_daemon.c 		if (strcmp(entry->d_name, ".") == 0 ||
entry              95 tools/hv/hv_vss_daemon.c 		    strcmp(entry->d_name, "..") == 0)
entry              99 tools/hv/hv_vss_daemon.c 			 entry->d_name);
entry             100 tools/include/linux/list.h static inline void __list_del_entry(struct list_head *entry)
entry             102 tools/include/linux/list.h 	__list_del(entry->prev, entry->next);
entry             105 tools/include/linux/list.h static inline void list_del(struct list_head *entry)
entry             107 tools/include/linux/list.h 	__list_del(entry->prev, entry->next);
entry             108 tools/include/linux/list.h 	entry->next = LIST_POISON1;
entry             109 tools/include/linux/list.h 	entry->prev = LIST_POISON2;
entry             112 tools/include/linux/list.h extern void __list_del_entry(struct list_head *entry);
entry             113 tools/include/linux/list.h extern void list_del(struct list_head *entry);
entry             143 tools/include/linux/list.h static inline void list_del_init(struct list_head *entry)
entry             145 tools/include/linux/list.h 	__list_del_entry(entry);
entry             146 tools/include/linux/list.h 	INIT_LIST_HEAD(entry);
entry             235 tools/include/linux/list.h 		struct list_head *head, struct list_head *entry)
entry             237 tools/include/linux/list.h 	struct list_head *new_first = entry->next;
entry             240 tools/include/linux/list.h 	list->prev = entry;
entry             241 tools/include/linux/list.h 	entry->next = list;
entry             261 tools/include/linux/list.h 		struct list_head *head, struct list_head *entry)
entry             266 tools/include/linux/list.h 		(head->next != entry && head != entry))
entry             268 tools/include/linux/list.h 	if (entry == head)
entry             271 tools/include/linux/list.h 		__list_cut_position(list, head, entry);
entry            1590 tools/include/uapi/linux/kvm.h 	__u16 entry; /* The index of entry in the MSI-X table */
entry               9 tools/lib/api/cpu.c 	char entry[PATH_MAX];
entry              15 tools/lib/api/cpu.c 	snprintf(entry, sizeof(entry),
entry              18 tools/lib/api/cpu.c 	return sysfs__read_ull(entry, freq);
entry             424 tools/lib/api/fs/fs.c int procfs__read_str(const char *entry, char **buf, size_t *sizep)
entry             432 tools/lib/api/fs/fs.c 	snprintf(path, sizeof(path), "%s/%s", procfs, entry);
entry             437 tools/lib/api/fs/fs.c static int sysfs__read_ull_base(const char *entry,
entry             446 tools/lib/api/fs/fs.c 	snprintf(path, sizeof(path), "%s/%s", sysfs, entry);
entry             451 tools/lib/api/fs/fs.c int sysfs__read_xll(const char *entry, unsigned long long *value)
entry             453 tools/lib/api/fs/fs.c 	return sysfs__read_ull_base(entry, value, 16);
entry             456 tools/lib/api/fs/fs.c int sysfs__read_ull(const char *entry, unsigned long long *value)
entry             458 tools/lib/api/fs/fs.c 	return sysfs__read_ull_base(entry, value, 0);
entry             461 tools/lib/api/fs/fs.c int sysfs__read_int(const char *entry, int *value)
entry             469 tools/lib/api/fs/fs.c 	snprintf(path, sizeof(path), "%s/%s", sysfs, entry);
entry             474 tools/lib/api/fs/fs.c int sysfs__read_str(const char *entry, char **buf, size_t *sizep)
entry             482 tools/lib/api/fs/fs.c 	snprintf(path, sizeof(path), "%s/%s", sysfs, entry);
entry             487 tools/lib/api/fs/fs.c int sysfs__read_bool(const char *entry, bool *value)
entry             493 tools/lib/api/fs/fs.c 	ret = sysfs__read_str(entry, &buf, &size);
entry             529 tools/lib/api/fs/fs.c int sysfs__write_int(const char *entry, int value)
entry             537 tools/lib/api/fs/fs.c 	if (snprintf(path, sizeof(path), "%s/%s", sysfs, entry) >= PATH_MAX)
entry              38 tools/lib/api/fs/fs.h int procfs__read_str(const char *entry, char **buf, size_t *sizep);
entry              41 tools/lib/api/fs/fs.h int sysfs__read_int(const char *entry, int *value);
entry              42 tools/lib/api/fs/fs.h int sysfs__read_ull(const char *entry, unsigned long long *value);
entry              43 tools/lib/api/fs/fs.h int sysfs__read_xll(const char *entry, unsigned long long *value);
entry              44 tools/lib/api/fs/fs.h int sysfs__read_str(const char *entry, char **buf, size_t *sizep);
entry              45 tools/lib/api/fs/fs.h int sysfs__read_bool(const char *entry, bool *value);
entry              47 tools/lib/api/fs/fs.h int sysfs__write_int(const char *entry, int value);
entry              19 tools/lib/bpf/hashmap.c 			      struct hashmap_entry *entry)
entry              21 tools/lib/bpf/hashmap.c 	entry->next = *pprev;
entry              22 tools/lib/bpf/hashmap.c 	*pprev = entry;
entry              26 tools/lib/bpf/hashmap.c 			      struct hashmap_entry *entry)
entry              28 tools/lib/bpf/hashmap.c 	*pprev = entry->next;
entry              29 tools/lib/bpf/hashmap.c 	entry->next = NULL;
entry             121 tools/lib/bpf/hashmap.c 			       struct hashmap_entry **entry)
entry             134 tools/lib/bpf/hashmap.c 			*entry = cur;
entry             146 tools/lib/bpf/hashmap.c 	struct hashmap_entry *entry;
entry             157 tools/lib/bpf/hashmap.c 	    hashmap_find_entry(map, key, h, NULL, &entry)) {
entry             159 tools/lib/bpf/hashmap.c 			*old_key = entry->key;
entry             161 tools/lib/bpf/hashmap.c 			*old_value = entry->value;
entry             164 tools/lib/bpf/hashmap.c 			entry->key = key;
entry             165 tools/lib/bpf/hashmap.c 			entry->value = value;
entry             182 tools/lib/bpf/hashmap.c 	entry = malloc(sizeof(struct hashmap_entry));
entry             183 tools/lib/bpf/hashmap.c 	if (!entry)
entry             186 tools/lib/bpf/hashmap.c 	entry->key = key;
entry             187 tools/lib/bpf/hashmap.c 	entry->value = value;
entry             188 tools/lib/bpf/hashmap.c 	hashmap_add_entry(&map->buckets[h], entry);
entry             196 tools/lib/bpf/hashmap.c 	struct hashmap_entry *entry;
entry             200 tools/lib/bpf/hashmap.c 	if (!hashmap_find_entry(map, key, h, NULL, &entry))
entry             204 tools/lib/bpf/hashmap.c 		*value = entry->value;
entry             211 tools/lib/bpf/hashmap.c 	struct hashmap_entry **pprev, *entry;
entry             215 tools/lib/bpf/hashmap.c 	if (!hashmap_find_entry(map, key, h, &pprev, &entry))
entry             219 tools/lib/bpf/hashmap.c 		*old_key = entry->key;
entry             221 tools/lib/bpf/hashmap.c 		*old_value = entry->value;
entry             223 tools/lib/bpf/hashmap.c 	hashmap_del_entry(pprev, entry);
entry             224 tools/lib/bpf/hashmap.c 	free(entry);
entry            3121 tools/lib/bpf/libbpf.c 	struct hashmap_entry *entry;
entry            3176 tools/lib/bpf/libbpf.c 		hashmap__for_each_entry(cand_cache, entry, i) {
entry            3177 tools/lib/bpf/libbpf.c 			bpf_core_free_cands(entry->value);
entry             211 tools/objtool/elf.c 	struct list_head *entry, *tmp;
entry             269 tools/objtool/elf.c 		entry = &sym->sec->symbol_list;
entry             276 tools/objtool/elf.c 				entry = tmp;
entry             285 tools/objtool/elf.c 					entry = tmp;
entry             291 tools/objtool/elf.c 		list_add(&sym->list, entry);
entry              71 tools/objtool/special.c static int get_alt_entry(struct elf *elf, struct special_entry *entry,
entry              78 tools/objtool/special.c 	offset = idx * entry->size;
entry              80 tools/objtool/special.c 	alt->group = entry->group;
entry              81 tools/objtool/special.c 	alt->jump_or_nop = entry->jump_or_nop;
entry              85 tools/objtool/special.c 						   entry->orig_len);
entry              87 tools/objtool/special.c 						  entry->new_len);
entry              90 tools/objtool/special.c 	if (entry->feature) {
entry              94 tools/objtool/special.c 					      entry->feature);
entry             121 tools/objtool/special.c 	orig_rela = find_rela_by_dest(sec, offset + entry->orig);
entry             123 tools/objtool/special.c 		WARN_FUNC("can't find orig rela", sec, offset + entry->orig);
entry             128 tools/objtool/special.c 			   sec, offset + entry->orig, orig_rela->sym->name);
entry             135 tools/objtool/special.c 	if (!entry->group || alt->new_len) {
entry             136 tools/objtool/special.c 		new_rela = find_rela_by_dest(sec, offset + entry->new);
entry             139 tools/objtool/special.c 				  sec, offset + entry->new);
entry             161 tools/objtool/special.c 	struct special_entry *entry;
entry             169 tools/objtool/special.c 	for (entry = entries; entry->sec; entry++) {
entry             170 tools/objtool/special.c 		sec = find_section_by_name(elf, entry->sec);
entry             174 tools/objtool/special.c 		if (sec->len % entry->size != 0) {
entry             176 tools/objtool/special.c 			     sec->name, entry->size);
entry             180 tools/objtool/special.c 		nr_entries = sec->len / entry->size;
entry             190 tools/objtool/special.c 			ret = get_alt_entry(elf, entry, sec, idx, alt);
entry              89 tools/perf/builtin-annotate.c 	struct block_range *entry;
entry             104 tools/perf/builtin-annotate.c 	entry = block_range_iter(&iter);
entry             105 tools/perf/builtin-annotate.c 	assert(entry->is_target);
entry             106 tools/perf/builtin-annotate.c 	entry->entry++;
entry             109 tools/perf/builtin-annotate.c 		entry = block_range_iter(&iter);
entry             111 tools/perf/builtin-annotate.c 		entry->coverage++;
entry             112 tools/perf/builtin-annotate.c 		entry->sym = sym;
entry             115 tools/perf/builtin-annotate.c 			notes->max_coverage = max(notes->max_coverage, entry->coverage);
entry             122 tools/perf/builtin-annotate.c 	entry = block_range_iter(&iter);
entry             123 tools/perf/builtin-annotate.c 	assert(entry->is_branch);
entry             124 tools/perf/builtin-annotate.c 	entry->taken++;
entry             126 tools/perf/builtin-annotate.c 		entry->pred++;
entry             402 tools/perf/builtin-c2c.c 	int   (*entry)(struct perf_hpp_fmt *fmt, struct perf_hpp *hpp,
entry            1280 tools/perf/builtin-c2c.c 	.entry		= dcacheline_entry,
entry            1288 tools/perf/builtin-c2c.c 	.entry		= dcacheline_node_entry,
entry            1296 tools/perf/builtin-c2c.c 	.entry		= dcacheline_node_count,
entry            1306 tools/perf/builtin-c2c.c 	.entry		= offset_entry,
entry            1314 tools/perf/builtin-c2c.c 	.entry		= dcacheline_node_entry,
entry            1322 tools/perf/builtin-c2c.c 	.entry		= iaddr_entry,
entry            1330 tools/perf/builtin-c2c.c 	.entry		= tot_hitm_entry,
entry            1338 tools/perf/builtin-c2c.c 	.entry		= lcl_hitm_entry,
entry            1346 tools/perf/builtin-c2c.c 	.entry		= rmt_hitm_entry,
entry            1354 tools/perf/builtin-c2c.c 	.entry		= rmt_hitm_entry,
entry            1362 tools/perf/builtin-c2c.c 	.entry		= lcl_hitm_entry,
entry            1370 tools/perf/builtin-c2c.c 	.entry		= store_entry,
entry            1378 tools/perf/builtin-c2c.c 	.entry		= st_l1hit_entry,
entry            1386 tools/perf/builtin-c2c.c 	.entry		= st_l1miss_entry,
entry            1394 tools/perf/builtin-c2c.c 	.entry		= st_l1hit_entry,
entry            1402 tools/perf/builtin-c2c.c 	.entry		= st_l1miss_entry,
entry            1410 tools/perf/builtin-c2c.c 	.entry		= ld_fbhit_entry,
entry            1418 tools/perf/builtin-c2c.c 	.entry		= ld_l1hit_entry,
entry            1426 tools/perf/builtin-c2c.c 	.entry		= ld_l2hit_entry,
entry            1434 tools/perf/builtin-c2c.c 	.entry		= ld_llchit_entry,
entry            1442 tools/perf/builtin-c2c.c 	.entry		= rmt_hit_entry,
entry            1450 tools/perf/builtin-c2c.c 	.entry		= ld_llcmiss_entry,
entry            1458 tools/perf/builtin-c2c.c 	.entry		= tot_recs_entry,
entry            1466 tools/perf/builtin-c2c.c 	.entry		= tot_loads_entry,
entry            1479 tools/perf/builtin-c2c.c 	.entry		= percent_hitm_entry,
entry            1488 tools/perf/builtin-c2c.c 	.entry		= percent_rmt_hitm_entry,
entry            1497 tools/perf/builtin-c2c.c 	.entry		= percent_lcl_hitm_entry,
entry            1506 tools/perf/builtin-c2c.c 	.entry		= percent_stores_l1hit_entry,
entry            1515 tools/perf/builtin-c2c.c 	.entry		= percent_stores_l1miss_entry,
entry            1524 tools/perf/builtin-c2c.c 	.entry		= lcl_dram_entry,
entry            1532 tools/perf/builtin-c2c.c 	.entry		= rmt_dram_entry,
entry            1540 tools/perf/builtin-c2c.c 	.entry		= pid_entry,
entry            1570 tools/perf/builtin-c2c.c 	.entry		= node_entry,
entry            1578 tools/perf/builtin-c2c.c 	.entry		= mean_rmt_entry,
entry            1586 tools/perf/builtin-c2c.c 	.entry		= mean_lcl_entry,
entry            1594 tools/perf/builtin-c2c.c 	.entry		= mean_load_entry,
entry            1602 tools/perf/builtin-c2c.c 	.entry		= cpucnt_entry,
entry            1615 tools/perf/builtin-c2c.c 	.entry		= cl_idx_entry,
entry            1623 tools/perf/builtin-c2c.c 	.entry		= cl_idx_entry,
entry            1631 tools/perf/builtin-c2c.c 	.entry		= cl_idx_empty_entry,
entry            1772 tools/perf/builtin-c2c.c 	fmt->entry	= dim->se ? c2c_se_entry : dim->entry;
entry            1640 tools/perf/builtin-diff.c 	fmt->entry  = hpp__entry_global;
entry             426 tools/perf/builtin-ftrace.c 	struct filter_entry *entry;
entry             428 tools/perf/builtin-ftrace.c 	entry = malloc(sizeof(*entry) + strlen(str) + 1);
entry             429 tools/perf/builtin-ftrace.c 	if (entry == NULL)
entry             432 tools/perf/builtin-ftrace.c 	strcpy(entry->name, str);
entry             433 tools/perf/builtin-ftrace.c 	list_add_tail(&entry->list, head);
entry             320 tools/perf/builtin-lock.c 	struct list_head *entry = lockhashentry(addr);
entry             323 tools/perf/builtin-lock.c 	list_for_each_entry(ret, entry, hash_entry) {
entry             342 tools/perf/builtin-lock.c 	list_add(&new->hash_entry, entry);
entry            2853 tools/perf/builtin-trace.c static void trace__init_bpf_map_syscall_args(struct trace *trace, int id, struct bpf_map_syscall_entry *entry)
entry            2862 tools/perf/builtin-trace.c 		entry->string_args_len[arg] = 0;
entry            2865 tools/perf/builtin-trace.c 			entry->string_args_len[arg] = PATH_MAX;
entry            2870 tools/perf/builtin-trace.c 		entry->string_args_len[arg] = 0;
entry            3676 tools/perf/builtin-trace.c 	entry->syscall = source->i;
entry            3677 tools/perf/builtin-trace.c 	entry->stats   = stats;
entry            3678 tools/perf/builtin-trace.c 	entry->msecs   = stats ? (u64)stats->n * (avg_stats(stats) / NSEC_PER_MSEC) : 0;
entry            3761 tools/perf/builtin-trace.c 	entry->thread = rb_entry(nd, struct thread, rb_node);
entry             146 tools/perf/lib/evsel.c 	int entry = sizeof(u64); /* value */
entry             157 tools/perf/lib/evsel.c 		entry += sizeof(u64);
entry             164 tools/perf/lib/evsel.c 	size += entry * nr;
entry              59 tools/perf/tests/dwarf-unwind.c static int unwind_entry(struct unwind_entry *entry, void *arg)
entry              62 tools/perf/tests/dwarf-unwind.c 	char *symbol = entry->sym ? entry->sym->name : NULL;
entry              87 tools/perf/tests/dwarf-unwind.c 			 entry->ip);
entry              93 tools/perf/tests/dwarf-unwind.c 		 symbol, entry->ip, funcs[idx]);
entry              29 tools/perf/ui/browser.h 	void	      (*write)(struct ui_browser *browser, void *entry, int row);
entry              31 tools/perf/ui/browser.h 	bool	      (*filter)(struct ui_browser *browser, void *entry);
entry              46 tools/perf/ui/browsers/annotate.c static bool disasm_line__filter(struct ui_browser *browser, void *entry)
entry              49 tools/perf/ui/browsers/annotate.c 	struct annotation_line *al = list_entry(entry, struct annotation_line, node);
entry              96 tools/perf/ui/browsers/annotate.c static void annotate_browser__write(struct ui_browser *browser, void *entry, int row)
entry             100 tools/perf/ui/browsers/annotate.c 	struct annotation_line *al = list_entry(entry, struct annotation_line, node);
entry              13 tools/perf/ui/browsers/header.c 				   void *entry, int row)
entry              15 tools/perf/ui/browsers/header.c 	char **arg = entry;
entry            1113 tools/perf/ui/browsers/hists.c 					struct hist_entry *entry, int level,
entry            1119 tools/perf/ui/browsers/hists.c 	u64 total = hists__total_period(entry->hists);
entry            1124 tools/perf/ui/browsers/hists.c 		parent_total = entry->stat_acc->period;
entry            1126 tools/perf/ui/browsers/hists.c 		parent_total = entry->stat.period;
entry            1130 tools/perf/ui/browsers/hists.c 						&entry->sorted_chain, row,
entry            1135 tools/perf/ui/browsers/hists.c 						&entry->sorted_chain, row,
entry            1140 tools/perf/ui/browsers/hists.c 						&entry->sorted_chain, level, row,
entry            1146 tools/perf/ui/browsers/hists.c 		browser->he_selection = entry;
entry            1245 tools/perf/ui/browsers/hists.c 				    struct hist_entry *entry,
entry            1252 tools/perf/ui/browsers/hists.c 	bool use_callchain = hist_entry__has_callchains(entry) && symbol_conf.use_callchain;
entry            1253 tools/perf/ui/browsers/hists.c 	off_t row_offset = entry->row_offset;
entry            1258 tools/perf/ui/browsers/hists.c 		browser->he_selection = entry;
entry            1259 tools/perf/ui/browsers/hists.c 		browser->selection = &entry->ms;
entry            1263 tools/perf/ui/browsers/hists.c 		hist_entry__init_have_children(entry);
entry            1264 tools/perf/ui/browsers/hists.c 		folded_sign = hist_entry__folded(entry);
entry            1285 tools/perf/ui/browsers/hists.c 			if (perf_hpp__should_skip(fmt, entry->hists) ||
entry            1309 tools/perf/ui/browsers/hists.c 				int ret = fmt->color(fmt, &hpp, entry);
entry            1310 tools/perf/ui/browsers/hists.c 				hist_entry__snprintf_alignment(entry, &hpp, fmt, ret);
entry            1317 tools/perf/ui/browsers/hists.c 				hist_entry__snprintf_alignment(entry, &hpp, fmt, fmt->entry(fmt, &hpp, entry));
entry            1341 tools/perf/ui/browsers/hists.c 				entry, 1, row,
entry            1351 tools/perf/ui/browsers/hists.c 					      struct hist_entry *entry,
entry            1359 tools/perf/ui/browsers/hists.c 	off_t row_offset = entry->row_offset;
entry            1368 tools/perf/ui/browsers/hists.c 	int hierarchy_indent = (entry->hists->nr_hpp_node - 2) * HIERARCHY_INDENT;
entry            1371 tools/perf/ui/browsers/hists.c 		browser->he_selection = entry;
entry            1372 tools/perf/ui/browsers/hists.c 		browser->selection = &entry->ms;
entry            1375 tools/perf/ui/browsers/hists.c 	hist_entry__init_have_children(entry);
entry            1376 tools/perf/ui/browsers/hists.c 	folded_sign = hist_entry__folded(entry);
entry            1379 tools/perf/ui/browsers/hists.c 	if (entry->leaf && row_offset) {
entry            1395 tools/perf/ui/browsers/hists.c 	fmt_node = list_first_entry(&entry->hists->hpp_formats,
entry            1405 tools/perf/ui/browsers/hists.c 		if (perf_hpp__should_skip(fmt, entry->hists) ||
entry            1427 tools/perf/ui/browsers/hists.c 			int ret = fmt->color(fmt, &hpp, entry);
entry            1428 tools/perf/ui/browsers/hists.c 			hist_entry__snprintf_alignment(entry, &hpp, fmt, ret);
entry            1435 tools/perf/ui/browsers/hists.c 			int ret = fmt->entry(fmt, &hpp, entry);
entry            1436 tools/perf/ui/browsers/hists.c 			hist_entry__snprintf_alignment(entry, &hpp, fmt, ret);
entry            1463 tools/perf/ui/browsers/hists.c 		perf_hpp_list__for_each_format(entry->hpp_list, fmt) {
entry            1479 tools/perf/ui/browsers/hists.c 				width -= fmt->color(fmt, &hpp, entry);
entry            1483 tools/perf/ui/browsers/hists.c 				width -= fmt->entry(fmt, &hpp, entry);
entry            1502 tools/perf/ui/browsers/hists.c 	if (entry->leaf && folded_sign == '-' && row != browser->b.rows) {
entry            1507 tools/perf/ui/browsers/hists.c 		printed += hist_browser__show_callchain(browser, entry,
entry            2024 tools/perf/ui/browsers/hists.c 		ret = fmt->entry(fmt, &hpp, he);
entry            2069 tools/perf/ui/browsers/hists.c 		ret = fmt->entry(fmt, &hpp, he);
entry            2080 tools/perf/ui/browsers/hists.c 		ret = fmt->entry(fmt, &hpp, he);
entry            3221 tools/perf/ui/browsers/hists.c 				   void *entry, int row)
entry            3225 tools/perf/ui/browsers/hists.c 	struct evsel *evsel = list_entry(entry, struct evsel, core.node);
entry            3362 tools/perf/ui/browsers/hists.c 				 void *entry)
entry            3364 tools/perf/ui/browsers/hists.c 	struct evsel *evsel = list_entry(entry, struct evsel, core.node);
entry             382 tools/perf/ui/gtk/hists.c 				fmt->entry(fmt, &hpp, h);
entry             444 tools/perf/ui/gtk/hists.c 				fmt->entry(fmt, hpp, he);
entry             457 tools/perf/ui/gtk/hists.c 				ret = fmt->entry(fmt, hpp, he);
entry             404 tools/perf/ui/hist.c 		.entry	= hpp__entry_ ## _fn,		\
entry             418 tools/perf/ui/hist.c 		.entry	= hpp__entry_ ## _fn,		\
entry             431 tools/perf/ui/hist.c 		.entry	= hpp__entry_ ## _fn,		\
entry             578 tools/perf/ui/hist.c 		if (!fmt->entry && !fmt->color)
entry             442 tools/perf/ui/stdio/hist.c 			ret = fmt->entry(fmt, hpp, he);
entry             492 tools/perf/ui/stdio/hist.c 			ret = fmt->entry(fmt, hpp, he);
entry             516 tools/perf/ui/stdio/hist.c 			fmt->entry(fmt, hpp, he);
entry              16 tools/perf/ui/tui/util.c 				   void *entry, int row)
entry              18 tools/perf/ui/tui/util.c 	char **arg = entry;
entry            1319 tools/perf/util/annotate.c 		p = 100 *(double)br->entry / branch->coverage;
entry            1410 tools/perf/util/auxtrace.c 	struct auxtrace_cache_entry *entry;
entry            1418 tools/perf/util/auxtrace.c 		hlist_for_each_entry_safe(entry, tmp, &c->hashtable[i], hash) {
entry            1419 tools/perf/util/auxtrace.c 			hlist_del(&entry->hash);
entry            1420 tools/perf/util/auxtrace.c 			auxtrace_cache__free_entry(c, entry);
entry            1443 tools/perf/util/auxtrace.c 				void *entry)
entry            1445 tools/perf/util/auxtrace.c 	free(entry);
entry            1449 tools/perf/util/auxtrace.c 			struct auxtrace_cache_entry *entry)
entry            1454 tools/perf/util/auxtrace.c 	entry->key = key;
entry            1455 tools/perf/util/auxtrace.c 	hlist_add_head(&entry->hash, &c->hashtable[hash_32(key, c->bits)]);
entry            1462 tools/perf/util/auxtrace.c 	struct auxtrace_cache_entry *entry;
entry            1469 tools/perf/util/auxtrace.c 	hlist_for_each_entry(entry, hlist, hash) {
entry            1470 tools/perf/util/auxtrace.c 		if (entry->key == key)
entry            1471 tools/perf/util/auxtrace.c 			return entry;
entry             489 tools/perf/util/auxtrace.h void auxtrace_cache__free_entry(struct auxtrace_cache *c, void *entry);
entry             491 tools/perf/util/auxtrace.h 			struct auxtrace_cache_entry *entry);
entry              23 tools/perf/util/block-range.c 		struct block_range *entry = rb_entry(rb, struct block_range, node);
entry              25 tools/perf/util/block-range.c 		assert(old < entry->start);
entry              26 tools/perf/util/block-range.c 		assert(entry->start <= entry->end); /* single instruction block; jump to a jump */
entry              28 tools/perf/util/block-range.c 		old = entry->end;
entry              37 tools/perf/util/block-range.c 	struct block_range *entry;
entry              41 tools/perf/util/block-range.c 		entry = rb_entry(parent, struct block_range, node);
entry              43 tools/perf/util/block-range.c 		if (addr < entry->start)
entry              45 tools/perf/util/block-range.c 		else if (addr > entry->end)
entry              48 tools/perf/util/block-range.c 			return entry;
entry              85 tools/perf/util/block-range.c 	struct block_range *next, *entry = NULL;
entry              90 tools/perf/util/block-range.c 		entry = rb_entry(parent, struct block_range, node);
entry              92 tools/perf/util/block-range.c 		if (start < entry->start)
entry              94 tools/perf/util/block-range.c 		else if (start > entry->end)
entry             105 tools/perf/util/block-range.c 		if (!entry) /* tree empty */
entry             112 tools/perf/util/block-range.c 		if (entry->end < start) {
entry             143 tools/perf/util/block-range.c 		entry = malloc(sizeof(struct block_range));
entry             144 tools/perf/util/block-range.c 		if (!entry)
entry             147 tools/perf/util/block-range.c 		*entry = (struct block_range){
entry             154 tools/perf/util/block-range.c 		rb_link_node(&entry->node, parent, p);
entry             155 tools/perf/util/block-range.c 		rb_insert_color(&entry->node, &block_ranges.root);
entry             158 tools/perf/util/block-range.c 		iter.start = entry;
entry             159 tools/perf/util/block-range.c 		iter.end   = entry;
entry             166 tools/perf/util/block-range.c 	if (entry->start < start) { /* split: [e->start...][start...] */
entry             172 tools/perf/util/block-range.c 			.start		= entry->start,
entry             174 tools/perf/util/block-range.c 			.is_target	= entry->is_target,
entry             177 tools/perf/util/block-range.c 			.coverage	= entry->coverage,
entry             178 tools/perf/util/block-range.c 			.entry		= entry->entry,
entry             181 tools/perf/util/block-range.c 		entry->start		= start;
entry             182 tools/perf/util/block-range.c 		entry->is_target	= 1;
entry             183 tools/perf/util/block-range.c 		entry->entry		= 0;
entry             185 tools/perf/util/block-range.c 		rb_link_left_of_node(&head->node, &entry->node);
entry             189 tools/perf/util/block-range.c 	} else if (entry->start == start)
entry             190 tools/perf/util/block-range.c 		entry->is_target = 1;
entry             192 tools/perf/util/block-range.c 	iter.start = entry;
entry             199 tools/perf/util/block-range.c 	entry = iter.start;
entry             204 tools/perf/util/block-range.c 		if (end < entry->end) { /* split: [...end][...e->end] */
entry             211 tools/perf/util/block-range.c 				.end		= entry->end,
entry             213 tools/perf/util/block-range.c 				.is_branch	= entry->is_branch,
entry             215 tools/perf/util/block-range.c 				.coverage	= entry->coverage,
entry             216 tools/perf/util/block-range.c 				.taken		= entry->taken,
entry             217 tools/perf/util/block-range.c 				.pred		= entry->pred,
entry             220 tools/perf/util/block-range.c 			entry->end		= end;
entry             221 tools/perf/util/block-range.c 			entry->is_branch	= 1;
entry             222 tools/perf/util/block-range.c 			entry->taken		= 0;
entry             223 tools/perf/util/block-range.c 			entry->pred		= 0;
entry             225 tools/perf/util/block-range.c 			rb_link_right_of_node(&tail->node, &entry->node);
entry             229 tools/perf/util/block-range.c 			iter.end = entry;
entry             236 tools/perf/util/block-range.c 		if (end == entry->end) {
entry             237 tools/perf/util/block-range.c 			entry->is_branch = 1;
entry             238 tools/perf/util/block-range.c 			iter.end = entry;
entry             242 tools/perf/util/block-range.c 		next = block_range__next(entry);
entry             257 tools/perf/util/block-range.c 				.start		= entry->end + 1,
entry             263 tools/perf/util/block-range.c 			rb_link_right_of_node(&tail->node, &entry->node);
entry             274 tools/perf/util/block-range.c 		if (entry->end + 1 != next->start) {
entry             280 tools/perf/util/block-range.c 				.start		= entry->end + 1,
entry             291 tools/perf/util/block-range.c 		entry = next;
entry              33 tools/perf/util/block-range.h 	u64 entry;
entry            1039 tools/perf/util/dwarf-aux.c 	Dwarf_Addr entry;
entry            1044 tools/perf/util/dwarf-aux.c 	ret = die_entrypc(sp_die, &entry);
entry            1062 tools/perf/util/dwarf-aux.c 		start -= entry;
entry            1063 tools/perf/util/dwarf-aux.c 		end -= entry;
entry            1099 tools/perf/util/dwarf-aux.c 	Dwarf_Addr entry;
entry            1107 tools/perf/util/dwarf-aux.c 	ret = die_entrypc(sp_die, &entry);
entry            1127 tools/perf/util/dwarf-aux.c 		start -= entry;
entry            1128 tools/perf/util/dwarf-aux.c 		end -= entry;
entry             160 tools/perf/util/evlist.c void evlist__add(struct evlist *evlist, struct evsel *entry)
entry             162 tools/perf/util/evlist.c 	entry->evlist = evlist;
entry             163 tools/perf/util/evlist.c 	entry->idx = evlist->core.nr_entries;
entry             164 tools/perf/util/evlist.c 	entry->tracking = !entry->idx;
entry             166 tools/perf/util/evlist.c 	perf_evlist__add(&evlist->core, &entry->core);
entry              92 tools/perf/util/evlist.h void evlist__add(struct evlist *evlist, struct evsel *entry);
entry             565 tools/perf/util/hist.c 					       struct hist_entry *entry,
entry             573 tools/perf/util/hist.c 	u64 period = entry->stat.period;
entry             574 tools/perf/util/hist.c 	u64 weight = entry->stat.weight;
entry             589 tools/perf/util/hist.c 		cmp = hist_entry__cmp(he, entry);
entry             603 tools/perf/util/hist.c 			mem_info__zput(entry->mem_info);
entry             605 tools/perf/util/hist.c 			block_info__zput(entry->block_info);
entry             613 tools/perf/util/hist.c 			if (he->ms.map != entry->ms.map) {
entry             615 tools/perf/util/hist.c 				he->ms.map = map__get(entry->ms.map);
entry             628 tools/perf/util/hist.c 	he = hist_entry__new(entry, sample_self);
entry             684 tools/perf/util/hist.c 	struct hist_entry entry = {
entry             717 tools/perf/util/hist.c 	}, *he = hists__findnew_entry(hists, &entry, al, sample_self);
entry             755 tools/perf/util/hist.c 	struct hist_entry entry = {
entry             758 tools/perf/util/hist.c 	}, *he = hists__findnew_entry(hists, &entry, al, false);
entry             264 tools/perf/util/hist.h 	int (*entry)(struct perf_hpp_fmt *fmt, struct perf_hpp *hpp,
entry             394 tools/perf/util/intel-pt.c 	struct auxtrace_cache_entry	entry;
entry             489 tools/perf/util/intel-pt.c 	err = auxtrace_cache__add(c, offset, &e->entry);
entry              14 tools/perf/util/intlist.c 					 const void *entry)
entry              16 tools/perf/util/intlist.c 	int i = (int)((long)entry);
entry              42 tools/perf/util/intlist.c static int intlist__node_cmp(struct rb_node *rb_node, const void *entry)
entry              44 tools/perf/util/intlist.c 	int i = (int)((long)entry);
entry              37 tools/perf/util/jitdump.c 	union jr_entry   *entry;
entry            2456 tools/perf/util/machine.c static int unwind_entry(struct unwind_entry *entry, void *arg)
entry            2460 tools/perf/util/machine.c 	u64 addr = entry->ip;
entry            2462 tools/perf/util/machine.c 	if (symbol_conf.hide_unresolved && entry->sym == NULL)
entry            2465 tools/perf/util/machine.c 	if (append_inlines(cursor, entry->map, entry->sym, entry->ip) == 0)
entry            2472 tools/perf/util/machine.c 	if (entry->map)
entry            2473 tools/perf/util/machine.c 		addr = map__map_ip(entry->map, entry->ip);
entry            2475 tools/perf/util/machine.c 	srcline = callchain_srcline(entry->map, entry->sym, addr);
entry            2476 tools/perf/util/machine.c 	return callchain_cursor_append(cursor, entry->ip,
entry            2477 tools/perf/util/machine.c 				       entry->map, entry->sym,
entry              17 tools/perf/util/mem2node.c static void phys_entry__insert(struct phys_entry *entry, struct rb_root *root)
entry              27 tools/perf/util/mem2node.c 		if (entry->start < e->start)
entry              33 tools/perf/util/mem2node.c 	rb_link_node(&entry->rb_node, parent, p);
entry              34 tools/perf/util/mem2node.c 	rb_insert_color(&entry->rb_node, root);
entry              38 tools/perf/util/mem2node.c phys_entry__init(struct phys_entry *entry, u64 start, u64 bsize, u64 node)
entry              40 tools/perf/util/mem2node.c 	entry->start = start;
entry              41 tools/perf/util/mem2node.c 	entry->end   = start + bsize;
entry              42 tools/perf/util/mem2node.c 	entry->node  = node;
entry              43 tools/perf/util/mem2node.c 	RB_CLEAR_NODE(&entry->rb_node);
entry             120 tools/perf/util/mem2node.c 	struct phys_entry *entry;
entry             125 tools/perf/util/mem2node.c 		entry = rb_entry(parent, struct phys_entry, rb_node);
entry             126 tools/perf/util/mem2node.c 		if (addr < entry->start)
entry             128 tools/perf/util/mem2node.c 		else if (addr >= entry->end)
entry             134 tools/perf/util/mem2node.c 	entry = NULL;
entry             136 tools/perf/util/mem2node.c 	return entry ? (int) entry->node : -1;
entry              50 tools/perf/util/metricgroup.c static int metric_event_cmp(struct rb_node *rb_node, const void *entry)
entry              55 tools/perf/util/metricgroup.c 	const struct metric_event *b = entry;
entry              65 tools/perf/util/metricgroup.c 					const void *entry)
entry              71 tools/perf/util/metricgroup.c 	memcpy(me, entry, sizeof(struct metric_event));
entry              72 tools/perf/util/metricgroup.c 	me->evsel = ((struct metric_event *)entry)->evsel;
entry             227 tools/perf/util/metricgroup.c static int mep_cmp(struct rb_node *rb_node, const void *entry)
entry             230 tools/perf/util/metricgroup.c 	struct mep *b = (struct mep *)entry;
entry             236 tools/perf/util/metricgroup.c 					const void *entry)
entry             242 tools/perf/util/metricgroup.c 	memcpy(me, entry, sizeof(struct mep));
entry            3177 tools/perf/util/probe-event.c 	struct probe_cache_entry *entry;
entry            3187 tools/perf/util/probe-event.c 	for_each_probe_cache_entry(entry, cache) {
entry            3189 tools/perf/util/probe-event.c 		if (!entry->pev.event || !entry->pev.group)
entry            3191 tools/perf/util/probe-event.c 		if ((!pev->group || strglobmatch(entry->pev.group, pev->group)) &&
entry            3192 tools/perf/util/probe-event.c 		    strglobmatch(entry->pev.event, pev->event)) {
entry            3193 tools/perf/util/probe-event.c 			ret = probe_cache_entry__get_event(entry, &tmp_tevs);
entry            3260 tools/perf/util/probe-event.c 	struct probe_cache_entry *entry;
entry            3276 tools/perf/util/probe-event.c 	entry = probe_cache__find(cache, pev);
entry            3277 tools/perf/util/probe-event.c 	if (!entry) {
entry            3283 tools/perf/util/probe-event.c 	ret = strlist__nr_entries(entry->tevlist);
entry            3298 tools/perf/util/probe-event.c 	strlist__for_each_entry(node, entry->tevlist) {
entry             349 tools/perf/util/probe-file.c static void probe_cache_entry__delete(struct probe_cache_entry *entry)
entry             351 tools/perf/util/probe-file.c 	if (entry) {
entry             352 tools/perf/util/probe-file.c 		BUG_ON(!list_empty(&entry->node));
entry             354 tools/perf/util/probe-file.c 		strlist__delete(entry->tevlist);
entry             355 tools/perf/util/probe-file.c 		clear_perf_probe_event(&entry->pev);
entry             356 tools/perf/util/probe-file.c 		zfree(&entry->spev);
entry             357 tools/perf/util/probe-file.c 		free(entry);
entry             364 tools/perf/util/probe-file.c 	struct probe_cache_entry *entry = zalloc(sizeof(*entry));
entry             366 tools/perf/util/probe-file.c 	if (entry) {
entry             367 tools/perf/util/probe-file.c 		INIT_LIST_HEAD(&entry->node);
entry             368 tools/perf/util/probe-file.c 		entry->tevlist = strlist__new(NULL, NULL);
entry             369 tools/perf/util/probe-file.c 		if (!entry->tevlist)
entry             370 tools/perf/util/probe-file.c 			zfree(&entry);
entry             372 tools/perf/util/probe-file.c 			entry->spev = synthesize_perf_probe_command(pev);
entry             373 tools/perf/util/probe-file.c 			if (!entry->spev ||
entry             374 tools/perf/util/probe-file.c 			    perf_probe_event__copy(&entry->pev, pev) < 0) {
entry             375 tools/perf/util/probe-file.c 				probe_cache_entry__delete(entry);
entry             381 tools/perf/util/probe-file.c 	return entry;
entry             384 tools/perf/util/probe-file.c int probe_cache_entry__get_event(struct probe_cache_entry *entry,
entry             391 tools/perf/util/probe-file.c 	ret = strlist__nr_entries(entry->tevlist);
entry             400 tools/perf/util/probe-file.c 	strlist__for_each_entry(node, entry->tevlist) {
entry             472 tools/perf/util/probe-file.c 	struct probe_cache_entry *entry = NULL;
entry             494 tools/perf/util/probe-file.c 			entry = probe_cache_entry__new(NULL);
entry             495 tools/perf/util/probe-file.c 			if (!entry) {
entry             500 tools/perf/util/probe-file.c 				entry->sdt = true;
entry             501 tools/perf/util/probe-file.c 			entry->spev = strdup(buf + 1);
entry             502 tools/perf/util/probe-file.c 			if (entry->spev)
entry             504 tools/perf/util/probe-file.c 								&entry->pev);
entry             508 tools/perf/util/probe-file.c 				probe_cache_entry__delete(entry);
entry             511 tools/perf/util/probe-file.c 			list_add_tail(&entry->node, &pcache->entries);
entry             513 tools/perf/util/probe-file.c 			if (!entry) {
entry             517 tools/perf/util/probe-file.c 			strlist__add(entry->tevlist, buf);
entry             538 tools/perf/util/probe-file.c 	struct probe_cache_entry *entry, *n;
entry             540 tools/perf/util/probe-file.c 	list_for_each_entry_safe(entry, n, &pcache->entries, node) {
entry             541 tools/perf/util/probe-file.c 		list_del_init(&entry->node);
entry             542 tools/perf/util/probe-file.c 		probe_cache_entry__delete(entry);
entry             598 tools/perf/util/probe-file.c 	struct probe_cache_entry *entry = NULL;
entry             604 tools/perf/util/probe-file.c 	for_each_probe_cache_entry(entry, pcache) {
entry             606 tools/perf/util/probe-file.c 			if (entry->pev.event &&
entry             607 tools/perf/util/probe-file.c 			    streql(entry->pev.event, pev->event) &&
entry             609 tools/perf/util/probe-file.c 			     streql(entry->pev.group, pev->group)))
entry             616 tools/perf/util/probe-file.c 		     (streql(entry->pev.group, pev->group) &&
entry             617 tools/perf/util/probe-file.c 		      streql(entry->pev.event, pev->event))) ||
entry             618 tools/perf/util/probe-file.c 		    (!strcmp(entry->spev, cmd)))
entry             621 tools/perf/util/probe-file.c 	entry = NULL;
entry             625 tools/perf/util/probe-file.c 	return entry;
entry             632 tools/perf/util/probe-file.c 	struct probe_cache_entry *entry = NULL;
entry             634 tools/perf/util/probe-file.c 	for_each_probe_cache_entry(entry, pcache) {
entry             636 tools/perf/util/probe-file.c 		if (streql(entry->pev.group, group) &&
entry             637 tools/perf/util/probe-file.c 		    streql(entry->pev.event, event))
entry             640 tools/perf/util/probe-file.c 	entry = NULL;
entry             643 tools/perf/util/probe-file.c 	return entry;
entry             650 tools/perf/util/probe-file.c 	struct probe_cache_entry *entry = NULL;
entry             660 tools/perf/util/probe-file.c 	entry = probe_cache__find(pcache, pev);
entry             661 tools/perf/util/probe-file.c 	if (entry) {
entry             662 tools/perf/util/probe-file.c 		list_del_init(&entry->node);
entry             663 tools/perf/util/probe-file.c 		probe_cache_entry__delete(entry);
entry             667 tools/perf/util/probe-file.c 	entry = probe_cache_entry__new(pev);
entry             668 tools/perf/util/probe-file.c 	if (!entry)
entry             678 tools/perf/util/probe-file.c 		strlist__add(entry->tevlist, command);
entry             681 tools/perf/util/probe-file.c 	list_add_tail(&entry->node, &pcache->entries);
entry             687 tools/perf/util/probe-file.c 	probe_cache_entry__delete(entry);
entry             819 tools/perf/util/probe-file.c 	struct probe_cache_entry *entry = NULL;
entry             837 tools/perf/util/probe-file.c 		entry = probe_cache__find_by_name(pcache, sdtgrp, note->name);
entry             838 tools/perf/util/probe-file.c 		if (!entry) {
entry             839 tools/perf/util/probe-file.c 			entry = probe_cache_entry__new(NULL);
entry             840 tools/perf/util/probe-file.c 			if (!entry) {
entry             844 tools/perf/util/probe-file.c 			entry->sdt = true;
entry             845 tools/perf/util/probe-file.c 			ret = asprintf(&entry->spev, "%s:%s=%s", sdtgrp,
entry             849 tools/perf/util/probe-file.c 			entry->pev.event = strdup(note->name);
entry             850 tools/perf/util/probe-file.c 			entry->pev.group = strdup(sdtgrp);
entry             851 tools/perf/util/probe-file.c 			list_add_tail(&entry->node, &pcache->entries);
entry             859 tools/perf/util/probe-file.c 		strlist__add(entry->tevlist, buf);
entry             861 tools/perf/util/probe-file.c 		entry = NULL;
entry             863 tools/perf/util/probe-file.c 	if (entry) {
entry             864 tools/perf/util/probe-file.c 		list_del_init(&entry->node);
entry             865 tools/perf/util/probe-file.c 		probe_cache_entry__delete(entry);
entry             872 tools/perf/util/probe-file.c static int probe_cache_entry__write(struct probe_cache_entry *entry, int fd)
entry             877 tools/perf/util/probe-file.c 	const char *prefix = entry->sdt ? "%" : "#";
entry             884 tools/perf/util/probe-file.c 	pr_debug("Writing cache: %s%s\n", prefix, entry->spev);
entry             886 tools/perf/util/probe-file.c 	iov[1].iov_base = entry->spev; iov[1].iov_len = strlen(entry->spev);
entry             892 tools/perf/util/probe-file.c 	strlist__for_each_entry(snode, entry->tevlist) {
entry             914 tools/perf/util/probe-file.c 	struct probe_cache_entry *entry;
entry             926 tools/perf/util/probe-file.c 	for_each_probe_cache_entry(entry, pcache) {
entry             927 tools/perf/util/probe-file.c 		ret = probe_cache_entry__write(entry, pcache->fd);
entry             936 tools/perf/util/probe-file.c static bool probe_cache_entry__compare(struct probe_cache_entry *entry,
entry             939 tools/perf/util/probe-file.c 	char buf[128], *ptr = entry->spev;
entry             941 tools/perf/util/probe-file.c 	if (entry->pev.event) {
entry             942 tools/perf/util/probe-file.c 		snprintf(buf, 128, "%s:%s", entry->pev.group, entry->pev.event);
entry             951 tools/perf/util/probe-file.c 	struct probe_cache_entry *entry, *tmp;
entry             953 tools/perf/util/probe-file.c 	list_for_each_entry_safe(entry, tmp, &pcache->entries, node) {
entry             954 tools/perf/util/probe-file.c 		if (probe_cache_entry__compare(entry, filter)) {
entry             955 tools/perf/util/probe-file.c 			pr_info("Removed cached event: %s\n", entry->spev);
entry             956 tools/perf/util/probe-file.c 			list_del_init(&entry->node);
entry             957 tools/perf/util/probe-file.c 			probe_cache_entry__delete(entry);
entry             966 tools/perf/util/probe-file.c 	struct probe_cache_entry *entry;
entry             968 tools/perf/util/probe-file.c 	for_each_probe_cache_entry(entry, pcache) {
entry             969 tools/perf/util/probe-file.c 		if (probe_cache_entry__compare(entry, filter))
entry             970 tools/perf/util/probe-file.c 			printf("%s\n", entry->spev);
entry              35 tools/perf/util/probe-file.h #define for_each_probe_cache_entry(entry, pcache) \
entry              36 tools/perf/util/probe-file.h 	list_for_each_entry(entry, &pcache->entries, node)
entry              52 tools/perf/util/probe-file.h int probe_cache_entry__get_event(struct probe_cache_entry *entry,
entry              61 tools/perf/util/rb_resort.h 					struct __name##_sorted_entry *entry);	\
entry             121 tools/perf/util/rb_resort.h 					struct __name##_sorted_entry *entry)
entry              54 tools/perf/util/rblist.c 					 const void *entry,
entry              66 tools/perf/util/rblist.c 		rc = rblist->node_cmp(parent, entry);
entry              78 tools/perf/util/rblist.c 		new_node = rblist->node_new(rblist, entry);
entry              90 tools/perf/util/rblist.c struct rb_node *rblist__find(struct rblist *rblist, const void *entry)
entry              92 tools/perf/util/rblist.c 	return __rblist__findnew(rblist, entry, false);
entry              95 tools/perf/util/rblist.c struct rb_node *rblist__findnew(struct rblist *rblist, const void *entry)
entry              97 tools/perf/util/rblist.c 	return __rblist__findnew(rblist, entry, true);
entry              26 tools/perf/util/rblist.h 	int (*node_cmp)(struct rb_node *rbn, const void *entry);
entry              36 tools/perf/util/rblist.h struct rb_node *rblist__find(struct rblist *rblist, const void *entry);
entry              37 tools/perf/util/rblist.h struct rb_node *rblist__findnew(struct rblist *rblist, const void *entry);
entry            1655 tools/perf/util/sort.c 	struct sort_entry	*entry;
entry            1659 tools/perf/util/sort.c #define DIM(d, n, func) [d] = { .name = n, .entry = &(func) }
entry            1684 tools/perf/util/sort.c #define DIM(d, n, func) [d - __SORT_BRANCH_STACK] = { .name = n, .entry = &(func) }
entry            1702 tools/perf/util/sort.c #define DIM(d, n, func) [d - __SORT_MEMORY_MODE] = { .name = n, .entry = &(func) }
entry            1889 tools/perf/util/sort.c 	hse->se = sd->entry;
entry            1890 tools/perf/util/sort.c 	hse->hpp.name = sd->entry->se_header;
entry            1893 tools/perf/util/sort.c 	hse->hpp.entry = __sort__hpp_entry;
entry            2242 tools/perf/util/sort.c 	hde->hpp.entry  = __sort__hde_entry;
entry            2510 tools/perf/util/sort.c 	if (sd->entry->se_collapse)
entry            2584 tools/perf/util/sort.c 		if (sd->entry == &sort_parent) {
entry            2594 tools/perf/util/sort.c 		} else if (sd->entry == &sort_sym) {
entry            2603 tools/perf/util/sort.c 				sd->entry->se_collapse = sort__sym_sort;
entry            2605 tools/perf/util/sort.c 		} else if (sd->entry == &sort_dso) {
entry            2607 tools/perf/util/sort.c 		} else if (sd->entry == &sort_socket) {
entry            2609 tools/perf/util/sort.c 		} else if (sd->entry == &sort_thread) {
entry            2611 tools/perf/util/sort.c 		} else if (sd->entry == &sort_comm) {
entry            2636 tools/perf/util/sort.c 		if (sd->entry == &sort_sym_from || sd->entry == &sort_sym_to)
entry            2652 tools/perf/util/sort.c 		if (sd->entry == &sort_mem_dcacheline && cacheline_size() == 0)
entry            2655 tools/perf/util/sort.c 		if (sd->entry == &sort_mem_daddr_sym)
entry              37 tools/perf/util/stat-shadow.c static int saved_value_cmp(struct rb_node *rb_node, const void *entry)
entry              42 tools/perf/util/stat-shadow.c 	const struct saved_value *b = entry;
entry              78 tools/perf/util/stat-shadow.c 				     const void *entry)
entry              84 tools/perf/util/stat-shadow.c 	memcpy(nd, entry, sizeof(struct saved_value));
entry              15 tools/perf/util/strlist.c struct rb_node *strlist__node_new(struct rblist *rblist, const void *entry)
entry              17 tools/perf/util/strlist.c 	const char *s = entry;
entry              55 tools/perf/util/strlist.c static int strlist__node_cmp(struct rb_node *rb_node, const void *entry)
entry              57 tools/perf/util/strlist.c 	const char *str = entry;
entry              70 tools/perf/util/strlist.c 	char entry[1024];
entry              77 tools/perf/util/strlist.c 	while (fgets(entry, sizeof(entry), fp) != NULL) {
entry              78 tools/perf/util/strlist.c 		const size_t len = strlen(entry);
entry              82 tools/perf/util/strlist.c 		entry[len - 1] = '\0';
entry              84 tools/perf/util/strlist.c 		err = strlist__add(slist, entry);
entry             100 tools/perf/util/strlist.c struct str_node *strlist__find(struct strlist *slist, const char *entry)
entry             103 tools/perf/util/strlist.c 	struct rb_node *rb_node = rblist__find(&slist->rblist, entry);
entry              40 tools/perf/util/strlist.h struct str_node *strlist__find(struct strlist *slist, const char *entry);
entry              42 tools/perf/util/strlist.h static inline bool strlist__has_entry(struct strlist *slist, const char *entry)
entry              44 tools/perf/util/strlist.h 	return strlist__find(slist, entry) != NULL;
entry             902 tools/perf/util/synthetic-events.c 		struct perf_record_thread_map_entry *entry = &event->thread_map.entries[i];
entry             908 tools/perf/util/synthetic-events.c 		entry->pid = perf_thread_map__pid(threads, i);
entry             909 tools/perf/util/synthetic-events.c 		strncpy((char *) &entry->comm, comm, sizeof(entry->comm));
entry              47 tools/perf/util/syscalltbl.c 	const struct syscall *entry = ventry;
entry              49 tools/perf/util/syscalltbl.c 	return strcmp(key, entry->name);
entry             190 tools/perf/util/unwind-libdw.c 	return entry(pc, ui) || !(--ui->max_stack) ?
entry             698 tools/perf/util/unwind-libunwind-local.c 		ret = ips[j] ? entry(ips[j], ui->thread, cb, arg) : 0;
entry              20 tools/perf/util/unwind.h typedef int (*unwind_entry_cb_t)(struct unwind_entry *entry, void *arg);
entry             226 tools/testing/radix-tree/idr-test.c 	void *entry;
entry             230 tools/testing/radix-tree/idr-test.c 		idr_for_each_entry(idr, entry, id);
entry             236 tools/testing/radix-tree/idr-test.c 		idr_for_each_entry(idr, entry, id);
entry             242 tools/testing/radix-tree/idr-test.c 		idr_for_each_entry(idr, entry, id);
entry             248 tools/testing/radix-tree/idr-test.c 		idr_for_each_entry(idr, entry, id);
entry             255 tools/testing/radix-tree/idr-test.c 		idr_for_each_entry(idr, entry, id);
entry             257 tools/testing/radix-tree/idr-test.c 		idr_for_each_entry(idr, entry, id);
entry             264 tools/testing/radix-tree/idr-test.c 		idr_for_each_entry(idr, entry, id);
entry             266 tools/testing/radix-tree/idr-test.c 		idr_for_each_entry(idr, entry, id);
entry             275 tools/testing/radix-tree/idr-test.c 		idr_for_each_entry(idr, entry, id);
entry             277 tools/testing/radix-tree/idr-test.c 		idr_for_each_entry(idr, entry, id);
entry             311 tools/testing/radix-tree/idr-test.c 		void *entry = idr_get_next(&find_idr, &id);
entry             312 tools/testing/radix-tree/idr-test.c 		BUG_ON(entry != xa_mk_value(id));
entry              70 tools/testing/radix-tree/iteration_check.c 	void *entry;
entry              77 tools/testing/radix-tree/iteration_check.c 		xas_for_each_marked(&xas, entry, ULONG_MAX, TAG) {
entry              78 tools/testing/radix-tree/iteration_check.c 			if (xas_retry(&xas, entry))
entry             103 tools/testing/radix-tree/iteration_check.c 	void *entry;
entry             110 tools/testing/radix-tree/iteration_check.c 		xas_for_each(&xas, entry, ULONG_MAX) {
entry             111 tools/testing/radix-tree/iteration_check.c 			if (xas_retry(&xas, entry))
entry              16 tools/testing/radix-tree/iteration_check_2.c 	void *entry;
entry              23 tools/testing/radix-tree/iteration_check_2.c 		xas_for_each_marked(&xas, entry, ULONG_MAX, XA_MARK_0)
entry              21 tools/testing/radix-tree/regression4.c 	void *entry;
entry              28 tools/testing/radix-tree/regression4.c 		entry = radix_tree_lookup(&mt_tree, 0);
entry              30 tools/testing/radix-tree/regression4.c 		if (entry != &obj0) {
entry              31 tools/testing/radix-tree/regression4.c 			printf("iteration %d bad entry = %p\n", i, entry);
entry             258 tools/testing/radix-tree/test.c 	void *entry;
entry             260 tools/testing/radix-tree/test.c 	xas_for_each(&xas, entry, ULONG_MAX) {
entry             261 tools/testing/radix-tree/test.c 		if (!xa_is_value(entry)) {
entry             262 tools/testing/radix-tree/test.c 			item_free(entry, xas.xa_index);
entry              54 tools/testing/selftests/bpf/test_hashmap.c 	struct hashmap_entry *entry, *tmp;
entry             101 tools/testing/selftests/bpf/test_hashmap.c 	hashmap__for_each_entry(map, entry, bkt) {
entry             102 tools/testing/selftests/bpf/test_hashmap.c 		long k = (long)entry->key;
entry             103 tools/testing/selftests/bpf/test_hashmap.c 		long v = (long)entry->value;
entry             144 tools/testing/selftests/bpf/test_hashmap.c 	hashmap__for_each_entry_safe(map, entry, tmp, bkt) {
entry             145 tools/testing/selftests/bpf/test_hashmap.c 		long k = (long)entry->key;
entry             146 tools/testing/selftests/bpf/test_hashmap.c 		long v = (long)entry->value;
entry             158 tools/testing/selftests/bpf/test_hashmap.c 	hashmap__for_each_key_entry(map, entry, (void *)0) {
entry             166 tools/testing/selftests/bpf/test_hashmap.c 	hashmap__for_each_key_entry_safe(map, entry, tmp, (void *)0) {
entry             170 tools/testing/selftests/bpf/test_hashmap.c 		k = entry->key;
entry             171 tools/testing/selftests/bpf/test_hashmap.c 		v = entry->value;
entry             201 tools/testing/selftests/bpf/test_hashmap.c 	hashmap__for_each_entry_safe(map, entry, tmp, bkt) {
entry             205 tools/testing/selftests/bpf/test_hashmap.c 		k = entry->key;
entry             206 tools/testing/selftests/bpf/test_hashmap.c 		v = entry->value;
entry             235 tools/testing/selftests/bpf/test_hashmap.c 	hashmap__for_each_entry(map, entry, bkt) {
entry             237 tools/testing/selftests/bpf/test_hashmap.c 			     (long)entry->key, (long)entry->value);
entry             242 tools/testing/selftests/bpf/test_hashmap.c 	hashmap__for_each_entry(map, entry, bkt) {
entry             244 tools/testing/selftests/bpf/test_hashmap.c 			     (long)entry->key, (long)entry->value);
entry             260 tools/testing/selftests/bpf/test_hashmap.c 	struct hashmap_entry *entry;
entry             303 tools/testing/selftests/bpf/test_hashmap.c 	hashmap__for_each_entry(map, entry, bkt) {
entry             304 tools/testing/selftests/bpf/test_hashmap.c 		found_msk |= (long)entry->value;
entry             312 tools/testing/selftests/bpf/test_hashmap.c 	hashmap__for_each_key_entry(map, entry, k1) {
entry             313 tools/testing/selftests/bpf/test_hashmap.c 		found_msk |= (long)entry->value;
entry             321 tools/testing/selftests/bpf/test_hashmap.c 	hashmap__for_each_key_entry(map, entry, k2) {
entry             322 tools/testing/selftests/bpf/test_hashmap.c 		found_msk |= (long)entry->value;
entry             334 tools/testing/selftests/bpf/test_hashmap.c 	struct hashmap_entry *entry;
entry             357 tools/testing/selftests/bpf/test_hashmap.c 	hashmap__for_each_entry(map, entry, bkt) {
entry             361 tools/testing/selftests/bpf/test_hashmap.c 	hashmap__for_each_key_entry(map, entry, k) {
entry              60 tools/testing/selftests/kvm/lib/aarch64/processor.c static uint64_t pte_addr(struct kvm_vm *vm, uint64_t entry)
entry              63 tools/testing/selftests/kvm/lib/aarch64/processor.c 	return entry & mask;
entry              74 tools/testing/selftests/kvm/lib/s390x/processor.c 	uint64_t *entry;
entry              94 tools/testing/selftests/kvm/lib/s390x/processor.c 	entry = addr_gpa2hva(vm, vm->pgd);
entry              97 tools/testing/selftests/kvm/lib/s390x/processor.c 		if (entry[idx] & REGION_ENTRY_INVALID)
entry              98 tools/testing/selftests/kvm/lib/s390x/processor.c 			entry[idx] = virt_alloc_region(vm, ri, memslot);
entry              99 tools/testing/selftests/kvm/lib/s390x/processor.c 		entry = addr_gpa2hva(vm, entry[idx] & REGION_ENTRY_ORIGIN);
entry             104 tools/testing/selftests/kvm/lib/s390x/processor.c 	if (!(entry[idx] & PAGE_INVALID))
entry             107 tools/testing/selftests/kvm/lib/s390x/processor.c 	entry[idx] = gpa;
entry             133 tools/testing/selftests/kvm/lib/s390x/processor.c 	uint64_t *entry;
entry             138 tools/testing/selftests/kvm/lib/s390x/processor.c 	entry = addr_gpa2hva(vm, vm->pgd);
entry             141 tools/testing/selftests/kvm/lib/s390x/processor.c 		TEST_ASSERT(!(entry[idx] & REGION_ENTRY_INVALID),
entry             144 tools/testing/selftests/kvm/lib/s390x/processor.c 		entry = addr_gpa2hva(vm, entry[idx] & REGION_ENTRY_ORIGIN);
entry             149 tools/testing/selftests/kvm/lib/s390x/processor.c 	TEST_ASSERT(!(entry[idx] & PAGE_INVALID),
entry             152 tools/testing/selftests/kvm/lib/s390x/processor.c 	return (entry[idx] & ~0xffful) + (gva & 0xffful);
entry             172 tools/testing/selftests/kvm/lib/s390x/processor.c 	uint64_t addr, *entry;
entry             175 tools/testing/selftests/kvm/lib/s390x/processor.c 		entry = addr_gpa2hva(vm, addr);
entry             176 tools/testing/selftests/kvm/lib/s390x/processor.c 		if (*entry & REGION_ENTRY_INVALID)
entry             179 tools/testing/selftests/kvm/lib/s390x/processor.c 			indent, "", 4 - ((*entry & REGION_ENTRY_TYPE) >> 2),
entry             180 tools/testing/selftests/kvm/lib/s390x/processor.c 			addr, *entry);
entry             181 tools/testing/selftests/kvm/lib/s390x/processor.c 		if (*entry & REGION_ENTRY_TYPE) {
entry             183 tools/testing/selftests/kvm/lib/s390x/processor.c 					 *entry & REGION_ENTRY_ORIGIN);
entry             186 tools/testing/selftests/kvm/lib/s390x/processor.c 				       *entry & REGION_ENTRY_ORIGIN);
entry             752 tools/testing/selftests/kvm/lib/x86_64/processor.c 	struct kvm_cpuid_entry2 *entry = NULL;
entry             759 tools/testing/selftests/kvm/lib/x86_64/processor.c 			entry = &cpuid->entries[i];
entry             764 tools/testing/selftests/kvm/lib/x86_64/processor.c 	TEST_ASSERT(entry, "Guest CPUID entry not found: (EAX=%x, ECX=%x).",
entry             766 tools/testing/selftests/kvm/lib/x86_64/processor.c 	return entry;
entry             858 tools/testing/selftests/kvm/lib/x86_64/processor.c 		struct kvm_msr_entry entry;
entry             864 tools/testing/selftests/kvm/lib/x86_64/processor.c 	buffer.entry.index = msr_index;
entry             869 tools/testing/selftests/kvm/lib/x86_64/processor.c 	return buffer.entry.data;
entry             892 tools/testing/selftests/kvm/lib/x86_64/processor.c 		struct kvm_msr_entry entry;
entry             899 tools/testing/selftests/kvm/lib/x86_64/processor.c 	buffer.entry.index = msr_index;
entry             900 tools/testing/selftests/kvm/lib/x86_64/processor.c 	buffer.entry.data = msr_value;
entry            1168 tools/testing/selftests/kvm/lib/x86_64/processor.c 	struct kvm_cpuid_entry2 *entry;
entry            1177 tools/testing/selftests/kvm/lib/x86_64/processor.c 		entry = kvm_get_supported_cpuid_entry(0x80000008);
entry            1178 tools/testing/selftests/kvm/lib/x86_64/processor.c 		*pa_bits = entry->eax & 0xff;
entry            1179 tools/testing/selftests/kvm/lib/x86_64/processor.c 		*va_bits = (entry->eax >> 8) & 0xff;
entry             384 tools/testing/selftests/kvm/lib/x86_64/vmx.c 	struct kvm_cpuid_entry2 *entry = kvm_get_supported_cpuid_entry(1);
entry             386 tools/testing/selftests/kvm/lib/x86_64/vmx.c 	if (!(entry->ecx & CPUID_VMX)) {
entry              69 tools/testing/selftests/kvm/x86_64/cr4_cpuid_sync_test.c 	struct kvm_cpuid_entry2 *entry;
entry              73 tools/testing/selftests/kvm/x86_64/cr4_cpuid_sync_test.c 	entry = kvm_get_supported_cpuid_entry(1);
entry              74 tools/testing/selftests/kvm/x86_64/cr4_cpuid_sync_test.c 	if (!(entry->ecx & X86_FEATURE_XSAVE)) {
entry              65 tools/testing/selftests/kvm/x86_64/hyperv_cpuid.c 		struct kvm_cpuid_entry2 *entry = &hv_cpuid_entries->entries[i];
entry              67 tools/testing/selftests/kvm/x86_64/hyperv_cpuid.c 		TEST_ASSERT((entry->function >= 0x40000000) &&
entry              68 tools/testing/selftests/kvm/x86_64/hyperv_cpuid.c 			    (entry->function <= 0x4000000A),
entry              70 tools/testing/selftests/kvm/x86_64/hyperv_cpuid.c 			    entry->function);
entry              72 tools/testing/selftests/kvm/x86_64/hyperv_cpuid.c 		TEST_ASSERT(entry->index == 0,
entry              75 tools/testing/selftests/kvm/x86_64/hyperv_cpuid.c 		TEST_ASSERT(entry->flags == 0,
entry              78 tools/testing/selftests/kvm/x86_64/hyperv_cpuid.c 		TEST_ASSERT(!entry->padding[0] && !entry->padding[1] &&
entry              79 tools/testing/selftests/kvm/x86_64/hyperv_cpuid.c 			    !entry->padding[2], "padding should be zero");
entry              81 tools/testing/selftests/kvm/x86_64/hyperv_cpuid.c 		if (entry->function == 0x40000004) {
entry              82 tools/testing/selftests/kvm/x86_64/hyperv_cpuid.c 			int nononarchcs = !!(entry->eax & (1UL << 18));
entry              57 tools/testing/selftests/kvm/x86_64/vmx_tsc_adjust_test.c 	struct kvm_msr_entry entry;
entry             264 tools/testing/selftests/networking/timestamping/timestamping.c 	struct iovec entry;
entry             273 tools/testing/selftests/networking/timestamping/timestamping.c 	msg.msg_iov = &entry;
entry             275 tools/testing/selftests/networking/timestamping/timestamping.c 	entry.iov_base = data;
entry             276 tools/testing/selftests/networking/timestamping/timestamping.c 	entry.iov_len = sizeof(data);
entry             271 tools/testing/selftests/networking/timestamping/txtimestamp.c 	struct iovec entry;
entry             280 tools/testing/selftests/networking/timestamping/txtimestamp.c 	memset(&entry, 0, sizeof(entry));
entry             283 tools/testing/selftests/networking/timestamping/txtimestamp.c 	entry.iov_base = data;
entry             284 tools/testing/selftests/networking/timestamping/txtimestamp.c 	entry.iov_len = cfg_payload_len;
entry             285 tools/testing/selftests/networking/timestamping/txtimestamp.c 	msg.msg_iov = &entry;
entry             133 tools/testing/selftests/powerpc/pmu/ebb/ebb.c 	u64 entry;
entry             136 tools/testing/selftests/powerpc/pmu/ebb/ebb.c 	entry = (u64)ebb_handler;
entry             140 tools/testing/selftests/powerpc/pmu/ebb/ebb.c 	    u64 entry;
entry             145 tools/testing/selftests/powerpc/pmu/ebb/ebb.c 	entry = opd->entry;
entry             147 tools/testing/selftests/powerpc/pmu/ebb/ebb.c 	printf("EBB Handler is at %#llx\n", entry);
entry             153 tools/testing/selftests/powerpc/pmu/ebb/ebb.c 	mtspr(SPRN_EBBHR, entry);
entry             266 tools/testing/selftests/powerpc/primitives/asm/ppc_asm.h #define _ASM_NOKPROBE_SYMBOL(entry)			\
entry             268 tools/testing/selftests/powerpc/primitives/asm/ppc_asm.h 	PPC_LONG (entry) ;				\
entry             271 tools/testing/selftests/powerpc/primitives/asm/ppc_asm.h #define _ASM_NOKPROBE_SYMBOL(entry)
entry              74 tools/testing/selftests/powerpc/primitives/load_unaligned_zeropad.c 	struct extbl_entry *entry = (struct extbl_entry *)__start___ex_table;
entry              76 tools/testing/selftests/powerpc/primitives/load_unaligned_zeropad.c 	while (entry < (struct extbl_entry *)__stop___ex_table) {
entry              79 tools/testing/selftests/powerpc/primitives/load_unaligned_zeropad.c 		insn  = (unsigned long)&entry->insn + entry->insn;
entry              80 tools/testing/selftests/powerpc/primitives/load_unaligned_zeropad.c 		fixup = (unsigned long)&entry->fixup + entry->fixup;
entry             188 tools/testing/selftests/powerpc/vphn/asm/lppaca.h extern void (*dtl_consumer)(struct dtl_entry *entry, u64 index);
entry             118 tools/testing/selftests/ptrace/get_syscall_info.c 			(void *) &info.entry - (void *) &info;
entry             120 tools/testing/selftests/ptrace/get_syscall_info.c 			(void *) &info.entry.args[6] - (void *) &info;
entry             203 tools/testing/selftests/ptrace/get_syscall_info.c 				ASSERT_EQ(exp_args[0], info.entry.nr) {
entry             206 tools/testing/selftests/ptrace/get_syscall_info.c 				ASSERT_EQ(exp_args[1], info.entry.args[0]) {
entry             209 tools/testing/selftests/ptrace/get_syscall_info.c 				ASSERT_EQ(exp_args[2], info.entry.args[1]) {
entry             212 tools/testing/selftests/ptrace/get_syscall_info.c 				ASSERT_EQ(exp_args[3], info.entry.args[2]) {
entry             215 tools/testing/selftests/ptrace/get_syscall_info.c 				ASSERT_EQ(exp_args[4], info.entry.args[3]) {
entry             218 tools/testing/selftests/ptrace/get_syscall_info.c 				ASSERT_EQ(exp_args[5], info.entry.args[4]) {
entry             221 tools/testing/selftests/ptrace/get_syscall_info.c 				ASSERT_EQ(exp_args[6], info.entry.args[5]) {
entry              23 tools/testing/selftests/rcutorture/formal/srcu-cbmc/src/workqueues.h 	struct list_head entry;
entry              31 tools/testing/selftests/rcutorture/formal/srcu-cbmc/src/workqueues.h 	struct hlist_node	entry;
entry              86 tools/testing/selftests/rcutorture/formal/srcu-cbmc/src/workqueues.h 		.entry = { &(n).entry, &(n).entry }, \
entry            1782 tools/testing/selftests/seccomp/seccomp_bpf.c 	static bool entry;
entry            1788 tools/testing/selftests/seccomp/seccomp_bpf.c 	entry = !entry;
entry            1793 tools/testing/selftests/seccomp/seccomp_bpf.c 	EXPECT_EQ(entry ? PTRACE_EVENTMSG_SYSCALL_ENTRY
entry            1796 tools/testing/selftests/seccomp/seccomp_bpf.c 	if (!entry)
entry             163 tools/testing/selftests/vm/compaction_test.c 	struct map_list *list, *entry;
entry             201 tools/testing/selftests/vm/compaction_test.c 		entry = malloc(sizeof(struct map_list));
entry             202 tools/testing/selftests/vm/compaction_test.c 		if (!entry) {
entry             206 tools/testing/selftests/vm/compaction_test.c 		entry->map = map;
entry             207 tools/testing/selftests/vm/compaction_test.c 		entry->next = list;
entry             208 tools/testing/selftests/vm/compaction_test.c 		list = entry;
entry             219 tools/testing/selftests/vm/compaction_test.c 	for (entry = list; entry != NULL; entry = entry->next) {
entry             220 tools/testing/selftests/vm/compaction_test.c 		munmap(entry->map, MAP_SIZE);
entry             221 tools/testing/selftests/vm/compaction_test.c 		if (!entry->next)
entry             223 tools/testing/selftests/vm/compaction_test.c 		entry = entry->next;
entry             237 tools/usb/testusb.c 	struct testdev			*entry;
entry             255 tools/usb/testusb.c 	entry = calloc(1, sizeof *entry);
entry             256 tools/usb/testusb.c 	if (!entry)
entry             259 tools/usb/testusb.c 	entry->name = strdup(name);
entry             260 tools/usb/testusb.c 	if (!entry->name) {
entry             261 tools/usb/testusb.c 		free(entry);
entry             267 tools/usb/testusb.c 	entry->ifnum = ifnum;
entry             272 tools/usb/testusb.c 		speed(entry->speed), entry->name, entry->ifnum);
entry             274 tools/usb/testusb.c 	entry->next = testdevs;
entry             275 tools/usb/testusb.c 	testdevs = entry;
entry             367 tools/usb/testusb.c 	struct testdev		*entry;
entry             485 tools/usb/testusb.c 	for (entry = testdevs; entry; entry = entry->next) {
entry             488 tools/usb/testusb.c 		entry->param = param;
entry             489 tools/usb/testusb.c 		entry->forever = forever;
entry             490 tools/usb/testusb.c 		entry->test = test;
entry             493 tools/usb/testusb.c 			if (strcmp (entry->name, device))
entry             495 tools/usb/testusb.c 			return handle_testdev (entry) != entry;
entry             497 tools/usb/testusb.c 		status = pthread_create (&entry->thread, 0, handle_testdev, entry);
entry             517 tools/usb/testusb.c 	for (entry = testdevs; entry; entry = entry->next) {
entry             520 tools/usb/testusb.c 		if (pthread_join (entry->thread, &retval))
entry              86 tools/usb/usbip/libsrc/list.h static inline void __list_del_entry(struct list_head *entry)
entry              88 tools/usb/usbip/libsrc/list.h 	__list_del(entry->prev, entry->next);
entry              91 tools/usb/usbip/libsrc/list.h static inline void list_del(struct list_head *entry)
entry              93 tools/usb/usbip/libsrc/list.h 	__list_del(entry->prev, entry->next);
entry              94 tools/usb/usbip/libsrc/list.h 	entry->next = LIST_POISON1;
entry              95 tools/usb/usbip/libsrc/list.h 	entry->prev = LIST_POISON2;
entry             142 virt/kvm/arm/vgic/vgic-its.c 	struct list_head	entry;
entry             544 virt/kvm/arm/vgic/vgic-its.c 	list_for_each_entry(cte, &dist->lpi_translation_cache, entry) {
entry             560 virt/kvm/arm/vgic/vgic-its.c 		if (!list_is_first(&cte->entry, &dist->lpi_translation_cache))
entry             561 virt/kvm/arm/vgic/vgic-its.c 			list_move(&cte->entry, &dist->lpi_translation_cache);
entry             612 virt/kvm/arm/vgic/vgic-its.c 			      typeof(*cte), entry);
entry             630 virt/kvm/arm/vgic/vgic-its.c 	list_move(&cte->entry, &dist->lpi_translation_cache);
entry             644 virt/kvm/arm/vgic/vgic-its.c 	list_for_each_entry(cte, &dist->lpi_translation_cache, entry) {
entry            1846 virt/kvm/arm/vgic/vgic-its.c 		INIT_LIST_HEAD(&cte->entry);
entry            1847 virt/kvm/arm/vgic/vgic-its.c 		list_add(&cte->entry, &dist->lpi_translation_cache);
entry            1859 virt/kvm/arm/vgic/vgic-its.c 				 &dist->lpi_translation_cache, entry) {
entry            1860 virt/kvm/arm/vgic/vgic-its.c 		list_del(&cte->entry);
entry            2061 virt/kvm/arm/vgic/vgic-its.c typedef int (*entry_fn_t)(struct vgic_its *its, u32 id, void *entry,
entry            2086 virt/kvm/arm/vgic/vgic-its.c 	char entry[ESZ_MAX];
entry            2089 virt/kvm/arm/vgic/vgic-its.c 	memset(entry, 0, esz);
entry            2095 virt/kvm/arm/vgic/vgic-its.c 		ret = kvm_read_guest_lock(kvm, gpa, entry, esz);
entry            2099 virt/kvm/arm/vgic/vgic-its.c 		next_offset = fn(its, id, entry, opaque);
entry            2293 virt/kvm/arm/vgic/vgic-its.c 	u64 entry = *(u64 *)ptr;
entry            2298 virt/kvm/arm/vgic/vgic-its.c 	entry = le64_to_cpu(entry);
entry            2300 virt/kvm/arm/vgic/vgic-its.c 	valid = entry >> KVM_ITS_DTE_VALID_SHIFT;
entry            2301 virt/kvm/arm/vgic/vgic-its.c 	num_eventid_bits = (entry & KVM_ITS_DTE_SIZE_MASK) + 1;
entry            2302 virt/kvm/arm/vgic/vgic-its.c 	itt_addr = ((entry & KVM_ITS_DTE_ITTADDR_MASK)
entry            2309 virt/kvm/arm/vgic/vgic-its.c 	offset = (entry & KVM_ITS_DTE_NEXT_MASK) >> KVM_ITS_DTE_NEXT_SHIFT;
entry            2391 virt/kvm/arm/vgic/vgic-its.c 	u64 entry = *(u64 *)addr;
entry            2396 virt/kvm/arm/vgic/vgic-its.c 	entry = le64_to_cpu(entry);
entry            2398 virt/kvm/arm/vgic/vgic-its.c 	if (!(entry & KVM_ITS_L1E_VALID_MASK))
entry            2401 virt/kvm/arm/vgic/vgic-its.c 	gpa = entry & KVM_ITS_L1E_ADDR_MASK;
entry            1782 virt/kvm/kvm_main.c 	gfn_t entry = 0;
entry            1784 virt/kvm/kvm_main.c 	addr = gfn_to_hva_many(slot, gfn, &entry);
entry            1788 virt/kvm/kvm_main.c 	if (entry < nr_pages)