walker            222 arch/x86/kvm/paging_tmpl.h 					     struct guest_walker *walker,
walker            235 arch/x86/kvm/paging_tmpl.h 	for (level = walker->max_level; level >= walker->level; --level) {
walker            236 arch/x86/kvm/paging_tmpl.h 		pte = orig_pte = walker->ptes[level - 1];
walker            237 arch/x86/kvm/paging_tmpl.h 		table_gfn = walker->table_gfn[level - 1];
walker            238 arch/x86/kvm/paging_tmpl.h 		ptep_user = walker->ptep_user[level - 1];
walker            244 arch/x86/kvm/paging_tmpl.h 		if (level == walker->level && write_fault &&
walker            269 arch/x86/kvm/paging_tmpl.h 		if (unlikely(!walker->pte_writable[level - 1]))
walker            277 arch/x86/kvm/paging_tmpl.h 		walker->ptes[level - 1] = pte;
walker            296 arch/x86/kvm/paging_tmpl.h static int FNAME(walk_addr_generic)(struct guest_walker *walker,
walker            320 arch/x86/kvm/paging_tmpl.h 	walker->level = mmu->root_level;
walker            326 arch/x86/kvm/paging_tmpl.h 	if (walker->level == PT32E_ROOT_LEVEL) {
walker            328 arch/x86/kvm/paging_tmpl.h 		trace_kvm_mmu_paging_element(pte, walker->level);
walker            331 arch/x86/kvm/paging_tmpl.h 		--walker->level;
walker            334 arch/x86/kvm/paging_tmpl.h 	walker->max_level = walker->level;
walker            345 arch/x86/kvm/paging_tmpl.h 	++walker->level;
walker            352 arch/x86/kvm/paging_tmpl.h 		--walker->level;
walker            354 arch/x86/kvm/paging_tmpl.h 		index = PT_INDEX(addr, walker->level);
walker            359 arch/x86/kvm/paging_tmpl.h 		BUG_ON(walker->level < 1);
walker            360 arch/x86/kvm/paging_tmpl.h 		walker->table_gfn[walker->level - 1] = table_gfn;
walker            361 arch/x86/kvm/paging_tmpl.h 		walker->pte_gpa[walker->level - 1] = pte_gpa;
walker            365 arch/x86/kvm/paging_tmpl.h 					      &walker->fault);
walker            383 arch/x86/kvm/paging_tmpl.h 					    &walker->pte_writable[walker->level - 1]);
walker            390 arch/x86/kvm/paging_tmpl.h 		walker->ptep_user[walker->level - 1] = ptep_user;
walker            392 arch/x86/kvm/paging_tmpl.h 		trace_kvm_mmu_paging_element(pte, walker->level);
walker            403 arch/x86/kvm/paging_tmpl.h 		if (unlikely(is_rsvd_bits_set(mmu, pte, walker->level))) {
walker            408 arch/x86/kvm/paging_tmpl.h 		walker->ptes[walker->level - 1] = pte;
walker            409 arch/x86/kvm/paging_tmpl.h 	} while (!is_last_gpte(mmu, walker->level, pte));
walker            415 arch/x86/kvm/paging_tmpl.h 	walker->pt_access = FNAME(gpte_access)(pt_access ^ walk_nx_mask);
walker            416 arch/x86/kvm/paging_tmpl.h 	walker->pte_access = FNAME(gpte_access)(pte_access ^ walk_nx_mask);
walker            417 arch/x86/kvm/paging_tmpl.h 	errcode = permission_fault(vcpu, mmu, walker->pte_access, pte_pkey, access);
walker            421 arch/x86/kvm/paging_tmpl.h 	gfn = gpte_to_gfn_lvl(pte, walker->level);
walker            422 arch/x86/kvm/paging_tmpl.h 	gfn += (addr & PT_LVL_OFFSET_MASK(walker->level)) >> PAGE_SHIFT;
walker            424 arch/x86/kvm/paging_tmpl.h 	if (PTTYPE == 32 && walker->level == PT_DIRECTORY_LEVEL && is_cpuid_PSE36())
walker            427 arch/x86/kvm/paging_tmpl.h 	real_gpa = mmu->translate_gpa(vcpu, gfn_to_gpa(gfn), access, &walker->fault);
walker            431 arch/x86/kvm/paging_tmpl.h 	walker->gfn = real_gpa >> PAGE_SHIFT;
walker            434 arch/x86/kvm/paging_tmpl.h 		FNAME(protect_clean_gpte)(mmu, &walker->pte_access, pte);
walker            445 arch/x86/kvm/paging_tmpl.h 		ret = FNAME(update_accessed_dirty_bits)(vcpu, mmu, walker, write_fault);
walker            453 arch/x86/kvm/paging_tmpl.h 		 __func__, (u64)pte, walker->pte_access, walker->pt_access);
walker            462 arch/x86/kvm/paging_tmpl.h 	walker->fault.vector = PF_VECTOR;
walker            463 arch/x86/kvm/paging_tmpl.h 	walker->fault.error_code_valid = true;
walker            464 arch/x86/kvm/paging_tmpl.h 	walker->fault.error_code = errcode;
walker            491 arch/x86/kvm/paging_tmpl.h 	walker->fault.address = addr;
walker            492 arch/x86/kvm/paging_tmpl.h 	walker->fault.nested_page_fault = mmu != vcpu->arch.walk_mmu;
walker            494 arch/x86/kvm/paging_tmpl.h 	trace_kvm_mmu_walker_error(walker->fault.error_code);
walker            498 arch/x86/kvm/paging_tmpl.h static int FNAME(walk_addr)(struct guest_walker *walker,
walker            501 arch/x86/kvm/paging_tmpl.h 	return FNAME(walk_addr_generic)(walker, vcpu, vcpu->arch.mmu, addr,
walker            506 arch/x86/kvm/paging_tmpl.h static int FNAME(walk_addr_nested)(struct guest_walker *walker,
walker            510 arch/x86/kvm/paging_tmpl.h 	return FNAME(walk_addr_generic)(walker, vcpu, &vcpu->arch.nested_mmu,
walker            733 arch/x86/kvm/paging_tmpl.h 			      struct guest_walker *walker, int user_fault,
walker            737 arch/x86/kvm/paging_tmpl.h 	gfn_t mask = ~(KVM_PAGES_PER_HPAGE(walker->level) - 1);
walker            740 arch/x86/kvm/paging_tmpl.h 	if (!(walker->pte_access & ACC_WRITE_MASK ||
walker            744 arch/x86/kvm/paging_tmpl.h 	for (level = walker->level; level <= walker->max_level; level++) {
walker            745 arch/x86/kvm/paging_tmpl.h 		gfn_t gfn = walker->gfn ^ walker->table_gfn[level - 1];
walker            773 arch/x86/kvm/paging_tmpl.h 	struct guest_walker walker;
walker            798 arch/x86/kvm/paging_tmpl.h 	r = FNAME(walk_addr)(&walker, vcpu, addr, error_code);
walker            806 arch/x86/kvm/paging_tmpl.h 			inject_page_fault(vcpu, &walker.fault);
walker            811 arch/x86/kvm/paging_tmpl.h 	if (page_fault_handle_page_track(vcpu, error_code, walker.gfn)) {
walker            819 arch/x86/kvm/paging_tmpl.h 	      &walker, user_fault, &vcpu->arch.write_fault_to_shadow_pgtable);
walker            821 arch/x86/kvm/paging_tmpl.h 	if (walker.level >= PT_DIRECTORY_LEVEL && !is_self_change_mapping) {
walker            822 arch/x86/kvm/paging_tmpl.h 		level = mapping_level(vcpu, walker.gfn, &force_pt_level);
walker            824 arch/x86/kvm/paging_tmpl.h 			level = min(walker.level, level);
walker            825 arch/x86/kvm/paging_tmpl.h 			walker.gfn = walker.gfn & ~(KVM_PAGES_PER_HPAGE(level) - 1);
walker            833 arch/x86/kvm/paging_tmpl.h 	if (try_async_pf(vcpu, prefault, walker.gfn, addr, &pfn, write_fault,
walker            837 arch/x86/kvm/paging_tmpl.h 	if (handle_abnormal_pfn(vcpu, addr, walker.gfn, pfn, walker.pte_access, &r))
walker            844 arch/x86/kvm/paging_tmpl.h 	if (write_fault && !(walker.pte_access & ACC_WRITE_MASK) &&
walker            847 arch/x86/kvm/paging_tmpl.h 		walker.pte_access |= ACC_WRITE_MASK;
walker            848 arch/x86/kvm/paging_tmpl.h 		walker.pte_access &= ~ACC_USER_MASK;
walker            857 arch/x86/kvm/paging_tmpl.h 			walker.pte_access &= ~ACC_EXEC_MASK;
walker            869 arch/x86/kvm/paging_tmpl.h 		transparent_hugepage_adjust(vcpu, walker.gfn, &pfn, &level);
walker            870 arch/x86/kvm/paging_tmpl.h 	r = FNAME(fetch)(vcpu, addr, &walker, write_fault,
walker            952 arch/x86/kvm/paging_tmpl.h 	struct guest_walker walker;
walker            956 arch/x86/kvm/paging_tmpl.h 	r = FNAME(walk_addr)(&walker, vcpu, addr, access);
walker            959 arch/x86/kvm/paging_tmpl.h 		gpa = gfn_to_gpa(walker.gfn);
walker            962 arch/x86/kvm/paging_tmpl.h 		*exception = walker.fault;
walker            973 arch/x86/kvm/paging_tmpl.h 	struct guest_walker walker;
walker            982 arch/x86/kvm/paging_tmpl.h 	r = FNAME(walk_addr_nested)(&walker, vcpu, vaddr, access);
walker            985 arch/x86/kvm/paging_tmpl.h 		gpa = gfn_to_gpa(walker.gfn);
walker            988 arch/x86/kvm/paging_tmpl.h 		*exception = walker.fault;
walker            150 drivers/pci/controller/dwc/pcie-hisi.c 	void *walker = &reg_val;
walker            153 drivers/pci/controller/dwc/pcie-hisi.c 	walker += (where & 0x3);
walker            158 drivers/pci/controller/dwc/pcie-hisi.c 		*val = *(u8 __force *) walker;
walker            160 drivers/pci/controller/dwc/pcie-hisi.c 		*val = *(u16 __force *) walker;
walker            175 drivers/pci/controller/dwc/pcie-hisi.c 	void *walker = &reg_val;
walker            178 drivers/pci/controller/dwc/pcie-hisi.c 	walker += (where & 0x3);
walker            184 drivers/pci/controller/dwc/pcie-hisi.c 		*(u16 __force *) walker = val;
walker            188 drivers/pci/controller/dwc/pcie-hisi.c 		*(u8 __force *) walker = val;
walker            489 fs/ecryptfs/keystore.c 	struct ecryptfs_global_auth_tok *walker;
walker            495 fs/ecryptfs/keystore.c 	list_for_each_entry(walker,
walker            498 fs/ecryptfs/keystore.c 		if (memcmp(walker->sig, sig, ECRYPTFS_SIG_SIZE_HEX))
walker            501 fs/ecryptfs/keystore.c 		if (walker->flags & ECRYPTFS_AUTH_TOK_INVALID) {
walker            506 fs/ecryptfs/keystore.c 		rc = key_validate(walker->global_auth_tok_key);
walker            513 fs/ecryptfs/keystore.c 		down_write(&(walker->global_auth_tok_key->sem));
walker            515 fs/ecryptfs/keystore.c 				walker->global_auth_tok_key, auth_tok);
walker            519 fs/ecryptfs/keystore.c 		(*auth_tok_key) = walker->global_auth_tok_key;
walker            526 fs/ecryptfs/keystore.c 	up_write(&(walker->global_auth_tok_key->sem));
walker            529 fs/ecryptfs/keystore.c 	walker->flags |= ECRYPTFS_AUTH_TOK_INVALID;
walker            530 fs/ecryptfs/keystore.c 	key_put(walker->global_auth_tok_key);
walker            531 fs/ecryptfs/keystore.c 	walker->global_auth_tok_key = NULL;
walker            137 fs/ecryptfs/super.c 	struct ecryptfs_global_auth_tok *walker;
walker            140 fs/ecryptfs/super.c 	list_for_each_entry(walker,
walker            143 fs/ecryptfs/super.c 		if (walker->flags & ECRYPTFS_AUTH_TOK_FNEK)
walker            144 fs/ecryptfs/super.c 			seq_printf(m, ",ecryptfs_fnek_sig=%s", walker->sig);
walker            146 fs/ecryptfs/super.c 			seq_printf(m, ",ecryptfs_sig=%s", walker->sig);
walker            471 fs/gfs2/bmap.c 		u64 max_len, gfs2_metadata_walker walker)
walker            501 fs/gfs2/bmap.c 		status = walker(mp, ptrs);
walker            124 include/linux/rhashtable-types.h 	struct rhashtable_walker walker;
walker            514 kernel/seccomp.c 	struct seccomp_filter *walker;
walker            520 kernel/seccomp.c 	for (walker = current->seccomp.filter; walker; walker = walker->prev)
walker            521 kernel/seccomp.c 		total_insns += walker->prog->len + 4;  /* 4 instr penalty */
walker            307 lib/rhashtable.c 	struct rhashtable_walker *walker;
walker            326 lib/rhashtable.c 	list_for_each_entry(walker, &old_tbl->walkers, list)
walker            327 lib/rhashtable.c 		walker->tbl = NULL;
walker            667 lib/rhashtable.c 	iter->walker.tbl =
walker            669 lib/rhashtable.c 	list_add(&iter->walker.list, &iter->walker.tbl->walkers);
walker            683 lib/rhashtable.c 	if (iter->walker.tbl)
walker            684 lib/rhashtable.c 		list_del(&iter->walker.list);
walker            716 lib/rhashtable.c 	if (iter->walker.tbl)
walker            717 lib/rhashtable.c 		list_del(&iter->walker.list);
walker            722 lib/rhashtable.c 	if (!iter->walker.tbl) {
walker            723 lib/rhashtable.c 		iter->walker.tbl = rht_dereference_rcu(ht->tbl, ht);
walker            736 lib/rhashtable.c 		rht_for_each_rcu(p, iter->walker.tbl, iter->slot) {
walker            751 lib/rhashtable.c 		rht_for_each_rcu(p, iter->walker.tbl, iter->slot) {
walker            782 lib/rhashtable.c 	struct bucket_table *tbl = iter->walker.tbl;
walker            828 lib/rhashtable.c 	iter->walker.tbl = rht_dereference_rcu(tbl->future_tbl, ht);
walker            829 lib/rhashtable.c 	if (iter->walker.tbl) {
walker            927 lib/rhashtable.c 	struct bucket_table *tbl = iter->walker.tbl;
walker            937 lib/rhashtable.c 		iter->walker.tbl = NULL;
walker            939 lib/rhashtable.c 		list_add(&iter->walker.list, &tbl->walkers);
walker            550 net/sched/cls_tcindex.c static void tcindex_walk(struct tcf_proto *tp, struct tcf_walker *walker,
walker            557 net/sched/cls_tcindex.c 	pr_debug("tcindex_walk(tp %p,walker %p),p %p\n", tp, walker, p);
walker            562 net/sched/cls_tcindex.c 			if (walker->count >= walker->skip) {
walker            563 net/sched/cls_tcindex.c 				if (walker->fn(tp, p->perfect + i, walker) < 0) {
walker            564 net/sched/cls_tcindex.c 					walker->stop = 1;
walker            568 net/sched/cls_tcindex.c 			walker->count++;
walker            576 net/sched/cls_tcindex.c 			if (walker->count >= walker->skip) {
walker            577 net/sched/cls_tcindex.c 				if (walker->fn(tp, &f->result, walker) < 0) {
walker            578 net/sched/cls_tcindex.c 					walker->stop = 1;
walker            582 net/sched/cls_tcindex.c 			walker->count++;
walker            347 net/sched/sch_atm.c static void atm_tc_walk(struct Qdisc *sch, struct qdisc_walker *walker)
walker            352 net/sched/sch_atm.c 	pr_debug("atm_tc_walk(sch %p,[qdisc %p],walker %p)\n", sch, p, walker);
walker            353 net/sched/sch_atm.c 	if (walker->stop)
walker            356 net/sched/sch_atm.c 		if (walker->count >= walker->skip &&
walker            357 net/sched/sch_atm.c 		    walker->fn(sch, (unsigned long)flow, walker) < 0) {
walker            358 net/sched/sch_atm.c 			walker->stop = 1;
walker            361 net/sched/sch_atm.c 		walker->count++;
walker            519 net/sched/sch_cbs.c static void cbs_walk(struct Qdisc *sch, struct qdisc_walker *walker)
walker            521 net/sched/sch_cbs.c 	if (!walker->stop) {
walker            522 net/sched/sch_cbs.c 		if (walker->count >= walker->skip) {
walker            523 net/sched/sch_cbs.c 			if (walker->fn(sch, 1, walker) < 0) {
walker            524 net/sched/sch_cbs.c 				walker->stop = 1;
walker            528 net/sched/sch_cbs.c 		walker->count++;
walker            166 net/sched/sch_dsmark.c static void dsmark_walk(struct Qdisc *sch, struct qdisc_walker *walker)
walker            172 net/sched/sch_dsmark.c 		 __func__, sch, p, walker);
walker            174 net/sched/sch_dsmark.c 	if (walker->stop)
walker            180 net/sched/sch_dsmark.c 		if (walker->count >= walker->skip) {
walker            181 net/sched/sch_dsmark.c 			if (walker->fn(sch, i + 1, walker) < 0) {
walker            182 net/sched/sch_dsmark.c 				walker->stop = 1;
walker            187 net/sched/sch_dsmark.c 		walker->count++;
walker             43 net/sched/sch_ingress.c static void ingress_walk(struct Qdisc *sch, struct qdisc_walker *walker)
walker           1247 net/sched/sch_netem.c static void netem_walk(struct Qdisc *sch, struct qdisc_walker *walker)
walker           1249 net/sched/sch_netem.c 	if (!walker->stop) {
walker           1250 net/sched/sch_netem.c 		if (walker->count >= walker->skip)
walker           1251 net/sched/sch_netem.c 			if (walker->fn(sch, 1, walker) < 0) {
walker           1252 net/sched/sch_netem.c 				walker->stop = 1;
walker           1255 net/sched/sch_netem.c 		walker->count++;
walker            407 net/sched/sch_red.c static void red_walk(struct Qdisc *sch, struct qdisc_walker *walker)
walker            409 net/sched/sch_red.c 	if (!walker->stop) {
walker            410 net/sched/sch_red.c 		if (walker->count >= walker->skip)
walker            411 net/sched/sch_red.c 			if (walker->fn(sch, 1, walker) < 0) {
walker            412 net/sched/sch_red.c 				walker->stop = 1;
walker            415 net/sched/sch_red.c 		walker->count++;
walker            657 net/sched/sch_sfb.c static void sfb_walk(struct Qdisc *sch, struct qdisc_walker *walker)
walker            659 net/sched/sch_sfb.c 	if (!walker->stop) {
walker            660 net/sched/sch_sfb.c 		if (walker->count >= walker->skip)
walker            661 net/sched/sch_sfb.c 			if (walker->fn(sch, 1, walker) < 0) {
walker            662 net/sched/sch_sfb.c 				walker->stop = 1;
walker            665 net/sched/sch_sfb.c 		walker->count++;
walker            512 net/sched/sch_tbf.c static void tbf_walk(struct Qdisc *sch, struct qdisc_walker *walker)
walker            514 net/sched/sch_tbf.c 	if (!walker->stop) {
walker            515 net/sched/sch_tbf.c 		if (walker->count >= walker->skip)
walker            516 net/sched/sch_tbf.c 			if (walker->fn(sch, 1, walker) < 0) {
walker            517 net/sched/sch_tbf.c 				walker->stop = 1;
walker            520 net/sched/sch_tbf.c 		walker->count++;
walker            276 security/yama/yama_lsm.c 	struct task_struct *walker = child;
walker            284 security/yama/yama_lsm.c 	while (walker->pid > 0) {
walker            285 security/yama/yama_lsm.c 		if (!thread_group_leader(walker))
walker            286 security/yama/yama_lsm.c 			walker = rcu_dereference(walker->group_leader);
walker            287 security/yama/yama_lsm.c 		if (walker == parent) {
walker            291 security/yama/yama_lsm.c 		walker = rcu_dereference(walker->real_parent);