walker 222 arch/x86/kvm/paging_tmpl.h struct guest_walker *walker, walker 235 arch/x86/kvm/paging_tmpl.h for (level = walker->max_level; level >= walker->level; --level) { walker 236 arch/x86/kvm/paging_tmpl.h pte = orig_pte = walker->ptes[level - 1]; walker 237 arch/x86/kvm/paging_tmpl.h table_gfn = walker->table_gfn[level - 1]; walker 238 arch/x86/kvm/paging_tmpl.h ptep_user = walker->ptep_user[level - 1]; walker 244 arch/x86/kvm/paging_tmpl.h if (level == walker->level && write_fault && walker 269 arch/x86/kvm/paging_tmpl.h if (unlikely(!walker->pte_writable[level - 1])) walker 277 arch/x86/kvm/paging_tmpl.h walker->ptes[level - 1] = pte; walker 296 arch/x86/kvm/paging_tmpl.h static int FNAME(walk_addr_generic)(struct guest_walker *walker, walker 320 arch/x86/kvm/paging_tmpl.h walker->level = mmu->root_level; walker 326 arch/x86/kvm/paging_tmpl.h if (walker->level == PT32E_ROOT_LEVEL) { walker 328 arch/x86/kvm/paging_tmpl.h trace_kvm_mmu_paging_element(pte, walker->level); walker 331 arch/x86/kvm/paging_tmpl.h --walker->level; walker 334 arch/x86/kvm/paging_tmpl.h walker->max_level = walker->level; walker 345 arch/x86/kvm/paging_tmpl.h ++walker->level; walker 352 arch/x86/kvm/paging_tmpl.h --walker->level; walker 354 arch/x86/kvm/paging_tmpl.h index = PT_INDEX(addr, walker->level); walker 359 arch/x86/kvm/paging_tmpl.h BUG_ON(walker->level < 1); walker 360 arch/x86/kvm/paging_tmpl.h walker->table_gfn[walker->level - 1] = table_gfn; walker 361 arch/x86/kvm/paging_tmpl.h walker->pte_gpa[walker->level - 1] = pte_gpa; walker 365 arch/x86/kvm/paging_tmpl.h &walker->fault); walker 383 arch/x86/kvm/paging_tmpl.h &walker->pte_writable[walker->level - 1]); walker 390 arch/x86/kvm/paging_tmpl.h walker->ptep_user[walker->level - 1] = ptep_user; walker 392 arch/x86/kvm/paging_tmpl.h trace_kvm_mmu_paging_element(pte, walker->level); walker 403 arch/x86/kvm/paging_tmpl.h if (unlikely(is_rsvd_bits_set(mmu, pte, walker->level))) { walker 408 arch/x86/kvm/paging_tmpl.h walker->ptes[walker->level - 1] = pte; walker 409 arch/x86/kvm/paging_tmpl.h } while (!is_last_gpte(mmu, walker->level, pte)); walker 415 arch/x86/kvm/paging_tmpl.h walker->pt_access = FNAME(gpte_access)(pt_access ^ walk_nx_mask); walker 416 arch/x86/kvm/paging_tmpl.h walker->pte_access = FNAME(gpte_access)(pte_access ^ walk_nx_mask); walker 417 arch/x86/kvm/paging_tmpl.h errcode = permission_fault(vcpu, mmu, walker->pte_access, pte_pkey, access); walker 421 arch/x86/kvm/paging_tmpl.h gfn = gpte_to_gfn_lvl(pte, walker->level); walker 422 arch/x86/kvm/paging_tmpl.h gfn += (addr & PT_LVL_OFFSET_MASK(walker->level)) >> PAGE_SHIFT; walker 424 arch/x86/kvm/paging_tmpl.h if (PTTYPE == 32 && walker->level == PT_DIRECTORY_LEVEL && is_cpuid_PSE36()) walker 427 arch/x86/kvm/paging_tmpl.h real_gpa = mmu->translate_gpa(vcpu, gfn_to_gpa(gfn), access, &walker->fault); walker 431 arch/x86/kvm/paging_tmpl.h walker->gfn = real_gpa >> PAGE_SHIFT; walker 434 arch/x86/kvm/paging_tmpl.h FNAME(protect_clean_gpte)(mmu, &walker->pte_access, pte); walker 445 arch/x86/kvm/paging_tmpl.h ret = FNAME(update_accessed_dirty_bits)(vcpu, mmu, walker, write_fault); walker 453 arch/x86/kvm/paging_tmpl.h __func__, (u64)pte, walker->pte_access, walker->pt_access); walker 462 arch/x86/kvm/paging_tmpl.h walker->fault.vector = PF_VECTOR; walker 463 arch/x86/kvm/paging_tmpl.h walker->fault.error_code_valid = true; walker 464 arch/x86/kvm/paging_tmpl.h walker->fault.error_code = errcode; walker 491 arch/x86/kvm/paging_tmpl.h walker->fault.address = addr; walker 492 arch/x86/kvm/paging_tmpl.h walker->fault.nested_page_fault = mmu != vcpu->arch.walk_mmu; walker 494 arch/x86/kvm/paging_tmpl.h trace_kvm_mmu_walker_error(walker->fault.error_code); walker 498 arch/x86/kvm/paging_tmpl.h static int FNAME(walk_addr)(struct guest_walker *walker, walker 501 arch/x86/kvm/paging_tmpl.h return FNAME(walk_addr_generic)(walker, vcpu, vcpu->arch.mmu, addr, walker 506 arch/x86/kvm/paging_tmpl.h static int FNAME(walk_addr_nested)(struct guest_walker *walker, walker 510 arch/x86/kvm/paging_tmpl.h return FNAME(walk_addr_generic)(walker, vcpu, &vcpu->arch.nested_mmu, walker 733 arch/x86/kvm/paging_tmpl.h struct guest_walker *walker, int user_fault, walker 737 arch/x86/kvm/paging_tmpl.h gfn_t mask = ~(KVM_PAGES_PER_HPAGE(walker->level) - 1); walker 740 arch/x86/kvm/paging_tmpl.h if (!(walker->pte_access & ACC_WRITE_MASK || walker 744 arch/x86/kvm/paging_tmpl.h for (level = walker->level; level <= walker->max_level; level++) { walker 745 arch/x86/kvm/paging_tmpl.h gfn_t gfn = walker->gfn ^ walker->table_gfn[level - 1]; walker 773 arch/x86/kvm/paging_tmpl.h struct guest_walker walker; walker 798 arch/x86/kvm/paging_tmpl.h r = FNAME(walk_addr)(&walker, vcpu, addr, error_code); walker 806 arch/x86/kvm/paging_tmpl.h inject_page_fault(vcpu, &walker.fault); walker 811 arch/x86/kvm/paging_tmpl.h if (page_fault_handle_page_track(vcpu, error_code, walker.gfn)) { walker 819 arch/x86/kvm/paging_tmpl.h &walker, user_fault, &vcpu->arch.write_fault_to_shadow_pgtable); walker 821 arch/x86/kvm/paging_tmpl.h if (walker.level >= PT_DIRECTORY_LEVEL && !is_self_change_mapping) { walker 822 arch/x86/kvm/paging_tmpl.h level = mapping_level(vcpu, walker.gfn, &force_pt_level); walker 824 arch/x86/kvm/paging_tmpl.h level = min(walker.level, level); walker 825 arch/x86/kvm/paging_tmpl.h walker.gfn = walker.gfn & ~(KVM_PAGES_PER_HPAGE(level) - 1); walker 833 arch/x86/kvm/paging_tmpl.h if (try_async_pf(vcpu, prefault, walker.gfn, addr, &pfn, write_fault, walker 837 arch/x86/kvm/paging_tmpl.h if (handle_abnormal_pfn(vcpu, addr, walker.gfn, pfn, walker.pte_access, &r)) walker 844 arch/x86/kvm/paging_tmpl.h if (write_fault && !(walker.pte_access & ACC_WRITE_MASK) && walker 847 arch/x86/kvm/paging_tmpl.h walker.pte_access |= ACC_WRITE_MASK; walker 848 arch/x86/kvm/paging_tmpl.h walker.pte_access &= ~ACC_USER_MASK; walker 857 arch/x86/kvm/paging_tmpl.h walker.pte_access &= ~ACC_EXEC_MASK; walker 869 arch/x86/kvm/paging_tmpl.h transparent_hugepage_adjust(vcpu, walker.gfn, &pfn, &level); walker 870 arch/x86/kvm/paging_tmpl.h r = FNAME(fetch)(vcpu, addr, &walker, write_fault, walker 952 arch/x86/kvm/paging_tmpl.h struct guest_walker walker; walker 956 arch/x86/kvm/paging_tmpl.h r = FNAME(walk_addr)(&walker, vcpu, addr, access); walker 959 arch/x86/kvm/paging_tmpl.h gpa = gfn_to_gpa(walker.gfn); walker 962 arch/x86/kvm/paging_tmpl.h *exception = walker.fault; walker 973 arch/x86/kvm/paging_tmpl.h struct guest_walker walker; walker 982 arch/x86/kvm/paging_tmpl.h r = FNAME(walk_addr_nested)(&walker, vcpu, vaddr, access); walker 985 arch/x86/kvm/paging_tmpl.h gpa = gfn_to_gpa(walker.gfn); walker 988 arch/x86/kvm/paging_tmpl.h *exception = walker.fault; walker 150 drivers/pci/controller/dwc/pcie-hisi.c void *walker = ®_val; walker 153 drivers/pci/controller/dwc/pcie-hisi.c walker += (where & 0x3); walker 158 drivers/pci/controller/dwc/pcie-hisi.c *val = *(u8 __force *) walker; walker 160 drivers/pci/controller/dwc/pcie-hisi.c *val = *(u16 __force *) walker; walker 175 drivers/pci/controller/dwc/pcie-hisi.c void *walker = ®_val; walker 178 drivers/pci/controller/dwc/pcie-hisi.c walker += (where & 0x3); walker 184 drivers/pci/controller/dwc/pcie-hisi.c *(u16 __force *) walker = val; walker 188 drivers/pci/controller/dwc/pcie-hisi.c *(u8 __force *) walker = val; walker 489 fs/ecryptfs/keystore.c struct ecryptfs_global_auth_tok *walker; walker 495 fs/ecryptfs/keystore.c list_for_each_entry(walker, walker 498 fs/ecryptfs/keystore.c if (memcmp(walker->sig, sig, ECRYPTFS_SIG_SIZE_HEX)) walker 501 fs/ecryptfs/keystore.c if (walker->flags & ECRYPTFS_AUTH_TOK_INVALID) { walker 506 fs/ecryptfs/keystore.c rc = key_validate(walker->global_auth_tok_key); walker 513 fs/ecryptfs/keystore.c down_write(&(walker->global_auth_tok_key->sem)); walker 515 fs/ecryptfs/keystore.c walker->global_auth_tok_key, auth_tok); walker 519 fs/ecryptfs/keystore.c (*auth_tok_key) = walker->global_auth_tok_key; walker 526 fs/ecryptfs/keystore.c up_write(&(walker->global_auth_tok_key->sem)); walker 529 fs/ecryptfs/keystore.c walker->flags |= ECRYPTFS_AUTH_TOK_INVALID; walker 530 fs/ecryptfs/keystore.c key_put(walker->global_auth_tok_key); walker 531 fs/ecryptfs/keystore.c walker->global_auth_tok_key = NULL; walker 137 fs/ecryptfs/super.c struct ecryptfs_global_auth_tok *walker; walker 140 fs/ecryptfs/super.c list_for_each_entry(walker, walker 143 fs/ecryptfs/super.c if (walker->flags & ECRYPTFS_AUTH_TOK_FNEK) walker 144 fs/ecryptfs/super.c seq_printf(m, ",ecryptfs_fnek_sig=%s", walker->sig); walker 146 fs/ecryptfs/super.c seq_printf(m, ",ecryptfs_sig=%s", walker->sig); walker 471 fs/gfs2/bmap.c u64 max_len, gfs2_metadata_walker walker) walker 501 fs/gfs2/bmap.c status = walker(mp, ptrs); walker 124 include/linux/rhashtable-types.h struct rhashtable_walker walker; walker 514 kernel/seccomp.c struct seccomp_filter *walker; walker 520 kernel/seccomp.c for (walker = current->seccomp.filter; walker; walker = walker->prev) walker 521 kernel/seccomp.c total_insns += walker->prog->len + 4; /* 4 instr penalty */ walker 307 lib/rhashtable.c struct rhashtable_walker *walker; walker 326 lib/rhashtable.c list_for_each_entry(walker, &old_tbl->walkers, list) walker 327 lib/rhashtable.c walker->tbl = NULL; walker 667 lib/rhashtable.c iter->walker.tbl = walker 669 lib/rhashtable.c list_add(&iter->walker.list, &iter->walker.tbl->walkers); walker 683 lib/rhashtable.c if (iter->walker.tbl) walker 684 lib/rhashtable.c list_del(&iter->walker.list); walker 716 lib/rhashtable.c if (iter->walker.tbl) walker 717 lib/rhashtable.c list_del(&iter->walker.list); walker 722 lib/rhashtable.c if (!iter->walker.tbl) { walker 723 lib/rhashtable.c iter->walker.tbl = rht_dereference_rcu(ht->tbl, ht); walker 736 lib/rhashtable.c rht_for_each_rcu(p, iter->walker.tbl, iter->slot) { walker 751 lib/rhashtable.c rht_for_each_rcu(p, iter->walker.tbl, iter->slot) { walker 782 lib/rhashtable.c struct bucket_table *tbl = iter->walker.tbl; walker 828 lib/rhashtable.c iter->walker.tbl = rht_dereference_rcu(tbl->future_tbl, ht); walker 829 lib/rhashtable.c if (iter->walker.tbl) { walker 927 lib/rhashtable.c struct bucket_table *tbl = iter->walker.tbl; walker 937 lib/rhashtable.c iter->walker.tbl = NULL; walker 939 lib/rhashtable.c list_add(&iter->walker.list, &tbl->walkers); walker 550 net/sched/cls_tcindex.c static void tcindex_walk(struct tcf_proto *tp, struct tcf_walker *walker, walker 557 net/sched/cls_tcindex.c pr_debug("tcindex_walk(tp %p,walker %p),p %p\n", tp, walker, p); walker 562 net/sched/cls_tcindex.c if (walker->count >= walker->skip) { walker 563 net/sched/cls_tcindex.c if (walker->fn(tp, p->perfect + i, walker) < 0) { walker 564 net/sched/cls_tcindex.c walker->stop = 1; walker 568 net/sched/cls_tcindex.c walker->count++; walker 576 net/sched/cls_tcindex.c if (walker->count >= walker->skip) { walker 577 net/sched/cls_tcindex.c if (walker->fn(tp, &f->result, walker) < 0) { walker 578 net/sched/cls_tcindex.c walker->stop = 1; walker 582 net/sched/cls_tcindex.c walker->count++; walker 347 net/sched/sch_atm.c static void atm_tc_walk(struct Qdisc *sch, struct qdisc_walker *walker) walker 352 net/sched/sch_atm.c pr_debug("atm_tc_walk(sch %p,[qdisc %p],walker %p)\n", sch, p, walker); walker 353 net/sched/sch_atm.c if (walker->stop) walker 356 net/sched/sch_atm.c if (walker->count >= walker->skip && walker 357 net/sched/sch_atm.c walker->fn(sch, (unsigned long)flow, walker) < 0) { walker 358 net/sched/sch_atm.c walker->stop = 1; walker 361 net/sched/sch_atm.c walker->count++; walker 519 net/sched/sch_cbs.c static void cbs_walk(struct Qdisc *sch, struct qdisc_walker *walker) walker 521 net/sched/sch_cbs.c if (!walker->stop) { walker 522 net/sched/sch_cbs.c if (walker->count >= walker->skip) { walker 523 net/sched/sch_cbs.c if (walker->fn(sch, 1, walker) < 0) { walker 524 net/sched/sch_cbs.c walker->stop = 1; walker 528 net/sched/sch_cbs.c walker->count++; walker 166 net/sched/sch_dsmark.c static void dsmark_walk(struct Qdisc *sch, struct qdisc_walker *walker) walker 172 net/sched/sch_dsmark.c __func__, sch, p, walker); walker 174 net/sched/sch_dsmark.c if (walker->stop) walker 180 net/sched/sch_dsmark.c if (walker->count >= walker->skip) { walker 181 net/sched/sch_dsmark.c if (walker->fn(sch, i + 1, walker) < 0) { walker 182 net/sched/sch_dsmark.c walker->stop = 1; walker 187 net/sched/sch_dsmark.c walker->count++; walker 43 net/sched/sch_ingress.c static void ingress_walk(struct Qdisc *sch, struct qdisc_walker *walker) walker 1247 net/sched/sch_netem.c static void netem_walk(struct Qdisc *sch, struct qdisc_walker *walker) walker 1249 net/sched/sch_netem.c if (!walker->stop) { walker 1250 net/sched/sch_netem.c if (walker->count >= walker->skip) walker 1251 net/sched/sch_netem.c if (walker->fn(sch, 1, walker) < 0) { walker 1252 net/sched/sch_netem.c walker->stop = 1; walker 1255 net/sched/sch_netem.c walker->count++; walker 407 net/sched/sch_red.c static void red_walk(struct Qdisc *sch, struct qdisc_walker *walker) walker 409 net/sched/sch_red.c if (!walker->stop) { walker 410 net/sched/sch_red.c if (walker->count >= walker->skip) walker 411 net/sched/sch_red.c if (walker->fn(sch, 1, walker) < 0) { walker 412 net/sched/sch_red.c walker->stop = 1; walker 415 net/sched/sch_red.c walker->count++; walker 657 net/sched/sch_sfb.c static void sfb_walk(struct Qdisc *sch, struct qdisc_walker *walker) walker 659 net/sched/sch_sfb.c if (!walker->stop) { walker 660 net/sched/sch_sfb.c if (walker->count >= walker->skip) walker 661 net/sched/sch_sfb.c if (walker->fn(sch, 1, walker) < 0) { walker 662 net/sched/sch_sfb.c walker->stop = 1; walker 665 net/sched/sch_sfb.c walker->count++; walker 512 net/sched/sch_tbf.c static void tbf_walk(struct Qdisc *sch, struct qdisc_walker *walker) walker 514 net/sched/sch_tbf.c if (!walker->stop) { walker 515 net/sched/sch_tbf.c if (walker->count >= walker->skip) walker 516 net/sched/sch_tbf.c if (walker->fn(sch, 1, walker) < 0) { walker 517 net/sched/sch_tbf.c walker->stop = 1; walker 520 net/sched/sch_tbf.c walker->count++; walker 276 security/yama/yama_lsm.c struct task_struct *walker = child; walker 284 security/yama/yama_lsm.c while (walker->pid > 0) { walker 285 security/yama/yama_lsm.c if (!thread_group_leader(walker)) walker 286 security/yama/yama_lsm.c walker = rcu_dereference(walker->group_leader); walker 287 security/yama/yama_lsm.c if (walker == parent) { walker 291 security/yama/yama_lsm.c walker = rcu_dereference(walker->real_parent);