Lines Matching refs:cr0
1170 svm->vcpu.arch.cr0 = 0; in init_vmcb()
1533 ulong gcr0 = svm->vcpu.arch.cr0; in update_cr0_intercept()
1534 u64 *hcr0 = &svm->vmcb->save.cr0; in update_cr0_intercept()
1553 static void svm_set_cr0(struct kvm_vcpu *vcpu, unsigned long cr0) in svm_set_cr0() argument
1559 if (!is_paging(vcpu) && (cr0 & X86_CR0_PG)) { in svm_set_cr0()
1564 if (is_paging(vcpu) && !(cr0 & X86_CR0_PG)) { in svm_set_cr0()
1570 vcpu->arch.cr0 = cr0; in svm_set_cr0()
1573 cr0 |= X86_CR0_PG | X86_CR0_WP; in svm_set_cr0()
1576 cr0 |= X86_CR0_TS; in svm_set_cr0()
1582 cr0 &= ~(X86_CR0_CD | X86_CR0_NW); in svm_set_cr0()
1583 svm->vmcb->save.cr0 = cr0; in svm_set_cr0()
2341 nested_vmcb->save.cr0 = kvm_read_cr0(&svm->vcpu); in nested_svm_vmexit()
2404 svm_set_cr0(&svm->vcpu, hsave->save.cr0 | X86_CR0_PE); in nested_svm_vmexit()
2528 hsave->save.cr0 = kvm_read_cr0(&svm->vcpu); in nested_svm_vmrun()
2561 svm_set_cr0(&svm->vcpu, nested_vmcb->save.cr0); in nested_svm_vmrun()
2914 unsigned long cr0 = svm->vcpu.arch.cr0; in check_selective_cr0_intercepted() local
2924 cr0 &= ~SVM_CR0_SELECTIVE_MASK; in check_selective_cr0_intercepted()
2927 if (cr0 ^ val) { in check_selective_cr0_intercepted()
3473 "cr0:", save->cr0, "cr2:", save->cr2); in dump_vmcb()
3516 vcpu->arch.cr0 = svm->vmcb->save.cr0; in handle_exit()
4238 unsigned long cr0, val; in svm_check_intercept() local
4253 cr0 = vcpu->arch.cr0 & ~SVM_CR0_SELECTIVE_MASK; in svm_check_intercept()
4257 cr0 &= 0xfUL; in svm_check_intercept()
4260 if (cr0 & X86_CR0_PE) in svm_check_intercept()
4264 if (cr0 ^ val) in svm_check_intercept()