Lines Matching refs:cr0

206 static void svm_set_cr0(struct kvm_vcpu *vcpu, unsigned long cr0);
1448 ulong gcr0 = svm->vcpu.arch.cr0; in update_cr0_intercept()
1449 u64 *hcr0 = &svm->vmcb->save.cr0; in update_cr0_intercept()
1468 static void svm_set_cr0(struct kvm_vcpu *vcpu, unsigned long cr0) in svm_set_cr0() argument
1474 if (!is_paging(vcpu) && (cr0 & X86_CR0_PG)) { in svm_set_cr0()
1479 if (is_paging(vcpu) && !(cr0 & X86_CR0_PG)) { in svm_set_cr0()
1485 vcpu->arch.cr0 = cr0; in svm_set_cr0()
1488 cr0 |= X86_CR0_PG | X86_CR0_WP; in svm_set_cr0()
1491 cr0 |= X86_CR0_TS; in svm_set_cr0()
1498 cr0 &= ~(X86_CR0_CD | X86_CR0_NW); in svm_set_cr0()
1499 svm->vmcb->save.cr0 = cr0; in svm_set_cr0()
2258 nested_vmcb->save.cr0 = kvm_read_cr0(&svm->vcpu); in nested_svm_vmexit()
2323 svm_set_cr0(&svm->vcpu, hsave->save.cr0 | X86_CR0_PE); in nested_svm_vmexit()
2447 hsave->save.cr0 = kvm_read_cr0(&svm->vcpu); in nested_svm_vmrun()
2480 svm_set_cr0(&svm->vcpu, nested_vmcb->save.cr0); in nested_svm_vmrun()
2833 unsigned long cr0 = svm->vcpu.arch.cr0; in check_selective_cr0_intercepted() local
2843 cr0 &= ~SVM_CR0_SELECTIVE_MASK; in check_selective_cr0_intercepted()
2846 if (cr0 ^ val) { in check_selective_cr0_intercepted()
3383 "cr0:", save->cr0, "cr2:", save->cr2); in dump_vmcb()
3428 vcpu->arch.cr0 = svm->vmcb->save.cr0; in handle_exit()
4157 unsigned long cr0, val; in svm_check_intercept() local
4172 cr0 = vcpu->arch.cr0 & ~SVM_CR0_SELECTIVE_MASK; in svm_check_intercept()
4176 cr0 &= 0xfUL; in svm_check_intercept()
4179 if (cr0 & X86_CR0_PE) in svm_check_intercept()
4183 if (cr0 ^ val) in svm_check_intercept()