Commit | Line | Data |
---|---|---|
1d737c8a ZX |
1 | #ifndef __KVM_X86_MMU_H |
2 | #define __KVM_X86_MMU_H | |
3 | ||
edf88417 | 4 | #include <linux/kvm_host.h> |
fc78f519 | 5 | #include "kvm_cache_regs.h" |
1d737c8a | 6 | |
8c6d6adc SY |
7 | #define PT64_PT_BITS 9 |
8 | #define PT64_ENT_PER_PAGE (1 << PT64_PT_BITS) | |
9 | #define PT32_PT_BITS 10 | |
10 | #define PT32_ENT_PER_PAGE (1 << PT32_PT_BITS) | |
11 | ||
12 | #define PT_WRITABLE_SHIFT 1 | |
13 | ||
14 | #define PT_PRESENT_MASK (1ULL << 0) | |
15 | #define PT_WRITABLE_MASK (1ULL << PT_WRITABLE_SHIFT) | |
16 | #define PT_USER_MASK (1ULL << 2) | |
17 | #define PT_PWT_MASK (1ULL << 3) | |
18 | #define PT_PCD_MASK (1ULL << 4) | |
1b7fcd32 AK |
19 | #define PT_ACCESSED_SHIFT 5 |
20 | #define PT_ACCESSED_MASK (1ULL << PT_ACCESSED_SHIFT) | |
8ea667f2 AK |
21 | #define PT_DIRTY_SHIFT 6 |
22 | #define PT_DIRTY_MASK (1ULL << PT_DIRTY_SHIFT) | |
6fd01b71 AK |
23 | #define PT_PAGE_SIZE_SHIFT 7 |
24 | #define PT_PAGE_SIZE_MASK (1ULL << PT_PAGE_SIZE_SHIFT) | |
8c6d6adc SY |
25 | #define PT_PAT_MASK (1ULL << 7) |
26 | #define PT_GLOBAL_MASK (1ULL << 8) | |
27 | #define PT64_NX_SHIFT 63 | |
28 | #define PT64_NX_MASK (1ULL << PT64_NX_SHIFT) | |
29 | ||
30 | #define PT_PAT_SHIFT 7 | |
31 | #define PT_DIR_PAT_SHIFT 12 | |
32 | #define PT_DIR_PAT_MASK (1ULL << PT_DIR_PAT_SHIFT) | |
33 | ||
34 | #define PT32_DIR_PSE36_SIZE 4 | |
35 | #define PT32_DIR_PSE36_SHIFT 13 | |
36 | #define PT32_DIR_PSE36_MASK \ | |
37 | (((1ULL << PT32_DIR_PSE36_SIZE) - 1) << PT32_DIR_PSE36_SHIFT) | |
38 | ||
39 | #define PT64_ROOT_LEVEL 4 | |
40 | #define PT32_ROOT_LEVEL 2 | |
41 | #define PT32E_ROOT_LEVEL 3 | |
42 | ||
c9c54174 SY |
43 | #define PT_PDPE_LEVEL 3 |
44 | #define PT_DIRECTORY_LEVEL 2 | |
45 | #define PT_PAGE_TABLE_LEVEL 1 | |
46 | ||
1871c602 GN |
47 | #define PFERR_PRESENT_MASK (1U << 0) |
48 | #define PFERR_WRITE_MASK (1U << 1) | |
49 | #define PFERR_USER_MASK (1U << 2) | |
50 | #define PFERR_RSVD_MASK (1U << 3) | |
51 | #define PFERR_FETCH_MASK (1U << 4) | |
52 | ||
94d8b056 | 53 | int kvm_mmu_get_spte_hierarchy(struct kvm_vcpu *vcpu, u64 addr, u64 sptes[4]); |
ce88decf XG |
54 | void kvm_mmu_set_mmio_spte_mask(u64 mmio_mask); |
55 | int handle_mmio_page_fault_common(struct kvm_vcpu *vcpu, u64 addr, bool direct); | |
52fde8df | 56 | int kvm_init_shadow_mmu(struct kvm_vcpu *vcpu, struct kvm_mmu *context); |
94d8b056 | 57 | |
e0df7b9f DH |
58 | static inline unsigned int kvm_mmu_available_pages(struct kvm *kvm) |
59 | { | |
49d5ca26 DH |
60 | return kvm->arch.n_max_mmu_pages - |
61 | kvm->arch.n_used_mmu_pages; | |
e0df7b9f DH |
62 | } |
63 | ||
1d737c8a ZX |
64 | static inline void kvm_mmu_free_some_pages(struct kvm_vcpu *vcpu) |
65 | { | |
e0df7b9f | 66 | if (unlikely(kvm_mmu_available_pages(vcpu->kvm)< KVM_MIN_FREE_MMU_PAGES)) |
1d737c8a ZX |
67 | __kvm_mmu_free_some_pages(vcpu); |
68 | } | |
69 | ||
70 | static inline int kvm_mmu_reload(struct kvm_vcpu *vcpu) | |
71 | { | |
72 | if (likely(vcpu->arch.mmu.root_hpa != INVALID_PAGE)) | |
73 | return 0; | |
74 | ||
75 | return kvm_mmu_load(vcpu); | |
76 | } | |
77 | ||
43a3795a | 78 | static inline int is_present_gpte(unsigned long pte) |
20c466b5 DE |
79 | { |
80 | return pte & PT_PRESENT_MASK; | |
81 | } | |
82 | ||
bebb106a XG |
83 | static inline int is_writable_pte(unsigned long pte) |
84 | { | |
85 | return pte & PT_WRITABLE_MASK; | |
86 | } | |
87 | ||
88 | static inline bool is_write_protection(struct kvm_vcpu *vcpu) | |
89 | { | |
90 | return kvm_read_cr0_bits(vcpu, X86_CR0_WP); | |
91 | } | |
92 | ||
97d64b78 AK |
93 | /* |
94 | * Will a fault with a given page-fault error code (pfec) cause a permission | |
95 | * fault with the given access (in ACC_* format)? | |
96 | */ | |
97 | static inline bool permission_fault(struct kvm_mmu *mmu, unsigned pte_access, | |
98 | unsigned pfec) | |
bebb106a | 99 | { |
97d64b78 | 100 | return (mmu->permissions[pfec >> 1] >> pte_access) & 1; |
bebb106a | 101 | } |
97d64b78 | 102 | |
1d737c8a | 103 | #endif |