Commit | Line | Data |
---|---|---|
982ed0de DW |
1 | // SPDX-License-Identifier: GPL-2.0-only |
2 | ||
3 | #ifndef __KVM_MM_H__ | |
4 | #define __KVM_MM_H__ 1 | |
5 | ||
6 | /* | |
7 | * Architectures can choose whether to use an rwlock or spinlock | |
8 | * for the mmu_lock. These macros, for use in common code | |
9 | * only, avoids using #ifdefs in places that must deal with | |
10 | * multiple architectures. | |
11 | */ | |
12 | ||
13 | #ifdef KVM_HAVE_MMU_RWLOCK | |
14 | #define KVM_MMU_LOCK_INIT(kvm) rwlock_init(&(kvm)->mmu_lock) | |
15 | #define KVM_MMU_LOCK(kvm) write_lock(&(kvm)->mmu_lock) | |
16 | #define KVM_MMU_UNLOCK(kvm) write_unlock(&(kvm)->mmu_lock) | |
17 | #define KVM_MMU_READ_LOCK(kvm) read_lock(&(kvm)->mmu_lock) | |
18 | #define KVM_MMU_READ_UNLOCK(kvm) read_unlock(&(kvm)->mmu_lock) | |
19 | #else | |
20 | #define KVM_MMU_LOCK_INIT(kvm) spin_lock_init(&(kvm)->mmu_lock) | |
21 | #define KVM_MMU_LOCK(kvm) spin_lock(&(kvm)->mmu_lock) | |
22 | #define KVM_MMU_UNLOCK(kvm) spin_unlock(&(kvm)->mmu_lock) | |
23 | #define KVM_MMU_READ_LOCK(kvm) spin_lock(&(kvm)->mmu_lock) | |
24 | #define KVM_MMU_READ_UNLOCK(kvm) spin_unlock(&(kvm)->mmu_lock) | |
25 | #endif /* KVM_HAVE_MMU_RWLOCK */ | |
26 | ||
27 | kvm_pfn_t hva_to_pfn(unsigned long addr, bool atomic, bool *async, | |
28 | bool write_fault, bool *writable); | |
29 | ||
30 | #ifdef CONFIG_HAVE_KVM_PFNCACHE | |
31 | void gfn_to_pfn_cache_invalidate_start(struct kvm *kvm, | |
32 | unsigned long start, | |
33 | unsigned long end, | |
34 | bool may_block); | |
35 | #else | |
36 | static inline void gfn_to_pfn_cache_invalidate_start(struct kvm *kvm, | |
37 | unsigned long start, | |
38 | unsigned long end, | |
39 | bool may_block) | |
40 | { | |
41 | } | |
42 | #endif /* HAVE_KVM_PFNCACHE */ | |
43 | ||
44 | #endif /* __KVM_MM_H__ */ |