KVM: Reinstate gfn_to_pfn_cache with invalidation support
[linux-block.git] / virt / kvm / kvm_mm.h
CommitLineData
982ed0de
DW
1// SPDX-License-Identifier: GPL-2.0-only
2
3#ifndef __KVM_MM_H__
4#define __KVM_MM_H__ 1
5
6/*
7 * Architectures can choose whether to use an rwlock or spinlock
8 * for the mmu_lock. These macros, for use in common code
9 * only, avoids using #ifdefs in places that must deal with
10 * multiple architectures.
11 */
12
13#ifdef KVM_HAVE_MMU_RWLOCK
14#define KVM_MMU_LOCK_INIT(kvm) rwlock_init(&(kvm)->mmu_lock)
15#define KVM_MMU_LOCK(kvm) write_lock(&(kvm)->mmu_lock)
16#define KVM_MMU_UNLOCK(kvm) write_unlock(&(kvm)->mmu_lock)
17#define KVM_MMU_READ_LOCK(kvm) read_lock(&(kvm)->mmu_lock)
18#define KVM_MMU_READ_UNLOCK(kvm) read_unlock(&(kvm)->mmu_lock)
19#else
20#define KVM_MMU_LOCK_INIT(kvm) spin_lock_init(&(kvm)->mmu_lock)
21#define KVM_MMU_LOCK(kvm) spin_lock(&(kvm)->mmu_lock)
22#define KVM_MMU_UNLOCK(kvm) spin_unlock(&(kvm)->mmu_lock)
23#define KVM_MMU_READ_LOCK(kvm) spin_lock(&(kvm)->mmu_lock)
24#define KVM_MMU_READ_UNLOCK(kvm) spin_unlock(&(kvm)->mmu_lock)
25#endif /* KVM_HAVE_MMU_RWLOCK */
26
27kvm_pfn_t hva_to_pfn(unsigned long addr, bool atomic, bool *async,
28 bool write_fault, bool *writable);
29
30#ifdef CONFIG_HAVE_KVM_PFNCACHE
31void gfn_to_pfn_cache_invalidate_start(struct kvm *kvm,
32 unsigned long start,
33 unsigned long end,
34 bool may_block);
35#else
36static inline void gfn_to_pfn_cache_invalidate_start(struct kvm *kvm,
37 unsigned long start,
38 unsigned long end,
39 bool may_block)
40{
41}
42#endif /* HAVE_KVM_PFNCACHE */
43
44#endif /* __KVM_MM_H__ */