Commit | Line | Data |
---|---|---|
d809aa23 | 1 | /* SPDX-License-Identifier: GPL-2.0 */ |
b0c632db | 2 | /* |
a53c8fab | 3 | * definition for kvm on s390 |
b0c632db | 4 | * |
a53c8fab | 5 | * Copyright IBM Corp. 2008, 2009 |
b0c632db | 6 | * |
b0c632db HC |
7 | * Author(s): Carsten Otte <cotte@de.ibm.com> |
8 | * Christian Borntraeger <borntraeger@de.ibm.com> | |
628eb9b8 | 9 | * Christian Ehrhardt <ehrhardt@de.ibm.com> |
b0c632db HC |
10 | */ |
11 | ||
12 | #ifndef ARCH_S390_KVM_S390_H | |
13 | #define ARCH_S390_KVM_S390_H | |
8f2abe6a | 14 | |
ca872302 | 15 | #include <linux/hrtimer.h> |
ba5c1e9b | 16 | #include <linux/kvm.h> |
8f2abe6a | 17 | #include <linux/kvm_host.h> |
9d8d5786 | 18 | #include <asm/facility.h> |
0e8bc06a | 19 | #include <asm/processor.h> |
a6940674 | 20 | #include <asm/sclp.h> |
8f2abe6a | 21 | |
7feb6bb8 | 22 | /* Transactional Memory Execution related macros */ |
0c9d8683 | 23 | #define IS_TE_ENABLED(vcpu) ((vcpu->arch.sie_block->ecb & ECB_TE)) |
7feb6bb8 MM |
24 | #define TDB_FORMAT1 1 |
25 | #define IS_ITDB_VALID(vcpu) ((*(char *)vcpu->arch.sie_block->itdba == TDB_FORMAT1)) | |
26 | ||
78f26131 CB |
27 | extern debug_info_t *kvm_s390_dbf; |
28 | #define KVM_EVENT(d_loglevel, d_string, d_args...)\ | |
29 | do { \ | |
30 | debug_sprintf_event(kvm_s390_dbf, d_loglevel, d_string "\n", \ | |
31 | d_args); \ | |
32 | } while (0) | |
33 | ||
b0c632db HC |
34 | #define VM_EVENT(d_kvm, d_loglevel, d_string, d_args...)\ |
35 | do { \ | |
36 | debug_sprintf_event(d_kvm->arch.dbf, d_loglevel, d_string "\n", \ | |
37 | d_args); \ | |
38 | } while (0) | |
39 | ||
40 | #define VCPU_EVENT(d_vcpu, d_loglevel, d_string, d_args...)\ | |
41 | do { \ | |
42 | debug_sprintf_event(d_vcpu->kvm->arch.dbf, d_loglevel, \ | |
43 | "%02d[%016lx-%016lx]: " d_string "\n", d_vcpu->vcpu_id, \ | |
44 | d_vcpu->arch.sie_block->gpsw.mask, d_vcpu->arch.sie_block->gpsw.addr,\ | |
45 | d_args); \ | |
46 | } while (0) | |
ba5c1e9b | 47 | |
2018224d DH |
48 | static inline void kvm_s390_set_cpuflags(struct kvm_vcpu *vcpu, u32 flags) |
49 | { | |
50 | atomic_or(flags, &vcpu->arch.sie_block->cpuflags); | |
51 | } | |
52 | ||
9daecfc6 DH |
53 | static inline void kvm_s390_clear_cpuflags(struct kvm_vcpu *vcpu, u32 flags) |
54 | { | |
55 | atomic_andnot(flags, &vcpu->arch.sie_block->cpuflags); | |
56 | } | |
57 | ||
8d5fb0dc DH |
58 | static inline bool kvm_s390_test_cpuflags(struct kvm_vcpu *vcpu, u32 flags) |
59 | { | |
60 | return (atomic_read(&vcpu->arch.sie_block->cpuflags) & flags) == flags; | |
61 | } | |
62 | ||
7a42fdc2 | 63 | static inline int is_vcpu_stopped(struct kvm_vcpu *vcpu) |
ba5c1e9b | 64 | { |
8d5fb0dc | 65 | return kvm_s390_test_cpuflags(vcpu, CPUSTAT_STOPPED); |
ba5c1e9b CO |
66 | } |
67 | ||
5ebda316 DH |
68 | static inline int is_vcpu_idle(struct kvm_vcpu *vcpu) |
69 | { | |
a9f6c9a9 | 70 | return test_bit(vcpu->vcpu_id, vcpu->kvm->arch.float_int.idle_mask); |
5ebda316 DH |
71 | } |
72 | ||
e08b9637 CO |
73 | static inline int kvm_is_ucontrol(struct kvm *kvm) |
74 | { | |
75 | #ifdef CONFIG_KVM_S390_UCONTROL | |
76 | if (kvm->arch.gmap) | |
77 | return 0; | |
78 | return 1; | |
79 | #else | |
80 | return 0; | |
81 | #endif | |
82 | } | |
8d26cf7b | 83 | |
fda902cb MM |
84 | #define GUEST_PREFIX_SHIFT 13 |
85 | static inline u32 kvm_s390_get_prefix(struct kvm_vcpu *vcpu) | |
86 | { | |
87 | return vcpu->arch.sie_block->prefix << GUEST_PREFIX_SHIFT; | |
88 | } | |
89 | ||
8d26cf7b CB |
90 | static inline void kvm_s390_set_prefix(struct kvm_vcpu *vcpu, u32 prefix) |
91 | { | |
71db35d2 CB |
92 | VCPU_EVENT(vcpu, 3, "set prefix of cpu %03u to 0x%x", vcpu->vcpu_id, |
93 | prefix); | |
fda902cb | 94 | vcpu->arch.sie_block->prefix = prefix >> GUEST_PREFIX_SHIFT; |
d3d692c8 | 95 | kvm_make_request(KVM_REQ_TLB_FLUSH, vcpu); |
2c70fe44 | 96 | kvm_make_request(KVM_REQ_MMU_RELOAD, vcpu); |
8d26cf7b CB |
97 | } |
98 | ||
27f67f87 | 99 | static inline u64 kvm_s390_get_base_disp_s(struct kvm_vcpu *vcpu, u8 *ar) |
b1c571a5 | 100 | { |
0c29b229 CB |
101 | u32 base2 = vcpu->arch.sie_block->ipb >> 28; |
102 | u32 disp2 = ((vcpu->arch.sie_block->ipb & 0x0fff0000) >> 16); | |
b1c571a5 | 103 | |
8ae04b8f AY |
104 | if (ar) |
105 | *ar = base2; | |
106 | ||
b1c571a5 CH |
107 | return (base2 ? vcpu->run->s.regs.gprs[base2] : 0) + disp2; |
108 | } | |
109 | ||
110 | static inline void kvm_s390_get_base_disp_sse(struct kvm_vcpu *vcpu, | |
8ae04b8f | 111 | u64 *address1, u64 *address2, |
27f67f87 | 112 | u8 *ar_b1, u8 *ar_b2) |
b1c571a5 | 113 | { |
0c29b229 CB |
114 | u32 base1 = (vcpu->arch.sie_block->ipb & 0xf0000000) >> 28; |
115 | u32 disp1 = (vcpu->arch.sie_block->ipb & 0x0fff0000) >> 16; | |
116 | u32 base2 = (vcpu->arch.sie_block->ipb & 0xf000) >> 12; | |
117 | u32 disp2 = vcpu->arch.sie_block->ipb & 0x0fff; | |
b1c571a5 CH |
118 | |
119 | *address1 = (base1 ? vcpu->run->s.regs.gprs[base1] : 0) + disp1; | |
120 | *address2 = (base2 ? vcpu->run->s.regs.gprs[base2] : 0) + disp2; | |
8ae04b8f AY |
121 | |
122 | if (ar_b1) | |
123 | *ar_b1 = base1; | |
124 | if (ar_b2) | |
125 | *ar_b2 = base2; | |
b1c571a5 CH |
126 | } |
127 | ||
69d0d3a3 CB |
128 | static inline void kvm_s390_get_regs_rre(struct kvm_vcpu *vcpu, int *r1, int *r2) |
129 | { | |
ff7158b2 TH |
130 | if (r1) |
131 | *r1 = (vcpu->arch.sie_block->ipb & 0x00f00000) >> 20; | |
132 | if (r2) | |
133 | *r2 = (vcpu->arch.sie_block->ipb & 0x000f0000) >> 16; | |
69d0d3a3 CB |
134 | } |
135 | ||
27f67f87 | 136 | static inline u64 kvm_s390_get_base_disp_rsy(struct kvm_vcpu *vcpu, u8 *ar) |
b1c571a5 | 137 | { |
0c29b229 CB |
138 | u32 base2 = vcpu->arch.sie_block->ipb >> 28; |
139 | u32 disp2 = ((vcpu->arch.sie_block->ipb & 0x0fff0000) >> 16) + | |
b1c571a5 | 140 | ((vcpu->arch.sie_block->ipb & 0xff00) << 4); |
0c29b229 CB |
141 | /* The displacement is a 20bit _SIGNED_ value */ |
142 | if (disp2 & 0x80000) | |
143 | disp2+=0xfff00000; | |
b1c571a5 | 144 | |
8ae04b8f AY |
145 | if (ar) |
146 | *ar = base2; | |
147 | ||
0c29b229 | 148 | return (base2 ? vcpu->run->s.regs.gprs[base2] : 0) + (long)(int)disp2; |
b1c571a5 CH |
149 | } |
150 | ||
27f67f87 | 151 | static inline u64 kvm_s390_get_base_disp_rs(struct kvm_vcpu *vcpu, u8 *ar) |
b1c571a5 | 152 | { |
0c29b229 CB |
153 | u32 base2 = vcpu->arch.sie_block->ipb >> 28; |
154 | u32 disp2 = ((vcpu->arch.sie_block->ipb & 0x0fff0000) >> 16); | |
b1c571a5 | 155 | |
8ae04b8f AY |
156 | if (ar) |
157 | *ar = base2; | |
158 | ||
b1c571a5 CH |
159 | return (base2 ? vcpu->run->s.regs.gprs[base2] : 0) + disp2; |
160 | } | |
161 | ||
ea828ebf TH |
162 | /* Set the condition code in the guest program status word */ |
163 | static inline void kvm_s390_set_psw_cc(struct kvm_vcpu *vcpu, unsigned long cc) | |
164 | { | |
165 | vcpu->arch.sie_block->gpsw.mask &= ~(3UL << 44); | |
166 | vcpu->arch.sie_block->gpsw.mask |= cc << 44; | |
167 | } | |
168 | ||
1e8d2424 | 169 | /* test availability of facility in a kvm instance */ |
9d8d5786 MM |
170 | static inline int test_kvm_facility(struct kvm *kvm, unsigned long nr) |
171 | { | |
c54f0d6a DH |
172 | return __test_facility(nr, kvm->arch.model.fac_mask) && |
173 | __test_facility(nr, kvm->arch.model.fac_list); | |
9d8d5786 MM |
174 | } |
175 | ||
18280d8b MM |
176 | static inline int set_kvm_facility(u64 *fac_list, unsigned long nr) |
177 | { | |
178 | unsigned char *ptr; | |
179 | ||
180 | if (nr >= MAX_FACILITY_BIT) | |
181 | return -EINVAL; | |
182 | ptr = (unsigned char *) fac_list + (nr >> 3); | |
183 | *ptr |= (0x80UL >> (nr & 7)); | |
184 | return 0; | |
185 | } | |
186 | ||
15c9705f DH |
187 | static inline int test_kvm_cpu_feat(struct kvm *kvm, unsigned long nr) |
188 | { | |
189 | WARN_ON_ONCE(nr >= KVM_S390_VM_CPU_FEAT_NR_BITS); | |
190 | return test_bit_inv(nr, kvm->arch.cpu_feat); | |
191 | } | |
192 | ||
6352e4d2 DH |
193 | /* are cpu states controlled by user space */ |
194 | static inline int kvm_s390_user_cpu_state_ctrl(struct kvm *kvm) | |
195 | { | |
196 | return kvm->arch.user_cpu_state_ctrl != 0; | |
197 | } | |
198 | ||
66933b78 | 199 | /* implemented in interrupt.c */ |
ba5c1e9b | 200 | int kvm_s390_handle_wait(struct kvm_vcpu *vcpu); |
0e9c85a5 | 201 | void kvm_s390_vcpu_wakeup(struct kvm_vcpu *vcpu); |
ca872302 | 202 | enum hrtimer_restart kvm_s390_idle_wakeup(struct hrtimer *timer); |
614aeab4 | 203 | int __must_check kvm_s390_deliver_pending_interrupts(struct kvm_vcpu *vcpu); |
2ed10cc1 | 204 | void kvm_s390_clear_local_irqs(struct kvm_vcpu *vcpu); |
67335e63 | 205 | void kvm_s390_clear_float_irqs(struct kvm *kvm); |
db4a29cb HC |
206 | int __must_check kvm_s390_inject_vm(struct kvm *kvm, |
207 | struct kvm_s390_interrupt *s390int); | |
208 | int __must_check kvm_s390_inject_vcpu(struct kvm_vcpu *vcpu, | |
383d0b05 | 209 | struct kvm_s390_irq *irq); |
66933b78 DH |
210 | static inline int kvm_s390_inject_prog_irq(struct kvm_vcpu *vcpu, |
211 | struct kvm_s390_pgm_info *pgm_info) | |
212 | { | |
213 | struct kvm_s390_irq irq = { | |
214 | .type = KVM_S390_PROGRAM_INT, | |
215 | .u.pgm = *pgm_info, | |
216 | }; | |
217 | ||
218 | return kvm_s390_inject_vcpu(vcpu, &irq); | |
219 | } | |
220 | static inline int kvm_s390_inject_program_int(struct kvm_vcpu *vcpu, u16 code) | |
221 | { | |
222 | struct kvm_s390_irq irq = { | |
223 | .type = KVM_S390_PROGRAM_INT, | |
224 | .u.pgm.code = code, | |
225 | }; | |
226 | ||
227 | return kvm_s390_inject_vcpu(vcpu, &irq); | |
228 | } | |
fa6b7fe9 | 229 | struct kvm_s390_interrupt_info *kvm_s390_get_io_int(struct kvm *kvm, |
6d3da241 | 230 | u64 isc_mask, u32 schid); |
15462e37 DH |
231 | int kvm_s390_reinject_io_int(struct kvm *kvm, |
232 | struct kvm_s390_interrupt_info *inti); | |
841b91c5 | 233 | int kvm_s390_mask_adapter(struct kvm *kvm, unsigned int id, bool masked); |
453423dc | 234 | |
04b41acd | 235 | /* implemented in intercept.c */ |
0e8bc06a | 236 | u8 kvm_s390_get_ilen(struct kvm_vcpu *vcpu); |
04b41acd | 237 | int kvm_handle_sie_intercept(struct kvm_vcpu *vcpu); |
0e8bc06a DH |
238 | static inline void kvm_s390_rewind_psw(struct kvm_vcpu *vcpu, int ilen) |
239 | { | |
240 | struct kvm_s390_sie_block *sie_block = vcpu->arch.sie_block; | |
241 | ||
242 | sie_block->gpsw.addr = __rewind_psw(sie_block->gpsw, ilen); | |
243 | } | |
244 | static inline void kvm_s390_forward_psw(struct kvm_vcpu *vcpu, int ilen) | |
245 | { | |
246 | kvm_s390_rewind_psw(vcpu, -ilen); | |
247 | } | |
248 | static inline void kvm_s390_retry_instr(struct kvm_vcpu *vcpu) | |
249 | { | |
5ffe466c DH |
250 | /* don't inject PER events if we re-execute the instruction */ |
251 | vcpu->arch.sie_block->icptstatus &= ~0x02; | |
0e8bc06a DH |
252 | kvm_s390_rewind_psw(vcpu, kvm_s390_get_ilen(vcpu)); |
253 | } | |
04b41acd | 254 | |
b7c92f1a QH |
255 | int handle_sthyi(struct kvm_vcpu *vcpu); |
256 | ||
453423dc | 257 | /* implemented in priv.c */ |
a3fb577e | 258 | int is_valid_psw(psw_t *psw); |
80cd8763 | 259 | int kvm_s390_handle_aa(struct kvm_vcpu *vcpu); |
70455a36 | 260 | int kvm_s390_handle_b2(struct kvm_vcpu *vcpu); |
4e0b1ab7 | 261 | int kvm_s390_handle_e3(struct kvm_vcpu *vcpu); |
bb25b9ba | 262 | int kvm_s390_handle_e5(struct kvm_vcpu *vcpu); |
8c3f61e2 | 263 | int kvm_s390_handle_01(struct kvm_vcpu *vcpu); |
48a3e950 CH |
264 | int kvm_s390_handle_b9(struct kvm_vcpu *vcpu); |
265 | int kvm_s390_handle_lpsw(struct kvm_vcpu *vcpu); | |
aba07508 | 266 | int kvm_s390_handle_stctl(struct kvm_vcpu *vcpu); |
953ed88d TH |
267 | int kvm_s390_handle_lctl(struct kvm_vcpu *vcpu); |
268 | int kvm_s390_handle_eb(struct kvm_vcpu *vcpu); | |
730cd632 | 269 | int kvm_s390_skey_check_enable(struct kvm_vcpu *vcpu); |
5288fbf0 | 270 | |
a3508fbe DH |
271 | /* implemented in vsie.c */ |
272 | int kvm_s390_handle_vsie(struct kvm_vcpu *vcpu); | |
adbf1698 | 273 | void kvm_s390_vsie_kick(struct kvm_vcpu *vcpu); |
a3508fbe DH |
274 | void kvm_s390_vsie_gmap_notifier(struct gmap *gmap, unsigned long start, |
275 | unsigned long end); | |
276 | void kvm_s390_vsie_init(struct kvm *kvm); | |
277 | void kvm_s390_vsie_destroy(struct kvm *kvm); | |
278 | ||
5288fbf0 CB |
279 | /* implemented in sigp.c */ |
280 | int kvm_s390_handle_sigp(struct kvm_vcpu *vcpu); | |
4953919f | 281 | int kvm_s390_handle_sigp_pei(struct kvm_vcpu *vcpu); |
5288fbf0 CB |
282 | |
283 | /* implemented in kvm-s390.c */ | |
0e7def5f DH |
284 | void kvm_s390_set_tod_clock(struct kvm *kvm, |
285 | const struct kvm_s390_vm_tod_clock *gtod); | |
fa576c58 | 286 | long kvm_arch_fault_in_page(struct kvm_vcpu *vcpu, gpa_t gpa, int writable); |
e879892c TH |
287 | int kvm_s390_store_status_unloaded(struct kvm_vcpu *vcpu, unsigned long addr); |
288 | int kvm_s390_vcpu_store_status(struct kvm_vcpu *vcpu, unsigned long addr); | |
6852d7b6 DH |
289 | void kvm_s390_vcpu_start(struct kvm_vcpu *vcpu); |
290 | void kvm_s390_vcpu_stop(struct kvm_vcpu *vcpu); | |
27406cd5 CB |
291 | void kvm_s390_vcpu_block(struct kvm_vcpu *vcpu); |
292 | void kvm_s390_vcpu_unblock(struct kvm_vcpu *vcpu); | |
49b99e1e | 293 | void exit_sie(struct kvm_vcpu *vcpu); |
8e236546 | 294 | void kvm_s390_sync_request(int req, struct kvm_vcpu *vcpu); |
b31605c1 DD |
295 | int kvm_s390_vcpu_setup_cmma(struct kvm_vcpu *vcpu); |
296 | void kvm_s390_vcpu_unsetup_cmma(struct kvm_vcpu *vcpu); | |
4287f247 DH |
297 | void kvm_s390_set_cpu_timer(struct kvm_vcpu *vcpu, __u64 cputm); |
298 | __u64 kvm_s390_get_cpu_timer(struct kvm_vcpu *vcpu); | |
b31605c1 | 299 | |
e28acfea CB |
300 | /* implemented in diag.c */ |
301 | int kvm_s390_handle_diag(struct kvm_vcpu *vcpu); | |
302 | ||
27406cd5 CB |
303 | static inline void kvm_s390_vcpu_block_all(struct kvm *kvm) |
304 | { | |
305 | int i; | |
306 | struct kvm_vcpu *vcpu; | |
307 | ||
308 | WARN_ON(!mutex_is_locked(&kvm->lock)); | |
309 | kvm_for_each_vcpu(i, vcpu, kvm) | |
310 | kvm_s390_vcpu_block(vcpu); | |
311 | } | |
312 | ||
313 | static inline void kvm_s390_vcpu_unblock_all(struct kvm *kvm) | |
314 | { | |
315 | int i; | |
316 | struct kvm_vcpu *vcpu; | |
317 | ||
318 | kvm_for_each_vcpu(i, vcpu, kvm) | |
319 | kvm_s390_vcpu_unblock(vcpu); | |
320 | } | |
321 | ||
60417fcc DH |
322 | static inline u64 kvm_s390_get_tod_clock_fast(struct kvm *kvm) |
323 | { | |
324 | u64 rc; | |
325 | ||
326 | preempt_disable(); | |
327 | rc = get_tod_clock_fast() + kvm->arch.epoch; | |
328 | preempt_enable(); | |
329 | return rc; | |
330 | } | |
331 | ||
1b0462e5 HC |
332 | /** |
333 | * kvm_s390_inject_prog_cond - conditionally inject a program check | |
334 | * @vcpu: virtual cpu | |
335 | * @rc: original return/error code | |
336 | * | |
337 | * This function is supposed to be used after regular guest access functions | |
338 | * failed, to conditionally inject a program check to a vcpu. The typical | |
339 | * pattern would look like | |
340 | * | |
341 | * rc = write_guest(vcpu, addr, data, len); | |
342 | * if (rc) | |
343 | * return kvm_s390_inject_prog_cond(vcpu, rc); | |
344 | * | |
345 | * A negative return code from guest access functions implies an internal error | |
346 | * like e.g. out of memory. In these cases no program check should be injected | |
347 | * to the guest. | |
348 | * A positive value implies that an exception happened while accessing a guest's | |
349 | * memory. In this case all data belonging to the corresponding program check | |
350 | * has been stored in vcpu->arch.pgm and can be injected with | |
351 | * kvm_s390_inject_prog_irq(). | |
352 | * | |
353 | * Returns: - the original @rc value if @rc was negative (internal error) | |
354 | * - zero if @rc was already zero | |
355 | * - zero or error code from injecting if @rc was positive | |
356 | * (program check injected to @vcpu) | |
357 | */ | |
358 | static inline int kvm_s390_inject_prog_cond(struct kvm_vcpu *vcpu, int rc) | |
359 | { | |
360 | if (rc <= 0) | |
361 | return rc; | |
362 | return kvm_s390_inject_prog_irq(vcpu, &vcpu->arch.pgm); | |
363 | } | |
364 | ||
383d0b05 JF |
365 | int s390int_to_s390irq(struct kvm_s390_interrupt *s390int, |
366 | struct kvm_s390_irq *s390irq); | |
367 | ||
3c038e6b | 368 | /* implemented in interrupt.c */ |
9a022067 | 369 | int kvm_s390_vcpu_has_irq(struct kvm_vcpu *vcpu, int exclude_stop); |
3c038e6b | 370 | int psw_extint_disabled(struct kvm_vcpu *vcpu); |
841b91c5 | 371 | void kvm_s390_destroy_adapters(struct kvm *kvm); |
ea5f4969 | 372 | int kvm_s390_ext_call_pending(struct kvm_vcpu *vcpu); |
84877d93 | 373 | extern struct kvm_device_ops kvm_flic_ops; |
6cddd432 DH |
374 | int kvm_s390_is_stop_irq_pending(struct kvm_vcpu *vcpu); |
375 | void kvm_s390_clear_stop_irq(struct kvm_vcpu *vcpu); | |
816c7667 JF |
376 | int kvm_s390_set_irq_state(struct kvm_vcpu *vcpu, |
377 | void __user *buf, int len); | |
378 | int kvm_s390_get_irq_state(struct kvm_vcpu *vcpu, | |
379 | __u8 __user *buf, int len); | |
d7c5cb01 MM |
380 | void kvm_s390_gisa_init(struct kvm *kvm); |
381 | void kvm_s390_gisa_clear(struct kvm *kvm); | |
382 | void kvm_s390_gisa_destroy(struct kvm *kvm); | |
3c038e6b | 383 | |
27291e21 DH |
384 | /* implemented in guestdbg.c */ |
385 | void kvm_s390_backup_guest_per_regs(struct kvm_vcpu *vcpu); | |
386 | void kvm_s390_restore_guest_per_regs(struct kvm_vcpu *vcpu); | |
387 | void kvm_s390_patch_guest_per_regs(struct kvm_vcpu *vcpu); | |
388 | int kvm_s390_import_bp_data(struct kvm_vcpu *vcpu, | |
389 | struct kvm_guest_debug *dbg); | |
390 | void kvm_s390_clear_bp_data(struct kvm_vcpu *vcpu); | |
391 | void kvm_s390_prepare_debug_exit(struct kvm_vcpu *vcpu); | |
5ffe466c | 392 | int kvm_s390_handle_per_ifetch_icpt(struct kvm_vcpu *vcpu); |
a69cbe81 | 393 | int kvm_s390_handle_per_event(struct kvm_vcpu *vcpu); |
27291e21 | 394 | |
60514510 ED |
395 | /* support for Basic/Extended SCA handling */ |
396 | static inline union ipte_control *kvm_s390_get_ipte_control(struct kvm *kvm) | |
397 | { | |
bc784cce ED |
398 | struct bsca_block *sca = kvm->arch.sca; /* SCA version doesn't matter */ |
399 | ||
400 | return &sca->ipte_control; | |
60514510 | 401 | } |
a6940674 DH |
402 | static inline int kvm_s390_use_sca_entries(void) |
403 | { | |
404 | /* | |
405 | * Without SIGP interpretation, only SRS interpretation (if available) | |
406 | * might use the entries. By not setting the entries and keeping them | |
407 | * invalid, hardware will not access them but intercept. | |
408 | */ | |
409 | return sclp.has_sigpif; | |
410 | } | |
4d62fcc0 QH |
411 | void kvm_s390_reinject_machine_check(struct kvm_vcpu *vcpu, |
412 | struct mcck_volatile_info *mcck_info); | |
b0c632db | 413 | #endif |