1 // SPDX-License-Identifier: GPL-2.0-only
5 * Copyright (C) 2021, Intel, Inc.
7 * Tests for amx #NM exception and save/restore.
10 #define _GNU_SOURCE /* for program_invocation_short_name */
15 #include <sys/ioctl.h>
16 #include <sys/syscall.h>
18 #include "test_util.h"
21 #include "processor.h"
25 # error This test is 64-bit only
29 #define TILE_SIZE 1024
30 #define XSAVE_SIZE ((NUM_TILES * TILE_SIZE) + PAGE_SIZE)
32 /* Tile configuration associated: */
33 #define PALETTE_TABLE_INDEX 1
35 #define RESERVED_BYTES 14
37 #define XFEATURE_XTILECFG 17
38 #define XFEATURE_XTILEDATA 18
39 #define XFEATURE_MASK_XTILECFG (1 << XFEATURE_XTILECFG)
40 #define XFEATURE_MASK_XTILEDATA (1 << XFEATURE_XTILEDATA)
41 #define XFEATURE_MASK_XTILE (XFEATURE_MASK_XTILECFG | XFEATURE_MASK_XTILEDATA)
43 #define XSAVE_HDR_OFFSET 512
48 u8 reserved[RESERVED_BYTES];
54 u8 data[NUM_TILES * TILE_SIZE];
66 static struct xtile_info xtile;
68 static inline void __ldtilecfg(void *cfg)
70 asm volatile(".byte 0xc4,0xe2,0x78,0x49,0x00"
74 static inline void __tileloadd(void *tile)
76 asm volatile(".byte 0xc4,0xe2,0x7b,0x4b,0x04,0x10"
77 : : "a"(tile), "d"(0));
80 static inline void __tilerelease(void)
82 asm volatile(".byte 0xc4, 0xe2, 0x78, 0x49, 0xc0" ::);
85 static inline void __xsavec(struct xstate *xstate, uint64_t rfbm)
87 uint32_t rfbm_lo = rfbm;
88 uint32_t rfbm_hi = rfbm >> 32;
90 asm volatile("xsavec (%%rdi)"
91 : : "D" (xstate), "a" (rfbm_lo), "d" (rfbm_hi)
95 static void check_xtile_info(void)
97 GUEST_ASSERT(this_cpu_has_p(X86_PROPERTY_XSTATE_MAX_SIZE_XCR0));
98 GUEST_ASSERT(this_cpu_property(X86_PROPERTY_XSTATE_MAX_SIZE_XCR0) <= XSAVE_SIZE);
100 xtile.xsave_offset = this_cpu_property(X86_PROPERTY_XSTATE_TILE_OFFSET);
101 GUEST_ASSERT(xtile.xsave_offset == 2816);
102 xtile.xsave_size = this_cpu_property(X86_PROPERTY_XSTATE_TILE_SIZE);
103 GUEST_ASSERT(xtile.xsave_size == 8192);
104 GUEST_ASSERT(sizeof(struct tile_data) >= xtile.xsave_size);
106 GUEST_ASSERT(this_cpu_has_p(X86_PROPERTY_AMX_MAX_PALETTE_TABLES));
107 GUEST_ASSERT(this_cpu_property(X86_PROPERTY_AMX_MAX_PALETTE_TABLES) >=
108 PALETTE_TABLE_INDEX);
110 GUEST_ASSERT(this_cpu_has_p(X86_PROPERTY_AMX_NR_TILE_REGS));
111 xtile.max_names = this_cpu_property(X86_PROPERTY_AMX_NR_TILE_REGS);
112 GUEST_ASSERT(xtile.max_names == 8);
113 xtile.bytes_per_tile = this_cpu_property(X86_PROPERTY_AMX_BYTES_PER_TILE);
114 GUEST_ASSERT(xtile.bytes_per_tile == 1024);
115 xtile.bytes_per_row = this_cpu_property(X86_PROPERTY_AMX_BYTES_PER_ROW);
116 GUEST_ASSERT(xtile.bytes_per_row == 64);
117 xtile.max_rows = this_cpu_property(X86_PROPERTY_AMX_MAX_ROWS);
118 GUEST_ASSERT(xtile.max_rows == 16);
121 static void set_tilecfg(struct tile_config *cfg)
125 /* Only palette id 1 */
127 for (i = 0; i < xtile.max_names; i++) {
128 cfg->colsb[i] = xtile.bytes_per_row;
129 cfg->rows[i] = xtile.max_rows;
133 static void init_regs(void)
137 GUEST_ASSERT(this_cpu_has(X86_FEATURE_XSAVE));
139 /* turn on CR4.OSXSAVE */
141 cr4 |= X86_CR4_OSXSAVE;
143 GUEST_ASSERT(this_cpu_has(X86_FEATURE_OSXSAVE));
146 xcr0 |= XFEATURE_MASK_XTILE;
148 GUEST_ASSERT((xgetbv(0) & XFEATURE_MASK_XTILE) == XFEATURE_MASK_XTILE);
151 static void __attribute__((__flatten__)) guest_code(struct tile_config *amx_cfg,
152 struct tile_data *tiledata,
153 struct xstate *xstate)
159 /* xfd=0, enable amx */
160 wrmsr(MSR_IA32_XFD, 0);
162 GUEST_ASSERT(rdmsr(MSR_IA32_XFD) == 0);
163 set_tilecfg(amx_cfg);
164 __ldtilecfg(amx_cfg);
166 /* Check save/restore when trap to userspace */
167 __tileloadd(tiledata);
172 * After XSAVEC, XTILEDATA is cleared in the xstate_bv but is set in
175 xstate->header.xstate_bv = XFEATURE_MASK_XTILEDATA;
176 __xsavec(xstate, XFEATURE_MASK_XTILEDATA);
177 GUEST_ASSERT(!(xstate->header.xstate_bv & XFEATURE_MASK_XTILEDATA));
178 GUEST_ASSERT(xstate->header.xcomp_bv & XFEATURE_MASK_XTILEDATA);
180 /* xfd=0x40000, disable amx tiledata */
181 wrmsr(MSR_IA32_XFD, XFEATURE_MASK_XTILEDATA);
184 * XTILEDATA is cleared in xstate_bv but set in xcomp_bv, this property
185 * remains the same even when amx tiledata is disabled by IA32_XFD.
187 xstate->header.xstate_bv = XFEATURE_MASK_XTILEDATA;
188 __xsavec(xstate, XFEATURE_MASK_XTILEDATA);
189 GUEST_ASSERT(!(xstate->header.xstate_bv & XFEATURE_MASK_XTILEDATA));
190 GUEST_ASSERT((xstate->header.xcomp_bv & XFEATURE_MASK_XTILEDATA));
193 GUEST_ASSERT(rdmsr(MSR_IA32_XFD) == XFEATURE_MASK_XTILEDATA);
194 set_tilecfg(amx_cfg);
195 __ldtilecfg(amx_cfg);
196 /* Trigger #NM exception */
197 __tileloadd(tiledata);
203 void guest_nm_handler(struct ex_regs *regs)
205 /* Check if #NM is triggered by XFEATURE_MASK_XTILEDATA */
207 GUEST_ASSERT(!(get_cr0() & X86_CR0_TS));
208 GUEST_ASSERT(rdmsr(MSR_IA32_XFD_ERR) == XFEATURE_MASK_XTILEDATA);
209 GUEST_ASSERT(rdmsr(MSR_IA32_XFD) == XFEATURE_MASK_XTILEDATA);
211 GUEST_ASSERT(rdmsr(MSR_IA32_XFD_ERR) == XFEATURE_MASK_XTILEDATA);
212 GUEST_ASSERT(rdmsr(MSR_IA32_XFD) == XFEATURE_MASK_XTILEDATA);
214 wrmsr(MSR_IA32_XFD_ERR, 0);
215 /* xfd=0, enable amx */
216 wrmsr(MSR_IA32_XFD, 0);
220 int main(int argc, char *argv[])
222 struct kvm_regs regs1, regs2;
223 struct kvm_vcpu *vcpu;
225 struct kvm_x86_state *state;
226 int xsave_restore_size;
227 vm_vaddr_t amx_cfg, tiledata, xstate;
233 * Note, all off-by-default features must be enabled before anything
234 * caches KVM_GET_SUPPORTED_CPUID, e.g. before using kvm_cpu_has().
236 vm_xsave_require_permission(XSTATE_XTILE_DATA_BIT);
238 TEST_REQUIRE(kvm_cpu_has(X86_FEATURE_XFD));
239 TEST_REQUIRE(kvm_cpu_has(X86_FEATURE_XSAVE));
240 TEST_REQUIRE(kvm_cpu_has(X86_FEATURE_AMX_TILE));
241 TEST_REQUIRE(kvm_cpu_has(X86_FEATURE_XTILECFG));
242 TEST_REQUIRE(kvm_cpu_has(X86_FEATURE_XTILEDATA));
243 TEST_REQUIRE(kvm_cpu_has(X86_FEATURE_XTILEDATA_XFD));
246 vm = vm_create_with_one_vcpu(&vcpu, guest_code);
248 TEST_ASSERT(kvm_cpu_has_p(X86_PROPERTY_XSTATE_MAX_SIZE),
249 "KVM should enumerate max XSAVE size when XSAVE is supported");
250 xsave_restore_size = kvm_cpu_property(X86_PROPERTY_XSTATE_MAX_SIZE);
252 vcpu_regs_get(vcpu, ®s1);
254 /* Register #NM handler */
255 vm_init_descriptor_tables(vm);
256 vcpu_init_descriptor_tables(vcpu);
257 vm_install_exception_handler(vm, NM_VECTOR, guest_nm_handler);
259 /* amx cfg for guest_code */
260 amx_cfg = vm_vaddr_alloc_page(vm);
261 memset(addr_gva2hva(vm, amx_cfg), 0x0, getpagesize());
263 /* amx tiledata for guest_code */
264 tiledata = vm_vaddr_alloc_pages(vm, 2);
265 memset(addr_gva2hva(vm, tiledata), rand() | 1, 2 * getpagesize());
267 /* XSAVE state for guest_code */
268 xstate = vm_vaddr_alloc_pages(vm, DIV_ROUND_UP(XSAVE_SIZE, PAGE_SIZE));
269 memset(addr_gva2hva(vm, xstate), 0, PAGE_SIZE * DIV_ROUND_UP(XSAVE_SIZE, PAGE_SIZE));
270 vcpu_args_set(vcpu, 3, amx_cfg, tiledata, xstate);
272 for (stage = 1; ; stage++) {
274 TEST_ASSERT_KVM_EXIT_REASON(vcpu, KVM_EXIT_IO);
276 switch (get_ucall(vcpu, &uc)) {
278 REPORT_GUEST_ASSERT(uc);
281 switch (uc.args[1]) {
289 fprintf(stderr, "GUEST_SYNC(%ld)\n", uc.args[1]);
294 "GUEST_SYNC(%ld), check save/restore status\n", uc.args[1]);
296 /* Compacted mode, get amx offset by xsave area
297 * size subtract 8K amx size.
299 amx_offset = xsave_restore_size - NUM_TILES*TILE_SIZE;
300 state = vcpu_save_state(vcpu);
301 void *amx_start = (void *)state->xsave + amx_offset;
302 void *tiles_data = (void *)addr_gva2hva(vm, tiledata);
303 /* Only check TMM0 register, 1 tile */
304 ret = memcmp(amx_start, tiles_data, TILE_SIZE);
305 TEST_ASSERT(ret == 0, "memcmp failed, ret=%d\n", ret);
306 kvm_x86_state_cleanup(state);
310 "GUEST_SYNC(%ld), #NM exception and enable amx\n", uc.args[1]);
315 fprintf(stderr, "UCALL_DONE\n");
318 TEST_FAIL("Unknown ucall %lu", uc.cmd);
321 state = vcpu_save_state(vcpu);
322 memset(®s1, 0, sizeof(regs1));
323 vcpu_regs_get(vcpu, ®s1);
327 /* Restore state in a new VM. */
328 vcpu = vm_recreate_with_one_vcpu(vm);
329 vcpu_load_state(vcpu, state);
330 kvm_x86_state_cleanup(state);
332 memset(®s2, 0, sizeof(regs2));
333 vcpu_regs_get(vcpu, ®s2);
334 TEST_ASSERT(!memcmp(®s1, ®s2, sizeof(regs2)),
335 "Unexpected register values after vcpu_load_state; rdi: %lx rsi: %lx",
336 (ulong) regs2.rdi, (ulong) regs2.rsi);