Commit | Line | Data |
---|---|---|
dff96888 | 1 | // SPDX-License-Identifier: GPL-2.0 OR MIT |
d80efd5c TH |
2 | /************************************************************************** |
3 | * | |
09881d29 | 4 | * Copyright 2014-2023 VMware, Inc., Palo Alto, CA., USA |
d80efd5c TH |
5 | * |
6 | * Permission is hereby granted, free of charge, to any person obtaining a | |
7 | * copy of this software and associated documentation files (the | |
8 | * "Software"), to deal in the Software without restriction, including | |
9 | * without limitation the rights to use, copy, modify, merge, publish, | |
10 | * distribute, sub license, and/or sell copies of the Software, and to | |
11 | * permit persons to whom the Software is furnished to do so, subject to | |
12 | * the following conditions: | |
13 | * | |
14 | * The above copyright notice and this permission notice (including the | |
15 | * next paragraph) shall be included in all copies or substantial portions | |
16 | * of the Software. | |
17 | * | |
18 | * THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR | |
19 | * IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY, | |
20 | * FITNESS FOR A PARTICULAR PURPOSE AND NON-INFRINGEMENT. IN NO EVENT SHALL | |
21 | * THE COPYRIGHT HOLDERS, AUTHORS AND/OR ITS SUPPLIERS BE LIABLE FOR ANY CLAIM, | |
22 | * DAMAGES OR OTHER LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR | |
23 | * OTHERWISE, ARISING FROM, OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE | |
24 | * USE OR OTHER DEALINGS IN THE SOFTWARE. | |
25 | * | |
26 | **************************************************************************/ | |
27 | /* | |
28 | * Treat context OTables as resources to make use of the resource | |
29 | * backing MOB eviction mechanism, that is used to read back the COTable | |
30 | * whenever the backing MOB is evicted. | |
31 | */ | |
32 | ||
09881d29 | 33 | #include "vmwgfx_bo.h" |
d80efd5c | 34 | #include "vmwgfx_drv.h" |
4bb50606 | 35 | #include "vmwgfx_mksstat.h" |
d80efd5c | 36 | #include "vmwgfx_resource_priv.h" |
d80efd5c TH |
37 | #include "vmwgfx_so.h" |
38 | ||
09881d29 ZR |
39 | #include <drm/ttm/ttm_placement.h> |
40 | ||
d80efd5c TH |
41 | /** |
42 | * struct vmw_cotable - Context Object Table resource | |
43 | * | |
44 | * @res: struct vmw_resource we are deriving from. | |
45 | * @ctx: non-refcounted pointer to the owning context. | |
46 | * @size_read_back: Size of data read back during eviction. | |
47 | * @seen_entries: Seen entries in command stream for this cotable. | |
48 | * @type: The cotable type. | |
49 | * @scrubbed: Whether the cotable has been scrubbed. | |
50 | * @resource_list: List of resources in the cotable. | |
51 | */ | |
52 | struct vmw_cotable { | |
53 | struct vmw_resource res; | |
54 | struct vmw_resource *ctx; | |
55 | size_t size_read_back; | |
56 | int seen_entries; | |
57 | u32 type; | |
58 | bool scrubbed; | |
59 | struct list_head resource_list; | |
60 | }; | |
61 | ||
62 | /** | |
63 | * struct vmw_cotable_info - Static info about cotable types | |
64 | * | |
65 | * @min_initial_entries: Min number of initial intries at cotable allocation | |
66 | * for this cotable type. | |
67 | * @size: Size of each entry. | |
3894709e | 68 | * @unbind_func: Unbind call-back function. |
d80efd5c TH |
69 | */ |
70 | struct vmw_cotable_info { | |
71 | u32 min_initial_entries; | |
72 | u32 size; | |
73 | void (*unbind_func)(struct vmw_private *, struct list_head *, | |
74 | bool); | |
75 | }; | |
76 | ||
148e5f55 ZR |
77 | |
78 | /* | |
79 | * Getting the initial size right is difficult because it all depends | |
80 | * on what the userspace is doing. The sizes will be aligned up to | |
81 | * a PAGE_SIZE so we just want to make sure that for majority of apps | |
82 | * the initial number of entries doesn't require an immediate resize. | |
83 | * For all cotables except SVGACOTableDXElementLayoutEntry and | |
84 | * SVGACOTableDXBlendStateEntry the initial number of entries fits | |
85 | * within the PAGE_SIZE. For SVGACOTableDXElementLayoutEntry and | |
86 | * SVGACOTableDXBlendStateEntry we want to reserve two pages, | |
87 | * because that's what all apps will require initially. | |
88 | */ | |
d80efd5c TH |
89 | static const struct vmw_cotable_info co_info[] = { |
90 | {1, sizeof(SVGACOTableDXRTViewEntry), &vmw_view_cotable_list_destroy}, | |
91 | {1, sizeof(SVGACOTableDXDSViewEntry), &vmw_view_cotable_list_destroy}, | |
92 | {1, sizeof(SVGACOTableDXSRViewEntry), &vmw_view_cotable_list_destroy}, | |
148e5f55 ZR |
93 | {PAGE_SIZE/sizeof(SVGACOTableDXElementLayoutEntry) + 1, sizeof(SVGACOTableDXElementLayoutEntry), NULL}, |
94 | {PAGE_SIZE/sizeof(SVGACOTableDXBlendStateEntry) + 1, sizeof(SVGACOTableDXBlendStateEntry), NULL}, | |
d80efd5c TH |
95 | {1, sizeof(SVGACOTableDXDepthStencilEntry), NULL}, |
96 | {1, sizeof(SVGACOTableDXRasterizerStateEntry), NULL}, | |
97 | {1, sizeof(SVGACOTableDXSamplerEntry), NULL}, | |
e8bead9c | 98 | {1, sizeof(SVGACOTableDXStreamOutputEntry), &vmw_dx_streamoutput_cotable_list_scrub}, |
d80efd5c | 99 | {1, sizeof(SVGACOTableDXQueryEntry), NULL}, |
5e8ec0d9 DR |
100 | {1, sizeof(SVGACOTableDXShaderEntry), &vmw_dx_shader_cotable_list_scrub}, |
101 | {1, sizeof(SVGACOTableDXUAViewEntry), &vmw_view_cotable_list_destroy} | |
d80efd5c TH |
102 | }; |
103 | ||
104 | /* | |
105 | * Cotables with bindings that we remove must be scrubbed first, | |
106 | * otherwise, the device will swap in an invalid context when we remove | |
107 | * bindings before scrubbing a cotable... | |
108 | */ | |
109 | const SVGACOTableType vmw_cotable_scrub_order[] = { | |
110 | SVGA_COTABLE_RTVIEW, | |
111 | SVGA_COTABLE_DSVIEW, | |
112 | SVGA_COTABLE_SRVIEW, | |
113 | SVGA_COTABLE_DXSHADER, | |
114 | SVGA_COTABLE_ELEMENTLAYOUT, | |
115 | SVGA_COTABLE_BLENDSTATE, | |
116 | SVGA_COTABLE_DEPTHSTENCIL, | |
117 | SVGA_COTABLE_RASTERIZERSTATE, | |
118 | SVGA_COTABLE_SAMPLER, | |
119 | SVGA_COTABLE_STREAMOUTPUT, | |
120 | SVGA_COTABLE_DXQUERY, | |
5e8ec0d9 | 121 | SVGA_COTABLE_UAVIEW, |
d80efd5c TH |
122 | }; |
123 | ||
124 | static int vmw_cotable_bind(struct vmw_resource *res, | |
125 | struct ttm_validate_buffer *val_buf); | |
126 | static int vmw_cotable_unbind(struct vmw_resource *res, | |
127 | bool readback, | |
128 | struct ttm_validate_buffer *val_buf); | |
129 | static int vmw_cotable_create(struct vmw_resource *res); | |
130 | static int vmw_cotable_destroy(struct vmw_resource *res); | |
131 | ||
132 | static const struct vmw_res_func vmw_cotable_func = { | |
133 | .res_type = vmw_res_cotable, | |
668b2066 | 134 | .needs_guest_memory = true, |
d80efd5c | 135 | .may_evict = true, |
a0a63940 TH |
136 | .prio = 3, |
137 | .dirty_prio = 3, | |
d80efd5c | 138 | .type_name = "context guest backed object tables", |
39985eea ZR |
139 | .domain = VMW_BO_DOMAIN_MOB, |
140 | .busy_domain = VMW_BO_DOMAIN_MOB, | |
d80efd5c TH |
141 | .create = vmw_cotable_create, |
142 | .destroy = vmw_cotable_destroy, | |
143 | .bind = vmw_cotable_bind, | |
144 | .unbind = vmw_cotable_unbind, | |
145 | }; | |
146 | ||
147 | /** | |
148 | * vmw_cotable - Convert a struct vmw_resource pointer to a struct | |
149 | * vmw_cotable pointer | |
150 | * | |
151 | * @res: Pointer to the resource. | |
152 | */ | |
153 | static struct vmw_cotable *vmw_cotable(struct vmw_resource *res) | |
154 | { | |
155 | return container_of(res, struct vmw_cotable, res); | |
156 | } | |
157 | ||
158 | /** | |
159 | * vmw_cotable_destroy - Cotable resource destroy callback | |
160 | * | |
161 | * @res: Pointer to the cotable resource. | |
162 | * | |
163 | * There is no device cotable destroy command, so this function only | |
164 | * makes sure that the resource id is set to invalid. | |
165 | */ | |
166 | static int vmw_cotable_destroy(struct vmw_resource *res) | |
167 | { | |
168 | res->id = -1; | |
169 | return 0; | |
170 | } | |
171 | ||
172 | /** | |
173 | * vmw_cotable_unscrub - Undo a cotable unscrub operation | |
174 | * | |
175 | * @res: Pointer to the cotable resource | |
176 | * | |
177 | * This function issues commands to (re)bind the cotable to | |
178 | * its backing mob, which needs to be validated and reserved at this point. | |
179 | * This is identical to bind() except the function interface looks different. | |
180 | */ | |
181 | static int vmw_cotable_unscrub(struct vmw_resource *res) | |
182 | { | |
183 | struct vmw_cotable *vcotbl = vmw_cotable(res); | |
184 | struct vmw_private *dev_priv = res->dev_priv; | |
668b2066 | 185 | struct ttm_buffer_object *bo = &res->guest_memory_bo->tbo; |
d80efd5c TH |
186 | struct { |
187 | SVGA3dCmdHeader header; | |
188 | SVGA3dCmdDXSetCOTable body; | |
189 | } *cmd; | |
190 | ||
d3116756 | 191 | WARN_ON_ONCE(bo->resource->mem_type != VMW_PL_MOB); |
52791eee | 192 | dma_resv_assert_held(bo->base.resv); |
d80efd5c | 193 | |
8426ed9c | 194 | cmd = VMW_CMD_RESERVE(dev_priv, sizeof(*cmd)); |
11c45419 | 195 | if (!cmd) |
d80efd5c | 196 | return -ENOMEM; |
d80efd5c TH |
197 | |
198 | WARN_ON(vcotbl->ctx->id == SVGA3D_INVALID_ID); | |
d3116756 | 199 | WARN_ON(bo->resource->mem_type != VMW_PL_MOB); |
d80efd5c TH |
200 | cmd->header.id = SVGA_3D_CMD_DX_SET_COTABLE; |
201 | cmd->header.size = sizeof(cmd->body); | |
202 | cmd->body.cid = vcotbl->ctx->id; | |
203 | cmd->body.type = vcotbl->type; | |
d3116756 | 204 | cmd->body.mobid = bo->resource->start; |
d80efd5c TH |
205 | cmd->body.validSizeInBytes = vcotbl->size_read_back; |
206 | ||
8426ed9c | 207 | vmw_cmd_commit_flush(dev_priv, sizeof(*cmd)); |
d80efd5c TH |
208 | vcotbl->scrubbed = false; |
209 | ||
210 | return 0; | |
211 | } | |
212 | ||
213 | /** | |
214 | * vmw_cotable_bind - Undo a cotable unscrub operation | |
215 | * | |
216 | * @res: Pointer to the cotable resource | |
217 | * @val_buf: Pointer to a struct ttm_validate_buffer prepared by the caller | |
218 | * for convenience / fencing. | |
219 | * | |
220 | * This function issues commands to (re)bind the cotable to | |
221 | * its backing mob, which needs to be validated and reserved at this point. | |
222 | */ | |
223 | static int vmw_cotable_bind(struct vmw_resource *res, | |
224 | struct ttm_validate_buffer *val_buf) | |
225 | { | |
226 | /* | |
227 | * The create() callback may have changed @res->backup without | |
228 | * the caller noticing, and with val_buf->bo still pointing to | |
229 | * the old backup buffer. Although hackish, and not used currently, | |
230 | * take the opportunity to correct the value here so that it's not | |
231 | * misused in the future. | |
232 | */ | |
668b2066 | 233 | val_buf->bo = &res->guest_memory_bo->tbo; |
d80efd5c TH |
234 | |
235 | return vmw_cotable_unscrub(res); | |
236 | } | |
237 | ||
238 | /** | |
239 | * vmw_cotable_scrub - Scrub the cotable from the device. | |
240 | * | |
241 | * @res: Pointer to the cotable resource. | |
242 | * @readback: Whether initiate a readback of the cotable data to the backup | |
243 | * buffer. | |
244 | * | |
245 | * In some situations (context swapouts) it might be desirable to make the | |
246 | * device forget about the cotable without performing a full unbind. A full | |
247 | * unbind requires reserved backup buffers and it might not be possible to | |
248 | * reserve them due to locking order violation issues. The vmw_cotable_scrub | |
249 | * function implements a partial unbind() without that requirement but with the | |
250 | * following restrictions. | |
251 | * 1) Before the cotable is again used by the GPU, vmw_cotable_unscrub() must | |
252 | * be called. | |
253 | * 2) Before the cotable backing buffer is used by the CPU, or during the | |
254 | * resource destruction, vmw_cotable_unbind() must be called. | |
255 | */ | |
256 | int vmw_cotable_scrub(struct vmw_resource *res, bool readback) | |
257 | { | |
258 | struct vmw_cotable *vcotbl = vmw_cotable(res); | |
259 | struct vmw_private *dev_priv = res->dev_priv; | |
260 | size_t submit_size; | |
261 | ||
262 | struct { | |
263 | SVGA3dCmdHeader header; | |
264 | SVGA3dCmdDXReadbackCOTable body; | |
265 | } *cmd0; | |
266 | struct { | |
267 | SVGA3dCmdHeader header; | |
268 | SVGA3dCmdDXSetCOTable body; | |
269 | } *cmd1; | |
270 | ||
271 | if (vcotbl->scrubbed) | |
272 | return 0; | |
273 | ||
274 | if (co_info[vcotbl->type].unbind_func) | |
275 | co_info[vcotbl->type].unbind_func(dev_priv, | |
276 | &vcotbl->resource_list, | |
277 | readback); | |
278 | submit_size = sizeof(*cmd1); | |
279 | if (readback) | |
280 | submit_size += sizeof(*cmd0); | |
281 | ||
8426ed9c | 282 | cmd1 = VMW_CMD_RESERVE(dev_priv, submit_size); |
11c45419 | 283 | if (!cmd1) |
d80efd5c | 284 | return -ENOMEM; |
d80efd5c TH |
285 | |
286 | vcotbl->size_read_back = 0; | |
287 | if (readback) { | |
288 | cmd0 = (void *) cmd1; | |
289 | cmd0->header.id = SVGA_3D_CMD_DX_READBACK_COTABLE; | |
290 | cmd0->header.size = sizeof(cmd0->body); | |
291 | cmd0->body.cid = vcotbl->ctx->id; | |
292 | cmd0->body.type = vcotbl->type; | |
293 | cmd1 = (void *) &cmd0[1]; | |
668b2066 | 294 | vcotbl->size_read_back = res->guest_memory_size; |
d80efd5c TH |
295 | } |
296 | cmd1->header.id = SVGA_3D_CMD_DX_SET_COTABLE; | |
297 | cmd1->header.size = sizeof(cmd1->body); | |
298 | cmd1->body.cid = vcotbl->ctx->id; | |
299 | cmd1->body.type = vcotbl->type; | |
300 | cmd1->body.mobid = SVGA3D_INVALID_ID; | |
301 | cmd1->body.validSizeInBytes = 0; | |
8426ed9c | 302 | vmw_cmd_commit_flush(dev_priv, submit_size); |
d80efd5c TH |
303 | vcotbl->scrubbed = true; |
304 | ||
305 | /* Trigger a create() on next validate. */ | |
306 | res->id = -1; | |
307 | ||
308 | return 0; | |
309 | } | |
310 | ||
311 | /** | |
312 | * vmw_cotable_unbind - Cotable resource unbind callback | |
313 | * | |
314 | * @res: Pointer to the cotable resource. | |
315 | * @readback: Whether to read back cotable data to the backup buffer. | |
3894709e | 316 | * @val_buf: Pointer to a struct ttm_validate_buffer prepared by the caller |
d80efd5c TH |
317 | * for convenience / fencing. |
318 | * | |
319 | * Unbinds the cotable from the device and fences the backup buffer. | |
320 | */ | |
321 | static int vmw_cotable_unbind(struct vmw_resource *res, | |
322 | bool readback, | |
323 | struct ttm_validate_buffer *val_buf) | |
324 | { | |
325 | struct vmw_cotable *vcotbl = vmw_cotable(res); | |
326 | struct vmw_private *dev_priv = res->dev_priv; | |
327 | struct ttm_buffer_object *bo = val_buf->bo; | |
328 | struct vmw_fence_obj *fence; | |
d80efd5c | 329 | |
a0a63940 | 330 | if (!vmw_resource_mob_attached(res)) |
d80efd5c TH |
331 | return 0; |
332 | ||
d3116756 | 333 | WARN_ON_ONCE(bo->resource->mem_type != VMW_PL_MOB); |
52791eee | 334 | dma_resv_assert_held(bo->base.resv); |
d80efd5c TH |
335 | |
336 | mutex_lock(&dev_priv->binding_mutex); | |
337 | if (!vcotbl->scrubbed) | |
338 | vmw_dx_context_scrub_cotables(vcotbl->ctx, readback); | |
339 | mutex_unlock(&dev_priv->binding_mutex); | |
340 | (void) vmw_execbuf_fence_commands(NULL, dev_priv, &fence, NULL); | |
e9431ea5 | 341 | vmw_bo_fence_single(bo, fence); |
d80efd5c TH |
342 | if (likely(fence != NULL)) |
343 | vmw_fence_obj_unreference(&fence); | |
344 | ||
49558b47 | 345 | return 0; |
d80efd5c TH |
346 | } |
347 | ||
348 | /** | |
349 | * vmw_cotable_readback - Read back a cotable without unbinding. | |
350 | * | |
351 | * @res: The cotable resource. | |
352 | * | |
353 | * Reads back a cotable to its backing mob without scrubbing the MOB from | |
354 | * the cotable. The MOB is fenced for subsequent CPU access. | |
355 | */ | |
356 | static int vmw_cotable_readback(struct vmw_resource *res) | |
357 | { | |
358 | struct vmw_cotable *vcotbl = vmw_cotable(res); | |
359 | struct vmw_private *dev_priv = res->dev_priv; | |
360 | ||
361 | struct { | |
362 | SVGA3dCmdHeader header; | |
363 | SVGA3dCmdDXReadbackCOTable body; | |
364 | } *cmd; | |
365 | struct vmw_fence_obj *fence; | |
366 | ||
367 | if (!vcotbl->scrubbed) { | |
8426ed9c | 368 | cmd = VMW_CMD_RESERVE(dev_priv, sizeof(*cmd)); |
11c45419 | 369 | if (!cmd) |
d80efd5c | 370 | return -ENOMEM; |
11c45419 | 371 | |
d80efd5c TH |
372 | cmd->header.id = SVGA_3D_CMD_DX_READBACK_COTABLE; |
373 | cmd->header.size = sizeof(cmd->body); | |
374 | cmd->body.cid = vcotbl->ctx->id; | |
375 | cmd->body.type = vcotbl->type; | |
668b2066 | 376 | vcotbl->size_read_back = res->guest_memory_size; |
8426ed9c | 377 | vmw_cmd_commit(dev_priv, sizeof(*cmd)); |
d80efd5c TH |
378 | } |
379 | ||
380 | (void) vmw_execbuf_fence_commands(NULL, dev_priv, &fence, NULL); | |
668b2066 | 381 | vmw_bo_fence_single(&res->guest_memory_bo->tbo, fence); |
d80efd5c TH |
382 | vmw_fence_obj_unreference(&fence); |
383 | ||
384 | return 0; | |
385 | } | |
386 | ||
387 | /** | |
388 | * vmw_cotable_resize - Resize a cotable. | |
389 | * | |
390 | * @res: The cotable resource. | |
391 | * @new_size: The new size. | |
392 | * | |
393 | * Resizes a cotable and binds the new backup buffer. | |
394 | * On failure the cotable is left intact. | |
395 | * Important! This function may not fail once the MOB switch has been | |
396 | * committed to hardware. That would put the device context in an | |
397 | * invalid state which we can't currently recover from. | |
398 | */ | |
399 | static int vmw_cotable_resize(struct vmw_resource *res, size_t new_size) | |
400 | { | |
19be5570 | 401 | struct ttm_operation_ctx ctx = { false, false }; |
d80efd5c TH |
402 | struct vmw_private *dev_priv = res->dev_priv; |
403 | struct vmw_cotable *vcotbl = vmw_cotable(res); | |
668b2066 ZR |
404 | struct vmw_bo *buf, *old_buf = res->guest_memory_bo; |
405 | struct ttm_buffer_object *bo, *old_bo = &res->guest_memory_bo->tbo; | |
406 | size_t old_size = res->guest_memory_size; | |
d80efd5c TH |
407 | size_t old_size_read_back = vcotbl->size_read_back; |
408 | size_t cur_size_read_back; | |
409 | struct ttm_bo_kmap_obj old_map, new_map; | |
410 | int ret; | |
411 | size_t i; | |
668b2066 ZR |
412 | struct vmw_bo_params bo_params = { |
413 | .domain = VMW_BO_DOMAIN_MOB, | |
414 | .busy_domain = VMW_BO_DOMAIN_MOB, | |
415 | .bo_type = ttm_bo_type_device, | |
416 | .size = new_size, | |
417 | .pin = true | |
418 | }; | |
d80efd5c | 419 | |
4bb50606 ZR |
420 | MKS_STAT_TIME_DECL(MKSSTAT_KERN_COTABLE_RESIZE); |
421 | MKS_STAT_TIME_PUSH(MKSSTAT_KERN_COTABLE_RESIZE); | |
422 | ||
d80efd5c TH |
423 | ret = vmw_cotable_readback(res); |
424 | if (ret) | |
4bb50606 | 425 | goto out_done; |
d80efd5c TH |
426 | |
427 | cur_size_read_back = vcotbl->size_read_back; | |
428 | vcotbl->size_read_back = old_size_read_back; | |
429 | ||
430 | /* | |
431 | * While device is processing, Allocate and reserve a buffer object | |
432 | * for the new COTable. Initially pin the buffer object to make sure | |
433 | * we can use tryreserve without failure. | |
434 | */ | |
91398b41 | 435 | ret = vmw_gem_object_create(dev_priv, &bo_params, &buf); |
d80efd5c TH |
436 | if (ret) { |
437 | DRM_ERROR("Failed initializing new cotable MOB.\n"); | |
4bb50606 | 438 | goto out_done; |
d80efd5c TH |
439 | } |
440 | ||
668b2066 | 441 | bo = &buf->tbo; |
dfd5e50e | 442 | WARN_ON_ONCE(ttm_bo_reserve(bo, false, true, NULL)); |
d80efd5c | 443 | |
8aa6d4fc | 444 | ret = ttm_bo_wait(old_bo, false, false); |
d80efd5c TH |
445 | if (unlikely(ret != 0)) { |
446 | DRM_ERROR("Failed waiting for cotable unbind.\n"); | |
447 | goto out_wait; | |
448 | } | |
449 | ||
450 | /* | |
451 | * Do a page by page copy of COTables. This eliminates slow vmap()s. | |
452 | * This should really be a TTM utility. | |
453 | */ | |
e3c92eb4 | 454 | for (i = 0; i < PFN_UP(old_bo->resource->size); ++i) { |
d80efd5c TH |
455 | bool dummy; |
456 | ||
457 | ret = ttm_bo_kmap(old_bo, i, 1, &old_map); | |
458 | if (unlikely(ret != 0)) { | |
459 | DRM_ERROR("Failed mapping old COTable on resize.\n"); | |
460 | goto out_wait; | |
461 | } | |
462 | ret = ttm_bo_kmap(bo, i, 1, &new_map); | |
463 | if (unlikely(ret != 0)) { | |
464 | DRM_ERROR("Failed mapping new COTable on resize.\n"); | |
465 | goto out_map_new; | |
466 | } | |
467 | memcpy(ttm_kmap_obj_virtual(&new_map, &dummy), | |
468 | ttm_kmap_obj_virtual(&old_map, &dummy), | |
469 | PAGE_SIZE); | |
470 | ttm_bo_kunmap(&new_map); | |
471 | ttm_bo_kunmap(&old_map); | |
472 | } | |
473 | ||
474 | /* Unpin new buffer, and switch backup buffers. */ | |
39985eea ZR |
475 | vmw_bo_placement_set(buf, |
476 | VMW_BO_DOMAIN_MOB, | |
477 | VMW_BO_DOMAIN_MOB); | |
478 | ret = ttm_bo_validate(bo, &buf->placement, &ctx); | |
d80efd5c TH |
479 | if (unlikely(ret != 0)) { |
480 | DRM_ERROR("Failed validating new COTable backup buffer.\n"); | |
481 | goto out_wait; | |
482 | } | |
483 | ||
a0a63940 | 484 | vmw_resource_mob_detach(res); |
668b2066 ZR |
485 | res->guest_memory_bo = buf; |
486 | res->guest_memory_size = new_size; | |
d80efd5c TH |
487 | vcotbl->size_read_back = cur_size_read_back; |
488 | ||
489 | /* | |
490 | * Now tell the device to switch. If this fails, then we need to | |
491 | * revert the full resize. | |
492 | */ | |
493 | ret = vmw_cotable_unscrub(res); | |
494 | if (ret) { | |
495 | DRM_ERROR("Failed switching COTable backup buffer.\n"); | |
668b2066 ZR |
496 | res->guest_memory_bo = old_buf; |
497 | res->guest_memory_size = old_size; | |
d80efd5c | 498 | vcotbl->size_read_back = old_size_read_back; |
a0a63940 | 499 | vmw_resource_mob_attach(res); |
d80efd5c TH |
500 | goto out_wait; |
501 | } | |
502 | ||
a0a63940 | 503 | vmw_resource_mob_attach(res); |
d80efd5c | 504 | /* Let go of the old mob. */ |
91398b41 | 505 | vmw_user_bo_unref(&old_buf); |
d80efd5c TH |
506 | res->id = vcotbl->type; |
507 | ||
544ccad7 ZR |
508 | ret = dma_resv_reserve_fences(bo->base.resv, 1); |
509 | if (unlikely(ret)) | |
510 | goto out_wait; | |
511 | ||
668b2066 | 512 | /* Release the pin acquired in vmw_bo_create */ |
2ef4fb92 ZR |
513 | ttm_bo_unpin(bo); |
514 | ||
4bb50606 ZR |
515 | MKS_STAT_TIME_POP(MKSSTAT_KERN_COTABLE_RESIZE); |
516 | ||
d80efd5c TH |
517 | return 0; |
518 | ||
519 | out_map_new: | |
520 | ttm_bo_kunmap(&old_map); | |
521 | out_wait: | |
2ef4fb92 | 522 | ttm_bo_unpin(bo); |
d80efd5c | 523 | ttm_bo_unreserve(bo); |
91398b41 | 524 | vmw_user_bo_unref(&buf); |
d80efd5c | 525 | |
4bb50606 ZR |
526 | out_done: |
527 | MKS_STAT_TIME_POP(MKSSTAT_KERN_COTABLE_RESIZE); | |
528 | ||
d80efd5c TH |
529 | return ret; |
530 | } | |
531 | ||
532 | /** | |
533 | * vmw_cotable_create - Cotable resource create callback | |
534 | * | |
535 | * @res: Pointer to a cotable resource. | |
536 | * | |
537 | * There is no separate create command for cotables, so this callback, which | |
538 | * is called before bind() in the validation sequence is instead used for two | |
539 | * things. | |
540 | * 1) Unscrub the cotable if it is scrubbed and still attached to a backup | |
a0a63940 | 541 | * buffer. |
d80efd5c TH |
542 | * 2) Resize the cotable if needed. |
543 | */ | |
544 | static int vmw_cotable_create(struct vmw_resource *res) | |
545 | { | |
546 | struct vmw_cotable *vcotbl = vmw_cotable(res); | |
668b2066 | 547 | size_t new_size = res->guest_memory_size; |
d80efd5c TH |
548 | size_t needed_size; |
549 | int ret; | |
550 | ||
551 | /* Check whether we need to resize the cotable */ | |
552 | needed_size = (vcotbl->seen_entries + 1) * co_info[vcotbl->type].size; | |
553 | while (needed_size > new_size) | |
554 | new_size *= 2; | |
555 | ||
668b2066 | 556 | if (likely(new_size <= res->guest_memory_size)) { |
a0a63940 | 557 | if (vcotbl->scrubbed && vmw_resource_mob_attached(res)) { |
d80efd5c TH |
558 | ret = vmw_cotable_unscrub(res); |
559 | if (ret) | |
560 | return ret; | |
561 | } | |
562 | res->id = vcotbl->type; | |
563 | return 0; | |
564 | } | |
565 | ||
566 | return vmw_cotable_resize(res, new_size); | |
567 | } | |
568 | ||
569 | /** | |
570 | * vmw_hw_cotable_destroy - Cotable hw_destroy callback | |
571 | * | |
572 | * @res: Pointer to a cotable resource. | |
573 | * | |
574 | * The final (part of resource destruction) destroy callback. | |
575 | */ | |
576 | static void vmw_hw_cotable_destroy(struct vmw_resource *res) | |
577 | { | |
578 | (void) vmw_cotable_destroy(res); | |
579 | } | |
580 | ||
d80efd5c TH |
581 | /** |
582 | * vmw_cotable_free - Cotable resource destructor | |
583 | * | |
584 | * @res: Pointer to a cotable resource. | |
585 | */ | |
586 | static void vmw_cotable_free(struct vmw_resource *res) | |
587 | { | |
d80efd5c | 588 | kfree(res); |
d80efd5c TH |
589 | } |
590 | ||
591 | /** | |
592 | * vmw_cotable_alloc - Create a cotable resource | |
593 | * | |
594 | * @dev_priv: Pointer to a device private struct. | |
595 | * @ctx: Pointer to the context resource. | |
596 | * The cotable resource will not add a refcount. | |
597 | * @type: The cotable type. | |
598 | */ | |
599 | struct vmw_resource *vmw_cotable_alloc(struct vmw_private *dev_priv, | |
600 | struct vmw_resource *ctx, | |
601 | u32 type) | |
602 | { | |
603 | struct vmw_cotable *vcotbl; | |
604 | int ret; | |
605 | u32 num_entries; | |
606 | ||
d80efd5c | 607 | vcotbl = kzalloc(sizeof(*vcotbl), GFP_KERNEL); |
1a4adb05 | 608 | if (unlikely(!vcotbl)) { |
d80efd5c TH |
609 | ret = -ENOMEM; |
610 | goto out_no_alloc; | |
611 | } | |
612 | ||
613 | ret = vmw_resource_init(dev_priv, &vcotbl->res, true, | |
614 | vmw_cotable_free, &vmw_cotable_func); | |
615 | if (unlikely(ret != 0)) | |
616 | goto out_no_init; | |
617 | ||
618 | INIT_LIST_HEAD(&vcotbl->resource_list); | |
619 | vcotbl->res.id = type; | |
668b2066 | 620 | vcotbl->res.guest_memory_size = PAGE_SIZE; |
d80efd5c TH |
621 | num_entries = PAGE_SIZE / co_info[type].size; |
622 | if (num_entries < co_info[type].min_initial_entries) { | |
668b2066 | 623 | vcotbl->res.guest_memory_size = co_info[type].min_initial_entries * |
d80efd5c | 624 | co_info[type].size; |
668b2066 | 625 | vcotbl->res.guest_memory_size = PFN_ALIGN(vcotbl->res.guest_memory_size); |
d80efd5c TH |
626 | } |
627 | ||
628 | vcotbl->scrubbed = true; | |
629 | vcotbl->seen_entries = -1; | |
630 | vcotbl->type = type; | |
631 | vcotbl->ctx = ctx; | |
632 | ||
13289241 | 633 | vcotbl->res.hw_destroy = vmw_hw_cotable_destroy; |
d80efd5c TH |
634 | |
635 | return &vcotbl->res; | |
636 | ||
637 | out_no_init: | |
638 | kfree(vcotbl); | |
639 | out_no_alloc: | |
d80efd5c TH |
640 | return ERR_PTR(ret); |
641 | } | |
642 | ||
643 | /** | |
644 | * vmw_cotable_notify - Notify the cotable about an item creation | |
645 | * | |
646 | * @res: Pointer to a cotable resource. | |
647 | * @id: Item id. | |
648 | */ | |
649 | int vmw_cotable_notify(struct vmw_resource *res, int id) | |
650 | { | |
651 | struct vmw_cotable *vcotbl = vmw_cotable(res); | |
652 | ||
653 | if (id < 0 || id >= SVGA_COTABLE_MAX_IDS) { | |
654 | DRM_ERROR("Illegal COTable id. Type is %u. Id is %d\n", | |
655 | (unsigned) vcotbl->type, id); | |
656 | return -EINVAL; | |
657 | } | |
658 | ||
659 | if (vcotbl->seen_entries < id) { | |
660 | /* Trigger a call to create() on next validate */ | |
661 | res->id = -1; | |
662 | vcotbl->seen_entries = id; | |
663 | } | |
664 | ||
665 | return 0; | |
666 | } | |
667 | ||
668 | /** | |
2cd80dbd | 669 | * vmw_cotable_add_resource - add a view to the cotable's list of active views. |
d80efd5c TH |
670 | * |
671 | * @res: pointer struct vmw_resource representing the cotable. | |
672 | * @head: pointer to the struct list_head member of the resource, dedicated | |
673 | * to the cotable active resource list. | |
674 | */ | |
675 | void vmw_cotable_add_resource(struct vmw_resource *res, struct list_head *head) | |
676 | { | |
677 | struct vmw_cotable *vcotbl = | |
678 | container_of(res, struct vmw_cotable, res); | |
679 | ||
680 | list_add_tail(head, &vcotbl->resource_list); | |
681 | } |