1 | // SPDX-License-Identifier: GPL-2.0 OR MIT |
2 | /************************************************************************** |
3 | * |
4 | * Copyright 2014-2023 VMware, Inc., Palo Alto, CA., USA |
5 | * |
6 | * Permission is hereby granted, free of charge, to any person obtaining a |
7 | * copy of this software and associated documentation files (the |
8 | * "Software"), to deal in the Software without restriction, including |
9 | * without limitation the rights to use, copy, modify, merge, publish, |
10 | * distribute, sub license, and/or sell copies of the Software, and to |
11 | * permit persons to whom the Software is furnished to do so, subject to |
12 | * the following conditions: |
13 | * |
14 | * The above copyright notice and this permission notice (including the |
15 | * next paragraph) shall be included in all copies or substantial portions |
16 | * of the Software. |
17 | * |
18 | * THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR |
19 | * IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY, |
20 | * FITNESS FOR A PARTICULAR PURPOSE AND NON-INFRINGEMENT. IN NO EVENT SHALL |
21 | * THE COPYRIGHT HOLDERS, AUTHORS AND/OR ITS SUPPLIERS BE LIABLE FOR ANY CLAIM, |
22 | * DAMAGES OR OTHER LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR |
23 | * OTHERWISE, ARISING FROM, OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE |
24 | * USE OR OTHER DEALINGS IN THE SOFTWARE. |
25 | * |
26 | **************************************************************************/ |
27 | /* |
28 | * Treat context OTables as resources to make use of the resource |
29 | * backing MOB eviction mechanism, that is used to read back the COTable |
30 | * whenever the backing MOB is evicted. |
31 | */ |
32 | |
33 | #include "vmwgfx_bo.h" |
34 | #include "vmwgfx_drv.h" |
35 | #include "vmwgfx_mksstat.h" |
36 | #include "vmwgfx_resource_priv.h" |
37 | #include "vmwgfx_so.h" |
38 | |
39 | #include <drm/ttm/ttm_placement.h> |
40 | |
41 | /** |
42 | * struct vmw_cotable - Context Object Table resource |
43 | * |
44 | * @res: struct vmw_resource we are deriving from. |
45 | * @ctx: non-refcounted pointer to the owning context. |
46 | * @size_read_back: Size of data read back during eviction. |
47 | * @seen_entries: Seen entries in command stream for this cotable. |
48 | * @type: The cotable type. |
49 | * @scrubbed: Whether the cotable has been scrubbed. |
50 | * @resource_list: List of resources in the cotable. |
51 | */ |
52 | struct vmw_cotable { |
53 | struct vmw_resource res; |
54 | struct vmw_resource *ctx; |
55 | size_t size_read_back; |
56 | int seen_entries; |
57 | u32 type; |
58 | bool scrubbed; |
59 | struct list_head resource_list; |
60 | }; |
61 | |
62 | /** |
63 | * struct vmw_cotable_info - Static info about cotable types |
64 | * |
65 | * @min_initial_entries: Min number of initial intries at cotable allocation |
66 | * for this cotable type. |
67 | * @size: Size of each entry. |
68 | * @unbind_func: Unbind call-back function. |
69 | */ |
70 | struct vmw_cotable_info { |
71 | u32 min_initial_entries; |
72 | u32 size; |
73 | void (*unbind_func)(struct vmw_private *, struct list_head *, |
74 | bool); |
75 | }; |
76 | |
77 | |
78 | /* |
79 | * Getting the initial size right is difficult because it all depends |
80 | * on what the userspace is doing. The sizes will be aligned up to |
81 | * a PAGE_SIZE so we just want to make sure that for majority of apps |
82 | * the initial number of entries doesn't require an immediate resize. |
83 | * For all cotables except SVGACOTableDXElementLayoutEntry and |
84 | * SVGACOTableDXBlendStateEntry the initial number of entries fits |
85 | * within the PAGE_SIZE. For SVGACOTableDXElementLayoutEntry and |
86 | * SVGACOTableDXBlendStateEntry we want to reserve two pages, |
87 | * because that's what all apps will require initially. |
88 | */ |
89 | static const struct vmw_cotable_info co_info[] = { |
90 | {1, sizeof(SVGACOTableDXRTViewEntry), &vmw_view_cotable_list_destroy}, |
91 | {1, sizeof(SVGACOTableDXDSViewEntry), &vmw_view_cotable_list_destroy}, |
92 | {1, sizeof(SVGACOTableDXSRViewEntry), &vmw_view_cotable_list_destroy}, |
93 | {PAGE_SIZE/sizeof(SVGACOTableDXElementLayoutEntry) + 1, sizeof(SVGACOTableDXElementLayoutEntry), NULL}, |
94 | {PAGE_SIZE/sizeof(SVGACOTableDXBlendStateEntry) + 1, sizeof(SVGACOTableDXBlendStateEntry), NULL}, |
95 | {1, sizeof(SVGACOTableDXDepthStencilEntry), NULL}, |
96 | {1, sizeof(SVGACOTableDXRasterizerStateEntry), NULL}, |
97 | {1, sizeof(SVGACOTableDXSamplerEntry), NULL}, |
98 | {1, sizeof(SVGACOTableDXStreamOutputEntry), &vmw_dx_streamoutput_cotable_list_scrub}, |
99 | {1, sizeof(SVGACOTableDXQueryEntry), NULL}, |
100 | {1, sizeof(SVGACOTableDXShaderEntry), &vmw_dx_shader_cotable_list_scrub}, |
101 | {1, sizeof(SVGACOTableDXUAViewEntry), &vmw_view_cotable_list_destroy} |
102 | }; |
103 | |
104 | /* |
105 | * Cotables with bindings that we remove must be scrubbed first, |
106 | * otherwise, the device will swap in an invalid context when we remove |
107 | * bindings before scrubbing a cotable... |
108 | */ |
109 | const SVGACOTableType vmw_cotable_scrub_order[] = { |
110 | SVGA_COTABLE_RTVIEW, |
111 | SVGA_COTABLE_DSVIEW, |
112 | SVGA_COTABLE_SRVIEW, |
113 | SVGA_COTABLE_DXSHADER, |
114 | SVGA_COTABLE_ELEMENTLAYOUT, |
115 | SVGA_COTABLE_BLENDSTATE, |
116 | SVGA_COTABLE_DEPTHSTENCIL, |
117 | SVGA_COTABLE_RASTERIZERSTATE, |
118 | SVGA_COTABLE_SAMPLER, |
119 | SVGA_COTABLE_STREAMOUTPUT, |
120 | SVGA_COTABLE_DXQUERY, |
121 | SVGA_COTABLE_UAVIEW, |
122 | }; |
123 | |
124 | static int vmw_cotable_bind(struct vmw_resource *res, |
125 | struct ttm_validate_buffer *val_buf); |
126 | static int vmw_cotable_unbind(struct vmw_resource *res, |
127 | bool readback, |
128 | struct ttm_validate_buffer *val_buf); |
129 | static int vmw_cotable_create(struct vmw_resource *res); |
130 | static int vmw_cotable_destroy(struct vmw_resource *res); |
131 | |
132 | static const struct vmw_res_func vmw_cotable_func = { |
133 | .res_type = vmw_res_cotable, |
134 | .needs_guest_memory = true, |
135 | .may_evict = true, |
136 | .prio = 3, |
137 | .dirty_prio = 3, |
138 | .type_name = "context guest backed object tables" , |
139 | .domain = VMW_BO_DOMAIN_MOB, |
140 | .busy_domain = VMW_BO_DOMAIN_MOB, |
141 | .create = vmw_cotable_create, |
142 | .destroy = vmw_cotable_destroy, |
143 | .bind = vmw_cotable_bind, |
144 | .unbind = vmw_cotable_unbind, |
145 | }; |
146 | |
147 | /** |
148 | * vmw_cotable - Convert a struct vmw_resource pointer to a struct |
149 | * vmw_cotable pointer |
150 | * |
151 | * @res: Pointer to the resource. |
152 | */ |
153 | static struct vmw_cotable *vmw_cotable(struct vmw_resource *res) |
154 | { |
155 | return container_of(res, struct vmw_cotable, res); |
156 | } |
157 | |
158 | /** |
159 | * vmw_cotable_destroy - Cotable resource destroy callback |
160 | * |
161 | * @res: Pointer to the cotable resource. |
162 | * |
163 | * There is no device cotable destroy command, so this function only |
164 | * makes sure that the resource id is set to invalid. |
165 | */ |
166 | static int vmw_cotable_destroy(struct vmw_resource *res) |
167 | { |
168 | res->id = -1; |
169 | return 0; |
170 | } |
171 | |
172 | /** |
173 | * vmw_cotable_unscrub - Undo a cotable unscrub operation |
174 | * |
175 | * @res: Pointer to the cotable resource |
176 | * |
177 | * This function issues commands to (re)bind the cotable to |
178 | * its backing mob, which needs to be validated and reserved at this point. |
179 | * This is identical to bind() except the function interface looks different. |
180 | */ |
181 | static int vmw_cotable_unscrub(struct vmw_resource *res) |
182 | { |
183 | struct vmw_cotable *vcotbl = vmw_cotable(res); |
184 | struct vmw_private *dev_priv = res->dev_priv; |
185 | struct ttm_buffer_object *bo = &res->guest_memory_bo->tbo; |
186 | struct { |
187 | SVGA3dCmdHeader ; |
188 | SVGA3dCmdDXSetCOTable body; |
189 | } *cmd; |
190 | |
191 | WARN_ON_ONCE(bo->resource->mem_type != VMW_PL_MOB); |
192 | dma_resv_assert_held(bo->base.resv); |
193 | |
194 | cmd = VMW_CMD_RESERVE(dev_priv, sizeof(*cmd)); |
195 | if (!cmd) |
196 | return -ENOMEM; |
197 | |
198 | WARN_ON(vcotbl->ctx->id == SVGA3D_INVALID_ID); |
199 | WARN_ON(bo->resource->mem_type != VMW_PL_MOB); |
200 | cmd->header.id = SVGA_3D_CMD_DX_SET_COTABLE; |
201 | cmd->header.size = sizeof(cmd->body); |
202 | cmd->body.cid = vcotbl->ctx->id; |
203 | cmd->body.type = vcotbl->type; |
204 | cmd->body.mobid = bo->resource->start; |
205 | cmd->body.validSizeInBytes = vcotbl->size_read_back; |
206 | |
207 | vmw_cmd_commit_flush(dev_priv, bytes: sizeof(*cmd)); |
208 | vcotbl->scrubbed = false; |
209 | |
210 | return 0; |
211 | } |
212 | |
213 | /** |
214 | * vmw_cotable_bind - Undo a cotable unscrub operation |
215 | * |
216 | * @res: Pointer to the cotable resource |
217 | * @val_buf: Pointer to a struct ttm_validate_buffer prepared by the caller |
218 | * for convenience / fencing. |
219 | * |
220 | * This function issues commands to (re)bind the cotable to |
221 | * its backing mob, which needs to be validated and reserved at this point. |
222 | */ |
223 | static int vmw_cotable_bind(struct vmw_resource *res, |
224 | struct ttm_validate_buffer *val_buf) |
225 | { |
226 | /* |
227 | * The create() callback may have changed @res->backup without |
228 | * the caller noticing, and with val_buf->bo still pointing to |
229 | * the old backup buffer. Although hackish, and not used currently, |
230 | * take the opportunity to correct the value here so that it's not |
231 | * misused in the future. |
232 | */ |
233 | val_buf->bo = &res->guest_memory_bo->tbo; |
234 | |
235 | return vmw_cotable_unscrub(res); |
236 | } |
237 | |
238 | /** |
239 | * vmw_cotable_scrub - Scrub the cotable from the device. |
240 | * |
241 | * @res: Pointer to the cotable resource. |
242 | * @readback: Whether initiate a readback of the cotable data to the backup |
243 | * buffer. |
244 | * |
245 | * In some situations (context swapouts) it might be desirable to make the |
246 | * device forget about the cotable without performing a full unbind. A full |
247 | * unbind requires reserved backup buffers and it might not be possible to |
248 | * reserve them due to locking order violation issues. The vmw_cotable_scrub |
249 | * function implements a partial unbind() without that requirement but with the |
250 | * following restrictions. |
251 | * 1) Before the cotable is again used by the GPU, vmw_cotable_unscrub() must |
252 | * be called. |
253 | * 2) Before the cotable backing buffer is used by the CPU, or during the |
254 | * resource destruction, vmw_cotable_unbind() must be called. |
255 | */ |
256 | int vmw_cotable_scrub(struct vmw_resource *res, bool readback) |
257 | { |
258 | struct vmw_cotable *vcotbl = vmw_cotable(res); |
259 | struct vmw_private *dev_priv = res->dev_priv; |
260 | size_t submit_size; |
261 | |
262 | struct { |
263 | SVGA3dCmdHeader ; |
264 | SVGA3dCmdDXReadbackCOTable body; |
265 | } *cmd0; |
266 | struct { |
267 | SVGA3dCmdHeader ; |
268 | SVGA3dCmdDXSetCOTable body; |
269 | } *cmd1; |
270 | |
271 | if (vcotbl->scrubbed) |
272 | return 0; |
273 | |
274 | if (co_info[vcotbl->type].unbind_func) |
275 | co_info[vcotbl->type].unbind_func(dev_priv, |
276 | &vcotbl->resource_list, |
277 | readback); |
278 | submit_size = sizeof(*cmd1); |
279 | if (readback) |
280 | submit_size += sizeof(*cmd0); |
281 | |
282 | cmd1 = VMW_CMD_RESERVE(dev_priv, submit_size); |
283 | if (!cmd1) |
284 | return -ENOMEM; |
285 | |
286 | vcotbl->size_read_back = 0; |
287 | if (readback) { |
288 | cmd0 = (void *) cmd1; |
289 | cmd0->header.id = SVGA_3D_CMD_DX_READBACK_COTABLE; |
290 | cmd0->header.size = sizeof(cmd0->body); |
291 | cmd0->body.cid = vcotbl->ctx->id; |
292 | cmd0->body.type = vcotbl->type; |
293 | cmd1 = (void *) &cmd0[1]; |
294 | vcotbl->size_read_back = res->guest_memory_size; |
295 | } |
296 | cmd1->header.id = SVGA_3D_CMD_DX_SET_COTABLE; |
297 | cmd1->header.size = sizeof(cmd1->body); |
298 | cmd1->body.cid = vcotbl->ctx->id; |
299 | cmd1->body.type = vcotbl->type; |
300 | cmd1->body.mobid = SVGA3D_INVALID_ID; |
301 | cmd1->body.validSizeInBytes = 0; |
302 | vmw_cmd_commit_flush(dev_priv, bytes: submit_size); |
303 | vcotbl->scrubbed = true; |
304 | |
305 | /* Trigger a create() on next validate. */ |
306 | res->id = -1; |
307 | |
308 | return 0; |
309 | } |
310 | |
311 | /** |
312 | * vmw_cotable_unbind - Cotable resource unbind callback |
313 | * |
314 | * @res: Pointer to the cotable resource. |
315 | * @readback: Whether to read back cotable data to the backup buffer. |
316 | * @val_buf: Pointer to a struct ttm_validate_buffer prepared by the caller |
317 | * for convenience / fencing. |
318 | * |
319 | * Unbinds the cotable from the device and fences the backup buffer. |
320 | */ |
321 | static int vmw_cotable_unbind(struct vmw_resource *res, |
322 | bool readback, |
323 | struct ttm_validate_buffer *val_buf) |
324 | { |
325 | struct vmw_cotable *vcotbl = vmw_cotable(res); |
326 | struct vmw_private *dev_priv = res->dev_priv; |
327 | struct ttm_buffer_object *bo = val_buf->bo; |
328 | struct vmw_fence_obj *fence; |
329 | |
330 | if (!vmw_resource_mob_attached(res)) |
331 | return 0; |
332 | |
333 | WARN_ON_ONCE(bo->resource->mem_type != VMW_PL_MOB); |
334 | dma_resv_assert_held(bo->base.resv); |
335 | |
336 | mutex_lock(&dev_priv->binding_mutex); |
337 | if (!vcotbl->scrubbed) |
338 | vmw_dx_context_scrub_cotables(ctx: vcotbl->ctx, readback); |
339 | mutex_unlock(lock: &dev_priv->binding_mutex); |
340 | (void) vmw_execbuf_fence_commands(NULL, dev_priv, p_fence: &fence, NULL); |
341 | vmw_bo_fence_single(bo, fence); |
342 | if (likely(fence != NULL)) |
343 | vmw_fence_obj_unreference(fence_p: &fence); |
344 | |
345 | return 0; |
346 | } |
347 | |
348 | /** |
349 | * vmw_cotable_readback - Read back a cotable without unbinding. |
350 | * |
351 | * @res: The cotable resource. |
352 | * |
353 | * Reads back a cotable to its backing mob without scrubbing the MOB from |
354 | * the cotable. The MOB is fenced for subsequent CPU access. |
355 | */ |
356 | static int vmw_cotable_readback(struct vmw_resource *res) |
357 | { |
358 | struct vmw_cotable *vcotbl = vmw_cotable(res); |
359 | struct vmw_private *dev_priv = res->dev_priv; |
360 | |
361 | struct { |
362 | SVGA3dCmdHeader ; |
363 | SVGA3dCmdDXReadbackCOTable body; |
364 | } *cmd; |
365 | struct vmw_fence_obj *fence; |
366 | |
367 | if (!vcotbl->scrubbed) { |
368 | cmd = VMW_CMD_RESERVE(dev_priv, sizeof(*cmd)); |
369 | if (!cmd) |
370 | return -ENOMEM; |
371 | |
372 | cmd->header.id = SVGA_3D_CMD_DX_READBACK_COTABLE; |
373 | cmd->header.size = sizeof(cmd->body); |
374 | cmd->body.cid = vcotbl->ctx->id; |
375 | cmd->body.type = vcotbl->type; |
376 | vcotbl->size_read_back = res->guest_memory_size; |
377 | vmw_cmd_commit(dev_priv, bytes: sizeof(*cmd)); |
378 | } |
379 | |
380 | (void) vmw_execbuf_fence_commands(NULL, dev_priv, p_fence: &fence, NULL); |
381 | vmw_bo_fence_single(bo: &res->guest_memory_bo->tbo, fence); |
382 | vmw_fence_obj_unreference(fence_p: &fence); |
383 | |
384 | return 0; |
385 | } |
386 | |
387 | /** |
388 | * vmw_cotable_resize - Resize a cotable. |
389 | * |
390 | * @res: The cotable resource. |
391 | * @new_size: The new size. |
392 | * |
393 | * Resizes a cotable and binds the new backup buffer. |
394 | * On failure the cotable is left intact. |
395 | * Important! This function may not fail once the MOB switch has been |
396 | * committed to hardware. That would put the device context in an |
397 | * invalid state which we can't currently recover from. |
398 | */ |
399 | static int vmw_cotable_resize(struct vmw_resource *res, size_t new_size) |
400 | { |
401 | struct ttm_operation_ctx ctx = { false, false }; |
402 | struct vmw_private *dev_priv = res->dev_priv; |
403 | struct vmw_cotable *vcotbl = vmw_cotable(res); |
404 | struct vmw_bo *buf, *old_buf = res->guest_memory_bo; |
405 | struct ttm_buffer_object *bo, *old_bo = &res->guest_memory_bo->tbo; |
406 | size_t old_size = res->guest_memory_size; |
407 | size_t old_size_read_back = vcotbl->size_read_back; |
408 | size_t cur_size_read_back; |
409 | struct ttm_bo_kmap_obj old_map, new_map; |
410 | int ret; |
411 | size_t i; |
412 | struct vmw_bo_params bo_params = { |
413 | .domain = VMW_BO_DOMAIN_MOB, |
414 | .busy_domain = VMW_BO_DOMAIN_MOB, |
415 | .bo_type = ttm_bo_type_device, |
416 | .size = new_size, |
417 | .pin = true |
418 | }; |
419 | |
420 | MKS_STAT_TIME_DECL(MKSSTAT_KERN_COTABLE_RESIZE); |
421 | MKS_STAT_TIME_PUSH(MKSSTAT_KERN_COTABLE_RESIZE); |
422 | |
423 | ret = vmw_cotable_readback(res); |
424 | if (ret) |
425 | goto out_done; |
426 | |
427 | cur_size_read_back = vcotbl->size_read_back; |
428 | vcotbl->size_read_back = old_size_read_back; |
429 | |
430 | /* |
431 | * While device is processing, Allocate and reserve a buffer object |
432 | * for the new COTable. Initially pin the buffer object to make sure |
433 | * we can use tryreserve without failure. |
434 | */ |
435 | ret = vmw_gem_object_create(vmw: dev_priv, params: &bo_params, p_vbo: &buf); |
436 | if (ret) { |
437 | DRM_ERROR("Failed initializing new cotable MOB.\n" ); |
438 | goto out_done; |
439 | } |
440 | |
441 | bo = &buf->tbo; |
442 | WARN_ON_ONCE(ttm_bo_reserve(bo, false, true, NULL)); |
443 | |
444 | ret = ttm_bo_wait(bo: old_bo, intr: false, no_wait: false); |
445 | if (unlikely(ret != 0)) { |
446 | DRM_ERROR("Failed waiting for cotable unbind.\n" ); |
447 | goto out_wait; |
448 | } |
449 | |
450 | /* |
451 | * Do a page by page copy of COTables. This eliminates slow vmap()s. |
452 | * This should really be a TTM utility. |
453 | */ |
454 | for (i = 0; i < PFN_UP(old_bo->resource->size); ++i) { |
455 | bool dummy; |
456 | |
457 | ret = ttm_bo_kmap(bo: old_bo, start_page: i, num_pages: 1, map: &old_map); |
458 | if (unlikely(ret != 0)) { |
459 | DRM_ERROR("Failed mapping old COTable on resize.\n" ); |
460 | goto out_wait; |
461 | } |
462 | ret = ttm_bo_kmap(bo, start_page: i, num_pages: 1, map: &new_map); |
463 | if (unlikely(ret != 0)) { |
464 | DRM_ERROR("Failed mapping new COTable on resize.\n" ); |
465 | goto out_map_new; |
466 | } |
467 | memcpy(ttm_kmap_obj_virtual(&new_map, &dummy), |
468 | ttm_kmap_obj_virtual(&old_map, &dummy), |
469 | PAGE_SIZE); |
470 | ttm_bo_kunmap(map: &new_map); |
471 | ttm_bo_kunmap(map: &old_map); |
472 | } |
473 | |
474 | /* Unpin new buffer, and switch backup buffers. */ |
475 | vmw_bo_placement_set(bo: buf, |
476 | domain: VMW_BO_DOMAIN_MOB, |
477 | busy_domain: VMW_BO_DOMAIN_MOB); |
478 | ret = ttm_bo_validate(bo, placement: &buf->placement, ctx: &ctx); |
479 | if (unlikely(ret != 0)) { |
480 | DRM_ERROR("Failed validating new COTable backup buffer.\n" ); |
481 | goto out_wait; |
482 | } |
483 | |
484 | vmw_resource_mob_detach(res); |
485 | res->guest_memory_bo = buf; |
486 | res->guest_memory_size = new_size; |
487 | vcotbl->size_read_back = cur_size_read_back; |
488 | |
489 | /* |
490 | * Now tell the device to switch. If this fails, then we need to |
491 | * revert the full resize. |
492 | */ |
493 | ret = vmw_cotable_unscrub(res); |
494 | if (ret) { |
495 | DRM_ERROR("Failed switching COTable backup buffer.\n" ); |
496 | res->guest_memory_bo = old_buf; |
497 | res->guest_memory_size = old_size; |
498 | vcotbl->size_read_back = old_size_read_back; |
499 | vmw_resource_mob_attach(res); |
500 | goto out_wait; |
501 | } |
502 | |
503 | vmw_resource_mob_attach(res); |
504 | /* Let go of the old mob. */ |
505 | vmw_user_bo_unref(buf: &old_buf); |
506 | res->id = vcotbl->type; |
507 | |
508 | ret = dma_resv_reserve_fences(obj: bo->base.resv, num_fences: 1); |
509 | if (unlikely(ret)) |
510 | goto out_wait; |
511 | |
512 | /* Release the pin acquired in vmw_bo_create */ |
513 | ttm_bo_unpin(bo); |
514 | |
515 | MKS_STAT_TIME_POP(MKSSTAT_KERN_COTABLE_RESIZE); |
516 | |
517 | return 0; |
518 | |
519 | out_map_new: |
520 | ttm_bo_kunmap(map: &old_map); |
521 | out_wait: |
522 | ttm_bo_unpin(bo); |
523 | ttm_bo_unreserve(bo); |
524 | vmw_user_bo_unref(buf: &buf); |
525 | |
526 | out_done: |
527 | MKS_STAT_TIME_POP(MKSSTAT_KERN_COTABLE_RESIZE); |
528 | |
529 | return ret; |
530 | } |
531 | |
532 | /** |
533 | * vmw_cotable_create - Cotable resource create callback |
534 | * |
535 | * @res: Pointer to a cotable resource. |
536 | * |
537 | * There is no separate create command for cotables, so this callback, which |
538 | * is called before bind() in the validation sequence is instead used for two |
539 | * things. |
540 | * 1) Unscrub the cotable if it is scrubbed and still attached to a backup |
541 | * buffer. |
542 | * 2) Resize the cotable if needed. |
543 | */ |
544 | static int vmw_cotable_create(struct vmw_resource *res) |
545 | { |
546 | struct vmw_cotable *vcotbl = vmw_cotable(res); |
547 | size_t new_size = res->guest_memory_size; |
548 | size_t needed_size; |
549 | int ret; |
550 | |
551 | /* Check whether we need to resize the cotable */ |
552 | needed_size = (vcotbl->seen_entries + 1) * co_info[vcotbl->type].size; |
553 | while (needed_size > new_size) |
554 | new_size *= 2; |
555 | |
556 | if (likely(new_size <= res->guest_memory_size)) { |
557 | if (vcotbl->scrubbed && vmw_resource_mob_attached(res)) { |
558 | ret = vmw_cotable_unscrub(res); |
559 | if (ret) |
560 | return ret; |
561 | } |
562 | res->id = vcotbl->type; |
563 | return 0; |
564 | } |
565 | |
566 | return vmw_cotable_resize(res, new_size); |
567 | } |
568 | |
569 | /** |
570 | * vmw_hw_cotable_destroy - Cotable hw_destroy callback |
571 | * |
572 | * @res: Pointer to a cotable resource. |
573 | * |
574 | * The final (part of resource destruction) destroy callback. |
575 | */ |
576 | static void vmw_hw_cotable_destroy(struct vmw_resource *res) |
577 | { |
578 | (void) vmw_cotable_destroy(res); |
579 | } |
580 | |
581 | /** |
582 | * vmw_cotable_free - Cotable resource destructor |
583 | * |
584 | * @res: Pointer to a cotable resource. |
585 | */ |
586 | static void vmw_cotable_free(struct vmw_resource *res) |
587 | { |
588 | kfree(objp: res); |
589 | } |
590 | |
591 | /** |
592 | * vmw_cotable_alloc - Create a cotable resource |
593 | * |
594 | * @dev_priv: Pointer to a device private struct. |
595 | * @ctx: Pointer to the context resource. |
596 | * The cotable resource will not add a refcount. |
597 | * @type: The cotable type. |
598 | */ |
599 | struct vmw_resource *vmw_cotable_alloc(struct vmw_private *dev_priv, |
600 | struct vmw_resource *ctx, |
601 | u32 type) |
602 | { |
603 | struct vmw_cotable *vcotbl; |
604 | int ret; |
605 | u32 num_entries; |
606 | |
607 | vcotbl = kzalloc(size: sizeof(*vcotbl), GFP_KERNEL); |
608 | if (unlikely(!vcotbl)) { |
609 | ret = -ENOMEM; |
610 | goto out_no_alloc; |
611 | } |
612 | |
613 | ret = vmw_resource_init(dev_priv, res: &vcotbl->res, delay_id: true, |
614 | res_free: vmw_cotable_free, func: &vmw_cotable_func); |
615 | if (unlikely(ret != 0)) |
616 | goto out_no_init; |
617 | |
618 | INIT_LIST_HEAD(list: &vcotbl->resource_list); |
619 | vcotbl->res.id = type; |
620 | vcotbl->res.guest_memory_size = PAGE_SIZE; |
621 | num_entries = PAGE_SIZE / co_info[type].size; |
622 | if (num_entries < co_info[type].min_initial_entries) { |
623 | vcotbl->res.guest_memory_size = co_info[type].min_initial_entries * |
624 | co_info[type].size; |
625 | vcotbl->res.guest_memory_size = PFN_ALIGN(vcotbl->res.guest_memory_size); |
626 | } |
627 | |
628 | vcotbl->scrubbed = true; |
629 | vcotbl->seen_entries = -1; |
630 | vcotbl->type = type; |
631 | vcotbl->ctx = ctx; |
632 | |
633 | vcotbl->res.hw_destroy = vmw_hw_cotable_destroy; |
634 | |
635 | return &vcotbl->res; |
636 | |
637 | out_no_init: |
638 | kfree(objp: vcotbl); |
639 | out_no_alloc: |
640 | return ERR_PTR(error: ret); |
641 | } |
642 | |
643 | /** |
644 | * vmw_cotable_notify - Notify the cotable about an item creation |
645 | * |
646 | * @res: Pointer to a cotable resource. |
647 | * @id: Item id. |
648 | */ |
649 | int vmw_cotable_notify(struct vmw_resource *res, int id) |
650 | { |
651 | struct vmw_cotable *vcotbl = vmw_cotable(res); |
652 | |
653 | if (id < 0 || id >= SVGA_COTABLE_MAX_IDS) { |
654 | DRM_ERROR("Illegal COTable id. Type is %u. Id is %d\n" , |
655 | (unsigned) vcotbl->type, id); |
656 | return -EINVAL; |
657 | } |
658 | |
659 | if (vcotbl->seen_entries < id) { |
660 | /* Trigger a call to create() on next validate */ |
661 | res->id = -1; |
662 | vcotbl->seen_entries = id; |
663 | } |
664 | |
665 | return 0; |
666 | } |
667 | |
668 | /** |
669 | * vmw_cotable_add_resource - add a view to the cotable's list of active views. |
670 | * |
671 | * @res: pointer struct vmw_resource representing the cotable. |
672 | * @head: pointer to the struct list_head member of the resource, dedicated |
673 | * to the cotable active resource list. |
674 | */ |
675 | void vmw_cotable_add_resource(struct vmw_resource *res, struct list_head *head) |
676 | { |
677 | struct vmw_cotable *vcotbl = |
678 | container_of(res, struct vmw_cotable, res); |
679 | |
680 | list_add_tail(new: head, head: &vcotbl->resource_list); |
681 | } |
682 | |