vmwgfx_drv.h revision 1.5 1 /* $NetBSD: vmwgfx_drv.h,v 1.5 2021/12/19 00:25:04 riastradh Exp $ */
2
3 /* SPDX-License-Identifier: GPL-2.0 OR MIT */
4 /**************************************************************************
5 *
6 * Copyright 2009-2015 VMware, Inc., Palo Alto, CA., USA
7 *
8 * Permission is hereby granted, free of charge, to any person obtaining a
9 * copy of this software and associated documentation files (the
10 * "Software"), to deal in the Software without restriction, including
11 * without limitation the rights to use, copy, modify, merge, publish,
12 * distribute, sub license, and/or sell copies of the Software, and to
13 * permit persons to whom the Software is furnished to do so, subject to
14 * the following conditions:
15 *
16 * The above copyright notice and this permission notice (including the
17 * next paragraph) shall be included in all copies or substantial portions
18 * of the Software.
19 *
20 * THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR
21 * IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY,
22 * FITNESS FOR A PARTICULAR PURPOSE AND NON-INFRINGEMENT. IN NO EVENT SHALL
23 * THE COPYRIGHT HOLDERS, AUTHORS AND/OR ITS SUPPLIERS BE LIABLE FOR ANY CLAIM,
24 * DAMAGES OR OTHER LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR
25 * OTHERWISE, ARISING FROM, OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE
26 * USE OR OTHER DEALINGS IN THE SOFTWARE.
27 *
28 **************************************************************************/
29
30 #ifndef _VMWGFX_DRV_H_
31 #define _VMWGFX_DRV_H_
32
33 #include <linux/suspend.h>
34 #include <linux/sync_file.h>
35
36 #include <drm/drm_auth.h>
37 #include <drm/drm_device.h>
38 #include <drm/drm_file.h>
39 #include <drm/drm_hashtab.h>
40 #include <drm/drm_rect.h>
41
42 #include <drm/ttm/ttm_bo_driver.h>
43 #include <drm/ttm/ttm_execbuf_util.h>
44 #include <drm/ttm/ttm_module.h>
45
46 #include "ttm_lock.h"
47 #include "ttm_object.h"
48
49 #include "vmwgfx_fence.h"
50 #include "vmwgfx_reg.h"
51 #include "vmwgfx_validation.h"
52
53 /*
54 * FIXME: vmwgfx_drm.h needs to be last due to dependencies.
55 * uapi headers should not depend on header files outside uapi/.
56 */
57 #include <drm/vmwgfx_drm.h>
58
59
60 #define VMWGFX_DRIVER_NAME "vmwgfx"
61 #define VMWGFX_DRIVER_DATE "20200114"
62 #define VMWGFX_DRIVER_MAJOR 2
63 #define VMWGFX_DRIVER_MINOR 17
64 #define VMWGFX_DRIVER_PATCHLEVEL 0
65 #define VMWGFX_FIFO_STATIC_SIZE (1024*1024)
66 #define VMWGFX_MAX_RELOCATIONS 2048
67 #define VMWGFX_MAX_VALIDATIONS 2048
68 #define VMWGFX_MAX_DISPLAYS 16
69 #define VMWGFX_CMD_BOUNCE_INIT_SIZE 32768
70 #define VMWGFX_ENABLE_SCREEN_TARGET_OTABLE 1
71
72 /*
73 * Perhaps we should have sysfs entries for these.
74 */
75 #define VMWGFX_NUM_GB_CONTEXT 256
76 #define VMWGFX_NUM_GB_SHADER 20000
77 #define VMWGFX_NUM_GB_SURFACE 32768
78 #define VMWGFX_NUM_GB_SCREEN_TARGET VMWGFX_MAX_DISPLAYS
79 #define VMWGFX_NUM_DXCONTEXT 256
80 #define VMWGFX_NUM_DXQUERY 512
81 #define VMWGFX_NUM_MOB (VMWGFX_NUM_GB_CONTEXT +\
82 VMWGFX_NUM_GB_SHADER +\
83 VMWGFX_NUM_GB_SURFACE +\
84 VMWGFX_NUM_GB_SCREEN_TARGET)
85
86 #define VMW_PL_GMR (TTM_PL_PRIV + 0)
87 #define VMW_PL_FLAG_GMR (TTM_PL_FLAG_PRIV << 0)
88 #define VMW_PL_MOB (TTM_PL_PRIV + 1)
89 #define VMW_PL_FLAG_MOB (TTM_PL_FLAG_PRIV << 1)
90
91 #define VMW_RES_CONTEXT ttm_driver_type0
92 #define VMW_RES_SURFACE ttm_driver_type1
93 #define VMW_RES_STREAM ttm_driver_type2
94 #define VMW_RES_FENCE ttm_driver_type3
95 #define VMW_RES_SHADER ttm_driver_type4
96
97 struct vmw_fpriv {
98 struct ttm_object_file *tfile;
99 bool gb_aware; /* user-space is guest-backed aware */
100 };
101
102 /**
103 * struct vmw_buffer_object - TTM buffer object with vmwgfx additions
104 * @base: The TTM buffer object
105 * @res_tree: RB tree of resources using this buffer object as a backing MOB
106 * @pin_count: pin depth
107 * @cpu_writers: Number of synccpu write grabs. Protected by reservation when
108 * increased. May be decreased without reservation.
109 * @dx_query_ctx: DX context if this buffer object is used as a DX query MOB
110 * @map: Kmap object for semi-persistent mappings
111 * @res_prios: Eviction priority counts for attached resources
112 * @dirty: structure for user-space dirty-tracking
113 */
114 struct vmw_buffer_object {
115 struct ttm_buffer_object base;
116 struct rb_root res_tree;
117 s32 pin_count;
118 atomic_t cpu_writers;
119 /* Not ref-counted. Protected by binding_mutex */
120 struct vmw_resource *dx_query_ctx;
121 /* Protected by reservation */
122 struct ttm_bo_kmap_obj map;
123 u32 res_prios[TTM_MAX_BO_PRIORITY];
124 struct vmw_bo_dirty *dirty;
125 };
126
127 /**
128 * struct vmw_validate_buffer - Carries validation info about buffers.
129 *
130 * @base: Validation info for TTM.
131 * @hash: Hash entry for quick lookup of the TTM buffer object.
132 *
133 * This structure contains also driver private validation info
134 * on top of the info needed by TTM.
135 */
136 struct vmw_validate_buffer {
137 struct ttm_validate_buffer base;
138 struct drm_hash_item hash;
139 bool validate_as_mob;
140 };
141
142 struct vmw_res_func;
143
144
145 /**
146 * struct vmw-resource - base class for hardware resources
147 *
148 * @kref: For refcounting.
149 * @dev_priv: Pointer to the device private for this resource. Immutable.
150 * @id: Device id. Protected by @dev_priv::resource_lock.
151 * @backup_size: Backup buffer size. Immutable.
152 * @res_dirty: Resource contains data not yet in the backup buffer. Protected
153 * by resource reserved.
154 * @backup_dirty: Backup buffer contains data not yet in the HW resource.
155 * Protected by resource reserved.
156 * @coherent: Emulate coherency by tracking vm accesses.
157 * @backup: The backup buffer if any. Protected by resource reserved.
158 * @backup_offset: Offset into the backup buffer if any. Protected by resource
159 * reserved. Note that only a few resource types can have a @backup_offset
160 * different from zero.
161 * @pin_count: The pin count for this resource. A pinned resource has a
162 * pin-count greater than zero. It is not on the resource LRU lists and its
163 * backup buffer is pinned. Hence it can't be evicted.
164 * @func: Method vtable for this resource. Immutable.
165 * @mob_node; Node for the MOB backup rbtree. Protected by @backup reserved.
166 * @lru_head: List head for the LRU list. Protected by @dev_priv::resource_lock.
167 * @binding_head: List head for the context binding list. Protected by
168 * the @dev_priv::binding_mutex
169 * @res_free: The resource destructor.
170 * @hw_destroy: Callback to destroy the resource on the device, as part of
171 * resource destruction.
172 */
173 struct vmw_resource_dirty;
174 struct vmw_resource {
175 struct kref kref;
176 struct vmw_private *dev_priv;
177 int id;
178 u32 used_prio;
179 unsigned long backup_size;
180 u32 res_dirty : 1;
181 u32 backup_dirty : 1;
182 u32 coherent : 1;
183 struct vmw_buffer_object *backup;
184 unsigned long backup_offset;
185 unsigned long pin_count;
186 const struct vmw_res_func *func;
187 struct rb_node mob_node;
188 struct list_head lru_head;
189 struct list_head binding_head;
190 struct vmw_resource_dirty *dirty;
191 void (*res_free) (struct vmw_resource *res);
192 void (*hw_destroy) (struct vmw_resource *res);
193 };
194
195
196 /*
197 * Resources that are managed using ioctls.
198 */
199 enum vmw_res_type {
200 vmw_res_context,
201 vmw_res_surface,
202 vmw_res_stream,
203 vmw_res_shader,
204 vmw_res_dx_context,
205 vmw_res_cotable,
206 vmw_res_view,
207 vmw_res_max
208 };
209
210 /*
211 * Resources that are managed using command streams.
212 */
213 enum vmw_cmdbuf_res_type {
214 vmw_cmdbuf_res_shader,
215 vmw_cmdbuf_res_view
216 };
217
218 struct vmw_cmdbuf_res_manager;
219
220 struct vmw_cursor_snooper {
221 size_t age;
222 uint32_t *image;
223 };
224
225 struct vmw_framebuffer;
226 struct vmw_surface_offset;
227
228 struct vmw_surface {
229 struct vmw_resource res;
230 SVGA3dSurfaceAllFlags flags;
231 uint32_t format;
232 uint32_t mip_levels[DRM_VMW_MAX_SURFACE_FACES];
233 struct drm_vmw_size base_size;
234 struct drm_vmw_size *sizes;
235 uint32_t num_sizes;
236 bool scanout;
237 uint32_t array_size;
238 /* TODO so far just a extra pointer */
239 struct vmw_cursor_snooper snooper;
240 struct vmw_surface_offset *offsets;
241 SVGA3dTextureFilter autogen_filter;
242 uint32_t multisample_count;
243 struct list_head view_list;
244 SVGA3dMSPattern multisample_pattern;
245 SVGA3dMSQualityLevel quality_level;
246 };
247
248 struct vmw_marker_queue {
249 struct list_head head;
250 u64 lag;
251 u64 lag_time;
252 spinlock_t lock;
253 };
254
255 struct vmw_fifo_state {
256 unsigned long reserved_size;
257 u32 *dynamic_buffer;
258 u32 *static_buffer;
259 unsigned long static_buffer_size;
260 bool using_bounce_buffer;
261 uint32_t capabilities;
262 struct mutex fifo_mutex;
263 struct rw_semaphore rwsem;
264 struct vmw_marker_queue marker_queue;
265 bool dx;
266 };
267
268 /**
269 * struct vmw_res_cache_entry - resource information cache entry
270 * @handle: User-space handle of a resource.
271 * @res: Non-ref-counted pointer to the resource.
272 * @valid_handle: Whether the @handle member is valid.
273 * @valid: Whether the entry is valid, which also implies that the execbuf
274 * code holds a reference to the resource, and it's placed on the
275 * validation list.
276 *
277 * Used to avoid frequent repeated user-space handle lookups of the
278 * same resource.
279 */
280 struct vmw_res_cache_entry {
281 uint32_t handle;
282 struct vmw_resource *res;
283 void *private;
284 unsigned short valid_handle;
285 unsigned short valid;
286 };
287
288 /**
289 * enum vmw_dma_map_mode - indicate how to perform TTM page dma mappings.
290 */
291 enum vmw_dma_map_mode {
292 vmw_dma_phys, /* Use physical page addresses */
293 vmw_dma_alloc_coherent, /* Use TTM coherent pages */
294 vmw_dma_map_populate, /* Unmap from DMA just after unpopulate */
295 vmw_dma_map_bind, /* Unmap from DMA just before unbind */
296 vmw_dma_map_max
297 };
298
299 /**
300 * struct vmw_sg_table - Scatter/gather table for binding, with additional
301 * device-specific information.
302 *
303 * @sgt: Pointer to a struct sg_table with binding information
304 * @num_regions: Number of regions with device-address contiguous pages
305 */
306 struct vmw_sg_table {
307 enum vmw_dma_map_mode mode;
308 struct page **pages;
309 const dma_addr_t *addrs;
310 struct sg_table *sgt;
311 unsigned long num_regions;
312 unsigned long num_pages;
313 };
314
315 /**
316 * struct vmw_piter - Page iterator that iterates over a list of pages
317 * and DMA addresses that could be either a scatter-gather list or
318 * arrays
319 *
320 * @pages: Array of page pointers to the pages.
321 * @addrs: DMA addresses to the pages if coherent pages are used.
322 * @iter: Scatter-gather page iterator. Current position in SG list.
323 * @i: Current position in arrays.
324 * @num_pages: Number of pages total.
325 * @next: Function to advance the iterator. Returns false if past the list
326 * of pages, true otherwise.
327 * @dma_address: Function to return the DMA address of the current page.
328 */
329 struct vmw_piter {
330 struct page **pages;
331 const dma_addr_t *addrs;
332 struct sg_dma_page_iter iter;
333 unsigned long i;
334 unsigned long num_pages;
335 bool (*next)(struct vmw_piter *);
336 dma_addr_t (*dma_address)(struct vmw_piter *);
337 struct page *(*page)(struct vmw_piter *);
338 };
339
340 /*
341 * enum vmw_display_unit_type - Describes the display unit
342 */
343 enum vmw_display_unit_type {
344 vmw_du_invalid = 0,
345 vmw_du_legacy,
346 vmw_du_screen_object,
347 vmw_du_screen_target
348 };
349
350 struct vmw_validation_context;
351 struct vmw_ctx_validation_info;
352
353 /**
354 * struct vmw_sw_context - Command submission context
355 * @res_ht: Pointer hash table used to find validation duplicates
356 * @kernel: Whether the command buffer originates from kernel code rather
357 * than from user-space
358 * @fp: If @kernel is false, points to the file of the client. Otherwise
359 * NULL
360 * @cmd_bounce: Command bounce buffer used for command validation before
361 * copying to fifo space
362 * @cmd_bounce_size: Current command bounce buffer size
363 * @cur_query_bo: Current buffer object used as query result buffer
364 * @bo_relocations: List of buffer object relocations
365 * @res_relocations: List of resource relocations
366 * @buf_start: Pointer to start of memory where command validation takes
367 * place
368 * @res_cache: Cache of recently looked up resources
369 * @last_query_ctx: Last context that submitted a query
370 * @needs_post_query_barrier: Whether a query barrier is needed after
371 * command submission
372 * @staged_bindings: Cached per-context binding tracker
373 * @staged_bindings_inuse: Whether the cached per-context binding tracker
374 * is in use
375 * @staged_cmd_res: List of staged command buffer managed resources in this
376 * command buffer
377 * @ctx_list: List of context resources referenced in this command buffer
378 * @dx_ctx_node: Validation metadata of the current DX context
379 * @dx_query_mob: The MOB used for DX queries
380 * @dx_query_ctx: The DX context used for the last DX query
381 * @man: Pointer to the command buffer managed resource manager
382 * @ctx: The validation context
383 */
384 struct vmw_sw_context{
385 struct drm_open_hash res_ht;
386 bool res_ht_initialized;
387 bool kernel;
388 struct vmw_fpriv *fp;
389 uint32_t *cmd_bounce;
390 uint32_t cmd_bounce_size;
391 struct vmw_buffer_object *cur_query_bo;
392 struct list_head bo_relocations;
393 struct list_head res_relocations;
394 uint32_t *buf_start;
395 struct vmw_res_cache_entry res_cache[vmw_res_max];
396 struct vmw_resource *last_query_ctx;
397 bool needs_post_query_barrier;
398 struct vmw_ctx_binding_state *staged_bindings;
399 bool staged_bindings_inuse;
400 struct list_head staged_cmd_res;
401 struct list_head ctx_list;
402 struct vmw_ctx_validation_info *dx_ctx_node;
403 struct vmw_buffer_object *dx_query_mob;
404 struct vmw_resource *dx_query_ctx;
405 struct vmw_cmdbuf_res_manager *man;
406 struct vmw_validation_context *ctx;
407 };
408
409 struct vmw_legacy_display;
410 struct vmw_overlay;
411
412 struct vmw_vga_topology_state {
413 uint32_t width;
414 uint32_t height;
415 uint32_t primary;
416 uint32_t pos_x;
417 uint32_t pos_y;
418 };
419
420
421 /*
422 * struct vmw_otable - Guest Memory OBject table metadata
423 *
424 * @size: Size of the table (page-aligned).
425 * @page_table: Pointer to a struct vmw_mob holding the page table.
426 */
427 struct vmw_otable {
428 unsigned long size;
429 struct vmw_mob *page_table;
430 bool enabled;
431 };
432
433 struct vmw_otable_batch {
434 unsigned num_otables;
435 struct vmw_otable *otables;
436 struct vmw_resource *context;
437 struct ttm_buffer_object *otable_bo;
438 };
439
440 enum {
441 VMW_IRQTHREAD_FENCE,
442 VMW_IRQTHREAD_CMDBUF,
443 VMW_IRQTHREAD_MAX
444 };
445
446 struct vmw_private {
447 struct ttm_bo_device bdev;
448
449 struct vmw_fifo_state fifo;
450
451 struct drm_device *dev;
452 struct drm_vma_offset_manager vma_manager;
453 unsigned long vmw_chipset;
454 unsigned int io_start;
455 uint32_t vram_start;
456 uint32_t vram_size;
457 uint32_t prim_bb_mem;
458 uint32_t mmio_start;
459 uint32_t mmio_size;
460 uint32_t fb_max_width;
461 uint32_t fb_max_height;
462 uint32_t texture_max_width;
463 uint32_t texture_max_height;
464 uint32_t stdu_max_width;
465 uint32_t stdu_max_height;
466 uint32_t initial_width;
467 uint32_t initial_height;
468 u32 *mmio_virt;
469 uint32_t capabilities;
470 uint32_t capabilities2;
471 uint32_t max_gmr_ids;
472 uint32_t max_gmr_pages;
473 uint32_t max_mob_pages;
474 uint32_t max_mob_size;
475 uint32_t memory_size;
476 bool has_gmr;
477 bool has_mob;
478 spinlock_t hw_lock;
479 spinlock_t cap_lock;
480 bool has_dx;
481 bool assume_16bpp;
482 bool has_sm4_1;
483
484 /*
485 * VGA registers.
486 */
487
488 struct vmw_vga_topology_state vga_save[VMWGFX_MAX_DISPLAYS];
489 uint32_t vga_width;
490 uint32_t vga_height;
491 uint32_t vga_bpp;
492 uint32_t vga_bpl;
493 uint32_t vga_pitchlock;
494
495 uint32_t num_displays;
496
497 /*
498 * Framebuffer info.
499 */
500
501 void *fb_info;
502 enum vmw_display_unit_type active_display_unit;
503 struct vmw_legacy_display *ldu_priv;
504 struct vmw_overlay *overlay_priv;
505 struct drm_property *hotplug_mode_update_property;
506 struct drm_property *implicit_placement_property;
507 struct mutex global_kms_state_mutex;
508 spinlock_t cursor_lock;
509 struct drm_atomic_state *suspend_state;
510
511 /*
512 * Context and surface management.
513 */
514
515 spinlock_t resource_lock;
516 struct idr res_idr[vmw_res_max];
517
518 /*
519 * A resource manager for kernel-only surfaces and
520 * contexts.
521 */
522
523 struct ttm_object_device *tdev;
524
525 /*
526 * Fencing and IRQs.
527 */
528
529 atomic_t marker_seq;
530 wait_queue_head_t fence_queue;
531 wait_queue_head_t fifo_queue;
532 spinlock_t waiter_lock;
533 int fence_queue_waiters; /* Protected by waiter_lock */
534 int goal_queue_waiters; /* Protected by waiter_lock */
535 int cmdbuf_waiters; /* Protected by waiter_lock */
536 int error_waiters; /* Protected by waiter_lock */
537 int fifo_queue_waiters; /* Protected by waiter_lock */
538 uint32_t last_read_seqno;
539 struct vmw_fence_manager *fman;
540 uint32_t irq_mask; /* Updates protected by waiter_lock */
541
542 /*
543 * Device state
544 */
545
546 uint32_t traces_state;
547 uint32_t enable_state;
548 uint32_t config_done_state;
549
550 /**
551 * Execbuf
552 */
553 /**
554 * Protected by the cmdbuf mutex.
555 */
556
557 struct vmw_sw_context ctx;
558 struct mutex cmdbuf_mutex;
559 struct mutex binding_mutex;
560
561 /**
562 * Operating mode.
563 */
564
565 bool stealth;
566 bool enable_fb;
567 spinlock_t svga_lock;
568
569 /**
570 * PM management.
571 */
572 struct notifier_block pm_nb;
573 bool refuse_hibernation;
574 bool suspend_locked;
575
576 struct mutex release_mutex;
577 atomic_t num_fifo_resources;
578
579 /*
580 * Replace this with an rwsem as soon as we have down_xx_interruptible()
581 */
582 struct ttm_lock reservation_sem;
583
584 /*
585 * Query processing. These members
586 * are protected by the cmdbuf mutex.
587 */
588
589 struct vmw_buffer_object *dummy_query_bo;
590 struct vmw_buffer_object *pinned_bo;
591 uint32_t query_cid;
592 uint32_t query_cid_valid;
593 bool dummy_query_bo_pinned;
594
595 /*
596 * Surface swapping. The "surface_lru" list is protected by the
597 * resource lock in order to be able to destroy a surface and take
598 * it off the lru atomically. "used_memory_size" is currently
599 * protected by the cmdbuf mutex for simplicity.
600 */
601
602 struct list_head res_lru[vmw_res_max];
603 uint32_t used_memory_size;
604
605 /*
606 * DMA mapping stuff.
607 */
608 enum vmw_dma_map_mode map_mode;
609
610 /*
611 * Guest Backed stuff
612 */
613 struct vmw_otable_batch otable_batch;
614
615 struct vmw_cmdbuf_man *cman;
616 DECLARE_BITMAP(irqthread_pending, VMW_IRQTHREAD_MAX);
617
618 /* Validation memory reservation */
619 struct vmw_validation_mem vvm;
620 };
621
622 static inline struct vmw_surface *vmw_res_to_srf(struct vmw_resource *res)
623 {
624 return container_of(res, struct vmw_surface, res);
625 }
626
627 static inline struct vmw_private *vmw_priv(struct drm_device *dev)
628 {
629 return (struct vmw_private *)dev->dev_private;
630 }
631
632 static inline struct vmw_fpriv *vmw_fpriv(struct drm_file *file_priv)
633 {
634 return (struct vmw_fpriv *)file_priv->driver_priv;
635 }
636
637 /*
638 * The locking here is fine-grained, so that it is performed once
639 * for every read- and write operation. This is of course costly, but we
640 * don't perform much register access in the timing critical paths anyway.
641 * Instead we have the extra benefit of being sure that we don't forget
642 * the hw lock around register accesses.
643 */
644 static inline void vmw_write(struct vmw_private *dev_priv,
645 unsigned int offset, uint32_t value)
646 {
647 spin_lock(&dev_priv->hw_lock);
648 outl(offset, dev_priv->io_start + VMWGFX_INDEX_PORT);
649 outl(value, dev_priv->io_start + VMWGFX_VALUE_PORT);
650 spin_unlock(&dev_priv->hw_lock);
651 }
652
653 static inline uint32_t vmw_read(struct vmw_private *dev_priv,
654 unsigned int offset)
655 {
656 u32 val;
657
658 spin_lock(&dev_priv->hw_lock);
659 outl(offset, dev_priv->io_start + VMWGFX_INDEX_PORT);
660 val = inl(dev_priv->io_start + VMWGFX_VALUE_PORT);
661 spin_unlock(&dev_priv->hw_lock);
662
663 return val;
664 }
665
666 extern void vmw_svga_enable(struct vmw_private *dev_priv);
667 extern void vmw_svga_disable(struct vmw_private *dev_priv);
668
669
670 /**
671 * GMR utilities - vmwgfx_gmr.c
672 */
673
674 extern int vmw_gmr_bind(struct vmw_private *dev_priv,
675 const struct vmw_sg_table *vsgt,
676 unsigned long num_pages,
677 int gmr_id);
678 extern void vmw_gmr_unbind(struct vmw_private *dev_priv, int gmr_id);
679
680 /**
681 * Resource utilities - vmwgfx_resource.c
682 */
683 struct vmw_user_resource_conv;
684
685 extern void vmw_resource_unreference(struct vmw_resource **p_res);
686 extern struct vmw_resource *vmw_resource_reference(struct vmw_resource *res);
687 extern struct vmw_resource *
688 vmw_resource_reference_unless_doomed(struct vmw_resource *res);
689 extern int vmw_resource_validate(struct vmw_resource *res, bool intr,
690 bool dirtying);
691 extern int vmw_resource_reserve(struct vmw_resource *res, bool interruptible,
692 bool no_backup);
693 extern bool vmw_resource_needs_backup(const struct vmw_resource *res);
694 extern int vmw_user_lookup_handle(struct vmw_private *dev_priv,
695 struct ttm_object_file *tfile,
696 uint32_t handle,
697 struct vmw_surface **out_surf,
698 struct vmw_buffer_object **out_buf);
699 extern int vmw_user_resource_lookup_handle(
700 struct vmw_private *dev_priv,
701 struct ttm_object_file *tfile,
702 uint32_t handle,
703 const struct vmw_user_resource_conv *converter,
704 struct vmw_resource **p_res);
705 extern struct vmw_resource *
706 vmw_user_resource_noref_lookup_handle(struct vmw_private *dev_priv,
707 struct ttm_object_file *tfile,
708 uint32_t handle,
709 const struct vmw_user_resource_conv *
710 converter);
711 extern int vmw_stream_claim_ioctl(struct drm_device *dev, void *data,
712 struct drm_file *file_priv);
713 extern int vmw_stream_unref_ioctl(struct drm_device *dev, void *data,
714 struct drm_file *file_priv);
715 extern int vmw_user_stream_lookup(struct vmw_private *dev_priv,
716 struct ttm_object_file *tfile,
717 uint32_t *inout_id,
718 struct vmw_resource **out);
719 extern void vmw_resource_unreserve(struct vmw_resource *res,
720 bool dirty_set,
721 bool dirty,
722 bool switch_backup,
723 struct vmw_buffer_object *new_backup,
724 unsigned long new_backup_offset);
725 extern void vmw_query_move_notify(struct ttm_buffer_object *bo,
726 struct ttm_mem_reg *mem);
727 extern int vmw_query_readback_all(struct vmw_buffer_object *dx_query_mob);
728 extern void vmw_resource_evict_all(struct vmw_private *dev_priv);
729 extern void vmw_resource_unbind_list(struct vmw_buffer_object *vbo);
730 void vmw_resource_mob_attach(struct vmw_resource *res);
731 void vmw_resource_mob_detach(struct vmw_resource *res);
732 void vmw_resource_dirty_update(struct vmw_resource *res, pgoff_t start,
733 pgoff_t end);
734 int vmw_resources_clean(struct vmw_buffer_object *vbo, pgoff_t start,
735 pgoff_t end, pgoff_t *num_prefault);
736
737 /**
738 * vmw_resource_mob_attached - Whether a resource currently has a mob attached
739 * @res: The resource
740 *
741 * Return: true if the resource has a mob attached, false otherwise.
742 */
743 static inline bool vmw_resource_mob_attached(const struct vmw_resource *res)
744 {
745 return !RB_EMPTY_NODE(&res->mob_node);
746 }
747
748 /**
749 * vmw_user_resource_noref_release - release a user resource pointer looked up
750 * without reference
751 */
752 static inline void vmw_user_resource_noref_release(void)
753 {
754 ttm_base_object_noref_release();
755 }
756
757 /**
758 * Buffer object helper functions - vmwgfx_bo.c
759 */
760 extern int vmw_bo_pin_in_placement(struct vmw_private *vmw_priv,
761 struct vmw_buffer_object *bo,
762 struct ttm_placement *placement,
763 bool interruptible);
764 extern int vmw_bo_pin_in_vram(struct vmw_private *dev_priv,
765 struct vmw_buffer_object *buf,
766 bool interruptible);
767 extern int vmw_bo_pin_in_vram_or_gmr(struct vmw_private *dev_priv,
768 struct vmw_buffer_object *buf,
769 bool interruptible);
770 extern int vmw_bo_pin_in_start_of_vram(struct vmw_private *vmw_priv,
771 struct vmw_buffer_object *bo,
772 bool interruptible);
773 extern int vmw_bo_unpin(struct vmw_private *vmw_priv,
774 struct vmw_buffer_object *bo,
775 bool interruptible);
776 extern void vmw_bo_get_guest_ptr(const struct ttm_buffer_object *buf,
777 SVGAGuestPtr *ptr);
778 extern void vmw_bo_pin_reserved(struct vmw_buffer_object *bo, bool pin);
779 extern void vmw_bo_bo_free(struct ttm_buffer_object *bo);
780 extern int vmw_bo_init(struct vmw_private *dev_priv,
781 struct vmw_buffer_object *vmw_bo,
782 size_t size, struct ttm_placement *placement,
783 bool interuptable,
784 void (*bo_free)(struct ttm_buffer_object *bo));
785 extern int vmw_user_bo_verify_access(struct ttm_buffer_object *bo,
786 struct ttm_object_file *tfile);
787 extern int vmw_user_bo_alloc(struct vmw_private *dev_priv,
788 struct ttm_object_file *tfile,
789 uint32_t size,
790 bool shareable,
791 uint32_t *handle,
792 struct vmw_buffer_object **p_dma_buf,
793 struct ttm_base_object **p_base);
794 extern int vmw_user_bo_reference(struct ttm_object_file *tfile,
795 struct vmw_buffer_object *dma_buf,
796 uint32_t *handle);
797 extern int vmw_bo_alloc_ioctl(struct drm_device *dev, void *data,
798 struct drm_file *file_priv);
799 extern int vmw_bo_unref_ioctl(struct drm_device *dev, void *data,
800 struct drm_file *file_priv);
801 extern int vmw_user_bo_synccpu_ioctl(struct drm_device *dev, void *data,
802 struct drm_file *file_priv);
803 extern int vmw_user_bo_lookup(struct ttm_object_file *tfile,
804 uint32_t id, struct vmw_buffer_object **out,
805 struct ttm_base_object **base);
806 extern void vmw_bo_fence_single(struct ttm_buffer_object *bo,
807 struct vmw_fence_obj *fence);
808 extern void *vmw_bo_map_and_cache(struct vmw_buffer_object *vbo);
809 extern void vmw_bo_unmap(struct vmw_buffer_object *vbo);
810 extern void vmw_bo_move_notify(struct ttm_buffer_object *bo,
811 struct ttm_mem_reg *mem);
812 extern void vmw_bo_swap_notify(struct ttm_buffer_object *bo);
813 extern struct vmw_buffer_object *
814 vmw_user_bo_noref_lookup(struct ttm_object_file *tfile, u32 handle);
815
816 /**
817 * vmw_user_bo_noref_release - release a buffer object pointer looked up
818 * without reference
819 */
820 static inline void vmw_user_bo_noref_release(void)
821 {
822 ttm_base_object_noref_release();
823 }
824
825 /**
826 * vmw_bo_adjust_prio - Adjust the buffer object eviction priority
827 * according to attached resources
828 * @vbo: The struct vmw_buffer_object
829 */
830 static inline void vmw_bo_prio_adjust(struct vmw_buffer_object *vbo)
831 {
832 int i = ARRAY_SIZE(vbo->res_prios);
833
834 while (i--) {
835 if (vbo->res_prios[i]) {
836 vbo->base.priority = i;
837 return;
838 }
839 }
840
841 vbo->base.priority = 3;
842 }
843
844 /**
845 * vmw_bo_prio_add - Notify a buffer object of a newly attached resource
846 * eviction priority
847 * @vbo: The struct vmw_buffer_object
848 * @prio: The resource priority
849 *
850 * After being notified, the code assigns the highest resource eviction priority
851 * to the backing buffer object (mob).
852 */
853 static inline void vmw_bo_prio_add(struct vmw_buffer_object *vbo, int prio)
854 {
855 if (vbo->res_prios[prio]++ == 0)
856 vmw_bo_prio_adjust(vbo);
857 }
858
859 /**
860 * vmw_bo_prio_del - Notify a buffer object of a resource with a certain
861 * priority being removed
862 * @vbo: The struct vmw_buffer_object
863 * @prio: The resource priority
864 *
865 * After being notified, the code assigns the highest resource eviction priority
866 * to the backing buffer object (mob).
867 */
868 static inline void vmw_bo_prio_del(struct vmw_buffer_object *vbo, int prio)
869 {
870 if (--vbo->res_prios[prio] == 0)
871 vmw_bo_prio_adjust(vbo);
872 }
873
874 /**
875 * Misc Ioctl functionality - vmwgfx_ioctl.c
876 */
877
878 extern int vmw_getparam_ioctl(struct drm_device *dev, void *data,
879 struct drm_file *file_priv);
880 extern int vmw_get_cap_3d_ioctl(struct drm_device *dev, void *data,
881 struct drm_file *file_priv);
882 extern int vmw_present_ioctl(struct drm_device *dev, void *data,
883 struct drm_file *file_priv);
884 extern int vmw_present_readback_ioctl(struct drm_device *dev, void *data,
885 struct drm_file *file_priv);
886 extern __poll_t vmw_fops_poll(struct file *filp,
887 struct poll_table_struct *wait);
888 extern ssize_t vmw_fops_read(struct file *filp, char __user *buffer,
889 size_t count, loff_t *offset);
890
891 /**
892 * Fifo utilities - vmwgfx_fifo.c
893 */
894
895 extern int vmw_fifo_init(struct vmw_private *dev_priv,
896 struct vmw_fifo_state *fifo);
897 extern void vmw_fifo_release(struct vmw_private *dev_priv,
898 struct vmw_fifo_state *fifo);
899 extern void *
900 vmw_fifo_reserve_dx(struct vmw_private *dev_priv, uint32_t bytes, int ctx_id);
901 extern void vmw_fifo_commit(struct vmw_private *dev_priv, uint32_t bytes);
902 extern void vmw_fifo_commit_flush(struct vmw_private *dev_priv, uint32_t bytes);
903 extern int vmw_fifo_send_fence(struct vmw_private *dev_priv,
904 uint32_t *seqno);
905 extern void vmw_fifo_ping_host_locked(struct vmw_private *, uint32_t reason);
906 extern void vmw_fifo_ping_host(struct vmw_private *dev_priv, uint32_t reason);
907 extern bool vmw_fifo_have_3d(struct vmw_private *dev_priv);
908 extern bool vmw_fifo_have_pitchlock(struct vmw_private *dev_priv);
909 extern int vmw_fifo_emit_dummy_query(struct vmw_private *dev_priv,
910 uint32_t cid);
911 extern int vmw_fifo_flush(struct vmw_private *dev_priv,
912 bool interruptible);
913
914 #define VMW_FIFO_RESERVE_DX(__priv, __bytes, __ctx_id) \
915 ({ \
916 vmw_fifo_reserve_dx(__priv, __bytes, __ctx_id) ? : ({ \
917 DRM_ERROR("FIFO reserve failed at %s for %u bytes\n", \
918 __func__, (unsigned int) __bytes); \
919 NULL; \
920 }); \
921 })
922
923 #define VMW_FIFO_RESERVE(__priv, __bytes) \
924 VMW_FIFO_RESERVE_DX(__priv, __bytes, SVGA3D_INVALID_ID)
925
926 /**
927 * TTM glue - vmwgfx_ttm_glue.c
928 */
929
930 extern int vmw_mmap(struct file *filp, struct vm_area_struct *vma);
931
932 extern void vmw_validation_mem_init_ttm(struct vmw_private *dev_priv,
933 size_t gran);
934 /**
935 * TTM buffer object driver - vmwgfx_ttm_buffer.c
936 */
937
938 extern const size_t vmw_tt_size;
939 extern struct ttm_placement vmw_vram_placement;
940 extern struct ttm_placement vmw_vram_ne_placement;
941 extern struct ttm_placement vmw_vram_sys_placement;
942 extern struct ttm_placement vmw_vram_gmr_placement;
943 extern struct ttm_placement vmw_vram_gmr_ne_placement;
944 extern struct ttm_placement vmw_sys_placement;
945 extern struct ttm_placement vmw_sys_ne_placement;
946 extern struct ttm_placement vmw_evictable_placement;
947 extern struct ttm_placement vmw_srf_placement;
948 extern struct ttm_placement vmw_mob_placement;
949 extern struct ttm_placement vmw_mob_ne_placement;
950 extern struct ttm_placement vmw_nonfixed_placement;
951 extern struct ttm_bo_driver vmw_bo_driver;
952 extern int vmw_dma_quiescent(struct drm_device *dev);
953 extern int vmw_bo_map_dma(struct ttm_buffer_object *bo);
954 extern void vmw_bo_unmap_dma(struct ttm_buffer_object *bo);
955 extern const struct vmw_sg_table *
956 vmw_bo_sg_table(struct ttm_buffer_object *bo);
957 extern void vmw_piter_start(struct vmw_piter *viter,
958 const struct vmw_sg_table *vsgt,
959 unsigned long p_offs);
960
961 /**
962 * vmw_piter_next - Advance the iterator one page.
963 *
964 * @viter: Pointer to the iterator to advance.
965 *
966 * Returns false if past the list of pages, true otherwise.
967 */
968 static inline bool vmw_piter_next(struct vmw_piter *viter)
969 {
970 return viter->next(viter);
971 }
972
973 /**
974 * vmw_piter_dma_addr - Return the DMA address of the current page.
975 *
976 * @viter: Pointer to the iterator
977 *
978 * Returns the DMA address of the page pointed to by @viter.
979 */
980 static inline dma_addr_t vmw_piter_dma_addr(struct vmw_piter *viter)
981 {
982 return viter->dma_address(viter);
983 }
984
985 /**
986 * vmw_piter_page - Return a pointer to the current page.
987 *
988 * @viter: Pointer to the iterator
989 *
990 * Returns the DMA address of the page pointed to by @viter.
991 */
992 static inline struct page *vmw_piter_page(struct vmw_piter *viter)
993 {
994 return viter->page(viter);
995 }
996
997 /**
998 * Command submission - vmwgfx_execbuf.c
999 */
1000
1001 extern int vmw_execbuf_ioctl(struct drm_device *dev, void *data,
1002 struct drm_file *file_priv);
1003 extern int vmw_execbuf_process(struct drm_file *file_priv,
1004 struct vmw_private *dev_priv,
1005 void __user *user_commands,
1006 void *kernel_commands,
1007 uint32_t command_size,
1008 uint64_t throttle_us,
1009 uint32_t dx_context_handle,
1010 struct drm_vmw_fence_rep __user
1011 *user_fence_rep,
1012 struct vmw_fence_obj **out_fence,
1013 uint32_t flags);
1014 extern void __vmw_execbuf_release_pinned_bo(struct vmw_private *dev_priv,
1015 struct vmw_fence_obj *fence);
1016 extern void vmw_execbuf_release_pinned_bo(struct vmw_private *dev_priv);
1017
1018 extern int vmw_execbuf_fence_commands(struct drm_file *file_priv,
1019 struct vmw_private *dev_priv,
1020 struct vmw_fence_obj **p_fence,
1021 uint32_t *p_handle);
1022 extern void vmw_execbuf_copy_fence_user(struct vmw_private *dev_priv,
1023 struct vmw_fpriv *vmw_fp,
1024 int ret,
1025 struct drm_vmw_fence_rep __user
1026 *user_fence_rep,
1027 struct vmw_fence_obj *fence,
1028 uint32_t fence_handle,
1029 int32_t out_fence_fd,
1030 struct sync_file *sync_file);
1031 bool vmw_cmd_describe(const void *buf, u32 *size, char const **cmd);
1032
1033 /**
1034 * IRQs and wating - vmwgfx_irq.c
1035 */
1036
1037 extern int vmw_wait_seqno(struct vmw_private *dev_priv, bool lazy,
1038 uint32_t seqno, bool interruptible,
1039 unsigned long timeout);
1040 extern int vmw_irq_install(struct drm_device *dev, int irq);
1041 extern void vmw_irq_uninstall(struct drm_device *dev);
1042 extern bool vmw_seqno_passed(struct vmw_private *dev_priv,
1043 uint32_t seqno);
1044 extern int vmw_fallback_wait(struct vmw_private *dev_priv,
1045 bool lazy,
1046 bool fifo_idle,
1047 uint32_t seqno,
1048 bool interruptible,
1049 unsigned long timeout);
1050 extern void vmw_update_seqno(struct vmw_private *dev_priv,
1051 struct vmw_fifo_state *fifo_state);
1052 extern void vmw_seqno_waiter_add(struct vmw_private *dev_priv);
1053 extern void vmw_seqno_waiter_remove(struct vmw_private *dev_priv);
1054 extern void vmw_goal_waiter_add(struct vmw_private *dev_priv);
1055 extern void vmw_goal_waiter_remove(struct vmw_private *dev_priv);
1056 extern void vmw_generic_waiter_add(struct vmw_private *dev_priv, u32 flag,
1057 int *waiter_count);
1058 extern void vmw_generic_waiter_remove(struct vmw_private *dev_priv,
1059 u32 flag, int *waiter_count);
1060
1061 /**
1062 * Rudimentary fence-like objects currently used only for throttling -
1063 * vmwgfx_marker.c
1064 */
1065
1066 extern void vmw_marker_queue_init(struct vmw_marker_queue *queue);
1067 extern void vmw_marker_queue_takedown(struct vmw_marker_queue *queue);
1068 extern int vmw_marker_push(struct vmw_marker_queue *queue,
1069 uint32_t seqno);
1070 extern int vmw_marker_pull(struct vmw_marker_queue *queue,
1071 uint32_t signaled_seqno);
1072 extern int vmw_wait_lag(struct vmw_private *dev_priv,
1073 struct vmw_marker_queue *queue, uint32_t us);
1074
1075 /**
1076 * Kernel framebuffer - vmwgfx_fb.c
1077 */
1078
1079 int vmw_fb_init(struct vmw_private *vmw_priv);
1080 int vmw_fb_close(struct vmw_private *dev_priv);
1081 int vmw_fb_off(struct vmw_private *vmw_priv);
1082 int vmw_fb_on(struct vmw_private *vmw_priv);
1083
1084 /**
1085 * Kernel modesetting - vmwgfx_kms.c
1086 */
1087
1088 int vmw_kms_init(struct vmw_private *dev_priv);
1089 int vmw_kms_close(struct vmw_private *dev_priv);
1090 int vmw_kms_save_vga(struct vmw_private *vmw_priv);
1091 int vmw_kms_restore_vga(struct vmw_private *vmw_priv);
1092 int vmw_kms_cursor_bypass_ioctl(struct drm_device *dev, void *data,
1093 struct drm_file *file_priv);
1094 void vmw_kms_cursor_post_execbuf(struct vmw_private *dev_priv);
1095 void vmw_kms_cursor_snoop(struct vmw_surface *srf,
1096 struct ttm_object_file *tfile,
1097 struct ttm_buffer_object *bo,
1098 SVGA3dCmdHeader *header);
1099 int vmw_kms_write_svga(struct vmw_private *vmw_priv,
1100 unsigned width, unsigned height, unsigned pitch,
1101 unsigned bpp, unsigned depth);
1102 bool vmw_kms_validate_mode_vram(struct vmw_private *dev_priv,
1103 uint32_t pitch,
1104 uint32_t height);
1105 u32 vmw_get_vblank_counter(struct drm_device *dev, unsigned int pipe);
1106 int vmw_enable_vblank(struct drm_device *dev, unsigned int pipe);
1107 void vmw_disable_vblank(struct drm_device *dev, unsigned int pipe);
1108 int vmw_kms_present(struct vmw_private *dev_priv,
1109 struct drm_file *file_priv,
1110 struct vmw_framebuffer *vfb,
1111 struct vmw_surface *surface,
1112 uint32_t sid, int32_t destX, int32_t destY,
1113 struct drm_vmw_rect *clips,
1114 uint32_t num_clips);
1115 int vmw_kms_update_layout_ioctl(struct drm_device *dev, void *data,
1116 struct drm_file *file_priv);
1117 void vmw_kms_legacy_hotspot_clear(struct vmw_private *dev_priv);
1118 int vmw_kms_suspend(struct drm_device *dev);
1119 int vmw_kms_resume(struct drm_device *dev);
1120 void vmw_kms_lost_device(struct drm_device *dev);
1121
1122 int vmw_dumb_create(struct drm_file *file_priv,
1123 struct drm_device *dev,
1124 struct drm_mode_create_dumb *args);
1125
1126 int vmw_dumb_map_offset(struct drm_file *file_priv,
1127 struct drm_device *dev, uint32_t handle,
1128 uint64_t *offset);
1129 int vmw_dumb_destroy(struct drm_file *file_priv,
1130 struct drm_device *dev,
1131 uint32_t handle);
1132 extern int vmw_resource_pin(struct vmw_resource *res, bool interruptible);
1133 extern void vmw_resource_unpin(struct vmw_resource *res);
1134 extern enum vmw_res_type vmw_res_type(const struct vmw_resource *res);
1135
1136 /**
1137 * Overlay control - vmwgfx_overlay.c
1138 */
1139
1140 int vmw_overlay_init(struct vmw_private *dev_priv);
1141 int vmw_overlay_close(struct vmw_private *dev_priv);
1142 int vmw_overlay_ioctl(struct drm_device *dev, void *data,
1143 struct drm_file *file_priv);
1144 int vmw_overlay_stop_all(struct vmw_private *dev_priv);
1145 int vmw_overlay_resume_all(struct vmw_private *dev_priv);
1146 int vmw_overlay_pause_all(struct vmw_private *dev_priv);
1147 int vmw_overlay_claim(struct vmw_private *dev_priv, uint32_t *out);
1148 int vmw_overlay_unref(struct vmw_private *dev_priv, uint32_t stream_id);
1149 int vmw_overlay_num_overlays(struct vmw_private *dev_priv);
1150 int vmw_overlay_num_free_overlays(struct vmw_private *dev_priv);
1151
1152 /**
1153 * GMR Id manager
1154 */
1155
1156 extern const struct ttm_mem_type_manager_func vmw_gmrid_manager_func;
1157
1158 /**
1159 * Prime - vmwgfx_prime.c
1160 */
1161
1162 extern const struct dma_buf_ops vmw_prime_dmabuf_ops;
1163 extern int vmw_prime_fd_to_handle(struct drm_device *dev,
1164 struct drm_file *file_priv,
1165 int fd, u32 *handle);
1166 extern int vmw_prime_handle_to_fd(struct drm_device *dev,
1167 struct drm_file *file_priv,
1168 uint32_t handle, uint32_t flags,
1169 int *prime_fd);
1170
1171 /*
1172 * MemoryOBject management - vmwgfx_mob.c
1173 */
1174 struct vmw_mob;
1175 extern int vmw_mob_bind(struct vmw_private *dev_priv, struct vmw_mob *mob,
1176 const struct vmw_sg_table *vsgt,
1177 unsigned long num_data_pages, int32_t mob_id);
1178 extern void vmw_mob_unbind(struct vmw_private *dev_priv,
1179 struct vmw_mob *mob);
1180 extern void vmw_mob_destroy(struct vmw_mob *mob);
1181 extern struct vmw_mob *vmw_mob_create(unsigned long data_pages);
1182 extern int vmw_otables_setup(struct vmw_private *dev_priv);
1183 extern void vmw_otables_takedown(struct vmw_private *dev_priv);
1184
1185 /*
1186 * Context management - vmwgfx_context.c
1187 */
1188
1189 extern const struct vmw_user_resource_conv *user_context_converter;
1190
1191 extern int vmw_context_check(struct vmw_private *dev_priv,
1192 struct ttm_object_file *tfile,
1193 int id,
1194 struct vmw_resource **p_res);
1195 extern int vmw_context_define_ioctl(struct drm_device *dev, void *data,
1196 struct drm_file *file_priv);
1197 extern int vmw_extended_context_define_ioctl(struct drm_device *dev, void *data,
1198 struct drm_file *file_priv);
1199 extern int vmw_context_destroy_ioctl(struct drm_device *dev, void *data,
1200 struct drm_file *file_priv);
1201 extern struct list_head *vmw_context_binding_list(struct vmw_resource *ctx);
1202 extern struct vmw_cmdbuf_res_manager *
1203 vmw_context_res_man(struct vmw_resource *ctx);
1204 extern struct vmw_resource *vmw_context_cotable(struct vmw_resource *ctx,
1205 SVGACOTableType cotable_type);
1206 extern struct list_head *vmw_context_binding_list(struct vmw_resource *ctx);
1207 struct vmw_ctx_binding_state;
1208 extern struct vmw_ctx_binding_state *
1209 vmw_context_binding_state(struct vmw_resource *ctx);
1210 extern void vmw_dx_context_scrub_cotables(struct vmw_resource *ctx,
1211 bool readback);
1212 extern int vmw_context_bind_dx_query(struct vmw_resource *ctx_res,
1213 struct vmw_buffer_object *mob);
1214 extern struct vmw_buffer_object *
1215 vmw_context_get_dx_query_mob(struct vmw_resource *ctx_res);
1216
1217
1218 /*
1219 * Surface management - vmwgfx_surface.c
1220 */
1221
1222 extern const struct vmw_user_resource_conv *user_surface_converter;
1223
1224 extern void vmw_surface_res_free(struct vmw_resource *res);
1225 extern int vmw_surface_destroy_ioctl(struct drm_device *dev, void *data,
1226 struct drm_file *file_priv);
1227 extern int vmw_surface_define_ioctl(struct drm_device *dev, void *data,
1228 struct drm_file *file_priv);
1229 extern int vmw_surface_reference_ioctl(struct drm_device *dev, void *data,
1230 struct drm_file *file_priv);
1231 extern int vmw_gb_surface_define_ioctl(struct drm_device *dev, void *data,
1232 struct drm_file *file_priv);
1233 extern int vmw_gb_surface_reference_ioctl(struct drm_device *dev, void *data,
1234 struct drm_file *file_priv);
1235 extern int vmw_surface_check(struct vmw_private *dev_priv,
1236 struct ttm_object_file *tfile,
1237 uint32_t handle, int *id);
1238 extern int vmw_surface_validate(struct vmw_private *dev_priv,
1239 struct vmw_surface *srf);
1240 int vmw_surface_gb_priv_define(struct drm_device *dev,
1241 uint32_t user_accounting_size,
1242 SVGA3dSurfaceAllFlags svga3d_flags,
1243 SVGA3dSurfaceFormat format,
1244 bool for_scanout,
1245 uint32_t num_mip_levels,
1246 uint32_t multisample_count,
1247 uint32_t array_size,
1248 struct drm_vmw_size size,
1249 SVGA3dMSPattern multisample_pattern,
1250 SVGA3dMSQualityLevel quality_level,
1251 struct vmw_surface **srf_out);
1252 extern int vmw_gb_surface_define_ext_ioctl(struct drm_device *dev,
1253 void *data,
1254 struct drm_file *file_priv);
1255 extern int vmw_gb_surface_reference_ext_ioctl(struct drm_device *dev,
1256 void *data,
1257 struct drm_file *file_priv);
1258
1259 /*
1260 * Shader management - vmwgfx_shader.c
1261 */
1262
1263 extern const struct vmw_user_resource_conv *user_shader_converter;
1264
1265 extern int vmw_shader_define_ioctl(struct drm_device *dev, void *data,
1266 struct drm_file *file_priv);
1267 extern int vmw_shader_destroy_ioctl(struct drm_device *dev, void *data,
1268 struct drm_file *file_priv);
1269 extern int vmw_compat_shader_add(struct vmw_private *dev_priv,
1270 struct vmw_cmdbuf_res_manager *man,
1271 u32 user_key, const void *bytecode,
1272 SVGA3dShaderType shader_type,
1273 size_t size,
1274 struct list_head *list);
1275 extern int vmw_shader_remove(struct vmw_cmdbuf_res_manager *man,
1276 u32 user_key, SVGA3dShaderType shader_type,
1277 struct list_head *list);
1278 extern int vmw_dx_shader_add(struct vmw_cmdbuf_res_manager *man,
1279 struct vmw_resource *ctx,
1280 u32 user_key,
1281 SVGA3dShaderType shader_type,
1282 struct list_head *list);
1283 extern void vmw_dx_shader_cotable_list_scrub(struct vmw_private *dev_priv,
1284 struct list_head *list,
1285 bool readback);
1286
1287 extern struct vmw_resource *
1288 vmw_shader_lookup(struct vmw_cmdbuf_res_manager *man,
1289 u32 user_key, SVGA3dShaderType shader_type);
1290
1291 /*
1292 * Command buffer managed resources - vmwgfx_cmdbuf_res.c
1293 */
1294
1295 extern struct vmw_cmdbuf_res_manager *
1296 vmw_cmdbuf_res_man_create(struct vmw_private *dev_priv);
1297 extern void vmw_cmdbuf_res_man_destroy(struct vmw_cmdbuf_res_manager *man);
1298 extern size_t vmw_cmdbuf_res_man_size(void);
1299 extern struct vmw_resource *
1300 vmw_cmdbuf_res_lookup(struct vmw_cmdbuf_res_manager *man,
1301 enum vmw_cmdbuf_res_type res_type,
1302 u32 user_key);
1303 extern void vmw_cmdbuf_res_revert(struct list_head *list);
1304 extern void vmw_cmdbuf_res_commit(struct list_head *list);
1305 extern int vmw_cmdbuf_res_add(struct vmw_cmdbuf_res_manager *man,
1306 enum vmw_cmdbuf_res_type res_type,
1307 u32 user_key,
1308 struct vmw_resource *res,
1309 struct list_head *list);
1310 extern int vmw_cmdbuf_res_remove(struct vmw_cmdbuf_res_manager *man,
1311 enum vmw_cmdbuf_res_type res_type,
1312 u32 user_key,
1313 struct list_head *list,
1314 struct vmw_resource **res);
1315
1316 /*
1317 * COTable management - vmwgfx_cotable.c
1318 */
1319 extern const SVGACOTableType vmw_cotable_scrub_order[];
1320 extern struct vmw_resource *vmw_cotable_alloc(struct vmw_private *dev_priv,
1321 struct vmw_resource *ctx,
1322 u32 type);
1323 extern int vmw_cotable_notify(struct vmw_resource *res, int id);
1324 extern int vmw_cotable_scrub(struct vmw_resource *res, bool readback);
1325 extern void vmw_cotable_add_resource(struct vmw_resource *ctx,
1326 struct list_head *head);
1327
1328 /*
1329 * Command buffer managerment vmwgfx_cmdbuf.c
1330 */
1331 struct vmw_cmdbuf_man;
1332 struct vmw_cmdbuf_header;
1333
1334 extern struct vmw_cmdbuf_man *
1335 vmw_cmdbuf_man_create(struct vmw_private *dev_priv);
1336 extern int vmw_cmdbuf_set_pool_size(struct vmw_cmdbuf_man *man,
1337 size_t size, size_t default_size);
1338 extern void vmw_cmdbuf_remove_pool(struct vmw_cmdbuf_man *man);
1339 extern void vmw_cmdbuf_man_destroy(struct vmw_cmdbuf_man *man);
1340 extern int vmw_cmdbuf_idle(struct vmw_cmdbuf_man *man, bool interruptible,
1341 unsigned long timeout);
1342 extern void *vmw_cmdbuf_reserve(struct vmw_cmdbuf_man *man, size_t size,
1343 int ctx_id, bool interruptible,
1344 struct vmw_cmdbuf_header *header);
1345 extern void vmw_cmdbuf_commit(struct vmw_cmdbuf_man *man, size_t size,
1346 struct vmw_cmdbuf_header *header,
1347 bool flush);
1348 extern void *vmw_cmdbuf_alloc(struct vmw_cmdbuf_man *man,
1349 size_t size, bool interruptible,
1350 struct vmw_cmdbuf_header **p_header);
1351 extern void vmw_cmdbuf_header_free(struct vmw_cmdbuf_header *header);
1352 extern int vmw_cmdbuf_cur_flush(struct vmw_cmdbuf_man *man,
1353 bool interruptible);
1354 extern void vmw_cmdbuf_irqthread(struct vmw_cmdbuf_man *man);
1355
1356 /* CPU blit utilities - vmwgfx_blit.c */
1357
1358 /**
1359 * struct vmw_diff_cpy - CPU blit information structure
1360 *
1361 * @rect: The output bounding box rectangle.
1362 * @line: The current line of the blit.
1363 * @line_offset: Offset of the current line segment.
1364 * @cpp: Bytes per pixel (granularity information).
1365 * @memcpy: Which memcpy function to use.
1366 */
1367 struct vmw_diff_cpy {
1368 struct drm_rect rect;
1369 size_t line;
1370 size_t line_offset;
1371 int cpp;
1372 void (*do_cpy)(struct vmw_diff_cpy *diff, u8 *dest, const u8 *src,
1373 size_t n);
1374 };
1375
1376 #define VMW_CPU_BLIT_INITIALIZER { \
1377 .do_cpy = vmw_memcpy, \
1378 }
1379
1380 #define VMW_CPU_BLIT_DIFF_INITIALIZER(_cpp) { \
1381 .line = 0, \
1382 .line_offset = 0, \
1383 .rect = { .x1 = INT_MAX/2, \
1384 .y1 = INT_MAX/2, \
1385 .x2 = INT_MIN/2, \
1386 .y2 = INT_MIN/2 \
1387 }, \
1388 .cpp = _cpp, \
1389 .do_cpy = vmw_diff_memcpy, \
1390 }
1391
1392 void vmw_diff_memcpy(struct vmw_diff_cpy *diff, u8 *dest, const u8 *src,
1393 size_t n);
1394
1395 void vmw_memcpy(struct vmw_diff_cpy *diff, u8 *dest, const u8 *src, size_t n);
1396
1397 int vmw_bo_cpu_blit(struct ttm_buffer_object *dst,
1398 u32 dst_offset, u32 dst_stride,
1399 struct ttm_buffer_object *src,
1400 u32 src_offset, u32 src_stride,
1401 u32 w, u32 h,
1402 struct vmw_diff_cpy *diff);
1403
1404 /* Host messaging -vmwgfx_msg.c: */
1405 int vmw_host_get_guestinfo(const char *guest_info_param,
1406 char *buffer, size_t *length);
1407 int vmw_host_log(const char *log);
1408 int vmw_msg_ioctl(struct drm_device *dev, void *data,
1409 struct drm_file *file_priv);
1410
1411 /* VMW logging */
1412
1413 /**
1414 * VMW_DEBUG_USER - Debug output for user-space debugging.
1415 *
1416 * @fmt: printf() like format string.
1417 *
1418 * This macro is for logging user-space error and debugging messages for e.g.
1419 * command buffer execution errors due to malformed commands, invalid context,
1420 * etc.
1421 */
1422 #define VMW_DEBUG_USER(fmt, ...) \
1423 DRM_DEBUG_DRIVER(fmt, ##__VA_ARGS__)
1424
1425 /* Resource dirtying - vmwgfx_page_dirty.c */
1426 void vmw_bo_dirty_scan(struct vmw_buffer_object *vbo);
1427 int vmw_bo_dirty_add(struct vmw_buffer_object *vbo);
1428 void vmw_bo_dirty_transfer_to_res(struct vmw_resource *res);
1429 void vmw_bo_dirty_clear_res(struct vmw_resource *res);
1430 void vmw_bo_dirty_release(struct vmw_buffer_object *vbo);
1431 void vmw_bo_dirty_unmap(struct vmw_buffer_object *vbo,
1432 pgoff_t start, pgoff_t end);
1433 vm_fault_t vmw_bo_vm_fault(struct vm_fault *vmf);
1434 vm_fault_t vmw_bo_vm_mkwrite(struct vm_fault *vmf);
1435
1436 /**
1437 * VMW_DEBUG_KMS - Debug output for kernel mode-setting
1438 *
1439 * This macro is for debugging vmwgfx mode-setting code.
1440 */
1441 #define VMW_DEBUG_KMS(fmt, ...) \
1442 DRM_DEBUG_DRIVER(fmt, ##__VA_ARGS__)
1443
1444 /**
1445 * Inline helper functions
1446 */
1447
1448 static inline void vmw_surface_unreference(struct vmw_surface **srf)
1449 {
1450 struct vmw_surface *tmp_srf = *srf;
1451 struct vmw_resource *res = &tmp_srf->res;
1452 *srf = NULL;
1453
1454 vmw_resource_unreference(&res);
1455 }
1456
1457 static inline struct vmw_surface *vmw_surface_reference(struct vmw_surface *srf)
1458 {
1459 (void) vmw_resource_reference(&srf->res);
1460 return srf;
1461 }
1462
1463 static inline void vmw_bo_unreference(struct vmw_buffer_object **buf)
1464 {
1465 struct vmw_buffer_object *tmp_buf = *buf;
1466
1467 *buf = NULL;
1468 if (tmp_buf != NULL) {
1469 ttm_bo_put(&tmp_buf->base);
1470 }
1471 }
1472
1473 static inline struct vmw_buffer_object *
1474 vmw_bo_reference(struct vmw_buffer_object *buf)
1475 {
1476 ttm_bo_get(&buf->base);
1477 return buf;
1478 }
1479
1480 static inline struct ttm_mem_global *vmw_mem_glob(struct vmw_private *dev_priv)
1481 {
1482 return &ttm_mem_glob;
1483 }
1484
1485 static inline void vmw_fifo_resource_inc(struct vmw_private *dev_priv)
1486 {
1487 atomic_inc(&dev_priv->num_fifo_resources);
1488 }
1489
1490 static inline void vmw_fifo_resource_dec(struct vmw_private *dev_priv)
1491 {
1492 atomic_dec(&dev_priv->num_fifo_resources);
1493 }
1494
1495 /**
1496 * vmw_mmio_read - Perform a MMIO read from volatile memory
1497 *
1498 * @addr: The address to read from
1499 *
1500 * This function is intended to be equivalent to ioread32() on
1501 * memremap'd memory, but without byteswapping.
1502 */
1503 static inline u32 vmw_mmio_read(u32 *addr)
1504 {
1505 return READ_ONCE(*addr);
1506 }
1507
1508 /**
1509 * vmw_mmio_write - Perform a MMIO write to volatile memory
1510 *
1511 * @addr: The address to write to
1512 *
1513 * This function is intended to be equivalent to iowrite32 on
1514 * memremap'd memory, but without byteswapping.
1515 */
1516 static inline void vmw_mmio_write(u32 value, u32 *addr)
1517 {
1518 WRITE_ONCE(*addr, value);
1519 }
1520 #endif
1521