1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28#ifndef _VMWGFX_DRV_H_
29#define _VMWGFX_DRV_H_
30
31#include <linux/suspend.h>
32#include <linux/sync_file.h>
33
34#include <drm/drm_auth.h>
35#include <drm/drm_device.h>
36#include <drm/drm_file.h>
37#include <drm/drm_hashtab.h>
38#include <drm/drm_rect.h>
39
40#include <drm/ttm/ttm_bo_driver.h>
41#include <drm/ttm/ttm_execbuf_util.h>
42#include <drm/ttm/ttm_module.h>
43
44#include "ttm_lock.h"
45#include "ttm_object.h"
46
47#include "vmwgfx_fence.h"
48#include "vmwgfx_reg.h"
49#include "vmwgfx_validation.h"
50
51
52
53
54
55#include <drm/vmwgfx_drm.h>
56
57
58#define VMWGFX_DRIVER_NAME "vmwgfx"
59#define VMWGFX_DRIVER_DATE "20200114"
60#define VMWGFX_DRIVER_MAJOR 2
61#define VMWGFX_DRIVER_MINOR 18
62#define VMWGFX_DRIVER_PATCHLEVEL 0
63#define VMWGFX_FIFO_STATIC_SIZE (1024*1024)
64#define VMWGFX_MAX_RELOCATIONS 2048
65#define VMWGFX_MAX_VALIDATIONS 2048
66#define VMWGFX_MAX_DISPLAYS 16
67#define VMWGFX_CMD_BOUNCE_INIT_SIZE 32768
68#define VMWGFX_ENABLE_SCREEN_TARGET_OTABLE 1
69
70
71
72
73#define VMWGFX_NUM_GB_CONTEXT 256
74#define VMWGFX_NUM_GB_SHADER 20000
75#define VMWGFX_NUM_GB_SURFACE 32768
76#define VMWGFX_NUM_GB_SCREEN_TARGET VMWGFX_MAX_DISPLAYS
77#define VMWGFX_NUM_DXCONTEXT 256
78#define VMWGFX_NUM_DXQUERY 512
79#define VMWGFX_NUM_MOB (VMWGFX_NUM_GB_CONTEXT +\
80 VMWGFX_NUM_GB_SHADER +\
81 VMWGFX_NUM_GB_SURFACE +\
82 VMWGFX_NUM_GB_SCREEN_TARGET)
83
84#define VMW_PL_GMR (TTM_PL_PRIV + 0)
85#define VMW_PL_MOB (TTM_PL_PRIV + 1)
86
87#define VMW_RES_CONTEXT ttm_driver_type0
88#define VMW_RES_SURFACE ttm_driver_type1
89#define VMW_RES_STREAM ttm_driver_type2
90#define VMW_RES_FENCE ttm_driver_type3
91#define VMW_RES_SHADER ttm_driver_type4
92
93struct vmw_fpriv {
94 struct ttm_object_file *tfile;
95 bool gb_aware;
96};
97
98
99
100
101
102
103
104
105
106
107
108
109
110struct vmw_buffer_object {
111 struct ttm_buffer_object base;
112 struct rb_root res_tree;
113 s32 pin_count;
114 atomic_t cpu_writers;
115
116 struct vmw_resource *dx_query_ctx;
117
118 struct ttm_bo_kmap_obj map;
119 u32 res_prios[TTM_MAX_BO_PRIORITY];
120 struct vmw_bo_dirty *dirty;
121};
122
123
124
125
126
127
128
129
130
131
132struct vmw_validate_buffer {
133 struct ttm_validate_buffer base;
134 struct drm_hash_item hash;
135 bool validate_as_mob;
136};
137
138struct vmw_res_func;
139
140
141
142
143
144
145
146
147
148
149
150
151
152
153
154
155
156
157
158
159
160
161
162
163
164
165
166
167
168
169struct vmw_resource_dirty;
170struct vmw_resource {
171 struct kref kref;
172 struct vmw_private *dev_priv;
173 int id;
174 u32 used_prio;
175 unsigned long backup_size;
176 u32 res_dirty : 1;
177 u32 backup_dirty : 1;
178 u32 coherent : 1;
179 struct vmw_buffer_object *backup;
180 unsigned long backup_offset;
181 unsigned long pin_count;
182 const struct vmw_res_func *func;
183 struct rb_node mob_node;
184 struct list_head lru_head;
185 struct list_head binding_head;
186 struct vmw_resource_dirty *dirty;
187 void (*res_free) (struct vmw_resource *res);
188 void (*hw_destroy) (struct vmw_resource *res);
189};
190
191
192
193
194
195enum vmw_res_type {
196 vmw_res_context,
197 vmw_res_surface,
198 vmw_res_stream,
199 vmw_res_shader,
200 vmw_res_dx_context,
201 vmw_res_cotable,
202 vmw_res_view,
203 vmw_res_streamoutput,
204 vmw_res_max
205};
206
207
208
209
210enum vmw_cmdbuf_res_type {
211 vmw_cmdbuf_res_shader,
212 vmw_cmdbuf_res_view,
213 vmw_cmdbuf_res_streamoutput
214};
215
216struct vmw_cmdbuf_res_manager;
217
218struct vmw_cursor_snooper {
219 size_t age;
220 uint32_t *image;
221};
222
223struct vmw_framebuffer;
224struct vmw_surface_offset;
225
226
227
228
229
230
231
232
233
234
235
236
237
238
239
240
241
242
243
244
245
246
247struct vmw_surface_metadata {
248 u64 flags;
249 u32 format;
250 u32 mip_levels[DRM_VMW_MAX_SURFACE_FACES];
251 u32 multisample_count;
252 u32 multisample_pattern;
253 u32 quality_level;
254 u32 autogen_filter;
255 u32 array_size;
256 u32 num_sizes;
257 u32 buffer_byte_stride;
258 struct drm_vmw_size base_size;
259 struct drm_vmw_size *sizes;
260 bool scanout;
261};
262
263
264
265
266
267
268
269
270
271
272struct vmw_surface {
273 struct vmw_resource res;
274 struct vmw_surface_metadata metadata;
275 struct vmw_cursor_snooper snooper;
276 struct vmw_surface_offset *offsets;
277 struct list_head view_list;
278};
279
280struct vmw_marker_queue {
281 struct list_head head;
282 u64 lag;
283 u64 lag_time;
284 spinlock_t lock;
285};
286
287struct vmw_fifo_state {
288 unsigned long reserved_size;
289 u32 *dynamic_buffer;
290 u32 *static_buffer;
291 unsigned long static_buffer_size;
292 bool using_bounce_buffer;
293 uint32_t capabilities;
294 struct mutex fifo_mutex;
295 struct rw_semaphore rwsem;
296 struct vmw_marker_queue marker_queue;
297 bool dx;
298};
299
300
301
302
303
304
305
306
307
308
309
310
311
312struct vmw_res_cache_entry {
313 uint32_t handle;
314 struct vmw_resource *res;
315 void *private;
316 unsigned short valid_handle;
317 unsigned short valid;
318};
319
320
321
322
323enum vmw_dma_map_mode {
324 vmw_dma_phys,
325 vmw_dma_alloc_coherent,
326 vmw_dma_map_populate,
327 vmw_dma_map_bind,
328 vmw_dma_map_max
329};
330
331
332
333
334
335
336
337
338struct vmw_sg_table {
339 enum vmw_dma_map_mode mode;
340 struct page **pages;
341 const dma_addr_t *addrs;
342 struct sg_table *sgt;
343 unsigned long num_regions;
344 unsigned long num_pages;
345};
346
347
348
349
350
351
352
353
354
355
356
357
358
359
360
361struct vmw_piter {
362 struct page **pages;
363 const dma_addr_t *addrs;
364 struct sg_dma_page_iter iter;
365 unsigned long i;
366 unsigned long num_pages;
367 bool (*next)(struct vmw_piter *);
368 dma_addr_t (*dma_address)(struct vmw_piter *);
369 struct page *(*page)(struct vmw_piter *);
370};
371
372
373
374
375enum vmw_display_unit_type {
376 vmw_du_invalid = 0,
377 vmw_du_legacy,
378 vmw_du_screen_object,
379 vmw_du_screen_target
380};
381
382struct vmw_validation_context;
383struct vmw_ctx_validation_info;
384
385
386
387
388
389
390
391
392
393
394
395
396
397
398
399
400
401
402
403
404
405
406
407
408
409
410
411
412
413
414
415
416struct vmw_sw_context{
417 struct drm_open_hash res_ht;
418 bool res_ht_initialized;
419 bool kernel;
420 struct vmw_fpriv *fp;
421 uint32_t *cmd_bounce;
422 uint32_t cmd_bounce_size;
423 struct vmw_buffer_object *cur_query_bo;
424 struct list_head bo_relocations;
425 struct list_head res_relocations;
426 uint32_t *buf_start;
427 struct vmw_res_cache_entry res_cache[vmw_res_max];
428 struct vmw_resource *last_query_ctx;
429 bool needs_post_query_barrier;
430 struct vmw_ctx_binding_state *staged_bindings;
431 bool staged_bindings_inuse;
432 struct list_head staged_cmd_res;
433 struct list_head ctx_list;
434 struct vmw_ctx_validation_info *dx_ctx_node;
435 struct vmw_buffer_object *dx_query_mob;
436 struct vmw_resource *dx_query_ctx;
437 struct vmw_cmdbuf_res_manager *man;
438 struct vmw_validation_context *ctx;
439};
440
441struct vmw_legacy_display;
442struct vmw_overlay;
443
444struct vmw_vga_topology_state {
445 uint32_t width;
446 uint32_t height;
447 uint32_t primary;
448 uint32_t pos_x;
449 uint32_t pos_y;
450};
451
452
453
454
455
456
457
458
459struct vmw_otable {
460 unsigned long size;
461 struct vmw_mob *page_table;
462 bool enabled;
463};
464
465struct vmw_otable_batch {
466 unsigned num_otables;
467 struct vmw_otable *otables;
468 struct vmw_resource *context;
469 struct ttm_buffer_object *otable_bo;
470};
471
472enum {
473 VMW_IRQTHREAD_FENCE,
474 VMW_IRQTHREAD_CMDBUF,
475 VMW_IRQTHREAD_MAX
476};
477
478
479
480
481
482
483
484
485
486enum vmw_sm_type {
487 VMW_SM_LEGACY = 0,
488 VMW_SM_4,
489 VMW_SM_4_1,
490 VMW_SM_5,
491 VMW_SM_MAX
492};
493
494struct vmw_private {
495 struct ttm_bo_device bdev;
496
497 struct vmw_fifo_state fifo;
498
499 struct drm_device *dev;
500 struct drm_vma_offset_manager vma_manager;
501 unsigned long vmw_chipset;
502 unsigned int io_start;
503 uint32_t vram_start;
504 uint32_t vram_size;
505 uint32_t prim_bb_mem;
506 uint32_t mmio_start;
507 uint32_t mmio_size;
508 uint32_t fb_max_width;
509 uint32_t fb_max_height;
510 uint32_t texture_max_width;
511 uint32_t texture_max_height;
512 uint32_t stdu_max_width;
513 uint32_t stdu_max_height;
514 uint32_t initial_width;
515 uint32_t initial_height;
516 u32 *mmio_virt;
517 uint32_t capabilities;
518 uint32_t capabilities2;
519 uint32_t max_gmr_ids;
520 uint32_t max_gmr_pages;
521 uint32_t max_mob_pages;
522 uint32_t max_mob_size;
523 uint32_t memory_size;
524 bool has_gmr;
525 bool has_mob;
526 spinlock_t hw_lock;
527 spinlock_t cap_lock;
528 bool assume_16bpp;
529
530 enum vmw_sm_type sm_type;
531
532
533
534
535
536 void *fb_info;
537 enum vmw_display_unit_type active_display_unit;
538 struct vmw_legacy_display *ldu_priv;
539 struct vmw_overlay *overlay_priv;
540 struct drm_property *hotplug_mode_update_property;
541 struct drm_property *implicit_placement_property;
542 struct mutex global_kms_state_mutex;
543 spinlock_t cursor_lock;
544 struct drm_atomic_state *suspend_state;
545
546
547
548
549
550 spinlock_t resource_lock;
551 struct idr res_idr[vmw_res_max];
552
553
554
555
556
557
558 struct ttm_object_device *tdev;
559
560
561
562
563
564 atomic_t marker_seq;
565 wait_queue_head_t fence_queue;
566 wait_queue_head_t fifo_queue;
567 spinlock_t waiter_lock;
568 int fence_queue_waiters;
569 int goal_queue_waiters;
570 int cmdbuf_waiters;
571 int error_waiters;
572 int fifo_queue_waiters;
573 uint32_t last_read_seqno;
574 struct vmw_fence_manager *fman;
575 uint32_t irq_mask;
576
577
578
579
580
581 uint32_t traces_state;
582 uint32_t enable_state;
583 uint32_t config_done_state;
584
585
586
587
588
589
590
591
592 struct vmw_sw_context ctx;
593 struct mutex cmdbuf_mutex;
594 struct mutex binding_mutex;
595
596
597
598
599
600 bool stealth;
601 bool enable_fb;
602 spinlock_t svga_lock;
603
604
605
606
607 struct notifier_block pm_nb;
608 bool refuse_hibernation;
609 bool suspend_locked;
610
611 struct mutex release_mutex;
612 atomic_t num_fifo_resources;
613
614
615
616
617 struct ttm_lock reservation_sem;
618
619
620
621
622
623
624 struct vmw_buffer_object *dummy_query_bo;
625 struct vmw_buffer_object *pinned_bo;
626 uint32_t query_cid;
627 uint32_t query_cid_valid;
628 bool dummy_query_bo_pinned;
629
630
631
632
633
634
635
636
637 struct list_head res_lru[vmw_res_max];
638 uint32_t used_memory_size;
639
640
641
642
643 enum vmw_dma_map_mode map_mode;
644
645
646
647
648 struct vmw_otable_batch otable_batch;
649
650 struct vmw_cmdbuf_man *cman;
651 DECLARE_BITMAP(irqthread_pending, VMW_IRQTHREAD_MAX);
652
653
654 struct vmw_validation_mem vvm;
655};
656
657static inline struct vmw_surface *vmw_res_to_srf(struct vmw_resource *res)
658{
659 return container_of(res, struct vmw_surface, res);
660}
661
662static inline struct vmw_private *vmw_priv(struct drm_device *dev)
663{
664 return (struct vmw_private *)dev->dev_private;
665}
666
667static inline struct vmw_fpriv *vmw_fpriv(struct drm_file *file_priv)
668{
669 return (struct vmw_fpriv *)file_priv->driver_priv;
670}
671
672
673
674
675
676
677
678
679static inline void vmw_write(struct vmw_private *dev_priv,
680 unsigned int offset, uint32_t value)
681{
682 spin_lock(&dev_priv->hw_lock);
683 outl(offset, dev_priv->io_start + VMWGFX_INDEX_PORT);
684 outl(value, dev_priv->io_start + VMWGFX_VALUE_PORT);
685 spin_unlock(&dev_priv->hw_lock);
686}
687
688static inline uint32_t vmw_read(struct vmw_private *dev_priv,
689 unsigned int offset)
690{
691 u32 val;
692
693 spin_lock(&dev_priv->hw_lock);
694 outl(offset, dev_priv->io_start + VMWGFX_INDEX_PORT);
695 val = inl(dev_priv->io_start + VMWGFX_VALUE_PORT);
696 spin_unlock(&dev_priv->hw_lock);
697
698 return val;
699}
700
701
702
703
704
705
706
707static inline bool has_sm4_context(const struct vmw_private *dev_priv)
708{
709 return (dev_priv->sm_type >= VMW_SM_4);
710}
711
712
713
714
715
716
717
718static inline bool has_sm4_1_context(const struct vmw_private *dev_priv)
719{
720 return (dev_priv->sm_type >= VMW_SM_4_1);
721}
722
723
724
725
726
727
728
729static inline bool has_sm5_context(const struct vmw_private *dev_priv)
730{
731 return (dev_priv->sm_type >= VMW_SM_5);
732}
733
734extern void vmw_svga_enable(struct vmw_private *dev_priv);
735extern void vmw_svga_disable(struct vmw_private *dev_priv);
736
737
738
739
740
741
742extern int vmw_gmr_bind(struct vmw_private *dev_priv,
743 const struct vmw_sg_table *vsgt,
744 unsigned long num_pages,
745 int gmr_id);
746extern void vmw_gmr_unbind(struct vmw_private *dev_priv, int gmr_id);
747
748
749
750
751struct vmw_user_resource_conv;
752
753extern void vmw_resource_unreference(struct vmw_resource **p_res);
754extern struct vmw_resource *vmw_resource_reference(struct vmw_resource *res);
755extern struct vmw_resource *
756vmw_resource_reference_unless_doomed(struct vmw_resource *res);
757extern int vmw_resource_validate(struct vmw_resource *res, bool intr,
758 bool dirtying);
759extern int vmw_resource_reserve(struct vmw_resource *res, bool interruptible,
760 bool no_backup);
761extern bool vmw_resource_needs_backup(const struct vmw_resource *res);
762extern int vmw_user_lookup_handle(struct vmw_private *dev_priv,
763 struct ttm_object_file *tfile,
764 uint32_t handle,
765 struct vmw_surface **out_surf,
766 struct vmw_buffer_object **out_buf);
767extern int vmw_user_resource_lookup_handle(
768 struct vmw_private *dev_priv,
769 struct ttm_object_file *tfile,
770 uint32_t handle,
771 const struct vmw_user_resource_conv *converter,
772 struct vmw_resource **p_res);
773extern struct vmw_resource *
774vmw_user_resource_noref_lookup_handle(struct vmw_private *dev_priv,
775 struct ttm_object_file *tfile,
776 uint32_t handle,
777 const struct vmw_user_resource_conv *
778 converter);
779extern int vmw_stream_claim_ioctl(struct drm_device *dev, void *data,
780 struct drm_file *file_priv);
781extern int vmw_stream_unref_ioctl(struct drm_device *dev, void *data,
782 struct drm_file *file_priv);
783extern int vmw_user_stream_lookup(struct vmw_private *dev_priv,
784 struct ttm_object_file *tfile,
785 uint32_t *inout_id,
786 struct vmw_resource **out);
787extern void vmw_resource_unreserve(struct vmw_resource *res,
788 bool dirty_set,
789 bool dirty,
790 bool switch_backup,
791 struct vmw_buffer_object *new_backup,
792 unsigned long new_backup_offset);
793extern void vmw_query_move_notify(struct ttm_buffer_object *bo,
794 struct ttm_resource *mem);
795extern int vmw_query_readback_all(struct vmw_buffer_object *dx_query_mob);
796extern void vmw_resource_evict_all(struct vmw_private *dev_priv);
797extern void vmw_resource_unbind_list(struct vmw_buffer_object *vbo);
798void vmw_resource_mob_attach(struct vmw_resource *res);
799void vmw_resource_mob_detach(struct vmw_resource *res);
800void vmw_resource_dirty_update(struct vmw_resource *res, pgoff_t start,
801 pgoff_t end);
802int vmw_resources_clean(struct vmw_buffer_object *vbo, pgoff_t start,
803 pgoff_t end, pgoff_t *num_prefault);
804
805
806
807
808
809
810
811static inline bool vmw_resource_mob_attached(const struct vmw_resource *res)
812{
813 return !RB_EMPTY_NODE(&res->mob_node);
814}
815
816
817
818
819
820static inline void vmw_user_resource_noref_release(void)
821{
822 ttm_base_object_noref_release();
823}
824
825
826
827
828extern int vmw_bo_pin_in_placement(struct vmw_private *vmw_priv,
829 struct vmw_buffer_object *bo,
830 struct ttm_placement *placement,
831 bool interruptible);
832extern int vmw_bo_pin_in_vram(struct vmw_private *dev_priv,
833 struct vmw_buffer_object *buf,
834 bool interruptible);
835extern int vmw_bo_pin_in_vram_or_gmr(struct vmw_private *dev_priv,
836 struct vmw_buffer_object *buf,
837 bool interruptible);
838extern int vmw_bo_pin_in_start_of_vram(struct vmw_private *vmw_priv,
839 struct vmw_buffer_object *bo,
840 bool interruptible);
841extern int vmw_bo_unpin(struct vmw_private *vmw_priv,
842 struct vmw_buffer_object *bo,
843 bool interruptible);
844extern void vmw_bo_get_guest_ptr(const struct ttm_buffer_object *buf,
845 SVGAGuestPtr *ptr);
846extern void vmw_bo_pin_reserved(struct vmw_buffer_object *bo, bool pin);
847extern void vmw_bo_bo_free(struct ttm_buffer_object *bo);
848extern int vmw_bo_init(struct vmw_private *dev_priv,
849 struct vmw_buffer_object *vmw_bo,
850 size_t size, struct ttm_placement *placement,
851 bool interruptible,
852 void (*bo_free)(struct ttm_buffer_object *bo));
853extern int vmw_user_bo_verify_access(struct ttm_buffer_object *bo,
854 struct ttm_object_file *tfile);
855extern int vmw_user_bo_alloc(struct vmw_private *dev_priv,
856 struct ttm_object_file *tfile,
857 uint32_t size,
858 bool shareable,
859 uint32_t *handle,
860 struct vmw_buffer_object **p_dma_buf,
861 struct ttm_base_object **p_base);
862extern int vmw_user_bo_reference(struct ttm_object_file *tfile,
863 struct vmw_buffer_object *dma_buf,
864 uint32_t *handle);
865extern int vmw_bo_alloc_ioctl(struct drm_device *dev, void *data,
866 struct drm_file *file_priv);
867extern int vmw_bo_unref_ioctl(struct drm_device *dev, void *data,
868 struct drm_file *file_priv);
869extern int vmw_user_bo_synccpu_ioctl(struct drm_device *dev, void *data,
870 struct drm_file *file_priv);
871extern int vmw_user_bo_lookup(struct ttm_object_file *tfile,
872 uint32_t id, struct vmw_buffer_object **out,
873 struct ttm_base_object **base);
874extern void vmw_bo_fence_single(struct ttm_buffer_object *bo,
875 struct vmw_fence_obj *fence);
876extern void *vmw_bo_map_and_cache(struct vmw_buffer_object *vbo);
877extern void vmw_bo_unmap(struct vmw_buffer_object *vbo);
878extern void vmw_bo_move_notify(struct ttm_buffer_object *bo,
879 struct ttm_resource *mem);
880extern void vmw_bo_swap_notify(struct ttm_buffer_object *bo);
881extern struct vmw_buffer_object *
882vmw_user_bo_noref_lookup(struct ttm_object_file *tfile, u32 handle);
883
884
885
886
887
888static inline void vmw_user_bo_noref_release(void)
889{
890 ttm_base_object_noref_release();
891}
892
893
894
895
896
897
898static inline void vmw_bo_prio_adjust(struct vmw_buffer_object *vbo)
899{
900 int i = ARRAY_SIZE(vbo->res_prios);
901
902 while (i--) {
903 if (vbo->res_prios[i]) {
904 vbo->base.priority = i;
905 return;
906 }
907 }
908
909 vbo->base.priority = 3;
910}
911
912
913
914
915
916
917
918
919
920
921static inline void vmw_bo_prio_add(struct vmw_buffer_object *vbo, int prio)
922{
923 if (vbo->res_prios[prio]++ == 0)
924 vmw_bo_prio_adjust(vbo);
925}
926
927
928
929
930
931
932
933
934
935
936static inline void vmw_bo_prio_del(struct vmw_buffer_object *vbo, int prio)
937{
938 if (--vbo->res_prios[prio] == 0)
939 vmw_bo_prio_adjust(vbo);
940}
941
942
943
944
945
946extern int vmw_getparam_ioctl(struct drm_device *dev, void *data,
947 struct drm_file *file_priv);
948extern int vmw_get_cap_3d_ioctl(struct drm_device *dev, void *data,
949 struct drm_file *file_priv);
950extern int vmw_present_ioctl(struct drm_device *dev, void *data,
951 struct drm_file *file_priv);
952extern int vmw_present_readback_ioctl(struct drm_device *dev, void *data,
953 struct drm_file *file_priv);
954extern __poll_t vmw_fops_poll(struct file *filp,
955 struct poll_table_struct *wait);
956extern ssize_t vmw_fops_read(struct file *filp, char __user *buffer,
957 size_t count, loff_t *offset);
958
959
960
961
962
963extern int vmw_fifo_init(struct vmw_private *dev_priv,
964 struct vmw_fifo_state *fifo);
965extern void vmw_fifo_release(struct vmw_private *dev_priv,
966 struct vmw_fifo_state *fifo);
967extern void *
968vmw_fifo_reserve_dx(struct vmw_private *dev_priv, uint32_t bytes, int ctx_id);
969extern void vmw_fifo_commit(struct vmw_private *dev_priv, uint32_t bytes);
970extern void vmw_fifo_commit_flush(struct vmw_private *dev_priv, uint32_t bytes);
971extern int vmw_fifo_send_fence(struct vmw_private *dev_priv,
972 uint32_t *seqno);
973extern void vmw_fifo_ping_host(struct vmw_private *dev_priv, uint32_t reason);
974extern bool vmw_fifo_have_3d(struct vmw_private *dev_priv);
975extern bool vmw_fifo_have_pitchlock(struct vmw_private *dev_priv);
976extern int vmw_fifo_emit_dummy_query(struct vmw_private *dev_priv,
977 uint32_t cid);
978extern int vmw_fifo_flush(struct vmw_private *dev_priv,
979 bool interruptible);
980
981#define VMW_FIFO_RESERVE_DX(__priv, __bytes, __ctx_id) \
982({ \
983 vmw_fifo_reserve_dx(__priv, __bytes, __ctx_id) ? : ({ \
984 DRM_ERROR("FIFO reserve failed at %s for %u bytes\n", \
985 __func__, (unsigned int) __bytes); \
986 NULL; \
987 }); \
988})
989
990#define VMW_FIFO_RESERVE(__priv, __bytes) \
991 VMW_FIFO_RESERVE_DX(__priv, __bytes, SVGA3D_INVALID_ID)
992
993
994
995
996
997extern int vmw_mmap(struct file *filp, struct vm_area_struct *vma);
998
999extern void vmw_validation_mem_init_ttm(struct vmw_private *dev_priv,
1000 size_t gran);
1001
1002
1003
1004
1005
1006extern const size_t vmw_tt_size;
1007extern struct ttm_placement vmw_vram_placement;
1008extern struct ttm_placement vmw_vram_ne_placement;
1009extern struct ttm_placement vmw_vram_sys_placement;
1010extern struct ttm_placement vmw_vram_gmr_placement;
1011extern struct ttm_placement vmw_vram_gmr_ne_placement;
1012extern struct ttm_placement vmw_sys_placement;
1013extern struct ttm_placement vmw_sys_ne_placement;
1014extern struct ttm_placement vmw_evictable_placement;
1015extern struct ttm_placement vmw_srf_placement;
1016extern struct ttm_placement vmw_mob_placement;
1017extern struct ttm_placement vmw_mob_ne_placement;
1018extern struct ttm_placement vmw_nonfixed_placement;
1019extern struct ttm_bo_driver vmw_bo_driver;
1020extern const struct vmw_sg_table *
1021vmw_bo_sg_table(struct ttm_buffer_object *bo);
1022extern int vmw_bo_create_and_populate(struct vmw_private *dev_priv,
1023 unsigned long bo_size,
1024 struct ttm_buffer_object **bo_p);
1025
1026extern void vmw_piter_start(struct vmw_piter *viter,
1027 const struct vmw_sg_table *vsgt,
1028 unsigned long p_offs);
1029
1030
1031
1032
1033
1034
1035
1036
1037static inline bool vmw_piter_next(struct vmw_piter *viter)
1038{
1039 return viter->next(viter);
1040}
1041
1042
1043
1044
1045
1046
1047
1048
1049static inline dma_addr_t vmw_piter_dma_addr(struct vmw_piter *viter)
1050{
1051 return viter->dma_address(viter);
1052}
1053
1054
1055
1056
1057
1058
1059
1060
1061static inline struct page *vmw_piter_page(struct vmw_piter *viter)
1062{
1063 return viter->page(viter);
1064}
1065
1066
1067
1068
1069
1070extern int vmw_execbuf_ioctl(struct drm_device *dev, void *data,
1071 struct drm_file *file_priv);
1072extern int vmw_execbuf_process(struct drm_file *file_priv,
1073 struct vmw_private *dev_priv,
1074 void __user *user_commands,
1075 void *kernel_commands,
1076 uint32_t command_size,
1077 uint64_t throttle_us,
1078 uint32_t dx_context_handle,
1079 struct drm_vmw_fence_rep __user
1080 *user_fence_rep,
1081 struct vmw_fence_obj **out_fence,
1082 uint32_t flags);
1083extern void __vmw_execbuf_release_pinned_bo(struct vmw_private *dev_priv,
1084 struct vmw_fence_obj *fence);
1085extern void vmw_execbuf_release_pinned_bo(struct vmw_private *dev_priv);
1086
1087extern int vmw_execbuf_fence_commands(struct drm_file *file_priv,
1088 struct vmw_private *dev_priv,
1089 struct vmw_fence_obj **p_fence,
1090 uint32_t *p_handle);
1091extern void vmw_execbuf_copy_fence_user(struct vmw_private *dev_priv,
1092 struct vmw_fpriv *vmw_fp,
1093 int ret,
1094 struct drm_vmw_fence_rep __user
1095 *user_fence_rep,
1096 struct vmw_fence_obj *fence,
1097 uint32_t fence_handle,
1098 int32_t out_fence_fd,
1099 struct sync_file *sync_file);
1100bool vmw_cmd_describe(const void *buf, u32 *size, char const **cmd);
1101
1102
1103
1104
1105
1106extern int vmw_wait_seqno(struct vmw_private *dev_priv, bool lazy,
1107 uint32_t seqno, bool interruptible,
1108 unsigned long timeout);
1109extern int vmw_irq_install(struct drm_device *dev, int irq);
1110extern void vmw_irq_uninstall(struct drm_device *dev);
1111extern bool vmw_seqno_passed(struct vmw_private *dev_priv,
1112 uint32_t seqno);
1113extern int vmw_fallback_wait(struct vmw_private *dev_priv,
1114 bool lazy,
1115 bool fifo_idle,
1116 uint32_t seqno,
1117 bool interruptible,
1118 unsigned long timeout);
1119extern void vmw_update_seqno(struct vmw_private *dev_priv,
1120 struct vmw_fifo_state *fifo_state);
1121extern void vmw_seqno_waiter_add(struct vmw_private *dev_priv);
1122extern void vmw_seqno_waiter_remove(struct vmw_private *dev_priv);
1123extern void vmw_goal_waiter_add(struct vmw_private *dev_priv);
1124extern void vmw_goal_waiter_remove(struct vmw_private *dev_priv);
1125extern void vmw_generic_waiter_add(struct vmw_private *dev_priv, u32 flag,
1126 int *waiter_count);
1127extern void vmw_generic_waiter_remove(struct vmw_private *dev_priv,
1128 u32 flag, int *waiter_count);
1129
1130
1131
1132
1133
1134
1135extern void vmw_marker_queue_init(struct vmw_marker_queue *queue);
1136extern void vmw_marker_queue_takedown(struct vmw_marker_queue *queue);
1137extern int vmw_marker_push(struct vmw_marker_queue *queue,
1138 uint32_t seqno);
1139extern int vmw_marker_pull(struct vmw_marker_queue *queue,
1140 uint32_t signaled_seqno);
1141extern int vmw_wait_lag(struct vmw_private *dev_priv,
1142 struct vmw_marker_queue *queue, uint32_t us);
1143
1144
1145
1146
1147
1148int vmw_fb_init(struct vmw_private *vmw_priv);
1149int vmw_fb_close(struct vmw_private *dev_priv);
1150int vmw_fb_off(struct vmw_private *vmw_priv);
1151int vmw_fb_on(struct vmw_private *vmw_priv);
1152
1153
1154
1155
1156
1157int vmw_kms_init(struct vmw_private *dev_priv);
1158int vmw_kms_close(struct vmw_private *dev_priv);
1159int vmw_kms_cursor_bypass_ioctl(struct drm_device *dev, void *data,
1160 struct drm_file *file_priv);
1161void vmw_kms_cursor_post_execbuf(struct vmw_private *dev_priv);
1162void vmw_kms_cursor_snoop(struct vmw_surface *srf,
1163 struct ttm_object_file *tfile,
1164 struct ttm_buffer_object *bo,
1165 SVGA3dCmdHeader *header);
1166int vmw_kms_write_svga(struct vmw_private *vmw_priv,
1167 unsigned width, unsigned height, unsigned pitch,
1168 unsigned bpp, unsigned depth);
1169bool vmw_kms_validate_mode_vram(struct vmw_private *dev_priv,
1170 uint32_t pitch,
1171 uint32_t height);
1172u32 vmw_get_vblank_counter(struct drm_crtc *crtc);
1173int vmw_enable_vblank(struct drm_crtc *crtc);
1174void vmw_disable_vblank(struct drm_crtc *crtc);
1175int vmw_kms_present(struct vmw_private *dev_priv,
1176 struct drm_file *file_priv,
1177 struct vmw_framebuffer *vfb,
1178 struct vmw_surface *surface,
1179 uint32_t sid, int32_t destX, int32_t destY,
1180 struct drm_vmw_rect *clips,
1181 uint32_t num_clips);
1182int vmw_kms_update_layout_ioctl(struct drm_device *dev, void *data,
1183 struct drm_file *file_priv);
1184void vmw_kms_legacy_hotspot_clear(struct vmw_private *dev_priv);
1185int vmw_kms_suspend(struct drm_device *dev);
1186int vmw_kms_resume(struct drm_device *dev);
1187void vmw_kms_lost_device(struct drm_device *dev);
1188
1189int vmw_dumb_create(struct drm_file *file_priv,
1190 struct drm_device *dev,
1191 struct drm_mode_create_dumb *args);
1192
1193int vmw_dumb_map_offset(struct drm_file *file_priv,
1194 struct drm_device *dev, uint32_t handle,
1195 uint64_t *offset);
1196int vmw_dumb_destroy(struct drm_file *file_priv,
1197 struct drm_device *dev,
1198 uint32_t handle);
1199extern int vmw_resource_pin(struct vmw_resource *res, bool interruptible);
1200extern void vmw_resource_unpin(struct vmw_resource *res);
1201extern enum vmw_res_type vmw_res_type(const struct vmw_resource *res);
1202
1203
1204
1205
1206
1207int vmw_overlay_init(struct vmw_private *dev_priv);
1208int vmw_overlay_close(struct vmw_private *dev_priv);
1209int vmw_overlay_ioctl(struct drm_device *dev, void *data,
1210 struct drm_file *file_priv);
1211int vmw_overlay_resume_all(struct vmw_private *dev_priv);
1212int vmw_overlay_pause_all(struct vmw_private *dev_priv);
1213int vmw_overlay_claim(struct vmw_private *dev_priv, uint32_t *out);
1214int vmw_overlay_unref(struct vmw_private *dev_priv, uint32_t stream_id);
1215int vmw_overlay_num_overlays(struct vmw_private *dev_priv);
1216int vmw_overlay_num_free_overlays(struct vmw_private *dev_priv);
1217
1218
1219
1220
1221
1222int vmw_gmrid_man_init(struct vmw_private *dev_priv, int type);
1223void vmw_gmrid_man_fini(struct vmw_private *dev_priv, int type);
1224
1225
1226
1227
1228
1229extern const struct dma_buf_ops vmw_prime_dmabuf_ops;
1230extern int vmw_prime_fd_to_handle(struct drm_device *dev,
1231 struct drm_file *file_priv,
1232 int fd, u32 *handle);
1233extern int vmw_prime_handle_to_fd(struct drm_device *dev,
1234 struct drm_file *file_priv,
1235 uint32_t handle, uint32_t flags,
1236 int *prime_fd);
1237
1238
1239
1240
1241struct vmw_mob;
1242extern int vmw_mob_bind(struct vmw_private *dev_priv, struct vmw_mob *mob,
1243 const struct vmw_sg_table *vsgt,
1244 unsigned long num_data_pages, int32_t mob_id);
1245extern void vmw_mob_unbind(struct vmw_private *dev_priv,
1246 struct vmw_mob *mob);
1247extern void vmw_mob_destroy(struct vmw_mob *mob);
1248extern struct vmw_mob *vmw_mob_create(unsigned long data_pages);
1249extern int vmw_otables_setup(struct vmw_private *dev_priv);
1250extern void vmw_otables_takedown(struct vmw_private *dev_priv);
1251
1252
1253
1254
1255
1256extern const struct vmw_user_resource_conv *user_context_converter;
1257
1258extern int vmw_context_define_ioctl(struct drm_device *dev, void *data,
1259 struct drm_file *file_priv);
1260extern int vmw_extended_context_define_ioctl(struct drm_device *dev, void *data,
1261 struct drm_file *file_priv);
1262extern int vmw_context_destroy_ioctl(struct drm_device *dev, void *data,
1263 struct drm_file *file_priv);
1264extern struct list_head *vmw_context_binding_list(struct vmw_resource *ctx);
1265extern struct vmw_cmdbuf_res_manager *
1266vmw_context_res_man(struct vmw_resource *ctx);
1267extern struct vmw_resource *vmw_context_cotable(struct vmw_resource *ctx,
1268 SVGACOTableType cotable_type);
1269extern struct list_head *vmw_context_binding_list(struct vmw_resource *ctx);
1270struct vmw_ctx_binding_state;
1271extern struct vmw_ctx_binding_state *
1272vmw_context_binding_state(struct vmw_resource *ctx);
1273extern void vmw_dx_context_scrub_cotables(struct vmw_resource *ctx,
1274 bool readback);
1275extern int vmw_context_bind_dx_query(struct vmw_resource *ctx_res,
1276 struct vmw_buffer_object *mob);
1277extern struct vmw_buffer_object *
1278vmw_context_get_dx_query_mob(struct vmw_resource *ctx_res);
1279
1280
1281
1282
1283
1284
1285extern const struct vmw_user_resource_conv *user_surface_converter;
1286
1287extern int vmw_surface_destroy_ioctl(struct drm_device *dev, void *data,
1288 struct drm_file *file_priv);
1289extern int vmw_surface_define_ioctl(struct drm_device *dev, void *data,
1290 struct drm_file *file_priv);
1291extern int vmw_surface_reference_ioctl(struct drm_device *dev, void *data,
1292 struct drm_file *file_priv);
1293extern int vmw_gb_surface_define_ioctl(struct drm_device *dev, void *data,
1294 struct drm_file *file_priv);
1295extern int vmw_gb_surface_reference_ioctl(struct drm_device *dev, void *data,
1296 struct drm_file *file_priv);
1297int vmw_surface_gb_priv_define(struct drm_device *dev,
1298 uint32_t user_accounting_size,
1299 SVGA3dSurfaceAllFlags svga3d_flags,
1300 SVGA3dSurfaceFormat format,
1301 bool for_scanout,
1302 uint32_t num_mip_levels,
1303 uint32_t multisample_count,
1304 uint32_t array_size,
1305 struct drm_vmw_size size,
1306 SVGA3dMSPattern multisample_pattern,
1307 SVGA3dMSQualityLevel quality_level,
1308 struct vmw_surface **srf_out);
1309extern int vmw_gb_surface_define_ext_ioctl(struct drm_device *dev,
1310 void *data,
1311 struct drm_file *file_priv);
1312extern int vmw_gb_surface_reference_ext_ioctl(struct drm_device *dev,
1313 void *data,
1314 struct drm_file *file_priv);
1315
1316int vmw_gb_surface_define(struct vmw_private *dev_priv,
1317 uint32_t user_accounting_size,
1318 const struct vmw_surface_metadata *req,
1319 struct vmw_surface **srf_out);
1320
1321
1322
1323
1324
1325extern const struct vmw_user_resource_conv *user_shader_converter;
1326
1327extern int vmw_shader_define_ioctl(struct drm_device *dev, void *data,
1328 struct drm_file *file_priv);
1329extern int vmw_shader_destroy_ioctl(struct drm_device *dev, void *data,
1330 struct drm_file *file_priv);
1331extern int vmw_compat_shader_add(struct vmw_private *dev_priv,
1332 struct vmw_cmdbuf_res_manager *man,
1333 u32 user_key, const void *bytecode,
1334 SVGA3dShaderType shader_type,
1335 size_t size,
1336 struct list_head *list);
1337extern int vmw_shader_remove(struct vmw_cmdbuf_res_manager *man,
1338 u32 user_key, SVGA3dShaderType shader_type,
1339 struct list_head *list);
1340extern int vmw_dx_shader_add(struct vmw_cmdbuf_res_manager *man,
1341 struct vmw_resource *ctx,
1342 u32 user_key,
1343 SVGA3dShaderType shader_type,
1344 struct list_head *list);
1345extern void vmw_dx_shader_cotable_list_scrub(struct vmw_private *dev_priv,
1346 struct list_head *list,
1347 bool readback);
1348
1349extern struct vmw_resource *
1350vmw_shader_lookup(struct vmw_cmdbuf_res_manager *man,
1351 u32 user_key, SVGA3dShaderType shader_type);
1352
1353
1354
1355
1356struct vmw_resource *
1357vmw_dx_streamoutput_lookup(struct vmw_cmdbuf_res_manager *man,
1358 u32 user_key);
1359int vmw_dx_streamoutput_add(struct vmw_cmdbuf_res_manager *man,
1360 struct vmw_resource *ctx,
1361 SVGA3dStreamOutputId user_key,
1362 struct list_head *list);
1363void vmw_dx_streamoutput_set_size(struct vmw_resource *res, u32 size);
1364int vmw_dx_streamoutput_remove(struct vmw_cmdbuf_res_manager *man,
1365 SVGA3dStreamOutputId user_key,
1366 struct list_head *list);
1367void vmw_dx_streamoutput_cotable_list_scrub(struct vmw_private *dev_priv,
1368 struct list_head *list,
1369 bool readback);
1370
1371
1372
1373
1374
1375extern struct vmw_cmdbuf_res_manager *
1376vmw_cmdbuf_res_man_create(struct vmw_private *dev_priv);
1377extern void vmw_cmdbuf_res_man_destroy(struct vmw_cmdbuf_res_manager *man);
1378extern size_t vmw_cmdbuf_res_man_size(void);
1379extern struct vmw_resource *
1380vmw_cmdbuf_res_lookup(struct vmw_cmdbuf_res_manager *man,
1381 enum vmw_cmdbuf_res_type res_type,
1382 u32 user_key);
1383extern void vmw_cmdbuf_res_revert(struct list_head *list);
1384extern void vmw_cmdbuf_res_commit(struct list_head *list);
1385extern int vmw_cmdbuf_res_add(struct vmw_cmdbuf_res_manager *man,
1386 enum vmw_cmdbuf_res_type res_type,
1387 u32 user_key,
1388 struct vmw_resource *res,
1389 struct list_head *list);
1390extern int vmw_cmdbuf_res_remove(struct vmw_cmdbuf_res_manager *man,
1391 enum vmw_cmdbuf_res_type res_type,
1392 u32 user_key,
1393 struct list_head *list,
1394 struct vmw_resource **res);
1395
1396
1397
1398
1399extern const SVGACOTableType vmw_cotable_scrub_order[];
1400extern struct vmw_resource *vmw_cotable_alloc(struct vmw_private *dev_priv,
1401 struct vmw_resource *ctx,
1402 u32 type);
1403extern int vmw_cotable_notify(struct vmw_resource *res, int id);
1404extern int vmw_cotable_scrub(struct vmw_resource *res, bool readback);
1405extern void vmw_cotable_add_resource(struct vmw_resource *ctx,
1406 struct list_head *head);
1407
1408
1409
1410
1411struct vmw_cmdbuf_man;
1412struct vmw_cmdbuf_header;
1413
1414extern struct vmw_cmdbuf_man *
1415vmw_cmdbuf_man_create(struct vmw_private *dev_priv);
1416extern int vmw_cmdbuf_set_pool_size(struct vmw_cmdbuf_man *man,
1417 size_t size, size_t default_size);
1418extern void vmw_cmdbuf_remove_pool(struct vmw_cmdbuf_man *man);
1419extern void vmw_cmdbuf_man_destroy(struct vmw_cmdbuf_man *man);
1420extern int vmw_cmdbuf_idle(struct vmw_cmdbuf_man *man, bool interruptible,
1421 unsigned long timeout);
1422extern void *vmw_cmdbuf_reserve(struct vmw_cmdbuf_man *man, size_t size,
1423 int ctx_id, bool interruptible,
1424 struct vmw_cmdbuf_header *header);
1425extern void vmw_cmdbuf_commit(struct vmw_cmdbuf_man *man, size_t size,
1426 struct vmw_cmdbuf_header *header,
1427 bool flush);
1428extern void *vmw_cmdbuf_alloc(struct vmw_cmdbuf_man *man,
1429 size_t size, bool interruptible,
1430 struct vmw_cmdbuf_header **p_header);
1431extern void vmw_cmdbuf_header_free(struct vmw_cmdbuf_header *header);
1432extern int vmw_cmdbuf_cur_flush(struct vmw_cmdbuf_man *man,
1433 bool interruptible);
1434extern void vmw_cmdbuf_irqthread(struct vmw_cmdbuf_man *man);
1435
1436
1437
1438
1439
1440
1441
1442
1443
1444
1445
1446
1447struct vmw_diff_cpy {
1448 struct drm_rect rect;
1449 size_t line;
1450 size_t line_offset;
1451 int cpp;
1452 void (*do_cpy)(struct vmw_diff_cpy *diff, u8 *dest, const u8 *src,
1453 size_t n);
1454};
1455
1456#define VMW_CPU_BLIT_INITIALIZER { \
1457 .do_cpy = vmw_memcpy, \
1458}
1459
1460#define VMW_CPU_BLIT_DIFF_INITIALIZER(_cpp) { \
1461 .line = 0, \
1462 .line_offset = 0, \
1463 .rect = { .x1 = INT_MAX/2, \
1464 .y1 = INT_MAX/2, \
1465 .x2 = INT_MIN/2, \
1466 .y2 = INT_MIN/2 \
1467 }, \
1468 .cpp = _cpp, \
1469 .do_cpy = vmw_diff_memcpy, \
1470}
1471
1472void vmw_diff_memcpy(struct vmw_diff_cpy *diff, u8 *dest, const u8 *src,
1473 size_t n);
1474
1475void vmw_memcpy(struct vmw_diff_cpy *diff, u8 *dest, const u8 *src, size_t n);
1476
1477int vmw_bo_cpu_blit(struct ttm_buffer_object *dst,
1478 u32 dst_offset, u32 dst_stride,
1479 struct ttm_buffer_object *src,
1480 u32 src_offset, u32 src_stride,
1481 u32 w, u32 h,
1482 struct vmw_diff_cpy *diff);
1483
1484
1485int vmw_host_get_guestinfo(const char *guest_info_param,
1486 char *buffer, size_t *length);
1487int vmw_host_log(const char *log);
1488int vmw_msg_ioctl(struct drm_device *dev, void *data,
1489 struct drm_file *file_priv);
1490
1491
1492
1493
1494
1495
1496
1497
1498
1499
1500
1501
1502#define VMW_DEBUG_USER(fmt, ...) \
1503 DRM_DEBUG_DRIVER(fmt, ##__VA_ARGS__)
1504
1505
1506void vmw_bo_dirty_scan(struct vmw_buffer_object *vbo);
1507int vmw_bo_dirty_add(struct vmw_buffer_object *vbo);
1508void vmw_bo_dirty_transfer_to_res(struct vmw_resource *res);
1509void vmw_bo_dirty_clear_res(struct vmw_resource *res);
1510void vmw_bo_dirty_release(struct vmw_buffer_object *vbo);
1511void vmw_bo_dirty_unmap(struct vmw_buffer_object *vbo,
1512 pgoff_t start, pgoff_t end);
1513vm_fault_t vmw_bo_vm_fault(struct vm_fault *vmf);
1514vm_fault_t vmw_bo_vm_mkwrite(struct vm_fault *vmf);
1515#ifdef CONFIG_TRANSPARENT_HUGEPAGE
1516vm_fault_t vmw_bo_vm_huge_fault(struct vm_fault *vmf,
1517 enum page_entry_size pe_size);
1518#endif
1519
1520
1521#ifdef CONFIG_TRANSPARENT_HUGEPAGE
1522extern int vmw_thp_init(struct vmw_private *dev_priv);
1523void vmw_thp_fini(struct vmw_private *dev_priv);
1524#endif
1525
1526
1527
1528
1529
1530
1531#define VMW_DEBUG_KMS(fmt, ...) \
1532 DRM_DEBUG_DRIVER(fmt, ##__VA_ARGS__)
1533
1534
1535
1536
1537
1538static inline void vmw_surface_unreference(struct vmw_surface **srf)
1539{
1540 struct vmw_surface *tmp_srf = *srf;
1541 struct vmw_resource *res = &tmp_srf->res;
1542 *srf = NULL;
1543
1544 vmw_resource_unreference(&res);
1545}
1546
1547static inline struct vmw_surface *vmw_surface_reference(struct vmw_surface *srf)
1548{
1549 (void) vmw_resource_reference(&srf->res);
1550 return srf;
1551}
1552
1553static inline void vmw_bo_unreference(struct vmw_buffer_object **buf)
1554{
1555 struct vmw_buffer_object *tmp_buf = *buf;
1556
1557 *buf = NULL;
1558 if (tmp_buf != NULL) {
1559 ttm_bo_put(&tmp_buf->base);
1560 }
1561}
1562
1563static inline struct vmw_buffer_object *
1564vmw_bo_reference(struct vmw_buffer_object *buf)
1565{
1566 ttm_bo_get(&buf->base);
1567 return buf;
1568}
1569
1570static inline struct ttm_mem_global *vmw_mem_glob(struct vmw_private *dev_priv)
1571{
1572 return &ttm_mem_glob;
1573}
1574
1575static inline void vmw_fifo_resource_inc(struct vmw_private *dev_priv)
1576{
1577 atomic_inc(&dev_priv->num_fifo_resources);
1578}
1579
1580static inline void vmw_fifo_resource_dec(struct vmw_private *dev_priv)
1581{
1582 atomic_dec(&dev_priv->num_fifo_resources);
1583}
1584
1585
1586
1587
1588
1589
1590
1591
1592
1593static inline u32 vmw_mmio_read(u32 *addr)
1594{
1595 return READ_ONCE(*addr);
1596}
1597
1598
1599
1600
1601
1602
1603
1604
1605
1606static inline void vmw_mmio_write(u32 value, u32 *addr)
1607{
1608 WRITE_ONCE(*addr, value);
1609}
1610#endif
1611