Lines Matching defs:qdev
38 struct qxl_device *qdev;
41 qdev = (struct qxl_device *)bo->tbo.base.dev->dev_private;
43 qxl_surface_evict(qdev, bo, false);
45 mutex_lock(&qdev->gem.mutex);
47 mutex_unlock(&qdev->gem.mutex);
103 int qxl_bo_create(struct qxl_device *qdev,
121 r = drm_gem_object_init(&qdev->ddev, &bo->tbo.base, size);
137 r = ttm_bo_init(&qdev->mman.bdev, &bo->tbo, size, type,
142 dev_err(qdev->ddev.dev,
172 void *qxl_bo_kmap_atomic_page(struct qxl_device *qdev,
180 map = qdev->vram_mapping;
182 map = qdev->surface_mapping;
214 void qxl_bo_kunmap_atomic_page(struct qxl_device *qdev,
319 void qxl_bo_force_delete(struct qxl_device *qdev)
323 if (list_empty(&qdev->gem.objects))
325 dev_err(qdev->ddev.dev, "Userspace still has active objects !\n");
326 list_for_each_entry_safe(bo, n, &qdev->gem.objects, list) {
327 dev_err(qdev->ddev.dev, "%p %p %lu %lu force free\n",
330 mutex_lock(&qdev->gem.mutex);
332 mutex_unlock(&qdev->gem.mutex);
338 int qxl_bo_init(struct qxl_device *qdev)
340 return qxl_ttm_init(qdev);
343 void qxl_bo_fini(struct qxl_device *qdev)
345 qxl_ttm_fini(qdev);
348 int qxl_bo_check_id(struct qxl_device *qdev, struct qxl_bo *bo)
354 ret = qxl_surface_id_alloc(qdev, bo);
358 ret = qxl_hw_surface_alloc(qdev, bo);
365 int qxl_surf_evict(struct qxl_device *qdev)
367 return ttm_bo_evict_mm(&qdev->mman.bdev, TTM_PL_PRIV);
370 int qxl_vram_evict(struct qxl_device *qdev)
372 return ttm_bo_evict_mm(&qdev->mman.bdev, TTM_PL_VRAM);