/src/sys/external/bsd/drm2/dist/drm/qxl/ |
qxl_gem.c | 40 struct ttm_buffer_object *tbo; local in function:qxl_gem_object_free 46 tbo = &qobj->tbo; 47 ttm_bo_put(tbo); 71 *obj = &qbo->tbo.base;
|
qxl_drv.h | 77 struct ttm_buffer_object tbo; member in struct:qxl_bo 81 /* Protected by tbo.reserved */ 99 #define gem_to_qxl_bo(gobj) container_of((gobj), struct qxl_bo, tbo.base) 100 #define to_qxl_bo(tobj) container_of((tobj), struct qxl_bo, tbo) 313 (bo->tbo.mem.mem_type == TTM_PL_VRAM) 316 WARN_ON_ONCE((bo->tbo.offset & slot->gpu_offset) != slot->gpu_offset); 318 /* TODO - need to hold one of the locks to read tbo.offset */ 319 return slot->high_bits | (bo->tbo.offset - slot->gpu_offset + offset);
|
/src/sys/external/bsd/drm2/dist/drm/amd/amdgpu/ |
amdgpu_gtt_mgr.c | 40 struct ttm_buffer_object *tbo; member in struct:amdgpu_gtt_node 180 * @tbo: TTM BO we need this range for 187 struct ttm_buffer_object *tbo, 231 * @tbo: TTM BO we need this range for 238 struct ttm_buffer_object *tbo, 247 if ((&tbo->mem == mem || tbo->mem.mem_type != TTM_PL_TT) && 263 node->tbo = tbo; 267 r = amdgpu_gtt_mgr_alloc(man, tbo, place, mem) [all...] |
amdgpu_object.c | 71 struct amdgpu_device *adev = amdgpu_ttm_adev(bo->tbo.bdev); 73 if (bo->tbo.mem.mem_type == TTM_PL_VRAM) { 77 } else if (bo->tbo.mem.mem_type == TTM_PL_TT) { 82 static void amdgpu_bo_destroy(struct ttm_buffer_object *tbo) 84 struct amdgpu_device *adev = amdgpu_ttm_adev(tbo->bdev); 85 struct amdgpu_bo *bo = ttm_to_amdgpu_bo(tbo); 92 if (bo->tbo.base.import_attach) 93 drm_prime_gem_destroy(&bo->tbo.base, bo->tbo.sg); 94 drm_gem_object_release(&bo->tbo.base) 860 struct ttm_buffer_object *tbo; local in function:amdgpu_bo_unref [all...] |
amdgpu_object.h | 86 /* Protected by tbo.reserved */ 91 struct ttm_buffer_object tbo; member in struct:amdgpu_bo 119 static inline struct amdgpu_bo *ttm_to_amdgpu_bo(struct ttm_buffer_object *tbo) 121 return container_of(tbo, struct amdgpu_bo, tbo); 162 struct amdgpu_device *adev = amdgpu_ttm_adev(bo->tbo.bdev); 165 r = __ttm_bo_reserve(&bo->tbo, !no_intr, false, NULL); 176 ttm_bo_unreserve(&bo->tbo); 181 return bo->tbo.num_pages << PAGE_SHIFT; 186 return (bo->tbo.num_pages << PAGE_SHIFT) / AMDGPU_GPU_PAGE_SIZE [all...] |
/src/sys/external/bsd/drm2/dist/drm/radeon/ |
radeon_object.c | 64 u64 size = (u64)bo->tbo.num_pages << PAGE_SHIFT; 82 static void radeon_ttm_bo_destroy(struct ttm_buffer_object *tbo) 86 bo = container_of(tbo, struct radeon_bo, tbo); 88 radeon_update_memory_usage(bo, bo->tbo.mem.mem_type, -1); 95 if (bo->tbo.base.import_attach) 96 drm_prime_gem_destroy(&bo->tbo.base, bo->tbo.sg); 97 drm_gem_object_release(&bo->tbo.base); 219 drm_gem_private_object_init(rdev->ddev, &bo->tbo.base, size) 326 struct ttm_buffer_object *tbo; local in function:radeon_bo_unref [all...] |
radeon.h | 506 /* Protected by tbo.reserved */ 510 struct ttm_buffer_object tbo; member in struct:radeon_bo 535 #define gem_to_radeon_bo(gobj) container_of((gobj), struct radeon_bo, tbo.base)
|