HomeSort by: relevance | last modified time | path
    Searched defs:bo (Results 1 - 25 of 89) sorted by relevancy

1 2 3 4

  /src/sys/external/bsd/drm2/dist/include/drm/ttm/
ttm_execbuf_util.h 44 * @bo: refcounted buffer object pointer.
50 struct ttm_buffer_object *bo; member in struct:ttm_validate_buffer
  /src/sys/external/bsd/drm2/dist/drm/amd/amdgpu/
amdgpu_mn.c 72 struct amdgpu_bo *bo = container_of(mni, struct amdgpu_bo, notifier); local in function:amdgpu_mn_invalidate_gfx
73 struct amdgpu_device *adev = amdgpu_ttm_adev(bo->tbo.bdev);
83 r = dma_resv_wait_timeout_rcu(bo->tbo.base.resv, true, false,
87 DRM_ERROR("(%ld) failed to wait for user bo\n", r);
102 * We temporarily evict the BO attached to this range. This necessitates
109 struct amdgpu_bo *bo = container_of(mni, struct amdgpu_bo, notifier); local in function:amdgpu_mn_invalidate_hsa
110 struct amdgpu_device *adev = amdgpu_ttm_adev(bo->tbo.bdev);
119 amdgpu_amdkfd_evict_userptr(bo->kfd_bo, bo->notifier.mm);
130 * amdgpu_mn_register - register a BO for notifier update
    [all...]
amdgpu_gart.h 49 struct amdgpu_bo *bo; member in struct:amdgpu_gart
amdgpu_bo_list.c 61 struct amdgpu_bo *bo = ttm_to_amdgpu_bo(e->tv.bo); local in function:amdgpu_bo_list_free
63 amdgpu_bo_unref(&bo);
102 struct amdgpu_bo *bo; local in function:amdgpu_bo_list_create
115 bo = amdgpu_bo_ref(gem_to_amdgpu_bo(gobj));
118 usermm = amdgpu_ttm_tt_get_usermm(bo->tbo.ttm);
121 amdgpu_bo_unref(&bo);
126 amdgpu_bo_unref(&bo);
138 entry->tv.bo = &bo->tbo
161 struct amdgpu_bo *bo = ttm_to_amdgpu_bo(array[i].tv.bo); local in function:amdgpu_bo_list_create
166 struct amdgpu_bo *bo = ttm_to_amdgpu_bo(array[i].tv.bo); local in function:amdgpu_bo_list_create
221 struct amdgpu_bo *bo = ttm_to_amdgpu_bo(e->tv.bo); local in function:amdgpu_bo_list_get_list
    [all...]
amdgpu_dma_buf.c 49 * @obj: GEM BO
51 * Sets up an in-kernel virtual mapping of the BO's memory.
58 struct amdgpu_bo *bo = gem_to_amdgpu_bo(obj); local in function:amdgpu_gem_prime_vmap
61 ret = ttm_bo_kmap(&bo->tbo, 0, bo->tbo.num_pages,
62 &bo->dma_buf_vmap);
66 return bo->dma_buf_vmap.virtual;
71 * @obj: GEM BO
74 * Tears down the in-kernel virtual mapping of the BO's memory.
78 struct amdgpu_bo *bo = gem_to_amdgpu_bo(obj) local in function:amdgpu_gem_prime_vunmap
198 struct amdgpu_bo *bo = gem_to_amdgpu_bo(obj); local in function:amdgpu_dma_buf_attach
242 struct amdgpu_bo *bo = gem_to_amdgpu_bo(obj); local in function:amdgpu_dma_buf_detach
272 struct amdgpu_bo *bo = gem_to_amdgpu_bo(obj); local in function:amdgpu_dma_buf_map
310 struct amdgpu_bo *bo = gem_to_amdgpu_bo(obj); local in function:amdgpu_dma_buf_unmap
334 struct amdgpu_bo *bo = gem_to_amdgpu_bo(gem); local in function:amdgpu_dma_buf_begin_cpu_access
385 struct amdgpu_bo *bo = gem_to_amdgpu_bo(gobj); local in function:amdgpu_gem_prime_export
416 struct amdgpu_bo *bo; local in function:amdgpu_dma_buf_create_obj
    [all...]
amdgpu_ttm.h 80 struct ttm_buffer_object *bo; member in struct:amdgpu_copy_mem
92 u64 amdgpu_vram_mgr_bo_visible_size(struct amdgpu_bo *bo);
113 int amdgpu_fill_buffer(struct amdgpu_bo *bo,
124 int amdgpu_ttm_alloc_gart(struct ttm_buffer_object *bo);
128 int amdgpu_ttm_tt_get_user_pages(struct amdgpu_bo *bo, struct page **pages);
131 static inline int amdgpu_ttm_tt_get_user_pages(struct amdgpu_bo *bo,
  /src/sys/external/bsd/drm2/dist/drm/nouveau/
nv10_fence.h 17 struct nouveau_bo *bo; member in struct:nv10_fence_priv
nouveau_bo.h 19 struct ttm_buffer_object bo; member in struct:nouveau_bo
52 nouveau_bo(struct ttm_buffer_object *bo)
54 return container_of(bo, struct nouveau_bo, bo);
67 ttm_bo_get(&ref->bo);
68 *pnvbo = nouveau_bo(&ref->bo);
73 ttm_bo_put(&prev->bo);
nouveau_fence.h 95 struct nouveau_bo *bo; member in struct:nv84_fence_priv
  /src/sys/external/bsd/drm2/dist/drm/qxl/
qxl_prime.c 38 struct qxl_bo *bo = gem_to_qxl_bo(obj); local in function:qxl_gem_prime_pin
40 return qxl_bo_pin(bo);
45 struct qxl_bo *bo = gem_to_qxl_bo(obj); local in function:qxl_gem_prime_unpin
47 qxl_bo_unpin(bo);
64 struct qxl_bo *bo = gem_to_qxl_bo(obj); local in function:qxl_gem_prime_vmap
68 ret = qxl_bo_kmap(bo, &ptr);
77 struct qxl_bo *bo = gem_to_qxl_bo(obj); local in function:qxl_gem_prime_vunmap
79 qxl_bo_kunmap(bo);
qxl_debugfs.c 62 struct qxl_bo *bo; local in function:qxl_debugfs_buffers_info
64 list_for_each_entry(bo, &qdev->gem.objects, list) {
69 fobj = rcu_dereference(bo->tbo.base.resv->fence);
74 (unsigned long)bo->tbo.base.size,
75 bo->pin_count, rel);
qxl_object.c 37 struct qxl_bo *bo; local in function:qxl_ttm_bo_destroy
40 bo = to_qxl_bo(tbo);
41 qdev = (struct qxl_device *)bo->tbo.base.dev->dev_private;
43 qxl_surface_evict(qdev, bo, false);
44 WARN_ON_ONCE(bo->map_count > 0);
46 list_del_init(&bo->list);
48 drm_gem_object_release(&bo->tbo.base);
49 kfree(bo);
52 bool qxl_ttm_bo_is_qxl_bo(struct ttm_buffer_object *bo)
54 if (bo->destroy == &qxl_ttm_bo_destroy
108 struct qxl_bo *bo; local in function:qxl_bo_create
321 struct qxl_bo *bo, *n; local in function:qxl_bo_force_delete
    [all...]
  /src/sys/external/bsd/drm2/dist/drm/radeon/
radeon_mn.c 57 struct radeon_bo *bo = container_of(mn, struct radeon_bo, notifier); local in function:radeon_mn_invalidate
61 if (!bo->tbo.ttm || bo->tbo.ttm->state != tt_bound)
67 r = radeon_bo_reserve(bo, true);
69 DRM_ERROR("(%ld) failed to reserve user bo\n", r);
73 r = dma_resv_wait_timeout_rcu(bo->tbo.base.resv, true, false,
76 DRM_ERROR("(%ld) failed to wait for user bo\n", r);
78 radeon_ttm_placement_from_domain(bo, RADEON_GEM_DOMAIN_CPU);
79 r = ttm_bo_validate(&bo->tbo, &bo->placement, &ctx)
    [all...]
radeon_prime.c 41 struct radeon_bo *bo = gem_to_radeon_bo(obj); local in function:radeon_gem_prime_get_sg_table
42 int npages = bo->tbo.num_pages;
44 return drm_prime_pages_to_sg(bo->tbo.ttm->pages, npages);
49 struct radeon_bo *bo = gem_to_radeon_bo(obj); local in function:radeon_gem_prime_vmap
52 ret = ttm_bo_kmap(&bo->tbo, 0, bo->tbo.num_pages,
53 &bo->dma_buf_vmap);
57 return bo->dma_buf_vmap.virtual;
62 struct radeon_bo *bo = gem_to_radeon_bo(obj); local in function:radeon_gem_prime_vunmap
64 ttm_bo_kunmap(&bo->dma_buf_vmap)
73 struct radeon_bo *bo; local in function:radeon_gem_prime_import_sg_table
93 struct radeon_bo *bo = gem_to_radeon_bo(obj); local in function:radeon_gem_prime_pin
111 struct radeon_bo *bo = gem_to_radeon_bo(obj); local in function:radeon_gem_prime_unpin
128 struct radeon_bo *bo = gem_to_radeon_bo(gobj); local in function:radeon_gem_prime_export
    [all...]
  /src/sys/external/bsd/drm2/dist/drm/ttm/
ttm_execbuf_util.c 45 struct ttm_buffer_object *bo = entry->bo; local in function:ttm_eu_backoff_reservation_reverse
47 dma_resv_unlock(bo->base.resv);
61 struct ttm_buffer_object *bo = entry->bo; local in function:ttm_eu_backoff_reservation
63 ttm_bo_move_to_lru_tail(bo, NULL);
64 dma_resv_unlock(bo->base.resv);
99 struct ttm_buffer_object *bo = entry->bo; local in function:ttm_eu_reserve_buffers
101 ret = __ttm_bo_reserve(bo, intr, (ticket == NULL), ticket)
172 struct ttm_buffer_object *bo = entry->bo; local in function:ttm_eu_fence_buffer_objects
    [all...]
ttm_bo_vm.c 50 static vm_fault_t ttm_bo_vm_fault_idle(struct ttm_buffer_object *bo,
56 if (likely(!bo->moving))
62 if (dma_fence_is_signaled(bo->moving))
74 ttm_bo_get(bo);
76 (void) dma_fence_wait(bo->moving, true);
77 dma_resv_unlock(bo->base.resv);
78 ttm_bo_put(bo);
85 err = dma_fence_wait(bo->moving, true);
93 dma_fence_put(bo->moving);
94 bo->moving = NULL
187 struct ttm_buffer_object *bo = vma->vm_private_data; local in function:ttm_bo_vm_fault_reserved
334 struct ttm_buffer_object *bo = vma->vm_private_data; local in function:ttm_bo_vm_fault
354 struct ttm_buffer_object *bo = vma->vm_private_data; local in function:ttm_bo_vm_open
364 struct ttm_buffer_object *bo = vma->vm_private_data; local in function:ttm_bo_vm_close
414 struct ttm_buffer_object *bo = vma->vm_private_data; local in function:ttm_bo_vm_access
461 struct ttm_buffer_object *bo = NULL; local in function:ttm_bo_vm_lookup
506 struct ttm_buffer_object *bo; local in function:ttm_bo_mmap
    [all...]
  /src/sys/external/bsd/drm2/dist/drm/virtio/
virtgpu_object.c 71 struct virtio_gpu_object *bo = gem_to_virtio_gpu_obj(obj); local in function:virtio_gpu_free_object
72 struct virtio_gpu_device *vgdev = bo->base.base.dev->dev_private;
74 if (bo->pages)
75 virtio_gpu_object_detach(vgdev, bo);
76 if (bo->created)
77 virtio_gpu_cmd_unref_resource(vgdev, bo->hw_res_handle);
78 virtio_gpu_resource_id_put(vgdev, bo->hw_res_handle);
100 struct virtio_gpu_object *bo; local in function:virtio_gpu_create_object
102 bo = kzalloc(sizeof(*bo), GFP_KERNEL)
117 struct virtio_gpu_object *bo; local in function:virtio_gpu_object_create
    [all...]
virtgpu_plane.c 115 struct virtio_gpu_object *bo = local in function:virtio_gpu_update_dumb_bo
128 virtio_gpu_array_add_obj(objs, &bo->base.base);
140 struct virtio_gpu_object *bo; local in function:virtio_gpu_primary_plane_update
164 bo = gem_to_virtio_gpu_obj(plane->state->fb->obj[0]);
165 if (bo->dumb)
174 bo->hw_res_handle,
182 bo->hw_res_handle,
189 virtio_gpu_cmd_resource_flush(vgdev, bo->hw_res_handle,
204 struct virtio_gpu_object *bo; local in function:virtio_gpu_cursor_prepare_fb
210 bo = gem_to_virtio_gpu_obj(vgfb->base.obj[0])
242 struct virtio_gpu_object *bo = NULL; local in function:virtio_gpu_cursor_plane_update
    [all...]
  /src/sys/external/bsd/drm2/dist/drm/nouveau/nvkm/engine/gr/
nouveau_nvkm_engine_gr_ctxgp100.c 60 u32 bo = ao + size; local in function:gp100_grctx_generate_attrib
86 mmio_wr32(info, o + 0xf4, bo);
88 bo += grctx->attrib_nr_max * gr->ppc_tpc_max;
nouveau_nvkm_engine_gr_ctxgp102.c 56 u32 bo = ao + size; local in function:gp102_grctx_generate_attrib
85 mmio_wr32(info, o + 0xf4, bo);
87 bo += gs;
nouveau_nvkm_engine_gr_ctxgf108.c 752 u32 bo = 0; local in function:gf108_grctx_generate_attrib
753 u32 ao = bo + grctx->attrib_nr_max * gr->tpc_total;
767 mmio_skip(info, o + 0x20, (t << 28) | (b << 16) | ++bo);
768 mmio_wr32(info, o + 0x20, (t << 28) | (b << 16) | --bo);
769 bo += grctx->attrib_nr_max;
nouveau_nvkm_engine_gr_ctxgf117.c 260 u32 bo = 0; local in function:gf117_grctx_generate_attrib
261 u32 ao = bo + grctx->attrib_nr_max * gr->tpc_total;
277 mmio_skip(info, o + 0xc0, (t << 28) | (b << 16) | ++bo);
278 mmio_wr32(info, o + 0xc0, (t << 28) | (b << 16) | --bo);
279 bo += grctx->attrib_nr_max * gr->ppc_tpc_nr[gpc][ppc];
nouveau_nvkm_engine_gr_ctxgv100.c 77 u32 bo = ao + size; local in function:gv100_grctx_generate_attrib
103 mmio_wr32(info, o + 0xf4, bo);
105 bo += gs;
  /src/usr.sbin/extattrctl/
extattrctl.c 187 const char *bo; local in function:showattr
215 bo = needswap ? "big-endian" : "little-endian";
217 bo = needswap ? "little-endian" : "big-endian";
221 argv[0], rw32(uef.uef_version), rw32(uef.uef_size), bo);
  /src/sys/fs/nfs/client/
nfs_clsubs.c 347 struct bufobj *bo; local in function:ncl_clearcommit
350 bo = &vp->v_bufobj;
353 BO_LOCK(bo);
354 TAILQ_FOREACH_SAFE(bp, &bo->bo_dirty.bv_hd, b_bobufs, nbp) {
360 BO_UNLOCK(bo);

Completed in 19 milliseconds

1 2 3 4