/src/sys/external/bsd/drm2/dist/drm/i915/ |
i915_mm.c | 42 struct sgt_iter sgt; member in struct:remap_pfn 62 return (r->sgt.dma + r->sgt.curr + r->iobase) >> PAGE_SHIFT; 64 return r->sgt.pfn + (r->sgt.curr >> PAGE_SHIFT); 71 if (GEM_WARN_ON(!r->sgt.pfn)) 79 r->sgt.curr += PAGE_SIZE; 80 if (r->sgt.curr >= r->sgt.max) 81 r->sgt = __sgt_iter(__sg_next(r->sgt.sgp), use_dma(r->iobase)) [all...] |
/src/sys/external/bsd/drm2/include/drm/ |
drm_gem_cma_helper.h | 56 struct sg_table *sgt; member in struct:drm_gem_cma_object
|
/src/sys/external/bsd/drm2/dist/drm/i915/gem/ |
i915_gem_internal.c | 47 struct sg_table *sgt = NULL; local in function:i915_gem_object_get_pages_internal 77 sgt = kmalloc(sizeof(*sgt), GFP_KERNEL); 78 if (sgt == NULL) { 82 if (sg_alloc_table_from_bus_dmamem(sgt, dmat, obj->mm.u.internal.segs, 89 ret = i915_gem_gtt_prepare_pages(obj, sgt); 95 __i915_gem_object_set_pages(obj, sgt, i915_sg_page_sizes(sgt->sgl)); 101 i915_gem_gtt_finish_pages(obj, sgt); 103 sg_free_table(sgt); [all...] |
/src/sys/external/bsd/drm2/dist/drm/amd/amdgpu/ |
amdgpu_dma_buf.c | 273 struct sg_table *sgt; local in function:amdgpu_dma_buf_map 280 sgt = drm_prime_pages_to_sg(bo->tbo.ttm->pages, bo->tbo.num_pages); 281 if (IS_ERR(sgt)) 282 return sgt; 284 if (!dma_map_sg_attrs(attach->dev, sgt->sgl, sgt->nents, dir, 288 return sgt; 291 sg_free_table(sgt); 292 kfree(sgt); 299 * @sgt: sg_table to unma [all...] |
amdgpu_ttm.c | 1340 struct sg_table *sgt; local in function:amdgpu_ttm_tt_populate 1343 sgt = dma_buf_map_attachment(attach, DMA_BIDIRECTIONAL); 1344 if (IS_ERR(sgt)) 1345 return PTR_ERR(sgt); 1347 ttm->sg = sgt;
|
/src/sys/external/bsd/drm2/dist/drm/i915/selftests/ |
scatterlist.c | 121 struct sgt_iter sgt; local in function:expect_pfn_sgtiter 126 for_each_sgt_page(page, sgt, &pt->st) {
|
i915_vma.c | 631 struct sgt_iter sgt; local in function:assert_partial 634 for_each_sgt_daddr(dma, sgt, vma->pages) {
|
/src/sys/dev/ic/ |
ld_aac.c | 147 struct aac_sg_tableraw *sgt; local in function:ld_aac_dobio 157 sgt = &raw->SgMapRaw; 161 sgt->SgCount = xfer->dm_nsegs; 162 sge = sgt->SgEntryRaw; 178 struct aac_sg_table *sgt; local in function:ld_aac_dobio 191 sgt = &bw->SgMap; 200 sgt = &br->SgMap; 204 sgt->SgCount = xfer->dm_nsegs; 205 sge = sgt->SgEntry; 223 struct aac_sg_table64 *sgt; local in function:ld_aac_dobio [all...] |
/src/sys/external/bsd/drm2/dist/drm/ |
drm_prime.c | 756 struct sg_table *sgt; local in function:drm_gem_map_dma_buf 762 sgt = obj->funcs->get_sg_table(obj); 764 sgt = obj->dev->driver->gem_prime_get_sg_table(obj); 766 if (!dma_map_sg_attrs(attach->dev, sgt->sgl, sgt->nents, dir, 768 sg_free_table(sgt); 769 kfree(sgt); 770 sgt = ERR_PTR(-ENOMEM); 773 return sgt; 780 * @sgt: scatterlist info of the buffer to unma 1059 struct sg_table *sgt; local in function:drm_gem_prime_import_dev [all...] |
/src/sys/external/bsd/drm2/dist/drm/vmwgfx/ |
vmwgfx_ttm_buffer.c | 250 struct sg_table sgt; member in struct:vmw_ttm_tt 350 __sg_page_iter_start(&viter->iter.base, vsgt->sgt->sgl, 351 vsgt->sgt->orig_nents, p_offset); 370 dma_unmap_sg(dev, vmw_tt->sgt.sgl, vmw_tt->sgt.nents, 372 vmw_tt->sgt.nents = vmw_tt->sgt.orig_nents; 393 ret = dma_map_sg(dev, vmw_tt->sgt.sgl, vmw_tt->sgt.orig_nents, 398 vmw_tt->sgt.nents = ret [all...] |
vmwgfx_drv.h | 310 * @sgt: Pointer to a struct sg_table with binding information 317 struct sg_table *sgt; member in struct:vmw_sg_table
|
/src/sys/external/bsd/drm2/dist/drm/virtio/ |
virtgpu_vq.c | 279 struct sg_table *sgt; local in function:vmalloc_to_sgt 286 sgt = kmalloc(sizeof(*sgt), GFP_KERNEL); 287 if (!sgt) 291 ret = sg_alloc_table(sgt, *sg_ents, GFP_KERNEL); 293 kfree(sgt); 297 for_each_sg(sgt->sgl, sg, *sg_ents, i) { 300 sg_free_table(sgt); 301 kfree(sgt); 312 return sgt; 368 struct sg_table *sgt = NULL; local in function:virtio_gpu_queue_fenced_ctrl_buffer [all...] |