| /src/sys/external/bsd/drm2/dist/drm/i915/ |
| i915_scatterlist.c | 21 if (orig_st->nents == orig_st->orig_nents) 24 if (sg_alloc_table(&new_st, orig_st->nents, GFP_KERNEL | __GFP_NOWARN)) 28 for_each_sg(orig_st->sgl, sg, orig_st->nents, i) { 35 GEM_BUG_ON(new_sg); /* Should walk exactly nents and hit the end */
|
| i915_gem_gtt.c | 50 pages->sgl, pages->nents, 56 pages->sgl, pages->nents, 99 dma_unmap_sg(kdev, pages->sgl, pages->nents, PCI_DMA_BIDIRECTIONAL);
|
| /src/sys/external/bsd/drm2/dist/drm/virtio/ |
| virtgpu_gem.c | 153 struct virtio_gpu_object_array *virtio_gpu_array_alloc(u32 nents) 156 size_t size = sizeof(*objs) + sizeof(objs->objs[0]) * nents; 162 objs->nents = 0; 163 objs->total = nents; 173 virtio_gpu_array_from_handles(struct drm_file *drm_file, u32 *handles, u32 nents) 178 objs = virtio_gpu_array_alloc(nents); 182 for (i = 0; i < nents; i++) { 185 objs->nents = i; 190 objs->nents = i; 197 if (WARN_ON_ONCE(objs->nents == objs->total) [all...] |
| virtgpu_vq.c | 599 bo->pages->sgl, bo->pages->nents, 621 uint32_t nents, 632 cmd_p->nr_entries = cpu_to_le32(nents); 635 vbuf->data_size = sizeof(*ents) * nents; 1014 bo->pages->sgl, bo->pages->nents, 1088 int si, nents, ret; local in function:virtio_gpu_object_attach 1107 obj->pages->sgl, obj->pages->nents, 1109 nents = obj->mapped; 1111 nents = obj->pages->nents; [all...] |
| virtgpu_drv.h | 85 u32 nents, total; member in struct:virtio_gpu_object_array 244 struct virtio_gpu_object_array *virtio_gpu_array_alloc(u32 nents); 246 virtio_gpu_array_from_handles(struct drm_file *drm_file, u32 *handles, u32 nents);
|
| /src/sys/external/bsd/drm2/linux/ |
| linux_sgt.c | 50 sgt->sgl->sg_npgs = sgt->nents = npgs; 141 dma_map_sg(bus_dma_tag_t dmat, struct scatterlist *sg, int nents, int dir) 144 return dma_map_sg_attrs(dmat, sg, nents, dir, 0); 148 dma_map_sg_attrs(bus_dma_tag_t dmat, struct scatterlist *sg, int nents, 157 KASSERT(nents >= 1); 175 nents, PAGE_SIZE, 0, BUS_DMA_WAITOK, &sg->sg_dmamap); 188 KASSERT(sg->sg_dmamap->dm_nsegs <= nents); 206 dma_unmap_sg(bus_dma_tag_t dmat, struct scatterlist *sg, int nents, int dir) 209 dma_unmap_sg_attrs(dmat, sg, nents, dir, 0); 213 dma_unmap_sg_attrs(bus_dma_tag_t dmat, struct scatterlist *sg, int nents, [all...] |
| /src/sys/external/bsd/drm2/dist/drm/i915/selftests/ |
| scatterlist.c | 56 for_each_sg(pt->st.sgl, sg, pt->st.nents, n) { 58 unsigned int npages = npages_fn(n, pt->st.nents, rnd); 94 for_each_sg_page(pt->st.sgl, &sgiter, pt->st.nents, 0) { 264 pt->st.nents = n; 351 pt.st.nents != prime) { 352 pr_err("i915_sg_trim failed (nents %u, orig_nents %u), expected %lu\n", 353 pt.st.nents, pt.st.orig_nents, prime);
|
| intel_memory_region.c | 156 if (obj->mm.pages->nents != 1) { 169 if (obj->mm.pages->nents != 1) { 195 if (obj->mm.pages->nents != 1) {
|
| /src/sys/external/bsd/drm2/dist/drm/i915/gem/selftests/ |
| mock_dmabuf.c | 36 if (!dma_map_sg(attachment->dev, st->sgl, st->nents, dir)) { 54 dma_unmap_sg(attachment->dev, st->sgl, st->nents, dir);
|
| huge_pages.c | 84 st->nents = 0; 107 st->nents++; 209 st->nents = 0; 225 st->nents++; 260 st->nents = 1;
|
| /src/sys/external/bsd/drm2/include/linux/ |
| scatterlist.h | 66 unsigned nents; member in struct:sg_table
|
| /src/sys/external/bsd/drm2/dist/drm/i915/gem/ |
| i915_gem_dmabuf.c | 46 ret = sg_alloc_table(st, obj->mm.pages->nents, GFP_KERNEL); 55 obj->mm.pages->nents * sizeof(st->sgl->sg_pgs[0])); 60 for (i = 0; i < obj->mm.pages->nents; i++) { 67 if (!dma_map_sg(attachment->dev, st->sgl, st->nents, dir)) { 90 dma_unmap_sg(attachment->dev, sg->sgl, sg->nents, dir);
|
| i915_gem_region.c | 122 st->nents = 0; 137 if (st->nents) { 147 st->nents++;
|
| i915_gem_shmem.c | 101 st->nents = 0; 158 KASSERT(st->nents == i); 159 sg->sg_pgs[st->nents++] = page; 169 st->nents++;
|
| i915_gem_internal.c | 163 st->nents = 0; 184 st->nents++;
|
| /src/sys/external/bsd/drm2/dist/drm/ |
| drm_cache.c | 135 for_each_sg_page(st->sgl, &sg_iter, st->nents, 0)
|
| drm_prime.c | 766 if (!dma_map_sg_attrs(attach->dev, sgt->sgl, sgt->nents, dir, 792 dma_unmap_sg_attrs(attach->dev, sgt->sgl, sgt->nents, dir, 1254 for_each_sg(sgt->sgl, sg, sgt->nents, count) {
|
| /src/sys/arch/i386/stand/bootxx/ |
| boot1.c | 202 uint32_t nents; local in function:gpt_lookup 245 nents = le32toh(hdr->hdr_entries); 267 for (i = 0; i < nents; i += entries_per_sector) {
|
| /src/sys/external/bsd/drm2/dist/drm/vmwgfx/ |
| vmwgfx_ttm_buffer.c | 370 dma_unmap_sg(dev, vmw_tt->sgt.sgl, vmw_tt->sgt.nents, 372 vmw_tt->sgt.nents = vmw_tt->sgt.orig_nents; 398 vmw_tt->sgt.nents = ret; 457 if (vsgt->num_pages > vmw_tt->sgt.nents) { 460 vmw_tt->sgt.nents);
|
| /src/sys/external/bsd/drm2/dist/drm/amd/amdgpu/ |
| amdgpu_dma_buf.c | 284 if (!dma_map_sg_attrs(attach->dev, sgt->sgl, sgt->nents, dir, 312 dma_unmap_sg(attach->dev, sgt->sgl, sgt->nents, dir);
|
| amdgpu_ttm.c | 995 unsigned nents; 1011 nents = dma_map_sg(adev->dev, ttm->sg->sgl, ttm->sg->nents, direction); 1012 if (nents != ttm->sg->nents) 1045 dma_unmap_sg(adev->dev, ttm->sg->sgl, ttm->sg->nents, direction);
|
| /src/sys/external/bsd/drm2/dist/drm/radeon/ |
| radeon_ttm.c | 521 unsigned pinned = 0, nents; local in function:radeon_ttm_tt_pin_userptr 640 nents = dma_map_sg(rdev->dev, ttm->sg->sgl, ttm->sg->nents, direction); 641 if (nents != ttm->sg->nents) 681 dma_unmap_sg(rdev->dev, ttm->sg->sgl, ttm->sg->nents, direction); 683 for_each_sg_page(ttm->sg->sgl, &sg_iter, ttm->sg->nents, 0) {
|
| /src/sys/external/bsd/drm2/dist/drm/xen/ |
| xen_drm_front_gem.c | 222 DRM_DEBUG("Imported buffer of size %zu with nents %u\n", 223 size, sgt->nents);
|
| /src/sys/external/bsd/drm2/dist/drm/i915/gt/ |
| intel_ggtt.c | 1452 st->nents++; 1489 st->nents = 0; 1535 st->nents++; 1571 st->nents = 0; 1724 st->nents = 0; 1735 st->nents++;
|
| /src/usr.bin/infocmp/ |
| infocmp.c | 144 print_ent(const TIENT *ents, size_t nents) 149 if (nents == 0) 154 for (i = 0; i < nents; i++) {
|