HomeSort by: relevance | last modified time | path
    Searched refs:nents (Results 1 - 25 of 27) sorted by relevancy

1 2

  /src/sys/external/bsd/drm2/dist/drm/i915/
i915_scatterlist.c 21 if (orig_st->nents == orig_st->orig_nents)
24 if (sg_alloc_table(&new_st, orig_st->nents, GFP_KERNEL | __GFP_NOWARN))
28 for_each_sg(orig_st->sgl, sg, orig_st->nents, i) {
35 GEM_BUG_ON(new_sg); /* Should walk exactly nents and hit the end */
i915_gem_gtt.c 50 pages->sgl, pages->nents,
56 pages->sgl, pages->nents,
99 dma_unmap_sg(kdev, pages->sgl, pages->nents, PCI_DMA_BIDIRECTIONAL);
  /src/sys/external/bsd/drm2/dist/drm/virtio/
virtgpu_gem.c 153 struct virtio_gpu_object_array *virtio_gpu_array_alloc(u32 nents)
156 size_t size = sizeof(*objs) + sizeof(objs->objs[0]) * nents;
162 objs->nents = 0;
163 objs->total = nents;
173 virtio_gpu_array_from_handles(struct drm_file *drm_file, u32 *handles, u32 nents)
178 objs = virtio_gpu_array_alloc(nents);
182 for (i = 0; i < nents; i++) {
185 objs->nents = i;
190 objs->nents = i;
197 if (WARN_ON_ONCE(objs->nents == objs->total)
    [all...]
virtgpu_vq.c 599 bo->pages->sgl, bo->pages->nents,
621 uint32_t nents,
632 cmd_p->nr_entries = cpu_to_le32(nents);
635 vbuf->data_size = sizeof(*ents) * nents;
1014 bo->pages->sgl, bo->pages->nents,
1088 int si, nents, ret; local in function:virtio_gpu_object_attach
1107 obj->pages->sgl, obj->pages->nents,
1109 nents = obj->mapped;
1111 nents = obj->pages->nents;
    [all...]
virtgpu_drv.h 85 u32 nents, total; member in struct:virtio_gpu_object_array
244 struct virtio_gpu_object_array *virtio_gpu_array_alloc(u32 nents);
246 virtio_gpu_array_from_handles(struct drm_file *drm_file, u32 *handles, u32 nents);
  /src/sys/external/bsd/drm2/linux/
linux_sgt.c 50 sgt->sgl->sg_npgs = sgt->nents = npgs;
141 dma_map_sg(bus_dma_tag_t dmat, struct scatterlist *sg, int nents, int dir)
144 return dma_map_sg_attrs(dmat, sg, nents, dir, 0);
148 dma_map_sg_attrs(bus_dma_tag_t dmat, struct scatterlist *sg, int nents,
157 KASSERT(nents >= 1);
175 nents, PAGE_SIZE, 0, BUS_DMA_WAITOK, &sg->sg_dmamap);
188 KASSERT(sg->sg_dmamap->dm_nsegs <= nents);
206 dma_unmap_sg(bus_dma_tag_t dmat, struct scatterlist *sg, int nents, int dir)
209 dma_unmap_sg_attrs(dmat, sg, nents, dir, 0);
213 dma_unmap_sg_attrs(bus_dma_tag_t dmat, struct scatterlist *sg, int nents,
    [all...]
  /src/sys/external/bsd/drm2/dist/drm/i915/selftests/
scatterlist.c 56 for_each_sg(pt->st.sgl, sg, pt->st.nents, n) {
58 unsigned int npages = npages_fn(n, pt->st.nents, rnd);
94 for_each_sg_page(pt->st.sgl, &sgiter, pt->st.nents, 0) {
264 pt->st.nents = n;
351 pt.st.nents != prime) {
352 pr_err("i915_sg_trim failed (nents %u, orig_nents %u), expected %lu\n",
353 pt.st.nents, pt.st.orig_nents, prime);
intel_memory_region.c 156 if (obj->mm.pages->nents != 1) {
169 if (obj->mm.pages->nents != 1) {
195 if (obj->mm.pages->nents != 1) {
  /src/sys/external/bsd/drm2/dist/drm/i915/gem/selftests/
mock_dmabuf.c 36 if (!dma_map_sg(attachment->dev, st->sgl, st->nents, dir)) {
54 dma_unmap_sg(attachment->dev, st->sgl, st->nents, dir);
huge_pages.c 84 st->nents = 0;
107 st->nents++;
209 st->nents = 0;
225 st->nents++;
260 st->nents = 1;
  /src/sys/external/bsd/drm2/include/linux/
scatterlist.h 66 unsigned nents; member in struct:sg_table
  /src/sys/external/bsd/drm2/dist/drm/i915/gem/
i915_gem_dmabuf.c 46 ret = sg_alloc_table(st, obj->mm.pages->nents, GFP_KERNEL);
55 obj->mm.pages->nents * sizeof(st->sgl->sg_pgs[0]));
60 for (i = 0; i < obj->mm.pages->nents; i++) {
67 if (!dma_map_sg(attachment->dev, st->sgl, st->nents, dir)) {
90 dma_unmap_sg(attachment->dev, sg->sgl, sg->nents, dir);
i915_gem_region.c 122 st->nents = 0;
137 if (st->nents) {
147 st->nents++;
i915_gem_shmem.c 101 st->nents = 0;
158 KASSERT(st->nents == i);
159 sg->sg_pgs[st->nents++] = page;
169 st->nents++;
i915_gem_internal.c 163 st->nents = 0;
184 st->nents++;
  /src/sys/external/bsd/drm2/dist/drm/
drm_cache.c 135 for_each_sg_page(st->sgl, &sg_iter, st->nents, 0)
drm_prime.c 766 if (!dma_map_sg_attrs(attach->dev, sgt->sgl, sgt->nents, dir,
792 dma_unmap_sg_attrs(attach->dev, sgt->sgl, sgt->nents, dir,
1254 for_each_sg(sgt->sgl, sg, sgt->nents, count) {
  /src/sys/arch/i386/stand/bootxx/
boot1.c 202 uint32_t nents; local in function:gpt_lookup
245 nents = le32toh(hdr->hdr_entries);
267 for (i = 0; i < nents; i += entries_per_sector) {
  /src/sys/external/bsd/drm2/dist/drm/vmwgfx/
vmwgfx_ttm_buffer.c 370 dma_unmap_sg(dev, vmw_tt->sgt.sgl, vmw_tt->sgt.nents,
372 vmw_tt->sgt.nents = vmw_tt->sgt.orig_nents;
398 vmw_tt->sgt.nents = ret;
457 if (vsgt->num_pages > vmw_tt->sgt.nents) {
460 vmw_tt->sgt.nents);
  /src/sys/external/bsd/drm2/dist/drm/amd/amdgpu/
amdgpu_dma_buf.c 284 if (!dma_map_sg_attrs(attach->dev, sgt->sgl, sgt->nents, dir,
312 dma_unmap_sg(attach->dev, sgt->sgl, sgt->nents, dir);
amdgpu_ttm.c 995 unsigned nents;
1011 nents = dma_map_sg(adev->dev, ttm->sg->sgl, ttm->sg->nents, direction);
1012 if (nents != ttm->sg->nents)
1045 dma_unmap_sg(adev->dev, ttm->sg->sgl, ttm->sg->nents, direction);
  /src/sys/external/bsd/drm2/dist/drm/radeon/
radeon_ttm.c 521 unsigned pinned = 0, nents; local in function:radeon_ttm_tt_pin_userptr
640 nents = dma_map_sg(rdev->dev, ttm->sg->sgl, ttm->sg->nents, direction);
641 if (nents != ttm->sg->nents)
681 dma_unmap_sg(rdev->dev, ttm->sg->sgl, ttm->sg->nents, direction);
683 for_each_sg_page(ttm->sg->sgl, &sg_iter, ttm->sg->nents, 0) {
  /src/sys/external/bsd/drm2/dist/drm/xen/
xen_drm_front_gem.c 222 DRM_DEBUG("Imported buffer of size %zu with nents %u\n",
223 size, sgt->nents);
  /src/sys/external/bsd/drm2/dist/drm/i915/gt/
intel_ggtt.c 1452 st->nents++;
1489 st->nents = 0;
1535 st->nents++;
1571 st->nents = 0;
1724 st->nents = 0;
1735 st->nents++;
  /src/usr.bin/infocmp/
infocmp.c 144 print_ent(const TIENT *ents, size_t nents)
149 if (nents == 0)
154 for (i = 0; i < nents; i++) {

Completed in 29 milliseconds

1 2