| /src/sys/external/bsd/drm2/dist/drm/i915/ |
| i915_scatterlist.c | 18 struct scatterlist *sg, *new_sg; local 28 for_each_sg(orig_st->sgl, sg, orig_st->nents, i) { 29 sg_set_page(new_sg, sg_page(sg), sg->length, 0); 30 sg_dma_address(new_sg) = sg_dma_address(sg); 31 sg_dma_len(new_sg) = sg_dma_len(sg);
|
| /src/sys/external/bsd/drm2/dist/drm/i915/gem/selftests/ |
| mock_dmabuf.c | 19 struct scatterlist *sg; local 30 sg = st->sgl; 32 sg_set_page(sg, mock->pages[i], PAGE_SIZE, 0); 33 sg = sg_next(sg);
|
| huge_gem_object.c | 38 struct scatterlist *sg, *src, *end; local 51 sg = pages->sgl; 57 sg_mark_end(sg); 61 sg_set_page(sg, page, PAGE_SIZE, 0); 62 sg = __sg_next(sg); 65 for (end = sg, src = pages->sgl; sg; sg = __sg_next(sg)) { [all...] |
| /src/sys/external/bsd/drm2/drm/ |
| drm_scatter.c | 59 struct drm_sg_mem *sg = NULL; local 73 if (dev->sg != NULL) 84 error = drm_sg_alloc_mem(dev, request->size, &sg); 89 dev->sg = sg; 92 request->handle = sg->handle; 100 struct drm_sg_mem *sg; local 110 sg = dev->sg; 111 if (sg == NULL 156 struct drm_sg_mem *const sg = local [all...] |
| drm_vm.c | 181 struct drm_sg_mem *const sg = dev->sg; local 184 KASSERT(sg == map->lm_data.sg); 187 return bus_dmamem_mmap(dev->dmat, sg->sg_segs, sg->sg_nsegs,
|
| /src/sys/external/bsd/drm2/dist/drm/i915/gem/ |
| i915_gem_internal.c | 30 struct scatterlist *sg; local 32 for (sg = st->sgl; sg; sg = __sg_next(sg)) { 33 if (sg_page(sg)) 34 __free_pages(sg_page(sg), get_order(sg->length)); 124 struct scatterlist *sg; 162 sg = st->sgl [all...] |
| i915_gem_region.c | 40 struct scatterlist *sg; local 72 sg = NULL; 98 sg = st->sgl; 102 &sg->sg_dmamap); 104 sg->sg_dmamap = NULL; 107 sg->sg_dmat = dmat; 110 ret = -bus_dmamap_load_raw(dmat, sg->sg_dmamap, segs, nsegs, size, 119 sg_page_sizes = i915_sg_page_sizes(sg); 121 sg = st->sgl; 133 GEM_BUG_ON(overflows_type(block_size, sg->length)) [all...] |
| i915_gem_phys.c | 120 struct scatterlist *sg; local 178 sg = st->sgl; 183 &sg->sg_dmamap); 185 sg->sg_dmamap = NULL; 188 sg->sg_dmat = dmat; 190 ret = -bus_dmamap_load_raw(dmat, sg->sg_dmamap, &obj->mm.u.phys.seg, 1, 196 sg->offset = 0; 197 sg->length = obj->base.size; 199 sg_assign_page(sg, (struct page *)vaddr); 200 sg_dma_address(sg) = dma [all...] |
| i915_gem_pages.c | 81 obj->mm.page_sizes.sg = 0; 84 obj->mm.page_sizes.sg |= BIT(i); 86 GEM_BUG_ON(!HAS_PAGE_SIZES(i915, obj->mm.page_sizes.sg)); 217 obj->mm.page_sizes.phys = obj->mm.page_sizes.sg = 0; 498 struct scatterlist *sg; local 505 /* As we iterate forward through the sg, we record each entry in a 519 /* We prefer to reuse the last sg so that repeated lookup of this 520 * (or the subsequent) sg are fast - comparing against the last 521 * sg is faster than going through the radixtree. 524 sg = iter->sg_pos [all...] |
| i915_gem_shmem.c | 47 struct scatterlist *sg; local 100 sg = st->sgl; 159 sg->sg_pgs[st->nents++] = page; 163 sg->length >= max_segment || 166 sg_page_sizes |= sg->length; 167 sg = sg_next(sg); 170 sg_set_page(sg, page, PAGE_SIZE, 0); 172 sg->length += PAGE_SIZE; 181 if (sg) { /* loop terminated early; short sg table * [all...] |
| i915_gem_stolen.c | 509 struct scatterlist *sg; local 545 sg = NULL; 549 DRM_ERROR("failed to alloc sg table for stolen object: %d\n", 554 sg = st->sgl; 581 if (sg && sg->sg_dmamap) 582 bus_dmamap_destroy(dmat, sg->sg_dmamap); 583 if (sg) 594 sg = st->sgl; 595 sg->offset = 0 [all...] |
| /src/sys/compat/common/ |
| tty_43.c | 129 struct sgttyb *sg = (struct sgttyb *)data; local 134 sg->sg_ospeed = (speed == -1) ? MAX_SPEED : speed; 136 sg->sg_ispeed = sg->sg_ospeed; 139 sg->sg_ispeed = (speed == -1) ? MAX_SPEED : speed; 141 sg->sg_erase = tty_getctrlchar(tp, VERASE); 142 sg->sg_kill = tty_getctrlchar(tp, VKILL); 143 sg->sg_flags = ttcompatgetflags(tp); 150 struct sgttyb *sg = (struct sgttyb *)data; local 156 if ((speed = sg->sg_ispeed) > MAX_SPEED || speed < 0 [all...] |
| /src/sys/external/bsd/drm2/linux/ |
| linux_dma_buf.c | 232 struct sg_table *sg; local 236 sg = attach->dmabuf->ops->map_dma_buf(attach, dir); 240 return sg; 245 struct sg_table *sg, enum dma_data_direction dir) 250 attach->dmabuf->ops->unmap_dma_buf(attach, sg, dir);
|
| /src/sys/external/bsd/drm2/dist/include/drm/ttm/ |
| ttm_tt.h | 113 struct sg_table *sg; /* for SG objects via dma-buf */ member in struct:ttm_tt
|
| /src/tests/crypto/opencrypto/ |
| h_ioctl.c | 100 struct crypt_sgop sg; local 111 memset(&sg, 0, sizeof(sg)); 112 sg.count = COUNT; 113 sg.sessions = css; 115 ret = ioctl(fd, CIOCNGSESSION, &sg);
|
| /src/sys/dev/ic/ |
| aic79xx_osm.c | 548 void *sg; local 554 /* Copy the segments into our SG list */ 555 for (i = nsegments, sg = scb->sg_list; i > 0; i--) { 557 sg = ahd_sg_setup(ahd, scb, sg, dm_segs->ds_addr,
|
| dpt.c | 833 struct eata_sg *sg; local 878 sg = ccb->ccb_sg; 880 for (i = 0; i < xfer->dm_nsegs; i++, sg++, ds++) { 881 sg->sg_addr = htobe32(ds->ds_addr); 882 sg->sg_len = htobe32(ds->ds_len); 1281 struct eata_sg *sg; local 1329 sg = ccb->ccb_sg; 1331 for (i = 0; i < xfer->dm_nsegs; i++, sg++, ds++) { 1332 sg->sg_addr = htobe32(ds->ds_addr); 1333 sg->sg_len = htobe32(ds->ds_len) [all...] |
| aic7xxx_osm.c | 538 struct ahc_dma_seg *sg; local 544 /* Copy the segments into our SG list */ 545 sg = scb->sg_list; 549 sg->addr = ahc_htole32(dm_segs->ds_addr); 552 sg->len = ahc_htole32(len); 553 sg++; 558 * Note where to find the SG entries in bus space. 573 sg--; 574 sg->len |= ahc_htole32(AHC_DMA_LAST_SEG); 576 /* Copy the first SG into the "current" data pointer area * [all...] |
| icp.c | 1075 struct icp_sg *sg; local 1091 sg = ic->ic_sg; 1093 if (sg != NULL) { 1094 for (i = 0; i < nsegs; i++, sg++) { 1095 sg->sg_addr = htole32(xfer->dm_segs[i].ds_addr); 1096 sg->sg_len = htole32(xfer->dm_segs[i].ds_len); 1099 panic("icp_ccb_map: no SG list specified, but nsegs > 1");
|
| /src/sys/external/bsd/drm2/dist/drm/i915/gvt/ |
| dmabuf.c | 68 struct scatterlist *sg; local 94 for_each_sg(st->sgl, sg, page_num, i) { 102 sg->offset = 0; 103 sg->length = PAGE_SIZE; 104 sg_dma_len(sg) = PAGE_SIZE; 105 sg_dma_address(sg) = dma_addr; 113 for_each_sg(st->sgl, sg, i, j) { 114 dma_addr = sg_dma_address(sg); 129 struct scatterlist *sg; local 137 for_each_sg(pages->sgl, sg, fb_info->size, i [all...] |
| /src/sys/external/bsd/drm2/dist/drm/i915/selftests/ |
| scatterlist.c | 52 struct scatterlist *sg; local 56 for_each_sg(pt->st.sgl, sg, pt->st.nents, n) { 57 struct page *page = sg_page(sg); 66 if (sg->length != npages * PAGE_SIZE) { 67 pr_err("%s: %s copied wrong sg length, expected size %lu, found %u (using for_each_sg)\n", 68 __func__, who, npages * PAGE_SIZE, sg->length); 225 struct scatterlist *sg; local 232 /* count should be less than 20 to prevent overflowing sg->length */ 233 GEM_BUG_ON(overflows_type(count * PAGE_SIZE, sg->length)); 241 sg = pt->st.sgl [all...] |
| /src/usr.bin/resize/ |
| resize.c | 333 struct sgttyb sg; local 469 sg = sgorig; 470 sg.sg_flags |= RAW; 471 UIntClr(sg.sg_flags, ECHO); 485 rc = ioctl(tty, TIOCSETP, &sg);
|
| /src/usr.sbin/syslogd/ |
| sign.h | 54 * 0 one global SG 55 * 1 one SG per PRI 59 * We use '3' and assign one SG to every destination (=struct filed) 163 * The fields .sg and .sig2_delims are set by init() and are always 171 int sg; member in struct:sign_global_t
|
| /src/sys/arch/amiga/amiga/ |
| amiga_init.c | 212 register st_entry_t sg_proto, *sg; local 444 sg = (st_entry_t *)RELOC(Sysseg_pa, u_int); 445 esg = &sg[kstsize * NPTEPG]; 446 while (sg < esg) 447 *sg++ = SG_NV; 457 sg = (st_entry_t *)RELOC(Sysseg_pa, u_int); 458 sg = &sg[SG4_LEV1SIZE]; 459 esg = &sg[nl2desc]; 461 while (sg < esg) [all...] |
| /src/sys/arch/sun3/include/ |
| mc68851.h | 67 * | RAL | WAL |SG | S | 0 | 0 | 0 | 0 | U |WP |DT (10)| 85 char sg:1; /* Shared Globally flag */ member in struct:mmu_long_dte_struct::__anon2116::__anon2117 128 * | RAL | WAL |SG | S | G |CI | L | M | U |WP |DT (01)| 142 char sg:1; /* Shared Globally flag */ member in struct:mmu_long_pte_struct::__anon2120::__anon2121
|