| /src/sys/external/bsd/drm2/dist/drm/i915/ |
| i915_scatterlist.c | 18 struct scatterlist *sg, *new_sg; local 28 for_each_sg(orig_st->sgl, sg, orig_st->nents, i) { 29 sg_set_page(new_sg, sg_page(sg), sg->length, 0); 30 sg_dma_address(new_sg) = sg_dma_address(sg); 31 sg_dma_len(new_sg) = sg_dma_len(sg);
|
| /src/external/gpl3/gdb/dist/gdb/testsuite/gdb.base/ |
| shr1.c | 6 float sg = 5.5; variable 20 sg = 6.6;
|
| /src/external/gpl3/gdb.old/dist/gdb/testsuite/gdb.base/ |
| shr1.c | 6 float sg = 5.5; variable 20 sg = 6.6;
|
| /src/sys/external/bsd/drm2/dist/drm/i915/gem/selftests/ |
| mock_dmabuf.c | 19 struct scatterlist *sg; local 30 sg = st->sgl; 32 sg_set_page(sg, mock->pages[i], PAGE_SIZE, 0); 33 sg = sg_next(sg);
|
| huge_gem_object.c | 38 struct scatterlist *sg, *src, *end; local 51 sg = pages->sgl; 57 sg_mark_end(sg); 61 sg_set_page(sg, page, PAGE_SIZE, 0); 62 sg = __sg_next(sg); 65 for (end = sg, src = pages->sgl; sg; sg = __sg_next(sg)) { [all...] |
| /src/sys/external/bsd/drm2/drm/ |
| drm_scatter.c | 59 struct drm_sg_mem *sg = NULL; local 73 if (dev->sg != NULL) 84 error = drm_sg_alloc_mem(dev, request->size, &sg); 89 dev->sg = sg; 92 request->handle = sg->handle; 100 struct drm_sg_mem *sg; local 110 sg = dev->sg; 111 if (sg == NULL 156 struct drm_sg_mem *const sg = local [all...] |
| drm_vm.c | 181 struct drm_sg_mem *const sg = dev->sg; local 184 KASSERT(sg == map->lm_data.sg); 187 return bus_dmamem_mmap(dev->dmat, sg->sg_segs, sg->sg_nsegs,
|
| /src/external/lgpl3/mpfr/dist/tests/ |
| tget_flt.c | 38 int sf, sg; local 45 sg = mpfr_signbit (z); 47 return !sf == !sg;
|
| /src/sys/external/bsd/drm2/dist/drm/i915/gem/ |
| i915_gem_internal.c | 30 struct scatterlist *sg; local 32 for (sg = st->sgl; sg; sg = __sg_next(sg)) { 33 if (sg_page(sg)) 34 __free_pages(sg_page(sg), get_order(sg->length)); 124 struct scatterlist *sg; 162 sg = st->sgl [all...] |
| i915_gem_region.c | 40 struct scatterlist *sg; local 72 sg = NULL; 98 sg = st->sgl; 102 &sg->sg_dmamap); 104 sg->sg_dmamap = NULL; 107 sg->sg_dmat = dmat; 110 ret = -bus_dmamap_load_raw(dmat, sg->sg_dmamap, segs, nsegs, size, 119 sg_page_sizes = i915_sg_page_sizes(sg); 121 sg = st->sgl; 133 GEM_BUG_ON(overflows_type(block_size, sg->length)) [all...] |
| i915_gem_phys.c | 120 struct scatterlist *sg; local 178 sg = st->sgl; 183 &sg->sg_dmamap); 185 sg->sg_dmamap = NULL; 188 sg->sg_dmat = dmat; 190 ret = -bus_dmamap_load_raw(dmat, sg->sg_dmamap, &obj->mm.u.phys.seg, 1, 196 sg->offset = 0; 197 sg->length = obj->base.size; 199 sg_assign_page(sg, (struct page *)vaddr); 200 sg_dma_address(sg) = dma [all...] |
| i915_gem_pages.c | 81 obj->mm.page_sizes.sg = 0; 84 obj->mm.page_sizes.sg |= BIT(i); 86 GEM_BUG_ON(!HAS_PAGE_SIZES(i915, obj->mm.page_sizes.sg)); 217 obj->mm.page_sizes.phys = obj->mm.page_sizes.sg = 0; 498 struct scatterlist *sg; local 505 /* As we iterate forward through the sg, we record each entry in a 519 /* We prefer to reuse the last sg so that repeated lookup of this 520 * (or the subsequent) sg are fast - comparing against the last 521 * sg is faster than going through the radixtree. 524 sg = iter->sg_pos [all...] |
| i915_gem_shmem.c | 47 struct scatterlist *sg; local 100 sg = st->sgl; 159 sg->sg_pgs[st->nents++] = page; 163 sg->length >= max_segment || 166 sg_page_sizes |= sg->length; 167 sg = sg_next(sg); 170 sg_set_page(sg, page, PAGE_SIZE, 0); 172 sg->length += PAGE_SIZE; 181 if (sg) { /* loop terminated early; short sg table * [all...] |
| i915_gem_stolen.c | 509 struct scatterlist *sg; local 545 sg = NULL; 549 DRM_ERROR("failed to alloc sg table for stolen object: %d\n", 554 sg = st->sgl; 581 if (sg && sg->sg_dmamap) 582 bus_dmamap_destroy(dmat, sg->sg_dmamap); 583 if (sg) 594 sg = st->sgl; 595 sg->offset = 0 [all...] |
| /src/sys/compat/common/ |
| tty_43.c | 129 struct sgttyb *sg = (struct sgttyb *)data; local 134 sg->sg_ospeed = (speed == -1) ? MAX_SPEED : speed; 136 sg->sg_ispeed = sg->sg_ospeed; 139 sg->sg_ispeed = (speed == -1) ? MAX_SPEED : speed; 141 sg->sg_erase = tty_getctrlchar(tp, VERASE); 142 sg->sg_kill = tty_getctrlchar(tp, VKILL); 143 sg->sg_flags = ttcompatgetflags(tp); 150 struct sgttyb *sg = (struct sgttyb *)data; local 156 if ((speed = sg->sg_ispeed) > MAX_SPEED || speed < 0 [all...] |
| /src/sys/external/bsd/drm2/linux/ |
| linux_dma_buf.c | 232 struct sg_table *sg; local 236 sg = attach->dmabuf->ops->map_dma_buf(attach, dir); 240 return sg; 245 struct sg_table *sg, enum dma_data_direction dir) 250 attach->dmabuf->ops->unmap_dma_buf(attach, sg, dir);
|
| /src/sys/external/bsd/drm2/dist/include/drm/ttm/ |
| ttm_tt.h | 113 struct sg_table *sg; /* for SG objects via dma-buf */ member in struct:ttm_tt
|
| /src/tests/crypto/opencrypto/ |
| h_ioctl.c | 100 struct crypt_sgop sg; local 111 memset(&sg, 0, sizeof(sg)); 112 sg.count = COUNT; 113 sg.sessions = css; 115 ret = ioctl(fd, CIOCNGSESSION, &sg);
|
| /src/external/bsd/iscsi/dist/src/osd/ |
| osd.c | 151 struct iovec *sg = (struct iovec *) arg; local 154 while (sg[i].iov_base != NULL) { 155 iscsi_free_atomic(sg[i].iov_base); 171 struct iovec sg[3]; local 253 sg[sg_len].iov_base = set_list; 254 sg[sg_len].iov_len = osd_args.set_attributes_list_length; 262 sg[sg_len].iov_base = get_list; 263 sg[sg_len].iov_len = osd_args.get_attributes_list_length; 271 sg[sg_len].iov_base = write_data; 272 sg[sg_len].iov_len = osd_args.length [all...] |
| /src/external/bsd/tmux/dist/ |
| cmd-new-session.c | 78 struct session_group *sg = NULL; local 136 sg = session_group_find(group); 138 sg = session_group_contains(groupwith); 139 if (sg != NULL) 140 prefix = xstrdup(sg->name); 306 if (sg == NULL) { 308 sg = session_group_new(groupwith->name); 309 session_group_add(sg, groupwith); 311 sg = session_group_new(group); 313 session_group_add(sg, s) [all...] |
| /src/external/lgpl3/mpfr/dist/src/ |
| rndna.c | 40 mpfr_sign_t sg; member in union:__anon23609 94 ext[OLD_SIGN].sg = MPFR_SIGN(rop); 155 MPFR_SIGN(rop) = ext[OLD_SIGN].sg;
|
| /src/sys/dev/ic/ |
| aic79xx_osm.c | 548 void *sg; local 554 /* Copy the segments into our SG list */ 555 for (i = nsegments, sg = scb->sg_list; i > 0; i--) { 557 sg = ahd_sg_setup(ahd, scb, sg, dm_segs->ds_addr,
|
| dpt.c | 833 struct eata_sg *sg; local 878 sg = ccb->ccb_sg; 880 for (i = 0; i < xfer->dm_nsegs; i++, sg++, ds++) { 881 sg->sg_addr = htobe32(ds->ds_addr); 882 sg->sg_len = htobe32(ds->ds_len); 1281 struct eata_sg *sg; local 1329 sg = ccb->ccb_sg; 1331 for (i = 0; i < xfer->dm_nsegs; i++, sg++, ds++) { 1332 sg->sg_addr = htobe32(ds->ds_addr); 1333 sg->sg_len = htobe32(ds->ds_len) [all...] |
| /src/sys/external/bsd/drm2/dist/drm/i915/gvt/ |
| dmabuf.c | 68 struct scatterlist *sg; local 94 for_each_sg(st->sgl, sg, page_num, i) { 102 sg->offset = 0; 103 sg->length = PAGE_SIZE; 104 sg_dma_len(sg) = PAGE_SIZE; 105 sg_dma_address(sg) = dma_addr; 113 for_each_sg(st->sgl, sg, i, j) { 114 dma_addr = sg_dma_address(sg); 129 struct scatterlist *sg; local 137 for_each_sg(pages->sgl, sg, fb_info->size, i [all...] |
| /src/sys/external/bsd/drm2/dist/drm/i915/selftests/ |
| scatterlist.c | 52 struct scatterlist *sg; local 56 for_each_sg(pt->st.sgl, sg, pt->st.nents, n) { 57 struct page *page = sg_page(sg); 66 if (sg->length != npages * PAGE_SIZE) { 67 pr_err("%s: %s copied wrong sg length, expected size %lu, found %u (using for_each_sg)\n", 68 __func__, who, npages * PAGE_SIZE, sg->length); 225 struct scatterlist *sg; local 232 /* count should be less than 20 to prevent overflowing sg->length */ 233 GEM_BUG_ON(overflows_type(count * PAGE_SIZE, sg->length)); 241 sg = pt->st.sgl [all...] |