HomeSort by: relevance | last modified time | path
    Searched defs:ring (Results 1 - 25 of 147) sorted by relevancy

1 2 3 4 5 6

  /src/sys/external/bsd/drm2/dist/drm/radeon/
radeon_semaphore.c 66 struct radeon_ring *ring = &rdev->ring[ridx]; local in function:radeon_semaphore_emit_signal
70 if (radeon_semaphore_ring_emit(rdev, ridx, ring, semaphore, false)) {
74 ring->last_semaphore_signal_addr = semaphore->gpu_addr;
83 struct radeon_ring *ring = &rdev->ring[ridx]; local in function:radeon_semaphore_emit_wait
87 if (radeon_semaphore_ring_emit(rdev, ridx, ring, semaphore, true)) {
91 ring->last_semaphore_wait_addr = semaphore->gpu_addr;
radeon_rv770_dma.c 55 struct radeon_ring *ring = &rdev->ring[ring_index]; local in function:rv770_copy_dma
64 r = radeon_ring_lock(rdev, ring, num_loops * 5 + 8);
72 radeon_sync_rings(rdev, &sync, ring->idx);
79 radeon_ring_write(ring, DMA_PACKET(DMA_PACKET_COPY, 0, 0, cur_size_in_dw));
80 radeon_ring_write(ring, dst_offset & 0xfffffffc);
81 radeon_ring_write(ring, src_offset & 0xfffffffc);
82 radeon_ring_write(ring, upper_32_bits(dst_offset) & 0xff);
83 radeon_ring_write(ring, upper_32_bits(src_offset) & 0xff);
88 r = radeon_fence_emit(rdev, &fence, ring->idx)
    [all...]
radeon_uvd_v2_2.c 42 * Write a fence and a trap command to the ring.
47 struct radeon_ring *ring = &rdev->ring[fence->ring]; local in function:uvd_v2_2_fence_emit
48 uint64_t addr = rdev->fence_drv[fence->ring].gpu_addr;
50 radeon_ring_write(ring, PACKET0(UVD_CONTEXT_ID, 0));
51 radeon_ring_write(ring, fence->seq);
52 radeon_ring_write(ring, PACKET0(UVD_GPCOM_VCPU_DATA0, 0));
53 radeon_ring_write(ring, lower_32_bits(addr));
54 radeon_ring_write(ring, PACKET0(UVD_GPCOM_VCPU_DATA1, 0))
    [all...]
radeon_evergreen_dma.c 37 * evergreen_dma_fence_ring_emit - emit a fence on the DMA ring
42 * Add a DMA fence packet to the ring to write
49 struct radeon_ring *ring = &rdev->ring[fence->ring]; local in function:evergreen_dma_fence_ring_emit
50 u64 addr = rdev->fence_drv[fence->ring].gpu_addr;
52 radeon_ring_write(ring, DMA_PACKET(DMA_PACKET_FENCE, 0, 0));
53 radeon_ring_write(ring, addr & 0xfffffffc);
54 radeon_ring_write(ring, (upper_32_bits(addr) & 0xff));
55 radeon_ring_write(ring, fence->seq)
75 struct radeon_ring *ring = &rdev->ring[ib->ring]; local in function:evergreen_dma_ring_ib_execute
121 struct radeon_ring *ring = &rdev->ring[ring_index]; local in function:evergreen_copy_dma
    [all...]
radeon_ib.c 44 * command ring and the hw will fetch the commands from the IB
47 * put in IBs for execution by the requested ring.
55 * @ring: ring index the IB is associated with
63 int radeon_ib_get(struct radeon_device *rdev, int ring,
77 ib->ring = ring;
110 * radeon_ib_schedule - schedule an IB (Indirect Buffer) on the ring
117 * Schedule an IB on the associated ring (all asics).
120 * On SI, there are two parallel engines fed from the primary ring,
133 struct radeon_ring *ring = &rdev->ring[ib->ring]; local in function:radeon_ib_schedule
271 struct radeon_ring *ring = &rdev->ring[i]; local in function:radeon_ib_ring_tests
    [all...]
radeon_si_dma.c 41 * @ring: radeon_ring structure holding ring information
46 bool si_dma_is_lockup(struct radeon_device *rdev, struct radeon_ring *ring)
51 if (ring->idx == R600_RING_TYPE_DMA_INDEX)
57 radeon_ring_lockup_update(rdev, ring);
60 return radeon_ring_test_lockup(rdev, ring);
192 void si_dma_vm_flush(struct radeon_device *rdev, struct radeon_ring *ring,
196 radeon_ring_write(ring, DMA_PACKET(DMA_PACKET_SRBM_WRITE, 0, 0, 0, 0));
198 radeon_ring_write(ring, (0xf << 16) | ((VM_CONTEXT0_PAGE_TABLE_BASE_ADDR + (vm_id << 2)) >> 2));
200 radeon_ring_write(ring, (0xf << 16) | ((VM_CONTEXT8_PAGE_TABLE_BASE_ADDR + ((vm_id - 8) << 2)) >> 2))
244 struct radeon_ring *ring = &rdev->ring[ring_index]; local in function:si_copy_dma
    [all...]
radeon_r200.c 94 struct radeon_ring *ring = &rdev->ring[RADEON_RING_TYPE_GFX_INDEX]; local in function:r200_copy_dma
104 r = radeon_ring_lock(rdev, ring, num_loops * 4 + 64);
110 radeon_ring_write(ring, PACKET0(RADEON_WAIT_UNTIL, 0));
111 radeon_ring_write(ring, (1 << 16));
118 radeon_ring_write(ring, PACKET0(0x720, 2));
119 radeon_ring_write(ring, src_offset);
120 radeon_ring_write(ring, dst_offset);
121 radeon_ring_write(ring, cur_size | (1 << 31) | (1 << 30));
125 radeon_ring_write(ring, PACKET0(RADEON_WAIT_UNTIL, 0))
    [all...]
radeon_r420.c 219 struct radeon_ring *ring = &rdev->ring[RADEON_RING_TYPE_GFX_INDEX]; local in function:r420_cp_errata_init
228 r = radeon_ring_lock(rdev, ring, 8);
230 radeon_ring_write(ring, PACKET0(R300_CP_RESYNC_ADDR, 1));
231 radeon_ring_write(ring, rdev->config.r300.resync_scratch);
232 radeon_ring_write(ring, 0xDEADBEEF);
233 radeon_ring_unlock_commit(rdev, ring, false);
239 struct radeon_ring *ring = &rdev->ring[RADEON_RING_TYPE_GFX_INDEX]; local in function:r420_cp_errata_fini
244 r = radeon_ring_lock(rdev, ring, 8)
    [all...]
radeon_uvd_v1_0.c 40 * @ring: radeon_ring pointer
45 struct radeon_ring *ring)
54 * @ring: radeon_ring pointer
59 struct radeon_ring *ring)
68 * @ring: radeon_ring pointer
73 struct radeon_ring *ring)
75 WREG32(UVD_RBC_RB_WPTR, ring->wptr);
84 * Write a fence and a trap command to the ring.
89 struct radeon_ring *ring = &rdev->ring[fence->ring] local in function:uvd_v1_0_fence_emit
164 struct radeon_ring *ring = &rdev->ring[R600_RING_TYPE_UVD_INDEX]; local in function:uvd_v1_0_init
255 struct radeon_ring *ring = &rdev->ring[R600_RING_TYPE_UVD_INDEX]; local in function:uvd_v1_0_fini
270 struct radeon_ring *ring = &rdev->ring[R600_RING_TYPE_UVD_INDEX]; local in function:uvd_v1_0_start
489 struct radeon_ring *ring = &rdev->ring[ib->ring]; local in function:uvd_v1_0_ib_execute
    [all...]
  /src/sys/external/bsd/drm2/dist/drm/amd/amdgpu/
amdgpu_test.c 39 struct amdgpu_ring *ring = adev->mman.buffer_funcs_ring; local in function:amdgpu_do_test_moves
50 * (Total GTT - IB pool - writeback page - ring buffers) / test size
131 r = amdgpu_copy_buffer(ring, gart_addr, vram_addr,
177 r = amdgpu_copy_buffer(ring, vram_addr, gart_addr,
amdgpu_benchmark.c 46 struct amdgpu_ring *ring = adev->mman.buffer_funcs_ring; local in function:amdgpu_benchmark_do_move
47 r = amdgpu_copy_buffer(ring, saddr, daddr, size, NULL, &fence,
amdgpu_ih.h 36 * R6xx+ IH ring
50 volatile uint32_t *ring; member in struct:amdgpu_ih_ring
66 /* ring read/write ptr handling, called from interrupt context */
amdgpu_job.c 38 struct amdgpu_ring *ring = to_amdgpu_ring(s_job->sched); local in function:amdgpu_job_timedout
44 if (amdgpu_ring_soft_recovery(ring, job->vmid, s_job->s_fence->parent)) {
45 DRM_ERROR("ring %s timeout, but soft recovered\n",
50 amdgpu_vm_get_task_info(ring->adev, job->pasid, &ti);
51 DRM_ERROR("ring %s timeout, signaled seq=%u, emitted seq=%u\n",
52 job->base.sched->name, atomic_read(&ring->fence_drv.last_seq),
53 ring->fence_drv.sync_seq);
57 if (amdgpu_device_should_recover_gpu(ring->adev))
58 amdgpu_device_gpu_recover(ring->adev, job);
60 drm_sched_suspend_timeout(&ring->sched)
112 struct amdgpu_ring *ring = to_amdgpu_ring(job->base.sched); local in function:amdgpu_job_free_resources
125 struct amdgpu_ring *ring = to_amdgpu_ring(s_job->sched); local in function:amdgpu_job_free_cb
151 struct amdgpu_ring *ring; local in function:amdgpu_job_submit
190 struct amdgpu_ring *ring = to_amdgpu_ring(s_entity->rq->sched); local in function:amdgpu_job_dependency
221 struct amdgpu_ring *ring = to_amdgpu_ring(sched_job->sched); local in function:amdgpu_job_run
    [all...]
amdgpu_psp_v10_0.c 145 struct psp_ring *ring; local in function:psp_v10_0_ring_init
148 ring = &psp->km_ring;
150 ring->ring_type = ring_type;
152 /* allocate 4k Page of Local Frame Buffer memory for ring */
153 ring->ring_size = 0x1000;
154 ret = amdgpu_bo_create_kernel(adev, ring->ring_size, PAGE_SIZE,
157 &ring->ring_mem_mc_addr,
158 (void **)&ring->ring_mem);
160 ring->ring_size = 0;
172 struct psp_ring *ring = &psp->km_ring local in function:psp_v10_0_ring_create
224 struct psp_ring *ring = &psp->km_ring; local in function:psp_v10_0_ring_destroy
    [all...]
amdgpu_vce.h 50 struct amdgpu_ring ring[AMDGPU_MAX_VCE_RINGS]; member in struct:amdgpu_vce
66 void amdgpu_vce_ring_emit_ib(struct amdgpu_ring *ring, struct amdgpu_job *job,
68 void amdgpu_vce_ring_emit_fence(struct amdgpu_ring *ring, u64 addr, u64 seq,
70 int amdgpu_vce_ring_test_ring(struct amdgpu_ring *ring);
71 int amdgpu_vce_ring_test_ib(struct amdgpu_ring *ring, long timeout);
72 void amdgpu_vce_ring_begin_use(struct amdgpu_ring *ring);
73 void amdgpu_vce_ring_end_use(struct amdgpu_ring *ring);
74 unsigned amdgpu_vce_ring_get_emit_ib_size(struct amdgpu_ring *ring);
75 unsigned amdgpu_vce_ring_get_dma_frame_size(struct amdgpu_ring *ring);
amdgpu_virt.c 58 struct amdgpu_ring *ring = &kiq->ring; local in function:amdgpu_virt_kiq_reg_write_reg_wait
64 amdgpu_ring_alloc(ring, 32);
65 amdgpu_ring_emit_reg_write_reg_wait(ring, reg0, reg1,
67 amdgpu_fence_emit_polling(ring, &seq);
68 amdgpu_ring_commit(ring);
71 r = amdgpu_fence_wait_polling(ring, seq, MAX_KIQ_REG_WAIT);
81 r = amdgpu_fence_wait_polling(ring, seq, MAX_KIQ_REG_WAIT);
amdgpu_vm_sdma.c 106 struct amdgpu_ring *ring; local in function:amdgpu_vm_sdma_commit
110 ring = container_of(entity->rq->sched, struct amdgpu_ring, sched);
113 amdgpu_ring_pad_ib(ring, ib);
amdgpu_gmc.c 301 /* If we don't have space left in the ring buffer return immediately */
324 /* Add the fault to the ring */
364 * Engine 0, 1, 4~16: amdgpu ring,
365 * subject to change when ring number changes
373 struct amdgpu_ring *ring; local in function:amdgpu_gmc_allocate_vm_inv_eng
381 ring = adev->rings[i];
382 vmhub = ring->funcs->vmhub;
386 dev_err(adev->dev, "no VM inv eng for ring %s\n",
387 ring->name);
391 ring->vm_inv_eng = inv_eng - 1
    [all...]
amdgpu_ib.c 51 * command ring and the hw will fetch the commands from the IB
54 * put in IBs for execution by the requested ring.
61 * @ring: ring index the IB is associated with
107 * amdgpu_ib_schedule - schedule an IB (Indirect Buffer) on the ring
114 * Schedule an IB on the associated ring (all asics).
117 * On SI, there are two parallel engines fed from the primary ring,
124 * a CONST_IB), it will be put on the ring prior to the DE IB. Prior
127 int amdgpu_ib_schedule(struct amdgpu_ring *ring, unsigned num_ibs,
131 struct amdgpu_device *adev = ring->adev
363 struct amdgpu_ring *ring = adev->rings[i]; local in function:amdgpu_ib_ring_tests
    [all...]
amdgpu_jpeg_v1_0.c 41 static void jpeg_v1_0_decode_ring_patch_wreg(struct amdgpu_ring *ring, uint32_t *ptr, uint32_t reg_offset, uint32_t val)
43 struct amdgpu_device *adev = ring->adev;
44 ring->ring[(*ptr)++] = PACKETJ(SOC15_REG_OFFSET(JPEG, 0, mmUVD_JRBC_EXTERNAL_REG_BASE), 0, 0, PACKETJ_TYPE0);
47 ring->ring[(*ptr)++] = 0;
48 ring->ring[(*ptr)++] = PACKETJ((reg_offset >> 2), 0, 0, PACKETJ_TYPE0);
50 ring->ring[(*ptr)++] = reg_offset
478 struct amdgpu_ring *ring; local in function:jpeg_v1_0_sw_init
521 struct amdgpu_ring *ring = &adev->jpeg.inst->ring_dec; local in function:jpeg_v1_0_start
    [all...]
amdgpu_jpeg_v2_5.c 58 * Set ring and irq function pointers
95 struct amdgpu_ring *ring; local in function:jpeg_v2_5_sw_init
122 ring = &adev->jpeg.inst[i].ring_dec;
123 ring->use_doorbell = true;
124 ring->doorbell_index = (adev->doorbell_index.vcn.vcn_ring0_1 << 1) + 1 + 8 * i;
125 snprintf(ring->name, sizeof(ring->name), "jpeg_dec_%d", i);
126 r = amdgpu_ring_init(adev, ring, 512, &adev->jpeg.inst[i].irq, 0);
167 struct amdgpu_ring *ring; local in function:jpeg_v2_5_hw_init
174 ring = &adev->jpeg.inst[i].ring_dec
198 struct amdgpu_ring *ring; local in function:jpeg_v2_5_hw_fini
310 struct amdgpu_ring *ring; local in function:jpeg_v2_5_start
    [all...]
amdgpu_psp_v12_0.c 182 /* Change IH ring for VMC */
195 /* Change IH ring for UMC */
212 struct psp_ring *ring; local in function:psp_v12_0_ring_init
217 ring = &psp->km_ring;
219 ring->ring_type = ring_type;
221 /* allocate 4k Page of Local Frame Buffer memory for ring */
222 ring->ring_size = 0x1000;
223 ret = amdgpu_bo_create_kernel(adev, ring->ring_size, PAGE_SIZE,
226 &ring->ring_mem_mc_addr,
227 (void **)&ring->ring_mem)
248 struct psp_ring *ring = &psp->km_ring; local in function:psp_v12_0_ring_create
328 struct psp_ring *ring = &psp->km_ring; local in function:psp_v12_0_ring_destroy
    [all...]
  /src/games/rogue/
ring.c 1 /* $NetBSD: ring.c,v 1.9 2008/01/14 03:50:02 dholland Exp $ */
38 static char sccsid[] = "@(#)ring.c 8.1 (Berkeley) 5/31/93";
40 __RCSID("$NetBSD: ring.c,v 1.9 2008/01/14 03:50:02 dholland Exp $");
45 * ring.c
59 static const char no_ring[] = "there's no ring on that hand";
78 object *ring; local in function:put_on_ring
84 if ((ch = pack_letter("put on what?", RING)) == CANCEL) {
87 if (!(ring = get_letter_object(ch))) {
91 if (!(ring->what_is & RING)) {
152 object *ring; local in function:remove_ring
279 object *ring; local in function:ring_stats
    [all...]
  /src/sys/external/bsd/drm2/dist/drm/i915/gt/
intel_ring.c 19 unsigned int intel_ring_update_space(struct intel_ring *ring)
23 space = __intel_ring_space(ring->head, ring->emit, ring->size);
25 ring->space = space;
29 int intel_ring_pin(struct intel_ring *ring)
31 struct i915_vma *vma = ring->vma;
36 if (atomic_fetch_inc(&ring->pin_count))
41 /* Ring wraparound at offset 0 sometimes hangs. No idea why. */
66 intel_ring_reset(ring, ring->emit)
141 struct intel_ring *ring; local in function:intel_engine_create_ring
178 struct intel_ring *ring = container_of(ref, typeof(*ring), ref); local in function:intel_ring_free
224 struct intel_ring *ring = rq->ring; local in function:intel_ring_begin
    [all...]
  /src/sys/external/mit/xen-include-public/dist/xen/include/public/io/
tpmif.h 63 struct tpmif_ring ring[TPMIF_TX_RING_SIZE]; member in struct:tpmif_tx_interface

Completed in 32 milliseconds

1 2 3 4 5 6