/src/sys/external/bsd/drm2/dist/drm/radeon/ |
radeon_semaphore.c | 66 struct radeon_ring *ring = &rdev->ring[ridx]; local in function:radeon_semaphore_emit_signal 70 if (radeon_semaphore_ring_emit(rdev, ridx, ring, semaphore, false)) { 74 ring->last_semaphore_signal_addr = semaphore->gpu_addr; 83 struct radeon_ring *ring = &rdev->ring[ridx]; local in function:radeon_semaphore_emit_wait 87 if (radeon_semaphore_ring_emit(rdev, ridx, ring, semaphore, true)) { 91 ring->last_semaphore_wait_addr = semaphore->gpu_addr;
|
radeon_semaphore.c | 66 struct radeon_ring *ring = &rdev->ring[ridx]; local in function:radeon_semaphore_emit_signal 70 if (radeon_semaphore_ring_emit(rdev, ridx, ring, semaphore, false)) { 74 ring->last_semaphore_signal_addr = semaphore->gpu_addr; 83 struct radeon_ring *ring = &rdev->ring[ridx]; local in function:radeon_semaphore_emit_wait 87 if (radeon_semaphore_ring_emit(rdev, ridx, ring, semaphore, true)) { 91 ring->last_semaphore_wait_addr = semaphore->gpu_addr;
|
radeon_rv770_dma.c | 55 struct radeon_ring *ring = &rdev->ring[ring_index]; local in function:rv770_copy_dma 64 r = radeon_ring_lock(rdev, ring, num_loops * 5 + 8); 72 radeon_sync_rings(rdev, &sync, ring->idx); 79 radeon_ring_write(ring, DMA_PACKET(DMA_PACKET_COPY, 0, 0, cur_size_in_dw)); 80 radeon_ring_write(ring, dst_offset & 0xfffffffc); 81 radeon_ring_write(ring, src_offset & 0xfffffffc); 82 radeon_ring_write(ring, upper_32_bits(dst_offset) & 0xff); 83 radeon_ring_write(ring, upper_32_bits(src_offset) & 0xff); 88 r = radeon_fence_emit(rdev, &fence, ring->idx) [all...] |
radeon_uvd_v2_2.c | 42 * Write a fence and a trap command to the ring. 47 struct radeon_ring *ring = &rdev->ring[fence->ring]; local in function:uvd_v2_2_fence_emit 48 uint64_t addr = rdev->fence_drv[fence->ring].gpu_addr; 50 radeon_ring_write(ring, PACKET0(UVD_CONTEXT_ID, 0)); 51 radeon_ring_write(ring, fence->seq); 52 radeon_ring_write(ring, PACKET0(UVD_GPCOM_VCPU_DATA0, 0)); 53 radeon_ring_write(ring, lower_32_bits(addr)); 54 radeon_ring_write(ring, PACKET0(UVD_GPCOM_VCPU_DATA1, 0)) [all...] |
radeon_rv770_dma.c | 55 struct radeon_ring *ring = &rdev->ring[ring_index]; local in function:rv770_copy_dma 64 r = radeon_ring_lock(rdev, ring, num_loops * 5 + 8); 72 radeon_sync_rings(rdev, &sync, ring->idx); 79 radeon_ring_write(ring, DMA_PACKET(DMA_PACKET_COPY, 0, 0, cur_size_in_dw)); 80 radeon_ring_write(ring, dst_offset & 0xfffffffc); 81 radeon_ring_write(ring, src_offset & 0xfffffffc); 82 radeon_ring_write(ring, upper_32_bits(dst_offset) & 0xff); 83 radeon_ring_write(ring, upper_32_bits(src_offset) & 0xff); 88 r = radeon_fence_emit(rdev, &fence, ring->idx) [all...] |
radeon_uvd_v2_2.c | 42 * Write a fence and a trap command to the ring. 47 struct radeon_ring *ring = &rdev->ring[fence->ring]; local in function:uvd_v2_2_fence_emit 48 uint64_t addr = rdev->fence_drv[fence->ring].gpu_addr; 50 radeon_ring_write(ring, PACKET0(UVD_CONTEXT_ID, 0)); 51 radeon_ring_write(ring, fence->seq); 52 radeon_ring_write(ring, PACKET0(UVD_GPCOM_VCPU_DATA0, 0)); 53 radeon_ring_write(ring, lower_32_bits(addr)); 54 radeon_ring_write(ring, PACKET0(UVD_GPCOM_VCPU_DATA1, 0)) [all...] |
radeon_evergreen_dma.c | 37 * evergreen_dma_fence_ring_emit - emit a fence on the DMA ring 42 * Add a DMA fence packet to the ring to write 49 struct radeon_ring *ring = &rdev->ring[fence->ring]; local in function:evergreen_dma_fence_ring_emit 50 u64 addr = rdev->fence_drv[fence->ring].gpu_addr; 52 radeon_ring_write(ring, DMA_PACKET(DMA_PACKET_FENCE, 0, 0)); 53 radeon_ring_write(ring, addr & 0xfffffffc); 54 radeon_ring_write(ring, (upper_32_bits(addr) & 0xff)); 55 radeon_ring_write(ring, fence->seq) 75 struct radeon_ring *ring = &rdev->ring[ib->ring]; local in function:evergreen_dma_ring_ib_execute 121 struct radeon_ring *ring = &rdev->ring[ring_index]; local in function:evergreen_copy_dma [all...] |
radeon_ib.c | 44 * command ring and the hw will fetch the commands from the IB 47 * put in IBs for execution by the requested ring. 55 * @ring: ring index the IB is associated with 63 int radeon_ib_get(struct radeon_device *rdev, int ring, 77 ib->ring = ring; 110 * radeon_ib_schedule - schedule an IB (Indirect Buffer) on the ring 117 * Schedule an IB on the associated ring (all asics). 120 * On SI, there are two parallel engines fed from the primary ring, 133 struct radeon_ring *ring = &rdev->ring[ib->ring]; local in function:radeon_ib_schedule 271 struct radeon_ring *ring = &rdev->ring[i]; local in function:radeon_ib_ring_tests [all...] |
radeon_evergreen_dma.c | 37 * evergreen_dma_fence_ring_emit - emit a fence on the DMA ring 42 * Add a DMA fence packet to the ring to write 49 struct radeon_ring *ring = &rdev->ring[fence->ring]; local in function:evergreen_dma_fence_ring_emit 50 u64 addr = rdev->fence_drv[fence->ring].gpu_addr; 52 radeon_ring_write(ring, DMA_PACKET(DMA_PACKET_FENCE, 0, 0)); 53 radeon_ring_write(ring, addr & 0xfffffffc); 54 radeon_ring_write(ring, (upper_32_bits(addr) & 0xff)); 55 radeon_ring_write(ring, fence->seq) 75 struct radeon_ring *ring = &rdev->ring[ib->ring]; local in function:evergreen_dma_ring_ib_execute 121 struct radeon_ring *ring = &rdev->ring[ring_index]; local in function:evergreen_copy_dma [all...] |
radeon_ib.c | 44 * command ring and the hw will fetch the commands from the IB 47 * put in IBs for execution by the requested ring. 55 * @ring: ring index the IB is associated with 63 int radeon_ib_get(struct radeon_device *rdev, int ring, 77 ib->ring = ring; 110 * radeon_ib_schedule - schedule an IB (Indirect Buffer) on the ring 117 * Schedule an IB on the associated ring (all asics). 120 * On SI, there are two parallel engines fed from the primary ring, 133 struct radeon_ring *ring = &rdev->ring[ib->ring]; local in function:radeon_ib_schedule 271 struct radeon_ring *ring = &rdev->ring[i]; local in function:radeon_ib_ring_tests [all...] |
radeon_si_dma.c | 41 * @ring: radeon_ring structure holding ring information 46 bool si_dma_is_lockup(struct radeon_device *rdev, struct radeon_ring *ring) 51 if (ring->idx == R600_RING_TYPE_DMA_INDEX) 57 radeon_ring_lockup_update(rdev, ring); 60 return radeon_ring_test_lockup(rdev, ring); 192 void si_dma_vm_flush(struct radeon_device *rdev, struct radeon_ring *ring, 196 radeon_ring_write(ring, DMA_PACKET(DMA_PACKET_SRBM_WRITE, 0, 0, 0, 0)); 198 radeon_ring_write(ring, (0xf << 16) | ((VM_CONTEXT0_PAGE_TABLE_BASE_ADDR + (vm_id << 2)) >> 2)); 200 radeon_ring_write(ring, (0xf << 16) | ((VM_CONTEXT8_PAGE_TABLE_BASE_ADDR + ((vm_id - 8) << 2)) >> 2)) 244 struct radeon_ring *ring = &rdev->ring[ring_index]; local in function:si_copy_dma [all...] |
radeon_si_dma.c | 41 * @ring: radeon_ring structure holding ring information 46 bool si_dma_is_lockup(struct radeon_device *rdev, struct radeon_ring *ring) 51 if (ring->idx == R600_RING_TYPE_DMA_INDEX) 57 radeon_ring_lockup_update(rdev, ring); 60 return radeon_ring_test_lockup(rdev, ring); 192 void si_dma_vm_flush(struct radeon_device *rdev, struct radeon_ring *ring, 196 radeon_ring_write(ring, DMA_PACKET(DMA_PACKET_SRBM_WRITE, 0, 0, 0, 0)); 198 radeon_ring_write(ring, (0xf << 16) | ((VM_CONTEXT0_PAGE_TABLE_BASE_ADDR + (vm_id << 2)) >> 2)); 200 radeon_ring_write(ring, (0xf << 16) | ((VM_CONTEXT8_PAGE_TABLE_BASE_ADDR + ((vm_id - 8) << 2)) >> 2)) 244 struct radeon_ring *ring = &rdev->ring[ring_index]; local in function:si_copy_dma [all...] |
radeon_r200.c | 94 struct radeon_ring *ring = &rdev->ring[RADEON_RING_TYPE_GFX_INDEX]; local in function:r200_copy_dma 104 r = radeon_ring_lock(rdev, ring, num_loops * 4 + 64); 110 radeon_ring_write(ring, PACKET0(RADEON_WAIT_UNTIL, 0)); 111 radeon_ring_write(ring, (1 << 16)); 118 radeon_ring_write(ring, PACKET0(0x720, 2)); 119 radeon_ring_write(ring, src_offset); 120 radeon_ring_write(ring, dst_offset); 121 radeon_ring_write(ring, cur_size | (1 << 31) | (1 << 30)); 125 radeon_ring_write(ring, PACKET0(RADEON_WAIT_UNTIL, 0)) [all...] |
radeon_r420.c | 219 struct radeon_ring *ring = &rdev->ring[RADEON_RING_TYPE_GFX_INDEX]; local in function:r420_cp_errata_init 228 r = radeon_ring_lock(rdev, ring, 8); 230 radeon_ring_write(ring, PACKET0(R300_CP_RESYNC_ADDR, 1)); 231 radeon_ring_write(ring, rdev->config.r300.resync_scratch); 232 radeon_ring_write(ring, 0xDEADBEEF); 233 radeon_ring_unlock_commit(rdev, ring, false); 239 struct radeon_ring *ring = &rdev->ring[RADEON_RING_TYPE_GFX_INDEX]; local in function:r420_cp_errata_fini 244 r = radeon_ring_lock(rdev, ring, 8) [all...] |
/src/sys/external/bsd/drm2/dist/drm/amd/amdgpu/ |
amdgpu_test.c | 39 struct amdgpu_ring *ring = adev->mman.buffer_funcs_ring; local in function:amdgpu_do_test_moves 50 * (Total GTT - IB pool - writeback page - ring buffers) / test size 131 r = amdgpu_copy_buffer(ring, gart_addr, vram_addr, 177 r = amdgpu_copy_buffer(ring, vram_addr, gart_addr,
|
amdgpu_test.c | 39 struct amdgpu_ring *ring = adev->mman.buffer_funcs_ring; local in function:amdgpu_do_test_moves 50 * (Total GTT - IB pool - writeback page - ring buffers) / test size 131 r = amdgpu_copy_buffer(ring, gart_addr, vram_addr, 177 r = amdgpu_copy_buffer(ring, vram_addr, gart_addr,
|
amdgpu_benchmark.c | 46 struct amdgpu_ring *ring = adev->mman.buffer_funcs_ring; local in function:amdgpu_benchmark_do_move 47 r = amdgpu_copy_buffer(ring, saddr, daddr, size, NULL, &fence,
|
amdgpu_ih.h | 36 * R6xx+ IH ring 50 volatile uint32_t *ring; member in struct:amdgpu_ih_ring 66 /* ring read/write ptr handling, called from interrupt context */
|
amdgpu_job.c | 38 struct amdgpu_ring *ring = to_amdgpu_ring(s_job->sched); local in function:amdgpu_job_timedout 44 if (amdgpu_ring_soft_recovery(ring, job->vmid, s_job->s_fence->parent)) { 45 DRM_ERROR("ring %s timeout, but soft recovered\n", 50 amdgpu_vm_get_task_info(ring->adev, job->pasid, &ti); 51 DRM_ERROR("ring %s timeout, signaled seq=%u, emitted seq=%u\n", 52 job->base.sched->name, atomic_read(&ring->fence_drv.last_seq), 53 ring->fence_drv.sync_seq); 57 if (amdgpu_device_should_recover_gpu(ring->adev)) 58 amdgpu_device_gpu_recover(ring->adev, job); 60 drm_sched_suspend_timeout(&ring->sched) 112 struct amdgpu_ring *ring = to_amdgpu_ring(job->base.sched); local in function:amdgpu_job_free_resources 125 struct amdgpu_ring *ring = to_amdgpu_ring(s_job->sched); local in function:amdgpu_job_free_cb 151 struct amdgpu_ring *ring; local in function:amdgpu_job_submit 190 struct amdgpu_ring *ring = to_amdgpu_ring(s_entity->rq->sched); local in function:amdgpu_job_dependency 221 struct amdgpu_ring *ring = to_amdgpu_ring(sched_job->sched); local in function:amdgpu_job_run [all...] |
amdgpu_psp_v10_0.c | 145 struct psp_ring *ring; local in function:psp_v10_0_ring_init 148 ring = &psp->km_ring; 150 ring->ring_type = ring_type; 152 /* allocate 4k Page of Local Frame Buffer memory for ring */ 153 ring->ring_size = 0x1000; 154 ret = amdgpu_bo_create_kernel(adev, ring->ring_size, PAGE_SIZE, 157 &ring->ring_mem_mc_addr, 158 (void **)&ring->ring_mem); 160 ring->ring_size = 0; 172 struct psp_ring *ring = &psp->km_ring local in function:psp_v10_0_ring_create 224 struct psp_ring *ring = &psp->km_ring; local in function:psp_v10_0_ring_destroy [all...] |
amdgpu_vce.h | 50 struct amdgpu_ring ring[AMDGPU_MAX_VCE_RINGS]; member in struct:amdgpu_vce 66 void amdgpu_vce_ring_emit_ib(struct amdgpu_ring *ring, struct amdgpu_job *job, 68 void amdgpu_vce_ring_emit_fence(struct amdgpu_ring *ring, u64 addr, u64 seq, 70 int amdgpu_vce_ring_test_ring(struct amdgpu_ring *ring); 71 int amdgpu_vce_ring_test_ib(struct amdgpu_ring *ring, long timeout); 72 void amdgpu_vce_ring_begin_use(struct amdgpu_ring *ring); 73 void amdgpu_vce_ring_end_use(struct amdgpu_ring *ring); 74 unsigned amdgpu_vce_ring_get_emit_ib_size(struct amdgpu_ring *ring); 75 unsigned amdgpu_vce_ring_get_dma_frame_size(struct amdgpu_ring *ring);
|
amdgpu_virt.c | 58 struct amdgpu_ring *ring = &kiq->ring; local in function:amdgpu_virt_kiq_reg_write_reg_wait 64 amdgpu_ring_alloc(ring, 32); 65 amdgpu_ring_emit_reg_write_reg_wait(ring, reg0, reg1, 67 amdgpu_fence_emit_polling(ring, &seq); 68 amdgpu_ring_commit(ring); 71 r = amdgpu_fence_wait_polling(ring, seq, MAX_KIQ_REG_WAIT); 81 r = amdgpu_fence_wait_polling(ring, seq, MAX_KIQ_REG_WAIT);
|
amdgpu_vm_sdma.c | 106 struct amdgpu_ring *ring; local in function:amdgpu_vm_sdma_commit 110 ring = container_of(entity->rq->sched, struct amdgpu_ring, sched); 113 amdgpu_ring_pad_ib(ring, ib);
|
/src/games/rogue/ |
ring.c | 1 /* $NetBSD: ring.c,v 1.9 2008/01/14 03:50:02 dholland Exp $ */ 38 static char sccsid[] = "@(#)ring.c 8.1 (Berkeley) 5/31/93"; 40 __RCSID("$NetBSD: ring.c,v 1.9 2008/01/14 03:50:02 dholland Exp $"); 45 * ring.c 59 static const char no_ring[] = "there's no ring on that hand"; 78 object *ring; local in function:put_on_ring 84 if ((ch = pack_letter("put on what?", RING)) == CANCEL) { 87 if (!(ring = get_letter_object(ch))) { 91 if (!(ring->what_is & RING)) { 152 object *ring; local in function:remove_ring 279 object *ring; local in function:ring_stats [all...] |
/src/sys/external/bsd/drm2/dist/drm/i915/gt/ |
intel_ring.c | 19 unsigned int intel_ring_update_space(struct intel_ring *ring) 23 space = __intel_ring_space(ring->head, ring->emit, ring->size); 25 ring->space = space; 29 int intel_ring_pin(struct intel_ring *ring) 31 struct i915_vma *vma = ring->vma; 36 if (atomic_fetch_inc(&ring->pin_count)) 41 /* Ring wraparound at offset 0 sometimes hangs. No idea why. */ 66 intel_ring_reset(ring, ring->emit) 141 struct intel_ring *ring; local in function:intel_engine_create_ring 178 struct intel_ring *ring = container_of(ref, typeof(*ring), ref); local in function:intel_ring_free 224 struct intel_ring *ring = rq->ring; local in function:intel_ring_begin [all...] |