/src/sys/external/bsd/drm2/dist/drm/radeon/ |
radeon_semaphore.c | 66 struct radeon_ring *ring = &rdev->ring[ridx]; local in function:radeon_semaphore_emit_signal 70 if (radeon_semaphore_ring_emit(rdev, ridx, ring, semaphore, false)) { 74 ring->last_semaphore_signal_addr = semaphore->gpu_addr; 83 struct radeon_ring *ring = &rdev->ring[ridx]; local in function:radeon_semaphore_emit_wait 87 if (radeon_semaphore_ring_emit(rdev, ridx, ring, semaphore, true)) { 91 ring->last_semaphore_wait_addr = semaphore->gpu_addr;
|
radeon_rv770_dma.c | 55 struct radeon_ring *ring = &rdev->ring[ring_index]; local in function:rv770_copy_dma 64 r = radeon_ring_lock(rdev, ring, num_loops * 5 + 8); 72 radeon_sync_rings(rdev, &sync, ring->idx); 79 radeon_ring_write(ring, DMA_PACKET(DMA_PACKET_COPY, 0, 0, cur_size_in_dw)); 80 radeon_ring_write(ring, dst_offset & 0xfffffffc); 81 radeon_ring_write(ring, src_offset & 0xfffffffc); 82 radeon_ring_write(ring, upper_32_bits(dst_offset) & 0xff); 83 radeon_ring_write(ring, upper_32_bits(src_offset) & 0xff); 88 r = radeon_fence_emit(rdev, &fence, ring->idx) [all...] |
radeon_uvd_v2_2.c | 42 * Write a fence and a trap command to the ring. 47 struct radeon_ring *ring = &rdev->ring[fence->ring]; local in function:uvd_v2_2_fence_emit 48 uint64_t addr = rdev->fence_drv[fence->ring].gpu_addr; 50 radeon_ring_write(ring, PACKET0(UVD_CONTEXT_ID, 0)); 51 radeon_ring_write(ring, fence->seq); 52 radeon_ring_write(ring, PACKET0(UVD_GPCOM_VCPU_DATA0, 0)); 53 radeon_ring_write(ring, lower_32_bits(addr)); 54 radeon_ring_write(ring, PACKET0(UVD_GPCOM_VCPU_DATA1, 0)) [all...] |
radeon_evergreen_dma.c | 37 * evergreen_dma_fence_ring_emit - emit a fence on the DMA ring 42 * Add a DMA fence packet to the ring to write 49 struct radeon_ring *ring = &rdev->ring[fence->ring]; local in function:evergreen_dma_fence_ring_emit 50 u64 addr = rdev->fence_drv[fence->ring].gpu_addr; 52 radeon_ring_write(ring, DMA_PACKET(DMA_PACKET_FENCE, 0, 0)); 53 radeon_ring_write(ring, addr & 0xfffffffc); 54 radeon_ring_write(ring, (upper_32_bits(addr) & 0xff)); 55 radeon_ring_write(ring, fence->seq) 75 struct radeon_ring *ring = &rdev->ring[ib->ring]; local in function:evergreen_dma_ring_ib_execute 121 struct radeon_ring *ring = &rdev->ring[ring_index]; local in function:evergreen_copy_dma [all...] |
radeon_ib.c | 44 * command ring and the hw will fetch the commands from the IB 47 * put in IBs for execution by the requested ring. 55 * @ring: ring index the IB is associated with 63 int radeon_ib_get(struct radeon_device *rdev, int ring, 77 ib->ring = ring; 110 * radeon_ib_schedule - schedule an IB (Indirect Buffer) on the ring 117 * Schedule an IB on the associated ring (all asics). 120 * On SI, there are two parallel engines fed from the primary ring, 133 struct radeon_ring *ring = &rdev->ring[ib->ring]; local in function:radeon_ib_schedule 271 struct radeon_ring *ring = &rdev->ring[i]; local in function:radeon_ib_ring_tests [all...] |
radeon_si_dma.c | 41 * @ring: radeon_ring structure holding ring information 46 bool si_dma_is_lockup(struct radeon_device *rdev, struct radeon_ring *ring) 51 if (ring->idx == R600_RING_TYPE_DMA_INDEX) 57 radeon_ring_lockup_update(rdev, ring); 60 return radeon_ring_test_lockup(rdev, ring); 192 void si_dma_vm_flush(struct radeon_device *rdev, struct radeon_ring *ring, 196 radeon_ring_write(ring, DMA_PACKET(DMA_PACKET_SRBM_WRITE, 0, 0, 0, 0)); 198 radeon_ring_write(ring, (0xf << 16) | ((VM_CONTEXT0_PAGE_TABLE_BASE_ADDR + (vm_id << 2)) >> 2)); 200 radeon_ring_write(ring, (0xf << 16) | ((VM_CONTEXT8_PAGE_TABLE_BASE_ADDR + ((vm_id - 8) << 2)) >> 2)) 244 struct radeon_ring *ring = &rdev->ring[ring_index]; local in function:si_copy_dma [all...] |
radeon_r200.c | 94 struct radeon_ring *ring = &rdev->ring[RADEON_RING_TYPE_GFX_INDEX]; local in function:r200_copy_dma 104 r = radeon_ring_lock(rdev, ring, num_loops * 4 + 64); 110 radeon_ring_write(ring, PACKET0(RADEON_WAIT_UNTIL, 0)); 111 radeon_ring_write(ring, (1 << 16)); 118 radeon_ring_write(ring, PACKET0(0x720, 2)); 119 radeon_ring_write(ring, src_offset); 120 radeon_ring_write(ring, dst_offset); 121 radeon_ring_write(ring, cur_size | (1 << 31) | (1 << 30)); 125 radeon_ring_write(ring, PACKET0(RADEON_WAIT_UNTIL, 0)) [all...] |
radeon_r420.c | 219 struct radeon_ring *ring = &rdev->ring[RADEON_RING_TYPE_GFX_INDEX]; local in function:r420_cp_errata_init 228 r = radeon_ring_lock(rdev, ring, 8); 230 radeon_ring_write(ring, PACKET0(R300_CP_RESYNC_ADDR, 1)); 231 radeon_ring_write(ring, rdev->config.r300.resync_scratch); 232 radeon_ring_write(ring, 0xDEADBEEF); 233 radeon_ring_unlock_commit(rdev, ring, false); 239 struct radeon_ring *ring = &rdev->ring[RADEON_RING_TYPE_GFX_INDEX]; local in function:r420_cp_errata_fini 244 r = radeon_ring_lock(rdev, ring, 8) [all...] |
radeon_uvd_v1_0.c | 40 * @ring: radeon_ring pointer 45 struct radeon_ring *ring) 54 * @ring: radeon_ring pointer 59 struct radeon_ring *ring) 68 * @ring: radeon_ring pointer 73 struct radeon_ring *ring) 75 WREG32(UVD_RBC_RB_WPTR, ring->wptr); 84 * Write a fence and a trap command to the ring. 89 struct radeon_ring *ring = &rdev->ring[fence->ring] local in function:uvd_v1_0_fence_emit 164 struct radeon_ring *ring = &rdev->ring[R600_RING_TYPE_UVD_INDEX]; local in function:uvd_v1_0_init 255 struct radeon_ring *ring = &rdev->ring[R600_RING_TYPE_UVD_INDEX]; local in function:uvd_v1_0_fini 270 struct radeon_ring *ring = &rdev->ring[R600_RING_TYPE_UVD_INDEX]; local in function:uvd_v1_0_start 489 struct radeon_ring *ring = &rdev->ring[ib->ring]; local in function:uvd_v1_0_ib_execute [all...] |
/src/sys/external/bsd/drm2/dist/drm/amd/amdgpu/ |
amdgpu_test.c | 39 struct amdgpu_ring *ring = adev->mman.buffer_funcs_ring; local in function:amdgpu_do_test_moves 50 * (Total GTT - IB pool - writeback page - ring buffers) / test size 131 r = amdgpu_copy_buffer(ring, gart_addr, vram_addr, 177 r = amdgpu_copy_buffer(ring, vram_addr, gart_addr,
|
amdgpu_benchmark.c | 46 struct amdgpu_ring *ring = adev->mman.buffer_funcs_ring; local in function:amdgpu_benchmark_do_move 47 r = amdgpu_copy_buffer(ring, saddr, daddr, size, NULL, &fence,
|
amdgpu_ih.h | 36 * R6xx+ IH ring 50 volatile uint32_t *ring; member in struct:amdgpu_ih_ring 66 /* ring read/write ptr handling, called from interrupt context */
|
amdgpu_job.c | 38 struct amdgpu_ring *ring = to_amdgpu_ring(s_job->sched); local in function:amdgpu_job_timedout 44 if (amdgpu_ring_soft_recovery(ring, job->vmid, s_job->s_fence->parent)) { 45 DRM_ERROR("ring %s timeout, but soft recovered\n", 50 amdgpu_vm_get_task_info(ring->adev, job->pasid, &ti); 51 DRM_ERROR("ring %s timeout, signaled seq=%u, emitted seq=%u\n", 52 job->base.sched->name, atomic_read(&ring->fence_drv.last_seq), 53 ring->fence_drv.sync_seq); 57 if (amdgpu_device_should_recover_gpu(ring->adev)) 58 amdgpu_device_gpu_recover(ring->adev, job); 60 drm_sched_suspend_timeout(&ring->sched) 112 struct amdgpu_ring *ring = to_amdgpu_ring(job->base.sched); local in function:amdgpu_job_free_resources 125 struct amdgpu_ring *ring = to_amdgpu_ring(s_job->sched); local in function:amdgpu_job_free_cb 151 struct amdgpu_ring *ring; local in function:amdgpu_job_submit 190 struct amdgpu_ring *ring = to_amdgpu_ring(s_entity->rq->sched); local in function:amdgpu_job_dependency 221 struct amdgpu_ring *ring = to_amdgpu_ring(sched_job->sched); local in function:amdgpu_job_run [all...] |
amdgpu_psp_v10_0.c | 145 struct psp_ring *ring; local in function:psp_v10_0_ring_init 148 ring = &psp->km_ring; 150 ring->ring_type = ring_type; 152 /* allocate 4k Page of Local Frame Buffer memory for ring */ 153 ring->ring_size = 0x1000; 154 ret = amdgpu_bo_create_kernel(adev, ring->ring_size, PAGE_SIZE, 157 &ring->ring_mem_mc_addr, 158 (void **)&ring->ring_mem); 160 ring->ring_size = 0; 172 struct psp_ring *ring = &psp->km_ring local in function:psp_v10_0_ring_create 224 struct psp_ring *ring = &psp->km_ring; local in function:psp_v10_0_ring_destroy [all...] |
amdgpu_vce.h | 50 struct amdgpu_ring ring[AMDGPU_MAX_VCE_RINGS]; member in struct:amdgpu_vce 66 void amdgpu_vce_ring_emit_ib(struct amdgpu_ring *ring, struct amdgpu_job *job, 68 void amdgpu_vce_ring_emit_fence(struct amdgpu_ring *ring, u64 addr, u64 seq, 70 int amdgpu_vce_ring_test_ring(struct amdgpu_ring *ring); 71 int amdgpu_vce_ring_test_ib(struct amdgpu_ring *ring, long timeout); 72 void amdgpu_vce_ring_begin_use(struct amdgpu_ring *ring); 73 void amdgpu_vce_ring_end_use(struct amdgpu_ring *ring); 74 unsigned amdgpu_vce_ring_get_emit_ib_size(struct amdgpu_ring *ring); 75 unsigned amdgpu_vce_ring_get_dma_frame_size(struct amdgpu_ring *ring);
|
amdgpu_virt.c | 58 struct amdgpu_ring *ring = &kiq->ring; local in function:amdgpu_virt_kiq_reg_write_reg_wait 64 amdgpu_ring_alloc(ring, 32); 65 amdgpu_ring_emit_reg_write_reg_wait(ring, reg0, reg1, 67 amdgpu_fence_emit_polling(ring, &seq); 68 amdgpu_ring_commit(ring); 71 r = amdgpu_fence_wait_polling(ring, seq, MAX_KIQ_REG_WAIT); 81 r = amdgpu_fence_wait_polling(ring, seq, MAX_KIQ_REG_WAIT);
|
amdgpu_vm_sdma.c | 106 struct amdgpu_ring *ring; local in function:amdgpu_vm_sdma_commit 110 ring = container_of(entity->rq->sched, struct amdgpu_ring, sched); 113 amdgpu_ring_pad_ib(ring, ib);
|
amdgpu_gmc.c | 301 /* If we don't have space left in the ring buffer return immediately */ 324 /* Add the fault to the ring */ 364 * Engine 0, 1, 4~16: amdgpu ring, 365 * subject to change when ring number changes 373 struct amdgpu_ring *ring; local in function:amdgpu_gmc_allocate_vm_inv_eng 381 ring = adev->rings[i]; 382 vmhub = ring->funcs->vmhub; 386 dev_err(adev->dev, "no VM inv eng for ring %s\n", 387 ring->name); 391 ring->vm_inv_eng = inv_eng - 1 [all...] |
amdgpu_ib.c | 51 * command ring and the hw will fetch the commands from the IB 54 * put in IBs for execution by the requested ring. 61 * @ring: ring index the IB is associated with 107 * amdgpu_ib_schedule - schedule an IB (Indirect Buffer) on the ring 114 * Schedule an IB on the associated ring (all asics). 117 * On SI, there are two parallel engines fed from the primary ring, 124 * a CONST_IB), it will be put on the ring prior to the DE IB. Prior 127 int amdgpu_ib_schedule(struct amdgpu_ring *ring, unsigned num_ibs, 131 struct amdgpu_device *adev = ring->adev 363 struct amdgpu_ring *ring = adev->rings[i]; local in function:amdgpu_ib_ring_tests [all...] |
amdgpu_jpeg_v1_0.c | 41 static void jpeg_v1_0_decode_ring_patch_wreg(struct amdgpu_ring *ring, uint32_t *ptr, uint32_t reg_offset, uint32_t val) 43 struct amdgpu_device *adev = ring->adev; 44 ring->ring[(*ptr)++] = PACKETJ(SOC15_REG_OFFSET(JPEG, 0, mmUVD_JRBC_EXTERNAL_REG_BASE), 0, 0, PACKETJ_TYPE0); 47 ring->ring[(*ptr)++] = 0; 48 ring->ring[(*ptr)++] = PACKETJ((reg_offset >> 2), 0, 0, PACKETJ_TYPE0); 50 ring->ring[(*ptr)++] = reg_offset 478 struct amdgpu_ring *ring; local in function:jpeg_v1_0_sw_init 521 struct amdgpu_ring *ring = &adev->jpeg.inst->ring_dec; local in function:jpeg_v1_0_start [all...] |
amdgpu_jpeg_v2_5.c | 58 * Set ring and irq function pointers 95 struct amdgpu_ring *ring; local in function:jpeg_v2_5_sw_init 122 ring = &adev->jpeg.inst[i].ring_dec; 123 ring->use_doorbell = true; 124 ring->doorbell_index = (adev->doorbell_index.vcn.vcn_ring0_1 << 1) + 1 + 8 * i; 125 snprintf(ring->name, sizeof(ring->name), "jpeg_dec_%d", i); 126 r = amdgpu_ring_init(adev, ring, 512, &adev->jpeg.inst[i].irq, 0); 167 struct amdgpu_ring *ring; local in function:jpeg_v2_5_hw_init 174 ring = &adev->jpeg.inst[i].ring_dec 198 struct amdgpu_ring *ring; local in function:jpeg_v2_5_hw_fini 310 struct amdgpu_ring *ring; local in function:jpeg_v2_5_start [all...] |
amdgpu_psp_v12_0.c | 182 /* Change IH ring for VMC */ 195 /* Change IH ring for UMC */ 212 struct psp_ring *ring; local in function:psp_v12_0_ring_init 217 ring = &psp->km_ring; 219 ring->ring_type = ring_type; 221 /* allocate 4k Page of Local Frame Buffer memory for ring */ 222 ring->ring_size = 0x1000; 223 ret = amdgpu_bo_create_kernel(adev, ring->ring_size, PAGE_SIZE, 226 &ring->ring_mem_mc_addr, 227 (void **)&ring->ring_mem) 248 struct psp_ring *ring = &psp->km_ring; local in function:psp_v12_0_ring_create 328 struct psp_ring *ring = &psp->km_ring; local in function:psp_v12_0_ring_destroy [all...] |
/src/games/rogue/ |
ring.c | 1 /* $NetBSD: ring.c,v 1.9 2008/01/14 03:50:02 dholland Exp $ */ 38 static char sccsid[] = "@(#)ring.c 8.1 (Berkeley) 5/31/93"; 40 __RCSID("$NetBSD: ring.c,v 1.9 2008/01/14 03:50:02 dholland Exp $"); 45 * ring.c 59 static const char no_ring[] = "there's no ring on that hand"; 78 object *ring; local in function:put_on_ring 84 if ((ch = pack_letter("put on what?", RING)) == CANCEL) { 87 if (!(ring = get_letter_object(ch))) { 91 if (!(ring->what_is & RING)) { 152 object *ring; local in function:remove_ring 279 object *ring; local in function:ring_stats [all...] |
/src/sys/external/bsd/drm2/dist/drm/i915/gt/ |
intel_ring.c | 19 unsigned int intel_ring_update_space(struct intel_ring *ring) 23 space = __intel_ring_space(ring->head, ring->emit, ring->size); 25 ring->space = space; 29 int intel_ring_pin(struct intel_ring *ring) 31 struct i915_vma *vma = ring->vma; 36 if (atomic_fetch_inc(&ring->pin_count)) 41 /* Ring wraparound at offset 0 sometimes hangs. No idea why. */ 66 intel_ring_reset(ring, ring->emit) 141 struct intel_ring *ring; local in function:intel_engine_create_ring 178 struct intel_ring *ring = container_of(ref, typeof(*ring), ref); local in function:intel_ring_free 224 struct intel_ring *ring = rq->ring; local in function:intel_ring_begin [all...] |
/src/sys/external/mit/xen-include-public/dist/xen/include/public/io/ |
tpmif.h | 63 struct tpmif_ring ring[TPMIF_TX_RING_SIZE]; member in struct:tpmif_tx_interface
|