HomeSort by: relevance | last modified time | path
    Searched refs:ring_id (Results 1 - 25 of 39) sorted by relevancy

1 2

  /src/sys/external/bsd/drm2/dist/drm/i915/gvt/
mmio_context.h 42 int ring_id; member in struct:engine_mmio
50 struct intel_vgpu *next, int ring_id);
execlist.c 47 #define execlist_ring_mmio(gvt, ring_id, offset) \
48 (gvt->dev_priv->engine[ring_id]->mmio_base + (offset))
62 static int ring_id_to_context_switch_event(unsigned int ring_id)
64 if (WARN_ON(ring_id >= ARRAY_SIZE(context_switch_events)))
67 return context_switch_events[ring_id];
101 int ring_id = execlist->ring_id; local in function:emulate_execlist_status
103 ring_id, _EL_OFFSET_STATUS);
136 int ring_id = execlist->ring_id; local in function:emulate_csb_update
269 int ring_id = execlist->ring_id; local in function:get_next_execlist_slot
387 int ring_id = workload->ring_id; local in function:prepare_execlist_workload
407 int ring_id = workload->ring_id; local in function:complete_execlist_workload
    [all...]
scheduler.c 106 if (workload->ring_id != RCS0)
136 int ring_id = workload->ring_id; local in function:populate_shadow_context
162 if (ring_id == RCS0) {
183 gvt_dbg_sched("ring id %d workload lrca %x", ring_id,
186 context_page_num = gvt->dev_priv->engine[ring_id]->context_size;
190 if (IS_BROADWELL(gvt->dev_priv) && ring_id == RCS0)
218 static void save_ring_hw_state(struct intel_vgpu *vgpu, int ring_id)
221 u32 ring_base = dev_priv->engine[ring_id]->mmio_base;
239 enum intel_engine_id ring_id = req->engine->id local in function:shadow_context_status_change
685 int ring_id = workload->ring_id; local in function:dispatch_workload
976 int ring_id; member in struct:workload_thread_param
983 int ring_id = p->ring_id; local in function:workload_thread
    [all...]
mmio_context.c 171 int ring_id, i; local in function:load_render_mocs
177 for (ring_id = 0; ring_id < cnt; ring_id++) {
178 if (!HAS_ENGINE(dev_priv, ring_id))
180 offset.reg = regs[ring_id];
182 gen9_render_mocs.control_table[ring_id][i] =
205 int ring_id = req->engine->id; local in function:restore_context_mmio_for_inhibit
206 int count = gvt->engine_mmio_list.ctx_mmio_count[ring_id];
222 if (mmio->ring_id != ring_id |
    [all...]
scheduler.h 84 int ring_id; member in struct:intel_vgpu_workload
134 #define workload_q_head(vgpu, ring_id) \
135 (&(vgpu->submission.workload_q_head[ring_id]))
160 intel_vgpu_create_workload(struct intel_vgpu *vgpu, int ring_id,
trace.h 118 TP_PROTO(int id, char *type, int ring_id, int root_entry_type,
121 TP_ARGS(id, type, ring_id, root_entry_type, gma, gpa),
130 id, type, ring_id, root_entry_type, gma, gpa);
231 TP_PROTO(u8 vgpu_id, u8 ring_id, u32 ip_gma, u32 *cmd_va,
235 TP_ARGS(vgpu_id, ring_id, ip_gma, cmd_va, cmd_len, buf_type,
240 __field(u8, ring_id)
252 __entry->ring_id = ring_id;
265 __entry->ring_id,
interrupt.h 234 int gvt_ring_id_to_pipe_control_notify_event(int ring_id);
235 int gvt_ring_id_to_mi_flush_dw_event(int ring_id);
236 int gvt_ring_id_to_mi_user_interrupt_event(int ring_id);
execlist.h 175 int ring_id; member in struct:intel_vgpu_execlist
184 int intel_vgpu_submit_execlist(struct intel_vgpu *vgpu, int ring_id);
sched_policy.c 452 int ring_id; local in function:intel_vgpu_stop_schedule
475 for (ring_id = 0; ring_id < I915_NUM_ENGINES; ring_id++) {
476 if (scheduler->engine_owner[ring_id] == vgpu) {
477 intel_gvt_switch_mmio(vgpu, NULL, ring_id);
478 scheduler->engine_owner[ring_id] = NULL;
cmd_parser.c 470 int ring_id; member in struct:parser_exec_state
643 static inline u32 get_opcode(u32 cmd, int ring_id)
647 d_info = ring_decode_info[ring_id][CMD_TYPE(cmd)];
655 unsigned int opcode, int ring_id)
660 if (opcode == e->info->opcode && e->info->rings & BIT(ring_id))
667 u32 cmd, int ring_id)
671 opcode = get_opcode(cmd, ring_id);
675 return find_cmd_entry(gvt, opcode, ring_id);
683 static inline void print_opcode(u32 cmd, int ring_id)
688 d_info = ring_decode_info[ring_id][CMD_TYPE(cmd)]
2863 int ring_id = workload->ring_id; local in function:shadow_workload_ring_buffer
    [all...]
handlers.c 522 int ring_id = intel_gvt_render_mmio_to_ring_id(vgpu->gvt, offset); local in function:force_nonpriv_write
527 if ((bytes != 4) || ((offset & (bytes - 1)) != 0) || ring_id < 0) {
529 vgpu->id, ring_id, offset, bytes);
533 ring_base = dev_priv->engine[ring_id]->mmio_base;
1486 int ring_id = intel_gvt_render_mmio_to_ring_id(vgpu->gvt, offset); local in function:hws_pga_write
1498 if (unlikely(ring_id < 0 || ring_id >= I915_NUM_ENGINES)) {
1503 vgpu->hws_pga[ring_id] = value;
1663 int ring_id; local in function:mmio_read_from_hw
1666 ring_id = intel_gvt_render_mmio_to_ring_id(gvt, offset)
1690 int ring_id = intel_gvt_render_mmio_to_ring_id(vgpu->gvt, offset); local in function:elsp_mmio_write
1717 int ring_id = intel_gvt_render_mmio_to_ring_id(vgpu->gvt, offset); local in function:ring_mode_mmio_write
    [all...]
  /src/sys/external/bsd/drm2/dist/drm/amd/amdkfd/
cik_int.h 33 uint32_t ring_id; member in struct:cik_ih_ring_entry
cik_event_interrupt.c 60 tmp_ihre->ring_id &= 0x000000ff;
61 tmp_ihre->ring_id |= vmid << 8;
62 tmp_ihre->ring_id |= pasid << 16;
70 vmid = (ihre->ring_id & 0x0000ff00) >> 8;
76 pasid = (ihre->ring_id & 0xffff0000) >> 16;
97 unsigned int vmid = (ihre->ring_id & 0x0000ff00) >> 8;
98 unsigned int pasid = (ihre->ring_id & 0xffff0000) >> 16;
kfd_int_process_v9.c 115 uint16_t ring_id = SOC15_RING_ID_FROM_IH_ENTRY(ih_ring_entry); local in function:event_interrupt_wq_v9
121 info.prot_valid = ring_id & 0x08;
122 info.prot_read = ring_id & 0x10;
123 info.prot_write = ring_id & 0x20;
  /src/sys/external/bsd/drm2/dist/drm/amd/amdgpu/
amdgpu_irq.h 51 unsigned ring_id; member in struct:amdgpu_iv_entry
amdgpu_trace.h 85 __field(unsigned, ring_id)
97 __entry->ring_id = iv->ring_id;
111 __entry->ring_id, __entry->vmid,
amdgpu_gfx_v10_0.c 1239 static int gfx_v10_0_gfx_ring_init(struct amdgpu_device *adev, int ring_id,
1246 ring = &adev->gfx.gfx_ring[ring_id];
1255 if (!ring_id)
1269 static int gfx_v10_0_compute_ring_init(struct amdgpu_device *adev, int ring_id,
1274 struct amdgpu_ring *ring = &adev->gfx.compute_ring[ring_id];
1276 ring = &adev->gfx.compute_ring[ring_id];
1285 ring->doorbell_index = (adev->doorbell_index.mec_ring0 + ring_id) << 1;
1287 + (ring_id * GFX10_MEC_HPD_SIZE);
1305 int i, j, k, r, ring_id = 0; local in function:gfx_v10_0_sw_init
1383 r = gfx_v10_0_gfx_ring_init(adev, ring_id,
    [all...]
amdgpu_cik_ih.c 262 entry->ring_id = dw[2] & 0xff;
amdgpu_sdma_v2_4.c 1062 instance_id = (entry->ring_id & 0x3) >> 0;
1063 queue_id = (entry->ring_id & 0xc) >> 2;
1103 instance_id = (entry->ring_id & 0x3) >> 0;
1104 queue_id = (entry->ring_id & 0xc) >> 2;
amdgpu_si_ih.c 144 entry->ring_id = dw[2] & 0xff;
amdgpu_vega10_ih.c 461 entry->ring_id = (dw[0] >> 16) & 0xff;
550 switch (entry->ring_id) {
amdgpu_cz_ih.c 241 entry->ring_id = dw[2] & 0xff;
amdgpu_iceland_ih.c 241 entry->ring_id = dw[2] & 0xff;
  /src/sys/dev/pci/ixgbe/
ixgbe_netmap.c 211 struct tx_ring *txr = &sc->tx_rings[kring->ring_id];
349 nic_i = IXGBE_READ_REG(&sc->hw, IXGBE_TDH(kring->ring_id));
395 struct rx_ring *rxr = &sc->rx_rings[kring->ring_id];
  /src/sys/dev/pci/
xmm7360.c 699 static void xmm7360_td_ring_activate(struct xmm_dev *xmm, u8 ring_id)
701 struct td_ring *ring = &xmm->td_ring[ring_id];
704 xmm->cp->s_rptr[ring_id] = xmm->cp->s_wptr[ring_id] = 0;
706 ret = xmm7360_cmd_ring_execute(xmm, CMD_RING_OPEN, ring_id, ring->depth, ring->tds_phys, 0x60);
710 static void xmm7360_td_ring_create(struct xmm_dev *xmm, u8 ring_id, u8 depth, u16 page_size)
712 struct td_ring *ring = &xmm->td_ring[ring_id];
732 xmm7360_td_ring_activate(xmm, ring_id);
735 static void xmm7360_td_ring_deactivate(struct xmm_dev *xmm, u8 ring_id)
737 xmm7360_cmd_ring_execute(xmm, CMD_RING_CLOSE, ring_id, 0, 0, 0)
    [all...]

Completed in 62 milliseconds

1 2