Lines Matching refs:measure

67    /* initialise list of measure structures that await rendering */
100 cmd_buffer->measure = NULL;
110 struct anv_measure_batch * measure =
115 memset(measure, 0, batch_bytes);
117 anv_device_alloc_bo(device, "measure data",
121 (struct anv_bo**)&measure->bo);
122 measure->base.timestamps = measure->bo->map;
125 cmd_buffer->measure = measure;
135 struct anv_measure_batch *measure = cmd_buffer->measure;
144 if (measure->base.frame == 0)
145 measure->base.frame = device_frame;
149 if (!measure->base.framebuffer &&
152 measure->base.framebuffer = framebuffer;
156 framebuffer == measure->base.framebuffer ||
159 unsigned index = measure->base.index++;
161 (*device->cmd_emit_timestamp)(batch, measure->bo, index * sizeof(uint64_t));
166 struct intel_measure_snapshot *snapshot = &(measure->base.snapshots[index]);
170 snapshot->event_count = measure->base.event_count;
192 struct anv_measure_batch *measure = cmd_buffer->measure;
195 unsigned index = measure->base.index++;
198 (*device->cmd_emit_timestamp)(batch, measure->bo, index * sizeof(uint64_t));
200 struct intel_measure_snapshot *snapshot = &(measure->base.snapshots[index]);
232 return intel_measure_state_changed(&cmd_buffer->measure->base,
243 struct anv_measure_batch *measure = cmd_buffer->measure;
246 if (measure == NULL)
256 ++measure->base.event_count;
257 if (measure->base.event_count == 1 ||
258 measure->base.event_count == config->event_interval + 1) {
261 if (measure->base.index % 2) {
263 anv_measure_end_snapshot(cmd_buffer, measure->base.event_count - 1);
265 measure->base.event_count = 1;
267 if (measure->base.index == config->batch_size) {
297 struct anv_measure_batch *measure = cmd_buffer->measure;
303 cmd_buffer->measure = NULL;
307 if (!measure) {
322 measure->base.index = 0;
323 measure->base.framebuffer = 0;
324 measure->base.frame = 0;
325 measure->base.event_count = 0;
326 list_inithead(&measure->base.link);
328 anv_device_release_bo(device, measure->bo);
330 anv_device_alloc_bo(device, "measure data",
334 (struct anv_bo**)&measure->bo);
335 measure->base.timestamps = measure->bo->map;
343 struct anv_measure_batch *measure = cmd_buffer->measure;
349 if (measure == NULL)
357 anv_device_release_bo(device, measure->bo);
358 vk_free(&cmd_buffer->pool->alloc, measure);
359 cmd_buffer->measure = NULL;
390 struct anv_measure_batch *measure = cmd_buffer->measure;
395 if (measure == NULL)
398 if (measure->base.index == 0)
404 measure->base.batch_count = p_atomic_inc_return(&cmd_buffer_count);
406 if (measure->base.index %2 == 1) {
407 anv_measure_end_snapshot(cmd_buffer, measure->base.event_count);
408 measure->base.event_count = 0;
413 list_addtail(&measure->base.link, &measure_device->queued_snapshots);
441 struct anv_measure_batch *measure = cmd_buffer->measure;
445 if (measure == NULL)
447 if (measure->base.index % 2 == 0)
450 anv_measure_end_snapshot(cmd_buffer, measure->base.event_count);
451 measure->base.event_count = 0;
458 struct anv_measure_batch *measure = cmd_buffer->measure;
462 if (measure == NULL)
465 if (measure->base.framebuffer == (uintptr_t) cmd_buffer->state.framebuffer)
471 if (filtering && measure->base.index % 2 == 1) {
474 measure->base.event_count);
475 measure->base.event_count = 0;
478 measure->base.framebuffer = (uintptr_t) cmd_buffer->state.framebuffer;
486 struct anv_measure_batch *measure = primary->measure;
489 if (measure == NULL)
505 if (measure->base.index % 2 == 1)
506 anv_measure_end_snapshot(primary, measure->base.event_count);
508 struct intel_measure_snapshot *snapshot = &(measure->base.snapshots[measure->base.index]);
511 snapshot->secondary = &secondary->measure->base;