| /xsrc/external/mit/MesaLib.old/dist/src/intel/vulkan/tests/ |
| H A D | state_pool_no_free.c | 39 } jobs[NUM_THREADS]; variable in typeref:struct:job[] 71 jobs[i].pool = &state_pool; 72 jobs[i].id = i; 73 pthread_create(&jobs[i].thread, NULL, alloc_states, &jobs[i]); 77 pthread_join(jobs[i].thread, NULL); 92 if (thread_max < jobs[i].offsets[next[i]]) { 93 thread_max = jobs[i].offsets[next[i]]; 105 assert(jobs[max_thread_idx].offsets[next[max_thread_idx]] > highest); 107 highest = jobs[max_thread_id [all...] |
| H A D | state_pool_test_helper.h | 30 } jobs[NUM_THREADS]; variable in typeref:struct:job[] 64 jobs[i].pool = state_pool; 65 jobs[i].id = i; 66 pthread_create(&jobs[i].thread, NULL, alloc_states, &jobs[i]); 70 pthread_join(jobs[i].thread, NULL);
|
| H A D | block_pool_no_free.c | 40 } jobs[NUM_THREADS]; variable in typeref:struct:job[] 46 uint32_t job_id = job - jobs; 124 jobs[i].pool = &pool; 125 jobs[i].id = i; 126 pthread_create(&jobs[i].thread, NULL, alloc_blocks, &jobs[i]); 130 pthread_join(jobs[i].thread, NULL); 135 block_ptrs[i] = jobs[i].blocks; 140 block_ptrs[i] = jobs[i].back_blocks;
|
| /xsrc/external/mit/MesaLib/dist/src/intel/vulkan/tests/ |
| H A D | state_pool_no_free.c | 38 } jobs[NUM_THREADS]; variable in typeref:struct:job[] 71 jobs[i].pool = &state_pool; 72 jobs[i].id = i; 73 pthread_create(&jobs[i].thread, NULL, alloc_states, &jobs[i]); 77 pthread_join(jobs[i].thread, NULL); 92 if (thread_max < jobs[i].offsets[next[i]]) { 93 thread_max = jobs[i].offsets[next[i]]; 105 ASSERT(jobs[max_thread_idx].offsets[next[max_thread_idx]] > highest); 107 highest = jobs[max_thread_id [all...] |
| H A D | state_pool_test_helper.h | 30 } jobs[NUM_THREADS]; variable in typeref:struct:job[] 64 jobs[i].pool = state_pool; 65 jobs[i].id = i; 66 pthread_create(&jobs[i].thread, NULL, alloc_states, &jobs[i]); 70 pthread_join(jobs[i].thread, NULL);
|
| H A D | block_pool_no_free.c | 39 } jobs[NUM_THREADS]; variable in typeref:struct:job[] 45 uint32_t job_id = job - jobs; 124 jobs[i].pool = &pool; 125 jobs[i].id = i; 126 pthread_create(&jobs[i].thread, NULL, alloc_blocks, &jobs[i]); 130 pthread_join(jobs[i].thread, NULL); 135 block_ptrs[i] = jobs[i].blocks; 140 block_ptrs[i] = jobs[i].back_blocks;
|
| /xsrc/external/mit/MesaLib.old/dist/src/gallium/drivers/panfrost/ |
| H A D | pan_job.c | 56 _mesa_hash_table_remove_key(ctx->jobs, &job->key); 80 struct hash_entry *entry = _mesa_hash_table_search(ctx->jobs, &key); 92 _mesa_hash_table_insert(ctx->jobs, &job->key, job); 152 hash_table_foreach(panfrost->jobs, entry) { 179 ctx->jobs = _mesa_hash_table_create(NULL,
|
| H A D | pan_context.h | 113 /* Bound job and map of panfrost_job_key to jobs */ 115 struct hash_table *jobs; member in struct:panfrost_context 153 * and tiler jobs, linked to the fragment job at the end. See the
|
| /xsrc/external/mit/libdrm/dist/tests/exynos/ |
| H A D | exynos_fimg2d_event.c | 103 * We need to wait until all G2D jobs are finished, otherwise we 110 static void wait_all_jobs(struct g2d_job* jobs, unsigned num_jobs) argument 115 while (jobs[i].busy) 121 static struct g2d_job* free_job(struct g2d_job* jobs, unsigned num_jobs) argument 126 if (jobs[i].busy == 0) 127 return &jobs[i]; 136 struct g2d_job *jobs = calloc(num_jobs, sizeof(struct g2d_job)); local in function:g2d_work 142 jobs[i].id = i; 150 j = free_job(jobs, num_jobs); 189 wait_all_jobs(jobs, num_job [all...] |
| /xsrc/external/mit/MesaLib/dist/src/util/ |
| H A D | u_queue.c | 318 job = queue->jobs[queue->read_idx]; 319 memset(&queue->jobs[queue->read_idx], 0, sizeof(struct util_queue_job)); 337 /* signal remaining jobs if all threads are being terminated */ 342 if (queue->jobs[i].job) { 343 if (queue->jobs[i].fence) 344 util_queue_fence_signal(queue->jobs[i].fence); 345 queue->jobs[i].job = NULL; 471 queue->jobs = (struct util_queue_job*) 473 if (!queue->jobs) 500 if (queue->jobs) { 604 struct util_queue_job *jobs = local in function:util_queue_add_job [all...] |
| H A D | u_queue.h | 219 size_t total_jobs_size; /* memory use of all jobs in the queue */ 220 struct util_queue_job *jobs; member in struct:util_queue
|
| /xsrc/external/mit/MesaLib.old/dist/src/util/ |
| H A D | u_queue.c | 309 job = queue->jobs[queue->read_idx]; 310 memset(&queue->jobs[queue->read_idx], 0, sizeof(struct util_queue_job)); 325 /* signal remaining jobs if all threads are being terminated */ 330 if (queue->jobs[i].job) { 331 util_queue_fence_signal(queue->jobs[i].fence); 332 queue->jobs[i].job = NULL; 448 queue->jobs = (struct util_queue_job*) 450 if (!queue->jobs) 484 if (queue->jobs) { 488 free(queue->jobs); 568 struct util_queue_job *jobs = local in function:util_queue_add_job [all...] |
| H A D | u_queue.h | 215 struct util_queue_job *jobs; member in struct:util_queue
|
| /xsrc/external/mit/MesaLib/dist/docs/ci/ |
| H A D | LAVA.rst | 17 taking in jobs for the hardware that the LAVA lab contains. The 19 connects to the LAVA lab using a predefined token to submit jobs under 22 The LAVA instance manages scheduling those jobs to the boards present. 30 some boards booting using test jobs. Start with the stock QEMU 37 that device type and the Mesa jobs will be scheduled to any of them. 50 to restrict the jobs it takes or it will grab random jobs from tasks 81 Now it's time to define your test jobs in the driver-specific
|
| H A D | docker.rst | 19 available jobs, with no inbound networking requirements. Jobs can 20 have tags, so we can have DUT-specific jobs that only run on runners 24 some level, which spawns multiple jobs from one definition, and then
|
| H A D | index.rst | 58 The CI replays application traces with various drivers in two different jobs. The first 86 So the jobs are created in personal repositories, the name of the user's account needs 135 able to handle a whole pipeline's worth of jobs in less than 15 minutes 154 issue, please push through an MR disabling that farm's jobs by adding 155 '.' to the front of the jobs names until the maintainer can bring 173 Mesa how many jobs it should serve (``concurrent=``) and how many 174 cores those jobs should use (``FDO_CI_CONCURRENT=``) by editing these
|
| /xsrc/external/mit/MesaLib/dist/.gitlab-ci/lava/ |
| H A D | lava_job_submitter.py | 243 (finished, data) = _call_proxy(proxy.scheduler.jobs.logs, job_id, line_count) 259 # `proxy.scheduler.jobs.logs` does not block, even when there is no 267 show = _call_proxy(proxy.scheduler.jobs.show, job_id) 274 return _call_proxy(proxy.scheduler.jobs.validate, job_file, True) 279 return _call_proxy(proxy.scheduler.jobs.submit, job_file) 311 proxy.scheduler.jobs.cancel(job_id)
|
| /xsrc/external/mit/MesaLib/dist/src/panfrost/lib/ |
| H A D | pan_indirect_draw.c | 104 struct jobs_data jobs; member in struct:indirect_draw_shader_builder 127 * the DDK seems to put all min/max search jobs at the beginning of the job chain 129 * the draw patching jobs which have the suppress_prefetch flag set. Merging the 326 builder->jobs.first_vertex_sysval = get_input_field(b, first_vertex_sysval); 327 builder->jobs.base_vertex_sysval = get_input_field(b, base_vertex_sysval); 328 builder->jobs.base_instance_sysval = get_input_field(b, base_instance_sysval); 329 builder->jobs.vertex_job = get_input_field(b, vertex_job); 330 builder->jobs.tiler_job = get_input_field(b, tiler_job); 391 builder->jobs.vertex_job : builder->jobs [all...] |
| H A D | pan_blitter.h | 97 struct panfrost_ptr *jobs);
|
| /xsrc/external/mit/MesaLib.old/dist/src/gallium/drivers/v3d/ |
| H A D | v3d_job.c | 26 * Functions for submitting VC5 render jobs to the kernel. 48 _mesa_hash_table_remove_key(v3d->jobs, &job->key); 169 hash_table_foreach(v3d->jobs, entry) { 174 /* Reminder: v3d->jobs is safe to keep iterating even 239 struct hash_entry *entry = _mesa_hash_table_search(v3d->jobs, 244 /* Creating a new job. Make sure that any previous jobs reading or 284 _mesa_hash_table_insert(v3d->jobs, &job->key, job); 305 * been bound, so set them all to ~0 when switching between jobs. We 443 v3d->jobs = _mesa_hash_table_create(v3d,
|
| /xsrc/external/mit/MesaLib/dist/src/gallium/drivers/v3d/ |
| H A D | v3d_job.c | 26 * Functions for submitting V3D render jobs to the kernel. 48 _mesa_hash_table_remove_key(v3d->jobs, &job->key); 154 hash_table_foreach(v3d->jobs, entry) { 198 * graphics-compute dependency but nowadays all compute jobs 252 hash_table_foreach(v3d->jobs, entry) { 272 /* Reminder: v3d->jobs is safe to keep iterating even 305 struct hash_entry *entry = _mesa_hash_table_search(v3d->jobs, 310 /* Creating a new job. Make sure that any previous jobs reading or 362 _mesa_hash_table_insert(v3d->jobs, &job->key, job); 388 * been bound, so set them all to ~0 when switching between jobs [all...] |
| /xsrc/external/mit/MesaLib/dist/src/panfrost/vulkan/ |
| H A D | panvk_vX_cmd_buffer.c | 53 util_dynarray_append(&batch->jobs, void *, job_ptr.cpu); 137 /* Batch has no jobs but is needed for synchronization, let's add a 142 util_dynarray_append(&batch->jobs, void *, ptr.cpu); 165 util_dynarray_append(&batch->jobs, void *, preload_jobs[i].cpu); 196 util_dynarray_append(&batch->jobs, void *, writeval_job.cpu); 671 util_dynarray_append(&batch->jobs, void *, ptr.cpu); 672 draw->jobs.vertex = ptr; 685 util_dynarray_append(&batch->jobs, void *, ptr.cpu); 686 draw->jobs.tiler = ptr; 706 * the 3 (2 in Bifrost) jobs i [all...] |
| /xsrc/external/mit/MesaLib.old/dist/src/gallium/drivers/vc4/ |
| H A D | vc4_job.c | 26 * Functions for submitting VC4 render jobs to the kernel. 42 _mesa_hash_table_remove_key(vc4->jobs, &job->key); 118 hash_table_foreach(vc4->jobs, entry) { 172 struct hash_entry *entry = _mesa_hash_table_search(vc4->jobs, 177 /* Creating a new job. Make sure that any previous jobs reading or 220 _mesa_hash_table_insert(vc4->jobs, &job->key, job); 236 * been bound, so set them all to ~0 when switching between jobs. We 541 vc4->jobs = _mesa_hash_table_create(vc4,
|
| /xsrc/external/mit/MesaLib/dist/src/gallium/drivers/vc4/ |
| H A D | vc4_job.c | 26 * Functions for submitting VC4 render jobs to the kernel. 42 _mesa_hash_table_remove_key(vc4->jobs, &job->key); 118 hash_table_foreach(vc4->jobs, entry) { 172 struct hash_entry *entry = _mesa_hash_table_search(vc4->jobs, 177 /* Creating a new job. Make sure that any previous jobs reading or 220 _mesa_hash_table_insert(vc4->jobs, &job->key, job); 236 * been bound, so set them all to ~0 when switching between jobs. We 541 vc4->jobs = _mesa_hash_table_create(vc4,
|
| /xsrc/external/mit/MesaLib/dist/src/gallium/drivers/radeonsi/ci/ |
| H A D | radeonsi-run-tests.py | 58 "--jobs", 317 "--jobs", 318 str(args.jobs), 357 "--jobs", 358 str(args.jobs), 423 "--jobs", 424 str(args.jobs),
|