Lines Matching defs:upload
396 /* special upload buffer */
516 /* and reset the upload counter */
2205 DBG(("%s: using magical upload buffer\n", __FUNCTION__));
2209 DBG(("%s: discarding cached upload buffer\n", __FUNCTION__));
2671 DBG(("%s: using magical upload buffer\n", __FUNCTION__));
2675 DBG(("%s: discarding cached upload buffer\n", __FUNCTION__));
3451 DBG(("%s: discarding cached upload buffer\n", __FUNCTION__));
3453 assert(priv->gpu_damage == NULL || DAMAGE_IS_ALL(priv->gpu_damage)); /* magical upload buffer */
3482 DBG(("%s: reusing cached upload\n", __FUNCTION__));
3499 DBG(("%s: using CPU bo for upload to GPU\n", __FUNCTION__));
3539 DBG(("%s: using CPU bo for upload to GPU\n", __FUNCTION__));
3568 DBG(("%s: using CPU bo for upload to GPU, %d boxes\n", __FUNCTION__, n));
3680 DBG(("%s: cached upload proxy, discard and revert to GPU\n", __FUNCTION__));
3682 assert(priv->gpu_damage == NULL || DAMAGE_IS_ALL(priv->gpu_damage)); /* magical upload buffer */
4345 DBG(("%s: promoting upload proxy handle=%d to GPU\n", __FUNCTION__, proxy->handle));
4376 DBG(("%s: reusing cached upload\n", __FUNCTION__));
4417 DBG(("%s: using CPU bo for upload to GPU\n", __FUNCTION__));
4791 DBG(("%s: upload(%d, %d, %d, %d) x %d\n", __FUNCTION__, x, y, w, h, n));
4899 DBG(("%s: discarding cached upload buffer\n", __FUNCTION__));
4901 assert(priv->gpu_damage == NULL || DAMAGE_IS_ALL(priv->gpu_damage)); /* magical upload buffer */
4995 DBG(("%s: upload(%d, %d, %d, %d) x %d\n", __FUNCTION__, x, y, w, h, n));
5095 DBG(("%s: upload(%d, %d, %d, %d) x %d through a temporary map\n",
5231 DBG(("%s: segmented, unclipped large upload (%d bytes), marking WHOLE_HINT\n",
5245 DBG(("%s: upload(%d, %d, %d, %d) x %d boxes\n", __FUNCTION__, x, y, w, h, n));
5344 DBG(("%s: upload(%d, %d, %d, %d)\n", __FUNCTION__, x, y, w, h));
5363 struct kgem_bo *upload;
5376 upload = kgem_create_buffer(&sna->kgem,
5380 if (!upload)
5400 assert(dst <= (uint8_t *)ptr + kgem_bo_size(upload));
5426 kgem_add_reloc64(&sna->kgem, sna->kgem.nbatch + 6, upload,
5452 b[5] = kgem_add_reloc(&sna->kgem, sna->kgem.nbatch + 5, upload,
5463 kgem_bo_destroy(&sna->kgem, upload);
5510 DBG(("%s: upload(%d, %d, %d, %d)\n", __FUNCTION__, x, y, w, h));
5535 struct kgem_bo *upload;
5548 upload = kgem_create_buffer(&sna->kgem,
5552 if (!upload)
5570 assert(dst <= (uint8_t *)ptr + kgem_bo_size(upload));
5598 kgem_add_reloc64(&sna->kgem, sna->kgem.nbatch + 6, upload,
5628 b[5] = kgem_add_reloc(&sna->kgem, sna->kgem.nbatch + 5, upload,
5642 kgem_bo_destroy(&sna->kgem, upload);
6436 DBG(("%s: discarding cached upload buffer\n", __FUNCTION__));
6438 assert(priv->gpu_damage == NULL || DAMAGE_IS_ALL(priv->gpu_damage)); /* magical upload buffer */
6675 DBG(("%s: region overlaps GPU damage, upload and copy\n",
6773 DBG(("%s: upload through a temporary map\n",
6808 DBG(("%s: creating temporary source upload for non-copy alu [%d]\n",
6855 DBG(("%s: caching upload for src bo\n",
8478 struct kgem_bo *upload;
8491 upload = kgem_create_buffer(&sna->kgem,
8495 if (!upload)
8514 kgem_add_reloc64(&sna->kgem, sna->kgem.nbatch + 6, upload,
8533 b[5] = kgem_add_reloc(&sna->kgem, sna->kgem.nbatch + 5, upload,
8558 assert(dst <= (uint8_t *)ptr + kgem_bo_size(upload));
8565 kgem_bo_destroy(&sna->kgem, upload);
8623 struct kgem_bo *upload;
8642 upload = kgem_create_buffer(&sna->kgem,
8646 if (!upload)
8757 kgem_bcs_set_tiling(&sna->kgem, upload, arg->bo);
8773 kgem_add_reloc64(&sna->kgem, sna->kgem.nbatch + 6, upload,
8791 b[5] = kgem_add_reloc(&sna->kgem, sna->kgem.nbatch + 5, upload,
8802 kgem_bo_destroy(&sna->kgem, upload);
12319 struct kgem_bo *upload;
12324 upload = kgem_upload_source_image(&to_sna_from_pixmap(pixmap)->kgem,
12328 if (upload == NULL)
12332 DBG(("%s: adding upload cache to pixmap=%ld\n",
12334 assert(upload->proxy != NULL);
12335 kgem_proxy_bo_attach(upload, &priv->gpu_bo);
12338 return upload;
12816 struct kgem_bo *upload;
12870 upload = kgem_create_buffer(&sna->kgem, 8*bpp, KGEM_BUFFER_WRITE, &ptr);
12871 if (upload == NULL)
12874 upload->pitch = bpp; /* for sanity checks */
12925 upload, gc, n, rect,
12930 kgem_bo_destroy(&sna->kgem, upload);
13282 DBG(("%s: alu=%d, upload (%d, %d), (%d, %d), origin (%d, %d)\n",
13700 DBG(("%s: upload (%d, %d), (%d, %d), origin (%d, %d), clipped=%x\n", __FUNCTION__,
13807 struct kgem_bo *upload;
13820 upload = kgem_create_buffer(&sna->kgem,
13824 if (!upload)
13859 kgem_add_reloc64(&sna->kgem, sna->kgem.nbatch + 6, upload,
13877 b[5] = kgem_add_reloc(&sna->kgem, sna->kgem.nbatch + 5, upload,
13889 kgem_bo_destroy(&sna->kgem, upload);
13918 struct kgem_bo *upload;
14019 upload = kgem_create_buffer(&sna->kgem,
14023 if (!upload)
14058 kgem_add_reloc64(&sna->kgem, sna->kgem.nbatch + 6, upload,
14076 b[5] = kgem_add_reloc(&sna->kgem, sna->kgem.nbatch + 5, upload,
14087 kgem_bo_destroy(&sna->kgem, upload);
14101 struct kgem_bo *upload;
14211 upload = kgem_create_buffer(&sna->kgem,
14215 if (!upload)
14250 kgem_add_reloc64(&sna->kgem, sna->kgem.nbatch + 6, upload,
14268 b[5] = kgem_add_reloc(&sna->kgem, sna->kgem.nbatch + 5, upload,
14279 kgem_bo_destroy(&sna->kgem, upload);
14538 struct kgem_bo *upload;
14543 upload = kgem_bo_reference(*tile);
14545 upload = kgem_create_buffer(&sna->kgem, bw*bh,
14548 if (!upload)
14566 kgem_add_reloc64(&sna->kgem, sna->kgem.nbatch + 6, upload,
14583 b[5] = kgem_add_reloc(&sna->kgem, sna->kgem.nbatch + 5, upload,
14608 *tile = kgem_bo_reference(upload);
14611 kgem_bo_destroy(&sna->kgem, upload);
14630 DBG(("%s: upload (%d, %d), (%d, %d), origin (%d, %d), clipped=%d, alu=%d, opaque=%d\n", __FUNCTION__,
14770 DBG(("%s: upload (%d, %d), (%d, %d), origin (%d, %d), clipped=%d, alu=%d, opaque=%d\n", __FUNCTION__,
16675 DBG(("%s: upload(%d, %d, %d, %d)\n", __FUNCTION__,
16692 struct kgem_bo *upload;
16705 upload = kgem_create_buffer(&sna->kgem,
16709 if (!upload)
16755 kgem_add_reloc64(&sna->kgem, sna->kgem.nbatch + 6, upload,
16780 b[5] = kgem_add_reloc(&sna->kgem, sna->kgem.nbatch + 5, upload,
16792 kgem_bo_destroy(&sna->kgem, upload);