Lines Matching defs:upload
385 /* special upload buffer */
505 /* and reset the upload counter */
2141 DBG(("%s: using magical upload buffer\n", __FUNCTION__));
2145 DBG(("%s: discarding cached upload buffer\n", __FUNCTION__));
2592 DBG(("%s: using magical upload buffer\n", __FUNCTION__));
2596 DBG(("%s: discarding cached upload buffer\n", __FUNCTION__));
3326 DBG(("%s: discarding cached upload buffer\n", __FUNCTION__));
3328 assert(priv->gpu_damage == NULL || DAMAGE_IS_ALL(priv->gpu_damage)); /* magical upload buffer */
3357 DBG(("%s: reusing cached upload\n", __FUNCTION__));
3377 DBG(("%s: using CPU bo for upload to GPU\n", __FUNCTION__));
3417 DBG(("%s: using CPU bo for upload to GPU\n", __FUNCTION__));
3446 DBG(("%s: using CPU bo for upload to GPU, %d boxes\n", __FUNCTION__, n));
3557 DBG(("%s: cached upload proxy, discard and revert to GPU\n", __FUNCTION__));
3559 assert(priv->gpu_damage == NULL || DAMAGE_IS_ALL(priv->gpu_damage)); /* magical upload buffer */
4225 DBG(("%s: promoting upload proxy handle=%d to GPU\n", __FUNCTION__, proxy->handle));
4256 DBG(("%s: reusing cached upload\n", __FUNCTION__));
4300 DBG(("%s: using CPU bo for upload to GPU\n", __FUNCTION__));
4669 DBG(("%s: upload(%d, %d, %d, %d) x %d\n", __FUNCTION__, x, y, w, h, n));
4777 DBG(("%s: discarding cached upload buffer\n", __FUNCTION__));
4779 assert(priv->gpu_damage == NULL || DAMAGE_IS_ALL(priv->gpu_damage)); /* magical upload buffer */
4872 DBG(("%s: upload(%d, %d, %d, %d) x %d\n", __FUNCTION__, x, y, w, h, n));
4973 DBG(("%s: upload(%d, %d, %d, %d) x %d through a temporary map\n",
5109 DBG(("%s: segmented, unclipped large upload (%d bytes), marking WHOLE_HINT\n",
5123 DBG(("%s: upload(%d, %d, %d, %d) x %d boxes\n", __FUNCTION__, x, y, w, h, n));
5212 DBG(("%s: upload(%d, %d, %d, %d)\n", __FUNCTION__, x, y, w, h));
5230 struct kgem_bo *upload;
5242 upload = kgem_create_buffer(&sna->kgem,
5246 if (!upload)
5266 assert(dst <= (uint8_t *)ptr + kgem_bo_size(upload));
5292 kgem_add_reloc64(&sna->kgem, sna->kgem.nbatch + 6, upload,
5318 b[5] = kgem_add_reloc(&sna->kgem, sna->kgem.nbatch + 5, upload,
5329 kgem_bo_destroy(&sna->kgem, upload);
5376 DBG(("%s: upload(%d, %d, %d, %d)\n", __FUNCTION__, x, y, w, h));
5400 struct kgem_bo *upload;
5412 upload = kgem_create_buffer(&sna->kgem,
5416 if (!upload)
5434 assert(dst <= (uint8_t *)ptr + kgem_bo_size(upload));
5462 kgem_add_reloc64(&sna->kgem, sna->kgem.nbatch + 6, upload,
5492 b[5] = kgem_add_reloc(&sna->kgem, sna->kgem.nbatch + 5, upload,
5506 kgem_bo_destroy(&sna->kgem, upload);
6288 DBG(("%s: discarding cached upload buffer\n", __FUNCTION__));
6290 assert(priv->gpu_damage == NULL || DAMAGE_IS_ALL(priv->gpu_damage)); /* magical upload buffer */
6518 DBG(("%s: region overlaps GPU damage, upload and copy\n",
6619 DBG(("%s: upload through a temporary map\n",
6655 DBG(("%s: creating temporary source upload for non-copy alu [%d]\n",
6702 DBG(("%s: caching upload for src bo\n",
8314 struct kgem_bo *upload;
8326 upload = kgem_create_buffer(&sna->kgem,
8330 if (!upload)
8349 kgem_add_reloc64(&sna->kgem, sna->kgem.nbatch + 6, upload,
8368 b[5] = kgem_add_reloc(&sna->kgem, sna->kgem.nbatch + 5, upload,
8393 assert(dst <= (uint8_t *)ptr + kgem_bo_size(upload));
8400 kgem_bo_destroy(&sna->kgem, upload);
8458 struct kgem_bo *upload;
8476 upload = kgem_create_buffer(&sna->kgem,
8480 if (!upload)
8605 kgem_add_reloc64(&sna->kgem, sna->kgem.nbatch + 6, upload,
8623 b[5] = kgem_add_reloc(&sna->kgem, sna->kgem.nbatch + 5, upload,
8634 kgem_bo_destroy(&sna->kgem, upload);
12120 struct kgem_bo *upload;
12125 upload = kgem_upload_source_image(&to_sna_from_pixmap(pixmap)->kgem,
12129 if (upload == NULL)
12133 DBG(("%s: adding upload cache to pixmap=%ld\n",
12135 assert(upload->proxy != NULL);
12136 kgem_proxy_bo_attach(upload, &priv->gpu_bo);
12139 return upload;
12613 struct kgem_bo *upload;
12667 upload = kgem_create_buffer(&sna->kgem, 8*bpp, KGEM_BUFFER_WRITE, &ptr);
12668 if (upload == NULL)
12671 upload->pitch = bpp; /* for sanity checks */
12722 upload, gc, n, rect,
12727 kgem_bo_destroy(&sna->kgem, upload);
13079 DBG(("%s: alu=%d, upload (%d, %d), (%d, %d), origin (%d, %d)\n",
13493 DBG(("%s: upload (%d, %d), (%d, %d), origin (%d, %d), clipped=%x\n", __FUNCTION__,
13598 struct kgem_bo *upload;
13610 upload = kgem_create_buffer(&sna->kgem,
13614 if (!upload)
13649 kgem_add_reloc64(&sna->kgem, sna->kgem.nbatch + 6, upload,
13667 b[5] = kgem_add_reloc(&sna->kgem, sna->kgem.nbatch + 5, upload,
13679 kgem_bo_destroy(&sna->kgem, upload);
13702 struct kgem_bo *upload;
13801 upload = kgem_create_buffer(&sna->kgem,
13805 if (!upload)
13840 kgem_add_reloc64(&sna->kgem, sna->kgem.nbatch + 6, upload,
13858 b[5] = kgem_add_reloc(&sna->kgem, sna->kgem.nbatch + 5, upload,
13869 kgem_bo_destroy(&sna->kgem, upload);
13883 struct kgem_bo *upload;
13991 upload = kgem_create_buffer(&sna->kgem,
13995 if (!upload)
14030 kgem_add_reloc64(&sna->kgem, sna->kgem.nbatch + 6, upload,
14048 b[5] = kgem_add_reloc(&sna->kgem, sna->kgem.nbatch + 5, upload,
14059 kgem_bo_destroy(&sna->kgem, upload);
14316 struct kgem_bo *upload;
14321 upload = kgem_bo_reference(*tile);
14323 upload = kgem_create_buffer(&sna->kgem, bw*bh,
14326 if (!upload)
14344 kgem_add_reloc64(&sna->kgem, sna->kgem.nbatch + 6, upload,
14361 b[5] = kgem_add_reloc(&sna->kgem, sna->kgem.nbatch + 5, upload,
14386 *tile = kgem_bo_reference(upload);
14389 kgem_bo_destroy(&sna->kgem, upload);
14408 DBG(("%s: upload (%d, %d), (%d, %d), origin (%d, %d), clipped=%d, alu=%d, opaque=%d\n", __FUNCTION__,
14547 DBG(("%s: upload (%d, %d), (%d, %d), origin (%d, %d), clipped=%d, alu=%d, opaque=%d\n", __FUNCTION__,
16447 DBG(("%s: upload(%d, %d, %d, %d)\n", __FUNCTION__,
16463 struct kgem_bo *upload;
16475 upload = kgem_create_buffer(&sna->kgem,
16479 if (!upload)
16525 kgem_add_reloc64(&sna->kgem, sna->kgem.nbatch + 6, upload,
16550 b[5] = kgem_add_reloc(&sna->kgem, sna->kgem.nbatch + 5, upload,
16562 kgem_bo_destroy(&sna->kgem, upload);