1/****************************************************************************
2 * Copyright (C) 2015 Intel Corporation.   All Rights Reserved.
3 *
4 * Permission is hereby granted, free of charge, to any person obtaining a
5 * copy of this software and associated documentation files (the "Software"),
6 * to deal in the Software without restriction, including without limitation
7 * the rights to use, copy, modify, merge, publish, distribute, sublicense,
8 * and/or sell copies of the Software, and to permit persons to whom the
9 * Software is furnished to do so, subject to the following conditions:
10 *
11 * The above copyright notice and this permission notice (including the next
12 * paragraph) shall be included in all copies or substantial portions of the
13 * Software.
14 *
15 * THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR
16 * IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY,
17 * FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT.  IN NO EVENT SHALL
18 * THE AUTHORS OR COPYRIGHT HOLDERS BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER
19 * LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING
20 * FROM, OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS
21 * IN THE SOFTWARE.
22 ***************************************************************************/
23
24#include "swr_context.h"
25#include "swr_memory.h"
26#include "swr_screen.h"
27#include "swr_resource.h"
28#include "swr_scratch.h"
29#include "swr_query.h"
30#include "swr_fence.h"
31
32#include "util/u_memory.h"
33#include "util/u_inlines.h"
34#include "util/u_format.h"
35#include "util/u_atomic.h"
36#include "util/u_upload_mgr.h"
37#include "util/u_transfer.h"
38#include "util/u_surface.h"
39
40#include "api.h"
41#include "backend.h"
42#include "knobs.h"
43
44static struct pipe_surface *
45swr_create_surface(struct pipe_context *pipe,
46                   struct pipe_resource *pt,
47                   const struct pipe_surface *surf_tmpl)
48{
49   struct pipe_surface *ps;
50
51   ps = CALLOC_STRUCT(pipe_surface);
52   if (ps) {
53      pipe_reference_init(&ps->reference, 1);
54      pipe_resource_reference(&ps->texture, pt);
55      ps->context = pipe;
56      ps->format = surf_tmpl->format;
57      if (pt->target != PIPE_BUFFER) {
58         assert(surf_tmpl->u.tex.level <= pt->last_level);
59         ps->width = u_minify(pt->width0, surf_tmpl->u.tex.level);
60         ps->height = u_minify(pt->height0, surf_tmpl->u.tex.level);
61         ps->u.tex.level = surf_tmpl->u.tex.level;
62         ps->u.tex.first_layer = surf_tmpl->u.tex.first_layer;
63         ps->u.tex.last_layer = surf_tmpl->u.tex.last_layer;
64      } else {
65         /* setting width as number of elements should get us correct
66          * renderbuffer width */
67         ps->width = surf_tmpl->u.buf.last_element
68            - surf_tmpl->u.buf.first_element + 1;
69         ps->height = pt->height0;
70         ps->u.buf.first_element = surf_tmpl->u.buf.first_element;
71         ps->u.buf.last_element = surf_tmpl->u.buf.last_element;
72         assert(ps->u.buf.first_element <= ps->u.buf.last_element);
73         assert(ps->u.buf.last_element < ps->width);
74      }
75   }
76   return ps;
77}
78
79static void
80swr_surface_destroy(struct pipe_context *pipe, struct pipe_surface *surf)
81{
82   assert(surf->texture);
83   struct pipe_resource *resource = surf->texture;
84
85   /* If the resource has been drawn to, store tiles. */
86   swr_store_dirty_resource(pipe, resource, SWR_TILE_RESOLVED);
87
88   pipe_resource_reference(&resource, NULL);
89   FREE(surf);
90}
91
92
93static void *
94swr_transfer_map(struct pipe_context *pipe,
95                 struct pipe_resource *resource,
96                 unsigned level,
97                 unsigned usage,
98                 const struct pipe_box *box,
99                 struct pipe_transfer **transfer)
100{
101   struct swr_screen *screen = swr_screen(pipe->screen);
102   struct swr_resource *spr = swr_resource(resource);
103   struct pipe_transfer *pt;
104   enum pipe_format format = resource->format;
105
106   assert(resource);
107   assert(level <= resource->last_level);
108
109   /* If mapping an attached rendertarget, store tiles to surface and set
110    * postStoreTileState to SWR_TILE_INVALID so tiles get reloaded on next use
111    * and nothing needs to be done at unmap. */
112   swr_store_dirty_resource(pipe, resource, SWR_TILE_INVALID);
113
114   if (!(usage & PIPE_TRANSFER_UNSYNCHRONIZED)) {
115      /* If resource is in use, finish fence before mapping.
116       * Unless requested not to block, then if not done return NULL map */
117      if (usage & PIPE_TRANSFER_DONTBLOCK) {
118         if (swr_is_fence_pending(screen->flush_fence))
119            return NULL;
120      } else {
121         if (spr->status) {
122            /* But, if there's no fence pending, submit one.
123             * XXX: Remove once draw timestamps are finished. */
124            if (!swr_is_fence_pending(screen->flush_fence))
125               swr_fence_submit(swr_context(pipe), screen->flush_fence);
126
127            swr_fence_finish(pipe->screen, NULL, screen->flush_fence, 0);
128            swr_resource_unused(resource);
129         }
130      }
131   }
132
133   pt = CALLOC_STRUCT(pipe_transfer);
134   if (!pt)
135      return NULL;
136   pipe_resource_reference(&pt->resource, resource);
137   pt->usage = (pipe_transfer_usage)usage;
138   pt->level = level;
139   pt->box = *box;
140   pt->stride = spr->swr.pitch;
141   pt->layer_stride = spr->swr.qpitch * spr->swr.pitch;
142
143   /* if we're mapping the depth/stencil, copy in stencil for the section
144    * being read in
145    */
146   if (usage & PIPE_TRANSFER_READ && spr->has_depth && spr->has_stencil) {
147      size_t zbase, sbase;
148      for (int z = box->z; z < box->z + box->depth; z++) {
149         zbase = (z * spr->swr.qpitch + box->y) * spr->swr.pitch +
150            spr->mip_offsets[level];
151         sbase = (z * spr->secondary.qpitch + box->y) * spr->secondary.pitch +
152            spr->secondary_mip_offsets[level];
153         for (int y = box->y; y < box->y + box->height; y++) {
154            if (spr->base.format == PIPE_FORMAT_Z24_UNORM_S8_UINT) {
155               for (int x = box->x; x < box->x + box->width; x++)
156                  ((uint8_t*)(spr->swr.xpBaseAddress))[zbase + 4 * x + 3] =
157                     ((uint8_t*)(spr->secondary.xpBaseAddress))[sbase + x];
158            } else if (spr->base.format == PIPE_FORMAT_Z32_FLOAT_S8X24_UINT) {
159               for (int x = box->x; x < box->x + box->width; x++)
160                  ((uint8_t*)(spr->swr.xpBaseAddress))[zbase + 8 * x + 4] =
161                     ((uint8_t*)(spr->secondary.xpBaseAddress))[sbase + x];
162            }
163            zbase += spr->swr.pitch;
164            sbase += spr->secondary.pitch;
165         }
166      }
167   }
168
169   unsigned offset = box->z * pt->layer_stride +
170      util_format_get_nblocksy(format, box->y) * pt->stride +
171      util_format_get_stride(format, box->x);
172
173   *transfer = pt;
174
175   return (void*)(spr->swr.xpBaseAddress + offset + spr->mip_offsets[level]);
176}
177
178static void
179swr_transfer_flush_region(struct pipe_context *pipe,
180                          struct pipe_transfer *transfer,
181                          const struct pipe_box *flush_box)
182{
183   assert(transfer->resource);
184   assert(transfer->usage & PIPE_TRANSFER_WRITE);
185
186   struct swr_resource *spr = swr_resource(transfer->resource);
187   if (!spr->has_depth || !spr->has_stencil)
188      return;
189
190   size_t zbase, sbase;
191   struct pipe_box box = *flush_box;
192   box.x += transfer->box.x;
193   box.y += transfer->box.y;
194   box.z += transfer->box.z;
195   for (int z = box.z; z < box.z + box.depth; z++) {
196      zbase = (z * spr->swr.qpitch + box.y) * spr->swr.pitch +
197         spr->mip_offsets[transfer->level];
198      sbase = (z * spr->secondary.qpitch + box.y) * spr->secondary.pitch +
199         spr->secondary_mip_offsets[transfer->level];
200      for (int y = box.y; y < box.y + box.height; y++) {
201         if (spr->base.format == PIPE_FORMAT_Z24_UNORM_S8_UINT) {
202            for (int x = box.x; x < box.x + box.width; x++)
203               ((uint8_t*)(spr->secondary.xpBaseAddress))[sbase + x] =
204                  ((uint8_t*)(spr->swr.xpBaseAddress))[zbase + 4 * x + 3];
205         } else if (spr->base.format == PIPE_FORMAT_Z32_FLOAT_S8X24_UINT) {
206            for (int x = box.x; x < box.x + box.width; x++)
207               ((uint8_t*)(spr->secondary.xpBaseAddress))[sbase + x] =
208                  ((uint8_t*)(spr->swr.xpBaseAddress))[zbase + 8 * x + 4];
209         }
210         zbase += spr->swr.pitch;
211         sbase += spr->secondary.pitch;
212      }
213   }
214}
215
216static void
217swr_transfer_unmap(struct pipe_context *pipe, struct pipe_transfer *transfer)
218{
219   assert(transfer->resource);
220
221   struct swr_resource *spr = swr_resource(transfer->resource);
222   /* if we're mapping the depth/stencil, copy in stencil for the section
223    * being written out
224    */
225   if (transfer->usage & PIPE_TRANSFER_WRITE &&
226       !(transfer->usage & PIPE_TRANSFER_FLUSH_EXPLICIT) &&
227       spr->has_depth && spr->has_stencil) {
228      struct pipe_box box;
229      u_box_3d(0, 0, 0, transfer->box.width, transfer->box.height,
230               transfer->box.depth, &box);
231      swr_transfer_flush_region(pipe, transfer, &box);
232   }
233
234   pipe_resource_reference(&transfer->resource, NULL);
235   FREE(transfer);
236}
237
238
239static void
240swr_resource_copy(struct pipe_context *pipe,
241                  struct pipe_resource *dst,
242                  unsigned dst_level,
243                  unsigned dstx,
244                  unsigned dsty,
245                  unsigned dstz,
246                  struct pipe_resource *src,
247                  unsigned src_level,
248                  const struct pipe_box *src_box)
249{
250   struct swr_screen *screen = swr_screen(pipe->screen);
251
252   /* If either the src or dst is a renderTarget, store tiles before copy */
253   swr_store_dirty_resource(pipe, src, SWR_TILE_RESOLVED);
254   swr_store_dirty_resource(pipe, dst, SWR_TILE_RESOLVED);
255
256   swr_fence_finish(pipe->screen, NULL, screen->flush_fence, 0);
257   swr_resource_unused(src);
258   swr_resource_unused(dst);
259
260   if ((dst->target == PIPE_BUFFER && src->target == PIPE_BUFFER)
261       || (dst->target != PIPE_BUFFER && src->target != PIPE_BUFFER)) {
262      util_resource_copy_region(
263         pipe, dst, dst_level, dstx, dsty, dstz, src, src_level, src_box);
264      return;
265   }
266
267   debug_printf("unhandled swr_resource_copy\n");
268}
269
270
271static void
272swr_blit(struct pipe_context *pipe, const struct pipe_blit_info *blit_info)
273{
274   struct swr_context *ctx = swr_context(pipe);
275   /* Make a copy of the const blit_info, so we can modify it */
276   struct pipe_blit_info info = *blit_info;
277
278   if (info.render_condition_enable && !swr_check_render_cond(pipe))
279      return;
280
281   if (info.src.resource->nr_samples > 1 && info.dst.resource->nr_samples <= 1
282       && !util_format_is_depth_or_stencil(info.src.resource->format)
283       && !util_format_is_pure_integer(info.src.resource->format)) {
284      debug_printf("swr_blit: color resolve : %d -> %d\n",
285            info.src.resource->nr_samples, info.dst.resource->nr_samples);
286
287      /* Resolve is done as part of the surface store. */
288      swr_store_dirty_resource(pipe, info.src.resource, SWR_TILE_RESOLVED);
289
290      struct pipe_resource *src_resource = info.src.resource;
291      struct pipe_resource *resolve_target =
292         swr_resource(src_resource)->resolve_target;
293
294      /* The resolve target becomes the new source for the blit. */
295      info.src.resource = resolve_target;
296   }
297
298   if (util_try_blit_via_copy_region(pipe, &info)) {
299      return; /* done */
300   }
301
302   if (info.mask & PIPE_MASK_S) {
303      debug_printf("swr: cannot blit stencil, skipping\n");
304      info.mask &= ~PIPE_MASK_S;
305   }
306
307   if (!util_blitter_is_blit_supported(ctx->blitter, &info)) {
308      debug_printf("swr: blit unsupported %s -> %s\n",
309                   util_format_short_name(info.src.resource->format),
310                   util_format_short_name(info.dst.resource->format));
311      return;
312   }
313
314   if (ctx->active_queries) {
315      ctx->api.pfnSwrEnableStatsFE(ctx->swrContext, FALSE);
316      ctx->api.pfnSwrEnableStatsBE(ctx->swrContext, FALSE);
317   }
318
319   util_blitter_save_vertex_buffer_slot(ctx->blitter, ctx->vertex_buffer);
320   util_blitter_save_vertex_elements(ctx->blitter, (void *)ctx->velems);
321   util_blitter_save_vertex_shader(ctx->blitter, (void *)ctx->vs);
322   util_blitter_save_geometry_shader(ctx->blitter, (void*)ctx->gs);
323   util_blitter_save_so_targets(
324      ctx->blitter,
325      ctx->num_so_targets,
326      (struct pipe_stream_output_target **)ctx->so_targets);
327   util_blitter_save_rasterizer(ctx->blitter, (void *)ctx->rasterizer);
328   util_blitter_save_viewport(ctx->blitter, &ctx->viewport);
329   util_blitter_save_scissor(ctx->blitter, &ctx->scissor);
330   util_blitter_save_fragment_shader(ctx->blitter, ctx->fs);
331   util_blitter_save_blend(ctx->blitter, (void *)ctx->blend);
332   util_blitter_save_depth_stencil_alpha(ctx->blitter,
333                                         (void *)ctx->depth_stencil);
334   util_blitter_save_stencil_ref(ctx->blitter, &ctx->stencil_ref);
335   util_blitter_save_sample_mask(ctx->blitter, ctx->sample_mask);
336   util_blitter_save_framebuffer(ctx->blitter, &ctx->framebuffer);
337   util_blitter_save_fragment_sampler_states(
338      ctx->blitter,
339      ctx->num_samplers[PIPE_SHADER_FRAGMENT],
340      (void **)ctx->samplers[PIPE_SHADER_FRAGMENT]);
341   util_blitter_save_fragment_sampler_views(
342      ctx->blitter,
343      ctx->num_sampler_views[PIPE_SHADER_FRAGMENT],
344      ctx->sampler_views[PIPE_SHADER_FRAGMENT]);
345   util_blitter_save_render_condition(ctx->blitter,
346                                      ctx->render_cond_query,
347                                      ctx->render_cond_cond,
348                                      ctx->render_cond_mode);
349
350   util_blitter_blit(ctx->blitter, &info);
351
352   if (ctx->active_queries) {
353      ctx->api.pfnSwrEnableStatsFE(ctx->swrContext, TRUE);
354      ctx->api.pfnSwrEnableStatsBE(ctx->swrContext, TRUE);
355   }
356}
357
358
359static void
360swr_destroy(struct pipe_context *pipe)
361{
362   struct swr_context *ctx = swr_context(pipe);
363   struct swr_screen *screen = swr_screen(pipe->screen);
364
365   if (ctx->blitter)
366      util_blitter_destroy(ctx->blitter);
367
368   for (unsigned i = 0; i < PIPE_MAX_COLOR_BUFS; i++) {
369      if (ctx->framebuffer.cbufs[i]) {
370         struct swr_resource *res = swr_resource(ctx->framebuffer.cbufs[i]->texture);
371         /* NULL curr_pipe, so we don't have a reference to a deleted pipe */
372         res->curr_pipe = NULL;
373         pipe_surface_reference(&ctx->framebuffer.cbufs[i], NULL);
374      }
375   }
376
377   if (ctx->framebuffer.zsbuf) {
378      struct swr_resource *res = swr_resource(ctx->framebuffer.zsbuf->texture);
379      /* NULL curr_pipe, so we don't have a reference to a deleted pipe */
380      res->curr_pipe = NULL;
381      pipe_surface_reference(&ctx->framebuffer.zsbuf, NULL);
382   }
383
384   for (unsigned i = 0; i < ARRAY_SIZE(ctx->sampler_views[0]); i++) {
385      pipe_sampler_view_reference(&ctx->sampler_views[PIPE_SHADER_FRAGMENT][i], NULL);
386   }
387
388   for (unsigned i = 0; i < ARRAY_SIZE(ctx->sampler_views[0]); i++) {
389      pipe_sampler_view_reference(&ctx->sampler_views[PIPE_SHADER_VERTEX][i], NULL);
390   }
391
392   if (ctx->pipe.stream_uploader)
393      u_upload_destroy(ctx->pipe.stream_uploader);
394
395   /* Idle core after destroying buffer resources, but before deleting
396    * context.  Destroying resources has potentially called StoreTiles.*/
397   ctx->api.pfnSwrWaitForIdle(ctx->swrContext);
398
399   if (ctx->swrContext)
400      ctx->api.pfnSwrDestroyContext(ctx->swrContext);
401
402   delete ctx->blendJIT;
403
404   swr_destroy_scratch_buffers(ctx);
405
406   /* Only update screen->pipe if current context is being destroyed */
407   assert(screen);
408   if (screen->pipe == pipe)
409      screen->pipe = NULL;
410
411   AlignedFree(ctx);
412}
413
414
415static void
416swr_render_condition(struct pipe_context *pipe,
417                     struct pipe_query *query,
418                     boolean condition,
419                     enum pipe_render_cond_flag mode)
420{
421   struct swr_context *ctx = swr_context(pipe);
422
423   ctx->render_cond_query = query;
424   ctx->render_cond_mode = mode;
425   ctx->render_cond_cond = condition;
426}
427
428static void
429swr_UpdateStats(HANDLE hPrivateContext, const SWR_STATS *pStats)
430{
431   swr_draw_context *pDC = (swr_draw_context*)hPrivateContext;
432
433   if (!pDC)
434      return;
435
436   struct swr_query_result *pqr = pDC->pStats;
437
438   SWR_STATS *pSwrStats = &pqr->core;
439
440   pSwrStats->DepthPassCount += pStats->DepthPassCount;
441   pSwrStats->PsInvocations += pStats->PsInvocations;
442   pSwrStats->CsInvocations += pStats->CsInvocations;
443}
444
445static void
446swr_UpdateStatsFE(HANDLE hPrivateContext, const SWR_STATS_FE *pStats)
447{
448   swr_draw_context *pDC = (swr_draw_context*)hPrivateContext;
449
450   if (!pDC)
451      return;
452
453   struct swr_query_result *pqr = pDC->pStats;
454
455   SWR_STATS_FE *pSwrStats = &pqr->coreFE;
456   p_atomic_add(&pSwrStats->IaVertices, pStats->IaVertices);
457   p_atomic_add(&pSwrStats->IaPrimitives, pStats->IaPrimitives);
458   p_atomic_add(&pSwrStats->VsInvocations, pStats->VsInvocations);
459   p_atomic_add(&pSwrStats->HsInvocations, pStats->HsInvocations);
460   p_atomic_add(&pSwrStats->DsInvocations, pStats->DsInvocations);
461   p_atomic_add(&pSwrStats->GsInvocations, pStats->GsInvocations);
462   p_atomic_add(&pSwrStats->CInvocations, pStats->CInvocations);
463   p_atomic_add(&pSwrStats->CPrimitives, pStats->CPrimitives);
464   p_atomic_add(&pSwrStats->GsPrimitives, pStats->GsPrimitives);
465
466   for (unsigned i = 0; i < 4; i++) {
467      p_atomic_add(&pSwrStats->SoPrimStorageNeeded[i],
468            pStats->SoPrimStorageNeeded[i]);
469      p_atomic_add(&pSwrStats->SoNumPrimsWritten[i],
470            pStats->SoNumPrimsWritten[i]);
471   }
472}
473
474struct pipe_context *
475swr_create_context(struct pipe_screen *p_screen, void *priv, unsigned flags)
476{
477   struct swr_context *ctx = (struct swr_context *)
478      AlignedMalloc(sizeof(struct swr_context), KNOB_SIMD_BYTES);
479   memset(ctx, 0, sizeof(struct swr_context));
480
481   swr_screen(p_screen)->pfnSwrGetInterface(ctx->api);
482   ctx->swrDC.pAPI = &ctx->api;
483
484   ctx->blendJIT =
485      new std::unordered_map<BLEND_COMPILE_STATE, PFN_BLEND_JIT_FUNC>;
486
487   ctx->max_draws_in_flight = KNOB_MAX_DRAWS_IN_FLIGHT;
488
489   SWR_CREATECONTEXT_INFO createInfo;
490   memset(&createInfo, 0, sizeof(createInfo));
491   createInfo.privateStateSize = sizeof(swr_draw_context);
492   createInfo.pfnLoadTile = swr_LoadHotTile;
493   createInfo.pfnStoreTile = swr_StoreHotTile;
494   createInfo.pfnClearTile = swr_StoreHotTileClear;
495   createInfo.pfnUpdateStats = swr_UpdateStats;
496   createInfo.pfnUpdateStatsFE = swr_UpdateStatsFE;
497   createInfo.pfnMakeGfxPtr = swr_MakeGfxPtr;
498
499   SWR_THREADING_INFO threadingInfo {0};
500
501   threadingInfo.MAX_WORKER_THREADS        = KNOB_MAX_WORKER_THREADS;
502   threadingInfo.MAX_NUMA_NODES            = KNOB_MAX_NUMA_NODES;
503   threadingInfo.MAX_CORES_PER_NUMA_NODE   = KNOB_MAX_CORES_PER_NUMA_NODE;
504   threadingInfo.MAX_THREADS_PER_CORE      = KNOB_MAX_THREADS_PER_CORE;
505   threadingInfo.SINGLE_THREADED           = KNOB_SINGLE_THREADED;
506
507   // Use non-standard settings for KNL
508   if (swr_screen(p_screen)->is_knl)
509   {
510      if (nullptr == getenv("KNOB_MAX_THREADS_PER_CORE"))
511         threadingInfo.MAX_THREADS_PER_CORE  = 2;
512
513      if (nullptr == getenv("KNOB_MAX_DRAWS_IN_FLIGHT"))
514      {
515         ctx->max_draws_in_flight = 2048;
516         createInfo.MAX_DRAWS_IN_FLIGHT = ctx->max_draws_in_flight;
517      }
518   }
519
520   createInfo.pThreadInfo = &threadingInfo;
521
522   ctx->swrContext = ctx->api.pfnSwrCreateContext(&createInfo);
523
524   ctx->api.pfnSwrInit();
525
526   if (ctx->swrContext == NULL)
527      goto fail;
528
529   ctx->pipe.screen = p_screen;
530   ctx->pipe.destroy = swr_destroy;
531   ctx->pipe.priv = priv;
532   ctx->pipe.create_surface = swr_create_surface;
533   ctx->pipe.surface_destroy = swr_surface_destroy;
534   ctx->pipe.transfer_map = swr_transfer_map;
535   ctx->pipe.transfer_unmap = swr_transfer_unmap;
536   ctx->pipe.transfer_flush_region = swr_transfer_flush_region;
537
538   ctx->pipe.buffer_subdata = u_default_buffer_subdata;
539   ctx->pipe.texture_subdata = u_default_texture_subdata;
540
541   ctx->pipe.clear_texture = util_clear_texture;
542   ctx->pipe.resource_copy_region = swr_resource_copy;
543   ctx->pipe.render_condition = swr_render_condition;
544
545   swr_state_init(&ctx->pipe);
546   swr_clear_init(&ctx->pipe);
547   swr_draw_init(&ctx->pipe);
548   swr_query_init(&ctx->pipe);
549
550   ctx->pipe.stream_uploader = u_upload_create_default(&ctx->pipe);
551   if (!ctx->pipe.stream_uploader)
552      goto fail;
553   ctx->pipe.const_uploader = ctx->pipe.stream_uploader;
554
555   ctx->pipe.blit = swr_blit;
556   ctx->blitter = util_blitter_create(&ctx->pipe);
557   if (!ctx->blitter)
558      goto fail;
559
560   swr_init_scratch_buffers(ctx);
561
562   return &ctx->pipe;
563
564fail:
565   /* Should really validate the init steps and fail gracefully */
566   swr_destroy(&ctx->pipe);
567   return NULL;
568}
569