1/*
2 * Copyright 2008 Ben Skeggs
3 *
4 * Permission is hereby granted, free of charge, to any person obtaining a
5 * copy of this software and associated documentation files (the "Software"),
6 * to deal in the Software without restriction, including without limitation
7 * the rights to use, copy, modify, merge, publish, distribute, sublicense,
8 * and/or sell copies of the Software, and to permit persons to whom the
9 * Software is furnished to do so, subject to the following conditions:
10 *
11 * The above copyright notice and this permission notice shall be included in
12 * all copies or substantial portions of the Software.
13 *
14 * THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR
15 * IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY,
16 * FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT.  IN NO EVENT SHALL
17 * THE AUTHORS OR COPYRIGHT HOLDERS BE LIABLE FOR ANY CLAIM, DAMAGES OR
18 * OTHER LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE,
19 * ARISING FROM, OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR
20 * OTHER DEALINGS IN THE SOFTWARE.
21 */
22
23#include "nvc0/nvc0_context.h"
24#include "nvc0/nvc0_resource.h"
25#include "nvc0/gm107_texture.xml.h"
26#include "nvc0/nvc0_compute.xml.h"
27#include "nv50/g80_texture.xml.h"
28#include "nv50/g80_defs.xml.h"
29
30#include "util/format/u_format.h"
31
32#define NVE4_TIC_ENTRY_INVALID 0x000fffff
33#define NVE4_TSC_ENTRY_INVALID 0xfff00000
34
35static inline uint32_t
36nv50_tic_swizzle(const struct nvc0_format *fmt, unsigned swz, bool tex_int)
37{
38   switch (swz) {
39   case PIPE_SWIZZLE_X  : return fmt->tic.src_x;
40   case PIPE_SWIZZLE_Y: return fmt->tic.src_y;
41   case PIPE_SWIZZLE_Z : return fmt->tic.src_z;
42   case PIPE_SWIZZLE_W: return fmt->tic.src_w;
43   case PIPE_SWIZZLE_1:
44      return tex_int ? G80_TIC_SOURCE_ONE_INT : G80_TIC_SOURCE_ONE_FLOAT;
45   case PIPE_SWIZZLE_0:
46   default:
47      return G80_TIC_SOURCE_ZERO;
48   }
49}
50
51struct pipe_sampler_view *
52nvc0_create_sampler_view(struct pipe_context *pipe,
53                         struct pipe_resource *res,
54                         const struct pipe_sampler_view *templ)
55{
56   uint32_t flags = 0;
57
58   if (templ->target == PIPE_TEXTURE_RECT || templ->target == PIPE_BUFFER)
59      flags |= NV50_TEXVIEW_SCALED_COORDS;
60
61   return nvc0_create_texture_view(pipe, res, templ, flags);
62}
63
64static struct pipe_sampler_view *
65gm107_create_texture_view(struct pipe_context *pipe,
66                          struct pipe_resource *texture,
67                          const struct pipe_sampler_view *templ,
68                          uint32_t flags)
69{
70   const struct util_format_description *desc;
71   const struct nvc0_format *fmt;
72   uint64_t address;
73   uint32_t *tic;
74   uint32_t swz[4];
75   uint32_t width, height;
76   uint32_t depth;
77   struct nv50_tic_entry *view;
78   struct nv50_miptree *mt;
79   bool tex_int;
80
81   view = MALLOC_STRUCT(nv50_tic_entry);
82   if (!view)
83      return NULL;
84   mt = nv50_miptree(texture);
85
86   view->pipe = *templ;
87   view->pipe.reference.count = 1;
88   view->pipe.texture = NULL;
89   view->pipe.context = pipe;
90
91   view->id = -1;
92   view->bindless = 0;
93
94   pipe_resource_reference(&view->pipe.texture, texture);
95
96   tic = &view->tic[0];
97
98   desc = util_format_description(view->pipe.format);
99   tex_int = util_format_is_pure_integer(view->pipe.format);
100
101   fmt = &nvc0_format_table[view->pipe.format];
102   swz[0] = nv50_tic_swizzle(fmt, view->pipe.swizzle_r, tex_int);
103   swz[1] = nv50_tic_swizzle(fmt, view->pipe.swizzle_g, tex_int);
104   swz[2] = nv50_tic_swizzle(fmt, view->pipe.swizzle_b, tex_int);
105   swz[3] = nv50_tic_swizzle(fmt, view->pipe.swizzle_a, tex_int);
106
107   tic[0]  = fmt->tic.format << GM107_TIC2_0_COMPONENTS_SIZES__SHIFT;
108   tic[0] |= fmt->tic.type_r << GM107_TIC2_0_R_DATA_TYPE__SHIFT;
109   tic[0] |= fmt->tic.type_g << GM107_TIC2_0_G_DATA_TYPE__SHIFT;
110   tic[0] |= fmt->tic.type_b << GM107_TIC2_0_B_DATA_TYPE__SHIFT;
111   tic[0] |= fmt->tic.type_a << GM107_TIC2_0_A_DATA_TYPE__SHIFT;
112   tic[0] |= swz[0] << GM107_TIC2_0_X_SOURCE__SHIFT;
113   tic[0] |= swz[1] << GM107_TIC2_0_Y_SOURCE__SHIFT;
114   tic[0] |= swz[2] << GM107_TIC2_0_Z_SOURCE__SHIFT;
115   tic[0] |= swz[3] << GM107_TIC2_0_W_SOURCE__SHIFT;
116
117   address = mt->base.address;
118
119   tic[3]  = GM107_TIC2_3_LOD_ANISO_QUALITY_2;
120   tic[4]  = GM107_TIC2_4_SECTOR_PROMOTION_PROMOTE_TO_2_V;
121   tic[4] |= GM107_TIC2_4_BORDER_SIZE_SAMPLER_COLOR;
122
123   if (desc->colorspace == UTIL_FORMAT_COLORSPACE_SRGB)
124      tic[4] |= GM107_TIC2_4_SRGB_CONVERSION;
125
126   if (!(flags & NV50_TEXVIEW_SCALED_COORDS))
127      tic[5] = GM107_TIC2_5_NORMALIZED_COORDS;
128   else
129      tic[5] = 0;
130
131   /* check for linear storage type */
132   if (unlikely(!nouveau_bo_memtype(nv04_resource(texture)->bo))) {
133      if (texture->target == PIPE_BUFFER) {
134         assert(!(tic[5] & GM107_TIC2_5_NORMALIZED_COORDS));
135         width = view->pipe.u.buf.size / (desc->block.bits / 8) - 1;
136         address +=
137            view->pipe.u.buf.offset;
138         tic[2]  = GM107_TIC2_2_HEADER_VERSION_ONE_D_BUFFER;
139         tic[3] |= width >> 16;
140         tic[4] |= GM107_TIC2_4_TEXTURE_TYPE_ONE_D_BUFFER;
141         tic[4] |= width & 0xffff;
142      } else {
143         assert(!(mt->level[0].pitch & 0x1f));
144         /* must be 2D texture without mip maps */
145         tic[2]  = GM107_TIC2_2_HEADER_VERSION_PITCH;
146         tic[4] |= GM107_TIC2_4_TEXTURE_TYPE_TWO_D_NO_MIPMAP;
147         tic[3] |= mt->level[0].pitch >> 5;
148         tic[4] |= mt->base.base.width0 - 1;
149         tic[5] |= 0 << GM107_TIC2_5_DEPTH_MINUS_ONE__SHIFT;
150         tic[5] |= mt->base.base.height0 - 1;
151      }
152      tic[1]  = address;
153      tic[2] |= address >> 32;
154      tic[6]  = 0;
155      tic[7]  = 0;
156      return &view->pipe;
157   }
158
159   tic[2]  = GM107_TIC2_2_HEADER_VERSION_BLOCKLINEAR;
160   tic[3] |=
161      ((mt->level[0].tile_mode & 0x0f0) >> 4 << 3) |
162      ((mt->level[0].tile_mode & 0xf00) >> 8 << 6);
163
164   depth = MAX2(mt->base.base.array_size, mt->base.base.depth0);
165
166   if (mt->base.base.array_size > 1) {
167      /* there doesn't seem to be a base layer field in TIC */
168      address += view->pipe.u.tex.first_layer * mt->layer_stride;
169      depth = view->pipe.u.tex.last_layer - view->pipe.u.tex.first_layer + 1;
170   }
171   tic[1]  = address;
172   tic[2] |= address >> 32;
173
174   switch (templ->target) {
175   case PIPE_TEXTURE_1D:
176      tic[4] |= GM107_TIC2_4_TEXTURE_TYPE_ONE_D;
177      break;
178   case PIPE_TEXTURE_2D:
179      tic[4] |= GM107_TIC2_4_TEXTURE_TYPE_TWO_D;
180      break;
181   case PIPE_TEXTURE_RECT:
182      tic[4] |= GM107_TIC2_4_TEXTURE_TYPE_TWO_D;
183      break;
184   case PIPE_TEXTURE_3D:
185      tic[4] |= GM107_TIC2_4_TEXTURE_TYPE_THREE_D;
186      break;
187   case PIPE_TEXTURE_CUBE:
188      depth /= 6;
189      tic[4] |= GM107_TIC2_4_TEXTURE_TYPE_CUBEMAP;
190      break;
191   case PIPE_TEXTURE_1D_ARRAY:
192      tic[4] |= GM107_TIC2_4_TEXTURE_TYPE_ONE_D_ARRAY;
193      break;
194   case PIPE_TEXTURE_2D_ARRAY:
195      tic[4] |= GM107_TIC2_4_TEXTURE_TYPE_TWO_D_ARRAY;
196      break;
197   case PIPE_TEXTURE_CUBE_ARRAY:
198      depth /= 6;
199      tic[4] |= GM107_TIC2_4_TEXTURE_TYPE_CUBE_ARRAY;
200      break;
201   default:
202      unreachable("unexpected/invalid texture target");
203   }
204
205   tic[3] |= (flags & NV50_TEXVIEW_FILTER_MSAA8) ?
206             GM107_TIC2_3_USE_HEADER_OPT_CONTROL :
207             GM107_TIC2_3_LOD_ANISO_QUALITY_HIGH |
208             GM107_TIC2_3_LOD_ISO_QUALITY_HIGH;
209
210   if (flags & (NV50_TEXVIEW_ACCESS_RESOLVE | NV50_TEXVIEW_IMAGE_GM107)) {
211      width = mt->base.base.width0 << mt->ms_x;
212      height = mt->base.base.height0 << mt->ms_y;
213   } else {
214      width = mt->base.base.width0;
215      height = mt->base.base.height0;
216   }
217
218   tic[4] |= width - 1;
219
220   tic[5] |= (height - 1) & 0xffff;
221   tic[5] |= (depth - 1) << GM107_TIC2_5_DEPTH_MINUS_ONE__SHIFT;
222   tic[3] |= mt->base.base.last_level << GM107_TIC2_3_MAX_MIP_LEVEL__SHIFT;
223
224   /* sampling points: (?) */
225   if ((flags & NV50_TEXVIEW_ACCESS_RESOLVE) && mt->ms_x > 1) {
226      tic[6]  = GM107_TIC2_6_ANISO_FINE_SPREAD_MODIFIER_CONST_TWO;
227      tic[6] |= GM107_TIC2_6_MAX_ANISOTROPY_2_TO_1;
228   } else {
229      tic[6]  = GM107_TIC2_6_ANISO_FINE_SPREAD_FUNC_TWO;
230      tic[6] |= GM107_TIC2_6_ANISO_COARSE_SPREAD_FUNC_ONE;
231   }
232
233   tic[7]  = (view->pipe.u.tex.last_level << 4) | view->pipe.u.tex.first_level;
234   tic[7] |= mt->ms_mode << GM107_TIC2_7_MULTI_SAMPLE_COUNT__SHIFT;
235
236   return &view->pipe;
237}
238
239struct pipe_sampler_view *
240gm107_create_texture_view_from_image(struct pipe_context *pipe,
241                                     const struct pipe_image_view *view)
242{
243   struct nv04_resource *res = nv04_resource(view->resource);
244   struct pipe_sampler_view templ = {};
245   enum pipe_texture_target target;
246   uint32_t flags = 0;
247
248   if (!res)
249      return NULL;
250   target = res->base.target;
251
252   if (target == PIPE_TEXTURE_CUBE || target == PIPE_TEXTURE_CUBE_ARRAY)
253      target = PIPE_TEXTURE_2D_ARRAY;
254
255   templ.target = target;
256   templ.format = view->format;
257   templ.swizzle_r = PIPE_SWIZZLE_X;
258   templ.swizzle_g = PIPE_SWIZZLE_Y;
259   templ.swizzle_b = PIPE_SWIZZLE_Z;
260   templ.swizzle_a = PIPE_SWIZZLE_W;
261
262   if (target == PIPE_BUFFER) {
263      templ.u.buf.offset = view->u.buf.offset;
264      templ.u.buf.size = view->u.buf.size;
265   } else {
266      templ.u.tex.first_layer = view->u.tex.first_layer;
267      templ.u.tex.last_layer = view->u.tex.last_layer;
268      templ.u.tex.first_level = templ.u.tex.last_level = view->u.tex.level;
269   }
270
271   flags = NV50_TEXVIEW_SCALED_COORDS | NV50_TEXVIEW_IMAGE_GM107;
272
273   return nvc0_create_texture_view(pipe, &res->base, &templ, flags);
274}
275
276static struct pipe_sampler_view *
277gf100_create_texture_view(struct pipe_context *pipe,
278                          struct pipe_resource *texture,
279                          const struct pipe_sampler_view *templ,
280                          uint32_t flags)
281{
282   const struct util_format_description *desc;
283   const struct nvc0_format *fmt;
284   uint64_t address;
285   uint32_t *tic;
286   uint32_t swz[4];
287   uint32_t width, height;
288   uint32_t depth;
289   uint32_t tex_fmt;
290   struct nv50_tic_entry *view;
291   struct nv50_miptree *mt;
292   bool tex_int;
293
294   view = MALLOC_STRUCT(nv50_tic_entry);
295   if (!view)
296      return NULL;
297   mt = nv50_miptree(texture);
298
299   view->pipe = *templ;
300   view->pipe.reference.count = 1;
301   view->pipe.texture = NULL;
302   view->pipe.context = pipe;
303
304   view->id = -1;
305   view->bindless = 0;
306
307   pipe_resource_reference(&view->pipe.texture, texture);
308
309   tic = &view->tic[0];
310
311   desc = util_format_description(view->pipe.format);
312
313   fmt = &nvc0_format_table[view->pipe.format];
314
315   tex_int = util_format_is_pure_integer(view->pipe.format);
316   tex_fmt = fmt->tic.format & 0x3f;
317
318   swz[0] = nv50_tic_swizzle(fmt, view->pipe.swizzle_r, tex_int);
319   swz[1] = nv50_tic_swizzle(fmt, view->pipe.swizzle_g, tex_int);
320   swz[2] = nv50_tic_swizzle(fmt, view->pipe.swizzle_b, tex_int);
321   swz[3] = nv50_tic_swizzle(fmt, view->pipe.swizzle_a, tex_int);
322   tic[0] = (tex_fmt << G80_TIC_0_COMPONENTS_SIZES__SHIFT) |
323            (fmt->tic.type_r << G80_TIC_0_R_DATA_TYPE__SHIFT) |
324            (fmt->tic.type_g << G80_TIC_0_G_DATA_TYPE__SHIFT) |
325            (fmt->tic.type_b << G80_TIC_0_B_DATA_TYPE__SHIFT) |
326            (fmt->tic.type_a << G80_TIC_0_A_DATA_TYPE__SHIFT) |
327            (swz[0] << G80_TIC_0_X_SOURCE__SHIFT) |
328            (swz[1] << G80_TIC_0_Y_SOURCE__SHIFT) |
329            (swz[2] << G80_TIC_0_Z_SOURCE__SHIFT) |
330            (swz[3] << G80_TIC_0_W_SOURCE__SHIFT) |
331            ((fmt->tic.format & 0x40) << (GK20A_TIC_0_USE_COMPONENT_SIZES_EXTENDED__SHIFT - 6));
332
333   address = mt->base.address;
334
335   tic[2] = 0x10001000 | G80_TIC_2_BORDER_SOURCE_COLOR;
336
337   if (desc->colorspace == UTIL_FORMAT_COLORSPACE_SRGB)
338      tic[2] |= G80_TIC_2_SRGB_CONVERSION;
339
340   if (!(flags & NV50_TEXVIEW_SCALED_COORDS))
341      tic[2] |= G80_TIC_2_NORMALIZED_COORDS;
342
343   /* check for linear storage type */
344   if (unlikely(!nouveau_bo_memtype(nv04_resource(texture)->bo))) {
345      if (texture->target == PIPE_BUFFER) {
346         assert(!(tic[2] & G80_TIC_2_NORMALIZED_COORDS));
347         address +=
348            view->pipe.u.buf.offset;
349         tic[2] |= G80_TIC_2_LAYOUT_PITCH | G80_TIC_2_TEXTURE_TYPE_ONE_D_BUFFER;
350         tic[3] = 0;
351         tic[4] = /* width */
352            view->pipe.u.buf.size / (desc->block.bits / 8);
353         tic[5] = 0;
354      } else {
355         /* must be 2D texture without mip maps */
356         tic[2] |= G80_TIC_2_LAYOUT_PITCH | G80_TIC_2_TEXTURE_TYPE_TWO_D_NO_MIPMAP;
357         tic[3] = mt->level[0].pitch;
358         tic[4] = mt->base.base.width0;
359         tic[5] = (1 << 16) | mt->base.base.height0;
360      }
361      tic[6] =
362      tic[7] = 0;
363      tic[1] = address;
364      tic[2] |= address >> 32;
365      return &view->pipe;
366   }
367
368   tic[2] |=
369      ((mt->level[0].tile_mode & 0x0f0) << (22 - 4)) |
370      ((mt->level[0].tile_mode & 0xf00) << (25 - 8));
371
372   depth = MAX2(mt->base.base.array_size, mt->base.base.depth0);
373
374   if (mt->base.base.array_size > 1) {
375      /* there doesn't seem to be a base layer field in TIC */
376      address += view->pipe.u.tex.first_layer * mt->layer_stride;
377      depth = view->pipe.u.tex.last_layer - view->pipe.u.tex.first_layer + 1;
378   }
379   tic[1] = address;
380   tic[2] |= address >> 32;
381
382   switch (templ->target) {
383   case PIPE_TEXTURE_1D:
384      tic[2] |= G80_TIC_2_TEXTURE_TYPE_ONE_D;
385      break;
386   case PIPE_TEXTURE_2D:
387      tic[2] |= G80_TIC_2_TEXTURE_TYPE_TWO_D;
388      break;
389   case PIPE_TEXTURE_RECT:
390      tic[2] |= G80_TIC_2_TEXTURE_TYPE_TWO_D;
391      break;
392   case PIPE_TEXTURE_3D:
393      tic[2] |= G80_TIC_2_TEXTURE_TYPE_THREE_D;
394      break;
395   case PIPE_TEXTURE_CUBE:
396      depth /= 6;
397      tic[2] |= G80_TIC_2_TEXTURE_TYPE_CUBEMAP;
398      break;
399   case PIPE_TEXTURE_1D_ARRAY:
400      tic[2] |= G80_TIC_2_TEXTURE_TYPE_ONE_D_ARRAY;
401      break;
402   case PIPE_TEXTURE_2D_ARRAY:
403      tic[2] |= G80_TIC_2_TEXTURE_TYPE_TWO_D_ARRAY;
404      break;
405   case PIPE_TEXTURE_CUBE_ARRAY:
406      depth /= 6;
407      tic[2] |= G80_TIC_2_TEXTURE_TYPE_CUBE_ARRAY;
408      break;
409   default:
410      unreachable("unexpected/invalid texture target");
411   }
412
413   tic[3] = (flags & NV50_TEXVIEW_FILTER_MSAA8) ? 0x20000000 : 0x00300000;
414
415   if (flags & NV50_TEXVIEW_ACCESS_RESOLVE) {
416      width = mt->base.base.width0 << mt->ms_x;
417      height = mt->base.base.height0 << mt->ms_y;
418   } else {
419      width = mt->base.base.width0;
420      height = mt->base.base.height0;
421   }
422
423   tic[4] = (1 << 31) | width;
424
425   tic[5] = height & 0xffff;
426   tic[5] |= depth << 16;
427   tic[5] |= mt->base.base.last_level << 28;
428
429   /* sampling points: (?) */
430   if (flags & NV50_TEXVIEW_ACCESS_RESOLVE)
431      tic[6] = (mt->ms_x > 1) ? 0x88000000 : 0x03000000;
432   else
433      tic[6] = 0x03000000;
434
435   tic[7] = (view->pipe.u.tex.last_level << 4) | view->pipe.u.tex.first_level;
436   tic[7] |= mt->ms_mode << 12;
437
438   return &view->pipe;
439}
440
441struct pipe_sampler_view *
442nvc0_create_texture_view(struct pipe_context *pipe,
443                         struct pipe_resource *texture,
444                         const struct pipe_sampler_view *templ,
445                         uint32_t flags)
446{
447   if (nvc0_context(pipe)->screen->tic.maxwell)
448      return gm107_create_texture_view(pipe, texture, templ, flags);
449   return gf100_create_texture_view(pipe, texture, templ, flags);
450}
451
452bool
453nvc0_update_tic(struct nvc0_context *nvc0, struct nv50_tic_entry *tic,
454                struct nv04_resource *res)
455{
456   uint64_t address = res->address;
457   if (res->base.target != PIPE_BUFFER)
458      return false;
459   address += tic->pipe.u.buf.offset;
460   if (tic->tic[1] == (uint32_t)address &&
461       (tic->tic[2] & 0xff) == address >> 32)
462      return false;
463
464   tic->tic[1] = address;
465   tic->tic[2] &= 0xffffff00;
466   tic->tic[2] |= address >> 32;
467
468   if (tic->id >= 0) {
469      nvc0->base.push_data(&nvc0->base, nvc0->screen->txc, tic->id * 32,
470                           NV_VRAM_DOMAIN(&nvc0->screen->base), 32,
471                           tic->tic);
472      return true;
473   }
474
475   return false;
476}
477
478bool
479nvc0_validate_tic(struct nvc0_context *nvc0, int s)
480{
481   uint32_t commands[32];
482   struct nouveau_pushbuf *push = nvc0->base.pushbuf;
483   unsigned i;
484   unsigned n = 0;
485   bool need_flush = false;
486
487   for (i = 0; i < nvc0->num_textures[s]; ++i) {
488      struct nv50_tic_entry *tic = nv50_tic_entry(nvc0->textures[s][i]);
489      struct nv04_resource *res;
490      const bool dirty = !!(nvc0->textures_dirty[s] & (1 << i));
491
492      if (!tic) {
493         if (dirty)
494            commands[n++] = (i << 1) | 0;
495         continue;
496      }
497      res = nv04_resource(tic->pipe.texture);
498      need_flush |= nvc0_update_tic(nvc0, tic, res);
499
500      if (tic->id < 0) {
501         tic->id = nvc0_screen_tic_alloc(nvc0->screen, tic);
502
503         nvc0->base.push_data(&nvc0->base, nvc0->screen->txc, tic->id * 32,
504                              NV_VRAM_DOMAIN(&nvc0->screen->base), 32,
505                              tic->tic);
506         need_flush = true;
507      } else
508      if (res->status & NOUVEAU_BUFFER_STATUS_GPU_WRITING) {
509         if (unlikely(s == 5))
510            BEGIN_NVC0(push, NVC0_CP(TEX_CACHE_CTL), 1);
511         else
512            BEGIN_NVC0(push, NVC0_3D(TEX_CACHE_CTL), 1);
513         PUSH_DATA (push, (tic->id << 4) | 1);
514         NOUVEAU_DRV_STAT(&nvc0->screen->base, tex_cache_flush_count, 1);
515      }
516      nvc0->screen->tic.lock[tic->id / 32] |= 1 << (tic->id % 32);
517
518      res->status &= ~NOUVEAU_BUFFER_STATUS_GPU_WRITING;
519      res->status |=  NOUVEAU_BUFFER_STATUS_GPU_READING;
520
521      if (!dirty)
522         continue;
523      commands[n++] = (tic->id << 9) | (i << 1) | 1;
524
525      if (unlikely(s == 5))
526         BCTX_REFN(nvc0->bufctx_cp, CP_TEX(i), res, RD);
527      else
528         BCTX_REFN(nvc0->bufctx_3d, 3D_TEX(s, i), res, RD);
529   }
530   for (; i < nvc0->state.num_textures[s]; ++i)
531      commands[n++] = (i << 1) | 0;
532
533   nvc0->state.num_textures[s] = nvc0->num_textures[s];
534
535   if (n) {
536      if (unlikely(s == 5))
537         BEGIN_NIC0(push, NVC0_CP(BIND_TIC), n);
538      else
539         BEGIN_NIC0(push, NVC0_3D(BIND_TIC(s)), n);
540      PUSH_DATAp(push, commands, n);
541   }
542   nvc0->textures_dirty[s] = 0;
543
544   return need_flush;
545}
546
547static bool
548nve4_validate_tic(struct nvc0_context *nvc0, unsigned s)
549{
550   struct nouveau_pushbuf *push = nvc0->base.pushbuf;
551   unsigned i;
552   bool need_flush = false;
553
554   for (i = 0; i < nvc0->num_textures[s]; ++i) {
555      struct nv50_tic_entry *tic = nv50_tic_entry(nvc0->textures[s][i]);
556      struct nv04_resource *res;
557      const bool dirty = !!(nvc0->textures_dirty[s] & (1 << i));
558
559      if (!tic) {
560         nvc0->tex_handles[s][i] |= NVE4_TIC_ENTRY_INVALID;
561         continue;
562      }
563      res = nv04_resource(tic->pipe.texture);
564      need_flush |= nvc0_update_tic(nvc0, tic, res);
565
566      if (tic->id < 0) {
567         tic->id = nvc0_screen_tic_alloc(nvc0->screen, tic);
568
569         nvc0->base.push_data(&nvc0->base, nvc0->screen->txc, tic->id * 32,
570                              NV_VRAM_DOMAIN(&nvc0->screen->base), 32,
571                              tic->tic);
572         need_flush = true;
573      } else
574      if (res->status & NOUVEAU_BUFFER_STATUS_GPU_WRITING) {
575         BEGIN_NVC0(push, NVC0_3D(TEX_CACHE_CTL), 1);
576         PUSH_DATA (push, (tic->id << 4) | 1);
577      }
578      nvc0->screen->tic.lock[tic->id / 32] |= 1 << (tic->id % 32);
579
580      res->status &= ~NOUVEAU_BUFFER_STATUS_GPU_WRITING;
581      res->status |=  NOUVEAU_BUFFER_STATUS_GPU_READING;
582
583      nvc0->tex_handles[s][i] &= ~NVE4_TIC_ENTRY_INVALID;
584      nvc0->tex_handles[s][i] |= tic->id;
585      if (dirty)
586         BCTX_REFN(nvc0->bufctx_3d, 3D_TEX(s, i), res, RD);
587   }
588   for (; i < nvc0->state.num_textures[s]; ++i) {
589      nvc0->tex_handles[s][i] |= NVE4_TIC_ENTRY_INVALID;
590      nvc0->textures_dirty[s] |= 1 << i;
591   }
592
593   nvc0->state.num_textures[s] = nvc0->num_textures[s];
594
595   return need_flush;
596}
597
598void nvc0_validate_textures(struct nvc0_context *nvc0)
599{
600   bool need_flush = false;
601   int i;
602
603   for (i = 0; i < 5; i++) {
604      if (nvc0->screen->base.class_3d >= NVE4_3D_CLASS)
605         need_flush |= nve4_validate_tic(nvc0, i);
606      else
607         need_flush |= nvc0_validate_tic(nvc0, i);
608   }
609
610   if (need_flush) {
611      BEGIN_NVC0(nvc0->base.pushbuf, NVC0_3D(TIC_FLUSH), 1);
612      PUSH_DATA (nvc0->base.pushbuf, 0);
613   }
614
615   /* Invalidate all CP textures because they are aliased. */
616   for (int i = 0; i < nvc0->num_textures[5]; i++)
617      nouveau_bufctx_reset(nvc0->bufctx_cp, NVC0_BIND_CP_TEX(i));
618   nvc0->textures_dirty[5] = ~0;
619   nvc0->dirty_cp |= NVC0_NEW_CP_TEXTURES;
620}
621
622bool
623nvc0_validate_tsc(struct nvc0_context *nvc0, int s)
624{
625   uint32_t commands[16];
626   struct nouveau_pushbuf *push = nvc0->base.pushbuf;
627   unsigned i;
628   unsigned n = 0;
629   bool need_flush = false;
630
631   for (i = 0; i < nvc0->num_samplers[s]; ++i) {
632      struct nv50_tsc_entry *tsc = nv50_tsc_entry(nvc0->samplers[s][i]);
633
634      if (!(nvc0->samplers_dirty[s] & (1 << i)))
635         continue;
636      if (!tsc) {
637         commands[n++] = (i << 4) | 0;
638         continue;
639      }
640      nvc0->seamless_cube_map = tsc->seamless_cube_map;
641      if (tsc->id < 0) {
642         tsc->id = nvc0_screen_tsc_alloc(nvc0->screen, tsc);
643
644         nvc0_m2mf_push_linear(&nvc0->base, nvc0->screen->txc,
645                               65536 + tsc->id * 32, NV_VRAM_DOMAIN(&nvc0->screen->base),
646                               32, tsc->tsc);
647         need_flush = true;
648      }
649      nvc0->screen->tsc.lock[tsc->id / 32] |= 1 << (tsc->id % 32);
650
651      commands[n++] = (tsc->id << 12) | (i << 4) | 1;
652   }
653   for (; i < nvc0->state.num_samplers[s]; ++i)
654      commands[n++] = (i << 4) | 0;
655
656   nvc0->state.num_samplers[s] = nvc0->num_samplers[s];
657
658   // TXF, in unlinked tsc mode, will always use sampler 0. So we have to
659   // ensure that it remains bound. Its contents don't matter, all samplers we
660   // ever create have the SRGB_CONVERSION bit set, so as long as the first
661   // entry is initialized, we're good to go. This is the only bit that has
662   // any effect on what TXF does.
663   if ((nvc0->samplers_dirty[s] & 1) && !nvc0->samplers[s][0]) {
664      if (n == 0)
665         n = 1;
666      // We're guaranteed that the first command refers to the first slot, so
667      // we're not overwriting a valid entry.
668      commands[0] = (0 << 12) | (0 << 4) | 1;
669   }
670
671   if (n) {
672      if (unlikely(s == 5))
673         BEGIN_NIC0(push, NVC0_CP(BIND_TSC), n);
674      else
675         BEGIN_NIC0(push, NVC0_3D(BIND_TSC(s)), n);
676      PUSH_DATAp(push, commands, n);
677   }
678   nvc0->samplers_dirty[s] = 0;
679
680   return need_flush;
681}
682
683bool
684nve4_validate_tsc(struct nvc0_context *nvc0, int s)
685{
686   unsigned i;
687   bool need_flush = false;
688
689   for (i = 0; i < nvc0->num_samplers[s]; ++i) {
690      struct nv50_tsc_entry *tsc = nv50_tsc_entry(nvc0->samplers[s][i]);
691
692      if (!tsc) {
693         nvc0->tex_handles[s][i] |= NVE4_TSC_ENTRY_INVALID;
694         continue;
695      }
696      if (tsc->id < 0) {
697         tsc->id = nvc0_screen_tsc_alloc(nvc0->screen, tsc);
698
699         nve4_p2mf_push_linear(&nvc0->base, nvc0->screen->txc,
700                               65536 + tsc->id * 32,
701                               NV_VRAM_DOMAIN(&nvc0->screen->base),
702                               32, tsc->tsc);
703         need_flush = true;
704      }
705      nvc0->screen->tsc.lock[tsc->id / 32] |= 1 << (tsc->id % 32);
706
707      nvc0->tex_handles[s][i] &= ~NVE4_TSC_ENTRY_INVALID;
708      nvc0->tex_handles[s][i] |= tsc->id << 20;
709   }
710   for (; i < nvc0->state.num_samplers[s]; ++i) {
711      nvc0->tex_handles[s][i] |= NVE4_TSC_ENTRY_INVALID;
712      nvc0->samplers_dirty[s] |= 1 << i;
713   }
714
715   nvc0->state.num_samplers[s] = nvc0->num_samplers[s];
716
717   return need_flush;
718}
719
720void nvc0_validate_samplers(struct nvc0_context *nvc0)
721{
722   bool need_flush = false;
723   int i;
724
725   for (i = 0; i < 5; i++) {
726      if (nvc0->screen->base.class_3d >= NVE4_3D_CLASS)
727         need_flush |= nve4_validate_tsc(nvc0, i);
728      else
729         need_flush |= nvc0_validate_tsc(nvc0, i);
730   }
731
732   if (need_flush) {
733      BEGIN_NVC0(nvc0->base.pushbuf, NVC0_3D(TSC_FLUSH), 1);
734      PUSH_DATA (nvc0->base.pushbuf, 0);
735   }
736
737   /* Invalidate all CP samplers because they are aliased. */
738   nvc0->samplers_dirty[5] = ~0;
739   nvc0->dirty_cp |= NVC0_NEW_CP_SAMPLERS;
740}
741
742void
743nvc0_upload_tsc0(struct nvc0_context *nvc0)
744{
745   struct nouveau_pushbuf *push = nvc0->base.pushbuf;
746   u32 data[8] = { G80_TSC_0_SRGB_CONVERSION };
747   nvc0->base.push_data(&nvc0->base, nvc0->screen->txc,
748                        65536 /*+ tsc->id * 32*/,
749                        NV_VRAM_DOMAIN(&nvc0->screen->base), 32, data);
750   BEGIN_NVC0(push, NVC0_3D(TSC_FLUSH), 1);
751   PUSH_DATA (push, 0);
752}
753
754/* Upload the "diagonal" entries for the possible texture sources ($t == $s).
755 * At some point we might want to get a list of the combinations used by a
756 * shader and fill in those entries instead of having it extract the handles.
757 */
758void
759nve4_set_tex_handles(struct nvc0_context *nvc0)
760{
761   struct nouveau_pushbuf *push = nvc0->base.pushbuf;
762   struct nvc0_screen *screen = nvc0->screen;
763   unsigned s;
764
765   if (nvc0->screen->base.class_3d < NVE4_3D_CLASS)
766      return;
767
768   for (s = 0; s < 5; ++s) {
769      uint32_t dirty = nvc0->textures_dirty[s] | nvc0->samplers_dirty[s];
770      if (!dirty)
771         continue;
772      BEGIN_NVC0(push, NVC0_3D(CB_SIZE), 3);
773      PUSH_DATA (push, NVC0_CB_AUX_SIZE);
774      PUSH_DATAh(push, screen->uniform_bo->offset + NVC0_CB_AUX_INFO(s));
775      PUSH_DATA (push, screen->uniform_bo->offset + NVC0_CB_AUX_INFO(s));
776      do {
777         int i = ffs(dirty) - 1;
778         dirty &= ~(1 << i);
779
780         BEGIN_NVC0(push, NVC0_3D(CB_POS), 2);
781         PUSH_DATA (push, NVC0_CB_AUX_TEX_INFO(i));
782         PUSH_DATA (push, nvc0->tex_handles[s][i]);
783      } while (dirty);
784
785      nvc0->textures_dirty[s] = 0;
786      nvc0->samplers_dirty[s] = 0;
787   }
788}
789
790static uint64_t
791nve4_create_texture_handle(struct pipe_context *pipe,
792                           struct pipe_sampler_view *view,
793                           const struct pipe_sampler_state *sampler)
794{
795   /* We have to create persistent handles that won't change for these objects
796    * That means that we have to upload them into place and lock them so that
797    * they can't be kicked out later.
798    */
799   struct nvc0_context *nvc0 = nvc0_context(pipe);
800   struct nouveau_pushbuf *push = nvc0->base.pushbuf;
801   struct nv50_tic_entry *tic = nv50_tic_entry(view);
802   struct nv50_tsc_entry *tsc = pipe->create_sampler_state(pipe, sampler);
803   struct pipe_sampler_view *v = NULL;
804
805   tsc->id = nvc0_screen_tsc_alloc(nvc0->screen, tsc);
806   if (tsc->id < 0)
807      goto fail;
808
809   if (tic->id < 0) {
810      tic->id = nvc0_screen_tic_alloc(nvc0->screen, tic);
811      if (tic->id < 0)
812         goto fail;
813
814      nve4_p2mf_push_linear(&nvc0->base, nvc0->screen->txc, tic->id * 32,
815                            NV_VRAM_DOMAIN(&nvc0->screen->base), 32,
816                            tic->tic);
817
818      IMMED_NVC0(push, NVC0_3D(TIC_FLUSH), 0);
819   }
820
821   nve4_p2mf_push_linear(&nvc0->base, nvc0->screen->txc,
822                         65536 + tsc->id * 32,
823                         NV_VRAM_DOMAIN(&nvc0->screen->base),
824                         32, tsc->tsc);
825
826   IMMED_NVC0(push, NVC0_3D(TSC_FLUSH), 0);
827
828   // Add an extra reference to this sampler view effectively held by this
829   // texture handle. This is to deal with the sampler view being dereferenced
830   // before the handle is. However we need the view to still be live until the
831   // handle to it is deleted.
832   pipe_sampler_view_reference(&v, view);
833   p_atomic_inc(&tic->bindless);
834
835   nvc0->screen->tic.lock[tic->id / 32] |= 1 << (tic->id % 32);
836   nvc0->screen->tsc.lock[tsc->id / 32] |= 1 << (tsc->id % 32);
837
838   return 0x100000000ULL | (tsc->id << 20) | tic->id;
839
840fail:
841   pipe->delete_sampler_state(pipe, tsc);
842   return 0;
843}
844
845static bool
846view_bound(struct nvc0_context *nvc0, struct pipe_sampler_view *view) {
847   for (int s = 0; s < 6; s++) {
848      for (int i = 0; i < nvc0->num_textures[s]; i++)
849         if (nvc0->textures[s][i] == view)
850            return true;
851   }
852   return false;
853}
854
855static void
856nve4_delete_texture_handle(struct pipe_context *pipe, uint64_t handle)
857{
858   struct nvc0_context *nvc0 = nvc0_context(pipe);
859   uint32_t tic = handle & NVE4_TIC_ENTRY_INVALID;
860   uint32_t tsc = (handle & NVE4_TSC_ENTRY_INVALID) >> 20;
861   struct nv50_tic_entry *entry = nvc0->screen->tic.entries[tic];
862
863   if (entry) {
864      struct pipe_sampler_view *view = &entry->pipe;
865      assert(entry->bindless);
866      p_atomic_dec(&entry->bindless);
867      if (!view_bound(nvc0, view))
868         nvc0_screen_tic_unlock(nvc0->screen, entry);
869      pipe_sampler_view_reference(&view, NULL);
870   }
871
872   pipe->delete_sampler_state(pipe, nvc0->screen->tsc.entries[tsc]);
873}
874
875static void
876nve4_make_texture_handle_resident(struct pipe_context *pipe,
877                                  uint64_t handle, bool resident)
878{
879   struct nvc0_context *nvc0 = nvc0_context(pipe);
880   if (resident) {
881      struct nvc0_resident *res = calloc(1, sizeof(struct nvc0_resident));
882      struct nv50_tic_entry *tic =
883         nvc0->screen->tic.entries[handle & NVE4_TIC_ENTRY_INVALID];
884      assert(tic);
885      assert(tic->bindless);
886
887      res->handle = handle;
888      res->buf = nv04_resource(tic->pipe.texture);
889      res->flags = NOUVEAU_BO_RD;
890      list_add(&res->list, &nvc0->tex_head);
891   } else {
892      list_for_each_entry_safe(struct nvc0_resident, pos, &nvc0->tex_head, list) {
893         if (pos->handle == handle) {
894            list_del(&pos->list);
895            free(pos);
896            break;
897         }
898      }
899   }
900}
901
902static const uint8_t nve4_su_format_map[PIPE_FORMAT_COUNT];
903static const uint16_t nve4_su_format_aux_map[PIPE_FORMAT_COUNT];
904static const uint16_t nve4_suldp_lib_offset[PIPE_FORMAT_COUNT];
905
906static void
907nvc0_get_surface_dims(const struct pipe_image_view *view,
908                      int *width, int *height, int *depth)
909{
910   struct nv04_resource *res = nv04_resource(view->resource);
911   int level;
912
913   *width = *height = *depth = 1;
914   if (res->base.target == PIPE_BUFFER) {
915      *width = view->u.buf.size / util_format_get_blocksize(view->format);
916      return;
917   }
918
919   level = view->u.tex.level;
920   *width = u_minify(view->resource->width0, level);
921   *height = u_minify(view->resource->height0, level);
922   *depth = u_minify(view->resource->depth0, level);
923
924   switch (res->base.target) {
925   case PIPE_TEXTURE_1D_ARRAY:
926   case PIPE_TEXTURE_2D_ARRAY:
927   case PIPE_TEXTURE_CUBE:
928   case PIPE_TEXTURE_CUBE_ARRAY:
929      *depth = view->u.tex.last_layer - view->u.tex.first_layer + 1;
930      break;
931   case PIPE_TEXTURE_1D:
932   case PIPE_TEXTURE_2D:
933   case PIPE_TEXTURE_RECT:
934   case PIPE_TEXTURE_3D:
935      break;
936   default:
937      assert(!"unexpected texture target");
938      break;
939   }
940}
941
942void
943nvc0_mark_image_range_valid(const struct pipe_image_view *view)
944{
945   struct nv04_resource *res = (struct nv04_resource *)view->resource;
946
947   assert(view->resource->target == PIPE_BUFFER);
948
949   util_range_add(&res->base, &res->valid_buffer_range,
950                  view->u.buf.offset,
951                  view->u.buf.offset + view->u.buf.size);
952}
953
954void
955nve4_set_surface_info(struct nouveau_pushbuf *push,
956                      const struct pipe_image_view *view,
957                      struct nvc0_context *nvc0)
958{
959   struct nvc0_screen *screen = nvc0->screen;
960   struct nv04_resource *res;
961   uint64_t address;
962   uint32_t *const info = push->cur;
963   int width, height, depth;
964   uint8_t log2cpp;
965
966   if (view && !nve4_su_format_map[view->format])
967      NOUVEAU_ERR("unsupported surface format, try is_format_supported() !\n");
968
969   push->cur += 16;
970
971   if (!view || !nve4_su_format_map[view->format]) {
972      memset(info, 0, 16 * sizeof(*info));
973
974      info[0] = 0xbadf0000;
975      info[1] = 0x80004000;
976      info[12] = nve4_suldp_lib_offset[PIPE_FORMAT_R32G32B32A32_UINT] +
977         screen->lib_code->start;
978      return;
979   }
980   res = nv04_resource(view->resource);
981
982   address = res->address;
983
984   /* get surface dimensions based on the target. */
985   nvc0_get_surface_dims(view, &width, &height, &depth);
986
987   info[8] = width;
988   info[9] = height;
989   info[10] = depth;
990   switch (res->base.target) {
991   case PIPE_TEXTURE_1D_ARRAY:
992      info[11] = 1;
993      break;
994   case PIPE_TEXTURE_2D:
995   case PIPE_TEXTURE_RECT:
996      info[11] = 2;
997      break;
998   case PIPE_TEXTURE_3D:
999      info[11] = 3;
1000      break;
1001   case PIPE_TEXTURE_2D_ARRAY:
1002   case PIPE_TEXTURE_CUBE:
1003   case PIPE_TEXTURE_CUBE_ARRAY:
1004      info[11] = 4;
1005      break;
1006   default:
1007      info[11] = 0;
1008      break;
1009   }
1010   log2cpp = (0xf000 & nve4_su_format_aux_map[view->format]) >> 12;
1011
1012   /* Stick the blockwidth (ie. number of bytes per pixel) to check if the
1013    * format doesn't mismatch. */
1014   info[12] = util_format_get_blocksize(view->format);
1015
1016   /* limit in bytes for raw access */
1017   info[13] = (0x06 << 22) | ((width << log2cpp) - 1);
1018
1019   info[1] = nve4_su_format_map[view->format];
1020
1021#if 0
1022   switch (util_format_get_blocksizebits(view->format)) {
1023   case  16: info[1] |= 1 << 16; break;
1024   case  32: info[1] |= 2 << 16; break;
1025   case  64: info[1] |= 3 << 16; break;
1026   case 128: info[1] |= 4 << 16; break;
1027   default:
1028      break;
1029   }
1030#else
1031   info[1] |= log2cpp << 16;
1032   info[1] |=  0x4000;
1033   info[1] |= (0x0f00 & nve4_su_format_aux_map[view->format]);
1034#endif
1035
1036   if (res->base.target == PIPE_BUFFER) {
1037      address += view->u.buf.offset;
1038
1039      info[0]  = address >> 8;
1040      info[2]  = width - 1;
1041      info[2] |= (0xff & nve4_su_format_aux_map[view->format]) << 22;
1042      info[3]  = 0;
1043      info[4]  = 0;
1044      info[5]  = 0;
1045      info[6]  = 0;
1046      info[7]  = 0;
1047      info[14] = 0;
1048      info[15] = 0;
1049   } else {
1050      struct nv50_miptree *mt = nv50_miptree(&res->base);
1051      struct nv50_miptree_level *lvl = &mt->level[view->u.tex.level];
1052      unsigned z = view->u.tex.first_layer;
1053
1054      if (!mt->layout_3d) {
1055         address += mt->layer_stride * z;
1056         z = 0;
1057      }
1058
1059      address += lvl->offset;
1060
1061      info[0]  = address >> 8;
1062      info[2]  = (width << mt->ms_x) - 1;
1063      /* NOTE: this is really important: */
1064      info[2] |= (0xff & nve4_su_format_aux_map[view->format]) << 22;
1065      info[3]  = (0x88 << 24) | (lvl->pitch / 64);
1066      info[4]  = (height << mt->ms_y) - 1;
1067      info[4] |= (lvl->tile_mode & 0x0f0) << 25;
1068      info[4] |= NVC0_TILE_SHIFT_Y(lvl->tile_mode) << 22;
1069      info[5]  = mt->layer_stride >> 8;
1070      info[6]  = depth - 1;
1071      info[6] |= (lvl->tile_mode & 0xf00) << 21;
1072      info[6] |= NVC0_TILE_SHIFT_Z(lvl->tile_mode) << 22;
1073      info[7]  = mt->layout_3d ? 1 : 0;
1074      info[7] |= z << 16;
1075      info[14] = mt->ms_x;
1076      info[15] = mt->ms_y;
1077   }
1078}
1079
1080static inline void
1081nvc0_set_surface_info(struct nouveau_pushbuf *push,
1082                      const struct pipe_image_view *view, uint64_t address,
1083                      int width, int height, int depth)
1084{
1085   struct nv04_resource *res;
1086   uint32_t *const info = push->cur;
1087
1088   push->cur += 16;
1089
1090   /* Make sure to always initialize the surface information area because it's
1091    * used to check if the given image is bound or not. */
1092   memset(info, 0, 16 * sizeof(*info));
1093
1094   if (!view || !view->resource)
1095      return;
1096   res = nv04_resource(view->resource);
1097
1098   /* Stick the image dimensions for the imageSize() builtin. */
1099   info[8] = width;
1100   info[9] = height;
1101   info[10] = depth;
1102
1103   /* Stick the blockwidth (ie. number of bytes per pixel) to calculate pixel
1104    * offset and to check if the format doesn't mismatch. */
1105   info[12] = ffs(util_format_get_blocksize(view->format)) - 1;
1106
1107   if (res->base.target == PIPE_BUFFER) {
1108      info[0]  = address >> 8;
1109      info[2]  = width;
1110   } else {
1111      struct nv50_miptree *mt = nv50_miptree(&res->base);
1112      struct nv50_miptree_level *lvl = &mt->level[view->u.tex.level];
1113      unsigned z = mt->layout_3d ? view->u.tex.first_layer : 0;
1114      unsigned nby = align(util_format_get_nblocksy(view->format, height),
1115                           NVC0_TILE_SIZE_Y(lvl->tile_mode));
1116
1117      /* NOTE: this does not precisely match nve4; the values are made to be
1118       * easier for the shader to consume.
1119       */
1120      info[0]  = address >> 8;
1121      info[2]  = (NVC0_TILE_SHIFT_X(lvl->tile_mode) - info[12]) << 24;
1122      info[4]  = NVC0_TILE_SHIFT_Y(lvl->tile_mode) << 24 | nby;
1123      info[5]  = mt->layer_stride >> 8;
1124      info[6]  = NVC0_TILE_SHIFT_Z(lvl->tile_mode) << 24;
1125      info[7]  = z;
1126      info[14] = mt->ms_x;
1127      info[15] = mt->ms_y;
1128   }
1129}
1130
1131void
1132nvc0_validate_suf(struct nvc0_context *nvc0, int s)
1133{
1134   struct nouveau_pushbuf *push = nvc0->base.pushbuf;
1135   struct nvc0_screen *screen = nvc0->screen;
1136
1137   for (int i = 0; i < NVC0_MAX_IMAGES; ++i) {
1138      struct pipe_image_view *view = &nvc0->images[s][i];
1139      int width, height, depth;
1140      uint64_t address = 0;
1141
1142      if (s == 5)
1143         BEGIN_NVC0(push, NVC0_CP(IMAGE(i)), 6);
1144      else
1145         BEGIN_NVC0(push, NVC0_3D(IMAGE(i)), 6);
1146
1147      if (view->resource) {
1148         struct nv04_resource *res = nv04_resource(view->resource);
1149         unsigned rt = nvc0_format_table[view->format].rt;
1150
1151         if (util_format_is_depth_or_stencil(view->format))
1152            rt = rt << 12;
1153         else
1154            rt = (rt << 4) | (0x14 << 12);
1155
1156         /* get surface dimensions based on the target. */
1157         nvc0_get_surface_dims(view, &width, &height, &depth);
1158
1159         address = res->address;
1160         if (res->base.target == PIPE_BUFFER) {
1161            unsigned blocksize = util_format_get_blocksize(view->format);
1162
1163            address += view->u.buf.offset;
1164            assert(!(address & 0xff));
1165
1166            if (view->access & PIPE_IMAGE_ACCESS_WRITE)
1167               nvc0_mark_image_range_valid(view);
1168
1169            PUSH_DATAh(push, address);
1170            PUSH_DATA (push, address);
1171            PUSH_DATA (push, align(width * blocksize, 0x100));
1172            PUSH_DATA (push, NVC0_3D_IMAGE_HEIGHT_LINEAR | 1);
1173            PUSH_DATA (push, rt);
1174            PUSH_DATA (push, 0);
1175         } else {
1176            struct nv50_miptree *mt = nv50_miptree(view->resource);
1177            struct nv50_miptree_level *lvl = &mt->level[view->u.tex.level];
1178            unsigned adjusted_width = width, adjusted_height = height;
1179
1180            if (mt->layout_3d) {
1181               // We have to adjust the size of the 3d surface to be
1182               // accessible within 2d limits. The size of each z tile goes
1183               // into the x direction, while the number of z tiles goes into
1184               // the y direction.
1185               const unsigned nbx = util_format_get_nblocksx(view->format, width);
1186               const unsigned nby = util_format_get_nblocksy(view->format, height);
1187               const unsigned tsx = NVC0_TILE_SIZE_X(lvl->tile_mode);
1188               const unsigned tsy = NVC0_TILE_SIZE_Y(lvl->tile_mode);
1189               const unsigned tsz = NVC0_TILE_SIZE_Z(lvl->tile_mode);
1190
1191               adjusted_width = align(nbx, tsx / util_format_get_blocksize(view->format)) * tsz;
1192               adjusted_height = align(nby, tsy) * align(depth, tsz) >> NVC0_TILE_SHIFT_Z(lvl->tile_mode);
1193            } else {
1194               const unsigned z = view->u.tex.first_layer;
1195               address += mt->layer_stride * z;
1196            }
1197            address += lvl->offset;
1198
1199            PUSH_DATAh(push, address);
1200            PUSH_DATA (push, address);
1201            PUSH_DATA (push, adjusted_width << mt->ms_x);
1202            PUSH_DATA (push, adjusted_height << mt->ms_y);
1203            PUSH_DATA (push, rt);
1204            PUSH_DATA (push, lvl->tile_mode & 0xff); /* mask out z-tiling */
1205         }
1206
1207         if (s == 5)
1208            BCTX_REFN(nvc0->bufctx_cp, CP_SUF, res, RDWR);
1209         else
1210            BCTX_REFN(nvc0->bufctx_3d, 3D_SUF, res, RDWR);
1211      } else {
1212         PUSH_DATA(push, 0);
1213         PUSH_DATA(push, 0);
1214         PUSH_DATA(push, 0);
1215         PUSH_DATA(push, 0);
1216         PUSH_DATA(push, 0x14000);
1217         PUSH_DATA(push, 0);
1218      }
1219
1220      /* stick surface information into the driver constant buffer */
1221      if (s == 5)
1222         BEGIN_NVC0(push, NVC0_CP(CB_SIZE), 3);
1223      else
1224         BEGIN_NVC0(push, NVC0_3D(CB_SIZE), 3);
1225      PUSH_DATA (push, NVC0_CB_AUX_SIZE);
1226      PUSH_DATAh(push, screen->uniform_bo->offset + NVC0_CB_AUX_INFO(s));
1227      PUSH_DATA (push, screen->uniform_bo->offset + NVC0_CB_AUX_INFO(s));
1228      if (s == 5)
1229         BEGIN_1IC0(push, NVC0_CP(CB_POS), 1 + 16);
1230      else
1231         BEGIN_1IC0(push, NVC0_3D(CB_POS), 1 + 16);
1232      PUSH_DATA (push, NVC0_CB_AUX_SU_INFO(i));
1233
1234      nvc0_set_surface_info(push, view, address, width, height, depth);
1235   }
1236}
1237
1238static inline void
1239nvc0_update_surface_bindings(struct nvc0_context *nvc0)
1240{
1241   nvc0_validate_suf(nvc0, 4);
1242
1243   /* Invalidate all COMPUTE images because they are aliased with FRAGMENT. */
1244   nouveau_bufctx_reset(nvc0->bufctx_cp, NVC0_BIND_CP_SUF);
1245   nvc0->dirty_cp |= NVC0_NEW_CP_SURFACES;
1246   nvc0->images_dirty[5] |= nvc0->images_valid[5];
1247}
1248
1249static void
1250gm107_validate_surfaces(struct nvc0_context *nvc0,
1251                        struct pipe_image_view *view, int stage, int slot)
1252{
1253   struct nv04_resource *res = nv04_resource(view->resource);
1254   struct nouveau_pushbuf *push = nvc0->base.pushbuf;
1255   struct nvc0_screen *screen = nvc0->screen;
1256   struct nv50_tic_entry *tic;
1257
1258   tic = nv50_tic_entry(nvc0->images_tic[stage][slot]);
1259
1260   res = nv04_resource(tic->pipe.texture);
1261   nvc0_update_tic(nvc0, tic, res);
1262
1263   if (tic->id < 0) {
1264      tic->id = nvc0_screen_tic_alloc(nvc0->screen, tic);
1265
1266      /* upload the texture view */
1267      nve4_p2mf_push_linear(&nvc0->base, nvc0->screen->txc, tic->id * 32,
1268                            NV_VRAM_DOMAIN(&nvc0->screen->base), 32, tic->tic);
1269
1270      BEGIN_NVC0(push, NVC0_3D(TIC_FLUSH), 1);
1271      PUSH_DATA (push, 0);
1272   } else
1273   if (res->status & NOUVEAU_BUFFER_STATUS_GPU_WRITING) {
1274      BEGIN_NVC0(push, NVC0_3D(TEX_CACHE_CTL), 1);
1275      PUSH_DATA (push, (tic->id << 4) | 1);
1276   }
1277   nvc0->screen->tic.lock[tic->id / 32] |= 1 << (tic->id % 32);
1278
1279   res->status &= ~NOUVEAU_BUFFER_STATUS_GPU_WRITING;
1280   res->status |= NOUVEAU_BUFFER_STATUS_GPU_READING;
1281
1282   BCTX_REFN(nvc0->bufctx_3d, 3D_SUF, res, RD);
1283
1284   /* upload the texture handle */
1285   BEGIN_NVC0(push, NVC0_3D(CB_SIZE), 3);
1286   PUSH_DATA (push, NVC0_CB_AUX_SIZE);
1287   PUSH_DATAh(push, screen->uniform_bo->offset + NVC0_CB_AUX_INFO(stage));
1288   PUSH_DATA (push, screen->uniform_bo->offset + NVC0_CB_AUX_INFO(stage));
1289   BEGIN_NVC0(push, NVC0_3D(CB_POS), 2);
1290   PUSH_DATA (push, NVC0_CB_AUX_TEX_INFO(slot + 32));
1291   PUSH_DATA (push, tic->id);
1292}
1293
1294static inline void
1295nve4_update_surface_bindings(struct nvc0_context *nvc0)
1296{
1297   struct nouveau_pushbuf *push = nvc0->base.pushbuf;
1298   struct nvc0_screen *screen = nvc0->screen;
1299   int i, j, s;
1300
1301   for (s = 0; s < 5; s++) {
1302      if (!nvc0->images_dirty[s])
1303         continue;
1304
1305      for (i = 0; i < NVC0_MAX_IMAGES; ++i) {
1306         struct pipe_image_view *view = &nvc0->images[s][i];
1307
1308         BEGIN_NVC0(push, NVC0_3D(CB_SIZE), 3);
1309         PUSH_DATA (push, NVC0_CB_AUX_SIZE);
1310         PUSH_DATAh(push, screen->uniform_bo->offset + NVC0_CB_AUX_INFO(s));
1311         PUSH_DATA (push, screen->uniform_bo->offset + NVC0_CB_AUX_INFO(s));
1312         BEGIN_1IC0(push, NVC0_3D(CB_POS), 1 + 16);
1313         PUSH_DATA (push, NVC0_CB_AUX_SU_INFO(i));
1314
1315         if (view->resource) {
1316            struct nv04_resource *res = nv04_resource(view->resource);
1317
1318            if (res->base.target == PIPE_BUFFER) {
1319               if (view->access & PIPE_IMAGE_ACCESS_WRITE)
1320                  nvc0_mark_image_range_valid(view);
1321            }
1322
1323            nve4_set_surface_info(push, view, nvc0);
1324            BCTX_REFN(nvc0->bufctx_3d, 3D_SUF, res, RDWR);
1325
1326            if (nvc0->screen->base.class_3d >= GM107_3D_CLASS)
1327               gm107_validate_surfaces(nvc0, view, s, i);
1328         } else {
1329            for (j = 0; j < 16; j++)
1330               PUSH_DATA(push, 0);
1331         }
1332      }
1333   }
1334}
1335
1336void
1337nvc0_validate_surfaces(struct nvc0_context *nvc0)
1338{
1339   if (nvc0->screen->base.class_3d >= NVE4_3D_CLASS) {
1340      nve4_update_surface_bindings(nvc0);
1341   } else {
1342      nvc0_update_surface_bindings(nvc0);
1343   }
1344}
1345
1346static uint64_t
1347nve4_create_image_handle(struct pipe_context *pipe,
1348                         const struct pipe_image_view *view)
1349{
1350   struct nvc0_context *nvc0 = nvc0_context(pipe);
1351   struct nouveau_pushbuf *push = nvc0->base.pushbuf;
1352   struct nvc0_screen *screen = nvc0->screen;
1353   int i = screen->img.next, s;
1354
1355   while (screen->img.entries[i]) {
1356      i = (i + 1) & (NVE4_IMG_MAX_HANDLES - 1);
1357      if (i == screen->img.next)
1358         return 0;
1359   }
1360
1361   screen->img.next = (i + 1) & (NVE4_IMG_MAX_HANDLES - 1);
1362   screen->img.entries[i] = calloc(1, sizeof(struct pipe_image_view));
1363   *screen->img.entries[i] = *view;
1364
1365   for (s = 0; s < 6; s++) {
1366      BEGIN_NVC0(push, NVC0_3D(CB_SIZE), 3);
1367      PUSH_DATA (push, NVC0_CB_AUX_SIZE);
1368      PUSH_DATAh(push, screen->uniform_bo->offset + NVC0_CB_AUX_INFO(s));
1369      PUSH_DATA (push, screen->uniform_bo->offset + NVC0_CB_AUX_INFO(s));
1370      BEGIN_1IC0(push, NVC0_3D(CB_POS), 1 + 16);
1371      PUSH_DATA (push, NVC0_CB_AUX_BINDLESS_INFO(i));
1372      nve4_set_surface_info(push, view, nvc0);
1373   }
1374
1375   return 0x100000000ULL | i;
1376}
1377
1378static void
1379nve4_delete_image_handle(struct pipe_context *pipe, uint64_t handle)
1380{
1381   struct nvc0_context *nvc0 = nvc0_context(pipe);
1382   struct nvc0_screen *screen = nvc0->screen;
1383   int i = handle & (NVE4_IMG_MAX_HANDLES - 1);
1384
1385   free(screen->img.entries[i]);
1386   screen->img.entries[i] = NULL;
1387}
1388
1389static void
1390nve4_make_image_handle_resident(struct pipe_context *pipe, uint64_t handle,
1391                                unsigned access, bool resident)
1392{
1393   struct nvc0_context *nvc0 = nvc0_context(pipe);
1394   struct nvc0_screen *screen = nvc0->screen;
1395
1396   if (resident) {
1397      struct nvc0_resident *res = calloc(1, sizeof(struct nvc0_resident));
1398      struct pipe_image_view *view =
1399         screen->img.entries[handle & (NVE4_IMG_MAX_HANDLES - 1)];
1400      assert(view);
1401
1402      if (view->resource->target == PIPE_BUFFER &&
1403          access & PIPE_IMAGE_ACCESS_WRITE)
1404         nvc0_mark_image_range_valid(view);
1405      res->handle = handle;
1406      res->buf = nv04_resource(view->resource);
1407      res->flags = (access & 3) << 8;
1408      list_add(&res->list, &nvc0->img_head);
1409   } else {
1410      list_for_each_entry_safe(struct nvc0_resident, pos, &nvc0->img_head, list) {
1411         if (pos->handle == handle) {
1412            list_del(&pos->list);
1413            free(pos);
1414            break;
1415         }
1416      }
1417   }
1418}
1419
1420static uint64_t
1421gm107_create_image_handle(struct pipe_context *pipe,
1422                          const struct pipe_image_view *view)
1423{
1424   /* GM107+ use TIC handles to reference images. As such, image handles are
1425    * just the TIC id.
1426    */
1427   struct nvc0_context *nvc0 = nvc0_context(pipe);
1428   struct nouveau_pushbuf *push = nvc0->base.pushbuf;
1429   struct pipe_sampler_view *sview =
1430      gm107_create_texture_view_from_image(pipe, view);
1431   struct nv50_tic_entry *tic = nv50_tic_entry(sview);
1432
1433   if (tic == NULL)
1434      goto fail;
1435
1436   tic->bindless = 1;
1437   tic->id = nvc0_screen_tic_alloc(nvc0->screen, tic);
1438   if (tic->id < 0)
1439      goto fail;
1440
1441   nve4_p2mf_push_linear(&nvc0->base, nvc0->screen->txc, tic->id * 32,
1442                         NV_VRAM_DOMAIN(&nvc0->screen->base), 32,
1443                         tic->tic);
1444
1445   IMMED_NVC0(push, NVC0_3D(TIC_FLUSH), 0);
1446
1447   nvc0->screen->tic.lock[tic->id / 32] |= 1 << (tic->id % 32);
1448
1449   // Compute handle. This will include the TIC as well as some additional
1450   // info regarding the bound 3d surface layer, if applicable.
1451   uint64_t handle = 0x100000000ULL | tic->id;
1452   struct nv04_resource *res = nv04_resource(view->resource);
1453   if (res->base.target == PIPE_TEXTURE_3D) {
1454      handle |= 1 << 11;
1455      handle |= view->u.tex.first_layer << (11 + 16);
1456   }
1457   return handle;
1458
1459fail:
1460   FREE(tic);
1461   return 0;
1462}
1463
1464static void
1465gm107_delete_image_handle(struct pipe_context *pipe, uint64_t handle)
1466{
1467   struct nvc0_context *nvc0 = nvc0_context(pipe);
1468   int tic = handle & NVE4_TIC_ENTRY_INVALID;
1469   struct nv50_tic_entry *entry = nvc0->screen->tic.entries[tic];
1470   struct pipe_sampler_view *view = &entry->pipe;
1471   assert(entry->bindless == 1);
1472   assert(!view_bound(nvc0, view));
1473   entry->bindless = 0;
1474   nvc0_screen_tic_unlock(nvc0->screen, entry);
1475   pipe_sampler_view_reference(&view, NULL);
1476}
1477
1478static void
1479gm107_make_image_handle_resident(struct pipe_context *pipe, uint64_t handle,
1480                                 unsigned access, bool resident)
1481{
1482   struct nvc0_context *nvc0 = nvc0_context(pipe);
1483
1484   if (resident) {
1485      struct nvc0_resident *res = calloc(1, sizeof(struct nvc0_resident));
1486      struct nv50_tic_entry *tic =
1487         nvc0->screen->tic.entries[handle & NVE4_TIC_ENTRY_INVALID];
1488      assert(tic);
1489      assert(tic->bindless);
1490
1491      res->handle = handle;
1492      res->buf = nv04_resource(tic->pipe.texture);
1493      res->flags = (access & 3) << 8;
1494      if (res->buf->base.target == PIPE_BUFFER &&
1495          access & PIPE_IMAGE_ACCESS_WRITE)
1496         util_range_add(&res->buf->base, &res->buf->valid_buffer_range,
1497                        tic->pipe.u.buf.offset,
1498                        tic->pipe.u.buf.offset + tic->pipe.u.buf.size);
1499      list_add(&res->list, &nvc0->img_head);
1500   } else {
1501      list_for_each_entry_safe(struct nvc0_resident, pos, &nvc0->img_head, list) {
1502         if (pos->handle == handle) {
1503            list_del(&pos->list);
1504            free(pos);
1505            break;
1506         }
1507      }
1508   }
1509}
1510
1511void
1512nvc0_init_bindless_functions(struct pipe_context *pipe) {
1513   pipe->create_texture_handle = nve4_create_texture_handle;
1514   pipe->delete_texture_handle = nve4_delete_texture_handle;
1515   pipe->make_texture_handle_resident = nve4_make_texture_handle_resident;
1516
1517   if (nvc0_context(pipe)->screen->base.class_3d < GM107_3D_CLASS) {
1518      pipe->create_image_handle = nve4_create_image_handle;
1519      pipe->delete_image_handle = nve4_delete_image_handle;
1520      pipe->make_image_handle_resident = nve4_make_image_handle_resident;
1521   } else {
1522      pipe->create_image_handle = gm107_create_image_handle;
1523      pipe->delete_image_handle = gm107_delete_image_handle;
1524      pipe->make_image_handle_resident = gm107_make_image_handle_resident;
1525   }
1526}
1527
1528
1529static const uint8_t nve4_su_format_map[PIPE_FORMAT_COUNT] =
1530{
1531   [PIPE_FORMAT_R32G32B32A32_FLOAT] = GK104_IMAGE_FORMAT_RGBA32_FLOAT,
1532   [PIPE_FORMAT_R32G32B32A32_SINT] = GK104_IMAGE_FORMAT_RGBA32_SINT,
1533   [PIPE_FORMAT_R32G32B32A32_UINT] = GK104_IMAGE_FORMAT_RGBA32_UINT,
1534   [PIPE_FORMAT_R16G16B16A16_FLOAT] = GK104_IMAGE_FORMAT_RGBA16_FLOAT,
1535   [PIPE_FORMAT_R16G16B16A16_UNORM] = GK104_IMAGE_FORMAT_RGBA16_UNORM,
1536   [PIPE_FORMAT_R16G16B16A16_SNORM] = GK104_IMAGE_FORMAT_RGBA16_SNORM,
1537   [PIPE_FORMAT_R16G16B16A16_SINT] = GK104_IMAGE_FORMAT_RGBA16_SINT,
1538   [PIPE_FORMAT_R16G16B16A16_UINT] = GK104_IMAGE_FORMAT_RGBA16_UINT,
1539   [PIPE_FORMAT_B8G8R8A8_UNORM] = GK104_IMAGE_FORMAT_BGRA8_UNORM,
1540   [PIPE_FORMAT_R8G8B8A8_UNORM] = GK104_IMAGE_FORMAT_RGBA8_UNORM,
1541   [PIPE_FORMAT_R8G8B8A8_SNORM] = GK104_IMAGE_FORMAT_RGBA8_SNORM,
1542   [PIPE_FORMAT_R8G8B8A8_SINT] = GK104_IMAGE_FORMAT_RGBA8_SINT,
1543   [PIPE_FORMAT_R8G8B8A8_UINT] = GK104_IMAGE_FORMAT_RGBA8_UINT,
1544   [PIPE_FORMAT_R11G11B10_FLOAT] = GK104_IMAGE_FORMAT_R11G11B10_FLOAT,
1545   [PIPE_FORMAT_R10G10B10A2_UNORM] = GK104_IMAGE_FORMAT_RGB10_A2_UNORM,
1546   [PIPE_FORMAT_R10G10B10A2_UINT] = GK104_IMAGE_FORMAT_RGB10_A2_UINT,
1547   [PIPE_FORMAT_R32G32_FLOAT] = GK104_IMAGE_FORMAT_RG32_FLOAT,
1548   [PIPE_FORMAT_R32G32_SINT] = GK104_IMAGE_FORMAT_RG32_SINT,
1549   [PIPE_FORMAT_R32G32_UINT] = GK104_IMAGE_FORMAT_RG32_UINT,
1550   [PIPE_FORMAT_R16G16_FLOAT] = GK104_IMAGE_FORMAT_RG16_FLOAT,
1551   [PIPE_FORMAT_R16G16_UNORM] = GK104_IMAGE_FORMAT_RG16_UNORM,
1552   [PIPE_FORMAT_R16G16_SNORM] = GK104_IMAGE_FORMAT_RG16_SNORM,
1553   [PIPE_FORMAT_R16G16_SINT] = GK104_IMAGE_FORMAT_RG16_SINT,
1554   [PIPE_FORMAT_R16G16_UINT] = GK104_IMAGE_FORMAT_RG16_UINT,
1555   [PIPE_FORMAT_R8G8_UNORM] = GK104_IMAGE_FORMAT_RG8_UNORM,
1556   [PIPE_FORMAT_R8G8_SNORM] = GK104_IMAGE_FORMAT_RG8_SNORM,
1557   [PIPE_FORMAT_R8G8_SINT] = GK104_IMAGE_FORMAT_RG8_SINT,
1558   [PIPE_FORMAT_R8G8_UINT] = GK104_IMAGE_FORMAT_RG8_UINT,
1559   [PIPE_FORMAT_R32_FLOAT] = GK104_IMAGE_FORMAT_R32_FLOAT,
1560   [PIPE_FORMAT_R32_SINT] = GK104_IMAGE_FORMAT_R32_SINT,
1561   [PIPE_FORMAT_R32_UINT] = GK104_IMAGE_FORMAT_R32_UINT,
1562   [PIPE_FORMAT_R16_FLOAT] = GK104_IMAGE_FORMAT_R16_FLOAT,
1563   [PIPE_FORMAT_R16_UNORM] = GK104_IMAGE_FORMAT_R16_UNORM,
1564   [PIPE_FORMAT_R16_SNORM] = GK104_IMAGE_FORMAT_R16_SNORM,
1565   [PIPE_FORMAT_R16_SINT] = GK104_IMAGE_FORMAT_R16_SINT,
1566   [PIPE_FORMAT_R16_UINT] = GK104_IMAGE_FORMAT_R16_UINT,
1567   [PIPE_FORMAT_R8_UNORM] = GK104_IMAGE_FORMAT_R8_UNORM,
1568   [PIPE_FORMAT_R8_SNORM] = GK104_IMAGE_FORMAT_R8_SNORM,
1569   [PIPE_FORMAT_R8_SINT] = GK104_IMAGE_FORMAT_R8_SINT,
1570   [PIPE_FORMAT_R8_UINT] = GK104_IMAGE_FORMAT_R8_UINT,
1571};
1572
1573/* Auxiliary format description values for surface instructions.
1574 * (log2(bytes per pixel) << 12) | (unk8 << 8) | unk22
1575 */
1576static const uint16_t nve4_su_format_aux_map[PIPE_FORMAT_COUNT] =
1577{
1578   [PIPE_FORMAT_R32G32B32A32_FLOAT] = 0x4842,
1579   [PIPE_FORMAT_R32G32B32A32_SINT] = 0x4842,
1580   [PIPE_FORMAT_R32G32B32A32_UINT] = 0x4842,
1581
1582   [PIPE_FORMAT_R16G16B16A16_UNORM] = 0x3933,
1583   [PIPE_FORMAT_R16G16B16A16_SNORM] = 0x3933,
1584   [PIPE_FORMAT_R16G16B16A16_SINT] = 0x3933,
1585   [PIPE_FORMAT_R16G16B16A16_UINT] = 0x3933,
1586   [PIPE_FORMAT_R16G16B16A16_FLOAT] = 0x3933,
1587
1588   [PIPE_FORMAT_R32G32_FLOAT] = 0x3433,
1589   [PIPE_FORMAT_R32G32_SINT] = 0x3433,
1590   [PIPE_FORMAT_R32G32_UINT] = 0x3433,
1591
1592   [PIPE_FORMAT_R10G10B10A2_UNORM] = 0x2a24,
1593   [PIPE_FORMAT_R10G10B10A2_UINT] = 0x2a24,
1594   [PIPE_FORMAT_B8G8R8A8_UNORM] = 0x2a24,
1595   [PIPE_FORMAT_R8G8B8A8_UNORM] = 0x2a24,
1596   [PIPE_FORMAT_R8G8B8A8_SNORM] = 0x2a24,
1597   [PIPE_FORMAT_R8G8B8A8_SINT] = 0x2a24,
1598   [PIPE_FORMAT_R8G8B8A8_UINT] = 0x2a24,
1599   [PIPE_FORMAT_R11G11B10_FLOAT] = 0x2a24,
1600
1601   [PIPE_FORMAT_R16G16_UNORM] = 0x2524,
1602   [PIPE_FORMAT_R16G16_SNORM] = 0x2524,
1603   [PIPE_FORMAT_R16G16_SINT] = 0x2524,
1604   [PIPE_FORMAT_R16G16_UINT] = 0x2524,
1605   [PIPE_FORMAT_R16G16_FLOAT] = 0x2524,
1606
1607   [PIPE_FORMAT_R32_SINT] = 0x2024,
1608   [PIPE_FORMAT_R32_UINT] = 0x2024,
1609   [PIPE_FORMAT_R32_FLOAT] = 0x2024,
1610
1611   [PIPE_FORMAT_R8G8_UNORM] = 0x1615,
1612   [PIPE_FORMAT_R8G8_SNORM] = 0x1615,
1613   [PIPE_FORMAT_R8G8_SINT] = 0x1615,
1614   [PIPE_FORMAT_R8G8_UINT] = 0x1615,
1615
1616   [PIPE_FORMAT_R16_UNORM] = 0x1115,
1617   [PIPE_FORMAT_R16_SNORM] = 0x1115,
1618   [PIPE_FORMAT_R16_SINT] = 0x1115,
1619   [PIPE_FORMAT_R16_UINT] = 0x1115,
1620   [PIPE_FORMAT_R16_FLOAT] = 0x1115,
1621
1622   [PIPE_FORMAT_R8_UNORM] = 0x0206,
1623   [PIPE_FORMAT_R8_SNORM] = 0x0206,
1624   [PIPE_FORMAT_R8_SINT] = 0x0206,
1625   [PIPE_FORMAT_R8_UINT] = 0x0206
1626};
1627
1628/* NOTE: These are hardcoded offsets for the shader library.
1629 * TODO: Automate them.
1630 */
1631static const uint16_t nve4_suldp_lib_offset[PIPE_FORMAT_COUNT] =
1632{
1633   [PIPE_FORMAT_R32G32B32A32_FLOAT] = 0x218,
1634   [PIPE_FORMAT_R32G32B32A32_SINT]  = 0x218,
1635   [PIPE_FORMAT_R32G32B32A32_UINT]  = 0x218,
1636   [PIPE_FORMAT_R16G16B16A16_UNORM] = 0x248,
1637   [PIPE_FORMAT_R16G16B16A16_SNORM] = 0x2b8,
1638   [PIPE_FORMAT_R16G16B16A16_SINT]  = 0x330,
1639   [PIPE_FORMAT_R16G16B16A16_UINT]  = 0x388,
1640   [PIPE_FORMAT_R16G16B16A16_FLOAT] = 0x3d8,
1641   [PIPE_FORMAT_R32G32_FLOAT]       = 0x428,
1642   [PIPE_FORMAT_R32G32_SINT]        = 0x468,
1643   [PIPE_FORMAT_R32G32_UINT]        = 0x468,
1644   [PIPE_FORMAT_R10G10B10A2_UNORM]  = 0x4a8,
1645   [PIPE_FORMAT_R10G10B10A2_UINT]   = 0x530,
1646   [PIPE_FORMAT_R8G8B8A8_UNORM]     = 0x588,
1647   [PIPE_FORMAT_R8G8B8A8_SNORM]     = 0x5f8,
1648   [PIPE_FORMAT_R8G8B8A8_SINT]      = 0x670,
1649   [PIPE_FORMAT_R8G8B8A8_UINT]      = 0x6c8,
1650   [PIPE_FORMAT_B5G6R5_UNORM]       = 0x718,
1651   [PIPE_FORMAT_B5G5R5X1_UNORM]     = 0x7a0,
1652   [PIPE_FORMAT_R16G16_UNORM]       = 0x828,
1653   [PIPE_FORMAT_R16G16_SNORM]       = 0x890,
1654   [PIPE_FORMAT_R16G16_SINT]        = 0x8f0,
1655   [PIPE_FORMAT_R16G16_UINT]        = 0x948,
1656   [PIPE_FORMAT_R16G16_FLOAT]       = 0x998,
1657   [PIPE_FORMAT_R32_FLOAT]          = 0x9e8,
1658   [PIPE_FORMAT_R32_SINT]           = 0xa30,
1659   [PIPE_FORMAT_R32_UINT]           = 0xa30,
1660   [PIPE_FORMAT_R8G8_UNORM]         = 0xa78,
1661   [PIPE_FORMAT_R8G8_SNORM]         = 0xae0,
1662   [PIPE_FORMAT_R8G8_UINT]          = 0xb48,
1663   [PIPE_FORMAT_R8G8_SINT]          = 0xb98,
1664   [PIPE_FORMAT_R16_UNORM]          = 0xbe8,
1665   [PIPE_FORMAT_R16_SNORM]          = 0xc48,
1666   [PIPE_FORMAT_R16_SINT]           = 0xca0,
1667   [PIPE_FORMAT_R16_UINT]           = 0xce8,
1668   [PIPE_FORMAT_R16_FLOAT]          = 0xd30,
1669   [PIPE_FORMAT_R8_UNORM]           = 0xd88,
1670   [PIPE_FORMAT_R8_SNORM]           = 0xde0,
1671   [PIPE_FORMAT_R8_SINT]            = 0xe38,
1672   [PIPE_FORMAT_R8_UINT]            = 0xe88,
1673   [PIPE_FORMAT_R11G11B10_FLOAT]    = 0xed0
1674};
1675