1/* 2 * Copyright 2012 Advanced Micro Devices, Inc. 3 * 4 * Permission is hereby granted, free of charge, to any person obtaining a 5 * copy of this software and associated documentation files (the "Software"), 6 * to deal in the Software without restriction, including without limitation 7 * the rights to use, copy, modify, merge, publish, distribute, sublicense, 8 * and/or sell copies of the Software, and to permit persons to whom the 9 * Software is furnished to do so, subject to the following conditions: 10 * 11 * The above copyright notice and this permission notice (including the next 12 * paragraph) shall be included in all copies or substantial portions of the 13 * Software. 14 * 15 * THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR 16 * IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY, 17 * FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. IN NO EVENT SHALL 18 * THE AUTHORS OR COPYRIGHT HOLDERS BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER 19 * LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING 20 * FROM, OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS 21 * IN THE SOFTWARE. 22 */ 23 24#include "ac_shader_util.h" 25#include "ac_gpu_info.h" 26 27#include "sid.h" 28#include "u_math.h" 29 30#include <assert.h> 31#include <stdlib.h> 32#include <string.h> 33 34unsigned ac_get_spi_shader_z_format(bool writes_z, bool writes_stencil, bool writes_samplemask) 35{ 36 if (writes_z) { 37 /* Z needs 32 bits. */ 38 if (writes_samplemask) 39 return V_028710_SPI_SHADER_32_ABGR; 40 else if (writes_stencil) 41 return V_028710_SPI_SHADER_32_GR; 42 else 43 return V_028710_SPI_SHADER_32_R; 44 } else if (writes_stencil || writes_samplemask) { 45 /* Both stencil and sample mask need only 16 bits. */ 46 return V_028710_SPI_SHADER_UINT16_ABGR; 47 } else { 48 return V_028710_SPI_SHADER_ZERO; 49 } 50} 51 52unsigned ac_get_cb_shader_mask(unsigned spi_shader_col_format) 53{ 54 unsigned i, cb_shader_mask = 0; 55 56 for (i = 0; i < 8; i++) { 57 switch ((spi_shader_col_format >> (i * 4)) & 0xf) { 58 case V_028714_SPI_SHADER_ZERO: 59 break; 60 case V_028714_SPI_SHADER_32_R: 61 cb_shader_mask |= 0x1 << (i * 4); 62 break; 63 case V_028714_SPI_SHADER_32_GR: 64 cb_shader_mask |= 0x3 << (i * 4); 65 break; 66 case V_028714_SPI_SHADER_32_AR: 67 cb_shader_mask |= 0x9u << (i * 4); 68 break; 69 case V_028714_SPI_SHADER_FP16_ABGR: 70 case V_028714_SPI_SHADER_UNORM16_ABGR: 71 case V_028714_SPI_SHADER_SNORM16_ABGR: 72 case V_028714_SPI_SHADER_UINT16_ABGR: 73 case V_028714_SPI_SHADER_SINT16_ABGR: 74 case V_028714_SPI_SHADER_32_ABGR: 75 cb_shader_mask |= 0xfu << (i * 4); 76 break; 77 default: 78 assert(0); 79 } 80 } 81 return cb_shader_mask; 82} 83 84/** 85 * Calculate the appropriate setting of VGT_GS_MODE when \p shader is a 86 * geometry shader. 87 */ 88uint32_t ac_vgt_gs_mode(unsigned gs_max_vert_out, enum chip_class chip_class) 89{ 90 unsigned cut_mode; 91 92 if (gs_max_vert_out <= 128) { 93 cut_mode = V_028A40_GS_CUT_128; 94 } else if (gs_max_vert_out <= 256) { 95 cut_mode = V_028A40_GS_CUT_256; 96 } else if (gs_max_vert_out <= 512) { 97 cut_mode = V_028A40_GS_CUT_512; 98 } else { 99 assert(gs_max_vert_out <= 1024); 100 cut_mode = V_028A40_GS_CUT_1024; 101 } 102 103 return S_028A40_MODE(V_028A40_GS_SCENARIO_G) | S_028A40_CUT_MODE(cut_mode) | 104 S_028A40_ES_WRITE_OPTIMIZE(chip_class <= GFX8) | S_028A40_GS_WRITE_OPTIMIZE(1) | 105 S_028A40_ONCHIP(chip_class >= GFX9 ? 1 : 0); 106} 107 108/// Translate a (dfmt, nfmt) pair into a chip-appropriate combined format 109/// value for LLVM8+ tbuffer intrinsics. 110unsigned ac_get_tbuffer_format(enum chip_class chip_class, unsigned dfmt, unsigned nfmt) 111{ 112 // Some games try to access vertex buffers without a valid format. 113 // This is a game bug, but we should still handle it gracefully. 114 if (dfmt == V_008F0C_GFX10_FORMAT_INVALID) 115 return V_008F0C_GFX10_FORMAT_INVALID; 116 117 if (chip_class >= GFX10) { 118 unsigned format; 119 switch (dfmt) { 120 default: 121 unreachable("bad dfmt"); 122 case V_008F0C_BUF_DATA_FORMAT_INVALID: 123 format = V_008F0C_GFX10_FORMAT_INVALID; 124 break; 125 case V_008F0C_BUF_DATA_FORMAT_8: 126 format = V_008F0C_GFX10_FORMAT_8_UINT; 127 break; 128 case V_008F0C_BUF_DATA_FORMAT_8_8: 129 format = V_008F0C_GFX10_FORMAT_8_8_UINT; 130 break; 131 case V_008F0C_BUF_DATA_FORMAT_8_8_8_8: 132 format = V_008F0C_GFX10_FORMAT_8_8_8_8_UINT; 133 break; 134 case V_008F0C_BUF_DATA_FORMAT_16: 135 format = V_008F0C_GFX10_FORMAT_16_UINT; 136 break; 137 case V_008F0C_BUF_DATA_FORMAT_16_16: 138 format = V_008F0C_GFX10_FORMAT_16_16_UINT; 139 break; 140 case V_008F0C_BUF_DATA_FORMAT_16_16_16_16: 141 format = V_008F0C_GFX10_FORMAT_16_16_16_16_UINT; 142 break; 143 case V_008F0C_BUF_DATA_FORMAT_32: 144 format = V_008F0C_GFX10_FORMAT_32_UINT; 145 break; 146 case V_008F0C_BUF_DATA_FORMAT_32_32: 147 format = V_008F0C_GFX10_FORMAT_32_32_UINT; 148 break; 149 case V_008F0C_BUF_DATA_FORMAT_32_32_32: 150 format = V_008F0C_GFX10_FORMAT_32_32_32_UINT; 151 break; 152 case V_008F0C_BUF_DATA_FORMAT_32_32_32_32: 153 format = V_008F0C_GFX10_FORMAT_32_32_32_32_UINT; 154 break; 155 case V_008F0C_BUF_DATA_FORMAT_2_10_10_10: 156 format = V_008F0C_GFX10_FORMAT_2_10_10_10_UINT; 157 break; 158 case V_008F0C_BUF_DATA_FORMAT_10_11_11: 159 format = V_008F0C_GFX10_FORMAT_10_11_11_UINT; 160 break; 161 } 162 163 // Use the regularity properties of the combined format enum. 164 // 165 // Note: float is incompatible with 8-bit data formats, 166 // [us]{norm,scaled} are incomparible with 32-bit data formats. 167 // [us]scaled are not writable. 168 switch (nfmt) { 169 case V_008F0C_BUF_NUM_FORMAT_UNORM: 170 format -= 4; 171 break; 172 case V_008F0C_BUF_NUM_FORMAT_SNORM: 173 format -= 3; 174 break; 175 case V_008F0C_BUF_NUM_FORMAT_USCALED: 176 format -= 2; 177 break; 178 case V_008F0C_BUF_NUM_FORMAT_SSCALED: 179 format -= 1; 180 break; 181 default: 182 unreachable("bad nfmt"); 183 case V_008F0C_BUF_NUM_FORMAT_UINT: 184 break; 185 case V_008F0C_BUF_NUM_FORMAT_SINT: 186 format += 1; 187 break; 188 case V_008F0C_BUF_NUM_FORMAT_FLOAT: 189 format += 2; 190 break; 191 } 192 193 return format; 194 } else { 195 return dfmt | (nfmt << 4); 196 } 197} 198 199static const struct ac_data_format_info data_format_table[] = { 200 [V_008F0C_BUF_DATA_FORMAT_INVALID] = {0, 4, 0, V_008F0C_BUF_DATA_FORMAT_INVALID}, 201 [V_008F0C_BUF_DATA_FORMAT_8] = {1, 1, 1, V_008F0C_BUF_DATA_FORMAT_8}, 202 [V_008F0C_BUF_DATA_FORMAT_16] = {2, 1, 2, V_008F0C_BUF_DATA_FORMAT_16}, 203 [V_008F0C_BUF_DATA_FORMAT_8_8] = {2, 2, 1, V_008F0C_BUF_DATA_FORMAT_8}, 204 [V_008F0C_BUF_DATA_FORMAT_32] = {4, 1, 4, V_008F0C_BUF_DATA_FORMAT_32}, 205 [V_008F0C_BUF_DATA_FORMAT_16_16] = {4, 2, 2, V_008F0C_BUF_DATA_FORMAT_16}, 206 [V_008F0C_BUF_DATA_FORMAT_10_11_11] = {4, 3, 0, V_008F0C_BUF_DATA_FORMAT_10_11_11}, 207 [V_008F0C_BUF_DATA_FORMAT_11_11_10] = {4, 3, 0, V_008F0C_BUF_DATA_FORMAT_11_11_10}, 208 [V_008F0C_BUF_DATA_FORMAT_10_10_10_2] = {4, 4, 0, V_008F0C_BUF_DATA_FORMAT_10_10_10_2}, 209 [V_008F0C_BUF_DATA_FORMAT_2_10_10_10] = {4, 4, 0, V_008F0C_BUF_DATA_FORMAT_2_10_10_10}, 210 [V_008F0C_BUF_DATA_FORMAT_8_8_8_8] = {4, 4, 1, V_008F0C_BUF_DATA_FORMAT_8}, 211 [V_008F0C_BUF_DATA_FORMAT_32_32] = {8, 2, 4, V_008F0C_BUF_DATA_FORMAT_32}, 212 [V_008F0C_BUF_DATA_FORMAT_16_16_16_16] = {8, 4, 2, V_008F0C_BUF_DATA_FORMAT_16}, 213 [V_008F0C_BUF_DATA_FORMAT_32_32_32] = {12, 3, 4, V_008F0C_BUF_DATA_FORMAT_32}, 214 [V_008F0C_BUF_DATA_FORMAT_32_32_32_32] = {16, 4, 4, V_008F0C_BUF_DATA_FORMAT_32}, 215}; 216 217const struct ac_data_format_info *ac_get_data_format_info(unsigned dfmt) 218{ 219 assert(dfmt < ARRAY_SIZE(data_format_table)); 220 return &data_format_table[dfmt]; 221} 222 223enum ac_image_dim ac_get_sampler_dim(enum chip_class chip_class, enum glsl_sampler_dim dim, 224 bool is_array) 225{ 226 switch (dim) { 227 case GLSL_SAMPLER_DIM_1D: 228 if (chip_class == GFX9) 229 return is_array ? ac_image_2darray : ac_image_2d; 230 return is_array ? ac_image_1darray : ac_image_1d; 231 case GLSL_SAMPLER_DIM_2D: 232 case GLSL_SAMPLER_DIM_RECT: 233 case GLSL_SAMPLER_DIM_EXTERNAL: 234 return is_array ? ac_image_2darray : ac_image_2d; 235 case GLSL_SAMPLER_DIM_3D: 236 return ac_image_3d; 237 case GLSL_SAMPLER_DIM_CUBE: 238 return ac_image_cube; 239 case GLSL_SAMPLER_DIM_MS: 240 return is_array ? ac_image_2darraymsaa : ac_image_2dmsaa; 241 case GLSL_SAMPLER_DIM_SUBPASS: 242 return ac_image_2darray; 243 case GLSL_SAMPLER_DIM_SUBPASS_MS: 244 return ac_image_2darraymsaa; 245 default: 246 unreachable("bad sampler dim"); 247 } 248} 249 250enum ac_image_dim ac_get_image_dim(enum chip_class chip_class, enum glsl_sampler_dim sdim, 251 bool is_array) 252{ 253 enum ac_image_dim dim = ac_get_sampler_dim(chip_class, sdim, is_array); 254 255 /* Match the resource type set in the descriptor. */ 256 if (dim == ac_image_cube || (chip_class <= GFX8 && dim == ac_image_3d)) 257 dim = ac_image_2darray; 258 else if (sdim == GLSL_SAMPLER_DIM_2D && !is_array && chip_class == GFX9) { 259 /* When a single layer of a 3D texture is bound, the shader 260 * will refer to a 2D target, but the descriptor has a 3D type. 261 * Since the HW ignores BASE_ARRAY in this case, we need to 262 * send 3 coordinates. This doesn't hurt when the underlying 263 * texture is non-3D. 264 */ 265 dim = ac_image_3d; 266 } 267 268 return dim; 269} 270 271unsigned ac_get_fs_input_vgpr_cnt(const struct ac_shader_config *config, 272 signed char *face_vgpr_index_ptr, 273 signed char *ancillary_vgpr_index_ptr) 274{ 275 unsigned num_input_vgprs = 0; 276 signed char face_vgpr_index = -1; 277 signed char ancillary_vgpr_index = -1; 278 279 if (G_0286CC_PERSP_SAMPLE_ENA(config->spi_ps_input_addr)) 280 num_input_vgprs += 2; 281 if (G_0286CC_PERSP_CENTER_ENA(config->spi_ps_input_addr)) 282 num_input_vgprs += 2; 283 if (G_0286CC_PERSP_CENTROID_ENA(config->spi_ps_input_addr)) 284 num_input_vgprs += 2; 285 if (G_0286CC_PERSP_PULL_MODEL_ENA(config->spi_ps_input_addr)) 286 num_input_vgprs += 3; 287 if (G_0286CC_LINEAR_SAMPLE_ENA(config->spi_ps_input_addr)) 288 num_input_vgprs += 2; 289 if (G_0286CC_LINEAR_CENTER_ENA(config->spi_ps_input_addr)) 290 num_input_vgprs += 2; 291 if (G_0286CC_LINEAR_CENTROID_ENA(config->spi_ps_input_addr)) 292 num_input_vgprs += 2; 293 if (G_0286CC_LINE_STIPPLE_TEX_ENA(config->spi_ps_input_addr)) 294 num_input_vgprs += 1; 295 if (G_0286CC_POS_X_FLOAT_ENA(config->spi_ps_input_addr)) 296 num_input_vgprs += 1; 297 if (G_0286CC_POS_Y_FLOAT_ENA(config->spi_ps_input_addr)) 298 num_input_vgprs += 1; 299 if (G_0286CC_POS_Z_FLOAT_ENA(config->spi_ps_input_addr)) 300 num_input_vgprs += 1; 301 if (G_0286CC_POS_W_FLOAT_ENA(config->spi_ps_input_addr)) 302 num_input_vgprs += 1; 303 if (G_0286CC_FRONT_FACE_ENA(config->spi_ps_input_addr)) { 304 face_vgpr_index = num_input_vgprs; 305 num_input_vgprs += 1; 306 } 307 if (G_0286CC_ANCILLARY_ENA(config->spi_ps_input_addr)) { 308 ancillary_vgpr_index = num_input_vgprs; 309 num_input_vgprs += 1; 310 } 311 if (G_0286CC_SAMPLE_COVERAGE_ENA(config->spi_ps_input_addr)) 312 num_input_vgprs += 1; 313 if (G_0286CC_POS_FIXED_PT_ENA(config->spi_ps_input_addr)) 314 num_input_vgprs += 1; 315 316 if (face_vgpr_index_ptr) 317 *face_vgpr_index_ptr = face_vgpr_index; 318 if (ancillary_vgpr_index_ptr) 319 *ancillary_vgpr_index_ptr = ancillary_vgpr_index; 320 321 return num_input_vgprs; 322} 323 324void ac_choose_spi_color_formats(unsigned format, unsigned swap, unsigned ntype, 325 bool is_depth, bool use_rbplus, 326 struct ac_spi_color_formats *formats) 327{ 328 /* Alpha is needed for alpha-to-coverage. 329 * Blending may be with or without alpha. 330 */ 331 unsigned normal = 0; /* most optimal, may not support blending or export alpha */ 332 unsigned alpha = 0; /* exports alpha, but may not support blending */ 333 unsigned blend = 0; /* supports blending, but may not export alpha */ 334 unsigned blend_alpha = 0; /* least optimal, supports blending and exports alpha */ 335 336 /* Choose the SPI color formats. These are required values for RB+. 337 * Other chips have multiple choices, though they are not necessarily better. 338 */ 339 switch (format) { 340 case V_028C70_COLOR_5_6_5: 341 case V_028C70_COLOR_1_5_5_5: 342 case V_028C70_COLOR_5_5_5_1: 343 case V_028C70_COLOR_4_4_4_4: 344 case V_028C70_COLOR_10_11_11: 345 case V_028C70_COLOR_11_11_10: 346 case V_028C70_COLOR_5_9_9_9: 347 case V_028C70_COLOR_8: 348 case V_028C70_COLOR_8_8: 349 case V_028C70_COLOR_8_8_8_8: 350 case V_028C70_COLOR_10_10_10_2: 351 case V_028C70_COLOR_2_10_10_10: 352 if (ntype == V_028C70_NUMBER_UINT) 353 alpha = blend = blend_alpha = normal = V_028714_SPI_SHADER_UINT16_ABGR; 354 else if (ntype == V_028C70_NUMBER_SINT) 355 alpha = blend = blend_alpha = normal = V_028714_SPI_SHADER_SINT16_ABGR; 356 else 357 alpha = blend = blend_alpha = normal = V_028714_SPI_SHADER_FP16_ABGR; 358 359 if (!use_rbplus && format == V_028C70_COLOR_8 && 360 ntype != V_028C70_NUMBER_SRGB && swap == V_028C70_SWAP_STD) /* R */ { 361 /* When RB+ is enabled, R8_UNORM should use FP16_ABGR for 2x 362 * exporting performance. Otherwise, use 32_R to remove useless 363 * instructions needed for 16-bit compressed exports. 364 */ 365 blend = normal = V_028714_SPI_SHADER_32_R; 366 } 367 break; 368 369 case V_028C70_COLOR_16: 370 case V_028C70_COLOR_16_16: 371 case V_028C70_COLOR_16_16_16_16: 372 if (ntype == V_028C70_NUMBER_UNORM || ntype == V_028C70_NUMBER_SNORM) { 373 /* UNORM16 and SNORM16 don't support blending */ 374 if (ntype == V_028C70_NUMBER_UNORM) 375 normal = alpha = V_028714_SPI_SHADER_UNORM16_ABGR; 376 else 377 normal = alpha = V_028714_SPI_SHADER_SNORM16_ABGR; 378 379 /* Use 32 bits per channel for blending. */ 380 if (format == V_028C70_COLOR_16) { 381 if (swap == V_028C70_SWAP_STD) { /* R */ 382 blend = V_028714_SPI_SHADER_32_R; 383 blend_alpha = V_028714_SPI_SHADER_32_AR; 384 } else if (swap == V_028C70_SWAP_ALT_REV) /* A */ 385 blend = blend_alpha = V_028714_SPI_SHADER_32_AR; 386 else 387 assert(0); 388 } else if (format == V_028C70_COLOR_16_16) { 389 if (swap == V_028C70_SWAP_STD) { /* RG */ 390 blend = V_028714_SPI_SHADER_32_GR; 391 blend_alpha = V_028714_SPI_SHADER_32_ABGR; 392 } else if (swap == V_028C70_SWAP_ALT) /* RA */ 393 blend = blend_alpha = V_028714_SPI_SHADER_32_AR; 394 else 395 assert(0); 396 } else /* 16_16_16_16 */ 397 blend = blend_alpha = V_028714_SPI_SHADER_32_ABGR; 398 } else if (ntype == V_028C70_NUMBER_UINT) 399 alpha = blend = blend_alpha = normal = V_028714_SPI_SHADER_UINT16_ABGR; 400 else if (ntype == V_028C70_NUMBER_SINT) 401 alpha = blend = blend_alpha = normal = V_028714_SPI_SHADER_SINT16_ABGR; 402 else if (ntype == V_028C70_NUMBER_FLOAT) 403 alpha = blend = blend_alpha = normal = V_028714_SPI_SHADER_FP16_ABGR; 404 else 405 assert(0); 406 break; 407 408 case V_028C70_COLOR_32: 409 if (swap == V_028C70_SWAP_STD) { /* R */ 410 blend = normal = V_028714_SPI_SHADER_32_R; 411 alpha = blend_alpha = V_028714_SPI_SHADER_32_AR; 412 } else if (swap == V_028C70_SWAP_ALT_REV) /* A */ 413 alpha = blend = blend_alpha = normal = V_028714_SPI_SHADER_32_AR; 414 else 415 assert(0); 416 break; 417 418 case V_028C70_COLOR_32_32: 419 if (swap == V_028C70_SWAP_STD) { /* RG */ 420 blend = normal = V_028714_SPI_SHADER_32_GR; 421 alpha = blend_alpha = V_028714_SPI_SHADER_32_ABGR; 422 } else if (swap == V_028C70_SWAP_ALT) /* RA */ 423 alpha = blend = blend_alpha = normal = V_028714_SPI_SHADER_32_AR; 424 else 425 assert(0); 426 break; 427 428 case V_028C70_COLOR_32_32_32_32: 429 case V_028C70_COLOR_8_24: 430 case V_028C70_COLOR_24_8: 431 case V_028C70_COLOR_X24_8_32_FLOAT: 432 alpha = blend = blend_alpha = normal = V_028714_SPI_SHADER_32_ABGR; 433 break; 434 435 default: 436 assert(0); 437 return; 438 } 439 440 /* The DB->CB copy needs 32_ABGR. */ 441 if (is_depth) 442 alpha = blend = blend_alpha = normal = V_028714_SPI_SHADER_32_ABGR; 443 444 formats->normal = normal; 445 formats->alpha = alpha; 446 formats->blend = blend; 447 formats->blend_alpha = blend_alpha; 448} 449 450void ac_compute_late_alloc(const struct radeon_info *info, bool ngg, bool ngg_culling, 451 bool uses_scratch, unsigned *late_alloc_wave64, unsigned *cu_mask) 452{ 453 *late_alloc_wave64 = 0; /* The limit is per SA. */ 454 *cu_mask = 0xffff; 455 456 /* CU masking can decrease performance and cause a hang with <= 2 CUs per SA. */ 457 if (info->min_good_cu_per_sa <= 2) 458 return; 459 460 /* If scratch is used with late alloc, the GPU could deadlock if PS uses scratch too. A more 461 * complicated computation is needed to enable late alloc with scratch (see PAL). 462 */ 463 if (uses_scratch) 464 return; 465 466 /* Late alloc is not used for NGG on Navi14 due to a hw bug. */ 467 if (ngg && info->family == CHIP_NAVI14) 468 return; 469 470 if (info->chip_class >= GFX10) { 471 /* For Wave32, the hw will launch twice the number of late alloc waves, so 1 == 2x wave32. 472 * These limits are estimated because they are all safe but they vary in performance. 473 */ 474 if (ngg_culling) 475 *late_alloc_wave64 = info->min_good_cu_per_sa * 10; 476 else 477 *late_alloc_wave64 = info->min_good_cu_per_sa * 4; 478 479 /* Limit LATE_ALLOC_GS to prevent a hang (hw bug) on gfx10. */ 480 if (info->chip_class == GFX10 && ngg) 481 *late_alloc_wave64 = MIN2(*late_alloc_wave64, 64); 482 483 /* Gfx10: CU2 & CU3 must be disabled to prevent a hw deadlock. 484 * Others: CU1 must be disabled to prevent a hw deadlock. 485 * 486 * The deadlock is caused by late alloc, which usually increases performance. 487 */ 488 *cu_mask &= info->chip_class == GFX10 ? ~BITFIELD_RANGE(2, 2) : 489 ~BITFIELD_RANGE(1, 1); 490 } else { 491 if (info->min_good_cu_per_sa <= 4) { 492 /* Too few available compute units per SA. Disallowing VS to run on one CU could hurt us 493 * more than late VS allocation would help. 494 * 495 * 2 is the highest safe number that allows us to keep all CUs enabled. 496 */ 497 *late_alloc_wave64 = 2; 498 } else { 499 /* This is a good initial value, allowing 1 late_alloc wave per SIMD on num_cu - 2. 500 */ 501 *late_alloc_wave64 = (info->min_good_cu_per_sa - 2) * 4; 502 } 503 504 /* VS can't execute on one CU if the limit is > 2. */ 505 if (*late_alloc_wave64 > 2) 506 *cu_mask = 0xfffe; /* 1 CU disabled */ 507 } 508 509 /* Max number that fits into the register field. */ 510 if (ngg) /* GS */ 511 *late_alloc_wave64 = MIN2(*late_alloc_wave64, G_00B204_SPI_SHADER_LATE_ALLOC_GS_GFX10(~0u)); 512 else /* VS */ 513 *late_alloc_wave64 = MIN2(*late_alloc_wave64, G_00B11C_LIMIT(~0u)); 514} 515 516unsigned ac_compute_cs_workgroup_size(uint16_t sizes[3], bool variable, unsigned max) 517{ 518 if (variable) 519 return max; 520 521 return sizes[0] * sizes[1] * sizes[2]; 522} 523 524unsigned ac_compute_lshs_workgroup_size(enum chip_class chip_class, gl_shader_stage stage, 525 unsigned tess_num_patches, 526 unsigned tess_patch_in_vtx, 527 unsigned tess_patch_out_vtx) 528{ 529 /* When tessellation is used, API VS runs on HW LS, API TCS runs on HW HS. 530 * These two HW stages are merged on GFX9+. 531 */ 532 533 bool merged_shaders = chip_class >= GFX9; 534 unsigned ls_workgroup_size = tess_num_patches * tess_patch_in_vtx; 535 unsigned hs_workgroup_size = tess_num_patches * tess_patch_out_vtx; 536 537 if (merged_shaders) 538 return MAX2(ls_workgroup_size, hs_workgroup_size); 539 else if (stage == MESA_SHADER_VERTEX) 540 return ls_workgroup_size; 541 else if (stage == MESA_SHADER_TESS_CTRL) 542 return hs_workgroup_size; 543 else 544 unreachable("invalid LSHS shader stage"); 545} 546 547unsigned ac_compute_esgs_workgroup_size(enum chip_class chip_class, unsigned wave_size, 548 unsigned es_verts, unsigned gs_inst_prims) 549{ 550 /* ESGS may operate in workgroups if on-chip GS (LDS rings) are enabled. 551 * 552 * GFX6: Not possible in the HW. 553 * GFX7-8 (unmerged): possible in the HW, but not implemented in Mesa. 554 * GFX9+ (merged): implemented in Mesa. 555 */ 556 557 if (chip_class <= GFX8) 558 return wave_size; 559 560 unsigned workgroup_size = MAX2(es_verts, gs_inst_prims); 561 return CLAMP(workgroup_size, 1, 256); 562} 563 564unsigned ac_compute_ngg_workgroup_size(unsigned es_verts, unsigned gs_inst_prims, 565 unsigned max_vtx_out, unsigned prim_amp_factor) 566{ 567 /* NGG always operates in workgroups. 568 * 569 * For API VS/TES/GS: 570 * - 1 invocation per input vertex 571 * - 1 invocation per input primitive 572 * 573 * The same invocation can process both an input vertex and primitive, 574 * however 1 invocation can only output up to 1 vertex and 1 primitive. 575 */ 576 577 unsigned max_vtx_in = es_verts < 256 ? es_verts : 3 * gs_inst_prims; 578 unsigned max_prim_in = gs_inst_prims; 579 unsigned max_prim_out = gs_inst_prims * prim_amp_factor; 580 unsigned workgroup_size = MAX4(max_vtx_in, max_vtx_out, max_prim_in, max_prim_out); 581 582 return CLAMP(workgroup_size, 1, 256); 583} 584