1/* 2 * Copyright © 2014 Intel Corporation 3 * 4 * Permission is hereby granted, free of charge, to any person obtaining a 5 * copy of this software and associated documentation files (the "Software"), 6 * to deal in the Software without restriction, including without limitation 7 * the rights to use, copy, modify, merge, publish, distribute, sublicense, 8 * and/or sell copies of the Software, and to permit persons to whom the 9 * Software is furnished to do so, subject to the following conditions: 10 * 11 * The above copyright notice and this permission notice (including the next 12 * paragraph) shall be included in all copies or substantial portions of the 13 * Software. 14 * 15 * THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR 16 * IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY, 17 * FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. IN NO EVENT SHALL 18 * THE AUTHORS OR COPYRIGHT HOLDERS BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER 19 * LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING 20 * FROM, OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS 21 * IN THE SOFTWARE. 22 * 23 * Authors: 24 * Jason Ekstrand (jason@jlekstrand.net) 25 * 26 */ 27 28#include "nir.h" 29#include "nir_builder.h" 30#include "nir_deref.h" 31#include "nir_phi_builder.h" 32#include "nir_vla.h" 33 34 35struct deref_node { 36 struct deref_node *parent; 37 const struct glsl_type *type; 38 39 bool lower_to_ssa; 40 41 /* Only valid for things that end up in the direct list. 42 * Note that multiple nir_deref_instrs may correspond to this node, but 43 * they will all be equivalent, so any is as good as the other. 44 */ 45 nir_deref_path path; 46 struct exec_node direct_derefs_link; 47 48 struct set *loads; 49 struct set *stores; 50 struct set *copies; 51 52 struct nir_phi_builder_value *pb_value; 53 54 /* True if this node is fully direct. If set, it must be in the children 55 * array of its parent. 56 */ 57 bool is_direct; 58 59 struct deref_node *wildcard; 60 struct deref_node *indirect; 61 struct deref_node *children[0]; 62}; 63 64struct lower_variables_state { 65 nir_shader *shader; 66 void *dead_ctx; 67 nir_function_impl *impl; 68 69 /* A hash table mapping variables to deref_node data */ 70 struct hash_table *deref_var_nodes; 71 72 /* A hash table mapping fully-qualified direct dereferences, i.e. 73 * dereferences with no indirect or wildcard array dereferences, to 74 * deref_node data. 75 * 76 * At the moment, we only lower loads, stores, and copies that can be 77 * trivially lowered to loads and stores, i.e. copies with no indirects 78 * and no wildcards. If a part of a variable that is being loaded from 79 * and/or stored into is also involved in a copy operation with 80 * wildcards, then we lower that copy operation to loads and stores, but 81 * otherwise we leave copies with wildcards alone. Since the only derefs 82 * used in these loads, stores, and trivial copies are ones with no 83 * wildcards and no indirects, these are precisely the derefs that we 84 * can actually consider lowering. 85 */ 86 struct exec_list direct_deref_nodes; 87 88 /* Controls whether get_deref_node will add variables to the 89 * direct_deref_nodes table. This is turned on when we are initially 90 * scanning for load/store instructions. It is then turned off so we 91 * don't accidentally change the direct_deref_nodes table while we're 92 * iterating throug it. 93 */ 94 bool add_to_direct_deref_nodes; 95 96 struct nir_phi_builder *phi_builder; 97}; 98 99static struct deref_node * 100deref_node_create(struct deref_node *parent, 101 const struct glsl_type *type, 102 bool is_direct, void *mem_ctx) 103{ 104 size_t size = sizeof(struct deref_node) + 105 glsl_get_length(type) * sizeof(struct deref_node *); 106 107 struct deref_node *node = rzalloc_size(mem_ctx, size); 108 node->type = type; 109 node->parent = parent; 110 exec_node_init(&node->direct_derefs_link); 111 node->is_direct = is_direct; 112 113 return node; 114} 115 116/* Returns the deref node associated with the given variable. This will be 117 * the root of the tree representing all of the derefs of the given variable. 118 */ 119static struct deref_node * 120get_deref_node_for_var(nir_variable *var, struct lower_variables_state *state) 121{ 122 struct deref_node *node; 123 124 struct hash_entry *var_entry = 125 _mesa_hash_table_search(state->deref_var_nodes, var); 126 127 if (var_entry) { 128 return var_entry->data; 129 } else { 130 node = deref_node_create(NULL, var->type, true, state->dead_ctx); 131 _mesa_hash_table_insert(state->deref_var_nodes, var, node); 132 return node; 133 } 134} 135 136/* Gets the deref_node for the given deref chain and creates it if it 137 * doesn't yet exist. If the deref is fully-qualified and direct and 138 * state->add_to_direct_deref_nodes is true, it will be added to the hash 139 * table of of fully-qualified direct derefs. 140 */ 141static struct deref_node * 142get_deref_node_recur(nir_deref_instr *deref, 143 struct lower_variables_state *state) 144{ 145 if (deref->deref_type == nir_deref_type_var) 146 return get_deref_node_for_var(deref->var, state); 147 148 struct deref_node *parent = 149 get_deref_node_recur(nir_deref_instr_parent(deref), state); 150 151 switch (deref->deref_type) { 152 case nir_deref_type_struct: 153 assert(glsl_type_is_struct_or_ifc(parent->type)); 154 assert(deref->strct.index < glsl_get_length(parent->type)); 155 156 if (parent->children[deref->strct.index] == NULL) { 157 parent->children[deref->strct.index] = 158 deref_node_create(parent, deref->type, parent->is_direct, 159 state->dead_ctx); 160 } 161 162 return parent->children[deref->strct.index]; 163 164 case nir_deref_type_array: { 165 if (nir_src_is_const(deref->arr.index)) { 166 uint32_t index = nir_src_as_uint(deref->arr.index); 167 /* This is possible if a loop unrolls and generates an 168 * out-of-bounds offset. We need to handle this at least 169 * somewhat gracefully. 170 */ 171 if (index >= glsl_get_length(parent->type)) 172 return NULL; 173 174 if (parent->children[index] == NULL) { 175 parent->children[index] = 176 deref_node_create(parent, deref->type, parent->is_direct, 177 state->dead_ctx); 178 } 179 180 return parent->children[index]; 181 } else { 182 if (parent->indirect == NULL) { 183 parent->indirect = 184 deref_node_create(parent, deref->type, false, state->dead_ctx); 185 } 186 187 return parent->indirect; 188 } 189 break; 190 } 191 192 case nir_deref_type_array_wildcard: 193 if (parent->wildcard == NULL) { 194 parent->wildcard = 195 deref_node_create(parent, deref->type, false, state->dead_ctx); 196 } 197 198 return parent->wildcard; 199 200 default: 201 unreachable("Invalid deref type"); 202 } 203} 204 205static struct deref_node * 206get_deref_node(nir_deref_instr *deref, struct lower_variables_state *state) 207{ 208 /* This pass only works on local variables. Just ignore any derefs with 209 * a non-local mode. 210 */ 211 if (deref->mode != nir_var_function_temp) 212 return NULL; 213 214 struct deref_node *node = get_deref_node_recur(deref, state); 215 if (!node) 216 return NULL; 217 218 /* Insert the node in the direct derefs list. We only do this if it's not 219 * already in the list and we only bother for deref nodes which are used 220 * directly in a load or store. 221 */ 222 if (node->is_direct && state->add_to_direct_deref_nodes && 223 node->direct_derefs_link.next == NULL) { 224 nir_deref_path_init(&node->path, deref, state->dead_ctx); 225 assert(deref->var != NULL); 226 exec_list_push_tail(&state->direct_deref_nodes, 227 &node->direct_derefs_link); 228 } 229 230 return node; 231} 232 233/* \sa foreach_deref_node_match */ 234static void 235foreach_deref_node_worker(struct deref_node *node, nir_deref_instr **path, 236 void (* cb)(struct deref_node *node, 237 struct lower_variables_state *state), 238 struct lower_variables_state *state) 239{ 240 if (*path == NULL) { 241 cb(node, state); 242 return; 243 } 244 245 switch ((*path)->deref_type) { 246 case nir_deref_type_struct: 247 if (node->children[(*path)->strct.index]) { 248 foreach_deref_node_worker(node->children[(*path)->strct.index], 249 path + 1, cb, state); 250 } 251 return; 252 253 case nir_deref_type_array: { 254 uint32_t index = nir_src_as_uint((*path)->arr.index); 255 256 if (node->children[index]) { 257 foreach_deref_node_worker(node->children[index], 258 path + 1, cb, state); 259 } 260 261 if (node->wildcard) { 262 foreach_deref_node_worker(node->wildcard, 263 path + 1, cb, state); 264 } 265 return; 266 } 267 268 default: 269 unreachable("Unsupported deref type"); 270 } 271} 272 273/* Walks over every "matching" deref_node and calls the callback. A node 274 * is considered to "match" if either refers to that deref or matches up t 275 * a wildcard. In other words, the following would match a[6].foo[3].bar: 276 * 277 * a[6].foo[3].bar 278 * a[*].foo[3].bar 279 * a[6].foo[*].bar 280 * a[*].foo[*].bar 281 * 282 * The given deref must be a full-length and fully qualified (no wildcards 283 * or indirects) deref chain. 284 */ 285static void 286foreach_deref_node_match(nir_deref_path *path, 287 void (* cb)(struct deref_node *node, 288 struct lower_variables_state *state), 289 struct lower_variables_state *state) 290{ 291 assert(path->path[0]->deref_type == nir_deref_type_var); 292 struct deref_node *node = get_deref_node_for_var(path->path[0]->var, state); 293 294 if (node == NULL) 295 return; 296 297 foreach_deref_node_worker(node, &path->path[1], cb, state); 298} 299 300/* \sa deref_may_be_aliased */ 301static bool 302path_may_be_aliased_node(struct deref_node *node, nir_deref_instr **path, 303 struct lower_variables_state *state) 304{ 305 if (*path == NULL) 306 return false; 307 308 switch ((*path)->deref_type) { 309 case nir_deref_type_struct: 310 if (node->children[(*path)->strct.index]) { 311 return path_may_be_aliased_node(node->children[(*path)->strct.index], 312 path + 1, state); 313 } else { 314 return false; 315 } 316 317 case nir_deref_type_array: { 318 if (!nir_src_is_const((*path)->arr.index)) 319 return true; 320 321 uint32_t index = nir_src_as_uint((*path)->arr.index); 322 323 /* If there is an indirect at this level, we're aliased. */ 324 if (node->indirect) 325 return true; 326 327 if (node->children[index] && 328 path_may_be_aliased_node(node->children[index], 329 path + 1, state)) 330 return true; 331 332 if (node->wildcard && 333 path_may_be_aliased_node(node->wildcard, path + 1, state)) 334 return true; 335 336 return false; 337 } 338 339 default: 340 unreachable("Unsupported deref type"); 341 } 342} 343 344/* Returns true if there are no indirects that can ever touch this deref. 345 * 346 * For example, if the given deref is a[6].foo, then any uses of a[i].foo 347 * would cause this to return false, but a[i].bar would not affect it 348 * because it's a different structure member. A var_copy involving of 349 * a[*].bar also doesn't affect it because that can be lowered to entirely 350 * direct load/stores. 351 * 352 * We only support asking this question about fully-qualified derefs. 353 * Obviously, it's pointless to ask this about indirects, but we also 354 * rule-out wildcards. Handling Wildcard dereferences would involve 355 * checking each array index to make sure that there aren't any indirect 356 * references. 357 */ 358static bool 359path_may_be_aliased(nir_deref_path *path, 360 struct lower_variables_state *state) 361{ 362 assert(path->path[0]->deref_type == nir_deref_type_var); 363 nir_variable *var = path->path[0]->var; 364 365 return path_may_be_aliased_node(get_deref_node_for_var(var, state), 366 &path->path[1], state); 367} 368 369static void 370register_load_instr(nir_intrinsic_instr *load_instr, 371 struct lower_variables_state *state) 372{ 373 nir_deref_instr *deref = nir_src_as_deref(load_instr->src[0]); 374 struct deref_node *node = get_deref_node(deref, state); 375 if (node == NULL) 376 return; 377 378 if (node->loads == NULL) 379 node->loads = _mesa_pointer_set_create(state->dead_ctx); 380 381 _mesa_set_add(node->loads, load_instr); 382} 383 384static void 385register_store_instr(nir_intrinsic_instr *store_instr, 386 struct lower_variables_state *state) 387{ 388 nir_deref_instr *deref = nir_src_as_deref(store_instr->src[0]); 389 struct deref_node *node = get_deref_node(deref, state); 390 if (node == NULL) 391 return; 392 393 if (node->stores == NULL) 394 node->stores = _mesa_pointer_set_create(state->dead_ctx); 395 396 _mesa_set_add(node->stores, store_instr); 397} 398 399static void 400register_copy_instr(nir_intrinsic_instr *copy_instr, 401 struct lower_variables_state *state) 402{ 403 for (unsigned idx = 0; idx < 2; idx++) { 404 nir_deref_instr *deref = nir_src_as_deref(copy_instr->src[idx]); 405 struct deref_node *node = get_deref_node(deref, state); 406 if (node == NULL) 407 continue; 408 409 if (node->copies == NULL) 410 node->copies = _mesa_pointer_set_create(state->dead_ctx); 411 412 _mesa_set_add(node->copies, copy_instr); 413 } 414} 415 416static void 417register_variable_uses(nir_function_impl *impl, 418 struct lower_variables_state *state) 419{ 420 nir_foreach_block(block, impl) { 421 nir_foreach_instr_safe(instr, block) { 422 if (instr->type != nir_instr_type_intrinsic) 423 continue; 424 425 nir_intrinsic_instr *intrin = nir_instr_as_intrinsic(instr); 426 427 switch (intrin->intrinsic) { 428 case nir_intrinsic_load_deref: 429 register_load_instr(intrin, state); 430 break; 431 432 case nir_intrinsic_store_deref: 433 register_store_instr(intrin, state); 434 break; 435 436 case nir_intrinsic_copy_deref: 437 register_copy_instr(intrin, state); 438 break; 439 440 default: 441 continue; 442 } 443 } 444 } 445} 446 447/* Walks over all of the copy instructions to or from the given deref_node 448 * and lowers them to load/store intrinsics. 449 */ 450static void 451lower_copies_to_load_store(struct deref_node *node, 452 struct lower_variables_state *state) 453{ 454 if (!node->copies) 455 return; 456 457 nir_builder b; 458 nir_builder_init(&b, state->impl); 459 460 set_foreach(node->copies, copy_entry) { 461 nir_intrinsic_instr *copy = (void *)copy_entry->key; 462 463 nir_lower_deref_copy_instr(&b, copy); 464 465 for (unsigned i = 0; i < 2; ++i) { 466 nir_deref_instr *arg_deref = nir_src_as_deref(copy->src[i]); 467 struct deref_node *arg_node = get_deref_node(arg_deref, state); 468 469 /* Only bother removing copy entries for other nodes */ 470 if (arg_node == NULL || arg_node == node) 471 continue; 472 473 struct set_entry *arg_entry = _mesa_set_search(arg_node->copies, copy); 474 assert(arg_entry); 475 _mesa_set_remove(arg_node->copies, arg_entry); 476 } 477 478 nir_instr_remove(©->instr); 479 } 480 481 node->copies = NULL; 482} 483 484/* Performs variable renaming 485 * 486 * This algorithm is very similar to the one outlined in "Efficiently 487 * Computing Static Single Assignment Form and the Control Dependence 488 * Graph" by Cytron et al. The primary difference is that we only put one 489 * SSA def on the stack per block. 490 */ 491static bool 492rename_variables(struct lower_variables_state *state) 493{ 494 nir_builder b; 495 nir_builder_init(&b, state->impl); 496 497 nir_foreach_block(block, state->impl) { 498 nir_foreach_instr_safe(instr, block) { 499 if (instr->type != nir_instr_type_intrinsic) 500 continue; 501 502 nir_intrinsic_instr *intrin = nir_instr_as_intrinsic(instr); 503 504 switch (intrin->intrinsic) { 505 case nir_intrinsic_load_deref: { 506 nir_deref_instr *deref = nir_src_as_deref(intrin->src[0]); 507 if (deref->mode != nir_var_function_temp) 508 continue; 509 510 struct deref_node *node = get_deref_node(deref, state); 511 if (node == NULL) { 512 /* If we hit this path then we are referencing an invalid 513 * value. Most likely, we unrolled something and are 514 * reading past the end of some array. In any case, this 515 * should result in an undefined value. 516 */ 517 nir_ssa_undef_instr *undef = 518 nir_ssa_undef_instr_create(state->shader, 519 intrin->num_components, 520 intrin->dest.ssa.bit_size); 521 522 nir_instr_insert_before(&intrin->instr, &undef->instr); 523 nir_instr_remove(&intrin->instr); 524 525 nir_ssa_def_rewrite_uses(&intrin->dest.ssa, 526 nir_src_for_ssa(&undef->def)); 527 continue; 528 } 529 530 if (!node->lower_to_ssa) 531 continue; 532 533 nir_alu_instr *mov = nir_alu_instr_create(state->shader, 534 nir_op_imov); 535 mov->src[0].src = nir_src_for_ssa( 536 nir_phi_builder_value_get_block_def(node->pb_value, block)); 537 for (unsigned i = intrin->num_components; i < NIR_MAX_VEC_COMPONENTS; i++) 538 mov->src[0].swizzle[i] = 0; 539 540 assert(intrin->dest.is_ssa); 541 542 mov->dest.write_mask = (1 << intrin->num_components) - 1; 543 nir_ssa_dest_init(&mov->instr, &mov->dest.dest, 544 intrin->num_components, 545 intrin->dest.ssa.bit_size, NULL); 546 547 nir_instr_insert_before(&intrin->instr, &mov->instr); 548 nir_instr_remove(&intrin->instr); 549 550 nir_ssa_def_rewrite_uses(&intrin->dest.ssa, 551 nir_src_for_ssa(&mov->dest.dest.ssa)); 552 break; 553 } 554 555 case nir_intrinsic_store_deref: { 556 nir_deref_instr *deref = nir_src_as_deref(intrin->src[0]); 557 if (deref->mode != nir_var_function_temp) 558 continue; 559 560 struct deref_node *node = get_deref_node(deref, state); 561 562 assert(intrin->src[1].is_ssa); 563 nir_ssa_def *value = intrin->src[1].ssa; 564 565 if (node == NULL) { 566 /* Probably an out-of-bounds array store. That should be a 567 * no-op. */ 568 nir_instr_remove(&intrin->instr); 569 continue; 570 } 571 572 if (!node->lower_to_ssa) 573 continue; 574 575 assert(intrin->num_components == 576 glsl_get_vector_elements(node->type)); 577 578 nir_ssa_def *new_def; 579 b.cursor = nir_before_instr(&intrin->instr); 580 581 unsigned wrmask = nir_intrinsic_write_mask(intrin); 582 if (wrmask == (1 << intrin->num_components) - 1) { 583 /* Whole variable store - just copy the source. Note that 584 * intrin->num_components and value->num_components 585 * may differ. 586 */ 587 unsigned swiz[NIR_MAX_VEC_COMPONENTS]; 588 for (unsigned i = 0; i < NIR_MAX_VEC_COMPONENTS; i++) 589 swiz[i] = i < intrin->num_components ? i : 0; 590 591 new_def = nir_swizzle(&b, value, swiz, 592 intrin->num_components, false); 593 } else { 594 nir_ssa_def *old_def = 595 nir_phi_builder_value_get_block_def(node->pb_value, block); 596 /* For writemasked store_var intrinsics, we combine the newly 597 * written values with the existing contents of unwritten 598 * channels, creating a new SSA value for the whole vector. 599 */ 600 nir_ssa_def *srcs[NIR_MAX_VEC_COMPONENTS]; 601 for (unsigned i = 0; i < intrin->num_components; i++) { 602 if (wrmask & (1 << i)) { 603 srcs[i] = nir_channel(&b, value, i); 604 } else { 605 srcs[i] = nir_channel(&b, old_def, i); 606 } 607 } 608 new_def = nir_vec(&b, srcs, intrin->num_components); 609 } 610 611 assert(new_def->num_components == intrin->num_components); 612 613 nir_phi_builder_value_set_block_def(node->pb_value, block, new_def); 614 nir_instr_remove(&intrin->instr); 615 break; 616 } 617 618 default: 619 break; 620 } 621 } 622 } 623 624 return true; 625} 626 627/** Implements a pass to lower variable uses to SSA values 628 * 629 * This path walks the list of instructions and tries to lower as many 630 * local variable load/store operations to SSA defs and uses as it can. 631 * The process involves four passes: 632 * 633 * 1) Iterate over all of the instructions and mark where each local 634 * variable deref is used in a load, store, or copy. While we're at 635 * it, we keep track of all of the fully-qualified (no wildcards) and 636 * fully-direct references we see and store them in the 637 * direct_deref_nodes hash table. 638 * 639 * 2) Walk over the list of fully-qualified direct derefs generated in 640 * the previous pass. For each deref, we determine if it can ever be 641 * aliased, i.e. if there is an indirect reference anywhere that may 642 * refer to it. If it cannot be aliased, we mark it for lowering to an 643 * SSA value. At this point, we lower any var_copy instructions that 644 * use the given deref to load/store operations. 645 * 646 * 3) Walk over the list of derefs we plan to lower to SSA values and 647 * insert phi nodes as needed. 648 * 649 * 4) Perform "variable renaming" by replacing the load/store instructions 650 * with SSA definitions and SSA uses. 651 */ 652static bool 653nir_lower_vars_to_ssa_impl(nir_function_impl *impl) 654{ 655 struct lower_variables_state state; 656 657 state.shader = impl->function->shader; 658 state.dead_ctx = ralloc_context(state.shader); 659 state.impl = impl; 660 661 state.deref_var_nodes = _mesa_pointer_hash_table_create(state.dead_ctx); 662 exec_list_make_empty(&state.direct_deref_nodes); 663 664 /* Build the initial deref structures and direct_deref_nodes table */ 665 state.add_to_direct_deref_nodes = true; 666 667 register_variable_uses(impl, &state); 668 669 bool progress = false; 670 671 nir_metadata_require(impl, nir_metadata_block_index); 672 673 /* We're about to iterate through direct_deref_nodes. Don't modify it. */ 674 state.add_to_direct_deref_nodes = false; 675 676 foreach_list_typed_safe(struct deref_node, node, direct_derefs_link, 677 &state.direct_deref_nodes) { 678 nir_deref_path *path = &node->path; 679 680 assert(path->path[0]->deref_type == nir_deref_type_var); 681 682 /* We don't build deref nodes for non-local variables */ 683 assert(path->path[0]->var->data.mode == nir_var_function_temp); 684 685 if (path_may_be_aliased(path, &state)) { 686 exec_node_remove(&node->direct_derefs_link); 687 continue; 688 } 689 690 node->lower_to_ssa = true; 691 progress = true; 692 693 foreach_deref_node_match(path, lower_copies_to_load_store, &state); 694 } 695 696 if (!progress) { 697#ifndef NDEBUG 698 impl->valid_metadata &= ~nir_metadata_not_properly_reset; 699#endif 700 return false; 701 } 702 703 nir_metadata_require(impl, nir_metadata_dominance); 704 705 /* We may have lowered some copy instructions to load/store 706 * instructions. The uses from the copy instructions hav already been 707 * removed but we need to rescan to ensure that the uses from the newly 708 * added load/store instructions are registered. We need this 709 * information for phi node insertion below. 710 */ 711 register_variable_uses(impl, &state); 712 713 state.phi_builder = nir_phi_builder_create(state.impl); 714 715 NIR_VLA(BITSET_WORD, store_blocks, BITSET_WORDS(state.impl->num_blocks)); 716 foreach_list_typed(struct deref_node, node, direct_derefs_link, 717 &state.direct_deref_nodes) { 718 if (!node->lower_to_ssa) 719 continue; 720 721 memset(store_blocks, 0, 722 BITSET_WORDS(state.impl->num_blocks) * sizeof(*store_blocks)); 723 724 assert(node->path.path[0]->var->constant_initializer == NULL); 725 726 if (node->stores) { 727 set_foreach(node->stores, store_entry) { 728 nir_intrinsic_instr *store = 729 (nir_intrinsic_instr *)store_entry->key; 730 BITSET_SET(store_blocks, store->instr.block->index); 731 } 732 } 733 734 node->pb_value = 735 nir_phi_builder_add_value(state.phi_builder, 736 glsl_get_vector_elements(node->type), 737 glsl_get_bit_size(node->type), 738 store_blocks); 739 } 740 741 rename_variables(&state); 742 743 nir_phi_builder_finish(state.phi_builder); 744 745 nir_metadata_preserve(impl, nir_metadata_block_index | 746 nir_metadata_dominance); 747 748 ralloc_free(state.dead_ctx); 749 750 return progress; 751} 752 753bool 754nir_lower_vars_to_ssa(nir_shader *shader) 755{ 756 bool progress = false; 757 758 nir_foreach_function(function, shader) { 759 if (function->impl) 760 progress |= nir_lower_vars_to_ssa_impl(function->impl); 761 } 762 763 return progress; 764} 765