// Copyright (c) 2018 Google LLC. // // Licensed under the Apache License, Version 2.0 (the "License"); // you may not use this file except in compliance with the License. // You may obtain a copy of the License at // // http://www.apache.org/licenses/LICENSE-2.0 // // Unless required by applicable law or agreed to in writing, software // distributed under the License is distributed on an "AS IS" BASIS, // WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. // See the License for the specific language governing permissions and // limitations under the License. #include <algorithm> #include <memory> #include <unordered_map> #include <unordered_set> #include <utility> #include <vector> #include "source/cfa.h" #include "source/opt/cfg.h" #include "source/opt/ir_builder.h" #include "source/opt/ir_context.h" #include "source/opt/loop_descriptor.h" #include "source/opt/loop_utils.h" namespace spvtools { namespace opt { namespace { // Return true if |bb| is dominated by at least one block in |exits| static inline bool DominatesAnExit(BasicBlock* bb, const std::unordered_set<BasicBlock*>& exits, const DominatorTree& dom_tree) { for (BasicBlock* e_bb : exits) if (dom_tree.Dominates(bb, e_bb)) return true; return false; } // Utility class to rewrite out-of-loop uses of an in-loop definition in terms // of phi instructions to achieve a LCSSA form. // For a given definition, the class user registers phi instructions using that // definition in all loop exit blocks by which the definition escapes. // Then, when rewriting a use of the definition, the rewriter walks the // paths from the use the loop exits. At each step, it will insert a phi // instruction to merge the incoming value according to exit blocks definition. class LCSSARewriter { public: LCSSARewriter(IRContext* context, const DominatorTree& dom_tree, const std::unordered_set<BasicBlock*>& exit_bb, BasicBlock* merge_block) : context_(context), cfg_(context_->cfg()), dom_tree_(dom_tree), exit_bb_(exit_bb), merge_block_id_(merge_block ? merge_block->id() : 0) {} struct UseRewriter { explicit UseRewriter(LCSSARewriter* base, const Instruction& def_insn) : base_(base), def_insn_(def_insn) {} // Rewrites the use of |def_insn_| by the instruction |user| at the index // |operand_index| in terms of phi instruction. This recursively builds new // phi instructions from |user| to the loop exit blocks' phis. The use of // |def_insn_| in |user| is replaced by the relevant phi instruction at the // end of the operation. // It is assumed that |user| does not dominates any of the loop exit basic // block. This operation does not update the def/use manager, instead it // records what needs to be updated. The actual update is performed by // UpdateManagers. void RewriteUse(BasicBlock* bb, Instruction* user, uint32_t operand_index) { assert( (user->opcode() != SpvOpPhi || bb != GetParent(user)) && "The root basic block must be the incoming edge if |user| is a phi " "instruction"); assert((user->opcode() == SpvOpPhi || bb == GetParent(user)) && "The root basic block must be the instruction parent if |user| is " "not " "phi instruction"); Instruction* new_def = GetOrBuildIncoming(bb->id()); user->SetOperand(operand_index, {new_def->result_id()}); rewritten_.insert(user); } // In-place update of some managers (avoid full invalidation). inline void UpdateManagers() { analysis::DefUseManager* def_use_mgr = base_->context_->get_def_use_mgr(); // Register all new definitions. for (Instruction* insn : rewritten_) { def_use_mgr->AnalyzeInstDef(insn); } // Register all new uses. for (Instruction* insn : rewritten_) { def_use_mgr->AnalyzeInstUse(insn); } } private: // Return the basic block that |instr| belongs to. BasicBlock* GetParent(Instruction* instr) { return base_->context_->get_instr_block(instr); } // Builds a phi instruction for the basic block |bb|. The function assumes // that |defining_blocks| contains the list of basic block that define the // usable value for each predecessor of |bb|. inline Instruction* CreatePhiInstruction( BasicBlock* bb, const std::vector<uint32_t>& defining_blocks) { std::vector<uint32_t> incomings; const std::vector<uint32_t>& bb_preds = base_->cfg_->preds(bb->id()); assert(bb_preds.size() == defining_blocks.size()); for (size_t i = 0; i < bb_preds.size(); i++) { incomings.push_back( GetOrBuildIncoming(defining_blocks[i])->result_id()); incomings.push_back(bb_preds[i]); } InstructionBuilder builder(base_->context_, &*bb->begin(), IRContext::kAnalysisInstrToBlockMapping); Instruction* incoming_phi = builder.AddPhi(def_insn_.type_id(), incomings); rewritten_.insert(incoming_phi); return incoming_phi; } // Builds a phi instruction for the basic block |bb|, all incoming values // will be |value|. inline Instruction* CreatePhiInstruction(BasicBlock* bb, const Instruction& value) { std::vector<uint32_t> incomings; const std::vector<uint32_t>& bb_preds = base_->cfg_->preds(bb->id()); for (size_t i = 0; i < bb_preds.size(); i++) { incomings.push_back(value.result_id()); incomings.push_back(bb_preds[i]); } InstructionBuilder builder(base_->context_, &*bb->begin(), IRContext::kAnalysisInstrToBlockMapping); Instruction* incoming_phi = builder.AddPhi(def_insn_.type_id(), incomings); rewritten_.insert(incoming_phi); return incoming_phi; } // Return the new def to use for the basic block |bb_id|. // If |bb_id| does not have a suitable def to use then we: // - return the common def used by all predecessors; // - if there is no common def, then we build a new phi instr at the // beginning of |bb_id| and return this new instruction. Instruction* GetOrBuildIncoming(uint32_t bb_id) { assert(base_->cfg_->block(bb_id) != nullptr && "Unknown basic block"); Instruction*& incoming_phi = bb_to_phi_[bb_id]; if (incoming_phi) { return incoming_phi; } BasicBlock* bb = &*base_->cfg_->block(bb_id); // If this is an exit basic block, look if there already is an eligible // phi instruction. An eligible phi has |def_insn_| as all incoming // values. if (base_->exit_bb_.count(bb)) { // Look if there is an eligible phi in this block. if (!bb->WhileEachPhiInst([&incoming_phi, this](Instruction* phi) { for (uint32_t i = 0; i < phi->NumInOperands(); i += 2) { if (phi->GetSingleWordInOperand(i) != def_insn_.result_id()) return true; } incoming_phi = phi; rewritten_.insert(incoming_phi); return false; })) { return incoming_phi; } incoming_phi = CreatePhiInstruction(bb, def_insn_); return incoming_phi; } // Get the block that defines the value to use for each predecessor. // If the vector has 1 value, then it means that this block does not need // to build a phi instruction unless |bb_id| is the loop merge block. const std::vector<uint32_t>& defining_blocks = base_->GetDefiningBlocks(bb_id); // Special case for structured loops: merge block might be different from // the exit block set. To maintain structured properties it will ease // transformations if the merge block also holds a phi instruction like // the exit ones. if (defining_blocks.size() > 1 || bb_id == base_->merge_block_id_) { if (defining_blocks.size() > 1) { incoming_phi = CreatePhiInstruction(bb, defining_blocks); } else { assert(bb_id == base_->merge_block_id_); incoming_phi = CreatePhiInstruction(bb, *GetOrBuildIncoming(defining_blocks[0])); } } else { incoming_phi = GetOrBuildIncoming(defining_blocks[0]); } return incoming_phi; } LCSSARewriter* base_; const Instruction& def_insn_; std::unordered_map<uint32_t, Instruction*> bb_to_phi_; std::unordered_set<Instruction*> rewritten_; }; private: // Return the new def to use for the basic block |bb_id|. // If |bb_id| does not have a suitable def to use then we: // - return the common def used by all predecessors; // - if there is no common def, then we build a new phi instr at the // beginning of |bb_id| and return this new instruction. const std::vector<uint32_t>& GetDefiningBlocks(uint32_t bb_id) { assert(cfg_->block(bb_id) != nullptr && "Unknown basic block"); std::vector<uint32_t>& defining_blocks = bb_to_defining_blocks_[bb_id]; if (defining_blocks.size()) return defining_blocks; // Check if one of the loop exit basic block dominates |bb_id|. for (const BasicBlock* e_bb : exit_bb_) { if (dom_tree_.Dominates(e_bb->id(), bb_id)) { defining_blocks.push_back(e_bb->id()); return defining_blocks; } } // Process parents, they will returns their suitable blocks. // If they are all the same, this means this basic block is dominated by a // common block, so we won't need to build a phi instruction. for (uint32_t pred_id : cfg_->preds(bb_id)) { const std::vector<uint32_t>& pred_blocks = GetDefiningBlocks(pred_id); if (pred_blocks.size() == 1) defining_blocks.push_back(pred_blocks[0]); else defining_blocks.push_back(pred_id); } assert(defining_blocks.size()); if (std::all_of(defining_blocks.begin(), defining_blocks.end(), [&defining_blocks](uint32_t id) { return id == defining_blocks[0]; })) { // No need for a phi. defining_blocks.resize(1); } return defining_blocks; } IRContext* context_; CFG* cfg_; const DominatorTree& dom_tree_; const std::unordered_set<BasicBlock*>& exit_bb_; uint32_t merge_block_id_; // This map represent the set of known paths. For each key, the vector // represent the set of blocks holding the definition to be used to build the // phi instruction. // If the vector has 0 value, then the path is unknown yet, and must be built. // If the vector has 1 value, then the value defined by that basic block // should be used. // If the vector has more than 1 value, then a phi node must be created, the // basic block ordering is the same as the predecessor ordering. std::unordered_map<uint32_t, std::vector<uint32_t>> bb_to_defining_blocks_; }; // Make the set |blocks| closed SSA. The set is closed SSA if all the uses // outside the set are phi instructions in exiting basic block set (hold by // |lcssa_rewriter|). inline void MakeSetClosedSSA(IRContext* context, Function* function, const std::unordered_set<uint32_t>& blocks, const std::unordered_set<BasicBlock*>& exit_bb, LCSSARewriter* lcssa_rewriter) { CFG& cfg = *context->cfg(); DominatorTree& dom_tree = context->GetDominatorAnalysis(function)->GetDomTree(); analysis::DefUseManager* def_use_manager = context->get_def_use_mgr(); for (uint32_t bb_id : blocks) { BasicBlock* bb = cfg.block(bb_id); // If bb does not dominate an exit block, then it cannot have escaping defs. if (!DominatesAnExit(bb, exit_bb, dom_tree)) continue; for (Instruction& inst : *bb) { LCSSARewriter::UseRewriter rewriter(lcssa_rewriter, inst); def_use_manager->ForEachUse( &inst, [&blocks, &rewriter, &exit_bb, context]( Instruction* use, uint32_t operand_index) { BasicBlock* use_parent = context->get_instr_block(use); assert(use_parent); if (blocks.count(use_parent->id())) return; if (use->opcode() == SpvOpPhi) { // If the use is a Phi instruction and the incoming block is // coming from the loop, then that's consistent with LCSSA form. if (exit_bb.count(use_parent)) { return; } else { // That's not an exit block, but the user is a phi instruction. // Consider the incoming branch only. use_parent = context->get_instr_block( use->GetSingleWordOperand(operand_index + 1)); } } // Rewrite the use. Note that this call does not invalidate the // def/use manager. So this operation is safe. rewriter.RewriteUse(use_parent, use, operand_index); }); rewriter.UpdateManagers(); } } } } // namespace void LoopUtils::CreateLoopDedicatedExits() { Function* function = loop_->GetHeaderBlock()->GetParent(); LoopDescriptor& loop_desc = *context_->GetLoopDescriptor(function); CFG& cfg = *context_->cfg(); analysis::DefUseManager* def_use_mgr = context_->get_def_use_mgr(); const IRContext::Analysis PreservedAnalyses = IRContext::kAnalysisDefUse | IRContext::kAnalysisInstrToBlockMapping; // Gathers the set of basic block that are not in this loop and have at least // one predecessor in the loop and one not in the loop. std::unordered_set<uint32_t> exit_bb_set; loop_->GetExitBlocks(&exit_bb_set); std::unordered_set<BasicBlock*> new_loop_exits; bool made_change = false; // For each block, we create a new one that gathers all branches from // the loop and fall into the block. for (uint32_t non_dedicate_id : exit_bb_set) { BasicBlock* non_dedicate = cfg.block(non_dedicate_id); const std::vector<uint32_t>& bb_pred = cfg.preds(non_dedicate_id); // Ignore the block if all the predecessors are in the loop. if (std::all_of(bb_pred.begin(), bb_pred.end(), [this](uint32_t id) { return loop_->IsInsideLoop(id); })) { new_loop_exits.insert(non_dedicate); continue; } made_change = true; Function::iterator insert_pt = function->begin(); for (; insert_pt != function->end() && &*insert_pt != non_dedicate; ++insert_pt) { } assert(insert_pt != function->end() && "Basic Block not found"); // Create the dedicate exit basic block. // TODO(1841): Handle id overflow. BasicBlock& exit = *insert_pt.InsertBefore(std::unique_ptr<BasicBlock>( new BasicBlock(std::unique_ptr<Instruction>(new Instruction( context_, SpvOpLabel, 0, context_->TakeNextId(), {}))))); exit.SetParent(function); // Redirect in loop predecessors to |exit| block. for (uint32_t exit_pred_id : bb_pred) { if (loop_->IsInsideLoop(exit_pred_id)) { BasicBlock* pred_block = cfg.block(exit_pred_id); pred_block->ForEachSuccessorLabel([non_dedicate, &exit](uint32_t* id) { if (*id == non_dedicate->id()) *id = exit.id(); }); // Update the CFG. // |non_dedicate|'s predecessor list will be updated at the end of the // loop. cfg.RegisterBlock(pred_block); } } // Register the label to the def/use manager, requires for the phi patching. def_use_mgr->AnalyzeInstDefUse(exit.GetLabelInst()); context_->set_instr_block(exit.GetLabelInst(), &exit); InstructionBuilder builder(context_, &exit, PreservedAnalyses); // Now jump from our dedicate basic block to the old exit. // We also reset the insert point so all instructions are inserted before // the branch. builder.SetInsertPoint(builder.AddBranch(non_dedicate->id())); non_dedicate->ForEachPhiInst( [&builder, &exit, def_use_mgr, this](Instruction* phi) { // New phi operands for this instruction. std::vector<uint32_t> new_phi_op; // Phi operands for the dedicated exit block. std::vector<uint32_t> exit_phi_op; for (uint32_t i = 0; i < phi->NumInOperands(); i += 2) { uint32_t def_id = phi->GetSingleWordInOperand(i); uint32_t incoming_id = phi->GetSingleWordInOperand(i + 1); if (loop_->IsInsideLoop(incoming_id)) { exit_phi_op.push_back(def_id); exit_phi_op.push_back(incoming_id); } else { new_phi_op.push_back(def_id); new_phi_op.push_back(incoming_id); } } // Build the new phi instruction dedicated exit block. Instruction* exit_phi = builder.AddPhi(phi->type_id(), exit_phi_op); // Build the new incoming branch. new_phi_op.push_back(exit_phi->result_id()); new_phi_op.push_back(exit.id()); // Rewrite operands. uint32_t idx = 0; for (; idx < new_phi_op.size(); idx++) phi->SetInOperand(idx, {new_phi_op[idx]}); // Remove extra operands, from last to first (more efficient). for (uint32_t j = phi->NumInOperands() - 1; j >= idx; j--) phi->RemoveInOperand(j); // Update the def/use manager for this |phi|. def_use_mgr->AnalyzeInstUse(phi); }); // Update the CFG. cfg.RegisterBlock(&exit); cfg.RemoveNonExistingEdges(non_dedicate->id()); new_loop_exits.insert(&exit); // If non_dedicate is in a loop, add the new dedicated exit in that loop. if (Loop* parent_loop = loop_desc[non_dedicate]) parent_loop->AddBasicBlock(&exit); } if (new_loop_exits.size() == 1) { loop_->SetMergeBlock(*new_loop_exits.begin()); } if (made_change) { context_->InvalidateAnalysesExceptFor( PreservedAnalyses | IRContext::kAnalysisCFG | IRContext::Analysis::kAnalysisLoopAnalysis); } } void LoopUtils::MakeLoopClosedSSA() { CreateLoopDedicatedExits(); Function* function = loop_->GetHeaderBlock()->GetParent(); CFG& cfg = *context_->cfg(); DominatorTree& dom_tree = context_->GetDominatorAnalysis(function)->GetDomTree(); std::unordered_set<BasicBlock*> exit_bb; { std::unordered_set<uint32_t> exit_bb_id; loop_->GetExitBlocks(&exit_bb_id); for (uint32_t bb_id : exit_bb_id) { exit_bb.insert(cfg.block(bb_id)); } } LCSSARewriter lcssa_rewriter(context_, dom_tree, exit_bb, loop_->GetMergeBlock()); MakeSetClosedSSA(context_, function, loop_->GetBlocks(), exit_bb, &lcssa_rewriter); // Make sure all defs post-dominated by the merge block have their last use no // further than the merge block. if (loop_->GetMergeBlock()) { std::unordered_set<uint32_t> merging_bb_id; loop_->GetMergingBlocks(&merging_bb_id); merging_bb_id.erase(loop_->GetMergeBlock()->id()); // Reset the exit set, now only the merge block is the exit. exit_bb.clear(); exit_bb.insert(loop_->GetMergeBlock()); // LCSSARewriter is reusable here only because it forces the creation of a // phi instruction in the merge block. MakeSetClosedSSA(context_, function, merging_bb_id, exit_bb, &lcssa_rewriter); } context_->InvalidateAnalysesExceptFor( IRContext::Analysis::kAnalysisCFG | IRContext::Analysis::kAnalysisDominatorAnalysis | IRContext::Analysis::kAnalysisLoopAnalysis); } Loop* LoopUtils::CloneLoop(LoopCloningResult* cloning_result) const { // Compute the structured order of the loop basic blocks and store it in the // vector ordered_loop_blocks. std::vector<BasicBlock*> ordered_loop_blocks; loop_->ComputeLoopStructuredOrder(&ordered_loop_blocks); // Clone the loop. return CloneLoop(cloning_result, ordered_loop_blocks); } Loop* LoopUtils::CloneAndAttachLoopToHeader(LoopCloningResult* cloning_result) { // Clone the loop. Loop* new_loop = CloneLoop(cloning_result); // Create a new exit block/label for the new loop. // TODO(1841): Handle id overflow. std::unique_ptr<Instruction> new_label{new Instruction( context_, SpvOp::SpvOpLabel, 0, context_->TakeNextId(), {})}; std::unique_ptr<BasicBlock> new_exit_bb{new BasicBlock(std::move(new_label))}; new_exit_bb->SetParent(loop_->GetMergeBlock()->GetParent()); // Create an unconditional branch to the header block. InstructionBuilder builder{context_, new_exit_bb.get()}; builder.AddBranch(loop_->GetHeaderBlock()->id()); // Save the ids of the new and old merge block. const uint32_t old_merge_block = loop_->GetMergeBlock()->id(); const uint32_t new_merge_block = new_exit_bb->id(); // Replace the uses of the old merge block in the new loop with the new merge // block. for (std::unique_ptr<BasicBlock>& basic_block : cloning_result->cloned_bb_) { for (Instruction& inst : *basic_block) { // For each operand in each instruction check if it is using the old merge // block and change it to be the new merge block. auto replace_merge_use = [old_merge_block, new_merge_block](uint32_t* id) { if (*id == old_merge_block) *id = new_merge_block; }; inst.ForEachInOperand(replace_merge_use); } } const uint32_t old_header = loop_->GetHeaderBlock()->id(); const uint32_t new_header = new_loop->GetHeaderBlock()->id(); analysis::DefUseManager* def_use = context_->get_def_use_mgr(); def_use->ForEachUse(old_header, [new_header, this](Instruction* inst, uint32_t operand) { if (!this->loop_->IsInsideLoop(inst)) inst->SetOperand(operand, {new_header}); }); // TODO(1841): Handle failure to create pre-header. def_use->ForEachUse( loop_->GetOrCreatePreHeaderBlock()->id(), [new_merge_block, this](Instruction* inst, uint32_t operand) { if (this->loop_->IsInsideLoop(inst)) inst->SetOperand(operand, {new_merge_block}); }); new_loop->SetMergeBlock(new_exit_bb.get()); new_loop->SetPreHeaderBlock(loop_->GetPreHeaderBlock()); // Add the new block into the cloned instructions. cloning_result->cloned_bb_.push_back(std::move(new_exit_bb)); return new_loop; } Loop* LoopUtils::CloneLoop( LoopCloningResult* cloning_result, const std::vector<BasicBlock*>& ordered_loop_blocks) const { analysis::DefUseManager* def_use_mgr = context_->get_def_use_mgr(); std::unique_ptr<Loop> new_loop = MakeUnique<Loop>(context_); CFG& cfg = *context_->cfg(); // Clone and place blocks in a SPIR-V compliant order (dominators first). for (BasicBlock* old_bb : ordered_loop_blocks) { // For each basic block in the loop, we clone it and register the mapping // between old and new ids. BasicBlock* new_bb = old_bb->Clone(context_); new_bb->SetParent(&function_); // TODO(1841): Handle id overflow. new_bb->GetLabelInst()->SetResultId(context_->TakeNextId()); def_use_mgr->AnalyzeInstDef(new_bb->GetLabelInst()); context_->set_instr_block(new_bb->GetLabelInst(), new_bb); cloning_result->cloned_bb_.emplace_back(new_bb); cloning_result->old_to_new_bb_[old_bb->id()] = new_bb; cloning_result->new_to_old_bb_[new_bb->id()] = old_bb; cloning_result->value_map_[old_bb->id()] = new_bb->id(); if (loop_->IsInsideLoop(old_bb)) new_loop->AddBasicBlock(new_bb); for (auto new_inst = new_bb->begin(), old_inst = old_bb->begin(); new_inst != new_bb->end(); ++new_inst, ++old_inst) { cloning_result->ptr_map_[&*new_inst] = &*old_inst; if (new_inst->HasResultId()) { // TODO(1841): Handle id overflow. new_inst->SetResultId(context_->TakeNextId()); cloning_result->value_map_[old_inst->result_id()] = new_inst->result_id(); // Only look at the defs for now, uses are not updated yet. def_use_mgr->AnalyzeInstDef(&*new_inst); } } } // All instructions (including all labels) have been cloned, // remap instruction operands id with the new ones. for (std::unique_ptr<BasicBlock>& bb_ref : cloning_result->cloned_bb_) { BasicBlock* bb = bb_ref.get(); for (Instruction& insn : *bb) { insn.ForEachInId([cloning_result](uint32_t* old_id) { // If the operand is defined in the loop, remap the id. auto id_it = cloning_result->value_map_.find(*old_id); if (id_it != cloning_result->value_map_.end()) { *old_id = id_it->second; } }); // Only look at what the instruction uses. All defs are register, so all // should be fine now. def_use_mgr->AnalyzeInstUse(&insn); context_->set_instr_block(&insn, bb); } cfg.RegisterBlock(bb); } PopulateLoopNest(new_loop.get(), *cloning_result); return new_loop.release(); } void LoopUtils::PopulateLoopNest( Loop* new_loop, const LoopCloningResult& cloning_result) const { std::unordered_map<Loop*, Loop*> loop_mapping; loop_mapping[loop_] = new_loop; if (loop_->HasParent()) loop_->GetParent()->AddNestedLoop(new_loop); PopulateLoopDesc(new_loop, loop_, cloning_result); for (Loop& sub_loop : make_range(++TreeDFIterator<Loop>(loop_), TreeDFIterator<Loop>())) { Loop* cloned = new Loop(context_); if (Loop* parent = loop_mapping[sub_loop.GetParent()]) parent->AddNestedLoop(cloned); loop_mapping[&sub_loop] = cloned; PopulateLoopDesc(cloned, &sub_loop, cloning_result); } loop_desc_->AddLoopNest(std::unique_ptr<Loop>(new_loop)); } // Populates |new_loop| descriptor according to |old_loop|'s one. void LoopUtils::PopulateLoopDesc( Loop* new_loop, Loop* old_loop, const LoopCloningResult& cloning_result) const { for (uint32_t bb_id : old_loop->GetBlocks()) { BasicBlock* bb = cloning_result.old_to_new_bb_.at(bb_id); new_loop->AddBasicBlock(bb); } new_loop->SetHeaderBlock( cloning_result.old_to_new_bb_.at(old_loop->GetHeaderBlock()->id())); if (old_loop->GetLatchBlock()) new_loop->SetLatchBlock( cloning_result.old_to_new_bb_.at(old_loop->GetLatchBlock()->id())); if (old_loop->GetContinueBlock()) new_loop->SetContinueBlock( cloning_result.old_to_new_bb_.at(old_loop->GetContinueBlock()->id())); if (old_loop->GetMergeBlock()) { auto it = cloning_result.old_to_new_bb_.find(old_loop->GetMergeBlock()->id()); BasicBlock* bb = it != cloning_result.old_to_new_bb_.end() ? it->second : old_loop->GetMergeBlock(); new_loop->SetMergeBlock(bb); } if (old_loop->GetPreHeaderBlock()) { auto it = cloning_result.old_to_new_bb_.find(old_loop->GetPreHeaderBlock()->id()); if (it != cloning_result.old_to_new_bb_.end()) { new_loop->SetPreHeaderBlock(it->second); } } } // Class to gather some metrics about a region of interest. void CodeMetrics::Analyze(const Loop& loop) { CFG& cfg = *loop.GetContext()->cfg(); roi_size_ = 0; block_sizes_.clear(); for (uint32_t id : loop.GetBlocks()) { const BasicBlock* bb = cfg.block(id); size_t bb_size = 0; bb->ForEachInst([&bb_size](const Instruction* insn) { if (insn->opcode() == SpvOpLabel) return; if (insn->IsNop()) return; if (insn->opcode() == SpvOpPhi) return; bb_size++; }); block_sizes_[bb->id()] = bb_size; roi_size_ += bb_size; } } } // namespace opt } // namespace spvtools