/* * Copyright (C) 2014 The Android Open Source Project * * Licensed under the Apache License, Version 2.0 (the "License"); * you may not use this file except in compliance with the License. * You may obtain a copy of the License at * * http://www.apache.org/licenses/LICENSE-2.0 * * Unless required by applicable law or agreed to in writing, software * distributed under the License is distributed on an "AS IS" BASIS, * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. * See the License for the specific language governing permissions and * limitations under the License. */ #include "prepare_for_register_allocation.h" #include "dex/dex_file_types.h" #include "driver/compiler_options.h" #include "jni/jni_internal.h" #include "optimizing_compiler_stats.h" #include "well_known_classes.h" namespace art { void PrepareForRegisterAllocation::Run() { // Order does not matter. for (HBasicBlock* block : GetGraph()->GetReversePostOrder()) { // No need to visit the phis. for (HInstructionIteratorHandleChanges inst_it(block->GetInstructions()); !inst_it.Done(); inst_it.Advance()) { inst_it.Current()->Accept(this); } } } void PrepareForRegisterAllocation::VisitCheckCast(HCheckCast* check_cast) { // Record only those bitstring type checks that make it to the codegen stage. if (check_cast->GetTypeCheckKind() == TypeCheckKind::kBitstringCheck) { MaybeRecordStat(stats_, MethodCompilationStat::kBitstringTypeCheck); } } void PrepareForRegisterAllocation::VisitInstanceOf(HInstanceOf* instance_of) { // Record only those bitstring type checks that make it to the codegen stage. if (instance_of->GetTypeCheckKind() == TypeCheckKind::kBitstringCheck) { MaybeRecordStat(stats_, MethodCompilationStat::kBitstringTypeCheck); } } void PrepareForRegisterAllocation::VisitNullCheck(HNullCheck* check) { check->ReplaceWith(check->InputAt(0)); if (compiler_options_.GetImplicitNullChecks()) { HInstruction* next = check->GetNext(); // The `PrepareForRegisterAllocation` pass removes `HBoundType` from the graph, // so do it ourselves now to not prevent optimizations. while (next->IsBoundType()) { next = next->GetNext(); VisitBoundType(next->GetPrevious()->AsBoundType()); } if (next->CanDoImplicitNullCheckOn(check->InputAt(0))) { check->MarkEmittedAtUseSite(); } } } void PrepareForRegisterAllocation::VisitDivZeroCheck(HDivZeroCheck* check) { check->ReplaceWith(check->InputAt(0)); } void PrepareForRegisterAllocation::VisitDeoptimize(HDeoptimize* deoptimize) { if (deoptimize->GuardsAnInput()) { // Replace the uses with the actual guarded instruction. deoptimize->ReplaceWith(deoptimize->GuardedInput()); deoptimize->RemoveGuard(); } } void PrepareForRegisterAllocation::VisitBoundsCheck(HBoundsCheck* check) { check->ReplaceWith(check->InputAt(0)); if (check->IsStringCharAt()) { // Add a fake environment for String.charAt() inline info as we want the exception // to appear as being thrown from there. Skip if we're compiling String.charAt() itself. ArtMethod* char_at_method = jni::DecodeArtMethod(WellKnownClasses::java_lang_String_charAt); if (GetGraph()->GetArtMethod() != char_at_method) { ArenaAllocator* allocator = GetGraph()->GetAllocator(); HEnvironment* environment = new (allocator) HEnvironment(allocator, /* number_of_vregs= */ 0u, char_at_method, /* dex_pc= */ dex::kDexNoIndex, check); check->InsertRawEnvironment(environment); } } } void PrepareForRegisterAllocation::VisitBoundType(HBoundType* bound_type) { bound_type->ReplaceWith(bound_type->InputAt(0)); bound_type->GetBlock()->RemoveInstruction(bound_type); } void PrepareForRegisterAllocation::VisitArraySet(HArraySet* instruction) { HInstruction* value = instruction->GetValue(); // PrepareForRegisterAllocation::VisitBoundType may have replaced a // BoundType (as value input of this ArraySet) with a NullConstant. // If so, this ArraySet no longer needs a type check. if (value->IsNullConstant()) { DCHECK_EQ(value->GetType(), DataType::Type::kReference); if (instruction->NeedsTypeCheck()) { instruction->ClearNeedsTypeCheck(); } } } void PrepareForRegisterAllocation::VisitClinitCheck(HClinitCheck* check) { // Try to find a static invoke or a new-instance from which this check originated. HInstruction* implicit_clinit = nullptr; for (const HUseListNode<HInstruction*>& use : check->GetUses()) { HInstruction* user = use.GetUser(); if ((user->IsInvokeStaticOrDirect() || user->IsNewInstance()) && CanMoveClinitCheck(check, user)) { implicit_clinit = user; if (user->IsInvokeStaticOrDirect()) { DCHECK(user->AsInvokeStaticOrDirect()->IsStaticWithExplicitClinitCheck()); user->AsInvokeStaticOrDirect()->RemoveExplicitClinitCheck( HInvokeStaticOrDirect::ClinitCheckRequirement::kImplicit); } else { DCHECK(user->IsNewInstance()); // We delegate the initialization duty to the allocation. if (user->AsNewInstance()->GetEntrypoint() == kQuickAllocObjectInitialized) { user->AsNewInstance()->SetEntrypoint(kQuickAllocObjectResolved); } } break; } } // If we found a static invoke or new-instance for merging, remove the check // from dominated static invokes. if (implicit_clinit != nullptr) { const HUseList<HInstruction*>& uses = check->GetUses(); for (auto it = uses.begin(), end = uses.end(); it != end; /* ++it below */) { HInstruction* user = it->GetUser(); // All other uses must be dominated. DCHECK(implicit_clinit->StrictlyDominates(user) || (implicit_clinit == user)); ++it; // Advance before we remove the node, reference to the next node is preserved. if (user->IsInvokeStaticOrDirect()) { user->AsInvokeStaticOrDirect()->RemoveExplicitClinitCheck( HInvokeStaticOrDirect::ClinitCheckRequirement::kNone); } } } HLoadClass* load_class = check->GetLoadClass(); bool can_merge_with_load_class = CanMoveClinitCheck(load_class, check); check->ReplaceWith(load_class); if (implicit_clinit != nullptr) { // Remove the check from the graph. It has been merged into the invoke or new-instance. check->GetBlock()->RemoveInstruction(check); // Check if we can merge the load class as well. if (can_merge_with_load_class && !load_class->HasUses()) { load_class->GetBlock()->RemoveInstruction(load_class); } } else if (can_merge_with_load_class && load_class->GetLoadKind() != HLoadClass::LoadKind::kRuntimeCall) { DCHECK(!load_class->NeedsAccessCheck()); // Pass the initialization duty to the `HLoadClass` instruction, // and remove the instruction from the graph. DCHECK(load_class->HasEnvironment()); load_class->SetMustGenerateClinitCheck(true); check->GetBlock()->RemoveInstruction(check); } } bool PrepareForRegisterAllocation::CanEmitConditionAt(HCondition* condition, HInstruction* user) const { if (condition->GetNext() != user) { return false; } if (user->IsIf() || user->IsDeoptimize()) { return true; } if (user->IsSelect() && user->AsSelect()->GetCondition() == condition) { return true; } return false; } void PrepareForRegisterAllocation::VisitCondition(HCondition* condition) { if (condition->HasOnlyOneNonEnvironmentUse()) { HInstruction* user = condition->GetUses().front().GetUser(); if (CanEmitConditionAt(condition, user)) { condition->MarkEmittedAtUseSite(); } } } void PrepareForRegisterAllocation::VisitConstructorFence(HConstructorFence* constructor_fence) { // Trivially remove redundant HConstructorFence when it immediately follows an HNewInstance // to an uninitialized class. In this special case, the art_quick_alloc_object_resolved // will already have the 'dmb' which is strictly stronger than an HConstructorFence. // // The instruction builder always emits "x = HNewInstance; HConstructorFence(x)" so this // is effectively pattern-matching that particular case and undoing the redundancy the builder // had introduced. // // TODO: Move this to a separate pass. HInstruction* allocation_inst = constructor_fence->GetAssociatedAllocation(); if (allocation_inst != nullptr && allocation_inst->IsNewInstance()) { HNewInstance* new_inst = allocation_inst->AsNewInstance(); // This relies on the entrypoint already being set to the more optimized version; // as that happens in this pass, this redundancy removal also cannot happen any earlier. if (new_inst != nullptr && new_inst->GetEntrypoint() == kQuickAllocObjectResolved) { // If this was done in an earlier pass, we would want to match that `previous` was an input // to the `constructor_fence`. However, since this pass removes the inputs to the fence, // we can ignore the inputs and just remove the instruction from its block. DCHECK_EQ(1u, constructor_fence->InputCount()); // TODO: GetAssociatedAllocation should not care about multiple inputs // if we are in prepare_for_register_allocation pass only. constructor_fence->GetBlock()->RemoveInstruction(constructor_fence); MaybeRecordStat(stats_, MethodCompilationStat::kConstructorFenceRemovedPFRA); return; } // HNewArray does not need this check because the art_quick_alloc_array does not itself // have a dmb in any normal situation (i.e. the array class is never exactly in the // "resolved" state). If the array class is not yet loaded, it will always go from // Unloaded->Initialized state. } // Remove all the inputs to the constructor fence; // they aren't used by the InstructionCodeGenerator and this lets us avoid creating a // LocationSummary in the LocationsBuilder. constructor_fence->RemoveAllInputs(); } void PrepareForRegisterAllocation::VisitInvokeStaticOrDirect(HInvokeStaticOrDirect* invoke) { if (invoke->IsStaticWithExplicitClinitCheck()) { HInstruction* last_input = invoke->GetInputs().back(); DCHECK(last_input->IsLoadClass()) << "Last input is not HLoadClass. It is " << last_input->DebugName(); // Detach the explicit class initialization check from the invoke. // Keeping track of the initializing instruction is no longer required // at this stage (i.e., after inlining has been performed). invoke->RemoveExplicitClinitCheck(HInvokeStaticOrDirect::ClinitCheckRequirement::kNone); // Merging with load class should have happened in VisitClinitCheck(). DCHECK(!CanMoveClinitCheck(last_input, invoke)); } } bool PrepareForRegisterAllocation::CanMoveClinitCheck(HInstruction* input, HInstruction* user) const { // Determine if input and user come from the same dex instruction, so that we can move // the clinit check responsibility from one to the other, i.e. from HClinitCheck (user) // to HLoadClass (input), or from HClinitCheck (input) to HInvokeStaticOrDirect (user), // or from HLoadClass (input) to HNewInstance (user). // Start with a quick dex pc check. if (user->GetDexPc() != input->GetDexPc()) { return false; } // Now do a thorough environment check that this is really coming from the same instruction in // the same inlined graph. Unfortunately, we have to go through the whole environment chain. HEnvironment* user_environment = user->GetEnvironment(); HEnvironment* input_environment = input->GetEnvironment(); while (user_environment != nullptr || input_environment != nullptr) { if (user_environment == nullptr || input_environment == nullptr) { // Different environment chain length. This happens when a method is called // once directly and once indirectly through another inlined method. return false; } if (user_environment->GetDexPc() != input_environment->GetDexPc() || user_environment->GetMethod() != input_environment->GetMethod()) { return false; } user_environment = user_environment->GetParent(); input_environment = input_environment->GetParent(); } // Check for code motion taking the input to a different block. if (user->GetBlock() != input->GetBlock()) { return false; } // In debug mode, check that we have not inserted a throwing instruction // or an instruction with side effects between input and user. if (kIsDebugBuild) { for (HInstruction* between = input->GetNext(); between != user; between = between->GetNext()) { CHECK(between != nullptr); // User must be after input in the same block. CHECK(!between->CanThrow()); CHECK(!between->HasSideEffects()); } } return true; } void PrepareForRegisterAllocation::VisitTypeConversion(HTypeConversion* instruction) { // For simplicity, our code generators don't handle implicit type conversion, so ensure // there are none before hitting codegen. if (instruction->IsImplicitConversion()) { instruction->ReplaceWith(instruction->GetInput()); instruction->GetBlock()->RemoveInstruction(instruction); } } } // namespace art