/*
* Copyright (C) 2011 The Android Open Source Project
*
* Licensed under the Apache License, Version 2.0 (the "License");
* you may not use this file except in compliance with the License.
* You may obtain a copy of the License at
*
* http://www.apache.org/licenses/LICENSE-2.0
*
* Unless required by applicable law or agreed to in writing, software
* distributed under the License is distributed on an "AS IS" BASIS,
* WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
* See the License for the specific language governing permissions and
* limitations under the License.
*/
#include <cstdint>
#include "compiler.h"
#include "compiler_internals.h"
#include "driver/compiler_driver.h"
#include "driver/compiler_options.h"
#include "dataflow_iterator-inl.h"
#include "leb128.h"
#include "mirror/object.h"
#include "pass_driver_me_opts.h"
#include "runtime.h"
#include "base/logging.h"
#include "base/timing_logger.h"
#include "driver/compiler_options.h"
#include "dex/quick/dex_file_to_method_inliner_map.h"
namespace art {
extern "C" void ArtInitQuickCompilerContext(art::CompilerDriver* driver) {
CHECK(driver->GetCompilerContext() == nullptr);
}
extern "C" void ArtUnInitQuickCompilerContext(art::CompilerDriver* driver) {
CHECK(driver->GetCompilerContext() == nullptr);
}
/* Default optimizer/debug setting for the compiler. */
static uint32_t kCompilerOptimizerDisableFlags = 0 | // Disable specific optimizations
(1 << kLoadStoreElimination) |
// (1 << kLoadHoisting) |
// (1 << kSuppressLoads) |
// (1 << kNullCheckElimination) |
// (1 << kClassInitCheckElimination) |
(1 << kGlobalValueNumbering) |
// (1 << kPromoteRegs) |
// (1 << kTrackLiveTemps) |
// (1 << kSafeOptimizations) |
// (1 << kBBOpt) |
// (1 << kMatch) |
// (1 << kPromoteCompilerTemps) |
// (1 << kSuppressExceptionEdges) |
// (1 << kSuppressMethodInlining) |
0;
static uint32_t kCompilerDebugFlags = 0 | // Enable debug/testing modes
// (1 << kDebugDisplayMissingTargets) |
// (1 << kDebugVerbose) |
// (1 << kDebugDumpCFG) |
// (1 << kDebugSlowFieldPath) |
// (1 << kDebugSlowInvokePath) |
// (1 << kDebugSlowStringPath) |
// (1 << kDebugSlowestFieldPath) |
// (1 << kDebugSlowestStringPath) |
// (1 << kDebugExerciseResolveMethod) |
// (1 << kDebugVerifyDataflow) |
// (1 << kDebugShowMemoryUsage) |
// (1 << kDebugShowNops) |
// (1 << kDebugCountOpcodes) |
// (1 << kDebugDumpCheckStats) |
// (1 << kDebugDumpBitcodeFile) |
// (1 << kDebugVerifyBitcode) |
// (1 << kDebugShowSummaryMemoryUsage) |
// (1 << kDebugShowFilterStats) |
// (1 << kDebugTimings) |
// (1 << kDebugCodegenDump) |
0;
COMPILE_ASSERT(0U == static_cast<size_t>(kNone), kNone_not_0);
COMPILE_ASSERT(1U == static_cast<size_t>(kArm), kArm_not_1);
COMPILE_ASSERT(2U == static_cast<size_t>(kArm64), kArm64_not_2);
COMPILE_ASSERT(3U == static_cast<size_t>(kThumb2), kThumb2_not_3);
COMPILE_ASSERT(4U == static_cast<size_t>(kX86), kX86_not_4);
COMPILE_ASSERT(5U == static_cast<size_t>(kX86_64), kX86_64_not_5);
COMPILE_ASSERT(6U == static_cast<size_t>(kMips), kMips_not_6);
COMPILE_ASSERT(7U == static_cast<size_t>(kMips64), kMips64_not_7);
// Additional disabled optimizations (over generally disabled) per instruction set.
static constexpr uint32_t kDisabledOptimizationsPerISA[] = {
// 0 = kNone.
~0U,
// 1 = kArm, unused (will use kThumb2).
~0U,
// 2 = kArm64.
0,
// 3 = kThumb2.
0,
// 4 = kX86.
(1 << kLoadStoreElimination) |
0,
// 5 = kX86_64.
(1 << kLoadStoreElimination) |
0,
// 6 = kMips.
(1 << kLoadStoreElimination) |
(1 << kLoadHoisting) |
(1 << kSuppressLoads) |
(1 << kNullCheckElimination) |
(1 << kPromoteRegs) |
(1 << kTrackLiveTemps) |
(1 << kSafeOptimizations) |
(1 << kBBOpt) |
(1 << kMatch) |
(1 << kPromoteCompilerTemps) |
0,
// 7 = kMips64.
~0U
};
COMPILE_ASSERT(sizeof(kDisabledOptimizationsPerISA) == 8 * sizeof(uint32_t), kDisabledOpts_unexp);
// Supported shorty types per instruction set. nullptr means that all are available.
// Z : boolean
// B : byte
// S : short
// C : char
// I : int
// J : long
// F : float
// D : double
// L : reference(object, array)
// V : void
static const char* kSupportedTypes[] = {
// 0 = kNone.
"",
// 1 = kArm, unused (will use kThumb2).
"",
// 2 = kArm64.
nullptr,
// 3 = kThumb2.
nullptr,
// 4 = kX86.
nullptr,
// 5 = kX86_64.
nullptr,
// 6 = kMips.
nullptr,
// 7 = kMips64.
""
};
COMPILE_ASSERT(sizeof(kSupportedTypes) == 8 * sizeof(char*), kSupportedTypes_unexp);
static int kAllOpcodes[] = {
Instruction::NOP,
Instruction::MOVE,
Instruction::MOVE_FROM16,
Instruction::MOVE_16,
Instruction::MOVE_WIDE,
Instruction::MOVE_WIDE_FROM16,
Instruction::MOVE_WIDE_16,
Instruction::MOVE_OBJECT,
Instruction::MOVE_OBJECT_FROM16,
Instruction::MOVE_OBJECT_16,
Instruction::MOVE_RESULT,
Instruction::MOVE_RESULT_WIDE,
Instruction::MOVE_RESULT_OBJECT,
Instruction::MOVE_EXCEPTION,
Instruction::RETURN_VOID,
Instruction::RETURN,
Instruction::RETURN_WIDE,
Instruction::RETURN_OBJECT,
Instruction::CONST_4,
Instruction::CONST_16,
Instruction::CONST,
Instruction::CONST_HIGH16,
Instruction::CONST_WIDE_16,
Instruction::CONST_WIDE_32,
Instruction::CONST_WIDE,
Instruction::CONST_WIDE_HIGH16,
Instruction::CONST_STRING,
Instruction::CONST_STRING_JUMBO,
Instruction::CONST_CLASS,
Instruction::MONITOR_ENTER,
Instruction::MONITOR_EXIT,
Instruction::CHECK_CAST,
Instruction::INSTANCE_OF,
Instruction::ARRAY_LENGTH,
Instruction::NEW_INSTANCE,
Instruction::NEW_ARRAY,
Instruction::FILLED_NEW_ARRAY,
Instruction::FILLED_NEW_ARRAY_RANGE,
Instruction::FILL_ARRAY_DATA,
Instruction::THROW,
Instruction::GOTO,
Instruction::GOTO_16,
Instruction::GOTO_32,
Instruction::PACKED_SWITCH,
Instruction::SPARSE_SWITCH,
Instruction::CMPL_FLOAT,
Instruction::CMPG_FLOAT,
Instruction::CMPL_DOUBLE,
Instruction::CMPG_DOUBLE,
Instruction::CMP_LONG,
Instruction::IF_EQ,
Instruction::IF_NE,
Instruction::IF_LT,
Instruction::IF_GE,
Instruction::IF_GT,
Instruction::IF_LE,
Instruction::IF_EQZ,
Instruction::IF_NEZ,
Instruction::IF_LTZ,
Instruction::IF_GEZ,
Instruction::IF_GTZ,
Instruction::IF_LEZ,
Instruction::UNUSED_3E,
Instruction::UNUSED_3F,
Instruction::UNUSED_40,
Instruction::UNUSED_41,
Instruction::UNUSED_42,
Instruction::UNUSED_43,
Instruction::AGET,
Instruction::AGET_WIDE,
Instruction::AGET_OBJECT,
Instruction::AGET_BOOLEAN,
Instruction::AGET_BYTE,
Instruction::AGET_CHAR,
Instruction::AGET_SHORT,
Instruction::APUT,
Instruction::APUT_WIDE,
Instruction::APUT_OBJECT,
Instruction::APUT_BOOLEAN,
Instruction::APUT_BYTE,
Instruction::APUT_CHAR,
Instruction::APUT_SHORT,
Instruction::IGET,
Instruction::IGET_WIDE,
Instruction::IGET_OBJECT,
Instruction::IGET_BOOLEAN,
Instruction::IGET_BYTE,
Instruction::IGET_CHAR,
Instruction::IGET_SHORT,
Instruction::IPUT,
Instruction::IPUT_WIDE,
Instruction::IPUT_OBJECT,
Instruction::IPUT_BOOLEAN,
Instruction::IPUT_BYTE,
Instruction::IPUT_CHAR,
Instruction::IPUT_SHORT,
Instruction::SGET,
Instruction::SGET_WIDE,
Instruction::SGET_OBJECT,
Instruction::SGET_BOOLEAN,
Instruction::SGET_BYTE,
Instruction::SGET_CHAR,
Instruction::SGET_SHORT,
Instruction::SPUT,
Instruction::SPUT_WIDE,
Instruction::SPUT_OBJECT,
Instruction::SPUT_BOOLEAN,
Instruction::SPUT_BYTE,
Instruction::SPUT_CHAR,
Instruction::SPUT_SHORT,
Instruction::INVOKE_VIRTUAL,
Instruction::INVOKE_SUPER,
Instruction::INVOKE_DIRECT,
Instruction::INVOKE_STATIC,
Instruction::INVOKE_INTERFACE,
Instruction::RETURN_VOID_BARRIER,
Instruction::INVOKE_VIRTUAL_RANGE,
Instruction::INVOKE_SUPER_RANGE,
Instruction::INVOKE_DIRECT_RANGE,
Instruction::INVOKE_STATIC_RANGE,
Instruction::INVOKE_INTERFACE_RANGE,
Instruction::UNUSED_79,
Instruction::UNUSED_7A,
Instruction::NEG_INT,
Instruction::NOT_INT,
Instruction::NEG_LONG,
Instruction::NOT_LONG,
Instruction::NEG_FLOAT,
Instruction::NEG_DOUBLE,
Instruction::INT_TO_LONG,
Instruction::INT_TO_FLOAT,
Instruction::INT_TO_DOUBLE,
Instruction::LONG_TO_INT,
Instruction::LONG_TO_FLOAT,
Instruction::LONG_TO_DOUBLE,
Instruction::FLOAT_TO_INT,
Instruction::FLOAT_TO_LONG,
Instruction::FLOAT_TO_DOUBLE,
Instruction::DOUBLE_TO_INT,
Instruction::DOUBLE_TO_LONG,
Instruction::DOUBLE_TO_FLOAT,
Instruction::INT_TO_BYTE,
Instruction::INT_TO_CHAR,
Instruction::INT_TO_SHORT,
Instruction::ADD_INT,
Instruction::SUB_INT,
Instruction::MUL_INT,
Instruction::DIV_INT,
Instruction::REM_INT,
Instruction::AND_INT,
Instruction::OR_INT,
Instruction::XOR_INT,
Instruction::SHL_INT,
Instruction::SHR_INT,
Instruction::USHR_INT,
Instruction::ADD_LONG,
Instruction::SUB_LONG,
Instruction::MUL_LONG,
Instruction::DIV_LONG,
Instruction::REM_LONG,
Instruction::AND_LONG,
Instruction::OR_LONG,
Instruction::XOR_LONG,
Instruction::SHL_LONG,
Instruction::SHR_LONG,
Instruction::USHR_LONG,
Instruction::ADD_FLOAT,
Instruction::SUB_FLOAT,
Instruction::MUL_FLOAT,
Instruction::DIV_FLOAT,
Instruction::REM_FLOAT,
Instruction::ADD_DOUBLE,
Instruction::SUB_DOUBLE,
Instruction::MUL_DOUBLE,
Instruction::DIV_DOUBLE,
Instruction::REM_DOUBLE,
Instruction::ADD_INT_2ADDR,
Instruction::SUB_INT_2ADDR,
Instruction::MUL_INT_2ADDR,
Instruction::DIV_INT_2ADDR,
Instruction::REM_INT_2ADDR,
Instruction::AND_INT_2ADDR,
Instruction::OR_INT_2ADDR,
Instruction::XOR_INT_2ADDR,
Instruction::SHL_INT_2ADDR,
Instruction::SHR_INT_2ADDR,
Instruction::USHR_INT_2ADDR,
Instruction::ADD_LONG_2ADDR,
Instruction::SUB_LONG_2ADDR,
Instruction::MUL_LONG_2ADDR,
Instruction::DIV_LONG_2ADDR,
Instruction::REM_LONG_2ADDR,
Instruction::AND_LONG_2ADDR,
Instruction::OR_LONG_2ADDR,
Instruction::XOR_LONG_2ADDR,
Instruction::SHL_LONG_2ADDR,
Instruction::SHR_LONG_2ADDR,
Instruction::USHR_LONG_2ADDR,
Instruction::ADD_FLOAT_2ADDR,
Instruction::SUB_FLOAT_2ADDR,
Instruction::MUL_FLOAT_2ADDR,
Instruction::DIV_FLOAT_2ADDR,
Instruction::REM_FLOAT_2ADDR,
Instruction::ADD_DOUBLE_2ADDR,
Instruction::SUB_DOUBLE_2ADDR,
Instruction::MUL_DOUBLE_2ADDR,
Instruction::DIV_DOUBLE_2ADDR,
Instruction::REM_DOUBLE_2ADDR,
Instruction::ADD_INT_LIT16,
Instruction::RSUB_INT,
Instruction::MUL_INT_LIT16,
Instruction::DIV_INT_LIT16,
Instruction::REM_INT_LIT16,
Instruction::AND_INT_LIT16,
Instruction::OR_INT_LIT16,
Instruction::XOR_INT_LIT16,
Instruction::ADD_INT_LIT8,
Instruction::RSUB_INT_LIT8,
Instruction::MUL_INT_LIT8,
Instruction::DIV_INT_LIT8,
Instruction::REM_INT_LIT8,
Instruction::AND_INT_LIT8,
Instruction::OR_INT_LIT8,
Instruction::XOR_INT_LIT8,
Instruction::SHL_INT_LIT8,
Instruction::SHR_INT_LIT8,
Instruction::USHR_INT_LIT8,
Instruction::IGET_QUICK,
Instruction::IGET_WIDE_QUICK,
Instruction::IGET_OBJECT_QUICK,
Instruction::IPUT_QUICK,
Instruction::IPUT_WIDE_QUICK,
Instruction::IPUT_OBJECT_QUICK,
Instruction::INVOKE_VIRTUAL_QUICK,
Instruction::INVOKE_VIRTUAL_RANGE_QUICK,
Instruction::UNUSED_EB,
Instruction::UNUSED_EC,
Instruction::UNUSED_ED,
Instruction::UNUSED_EE,
Instruction::UNUSED_EF,
Instruction::UNUSED_F0,
Instruction::UNUSED_F1,
Instruction::UNUSED_F2,
Instruction::UNUSED_F3,
Instruction::UNUSED_F4,
Instruction::UNUSED_F5,
Instruction::UNUSED_F6,
Instruction::UNUSED_F7,
Instruction::UNUSED_F8,
Instruction::UNUSED_F9,
Instruction::UNUSED_FA,
Instruction::UNUSED_FB,
Instruction::UNUSED_FC,
Instruction::UNUSED_FD,
Instruction::UNUSED_FE,
Instruction::UNUSED_FF,
// ----- ExtendedMIROpcode -----
kMirOpPhi,
kMirOpCopy,
kMirOpFusedCmplFloat,
kMirOpFusedCmpgFloat,
kMirOpFusedCmplDouble,
kMirOpFusedCmpgDouble,
kMirOpFusedCmpLong,
kMirOpNop,
kMirOpNullCheck,
kMirOpRangeCheck,
kMirOpDivZeroCheck,
kMirOpCheck,
kMirOpCheckPart2,
kMirOpSelect,
};
// Unsupported opcodes. nullptr can be used when everything is supported. Size of the lists is
// recorded below.
static const int* kUnsupportedOpcodes[] = {
// 0 = kNone.
kAllOpcodes,
// 1 = kArm, unused (will use kThumb2).
kAllOpcodes,
// 2 = kArm64.
nullptr,
// 3 = kThumb2.
nullptr,
// 4 = kX86.
nullptr,
// 5 = kX86_64.
nullptr,
// 6 = kMips.
nullptr,
// 7 = kMips64.
kAllOpcodes
};
COMPILE_ASSERT(sizeof(kUnsupportedOpcodes) == 8 * sizeof(int*), kUnsupportedOpcodes_unexp);
// Size of the arrays stored above.
static const size_t kUnsupportedOpcodesSize[] = {
// 0 = kNone.
arraysize(kAllOpcodes),
// 1 = kArm, unused (will use kThumb2).
arraysize(kAllOpcodes),
// 2 = kArm64.
0,
// 3 = kThumb2.
0,
// 4 = kX86.
0,
// 5 = kX86_64.
0,
// 6 = kMips.
0,
// 7 = kMips64.
arraysize(kAllOpcodes),
};
COMPILE_ASSERT(sizeof(kUnsupportedOpcodesSize) == 8 * sizeof(size_t),
kUnsupportedOpcodesSize_unexp);
// The maximum amount of Dalvik register in a method for which we will start compiling. Tries to
// avoid an abort when we need to manage more SSA registers than we can.
static constexpr size_t kMaxAllowedDalvikRegisters = INT16_MAX / 2;
CompilationUnit::CompilationUnit(ArenaPool* pool)
: compiler_driver(nullptr),
class_linker(nullptr),
dex_file(nullptr),
class_loader(nullptr),
class_def_idx(0),
method_idx(0),
code_item(nullptr),
access_flags(0),
invoke_type(kDirect),
shorty(nullptr),
disable_opt(0),
enable_debug(0),
verbose(false),
compiler(nullptr),
instruction_set(kNone),
target64(false),
num_dalvik_registers(0),
insns(nullptr),
num_ins(0),
num_outs(0),
num_regs(0),
compiler_flip_match(false),
arena(pool),
arena_stack(pool),
mir_graph(nullptr),
cg(nullptr),
timings("QuickCompiler", true, false),
print_pass(false) {
}
CompilationUnit::~CompilationUnit() {
}
void CompilationUnit::StartTimingSplit(const char* label) {
if (compiler_driver->GetDumpPasses()) {
timings.StartTiming(label);
}
}
void CompilationUnit::NewTimingSplit(const char* label) {
if (compiler_driver->GetDumpPasses()) {
timings.EndTiming();
timings.StartTiming(label);
}
}
void CompilationUnit::EndTiming() {
if (compiler_driver->GetDumpPasses()) {
timings.EndTiming();
if (enable_debug & (1 << kDebugTimings)) {
LOG(INFO) << "TIMINGS " << PrettyMethod(method_idx, *dex_file);
LOG(INFO) << Dumpable<TimingLogger>(timings);
}
}
}
static bool CanCompileShorty(const char* shorty, InstructionSet instruction_set) {
const char* supported_types = kSupportedTypes[instruction_set];
if (supported_types == nullptr) {
// Everything available.
return true;
}
uint32_t shorty_size = strlen(shorty);
CHECK_GE(shorty_size, 1u);
for (uint32_t i = 0; i < shorty_size; i++) {
if (strchr(supported_types, shorty[i]) == nullptr) {
return false;
}
}
return true;
};
// Skip the method that we do not support currently.
static bool CanCompileMethod(uint32_t method_idx, const DexFile& dex_file,
CompilationUnit& cu) {
// This is a limitation in mir_graph. See MirGraph::SetNumSSARegs.
if (cu.num_dalvik_registers > kMaxAllowedDalvikRegisters) {
VLOG(compiler) << "Too many dalvik registers : " << cu.num_dalvik_registers;
return false;
}
// Check whether we do have limitations at all.
if (kSupportedTypes[cu.instruction_set] == nullptr &&
kUnsupportedOpcodesSize[cu.instruction_set] == 0U) {
return true;
}
// Check if we can compile the prototype.
const char* shorty = dex_file.GetMethodShorty(dex_file.GetMethodId(method_idx));
if (!CanCompileShorty(shorty, cu.instruction_set)) {
VLOG(compiler) << "Unsupported shorty : " << shorty;
return false;
}
const int *unsupport_list = kUnsupportedOpcodes[cu.instruction_set];
int unsupport_list_size = kUnsupportedOpcodesSize[cu.instruction_set];
for (unsigned int idx = 0; idx < cu.mir_graph->GetNumBlocks(); idx++) {
BasicBlock* bb = cu.mir_graph->GetBasicBlock(idx);
if (bb == NULL) continue;
if (bb->block_type == kDead) continue;
for (MIR* mir = bb->first_mir_insn; mir != nullptr; mir = mir->next) {
int opcode = mir->dalvikInsn.opcode;
// Check if we support the byte code.
if (std::find(unsupport_list, unsupport_list + unsupport_list_size,
opcode) != unsupport_list + unsupport_list_size) {
if (!MIR::DecodedInstruction::IsPseudoMirOp(opcode)) {
VLOG(compiler) << "Unsupported dalvik byte code : "
<< mir->dalvikInsn.opcode;
} else {
VLOG(compiler) << "Unsupported extended MIR opcode : "
<< MIRGraph::extended_mir_op_names_[opcode - kMirOpFirst];
}
return false;
}
// Check if it invokes a prototype that we cannot support.
if (Instruction::INVOKE_VIRTUAL == opcode ||
Instruction::INVOKE_SUPER == opcode ||
Instruction::INVOKE_DIRECT == opcode ||
Instruction::INVOKE_STATIC == opcode ||
Instruction::INVOKE_INTERFACE == opcode) {
uint32_t invoke_method_idx = mir->dalvikInsn.vB;
const char* invoke_method_shorty = dex_file.GetMethodShorty(
dex_file.GetMethodId(invoke_method_idx));
if (!CanCompileShorty(invoke_method_shorty, cu.instruction_set)) {
VLOG(compiler) << "Unsupported to invoke '"
<< PrettyMethod(invoke_method_idx, dex_file)
<< "' with shorty : " << invoke_method_shorty;
return false;
}
}
}
}
return true;
}
static CompiledMethod* CompileMethod(CompilerDriver& driver,
Compiler* compiler,
const DexFile::CodeItem* code_item,
uint32_t access_flags, InvokeType invoke_type,
uint16_t class_def_idx, uint32_t method_idx,
jobject class_loader, const DexFile& dex_file,
void* llvm_compilation_unit) {
VLOG(compiler) << "Compiling " << PrettyMethod(method_idx, dex_file) << "...";
/*
* Skip compilation for pathologically large methods - either by instruction count or num vregs.
* Dalvik uses 16-bit uints for instruction and register counts. We'll limit to a quarter
* of that, which also guarantees we cannot overflow our 16-bit internal SSA name space.
*/
if (code_item->insns_size_in_code_units_ >= UINT16_MAX / 4) {
LOG(INFO) << "Method exceeds compiler instruction limit: "
<< code_item->insns_size_in_code_units_
<< " in " << PrettyMethod(method_idx, dex_file);
return NULL;
}
if (code_item->registers_size_ >= UINT16_MAX / 4) {
LOG(INFO) << "Method exceeds compiler virtual register limit: "
<< code_item->registers_size_ << " in " << PrettyMethod(method_idx, dex_file);
return NULL;
}
if (!driver.GetCompilerOptions().IsCompilationEnabled()) {
return nullptr;
}
ClassLinker* class_linker = Runtime::Current()->GetClassLinker();
CompilationUnit cu(driver.GetArenaPool());
cu.compiler_driver = &driver;
cu.class_linker = class_linker;
cu.instruction_set = driver.GetInstructionSet();
if (cu.instruction_set == kArm) {
cu.instruction_set = kThumb2;
}
cu.target64 = Is64BitInstructionSet(cu.instruction_set);
cu.compiler = compiler;
// TODO: Mips64 is not yet implemented.
CHECK((cu.instruction_set == kThumb2) ||
(cu.instruction_set == kArm64) ||
(cu.instruction_set == kX86) ||
(cu.instruction_set == kX86_64) ||
(cu.instruction_set == kMips));
/* Adjust this value accordingly once inlining is performed */
cu.num_dalvik_registers = code_item->registers_size_;
// TODO: set this from command line
cu.compiler_flip_match = false;
bool use_match = !cu.compiler_method_match.empty();
bool match = use_match && (cu.compiler_flip_match ^
(PrettyMethod(method_idx, dex_file).find(cu.compiler_method_match) != std::string::npos));
if (!use_match || match) {
cu.disable_opt = kCompilerOptimizerDisableFlags;
cu.enable_debug = kCompilerDebugFlags;
cu.verbose = VLOG_IS_ON(compiler) ||
(cu.enable_debug & (1 << kDebugVerbose));
}
if (gVerboseMethods.size() != 0) {
cu.verbose = false;
for (size_t i = 0; i < gVerboseMethods.size(); ++i) {
if (PrettyMethod(method_idx, dex_file).find(gVerboseMethods[i])
!= std::string::npos) {
cu.verbose = true;
break;
}
}
}
if (cu.verbose) {
cu.enable_debug |= (1 << kDebugCodegenDump);
}
/*
* TODO: rework handling of optimization and debug flags. Should we split out
* MIR and backend flags? Need command-line setting as well.
*/
compiler->InitCompilationUnit(cu);
// Disable optimizations according to instruction set.
cu.disable_opt |= kDisabledOptimizationsPerISA[cu.instruction_set];
cu.StartTimingSplit("BuildMIRGraph");
cu.mir_graph.reset(new MIRGraph(&cu, &cu.arena));
/*
* After creation of the MIR graph, also create the code generator.
* The reason we do this is that optimizations on the MIR graph may need to get information
* that is only available if a CG exists.
*/
cu.cg.reset(compiler->GetCodeGenerator(&cu, llvm_compilation_unit));
/* Gathering opcode stats? */
if (kCompilerDebugFlags & (1 << kDebugCountOpcodes)) {
cu.mir_graph->EnableOpcodeCounting();
}
/* Build the raw MIR graph */
cu.mir_graph->InlineMethod(code_item, access_flags, invoke_type, class_def_idx, method_idx,
class_loader, dex_file);
if (!CanCompileMethod(method_idx, dex_file, cu)) {
VLOG(compiler) << cu.instruction_set << ": Cannot compile method : "
<< PrettyMethod(method_idx, dex_file);
return nullptr;
}
cu.NewTimingSplit("MIROpt:CheckFilters");
std::string skip_message;
if (cu.mir_graph->SkipCompilation(&skip_message)) {
VLOG(compiler) << cu.instruction_set << ": Skipping method : "
<< PrettyMethod(method_idx, dex_file) << " Reason = " << skip_message;
return nullptr;
}
/* Create the pass driver and launch it */
PassDriverMEOpts pass_driver(&cu);
pass_driver.Launch();
/* For non-leaf methods check if we should skip compilation when the profiler is enabled. */
if (cu.compiler_driver->ProfilePresent()
&& !cu.mir_graph->MethodIsLeaf()
&& cu.mir_graph->SkipCompilationByName(PrettyMethod(method_idx, dex_file))) {
return nullptr;
}
if (cu.enable_debug & (1 << kDebugDumpCheckStats)) {
cu.mir_graph->DumpCheckStats();
}
if (kCompilerDebugFlags & (1 << kDebugCountOpcodes)) {
cu.mir_graph->ShowOpcodeStats();
}
/* Reassociate sreg names with original Dalvik vreg names. */
cu.mir_graph->RemapRegLocations();
/* Free Arenas from the cu.arena_stack for reuse by the cu.arena in the codegen. */
if (cu.enable_debug & (1 << kDebugShowMemoryUsage)) {
if (cu.arena_stack.PeakBytesAllocated() > 1 * 1024 * 1024) {
MemStats stack_stats(cu.arena_stack.GetPeakStats());
LOG(INFO) << PrettyMethod(method_idx, dex_file) << " " << Dumpable<MemStats>(stack_stats);
}
}
cu.arena_stack.Reset();
CompiledMethod* result = NULL;
if (cu.mir_graph->PuntToInterpreter()) {
VLOG(compiler) << cu.instruction_set << ": Punted method to interpreter: "
<< PrettyMethod(method_idx, dex_file);
return nullptr;
}
cu.cg->Materialize();
cu.NewTimingSplit("Dedupe"); /* deduping takes up the vast majority of time in GetCompiledMethod(). */
result = cu.cg->GetCompiledMethod();
cu.NewTimingSplit("Cleanup");
if (result) {
VLOG(compiler) << cu.instruction_set << ": Compiled " << PrettyMethod(method_idx, dex_file);
} else {
VLOG(compiler) << cu.instruction_set << ": Deferred " << PrettyMethod(method_idx, dex_file);
}
if (cu.enable_debug & (1 << kDebugShowMemoryUsage)) {
if (cu.arena.BytesAllocated() > (1 * 1024 *1024)) {
MemStats mem_stats(cu.arena.GetMemStats());
LOG(INFO) << PrettyMethod(method_idx, dex_file) << " " << Dumpable<MemStats>(mem_stats);
}
}
if (cu.enable_debug & (1 << kDebugShowSummaryMemoryUsage)) {
LOG(INFO) << "MEMINFO " << cu.arena.BytesAllocated() << " " << cu.mir_graph->GetNumBlocks()
<< " " << PrettyMethod(method_idx, dex_file);
}
cu.EndTiming();
driver.GetTimingsLogger()->AddLogger(cu.timings);
return result;
}
CompiledMethod* CompileOneMethod(CompilerDriver& driver,
Compiler* compiler,
const DexFile::CodeItem* code_item,
uint32_t access_flags,
InvokeType invoke_type,
uint16_t class_def_idx,
uint32_t method_idx,
jobject class_loader,
const DexFile& dex_file,
void* compilation_unit) {
return CompileMethod(driver, compiler, code_item, access_flags, invoke_type, class_def_idx,
method_idx, class_loader, dex_file, compilation_unit);
}
} // namespace art
extern "C" art::CompiledMethod*
ArtQuickCompileMethod(art::CompilerDriver& driver,
const art::DexFile::CodeItem* code_item,
uint32_t access_flags, art::InvokeType invoke_type,
uint16_t class_def_idx, uint32_t method_idx, jobject class_loader,
const art::DexFile& dex_file) {
// TODO: check method fingerprint here to determine appropriate backend type. Until then, use
// build default.
art::Compiler* compiler = driver.GetCompiler();
return art::CompileOneMethod(driver, compiler, code_item, access_flags, invoke_type,
class_def_idx, method_idx, class_loader, dex_file,
NULL /* use thread llvm_info */);
}