/*
* Copyright (C) 2017 The Android Open Source Project
*
* Licensed under the Apache License, Version 2.0 (the "License");
* you may not use this file except in compliance with the License.
* You may obtain a copy of the License at
*
* http://www.apache.org/licenses/LICENSE-2.0
*
* Unless required by applicable law or agreed to in writing, software
* distributed under the License is distributed on an "AS IS" BASIS,
* WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
* See the License for the specific language governing permissions and
* limitations under the License.
*/
#ifndef ART_COMPILER_OPTIMIZING_SCHEDULER_ARM_H_
#define ART_COMPILER_OPTIMIZING_SCHEDULER_ARM_H_
#include "code_generator_arm_vixl.h"
#include "scheduler.h"
namespace art {
namespace arm {
// TODO: Replace CodeGeneratorARMType with CodeGeneratorARMVIXL everywhere?
typedef CodeGeneratorARMVIXL CodeGeneratorARMType;
// AArch32 instruction latencies.
// We currently assume that all ARM CPUs share the same instruction latency list.
// The following latencies were tuned based on performance experiments and
// automatic tuning using differential evolution approach on various benchmarks.
static constexpr uint32_t kArmIntegerOpLatency = 2;
static constexpr uint32_t kArmFloatingPointOpLatency = 11;
static constexpr uint32_t kArmDataProcWithShifterOpLatency = 4;
static constexpr uint32_t kArmMulIntegerLatency = 6;
static constexpr uint32_t kArmMulFloatingPointLatency = 11;
static constexpr uint32_t kArmDivIntegerLatency = 10;
static constexpr uint32_t kArmDivFloatLatency = 20;
static constexpr uint32_t kArmDivDoubleLatency = 25;
static constexpr uint32_t kArmTypeConversionFloatingPointIntegerLatency = 11;
static constexpr uint32_t kArmMemoryLoadLatency = 9;
static constexpr uint32_t kArmMemoryStoreLatency = 9;
static constexpr uint32_t kArmMemoryBarrierLatency = 6;
static constexpr uint32_t kArmBranchLatency = 4;
static constexpr uint32_t kArmCallLatency = 5;
static constexpr uint32_t kArmCallInternalLatency = 29;
static constexpr uint32_t kArmLoadStringInternalLatency = 10;
static constexpr uint32_t kArmNopLatency = 2;
static constexpr uint32_t kArmLoadWithBakerReadBarrierLatency = 18;
static constexpr uint32_t kArmRuntimeTypeCheckLatency = 46;
class SchedulingLatencyVisitorARM : public SchedulingLatencyVisitor {
public:
explicit SchedulingLatencyVisitorARM(CodeGenerator* codegen)
: codegen_(down_cast<CodeGeneratorARMType*>(codegen)) {}
// Default visitor for instructions not handled specifically below.
void VisitInstruction(HInstruction* ATTRIBUTE_UNUSED) {
last_visited_latency_ = kArmIntegerOpLatency;
}
// We add a second unused parameter to be able to use this macro like the others
// defined in `nodes.h`.
#define FOR_EACH_SCHEDULED_ARM_INSTRUCTION(M) \
M(ArrayGet , unused) \
M(ArrayLength , unused) \
M(ArraySet , unused) \
M(Add , unused) \
M(Sub , unused) \
M(And , unused) \
M(Or , unused) \
M(Ror , unused) \
M(Xor , unused) \
M(Shl , unused) \
M(Shr , unused) \
M(UShr , unused) \
M(Mul , unused) \
M(Div , unused) \
M(Condition , unused) \
M(Compare , unused) \
M(BoundsCheck , unused) \
M(InstanceFieldGet , unused) \
M(InstanceFieldSet , unused) \
M(InstanceOf , unused) \
M(Invoke , unused) \
M(LoadString , unused) \
M(NewArray , unused) \
M(NewInstance , unused) \
M(Rem , unused) \
M(StaticFieldGet , unused) \
M(StaticFieldSet , unused) \
M(SuspendCheck , unused) \
M(TypeConversion , unused)
#define FOR_EACH_SCHEDULED_SHARED_INSTRUCTION(M) \
M(BitwiseNegatedRight, unused) \
M(MultiplyAccumulate, unused) \
M(IntermediateAddress, unused) \
M(IntermediateAddressIndex, unused) \
M(DataProcWithShifterOp, unused)
#define DECLARE_VISIT_INSTRUCTION(type, unused) \
void Visit##type(H##type* instruction) OVERRIDE;
FOR_EACH_SCHEDULED_ARM_INSTRUCTION(DECLARE_VISIT_INSTRUCTION)
FOR_EACH_SCHEDULED_SHARED_INSTRUCTION(DECLARE_VISIT_INSTRUCTION)
FOR_EACH_CONCRETE_INSTRUCTION_ARM(DECLARE_VISIT_INSTRUCTION)
#undef DECLARE_VISIT_INSTRUCTION
private:
bool CanGenerateTest(HCondition* cond);
void HandleGenerateConditionWithZero(IfCondition cond);
void HandleGenerateLongTestConstant(HCondition* cond);
void HandleGenerateLongTest(HCondition* cond);
void HandleGenerateLongComparesAndJumps();
void HandleGenerateTest(HCondition* cond);
void HandleGenerateConditionGeneric(HCondition* cond);
void HandleGenerateEqualLong(HCondition* cond);
void HandleGenerateConditionLong(HCondition* cond);
void HandleGenerateConditionIntegralOrNonPrimitive(HCondition* cond);
void HandleCondition(HCondition* instr);
void HandleBinaryOperationLantencies(HBinaryOperation* instr);
void HandleBitwiseOperationLantencies(HBinaryOperation* instr);
void HandleShiftLatencies(HBinaryOperation* instr);
void HandleDivRemConstantIntegralLatencies(int32_t imm);
void HandleFieldSetLatencies(HInstruction* instruction, const FieldInfo& field_info);
void HandleFieldGetLatencies(HInstruction* instruction, const FieldInfo& field_info);
void HandleGenerateDataProcInstruction(bool internal_latency = false);
void HandleGenerateDataProc(HDataProcWithShifterOp* instruction);
void HandleGenerateLongDataProc(HDataProcWithShifterOp* instruction);
// The latency setting for each HInstruction depends on how CodeGenerator may generate code,
// latency visitors may query CodeGenerator for such information for accurate latency settings.
CodeGeneratorARMType* codegen_;
};
class HSchedulerARM : public HScheduler {
public:
HSchedulerARM(ScopedArenaAllocator* allocator,
SchedulingNodeSelector* selector,
SchedulingLatencyVisitorARM* arm_latency_visitor)
: HScheduler(allocator, arm_latency_visitor, selector) {}
~HSchedulerARM() OVERRIDE {}
bool IsSchedulable(const HInstruction* instruction) const OVERRIDE {
#define CASE_INSTRUCTION_KIND(type, unused) case \
HInstruction::InstructionKind::k##type:
switch (instruction->GetKind()) {
FOR_EACH_SCHEDULED_SHARED_INSTRUCTION(CASE_INSTRUCTION_KIND)
return true;
FOR_EACH_CONCRETE_INSTRUCTION_ARM(CASE_INSTRUCTION_KIND)
return true;
default:
return HScheduler::IsSchedulable(instruction);
}
#undef CASE_INSTRUCTION_KIND
}
private:
DISALLOW_COPY_AND_ASSIGN(HSchedulerARM);
};
} // namespace arm
} // namespace art
#endif // ART_COMPILER_OPTIMIZING_SCHEDULER_ARM_H_