/* * Copyright (C) 2013 The Android Open Source Project * * Licensed under the Apache License, Version 2.0 (the "License"); * you may not use this file except in compliance with the License. * You may obtain a copy of the License at * * http://www.apache.org/licenses/LICENSE-2.0 * * Unless required by applicable law or agreed to in writing, software * distributed under the License is distributed on an "AS IS" BASIS, * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. * See the License for the specific language governing permissions and * limitations under the License. */ #include "trampoline_compiler.h" #include "base/arena_allocator.h" #include "jni_env_ext.h" #ifdef ART_ENABLE_CODEGEN_arm #include "utils/arm/assembler_arm_vixl.h" #endif #ifdef ART_ENABLE_CODEGEN_arm64 #include "utils/arm64/assembler_arm64.h" #endif #ifdef ART_ENABLE_CODEGEN_mips #include "utils/mips/assembler_mips.h" #endif #ifdef ART_ENABLE_CODEGEN_mips64 #include "utils/mips64/assembler_mips64.h" #endif #ifdef ART_ENABLE_CODEGEN_x86 #include "utils/x86/assembler_x86.h" #endif #ifdef ART_ENABLE_CODEGEN_x86_64 #include "utils/x86_64/assembler_x86_64.h" #endif #define __ assembler. namespace art { #ifdef ART_ENABLE_CODEGEN_arm namespace arm { #ifdef ___ #error "ARM Assembler macro already defined." #else #define ___ assembler.GetVIXLAssembler()-> #endif static std::unique_ptr<const std::vector<uint8_t>> CreateTrampoline( ArenaAllocator* arena, EntryPointCallingConvention abi, ThreadOffset32 offset) { using vixl::aarch32::MemOperand; using vixl::aarch32::pc; using vixl::aarch32::r0; ArmVIXLAssembler assembler(arena); switch (abi) { case kInterpreterAbi: // Thread* is first argument (R0) in interpreter ABI. ___ Ldr(pc, MemOperand(r0, offset.Int32Value())); break; case kJniAbi: { // Load via Thread* held in JNIEnv* in first argument (R0). vixl::aarch32::UseScratchRegisterScope temps(assembler.GetVIXLAssembler()); const vixl::aarch32::Register temp_reg = temps.Acquire(); // VIXL will use the destination as a scratch register if // the offset is not encodable as an immediate operand. ___ Ldr(temp_reg, MemOperand(r0, JNIEnvExt::SelfOffset(4).Int32Value())); ___ Ldr(pc, MemOperand(temp_reg, offset.Int32Value())); break; } case kQuickAbi: // TR holds Thread*. ___ Ldr(pc, MemOperand(tr, offset.Int32Value())); } __ FinalizeCode(); size_t cs = __ CodeSize(); std::unique_ptr<std::vector<uint8_t>> entry_stub(new std::vector<uint8_t>(cs)); MemoryRegion code(entry_stub->data(), entry_stub->size()); __ FinalizeInstructions(code); return std::move(entry_stub); } #undef ___ } // namespace arm #endif // ART_ENABLE_CODEGEN_arm #ifdef ART_ENABLE_CODEGEN_arm64 namespace arm64 { static std::unique_ptr<const std::vector<uint8_t>> CreateTrampoline( ArenaAllocator* arena, EntryPointCallingConvention abi, ThreadOffset64 offset) { Arm64Assembler assembler(arena); switch (abi) { case kInterpreterAbi: // Thread* is first argument (X0) in interpreter ABI. __ JumpTo(Arm64ManagedRegister::FromXRegister(X0), Offset(offset.Int32Value()), Arm64ManagedRegister::FromXRegister(IP1)); break; case kJniAbi: // Load via Thread* held in JNIEnv* in first argument (X0). __ LoadRawPtr(Arm64ManagedRegister::FromXRegister(IP1), Arm64ManagedRegister::FromXRegister(X0), Offset(JNIEnvExt::SelfOffset(8).Int32Value())); __ JumpTo(Arm64ManagedRegister::FromXRegister(IP1), Offset(offset.Int32Value()), Arm64ManagedRegister::FromXRegister(IP0)); break; case kQuickAbi: // X18 holds Thread*. __ JumpTo(Arm64ManagedRegister::FromXRegister(TR), Offset(offset.Int32Value()), Arm64ManagedRegister::FromXRegister(IP0)); break; } __ FinalizeCode(); size_t cs = __ CodeSize(); std::unique_ptr<std::vector<uint8_t>> entry_stub(new std::vector<uint8_t>(cs)); MemoryRegion code(entry_stub->data(), entry_stub->size()); __ FinalizeInstructions(code); return std::move(entry_stub); } } // namespace arm64 #endif // ART_ENABLE_CODEGEN_arm64 #ifdef ART_ENABLE_CODEGEN_mips namespace mips { static std::unique_ptr<const std::vector<uint8_t>> CreateTrampoline( ArenaAllocator* arena, EntryPointCallingConvention abi, ThreadOffset32 offset) { MipsAssembler assembler(arena); switch (abi) { case kInterpreterAbi: // Thread* is first argument (A0) in interpreter ABI. __ LoadFromOffset(kLoadWord, T9, A0, offset.Int32Value()); break; case kJniAbi: // Load via Thread* held in JNIEnv* in first argument (A0). __ LoadFromOffset(kLoadWord, T9, A0, JNIEnvExt::SelfOffset(4).Int32Value()); __ LoadFromOffset(kLoadWord, T9, T9, offset.Int32Value()); break; case kQuickAbi: // S1 holds Thread*. __ LoadFromOffset(kLoadWord, T9, S1, offset.Int32Value()); } __ Jr(T9); __ NopIfNoReordering(); __ Break(); __ FinalizeCode(); size_t cs = __ CodeSize(); std::unique_ptr<std::vector<uint8_t>> entry_stub(new std::vector<uint8_t>(cs)); MemoryRegion code(entry_stub->data(), entry_stub->size()); __ FinalizeInstructions(code); return std::move(entry_stub); } } // namespace mips #endif // ART_ENABLE_CODEGEN_mips #ifdef ART_ENABLE_CODEGEN_mips64 namespace mips64 { static std::unique_ptr<const std::vector<uint8_t>> CreateTrampoline( ArenaAllocator* arena, EntryPointCallingConvention abi, ThreadOffset64 offset) { Mips64Assembler assembler(arena); switch (abi) { case kInterpreterAbi: // Thread* is first argument (A0) in interpreter ABI. __ LoadFromOffset(kLoadDoubleword, T9, A0, offset.Int32Value()); break; case kJniAbi: // Load via Thread* held in JNIEnv* in first argument (A0). __ LoadFromOffset(kLoadDoubleword, T9, A0, JNIEnvExt::SelfOffset(8).Int32Value()); __ LoadFromOffset(kLoadDoubleword, T9, T9, offset.Int32Value()); break; case kQuickAbi: // Fall-through. __ LoadFromOffset(kLoadDoubleword, T9, S1, offset.Int32Value()); } __ Jr(T9); __ Nop(); __ Break(); __ FinalizeCode(); size_t cs = __ CodeSize(); std::unique_ptr<std::vector<uint8_t>> entry_stub(new std::vector<uint8_t>(cs)); MemoryRegion code(entry_stub->data(), entry_stub->size()); __ FinalizeInstructions(code); return std::move(entry_stub); } } // namespace mips64 #endif // ART_ENABLE_CODEGEN_mips #ifdef ART_ENABLE_CODEGEN_x86 namespace x86 { static std::unique_ptr<const std::vector<uint8_t>> CreateTrampoline(ArenaAllocator* arena, ThreadOffset32 offset) { X86Assembler assembler(arena); // All x86 trampolines call via the Thread* held in fs. __ fs()->jmp(Address::Absolute(offset)); __ int3(); __ FinalizeCode(); size_t cs = __ CodeSize(); std::unique_ptr<std::vector<uint8_t>> entry_stub(new std::vector<uint8_t>(cs)); MemoryRegion code(entry_stub->data(), entry_stub->size()); __ FinalizeInstructions(code); return std::move(entry_stub); } } // namespace x86 #endif // ART_ENABLE_CODEGEN_x86 #ifdef ART_ENABLE_CODEGEN_x86_64 namespace x86_64 { static std::unique_ptr<const std::vector<uint8_t>> CreateTrampoline(ArenaAllocator* arena, ThreadOffset64 offset) { x86_64::X86_64Assembler assembler(arena); // All x86 trampolines call via the Thread* held in gs. __ gs()->jmp(x86_64::Address::Absolute(offset, true)); __ int3(); __ FinalizeCode(); size_t cs = __ CodeSize(); std::unique_ptr<std::vector<uint8_t>> entry_stub(new std::vector<uint8_t>(cs)); MemoryRegion code(entry_stub->data(), entry_stub->size()); __ FinalizeInstructions(code); return std::move(entry_stub); } } // namespace x86_64 #endif // ART_ENABLE_CODEGEN_x86_64 std::unique_ptr<const std::vector<uint8_t>> CreateTrampoline64(InstructionSet isa, EntryPointCallingConvention abi, ThreadOffset64 offset) { ArenaPool pool; ArenaAllocator arena(&pool); switch (isa) { #ifdef ART_ENABLE_CODEGEN_arm64 case kArm64: return arm64::CreateTrampoline(&arena, abi, offset); #endif #ifdef ART_ENABLE_CODEGEN_mips64 case kMips64: return mips64::CreateTrampoline(&arena, abi, offset); #endif #ifdef ART_ENABLE_CODEGEN_x86_64 case kX86_64: return x86_64::CreateTrampoline(&arena, offset); #endif default: UNUSED(abi); UNUSED(offset); LOG(FATAL) << "Unexpected InstructionSet: " << isa; UNREACHABLE(); } } std::unique_ptr<const std::vector<uint8_t>> CreateTrampoline32(InstructionSet isa, EntryPointCallingConvention abi, ThreadOffset32 offset) { ArenaPool pool; ArenaAllocator arena(&pool); switch (isa) { #ifdef ART_ENABLE_CODEGEN_arm case kArm: case kThumb2: return arm::CreateTrampoline(&arena, abi, offset); #endif #ifdef ART_ENABLE_CODEGEN_mips case kMips: return mips::CreateTrampoline(&arena, abi, offset); #endif #ifdef ART_ENABLE_CODEGEN_x86 case kX86: UNUSED(abi); return x86::CreateTrampoline(&arena, offset); #endif default: LOG(FATAL) << "Unexpected InstructionSet: " << isa; UNREACHABLE(); } } } // namespace art