/* * Copyright (C) 2013 The Android Open Source Project * * Licensed under the Apache License, Version 2.0 (the "License"); * you may not use this file except in compliance with the License. * You may obtain a copy of the License at * * http://www.apache.org/licenses/LICENSE-2.0 * * Unless required by applicable law or agreed to in writing, software * distributed under the License is distributed on an "AS IS" BASIS, * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. * See the License for the specific language governing permissions and * limitations under the License. */ #include "trampoline_compiler.h" #include "jni_env_ext.h" #include "utils/arm/assembler_arm.h" #include "utils/arm64/assembler_arm64.h" #include "utils/mips/assembler_mips.h" #include "utils/mips64/assembler_mips64.h" #include "utils/x86/assembler_x86.h" #include "utils/x86_64/assembler_x86_64.h" #define __ assembler-> namespace art { namespace arm { static const std::vector<uint8_t>* CreateTrampoline(EntryPointCallingConvention abi, ThreadOffset<4> offset) { std::unique_ptr<ArmAssembler> assembler(static_cast<ArmAssembler*>(Assembler::Create(kThumb2))); switch (abi) { case kInterpreterAbi: // Thread* is first argument (R0) in interpreter ABI. __ LoadFromOffset(kLoadWord, PC, R0, offset.Int32Value()); break; case kJniAbi: // Load via Thread* held in JNIEnv* in first argument (R0). __ LoadFromOffset(kLoadWord, IP, R0, JNIEnvExt::SelfOffset().Int32Value()); __ LoadFromOffset(kLoadWord, PC, IP, offset.Int32Value()); break; case kQuickAbi: // R9 holds Thread*. __ LoadFromOffset(kLoadWord, PC, R9, offset.Int32Value()); } __ bkpt(0); size_t cs = assembler->CodeSize(); std::unique_ptr<std::vector<uint8_t>> entry_stub(new std::vector<uint8_t>(cs)); MemoryRegion code(&(*entry_stub)[0], entry_stub->size()); assembler->FinalizeInstructions(code); return entry_stub.release(); } } // namespace arm namespace arm64 { static const std::vector<uint8_t>* CreateTrampoline(EntryPointCallingConvention abi, ThreadOffset<8> offset) { std::unique_ptr<Arm64Assembler> assembler(static_cast<Arm64Assembler*>(Assembler::Create(kArm64))); switch (abi) { case kInterpreterAbi: // Thread* is first argument (X0) in interpreter ABI. __ JumpTo(Arm64ManagedRegister::FromXRegister(X0), Offset(offset.Int32Value()), Arm64ManagedRegister::FromXRegister(IP1)); break; case kJniAbi: // Load via Thread* held in JNIEnv* in first argument (X0). __ LoadRawPtr(Arm64ManagedRegister::FromXRegister(IP1), Arm64ManagedRegister::FromXRegister(X0), Offset(JNIEnvExt::SelfOffset().Int32Value())); __ JumpTo(Arm64ManagedRegister::FromXRegister(IP1), Offset(offset.Int32Value()), Arm64ManagedRegister::FromXRegister(IP0)); break; case kQuickAbi: // X18 holds Thread*. __ JumpTo(Arm64ManagedRegister::FromXRegister(TR), Offset(offset.Int32Value()), Arm64ManagedRegister::FromXRegister(IP0)); break; } assembler->EmitSlowPaths(); size_t cs = assembler->CodeSize(); std::unique_ptr<std::vector<uint8_t>> entry_stub(new std::vector<uint8_t>(cs)); MemoryRegion code(&(*entry_stub)[0], entry_stub->size()); assembler->FinalizeInstructions(code); return entry_stub.release(); } } // namespace arm64 namespace mips { static const std::vector<uint8_t>* CreateTrampoline(EntryPointCallingConvention abi, ThreadOffset<4> offset) { std::unique_ptr<MipsAssembler> assembler(static_cast<MipsAssembler*>(Assembler::Create(kMips))); switch (abi) { case kInterpreterAbi: // Thread* is first argument (A0) in interpreter ABI. __ LoadFromOffset(kLoadWord, T9, A0, offset.Int32Value()); break; case kJniAbi: // Load via Thread* held in JNIEnv* in first argument (A0). __ LoadFromOffset(kLoadWord, T9, A0, JNIEnvExt::SelfOffset().Int32Value()); __ LoadFromOffset(kLoadWord, T9, T9, offset.Int32Value()); break; case kQuickAbi: // S1 holds Thread*. __ LoadFromOffset(kLoadWord, T9, S1, offset.Int32Value()); } __ Jr(T9); __ Nop(); __ Break(); size_t cs = assembler->CodeSize(); std::unique_ptr<std::vector<uint8_t>> entry_stub(new std::vector<uint8_t>(cs)); MemoryRegion code(&(*entry_stub)[0], entry_stub->size()); assembler->FinalizeInstructions(code); return entry_stub.release(); } } // namespace mips namespace mips64 { static const std::vector<uint8_t>* CreateTrampoline(EntryPointCallingConvention abi, ThreadOffset<8> offset) { std::unique_ptr<Mips64Assembler> assembler(static_cast<Mips64Assembler*>(Assembler::Create(kMips64))); switch (abi) { case kInterpreterAbi: // Thread* is first argument (A0) in interpreter ABI. __ LoadFromOffset(kLoadDoubleword, T9, A0, offset.Int32Value()); break; case kJniAbi: // Load via Thread* held in JNIEnv* in first argument (A0). __ LoadFromOffset(kLoadDoubleword, T9, A0, JNIEnvExt::SelfOffset().Int32Value()); __ LoadFromOffset(kLoadDoubleword, T9, T9, offset.Int32Value()); break; case kQuickAbi: // Fall-through. __ LoadFromOffset(kLoadDoubleword, T9, S1, offset.Int32Value()); } __ Jr(T9); __ Nop(); __ Break(); size_t cs = assembler->CodeSize(); std::unique_ptr<std::vector<uint8_t>> entry_stub(new std::vector<uint8_t>(cs)); MemoryRegion code(&(*entry_stub)[0], entry_stub->size()); assembler->FinalizeInstructions(code); return entry_stub.release(); } } // namespace mips64 namespace x86 { static const std::vector<uint8_t>* CreateTrampoline(ThreadOffset<4> offset) { std::unique_ptr<X86Assembler> assembler(static_cast<X86Assembler*>(Assembler::Create(kX86))); // All x86 trampolines call via the Thread* held in fs. __ fs()->jmp(Address::Absolute(offset)); __ int3(); size_t cs = assembler->CodeSize(); std::unique_ptr<std::vector<uint8_t>> entry_stub(new std::vector<uint8_t>(cs)); MemoryRegion code(&(*entry_stub)[0], entry_stub->size()); assembler->FinalizeInstructions(code); return entry_stub.release(); } } // namespace x86 namespace x86_64 { static const std::vector<uint8_t>* CreateTrampoline(ThreadOffset<8> offset) { std::unique_ptr<x86_64::X86_64Assembler> assembler(static_cast<x86_64::X86_64Assembler*>(Assembler::Create(kX86_64))); // All x86 trampolines call via the Thread* held in gs. __ gs()->jmp(x86_64::Address::Absolute(offset, true)); __ int3(); size_t cs = assembler->CodeSize(); std::unique_ptr<std::vector<uint8_t>> entry_stub(new std::vector<uint8_t>(cs)); MemoryRegion code(&(*entry_stub)[0], entry_stub->size()); assembler->FinalizeInstructions(code); return entry_stub.release(); } } // namespace x86_64 const std::vector<uint8_t>* CreateTrampoline64(InstructionSet isa, EntryPointCallingConvention abi, ThreadOffset<8> offset) { switch (isa) { case kArm64: return arm64::CreateTrampoline(abi, offset); case kMips64: return mips64::CreateTrampoline(abi, offset); case kX86_64: return x86_64::CreateTrampoline(offset); default: LOG(FATAL) << "Unexpected InstructionSet: " << isa; UNREACHABLE(); } } const std::vector<uint8_t>* CreateTrampoline32(InstructionSet isa, EntryPointCallingConvention abi, ThreadOffset<4> offset) { switch (isa) { case kArm: case kThumb2: return arm::CreateTrampoline(abi, offset); case kMips: return mips::CreateTrampoline(abi, offset); case kX86: return x86::CreateTrampoline(offset); default: LOG(FATAL) << "Unexpected InstructionSet: " << isa; UNREACHABLE(); } } } // namespace art