| // Copyright (c) 2012, the Dart project authors. Please see the AUTHORS file |
| // for details. All rights reserved. Use of this source code is governed by a |
| // BSD-style license that can be found in the LICENSE file. |
| |
| #include "vm/globals.h" // Needed here to get TARGET_ARCH_X64. |
| #if defined(TARGET_ARCH_X64) |
| |
| #include "vm/assembler.h" |
| #include "vm/code_patcher.h" |
| #include "vm/cpu.h" |
| #include "vm/dart_entry.h" |
| #include "vm/flow_graph_compiler.h" |
| #include "vm/instructions.h" |
| #include "vm/object.h" |
| #include "vm/raw_object.h" |
| |
| namespace dart { |
| |
| |
| static bool MatchesPattern(uword addr, int16_t* pattern, intptr_t size) { |
| uint8_t* bytes = reinterpret_cast<uint8_t*>(addr); |
| for (intptr_t i = 0; i < size; i++) { |
| int16_t val = pattern[i]; |
| if ((val >= 0) && (val != bytes[i])) { |
| return false; |
| } |
| } |
| return true; |
| } |
| |
| |
| intptr_t IndexFromPPLoad(uword start) { |
| int32_t offset = *reinterpret_cast<int32_t*>(start); |
| return ObjectPool::IndexFromOffset(offset); |
| } |
| |
| |
| class UnoptimizedCall : public ValueObject { |
| public: |
| UnoptimizedCall(uword return_address, const Code& code) |
| : object_pool_(ObjectPool::Handle(code.GetObjectPool())), |
| start_(return_address - kCallPatternSize) { |
| ASSERT((kCallPatternSize - 7) == Assembler::kCallExternalLabelSize); |
| ASSERT(IsValid()); |
| } |
| |
| static const int kCallPatternSize = 22; |
| |
| bool IsValid() const { |
| static int16_t pattern[kCallPatternSize] = { |
| 0x49, 0x8b, 0x9f, -1, -1, -1, -1, // movq RBX, [PP + offs] |
| 0x4d, 0x8b, 0xa7, -1, -1, -1, -1, // movq CR, [PP + offs] |
| 0x4d, 0x8b, 0x5c, 0x24, 0x07, // movq TMP, [CR + entry_point_offs] |
| 0x41, 0xff, 0xd3 // callq TMP |
| }; |
| return MatchesPattern(start_, pattern, kCallPatternSize); |
| } |
| |
| intptr_t argument_index() const { |
| return IndexFromPPLoad(start_ + 3); |
| } |
| |
| RawObject* ic_data() const { |
| return object_pool_.ObjectAt(argument_index()); |
| } |
| |
| RawCode* target() const { |
| intptr_t index = IndexFromPPLoad(start_ + 10); |
| Code& code = Code::Handle(); |
| code ^= object_pool_.ObjectAt(index); |
| return code.raw(); |
| } |
| |
| void set_target(const Code& target) const { |
| intptr_t index = IndexFromPPLoad(start_ + 10); |
| object_pool_.SetObjectAt(index, target); |
| // No need to flush the instruction cache, since the code is not modified. |
| } |
| |
| protected: |
| const ObjectPool& object_pool_; |
| |
| private: |
| uword start_; |
| DISALLOW_IMPLICIT_CONSTRUCTORS(UnoptimizedCall); |
| }; |
| |
| |
| class NativeCall : public UnoptimizedCall { |
| public: |
| NativeCall(uword return_address, const Code& code) |
| : UnoptimizedCall(return_address, code) { |
| } |
| |
| NativeFunction native_function() const { |
| return reinterpret_cast<NativeFunction>( |
| object_pool_.RawValueAt(argument_index())); |
| } |
| |
| void set_native_function(NativeFunction func) const { |
| object_pool_.SetRawValueAt(argument_index(), |
| reinterpret_cast<uword>(func)); |
| } |
| |
| private: |
| DISALLOW_IMPLICIT_CONSTRUCTORS(NativeCall); |
| }; |
| |
| |
| class InstanceCall : public UnoptimizedCall { |
| public: |
| InstanceCall(uword return_address, const Code& code) |
| : UnoptimizedCall(return_address, code) { |
| #if defined(DEBUG) |
| ICData& test_ic_data = ICData::Handle(); |
| test_ic_data ^= ic_data(); |
| ASSERT(test_ic_data.NumArgsTested() > 0); |
| #endif // DEBUG |
| } |
| |
| private: |
| DISALLOW_IMPLICIT_CONSTRUCTORS(InstanceCall); |
| }; |
| |
| |
| class UnoptimizedStaticCall : public UnoptimizedCall { |
| public: |
| UnoptimizedStaticCall(uword return_address, const Code& code) |
| : UnoptimizedCall(return_address, code) { |
| #if defined(DEBUG) |
| ICData& test_ic_data = ICData::Handle(); |
| test_ic_data ^= ic_data(); |
| ASSERT(test_ic_data.NumArgsTested() >= 0); |
| #endif // DEBUG |
| } |
| |
| private: |
| DISALLOW_IMPLICIT_CONSTRUCTORS(UnoptimizedStaticCall); |
| }; |
| |
| |
| // The expected pattern of a call where the target is loaded from |
| // the object pool. |
| class PoolPointerCall : public ValueObject { |
| public: |
| explicit PoolPointerCall(uword return_address, const Code& code) |
| : start_(return_address - kCallPatternSize), |
| object_pool_(ObjectPool::Handle(code.GetObjectPool())) { |
| ASSERT(IsValid()); |
| } |
| |
| static const int kCallPatternSize = 15; |
| |
| bool IsValid() const { |
| static int16_t pattern[kCallPatternSize] = { |
| 0x4d, 0x8b, 0xa7, -1, -1, -1, -1, // movq CR, [PP + offs] |
| 0x4d, 0x8b, 0x5c, 0x24, 0x07, // movq TMP, [CR + entry_point_off] |
| 0x41, 0xff, 0xd3 // callq TMP |
| }; |
| return MatchesPattern(start_, pattern, kCallPatternSize); |
| } |
| |
| intptr_t pp_index() const { |
| return IndexFromPPLoad(start_ + 3); |
| } |
| |
| RawCode* Target() const { |
| Code& code = Code::Handle(); |
| code ^= object_pool_.ObjectAt(pp_index()); |
| return code.raw(); |
| } |
| |
| void SetTarget(const Code& target) const { |
| object_pool_.SetObjectAt(pp_index(), target); |
| // No need to flush the instruction cache, since the code is not modified. |
| } |
| |
| protected: |
| uword start_; |
| const ObjectPool& object_pool_; |
| |
| private: |
| DISALLOW_IMPLICIT_CONSTRUCTORS(PoolPointerCall); |
| }; |
| |
| |
| RawCode* CodePatcher::GetStaticCallTargetAt(uword return_address, |
| const Code& code) { |
| ASSERT(code.ContainsInstructionAt(return_address)); |
| PoolPointerCall call(return_address, code); |
| return call.Target(); |
| } |
| |
| |
| void CodePatcher::PatchStaticCallAt(uword return_address, |
| const Code& code, |
| const Code& new_target) { |
| PatchPoolPointerCallAt(return_address, code, new_target); |
| } |
| |
| |
| void CodePatcher::PatchPoolPointerCallAt(uword return_address, |
| const Code& code, |
| const Code& new_target) { |
| ASSERT(code.ContainsInstructionAt(return_address)); |
| PoolPointerCall call(return_address, code); |
| call.SetTarget(new_target); |
| } |
| |
| |
| RawCode* CodePatcher::GetInstanceCallAt(uword return_address, |
| const Code& code, |
| ICData* ic_data) { |
| ASSERT(code.ContainsInstructionAt(return_address)); |
| InstanceCall call(return_address, code); |
| if (ic_data != NULL) { |
| *ic_data ^= call.ic_data(); |
| } |
| return call.target(); |
| } |
| |
| |
| intptr_t CodePatcher::InstanceCallSizeInBytes() { |
| return InstanceCall::kCallPatternSize; |
| } |
| |
| |
| void CodePatcher::InsertDeoptimizationCallAt(uword start, uword target) { |
| // The inserted call should not overlap the lazy deopt jump code. |
| ASSERT(start + ShortCallPattern::pattern_length_in_bytes() <= target); |
| *reinterpret_cast<uint8_t*>(start) = 0xE8; |
| ShortCallPattern call(start); |
| call.SetTargetAddress(target); |
| CPU::FlushICache(start, ShortCallPattern::pattern_length_in_bytes()); |
| } |
| |
| |
| RawFunction* CodePatcher::GetUnoptimizedStaticCallAt( |
| uword return_address, const Code& code, ICData* ic_data_result) { |
| ASSERT(code.ContainsInstructionAt(return_address)); |
| UnoptimizedStaticCall static_call(return_address, code); |
| ICData& ic_data = ICData::Handle(); |
| ic_data ^= static_call.ic_data(); |
| if (ic_data_result != NULL) { |
| *ic_data_result = ic_data.raw(); |
| } |
| return ic_data.GetTargetAt(0); |
| } |
| |
| |
| void CodePatcher::PatchNativeCallAt(uword return_address, |
| const Code& code, |
| NativeFunction target, |
| const Code& trampoline) { |
| ASSERT(code.ContainsInstructionAt(return_address)); |
| NativeCall call(return_address, code); |
| call.set_target(trampoline); |
| call.set_native_function(target); |
| } |
| |
| |
| RawCode* CodePatcher::GetNativeCallAt(uword return_address, |
| const Code& code, |
| NativeFunction* target) { |
| ASSERT(code.ContainsInstructionAt(return_address)); |
| NativeCall call(return_address, code); |
| *target = call.native_function(); |
| return call.target(); |
| } |
| |
| } // namespace dart |
| |
| #endif // defined TARGET_ARCH_X64 |