| // Copyright (c) 2012, the Dart project authors. Please see the AUTHORS file |
| // for details. All rights reserved. Use of this source code is governed by a |
| // BSD-style license that can be found in the LICENSE file. |
| |
| #include "vm/debugger.h" |
| |
| #include "include/dart_api.h" |
| |
| #include "platform/address_sanitizer.h" |
| |
| #include "vm/code_patcher.h" |
| #include "vm/compiler/assembler/disassembler.h" |
| #include "vm/compiler/jit/compiler.h" |
| #include "vm/dart_entry.h" |
| #include "vm/deopt_instructions.h" |
| #include "vm/flags.h" |
| #include "vm/globals.h" |
| #include "vm/json_stream.h" |
| #include "vm/kernel.h" |
| #include "vm/longjump.h" |
| #include "vm/message_handler.h" |
| #include "vm/object.h" |
| #include "vm/object_store.h" |
| #include "vm/os.h" |
| #include "vm/parser.h" |
| #include "vm/port.h" |
| #include "vm/runtime_entry.h" |
| #include "vm/service.h" |
| #include "vm/service_event.h" |
| #include "vm/service_isolate.h" |
| #include "vm/stack_frame.h" |
| #include "vm/stack_trace.h" |
| #include "vm/stub_code.h" |
| #include "vm/symbols.h" |
| #include "vm/thread_interrupter.h" |
| #include "vm/timeline.h" |
| #include "vm/token_position.h" |
| #include "vm/visitor.h" |
| |
| namespace dart { |
| |
| DEFINE_FLAG(bool, |
| show_invisible_frames, |
| false, |
| "Show invisible frames in debugger stack traces"); |
| DEFINE_FLAG(bool, |
| trace_debugger_stacktrace, |
| false, |
| "Trace debugger stacktrace collection"); |
| DEFINE_FLAG(bool, trace_rewind, false, "Trace frame rewind"); |
| DEFINE_FLAG(bool, verbose_debug, false, "Verbose debugger messages"); |
| |
| DECLARE_FLAG(bool, enable_interpreter); |
| DECLARE_FLAG(bool, trace_deoptimization); |
| DECLARE_FLAG(bool, warn_on_pause_with_no_debugger); |
| |
| #ifndef PRODUCT |
| |
| // Create an unresolved breakpoint in given token range and script. |
| BreakpointLocation::BreakpointLocation(const Script& script, |
| TokenPosition token_pos, |
| TokenPosition end_token_pos, |
| intptr_t requested_line_number, |
| intptr_t requested_column_number) |
| : script_(script.raw()), |
| url_(script.url()), |
| token_pos_(token_pos), |
| end_token_pos_(end_token_pos), |
| is_resolved_(false), |
| next_(NULL), |
| conditions_(NULL), |
| requested_line_number_(requested_line_number), |
| requested_column_number_(requested_column_number), |
| function_(Function::null()), |
| line_number_(-1), |
| column_number_(-1) { |
| ASSERT(!script.IsNull()); |
| ASSERT(token_pos_.IsReal()); |
| } |
| |
| // Create a latent breakpoint at given url and line number. |
| BreakpointLocation::BreakpointLocation(const String& url, |
| intptr_t requested_line_number, |
| intptr_t requested_column_number) |
| : script_(Script::null()), |
| url_(url.raw()), |
| token_pos_(TokenPosition::kNoSource), |
| end_token_pos_(TokenPosition::kNoSource), |
| is_resolved_(false), |
| next_(NULL), |
| conditions_(NULL), |
| requested_line_number_(requested_line_number), |
| requested_column_number_(requested_column_number), |
| function_(Function::null()), |
| line_number_(-1), |
| column_number_(-1) { |
| ASSERT(requested_line_number_ >= 0); |
| } |
| |
| BreakpointLocation::~BreakpointLocation() { |
| Breakpoint* bpt = breakpoints(); |
| while (bpt != NULL) { |
| Breakpoint* temp = bpt; |
| bpt = bpt->next(); |
| delete temp; |
| } |
| } |
| |
| bool BreakpointLocation::AnyEnabled() const { |
| return breakpoints() != NULL; |
| } |
| |
| void BreakpointLocation::SetResolved(const Function& func, |
| TokenPosition token_pos) { |
| ASSERT(!IsLatent()); |
| ASSERT(func.script() == script_); |
| ASSERT((func.token_pos() <= token_pos) && |
| (token_pos <= func.end_token_pos())); |
| ASSERT(func.is_debuggable()); |
| function_ = func.raw(); |
| token_pos_ = token_pos; |
| end_token_pos_ = token_pos; |
| is_resolved_ = true; |
| } |
| |
| // TODO(hausner): Get rid of library parameter. A source breakpoint location |
| // does not imply a library, since the same source code can be included |
| // in more than one library, e.g. the text location of mixin functions. |
| void BreakpointLocation::GetCodeLocation(Library* lib, |
| Script* script, |
| TokenPosition* pos) const { |
| if (IsLatent()) { |
| *lib = Library::null(); |
| *script = Script::null(); |
| *pos = TokenPosition::kNoSource; |
| } else { |
| *script = this->script(); |
| *pos = token_pos_; |
| if (IsResolved()) { |
| const Function& func = Function::Handle(function_); |
| ASSERT(!func.IsNull()); |
| const Class& cls = Class::Handle(func.origin()); |
| *lib = cls.library(); |
| } else { |
| *lib = Library::null(); |
| } |
| } |
| } |
| |
| intptr_t BreakpointLocation::LineNumber() { |
| ASSERT(IsResolved()); |
| // Compute line number lazily since it causes scanning of the script. |
| if (line_number_ < 0) { |
| const Script& script = Script::Handle(this->script()); |
| script.GetTokenLocation(token_pos_, &line_number_, NULL); |
| } |
| return line_number_; |
| } |
| |
| intptr_t BreakpointLocation::ColumnNumber() { |
| ASSERT(IsResolved()); |
| // Compute column number lazily since it causes scanning of the script. |
| if (column_number_ < 0) { |
| const Script& script = Script::Handle(this->script()); |
| script.GetTokenLocation(token_pos_, &line_number_, &column_number_); |
| } |
| return column_number_; |
| } |
| |
| void Breakpoint::set_bpt_location(BreakpointLocation* new_bpt_location) { |
| // Only latent breakpoints can be moved. |
| ASSERT((new_bpt_location == NULL) || bpt_location_->IsLatent()); |
| bpt_location_ = new_bpt_location; |
| } |
| |
| void Breakpoint::VisitObjectPointers(ObjectPointerVisitor* visitor) { |
| visitor->VisitPointer(reinterpret_cast<RawObject**>(&closure_)); |
| } |
| |
| void BreakpointLocation::VisitObjectPointers(ObjectPointerVisitor* visitor) { |
| visitor->VisitPointer(reinterpret_cast<RawObject**>(&script_)); |
| visitor->VisitPointer(reinterpret_cast<RawObject**>(&url_)); |
| visitor->VisitPointer(reinterpret_cast<RawObject**>(&function_)); |
| |
| Breakpoint* bpt = conditions_; |
| while (bpt != NULL) { |
| bpt->VisitObjectPointers(visitor); |
| bpt = bpt->next(); |
| } |
| } |
| |
| void Breakpoint::PrintJSON(JSONStream* stream) { |
| JSONObject jsobj(stream); |
| jsobj.AddProperty("type", "Breakpoint"); |
| |
| jsobj.AddFixedServiceId("breakpoints/%" Pd "", id()); |
| jsobj.AddProperty("breakpointNumber", id()); |
| if (is_synthetic_async()) { |
| jsobj.AddProperty("isSyntheticAsyncContinuation", is_synthetic_async()); |
| } |
| jsobj.AddProperty("resolved", bpt_location_->IsResolved()); |
| if (bpt_location_->IsResolved()) { |
| jsobj.AddLocation(bpt_location_); |
| } else { |
| jsobj.AddUnresolvedLocation(bpt_location_); |
| } |
| } |
| |
| void CodeBreakpoint::VisitObjectPointers(ObjectPointerVisitor* visitor) { |
| visitor->VisitPointer(reinterpret_cast<RawObject**>(&code_)); |
| #if !defined(TARGET_ARCH_DBC) |
| visitor->VisitPointer(reinterpret_cast<RawObject**>(&saved_value_)); |
| #endif |
| } |
| |
| ActivationFrame::ActivationFrame(uword pc, |
| uword fp, |
| uword sp, |
| const Code& code, |
| const Array& deopt_frame, |
| intptr_t deopt_frame_offset, |
| ActivationFrame::Kind kind) |
| : pc_(pc), |
| fp_(fp), |
| sp_(sp), |
| ctx_(Context::ZoneHandle()), |
| code_(Code::ZoneHandle(code.raw())), |
| function_(Function::ZoneHandle(code.function())), |
| live_frame_((kind == kRegular) || (kind == kAsyncActivation)), |
| token_pos_initialized_(false), |
| token_pos_(TokenPosition::kNoSource), |
| try_index_(-1), |
| deopt_id_(DeoptId::kNone), |
| line_number_(-1), |
| column_number_(-1), |
| context_level_(-1), |
| deopt_frame_(Array::ZoneHandle(deopt_frame.raw())), |
| deopt_frame_offset_(deopt_frame_offset), |
| kind_(kind), |
| is_interpreted_(false), // TODO(regis): No bytecode debugging support. |
| vars_initialized_(false), |
| var_descriptors_(LocalVarDescriptors::ZoneHandle()), |
| desc_indices_(8), |
| pc_desc_(PcDescriptors::ZoneHandle()) {} |
| |
| ActivationFrame::ActivationFrame(Kind kind) |
| : pc_(0), |
| fp_(0), |
| sp_(0), |
| ctx_(Context::ZoneHandle()), |
| code_(Code::ZoneHandle()), |
| function_(Function::ZoneHandle()), |
| live_frame_(kind == kRegular), |
| token_pos_initialized_(false), |
| token_pos_(TokenPosition::kNoSource), |
| try_index_(-1), |
| line_number_(-1), |
| column_number_(-1), |
| context_level_(-1), |
| deopt_frame_(Array::ZoneHandle()), |
| deopt_frame_offset_(0), |
| kind_(kind), |
| is_interpreted_(false), |
| vars_initialized_(false), |
| var_descriptors_(LocalVarDescriptors::ZoneHandle()), |
| desc_indices_(8), |
| pc_desc_(PcDescriptors::ZoneHandle()) {} |
| |
| ActivationFrame::ActivationFrame(const Closure& async_activation) |
| : pc_(0), |
| fp_(0), |
| sp_(0), |
| ctx_(Context::ZoneHandle()), |
| code_(Code::ZoneHandle()), |
| function_(Function::ZoneHandle()), |
| live_frame_(false), |
| token_pos_initialized_(false), |
| token_pos_(TokenPosition::kNoSource), |
| try_index_(-1), |
| line_number_(-1), |
| column_number_(-1), |
| context_level_(-1), |
| deopt_frame_(Array::ZoneHandle()), |
| deopt_frame_offset_(0), |
| kind_(kAsyncActivation), |
| is_interpreted_(false), |
| vars_initialized_(false), |
| var_descriptors_(LocalVarDescriptors::ZoneHandle()), |
| desc_indices_(8), |
| pc_desc_(PcDescriptors::ZoneHandle()) { |
| // Extract the function and the code from the asynchronous activation. |
| function_ = async_activation.function(); |
| function_.EnsureHasCompiledUnoptimizedCode(); |
| code_ = function_.unoptimized_code(); |
| ctx_ = async_activation.context(); |
| ASSERT(fp_ == 0); |
| ASSERT(!ctx_.IsNull()); |
| } |
| |
| bool Debugger::NeedsIsolateEvents() { |
| return !Isolate::IsVMInternalIsolate(isolate_) && |
| Service::isolate_stream.enabled(); |
| } |
| |
| bool Debugger::NeedsDebugEvents() { |
| ASSERT(!Isolate::IsVMInternalIsolate(isolate_)); |
| return FLAG_warn_on_pause_with_no_debugger || Service::debug_stream.enabled(); |
| } |
| |
| void Debugger::InvokeEventHandler(ServiceEvent* event) { |
| ASSERT(!event->IsPause()); // For pause events, call Pause instead. |
| Service::HandleEvent(event); |
| } |
| |
| RawError* Debugger::PauseInterrupted() { |
| return PauseRequest(ServiceEvent::kPauseInterrupted); |
| } |
| |
| RawError* Debugger::PausePostRequest() { |
| return PauseRequest(ServiceEvent::kPausePostRequest); |
| } |
| |
| RawError* Debugger::PauseRequest(ServiceEvent::EventKind kind) { |
| if (ignore_breakpoints_ || IsPaused()) { |
| // We don't let the isolate get interrupted if we are already |
| // paused or ignoring breakpoints. |
| return Error::null(); |
| } |
| ServiceEvent event(isolate_, kind); |
| DebuggerStackTrace* trace = CollectStackTrace(); |
| if (trace->Length() > 0) { |
| event.set_top_frame(trace->FrameAt(0)); |
| } |
| CacheStackTraces(trace, CollectAsyncCausalStackTrace(), |
| CollectAwaiterReturnStackTrace()); |
| resume_action_ = kContinue; |
| Pause(&event); |
| HandleSteppingRequest(trace); |
| ClearCachedStackTraces(); |
| |
| // If any error occurred while in the debug message loop, return it here. |
| const Error& error = Error::Handle(Thread::Current()->sticky_error()); |
| ASSERT(error.IsNull() || error.IsUnwindError()); |
| Thread::Current()->clear_sticky_error(); |
| return error.raw(); |
| } |
| |
| void Debugger::SendBreakpointEvent(ServiceEvent::EventKind kind, |
| Breakpoint* bpt) { |
| if (NeedsDebugEvents()) { |
| // TODO(turnidge): Currently we send single-shot breakpoint events |
| // to the vm service. Do we want to change this? |
| ServiceEvent event(isolate_, kind); |
| event.set_breakpoint(bpt); |
| InvokeEventHandler(&event); |
| } |
| } |
| |
| void BreakpointLocation::AddBreakpoint(Breakpoint* bpt, Debugger* dbg) { |
| bpt->set_next(breakpoints()); |
| set_breakpoints(bpt); |
| |
| dbg->SyncBreakpointLocation(this); |
| dbg->SendBreakpointEvent(ServiceEvent::kBreakpointAdded, bpt); |
| } |
| |
| Breakpoint* BreakpointLocation::AddRepeated(Debugger* dbg) { |
| Breakpoint* bpt = breakpoints(); |
| while (bpt != NULL) { |
| if (bpt->IsRepeated()) break; |
| bpt = bpt->next(); |
| } |
| if (bpt == NULL) { |
| bpt = new Breakpoint(dbg->nextId(), this); |
| bpt->SetIsRepeated(); |
| AddBreakpoint(bpt, dbg); |
| } |
| return bpt; |
| } |
| |
| Breakpoint* BreakpointLocation::AddSingleShot(Debugger* dbg) { |
| Breakpoint* bpt = breakpoints(); |
| while (bpt != NULL) { |
| if (bpt->IsSingleShot()) break; |
| bpt = bpt->next(); |
| } |
| if (bpt == NULL) { |
| bpt = new Breakpoint(dbg->nextId(), this); |
| bpt->SetIsSingleShot(); |
| AddBreakpoint(bpt, dbg); |
| } |
| return bpt; |
| } |
| |
| Breakpoint* BreakpointLocation::AddPerClosure(Debugger* dbg, |
| const Instance& closure, |
| bool for_over_await) { |
| Breakpoint* bpt = NULL; |
| // Do not reuse existing breakpoints for stepping over await clauses. |
| // A second async step-over command will set a new breakpoint before |
| // the existing one gets deleted when first async step-over resumes. |
| if (!for_over_await) { |
| bpt = breakpoints(); |
| while (bpt != NULL) { |
| if (bpt->IsPerClosure() && (bpt->closure() == closure.raw())) break; |
| bpt = bpt->next(); |
| } |
| } |
| if (bpt == NULL) { |
| bpt = new Breakpoint(dbg->nextId(), this); |
| bpt->SetIsPerClosure(closure); |
| bpt->set_is_synthetic_async(for_over_await); |
| AddBreakpoint(bpt, dbg); |
| } |
| return bpt; |
| } |
| |
| const char* Debugger::QualifiedFunctionName(const Function& func) { |
| const String& func_name = String::Handle(func.name()); |
| Class& func_class = Class::Handle(func.Owner()); |
| String& class_name = String::Handle(func_class.Name()); |
| |
| return OS::SCreate(Thread::Current()->zone(), "%s%s%s", |
| func_class.IsTopLevel() ? "" : class_name.ToCString(), |
| func_class.IsTopLevel() ? "" : ".", func_name.ToCString()); |
| } |
| |
| // Returns true if the function |func| overlaps the token range |
| // [|token_pos|, |end_token_pos|] in |script|. |
| static bool FunctionOverlaps(const Function& func, |
| const Script& script, |
| TokenPosition token_pos, |
| TokenPosition end_token_pos) { |
| TokenPosition func_start = func.token_pos(); |
| if (((func_start <= token_pos) && (token_pos <= func.end_token_pos())) || |
| ((token_pos <= func_start) && (func_start <= end_token_pos))) { |
| // Check script equality second because it allocates |
| // handles as a side effect. |
| return func.script() == script.raw(); |
| } |
| return false; |
| } |
| |
| static bool IsImplicitFunction(const Function& func) { |
| switch (func.kind()) { |
| case RawFunction::kImplicitGetter: |
| case RawFunction::kImplicitSetter: |
| case RawFunction::kImplicitStaticFinalGetter: |
| case RawFunction::kMethodExtractor: |
| case RawFunction::kNoSuchMethodDispatcher: |
| case RawFunction::kInvokeFieldDispatcher: |
| case RawFunction::kIrregexpFunction: |
| return true; |
| default: |
| if (func.token_pos() == func.end_token_pos()) { |
| // |func| could be an implicit constructor for example. |
| return true; |
| } |
| } |
| return false; |
| } |
| |
| bool Debugger::HasBreakpoint(const Function& func, Zone* zone) { |
| if (!func.HasCode()) { |
| // If the function is not compiled yet, just check whether there |
| // is a user-defined breakpoint that falls into the token |
| // range of the function. This may be a false positive: the breakpoint |
| // might be inside a local closure. |
| Script& script = Script::Handle(zone); |
| BreakpointLocation* sbpt = breakpoint_locations_; |
| while (sbpt != NULL) { |
| script = sbpt->script(); |
| if (FunctionOverlaps(func, script, sbpt->token_pos(), |
| sbpt->end_token_pos())) { |
| return true; |
| } |
| sbpt = sbpt->next_; |
| } |
| return false; |
| } |
| CodeBreakpoint* cbpt = code_breakpoints_; |
| while (cbpt != NULL) { |
| if (func.raw() == cbpt->function()) { |
| return true; |
| } |
| cbpt = cbpt->next_; |
| } |
| return false; |
| } |
| |
| bool Debugger::HasBreakpoint(const Code& code) { |
| CodeBreakpoint* cbpt = code_breakpoints_; |
| while (cbpt != NULL) { |
| if (code.raw() == cbpt->code_) { |
| return true; |
| } |
| cbpt = cbpt->next_; |
| } |
| return false; |
| } |
| |
| void Debugger::PrintBreakpointsToJSONArray(JSONArray* jsarr) const { |
| PrintBreakpointsListToJSONArray(breakpoint_locations_, jsarr); |
| PrintBreakpointsListToJSONArray(latent_locations_, jsarr); |
| } |
| |
| void Debugger::PrintBreakpointsListToJSONArray(BreakpointLocation* sbpt, |
| JSONArray* jsarr) const { |
| while (sbpt != NULL) { |
| Breakpoint* bpt = sbpt->breakpoints(); |
| while (bpt != NULL) { |
| jsarr->AddValue(bpt); |
| bpt = bpt->next(); |
| } |
| sbpt = sbpt->next_; |
| } |
| } |
| |
| void Debugger::PrintSettingsToJSONObject(JSONObject* jsobj) const { |
| // This won't cut it when we support filtering by class, etc. |
| switch (GetExceptionPauseInfo()) { |
| case kNoPauseOnExceptions: |
| jsobj->AddProperty("_exceptions", "none"); |
| break; |
| case kPauseOnAllExceptions: |
| jsobj->AddProperty("_exceptions", "all"); |
| break; |
| case kPauseOnUnhandledExceptions: |
| jsobj->AddProperty("_exceptions", "unhandled"); |
| break; |
| default: |
| UNREACHABLE(); |
| } |
| } |
| |
| RawString* ActivationFrame::QualifiedFunctionName() { |
| return String::New(Debugger::QualifiedFunctionName(function())); |
| } |
| |
| RawString* ActivationFrame::SourceUrl() { |
| const Script& script = Script::Handle(SourceScript()); |
| return script.url(); |
| } |
| |
| RawScript* ActivationFrame::SourceScript() { |
| return function().script(); |
| } |
| |
| RawLibrary* ActivationFrame::Library() { |
| const Class& cls = Class::Handle(function().origin()); |
| return cls.library(); |
| } |
| |
| void ActivationFrame::GetPcDescriptors() { |
| if (pc_desc_.IsNull()) { |
| pc_desc_ = code().pc_descriptors(); |
| ASSERT(!pc_desc_.IsNull()); |
| } |
| } |
| |
| // Compute token_pos_ and try_index_ and token_pos_initialized_. |
| TokenPosition ActivationFrame::TokenPos() { |
| if (!token_pos_initialized_) { |
| token_pos_initialized_ = true; |
| token_pos_ = TokenPosition::kNoSource; |
| GetPcDescriptors(); |
| PcDescriptors::Iterator iter(pc_desc_, RawPcDescriptors::kAnyKind); |
| uword pc_offset = pc_ - code().PayloadStart(); |
| while (iter.MoveNext()) { |
| if (iter.PcOffset() == pc_offset) { |
| try_index_ = iter.TryIndex(); |
| token_pos_ = iter.TokenPos(); |
| deopt_id_ = iter.DeoptId(); |
| break; |
| } |
| } |
| } |
| return token_pos_; |
| } |
| |
| intptr_t ActivationFrame::TryIndex() { |
| if (!token_pos_initialized_) { |
| TokenPos(); // Side effect: computes token_pos_initialized_, try_index_. |
| } |
| return try_index_; |
| } |
| |
| intptr_t ActivationFrame::DeoptId() { |
| if (!token_pos_initialized_) { |
| TokenPos(); // Side effect: computes token_pos_initialized_, try_index_. |
| } |
| return deopt_id_; |
| } |
| |
| intptr_t ActivationFrame::LineNumber() { |
| // Compute line number lazily since it causes scanning of the script. |
| if ((line_number_ < 0) && TokenPos().IsSourcePosition()) { |
| const TokenPosition token_pos = TokenPos().SourcePosition(); |
| const Script& script = Script::Handle(SourceScript()); |
| script.GetTokenLocation(token_pos, &line_number_, NULL); |
| } |
| return line_number_; |
| } |
| |
| intptr_t ActivationFrame::ColumnNumber() { |
| // Compute column number lazily since it causes scanning of the script. |
| if ((column_number_ < 0) && TokenPos().IsSourcePosition()) { |
| const TokenPosition token_pos = TokenPos().SourcePosition(); |
| const Script& script = Script::Handle(SourceScript()); |
| if (script.HasSource()) { |
| script.GetTokenLocation(token_pos, &line_number_, &column_number_); |
| } else { |
| column_number_ = -1; |
| } |
| } |
| return column_number_; |
| } |
| |
| void ActivationFrame::GetVarDescriptors() { |
| if (var_descriptors_.IsNull()) { |
| if (is_interpreted()) { |
| // TODO(regis): Kernel bytecode does not yet provide var descriptors. |
| var_descriptors_ = Object::empty_var_descriptors().raw(); |
| return; |
| } |
| Code& unoptimized_code = Code::Handle(function().unoptimized_code()); |
| if (unoptimized_code.IsNull()) { |
| Thread* thread = Thread::Current(); |
| Zone* zone = thread->zone(); |
| const Error& error = Error::Handle( |
| zone, Compiler::EnsureUnoptimizedCode(thread, function())); |
| if (!error.IsNull()) { |
| Exceptions::PropagateError(error); |
| } |
| unoptimized_code ^= function().unoptimized_code(); |
| } |
| ASSERT(!unoptimized_code.IsNull()); |
| var_descriptors_ = unoptimized_code.GetLocalVarDescriptors(); |
| ASSERT(!var_descriptors_.IsNull()); |
| } |
| } |
| |
| bool ActivationFrame::IsDebuggable() const { |
| return Debugger::IsDebuggable(function()); |
| } |
| |
| void ActivationFrame::PrintDescriptorsError(const char* message) { |
| OS::PrintErr("Bad descriptors: %s\n", message); |
| OS::PrintErr("function %s\n", function().ToQualifiedCString()); |
| OS::PrintErr("pc_ %" Px "\n", pc_); |
| OS::PrintErr("deopt_id_ %" Px "\n", deopt_id_); |
| OS::PrintErr("context_level_ %" Px "\n", context_level_); |
| DisassembleToStdout formatter; |
| code().Disassemble(&formatter); |
| PcDescriptors::Handle(code().pc_descriptors()).Print(); |
| StackFrameIterator frames(ValidationPolicy::kDontValidateFrames, |
| Thread::Current(), |
| StackFrameIterator::kNoCrossThreadIteration); |
| StackFrame* frame = frames.NextFrame(); |
| while (frame != NULL) { |
| OS::PrintErr("%s\n", frame->ToCString()); |
| frame = frames.NextFrame(); |
| } |
| OS::Abort(); |
| } |
| |
| // Calculate the context level at the current token index of the frame. |
| intptr_t ActivationFrame::ContextLevel() { |
| const Context& ctx = GetSavedCurrentContext(); |
| if (context_level_ < 0 && !ctx.IsNull()) { |
| ASSERT(!code_.is_optimized()); |
| |
| GetVarDescriptors(); |
| intptr_t deopt_id = DeoptId(); |
| if (deopt_id == DeoptId::kNone) { |
| PrintDescriptorsError("Missing deopt id"); |
| } |
| intptr_t var_desc_len = var_descriptors_.Length(); |
| bool found = false; |
| for (intptr_t cur_idx = 0; cur_idx < var_desc_len; cur_idx++) { |
| RawLocalVarDescriptors::VarInfo var_info; |
| var_descriptors_.GetInfo(cur_idx, &var_info); |
| const int8_t kind = var_info.kind(); |
| if ((kind == RawLocalVarDescriptors::kContextLevel) && |
| (deopt_id >= var_info.begin_pos.value()) && |
| (deopt_id <= var_info.end_pos.value())) { |
| context_level_ = var_info.index(); |
| found = true; |
| break; |
| } |
| } |
| if (!found) { |
| PrintDescriptorsError("Missing context level"); |
| } |
| ASSERT(context_level_ >= 0); |
| } |
| return context_level_; |
| } |
| |
| RawObject* ActivationFrame::GetAsyncContextVariable(const String& name) { |
| if (!function_.IsAsyncClosure() && !function_.IsAsyncGenClosure()) { |
| return Object::null(); |
| } |
| GetVarDescriptors(); |
| intptr_t var_desc_len = var_descriptors_.Length(); |
| for (intptr_t i = 0; i < var_desc_len; i++) { |
| RawLocalVarDescriptors::VarInfo var_info; |
| var_descriptors_.GetInfo(i, &var_info); |
| if (var_descriptors_.GetName(i) == name.raw()) { |
| const int8_t kind = var_info.kind(); |
| if (!live_frame_) { |
| ASSERT(kind == RawLocalVarDescriptors::kContextVar); |
| } |
| const auto variable_index = VariableIndex(var_info.index()); |
| if (kind == RawLocalVarDescriptors::kStackVar) { |
| return GetStackVar(variable_index); |
| } else { |
| ASSERT(kind == RawLocalVarDescriptors::kContextVar); |
| if (!live_frame_) { |
| ASSERT(!ctx_.IsNull()); |
| return ctx_.At(variable_index.value()); |
| } |
| return GetContextVar(var_info.scope_id, variable_index.value()); |
| } |
| } |
| } |
| return Object::null(); |
| } |
| |
| RawObject* ActivationFrame::GetAsyncCompleter() { |
| return GetAsyncContextVariable(Symbols::AsyncCompleter()); |
| } |
| |
| RawObject* ActivationFrame::GetAsyncCompleterAwaiter(const Object& completer) { |
| Instance& future = Instance::Handle(); |
| if (FLAG_sync_async) { |
| const Class& completer_cls = Class::Handle(completer.clazz()); |
| ASSERT(!completer_cls.IsNull()); |
| const Function& future_getter = Function::Handle( |
| completer_cls.LookupGetterFunction(Symbols::CompleterFuture())); |
| ASSERT(!future_getter.IsNull()); |
| const Array& args = Array::Handle(Array::New(1)); |
| args.SetAt(0, Instance::Cast(completer)); |
| future ^= DartEntry::InvokeFunction(future_getter, args); |
| } else { |
| const Class& sync_completer_cls = Class::Handle(completer.clazz()); |
| ASSERT(!sync_completer_cls.IsNull()); |
| const Class& completer_cls = Class::Handle(sync_completer_cls.SuperClass()); |
| const Field& future_field = |
| Field::Handle(completer_cls.LookupInstanceFieldAllowPrivate( |
| Symbols::CompleterFuture())); |
| ASSERT(!future_field.IsNull()); |
| future ^= Instance::Cast(completer).GetField(future_field); |
| } |
| if (future.IsNull()) { |
| // The completer object may not be fully initialized yet. |
| return Object::null(); |
| } |
| const Class& future_cls = Class::Handle(future.clazz()); |
| ASSERT(!future_cls.IsNull()); |
| const Field& awaiter_field = Field::Handle( |
| future_cls.LookupInstanceFieldAllowPrivate(Symbols::_Awaiter())); |
| ASSERT(!awaiter_field.IsNull()); |
| return future.GetField(awaiter_field); |
| } |
| |
| RawObject* ActivationFrame::GetAsyncStreamControllerStream() { |
| return GetAsyncContextVariable(Symbols::ControllerStream()); |
| } |
| |
| RawObject* ActivationFrame::GetAsyncStreamControllerStreamAwaiter( |
| const Object& stream) { |
| const Class& stream_cls = Class::Handle(stream.clazz()); |
| ASSERT(!stream_cls.IsNull()); |
| const Class& stream_impl_cls = Class::Handle(stream_cls.SuperClass()); |
| const Field& awaiter_field = Field::Handle( |
| stream_impl_cls.LookupInstanceFieldAllowPrivate(Symbols::_Awaiter())); |
| ASSERT(!awaiter_field.IsNull()); |
| return Instance::Cast(stream).GetField(awaiter_field); |
| } |
| |
| RawObject* ActivationFrame::GetAsyncAwaiter() { |
| const Object& async_stream_controller_stream = |
| Object::Handle(GetAsyncStreamControllerStream()); |
| if (!async_stream_controller_stream.IsNull()) { |
| return GetAsyncStreamControllerStreamAwaiter( |
| async_stream_controller_stream); |
| } |
| const Object& completer = Object::Handle(GetAsyncCompleter()); |
| if (!completer.IsNull()) { |
| return GetAsyncCompleterAwaiter(completer); |
| } |
| return Object::null(); |
| } |
| |
| RawObject* ActivationFrame::GetCausalStack() { |
| return GetAsyncContextVariable(Symbols::AsyncStackTraceVar()); |
| } |
| |
| bool ActivationFrame::HandlesException(const Instance& exc_obj) { |
| if ((kind_ == kAsyncSuspensionMarker) || (kind_ == kAsyncCausal)) { |
| // These frames are historical. |
| return false; |
| } |
| intptr_t try_index = TryIndex(); |
| if (try_index < 0) { |
| return false; |
| } |
| ExceptionHandlers& handlers = ExceptionHandlers::Handle(); |
| Array& handled_types = Array::Handle(); |
| AbstractType& type = Type::Handle(); |
| const bool is_async = |
| function().IsAsyncClosure() || function().IsAsyncGenClosure(); |
| handlers = code().exception_handlers(); |
| ASSERT(!handlers.IsNull()); |
| intptr_t num_handlers_checked = 0; |
| while (try_index != kInvalidTryIndex) { |
| // Detect circles in the exception handler data. |
| num_handlers_checked++; |
| ASSERT(num_handlers_checked <= handlers.num_entries()); |
| // Only consider user written handlers for async methods. |
| if (!is_async || !handlers.IsGenerated(try_index)) { |
| handled_types = handlers.GetHandledTypes(try_index); |
| const intptr_t num_types = handled_types.Length(); |
| for (intptr_t k = 0; k < num_types; k++) { |
| type ^= handled_types.At(k); |
| ASSERT(!type.IsNull()); |
| // Uninstantiated types are not added to ExceptionHandlers data. |
| ASSERT(type.IsInstantiated()); |
| if (type.IsMalformed()) { |
| continue; |
| } |
| if (type.IsDynamicType()) { |
| return true; |
| } |
| if (exc_obj.IsInstanceOf(type, Object::null_type_arguments(), |
| Object::null_type_arguments(), NULL)) { |
| return true; |
| } |
| } |
| } |
| try_index = handlers.OuterTryIndex(try_index); |
| } |
| return false; |
| } |
| |
| void ActivationFrame::ExtractTokenPositionFromAsyncClosure() { |
| // Attempt to determine the token position from the async closure. |
| Thread* thread = Thread::Current(); |
| Zone* zone = thread->zone(); |
| const Script& script = Script::Handle(zone, function().script()); |
| |
| ASSERT(function_.IsAsyncGenClosure() || function_.IsAsyncClosure()); |
| // This should only be called on frames that aren't active on the stack. |
| ASSERT(fp() == 0); |
| |
| const Array& await_to_token_map = |
| Array::Handle(zone, script.kind() == RawScript::kKernelTag |
| ? script.yield_positions() |
| : code_.await_token_positions()); |
| if (await_to_token_map.IsNull()) { |
| // No mapping. |
| return; |
| } |
| GetVarDescriptors(); |
| GetPcDescriptors(); |
| intptr_t var_desc_len = var_descriptors_.Length(); |
| intptr_t await_jump_var = -1; |
| for (intptr_t i = 0; i < var_desc_len; i++) { |
| RawLocalVarDescriptors::VarInfo var_info; |
| var_descriptors_.GetInfo(i, &var_info); |
| const int8_t kind = var_info.kind(); |
| if (var_descriptors_.GetName(i) == Symbols::AwaitJumpVar().raw()) { |
| ASSERT(kind == RawLocalVarDescriptors::kContextVar); |
| ASSERT(!ctx_.IsNull()); |
| Object& await_jump_index = Object::Handle(ctx_.At(var_info.index())); |
| ASSERT(await_jump_index.IsSmi()); |
| await_jump_var = Smi::Cast(await_jump_index).Value(); |
| } |
| } |
| if (await_jump_var < 0) { |
| return; |
| } |
| intptr_t await_to_token_map_index = |
| script.kind() == RawScript::kKernelTag |
| ? await_jump_var - 1 |
| : |
| // source script tokens array has first element duplicated |
| await_jump_var; |
| |
| if (script.kind() == RawScript::kKernelTag) { |
| // yield_positions returns all yield positions for the script (in sorted |
| // order). |
| // We thus need to offset the function start to get the actual index. |
| if (!function_.token_pos().IsReal()) { |
| return; |
| } |
| const intptr_t function_start = function_.token_pos().value(); |
| for (intptr_t i = 0; |
| i < await_to_token_map.Length() && |
| Smi::Value(reinterpret_cast<RawSmi*>(await_to_token_map.At(i))) < |
| function_start; |
| i++) { |
| await_to_token_map_index++; |
| } |
| } |
| |
| if (await_to_token_map_index >= await_to_token_map.Length()) { |
| return; |
| } |
| |
| const Object& token_pos = |
| Object::Handle(await_to_token_map.At(await_to_token_map_index)); |
| if (token_pos.IsNull()) { |
| return; |
| } |
| ASSERT(token_pos.IsSmi()); |
| token_pos_ = TokenPosition(Smi::Cast(token_pos).Value()); |
| token_pos_initialized_ = true; |
| PcDescriptors::Iterator iter(pc_desc_, RawPcDescriptors::kAnyKind); |
| while (iter.MoveNext()) { |
| if (iter.TokenPos() == token_pos_) { |
| // Match the lowest try index at this token position. |
| // TODO(johnmccutchan): Is this heuristic precise enough? |
| if (iter.TryIndex() != kInvalidTryIndex) { |
| if ((try_index_ == -1) || (iter.TryIndex() < try_index_)) { |
| try_index_ = iter.TryIndex(); |
| } |
| } |
| } |
| } |
| } |
| |
| bool ActivationFrame::IsAsyncMachinery() const { |
| Isolate* isolate = Isolate::Current(); |
| if (function_.raw() == isolate->object_store()->complete_on_async_return()) { |
| // We are completing an async function's completer. |
| return true; |
| } |
| if (function_.Owner() == |
| isolate->object_store()->async_star_stream_controller()) { |
| // We are inside the async* stream controller code. |
| return true; |
| } |
| return false; |
| } |
| |
| // Get the saved current context of this activation. |
| const Context& ActivationFrame::GetSavedCurrentContext() { |
| if (!ctx_.IsNull()) return ctx_; |
| GetVarDescriptors(); |
| intptr_t var_desc_len = var_descriptors_.Length(); |
| Object& obj = Object::Handle(); |
| for (intptr_t i = 0; i < var_desc_len; i++) { |
| RawLocalVarDescriptors::VarInfo var_info; |
| var_descriptors_.GetInfo(i, &var_info); |
| const int8_t kind = var_info.kind(); |
| if (kind == RawLocalVarDescriptors::kSavedCurrentContext) { |
| if (FLAG_trace_debugger_stacktrace) { |
| OS::PrintErr("\tFound saved current ctx at index %d\n", |
| var_info.index()); |
| } |
| const auto variable_index = VariableIndex(var_info.index()); |
| obj = GetStackVar(variable_index); |
| if (obj.IsClosure()) { |
| ASSERT(function().name() == Symbols::Call().raw()); |
| ASSERT(function().IsInvokeFieldDispatcher()); |
| // Closure.call frames. |
| ctx_ ^= Closure::Cast(obj).context(); |
| } else if (obj.IsContext()) { |
| ctx_ ^= Context::Cast(obj).raw(); |
| } else { |
| ASSERT(obj.IsNull()); |
| } |
| return ctx_; |
| } |
| } |
| return ctx_; |
| } |
| |
| RawObject* ActivationFrame::GetAsyncOperation() { |
| GetVarDescriptors(); |
| intptr_t var_desc_len = var_descriptors_.Length(); |
| for (intptr_t i = 0; i < var_desc_len; i++) { |
| RawLocalVarDescriptors::VarInfo var_info; |
| var_descriptors_.GetInfo(i, &var_info); |
| if (var_descriptors_.GetName(i) == Symbols::AsyncOperation().raw()) { |
| const int8_t kind = var_info.kind(); |
| const auto variable_index = VariableIndex(var_info.index()); |
| if (kind == RawLocalVarDescriptors::kStackVar) { |
| return GetStackVar(variable_index); |
| } else { |
| ASSERT(kind == RawLocalVarDescriptors::kContextVar); |
| return GetContextVar(var_info.scope_id, variable_index.value()); |
| } |
| } |
| } |
| return Object::null(); |
| } |
| |
| ActivationFrame* DebuggerStackTrace::GetHandlerFrame( |
| const Instance& exc_obj) const { |
| for (intptr_t frame_index = 0; frame_index < Length(); frame_index++) { |
| ActivationFrame* frame = FrameAt(frame_index); |
| if (frame->HandlesException(exc_obj)) { |
| return frame; |
| } |
| } |
| return NULL; |
| } |
| |
| void ActivationFrame::GetDescIndices() { |
| if (vars_initialized_) { |
| return; |
| } |
| GetVarDescriptors(); |
| |
| TokenPosition activation_token_pos = TokenPos(); |
| if (!activation_token_pos.IsDebugPause() || !live_frame_) { |
| // We don't have a token position for this frame, so can't determine |
| // which variables are visible. |
| vars_initialized_ = true; |
| return; |
| } |
| |
| GrowableArray<String*> var_names(8); |
| intptr_t var_desc_len = var_descriptors_.Length(); |
| for (intptr_t cur_idx = 0; cur_idx < var_desc_len; cur_idx++) { |
| ASSERT(var_names.length() == desc_indices_.length()); |
| RawLocalVarDescriptors::VarInfo var_info; |
| var_descriptors_.GetInfo(cur_idx, &var_info); |
| const int8_t kind = var_info.kind(); |
| if ((kind != RawLocalVarDescriptors::kStackVar) && |
| (kind != RawLocalVarDescriptors::kContextVar)) { |
| continue; |
| } |
| if ((var_info.begin_pos <= activation_token_pos) && |
| (activation_token_pos <= var_info.end_pos)) { |
| if ((kind == RawLocalVarDescriptors::kContextVar) && |
| (ContextLevel() < var_info.scope_id)) { |
| // The variable is textually in scope but the context level |
| // at the activation frame's PC is lower than the context |
| // level of the variable. The context containing the variable |
| // has already been removed from the chain. This can happen when we |
| // break at a return statement, since the contexts get discarded |
| // before the debugger gets called. |
| continue; |
| } |
| // The current variable is textually in scope. Now check whether |
| // there is another local variable with the same name that shadows |
| // or is shadowed by this variable. |
| String& var_name = String::Handle(var_descriptors_.GetName(cur_idx)); |
| intptr_t indices_len = desc_indices_.length(); |
| bool name_match_found = false; |
| for (intptr_t i = 0; i < indices_len; i++) { |
| if (var_name.Equals(*var_names[i])) { |
| // Found two local variables with the same name. Now determine |
| // which one is shadowed. |
| name_match_found = true; |
| RawLocalVarDescriptors::VarInfo i_var_info; |
| var_descriptors_.GetInfo(desc_indices_[i], &i_var_info); |
| if (i_var_info.begin_pos < var_info.begin_pos) { |
| // The variable we found earlier is in an outer scope |
| // and is shadowed by the current variable. Replace the |
| // descriptor index of the previously found variable |
| // with the descriptor index of the current variable. |
| desc_indices_[i] = cur_idx; |
| } else { |
| // The variable we found earlier is in an inner scope |
| // and shadows the current variable. Skip the current |
| // variable. (Nothing to do.) |
| } |
| break; // Stop looking for name matches. |
| } |
| } |
| if (!name_match_found) { |
| // No duplicate name found. Add the current descriptor index to the |
| // list of visible variables. |
| desc_indices_.Add(cur_idx); |
| var_names.Add(&var_name); |
| } |
| } |
| } |
| vars_initialized_ = true; |
| } |
| |
| intptr_t ActivationFrame::NumLocalVariables() { |
| GetDescIndices(); |
| return desc_indices_.length(); |
| } |
| |
| DART_FORCE_INLINE static RawObject* GetVariableValue(uword addr) { |
| return *reinterpret_cast<RawObject**>(addr); |
| } |
| |
| RawObject* ActivationFrame::GetParameter(intptr_t index) { |
| intptr_t num_parameters = function().num_fixed_parameters(); |
| ASSERT(0 <= index && index < num_parameters); |
| |
| if (function().NumOptionalParameters() > 0) { |
| // If the function has optional parameters, the first positional parameter |
| // can be in a number of places in the caller's frame depending on how many |
| // were actually supplied at the call site, but they are copied to a fixed |
| // place in the callee's frame. |
| |
| return GetVariableValue(LocalVarAddress( |
| fp(), runtime_frame_layout.FrameSlotForVariableIndex(-index))); |
| } else { |
| intptr_t reverse_index = num_parameters - index; |
| return GetVariableValue(ParamAddress(fp(), reverse_index)); |
| } |
| } |
| |
| RawObject* ActivationFrame::GetClosure() { |
| ASSERT(function().IsClosureFunction()); |
| return GetParameter(0); |
| } |
| |
| RawObject* ActivationFrame::GetStackVar(VariableIndex variable_index) { |
| const intptr_t slot_index = |
| runtime_frame_layout.FrameSlotForVariableIndex(variable_index.value()); |
| if (deopt_frame_.IsNull()) { |
| return GetVariableValue(LocalVarAddress(fp(), slot_index)); |
| } else { |
| return deopt_frame_.At(LocalVarIndex(deopt_frame_offset_, slot_index)); |
| } |
| } |
| |
| bool ActivationFrame::IsRewindable() const { |
| if (deopt_frame_.IsNull()) { |
| return true; |
| } |
| // TODO(turnidge): This is conservative. It looks at all values in |
| // the deopt_frame_ even though some of them may correspond to other |
| // inlined frames. |
| Object& obj = Object::Handle(); |
| for (int i = 0; i < deopt_frame_.Length(); i++) { |
| obj = deopt_frame_.At(i); |
| if (obj.raw() == Symbols::OptimizedOut().raw()) { |
| return false; |
| } |
| } |
| return true; |
| } |
| |
| void ActivationFrame::PrintContextMismatchError(intptr_t ctx_slot, |
| intptr_t frame_ctx_level, |
| intptr_t var_ctx_level) { |
| OS::PrintErr( |
| "-------------------------\n" |
| "Encountered context mismatch\n" |
| "\tctx_slot: %" Pd |
| "\n" |
| "\tframe_ctx_level: %" Pd |
| "\n" |
| "\tvar_ctx_level: %" Pd "\n\n", |
| ctx_slot, frame_ctx_level, var_ctx_level); |
| |
| OS::PrintErr( |
| "-------------------------\n" |
| "Current frame:\n%s\n", |
| this->ToCString()); |
| |
| OS::PrintErr( |
| "-------------------------\n" |
| "Context contents:\n"); |
| const Context& ctx = GetSavedCurrentContext(); |
| ctx.Dump(8); |
| |
| OS::PrintErr( |
| "-------------------------\n" |
| "Debugger stack trace...\n\n"); |
| DebuggerStackTrace* stack = Isolate::Current()->debugger()->StackTrace(); |
| intptr_t num_frames = stack->Length(); |
| for (intptr_t i = 0; i < num_frames; i++) { |
| ActivationFrame* frame = stack->FrameAt(i); |
| OS::PrintErr("#%04" Pd " %s", i, frame->ToCString()); |
| } |
| |
| OS::PrintErr( |
| "-------------------------\n" |
| "All frames...\n\n"); |
| StackFrameIterator iterator(ValidationPolicy::kDontValidateFrames, |
| Thread::Current(), |
| StackFrameIterator::kNoCrossThreadIteration); |
| StackFrame* frame = iterator.NextFrame(); |
| intptr_t num = 0; |
| while ((frame != NULL)) { |
| OS::PrintErr("#%04" Pd " %s\n", num++, frame->ToCString()); |
| frame = iterator.NextFrame(); |
| } |
| } |
| |
| void ActivationFrame::VariableAt(intptr_t i, |
| String* name, |
| TokenPosition* declaration_token_pos, |
| TokenPosition* visible_start_token_pos, |
| TokenPosition* visible_end_token_pos, |
| Object* value) { |
| GetDescIndices(); |
| ASSERT(i < desc_indices_.length()); |
| intptr_t desc_index = desc_indices_[i]; |
| ASSERT(name != NULL); |
| |
| *name = var_descriptors_.GetName(desc_index); |
| |
| RawLocalVarDescriptors::VarInfo var_info; |
| var_descriptors_.GetInfo(desc_index, &var_info); |
| ASSERT(declaration_token_pos != NULL); |
| *declaration_token_pos = var_info.declaration_pos; |
| ASSERT(visible_start_token_pos != NULL); |
| *visible_start_token_pos = var_info.begin_pos; |
| ASSERT(visible_end_token_pos != NULL); |
| *visible_end_token_pos = var_info.end_pos; |
| ASSERT(value != NULL); |
| const int8_t kind = var_info.kind(); |
| const auto variable_index = VariableIndex(var_info.index()); |
| if (kind == RawLocalVarDescriptors::kStackVar) { |
| *value = GetStackVar(variable_index); |
| } else { |
| ASSERT(kind == RawLocalVarDescriptors::kContextVar); |
| *value = GetContextVar(var_info.scope_id, variable_index.value()); |
| } |
| } |
| |
| RawObject* ActivationFrame::GetContextVar(intptr_t var_ctx_level, |
| intptr_t ctx_slot) { |
| const Context& ctx = GetSavedCurrentContext(); |
| ASSERT(!ctx.IsNull()); |
| |
| // The context level at the PC/token index of this activation frame. |
| intptr_t frame_ctx_level = ContextLevel(); |
| |
| intptr_t level_diff = frame_ctx_level - var_ctx_level; |
| if (level_diff == 0) { |
| if ((ctx_slot < 0) || (ctx_slot >= ctx.num_variables())) { |
| PrintContextMismatchError(ctx_slot, frame_ctx_level, var_ctx_level); |
| } |
| ASSERT((ctx_slot >= 0) && (ctx_slot < ctx.num_variables())); |
| return ctx.At(ctx_slot); |
| } else { |
| ASSERT(level_diff > 0); |
| Context& var_ctx = Context::Handle(ctx.raw()); |
| while (level_diff > 0 && !var_ctx.IsNull()) { |
| level_diff--; |
| var_ctx = var_ctx.parent(); |
| } |
| if (var_ctx.IsNull() || (ctx_slot < 0) || |
| (ctx_slot >= var_ctx.num_variables())) { |
| PrintContextMismatchError(ctx_slot, frame_ctx_level, var_ctx_level); |
| } |
| ASSERT(!var_ctx.IsNull()); |
| ASSERT((ctx_slot >= 0) && (ctx_slot < var_ctx.num_variables())); |
| return var_ctx.At(ctx_slot); |
| } |
| } |
| |
| RawArray* ActivationFrame::GetLocalVariables() { |
| GetDescIndices(); |
| intptr_t num_variables = desc_indices_.length(); |
| String& var_name = String::Handle(); |
| Object& value = Instance::Handle(); |
| const Array& list = Array::Handle(Array::New(2 * num_variables)); |
| for (intptr_t i = 0; i < num_variables; i++) { |
| TokenPosition ignore; |
| VariableAt(i, &var_name, &ignore, &ignore, &ignore, &value); |
| list.SetAt(2 * i, var_name); |
| list.SetAt((2 * i) + 1, value); |
| } |
| return list.raw(); |
| } |
| |
| RawObject* ActivationFrame::GetReceiver() { |
| GetDescIndices(); |
| intptr_t num_variables = desc_indices_.length(); |
| String& var_name = String::Handle(); |
| Instance& value = Instance::Handle(); |
| for (intptr_t i = 0; i < num_variables; i++) { |
| TokenPosition ignore; |
| VariableAt(i, &var_name, &ignore, &ignore, &ignore, &value); |
| if (var_name.Equals(Symbols::This())) { |
| return value.raw(); |
| } |
| } |
| return Symbols::OptimizedOut().raw(); |
| } |
| |
| static bool IsSyntheticVariableName(const String& var_name) { |
| return (var_name.Length() >= 1) && (var_name.CharAt(0) == ':'); |
| } |
| |
| static bool IsPrivateVariableName(const String& var_name) { |
| return (var_name.Length() >= 1) && (var_name.CharAt(0) == '_'); |
| } |
| |
| RawObject* ActivationFrame::EvaluateCompiledExpression( |
| const uint8_t* kernel_bytes, |
| intptr_t kernel_length, |
| const Array& type_definitions, |
| const Array& arguments, |
| const TypeArguments& type_arguments) { |
| if (function().is_static()) { |
| const Class& cls = Class::Handle(function().Owner()); |
| return cls.EvaluateCompiledExpression(kernel_bytes, kernel_length, |
| type_definitions, arguments, |
| type_arguments); |
| } else { |
| const Object& receiver = Object::Handle(GetReceiver()); |
| const Class& method_cls = Class::Handle(function().origin()); |
| ASSERT(receiver.IsInstance() || receiver.IsNull()); |
| if (!(receiver.IsInstance() || receiver.IsNull())) { |
| return Object::null(); |
| } |
| const Instance& inst = Instance::Cast(receiver); |
| return inst.EvaluateCompiledExpression(method_cls, kernel_bytes, |
| kernel_length, type_definitions, |
| arguments, type_arguments); |
| } |
| } |
| |
| RawTypeArguments* ActivationFrame::BuildParameters( |
| const GrowableObjectArray& param_names, |
| const GrowableObjectArray& param_values, |
| const GrowableObjectArray& type_params_names) { |
| GetDescIndices(); |
| bool type_arguments_available = false; |
| String& name = String::Handle(); |
| String& existing_name = String::Handle(); |
| Object& value = Instance::Handle(); |
| TypeArguments& type_arguments = TypeArguments::Handle(); |
| intptr_t num_variables = desc_indices_.length(); |
| for (intptr_t i = 0; i < num_variables; i++) { |
| TokenPosition ignore; |
| VariableAt(i, &name, &ignore, &ignore, &ignore, &value); |
| if (name.Equals(Symbols::FunctionTypeArgumentsVar())) { |
| type_arguments_available = true; |
| type_arguments ^= value.raw(); |
| } else if (!name.Equals(Symbols::This()) && |
| !IsSyntheticVariableName(name)) { |
| if (IsPrivateVariableName(name)) { |
| name = String::ScrubName(name); |
| } |
| bool conflict = false; |
| for (intptr_t j = 0; j < param_names.Length(); j++) { |
| existing_name ^= param_names.At(j); |
| if (name.Equals(existing_name)) { |
| conflict = true; |
| break; |
| } |
| } |
| // If local has the same name as a binding in the incoming scope, prefer |
| // the one from the incoming scope, since it is logically a child scope |
| // of the activation's current scope. |
| if (!conflict) { |
| param_names.Add(name); |
| param_values.Add(value); |
| } |
| } |
| } |
| |
| if ((function().IsGeneric() || function().HasGenericParent()) && |
| type_arguments_available) { |
| intptr_t num_vars = |
| function().NumTypeParameters() + function().NumParentTypeParameters(); |
| type_params_names.Grow(num_vars); |
| type_params_names.SetLength(num_vars); |
| TypeArguments& type_params = TypeArguments::Handle(); |
| TypeParameter& type_param = TypeParameter::Handle(); |
| Function& current = Function::Handle(function().raw()); |
| for (intptr_t i = 0; !current.IsNull(); i += current.NumTypeParameters(), |
| current = current.parent_function()) { |
| type_params = current.type_parameters(); |
| for (intptr_t j = 0; j < current.NumTypeParameters(); ++j) { |
| type_param = TypeParameter::RawCast(type_params.TypeAt(j)); |
| name = type_param.Name(); |
| // Write the names in backwards so they match up with the order of the |
| // types in 'type_arguments'. |
| type_params_names.SetAt(num_vars - (i + j) - 1, name); |
| } |
| } |
| if (!type_arguments.IsNull()) { |
| if (type_arguments.Length() == 0) { |
| for (intptr_t i = 0; i < num_vars; ++i) { |
| type_arguments.SetTypeAt(i, Object::dynamic_type()); |
| } |
| } |
| ASSERT(type_arguments.Length() == num_vars); |
| } |
| } |
| |
| return type_arguments.raw(); |
| } |
| |
| const char* ActivationFrame::ToCString() { |
| const String& url = String::Handle(SourceUrl()); |
| intptr_t line = LineNumber(); |
| const char* func_name = Debugger::QualifiedFunctionName(function()); |
| return Thread::Current()->zone()->PrintToString( |
| "[ Frame pc(0x%" Px ") fp(0x%" Px ") sp(0x%" Px |
| ")\n" |
| "\tfunction = %s\n" |
| "\turl = %s\n" |
| "\tline = %" Pd |
| "\n" |
| "\tcontext = %s\n" |
| "\tcontext level = %" Pd " ]\n", |
| pc(), fp(), sp(), func_name, url.ToCString(), line, ctx_.ToCString(), |
| ContextLevel()); |
| } |
| |
| void ActivationFrame::PrintToJSONObject(JSONObject* jsobj, bool full) { |
| if (kind_ == kRegular) { |
| PrintToJSONObjectRegular(jsobj, full); |
| } else if (kind_ == kAsyncCausal) { |
| PrintToJSONObjectAsyncCausal(jsobj, full); |
| } else if (kind_ == kAsyncSuspensionMarker) { |
| PrintToJSONObjectAsyncSuspensionMarker(jsobj, full); |
| } else if (kind_ == kAsyncActivation) { |
| PrintToJSONObjectAsyncActivation(jsobj, full); |
| } else { |
| UNIMPLEMENTED(); |
| } |
| } |
| |
| void ActivationFrame::PrintToJSONObjectRegular(JSONObject* jsobj, bool full) { |
| const Script& script = Script::Handle(SourceScript()); |
| jsobj->AddProperty("type", "Frame"); |
| jsobj->AddProperty("kind", KindToCString(kind_)); |
| const TokenPosition pos = TokenPos().SourcePosition(); |
| jsobj->AddLocation(script, pos); |
| jsobj->AddProperty("function", function(), !full); |
| jsobj->AddProperty("code", code()); |
| if (full) { |
| // TODO(cutch): The old "full" script usage no longer fits |
| // in the world where we pass the script as part of the |
| // location. |
| jsobj->AddProperty("script", script, !full); |
| } |
| { |
| JSONArray jsvars(jsobj, "vars"); |
| const int num_vars = NumLocalVariables(); |
| for (intptr_t v = 0; v < num_vars; v++) { |
| String& var_name = String::Handle(); |
| Instance& var_value = Instance::Handle(); |
| TokenPosition declaration_token_pos; |
| TokenPosition visible_start_token_pos; |
| TokenPosition visible_end_token_pos; |
| VariableAt(v, &var_name, &declaration_token_pos, &visible_start_token_pos, |
| &visible_end_token_pos, &var_value); |
| if ((var_name.raw() != Symbols::AsyncOperation().raw()) && |
| (var_name.raw() != Symbols::AsyncCompleter().raw()) && |
| (var_name.raw() != Symbols::ControllerStream().raw()) && |
| (var_name.raw() != Symbols::AwaitJumpVar().raw()) && |
| (var_name.raw() != Symbols::AsyncStackTraceVar().raw())) { |
| JSONObject jsvar(&jsvars); |
| jsvar.AddProperty("type", "BoundVariable"); |
| var_name = String::ScrubName(var_name); |
| jsvar.AddProperty("name", var_name.ToCString()); |
| jsvar.AddProperty("value", var_value, !full); |
| // Where was the variable declared? |
| jsvar.AddProperty("declarationTokenPos", declaration_token_pos); |
| // When the variable becomes visible to the scope. |
| jsvar.AddProperty("scopeStartTokenPos", visible_start_token_pos); |
| // When the variable stops being visible to the scope. |
| jsvar.AddProperty("scopeEndTokenPos", visible_end_token_pos); |
| } |
| } |
| } |
| } |
| |
| void ActivationFrame::PrintToJSONObjectAsyncCausal(JSONObject* jsobj, |
| bool full) { |
| jsobj->AddProperty("type", "Frame"); |
| jsobj->AddProperty("kind", KindToCString(kind_)); |
| const Script& script = Script::Handle(SourceScript()); |
| const TokenPosition pos = TokenPos().SourcePosition(); |
| jsobj->AddLocation(script, pos); |
| jsobj->AddProperty("function", function(), !full); |
| jsobj->AddProperty("code", code()); |
| if (full) { |
| // TODO(cutch): The old "full" script usage no longer fits |
| // in the world where we pass the script as part of the |
| // location. |
| jsobj->AddProperty("script", script, !full); |
| } |
| } |
| |
| void ActivationFrame::PrintToJSONObjectAsyncSuspensionMarker(JSONObject* jsobj, |
| bool full) { |
| jsobj->AddProperty("type", "Frame"); |
| jsobj->AddProperty("kind", KindToCString(kind_)); |
| jsobj->AddProperty("marker", "AsynchronousSuspension"); |
| } |
| |
| void ActivationFrame::PrintToJSONObjectAsyncActivation(JSONObject* jsobj, |
| bool full) { |
| jsobj->AddProperty("type", "Frame"); |
| jsobj->AddProperty("kind", KindToCString(kind_)); |
| const Script& script = Script::Handle(SourceScript()); |
| const TokenPosition pos = TokenPos().SourcePosition(); |
| jsobj->AddLocation(script, pos); |
| jsobj->AddProperty("function", function(), !full); |
| jsobj->AddProperty("code", code()); |
| if (full) { |
| // TODO(cutch): The old "full" script usage no longer fits |
| // in the world where we pass the script as part of the |
| // location. |
| jsobj->AddProperty("script", script, !full); |
| } |
| } |
| |
| static bool IsFunctionVisible(const Function& function) { |
| return FLAG_show_invisible_frames || function.is_visible(); |
| } |
| |
| void DebuggerStackTrace::AddActivation(ActivationFrame* frame) { |
| if (IsFunctionVisible(frame->function())) { |
| trace_.Add(frame); |
| } |
| } |
| |
| void DebuggerStackTrace::AddMarker(ActivationFrame::Kind marker) { |
| ASSERT(marker == ActivationFrame::kAsyncSuspensionMarker); |
| trace_.Add(new ActivationFrame(marker)); |
| } |
| |
| void DebuggerStackTrace::AddAsyncCausalFrame(uword pc, const Code& code) { |
| trace_.Add(new ActivationFrame(pc, 0, 0, code, Array::Handle(), 0, |
| ActivationFrame::kAsyncCausal)); |
| } |
| |
| const uint8_t kSafepointKind = RawPcDescriptors::kIcCall | |
| RawPcDescriptors::kUnoptStaticCall | |
| RawPcDescriptors::kRuntimeCall; |
| |
| CodeBreakpoint::CodeBreakpoint(const Code& code, |
| TokenPosition token_pos, |
| uword pc, |
| RawPcDescriptors::Kind kind) |
| : code_(code.raw()), |
| token_pos_(token_pos), |
| pc_(pc), |
| line_number_(-1), |
| is_enabled_(false), |
| bpt_location_(NULL), |
| next_(NULL), |
| breakpoint_kind_(kind), |
| #if !defined(TARGET_ARCH_DBC) |
| saved_value_(Code::null()) |
| #else |
| saved_value_(SimulatorBytecode::kTrap), |
| saved_value_fastsmi_(SimulatorBytecode::kTrap) |
| #endif |
| { |
| ASSERT(!code.IsNull()); |
| ASSERT(token_pos_.IsReal()); |
| ASSERT(pc_ != 0); |
| ASSERT((breakpoint_kind_ & kSafepointKind) != 0); |
| } |
| |
| CodeBreakpoint::~CodeBreakpoint() { |
| // Make sure we don't leave patched code behind. |
| ASSERT(!IsEnabled()); |
| // Poison the data so we catch use after free errors. |
| #ifdef DEBUG |
| code_ = Code::null(); |
| pc_ = 0ul; |
| bpt_location_ = NULL; |
| next_ = NULL; |
| breakpoint_kind_ = RawPcDescriptors::kOther; |
| #endif |
| } |
| |
| RawFunction* CodeBreakpoint::function() const { |
| return Code::Handle(code_).function(); |
| } |
| |
| RawScript* CodeBreakpoint::SourceCode() { |
| const Function& func = Function::Handle(this->function()); |
| return func.script(); |
| } |
| |
| RawString* CodeBreakpoint::SourceUrl() { |
| const Script& script = Script::Handle(SourceCode()); |
| return script.url(); |
| } |
| |
| intptr_t CodeBreakpoint::LineNumber() { |
| // Compute line number lazily since it causes scanning of the script. |
| if (line_number_ < 0) { |
| const Script& script = Script::Handle(SourceCode()); |
| script.GetTokenLocation(token_pos_, &line_number_, NULL); |
| } |
| return line_number_; |
| } |
| |
| void CodeBreakpoint::Enable() { |
| if (!is_enabled_) { |
| PatchCode(); |
| } |
| ASSERT(is_enabled_); |
| } |
| |
| void CodeBreakpoint::Disable() { |
| if (is_enabled_) { |
| RestoreCode(); |
| } |
| ASSERT(!is_enabled_); |
| } |
| |
| Debugger::Debugger(Isolate* isolate) |
| : isolate_(isolate), |
| next_id_(1), |
| latent_locations_(NULL), |
| breakpoint_locations_(NULL), |
| code_breakpoints_(NULL), |
| resume_action_(kContinue), |
| resume_frame_index_(-1), |
| post_deopt_frame_index_(-1), |
| ignore_breakpoints_(false), |
| pause_event_(NULL), |
| stack_trace_(NULL), |
| async_causal_stack_trace_(NULL), |
| awaiter_stack_trace_(NULL), |
| stepping_fp_(0), |
| async_stepping_fp_(0), |
| top_frame_awaiter_(Object::null()), |
| skip_next_step_(false), |
| needs_breakpoint_cleanup_(false), |
| synthetic_async_breakpoint_(NULL), |
| exc_pause_info_(kNoPauseOnExceptions) {} |
| |
| Debugger::~Debugger() { |
| ASSERT(!IsPaused()); |
| ASSERT(latent_locations_ == NULL); |
| ASSERT(breakpoint_locations_ == NULL); |
| ASSERT(code_breakpoints_ == NULL); |
| ASSERT(stack_trace_ == NULL); |
| ASSERT(async_causal_stack_trace_ == NULL); |
| ASSERT(synthetic_async_breakpoint_ == NULL); |
| } |
| |
| void Debugger::Shutdown() { |
| // TODO(johnmccutchan): Do not create a debugger for isolates that don't need |
| // them. Then, assert here that isolate_ is not one of those isolates. |
| if (Isolate::IsVMInternalIsolate(isolate_)) { |
| return; |
| } |
| while (breakpoint_locations_ != NULL) { |
| BreakpointLocation* bpt = breakpoint_locations_; |
| breakpoint_locations_ = breakpoint_locations_->next(); |
| delete bpt; |
| } |
| while (latent_locations_ != NULL) { |
| BreakpointLocation* bpt = latent_locations_; |
| latent_locations_ = latent_locations_->next(); |
| delete bpt; |
| } |
| while (code_breakpoints_ != NULL) { |
| CodeBreakpoint* bpt = code_breakpoints_; |
| code_breakpoints_ = code_breakpoints_->next(); |
| bpt->Disable(); |
| delete bpt; |
| } |
| if (NeedsIsolateEvents()) { |
| ServiceEvent event(isolate_, ServiceEvent::kIsolateExit); |
| InvokeEventHandler(&event); |
| } |
| } |
| |
| void Debugger::OnIsolateRunnable() {} |
| |
| static RawFunction* ResolveLibraryFunction(const Library& library, |
| const String& fname) { |
| ASSERT(!library.IsNull()); |
| const Object& object = Object::Handle(library.ResolveName(fname)); |
| if (!object.IsNull() && object.IsFunction()) { |
| return Function::Cast(object).raw(); |
| } |
| return Function::null(); |
| } |
| |
| bool Debugger::SetupStepOverAsyncSuspension(const char** error) { |
| ActivationFrame* top_frame = TopDartFrame(); |
| if (!IsAtAsyncJump(top_frame)) { |
| // Not at an async operation. |
| if (error) { |
| *error = "Isolate must be paused at an async suspension point"; |
| } |
| return false; |
| } |
| Object& closure = Object::Handle(top_frame->GetAsyncOperation()); |
| ASSERT(!closure.IsNull()); |
| ASSERT(closure.IsInstance()); |
| ASSERT(Instance::Cast(closure).IsClosure()); |
| Breakpoint* bpt = SetBreakpointAtActivation(Instance::Cast(closure), true); |
| if (bpt == NULL) { |
| // Unable to set the breakpoint. |
| if (error) { |
| *error = "Unable to set breakpoint at async suspension point"; |
| } |
| return false; |
| } |
| return true; |
| } |
| |
| bool Debugger::SetResumeAction(ResumeAction action, |
| intptr_t frame_index, |
| const char** error) { |
| if (error) { |
| *error = NULL; |
| } |
| resume_frame_index_ = -1; |
| switch (action) { |
| case kStepInto: |
| case kStepOver: |
| case kStepOut: |
| case kContinue: |
| resume_action_ = action; |
| return true; |
| case kStepRewind: |
| if (!CanRewindFrame(frame_index, error)) { |
| return false; |
| } |
| resume_action_ = kStepRewind; |
| resume_frame_index_ = frame_index; |
| return true; |
| case kStepOverAsyncSuspension: |
| return SetupStepOverAsyncSuspension(error); |
| default: |
| UNREACHABLE(); |
| return false; |
| } |
| } |
| |
| RawFunction* Debugger::ResolveFunction(const Library& library, |
| const String& class_name, |
| const String& function_name) { |
| ASSERT(!library.IsNull()); |
| ASSERT(!class_name.IsNull()); |
| ASSERT(!function_name.IsNull()); |
| if (class_name.Length() == 0) { |
| return ResolveLibraryFunction(library, function_name); |
| } |
| const Class& cls = Class::Handle(library.LookupClass(class_name)); |
| Function& function = Function::Handle(); |
| if (!cls.IsNull()) { |
| function = cls.LookupStaticFunction(function_name); |
| if (function.IsNull()) { |
| function = cls.LookupDynamicFunction(function_name); |
| } |
| } |
| return function.raw(); |
| } |
| |
| // Deoptimize all functions in the isolate. |
| // TODO(hausner): Actually we only need to deoptimize those functions |
| // that inline the function that contains the newly created breakpoint. |
| // We currently don't have this info so we deoptimize all functions. |
| void Debugger::DeoptimizeWorld() { |
| BackgroundCompiler::Stop(isolate_); |
| if (FLAG_trace_deoptimization) { |
| THR_Print("Deopt for debugger\n"); |
| } |
| DeoptimizeFunctionsOnStack(); |
| // Iterate over all classes, deoptimize functions. |
| // TODO(hausner): Could possibly be combined with RemoveOptimizedCode() |
| const ClassTable& class_table = *isolate_->class_table(); |
| Class& cls = Class::Handle(); |
| Array& functions = Array::Handle(); |
| GrowableObjectArray& closures = GrowableObjectArray::Handle(); |
| Function& function = Function::Handle(); |
| intptr_t num_classes = class_table.NumCids(); |
| for (intptr_t i = 1; i < num_classes; i++) { |
| if (class_table.HasValidClassAt(i)) { |
| cls = class_table.At(i); |
| |
| // Disable optimized functions. |
| functions = cls.functions(); |
| if (!functions.IsNull()) { |
| intptr_t num_functions = functions.Length(); |
| for (intptr_t pos = 0; pos < num_functions; pos++) { |
| function ^= functions.At(pos); |
| ASSERT(!function.IsNull()); |
| if (function.HasOptimizedCode()) { |
| function.SwitchToUnoptimizedCode(); |
| } |
| // Also disable any optimized implicit closure functions. |
| if (function.HasImplicitClosureFunction()) { |
| function = function.ImplicitClosureFunction(); |
| if (function.HasOptimizedCode()) { |
| function.SwitchToUnoptimizedCode(); |
| } |
| } |
| } |
| } |
| } |
| } |
| |
| // Disable optimized closure functions. |
| closures = isolate_->object_store()->closure_functions(); |
| const intptr_t num_closures = closures.Length(); |
| for (intptr_t pos = 0; pos < num_closures; pos++) { |
| function ^= closures.At(pos); |
| ASSERT(!function.IsNull()); |
| if (function.HasOptimizedCode()) { |
| function.SwitchToUnoptimizedCode(); |
| } |
| } |
| } |
| |
| ActivationFrame* Debugger::CollectDartFrame(Isolate* isolate, |
| uword pc, |
| StackFrame* frame, |
| const Code& code, |
| const Array& deopt_frame, |
| intptr_t deopt_frame_offset, |
| ActivationFrame::Kind kind) { |
| ASSERT(code.ContainsInstructionAt(pc)); |
| ActivationFrame* activation = |
| new ActivationFrame(pc, frame->fp(), frame->sp(), code, deopt_frame, |
| deopt_frame_offset, kind); |
| if (FLAG_trace_debugger_stacktrace) { |
| const Context& ctx = activation->GetSavedCurrentContext(); |
| OS::PrintErr("\tUsing saved context: %s\n", ctx.ToCString()); |
| } |
| if (FLAG_trace_debugger_stacktrace) { |
| OS::PrintErr("\tLine number: %" Pd "\n", activation->LineNumber()); |
| } |
| return activation; |
| } |
| |
| #if !defined(DART_PRECOMPILED_RUNTIME) |
| RawArray* Debugger::DeoptimizeToArray(Thread* thread, |
| StackFrame* frame, |
| const Code& code) { |
| ASSERT(code.is_optimized()); |
| Isolate* isolate = thread->isolate(); |
| // Create the DeoptContext for this deoptimization. |
| DeoptContext* deopt_context = |
| new DeoptContext(frame, code, DeoptContext::kDestIsAllocated, NULL, NULL, |
| true, false /* deoptimizing_code */); |
| isolate->set_deopt_context(deopt_context); |
| |
| deopt_context->FillDestFrame(); |
| deopt_context->MaterializeDeferredObjects(); |
| const Array& dest_frame = |
| Array::Handle(thread->zone(), deopt_context->DestFrameAsArray()); |
| |
| isolate->set_deopt_context(NULL); |
| delete deopt_context; |
| |
| return dest_frame.raw(); |
| } |
| #endif // !defined(DART_PRECOMPILED_RUNTIME) |
| |
| DebuggerStackTrace* Debugger::CollectStackTrace() { |
| Thread* thread = Thread::Current(); |
| Zone* zone = thread->zone(); |
| Isolate* isolate = thread->isolate(); |
| DebuggerStackTrace* stack_trace = new DebuggerStackTrace(8); |
| StackFrameIterator iterator(ValidationPolicy::kDontValidateFrames, |
| Thread::Current(), |
| StackFrameIterator::kNoCrossThreadIteration); |
| Code& code = Code::Handle(zone); |
| Code& inlined_code = Code::Handle(zone); |
| Array& deopt_frame = Array::Handle(zone); |
| |
| for (StackFrame* frame = iterator.NextFrame(); frame != NULL; |
| frame = iterator.NextFrame()) { |
| ASSERT(frame->IsValid()); |
| if (FLAG_trace_debugger_stacktrace) { |
| OS::PrintErr("CollectStackTrace: visiting frame:\n\t%s\n", |
| frame->ToCString()); |
| } |
| if (frame->IsDartFrame()) { |
| if (frame->is_interpreted()) { |
| // TODO(regis): Support debugging of interpreted frames. |
| // For now, do not abort, but skip the frame, as this code is run |
| // while handling a stack overflow. See HandleStackOverflowTestCases. |
| continue; |
| } |
| code = frame->LookupDartCode(); |
| AppendCodeFrames(thread, isolate, zone, stack_trace, frame, &code, |
| &inlined_code, &deopt_frame); |
| } |
| } |
| return stack_trace; |
| } |
| |
| void Debugger::AppendCodeFrames(Thread* thread, |
| Isolate* isolate, |
| Zone* zone, |
| DebuggerStackTrace* stack_trace, |
| StackFrame* frame, |
| Code* code, |
| Code* inlined_code, |
| Array* deopt_frame) { |
| #if !defined(DART_PRECOMPILED_RUNTIME) |
| if (code->is_optimized()) { |
| // TODO(rmacnak): Use CodeSourceMap |
| *deopt_frame = DeoptimizeToArray(thread, frame, *code); |
| for (InlinedFunctionsIterator it(*code, frame->pc()); !it.Done(); |
| it.Advance()) { |
| *inlined_code = it.code(); |
| if (FLAG_trace_debugger_stacktrace) { |
| const Function& function = Function::Handle(zone, it.function()); |
| ASSERT(!function.IsNull()); |
| OS::PrintErr("CollectStackTrace: visiting inlined function: %s\n", |
| function.ToFullyQualifiedCString()); |
| } |
| intptr_t deopt_frame_offset = it.GetDeoptFpOffset(); |
| stack_trace->AddActivation(CollectDartFrame(isolate, it.pc(), frame, |
| *inlined_code, *deopt_frame, |
| deopt_frame_offset)); |
| } |
| return; |
| } |
| #endif // !defined(DART_PRECOMPILED_RUNTIME) |
| stack_trace->AddActivation(CollectDartFrame(isolate, frame->pc(), frame, |
| *code, Object::null_array(), 0)); |
| } |
| |
| DebuggerStackTrace* Debugger::CollectAsyncCausalStackTrace() { |
| if (!FLAG_causal_async_stacks) { |
| return NULL; |
| } |
| Thread* thread = Thread::Current(); |
| Zone* zone = thread->zone(); |
| Isolate* isolate = thread->isolate(); |
| DebuggerStackTrace* stack_trace = new DebuggerStackTrace(8); |
| |
| Code& code = Code::Handle(zone); |
| Smi& offset = Smi::Handle(); |
| Code& inlined_code = Code::Handle(zone); |
| Array& deopt_frame = Array::Handle(zone); |
| |
| Function& async_function = Function::Handle(zone); |
| class StackTrace& async_stack_trace = StackTrace::Handle(zone); |
| Array& async_code_array = Array::Handle(zone); |
| Array& async_pc_offset_array = Array::Handle(zone); |
| StackTraceUtils::ExtractAsyncStackTraceInfo( |
| thread, &async_function, &async_stack_trace, &async_code_array, |
| &async_pc_offset_array); |
| |
| if (async_function.IsNull()) { |
| return NULL; |
| } |
| |
| intptr_t synchronous_stack_trace_length = |
| StackTraceUtils::CountFrames(thread, 0, async_function); |
| |
| // Append the top frames from the synchronous stack trace, up until the active |
| // asynchronous function. We truncate the remainder of the synchronous |
| // stack trace because it contains activations that are part of the |
| // asynchronous dispatch mechanisms. |
| StackFrameIterator iterator(ValidationPolicy::kDontValidateFrames, |
| Thread::Current(), |
| StackFrameIterator::kNoCrossThreadIteration); |
| StackFrame* frame = iterator.NextFrame(); |
| while (synchronous_stack_trace_length > 0) { |
| ASSERT(frame != NULL); |
| if (frame->IsDartFrame()) { |
| if (frame->is_interpreted()) { |
| // TODO(regis): Support debugging of interpreted frames. |
| UNIMPLEMENTED(); |
| } |
| code = frame->LookupDartCode(); |
| AppendCodeFrames(thread, isolate, zone, stack_trace, frame, &code, |
| &inlined_code, &deopt_frame); |
| synchronous_stack_trace_length--; |
| } |
| frame = iterator.NextFrame(); |
| } |
| |
| // Now we append the asynchronous causal stack trace. These are not active |
| // frames but a historical record of how this asynchronous function was |
| // activated. |
| while (!async_stack_trace.IsNull()) { |
| for (intptr_t i = 0; i < async_stack_trace.Length(); i++) { |
| if (async_stack_trace.CodeAtFrame(i) == Code::null()) { |
| break; |
| } |
| if (async_stack_trace.CodeAtFrame(i) == |
| StubCode::AsynchronousGapMarker().raw()) { |
| stack_trace->AddMarker(ActivationFrame::kAsyncSuspensionMarker); |
| // The frame immediately below the asynchronous gap marker is the |
| // identical to the frame above the marker. Skip the frame to enhance |
| // the readability of the trace. |
| i++; |
| } else { |
| code = Code::RawCast(async_stack_trace.CodeAtFrame(i)); |
| offset = Smi::RawCast(async_stack_trace.PcOffsetAtFrame(i)); |
| uword pc = code.PayloadStart() + offset.Value(); |
| if (code.is_optimized()) { |
| for (InlinedFunctionsIterator it(code, pc); !it.Done(); |
| it.Advance()) { |
| inlined_code = it.code(); |
| stack_trace->AddAsyncCausalFrame(it.pc(), inlined_code); |
| } |
| } else { |
| stack_trace->AddAsyncCausalFrame(pc, code); |
| } |
| } |
| } |
| // Follow the link. |
| async_stack_trace = async_stack_trace.async_link(); |
| } |
| |
| return stack_trace; |
| } |
| |
| #if !defined(DART_PRECOMPILED_RUNTIME) |
| static bool CheckAndSkipAsync(int skip_sync_async_frames_count, |
| const String& function_name) { |
| return (skip_sync_async_frames_count == 2 && |
| function_name.Equals(Symbols::_ClosureCall())) || |
| (skip_sync_async_frames_count == 1 && |
| function_name.Equals(Symbols::_AsyncAwaitCompleterStart())); |
| } |
| #endif |
| |
| DebuggerStackTrace* Debugger::CollectAwaiterReturnStackTrace() { |
| #if defined(DART_PRECOMPILED_RUNTIME) |
| // Causal async stacks are not supported in the AOT runtime. |
| ASSERT(!FLAG_async_debugger); |
| return NULL; |
| #else |
| if (!FLAG_async_debugger) { |
| return NULL; |
| } |
| |
| Thread* thread = Thread::Current(); |
| Zone* zone = thread->zone(); |
| Isolate* isolate = thread->isolate(); |
| DebuggerStackTrace* stack_trace = new DebuggerStackTrace(8); |
| |
| StackFrameIterator iterator(ValidationPolicy::kDontValidateFrames, |
| Thread::Current(), |
| StackFrameIterator::kNoCrossThreadIteration); |
| |
| Code& code = Code::Handle(zone); |
| Smi& offset = Smi::Handle(zone); |
| Function& function = Function::Handle(zone); |
| Code& inlined_code = Code::Handle(zone); |
| Closure& async_activation = Closure::Handle(zone); |
| Object& next_async_activation = Object::Handle(zone); |
| Array& deopt_frame = Array::Handle(zone); |
| class StackTrace& async_stack_trace = StackTrace::Handle(zone); |
| bool stack_has_async_function = false; |
| |
| // Number of frames we are trying to skip that form "sync async" entry. |
| int skip_sync_async_frames_count = -1; |
| String& function_name = String::Handle(zone); |
| for (StackFrame* frame = iterator.NextFrame(); frame != NULL; |
| frame = iterator.NextFrame()) { |
| ASSERT(frame->IsValid()); |
| if (FLAG_trace_debugger_stacktrace) { |
| OS::PrintErr("CollectAwaiterReturnStackTrace: visiting frame:\n\t%s\n", |
| frame->ToCString()); |
| } |
| if (frame->IsDartFrame()) { |
| if (frame->is_interpreted()) { |
| // TODO(regis): Support debugging of interpreted frames. |
| UNIMPLEMENTED(); |
| } |
| code = frame->LookupDartCode(); |
| if (code.is_optimized()) { |
| deopt_frame = DeoptimizeToArray(thread, frame, code); |
| bool found_async_awaiter = false; |
| bool abort_attempt_to_navigate_through_sync_async = false; |
| for (InlinedFunctionsIterator it(code, frame->pc()); !it.Done(); |
| it.Advance()) { |
| inlined_code = it.code(); |
| function = it.function(); |
| |
| if (skip_sync_async_frames_count > 0) { |
| function_name ^= function.QualifiedScrubbedName(); |
| if (CheckAndSkipAsync(skip_sync_async_frames_count, |
| function_name)) { |
| skip_sync_async_frames_count--; |
| } else { |
| // Unexpected function in sync async call |
| skip_sync_async_frames_count = -1; |
| abort_attempt_to_navigate_through_sync_async = true; |
| break; |
| } |
| } |
| |
| if (FLAG_trace_debugger_stacktrace) { |
| ASSERT(!function.IsNull()); |
| OS::PrintErr( |
| "CollectAwaiterReturnStackTrace: visiting inlined function: " |
| "%s\n", |
| function.ToFullyQualifiedCString()); |
| } |
| intptr_t deopt_frame_offset = it.GetDeoptFpOffset(); |
| if (function.IsAsyncClosure() || function.IsAsyncGenClosure()) { |
| ActivationFrame* activation = CollectDartFrame( |
| isolate, it.pc(), frame, inlined_code, deopt_frame, |
| deopt_frame_offset, ActivationFrame::kAsyncActivation); |
| ASSERT(activation != NULL); |
| stack_trace->AddActivation(activation); |
| stack_has_async_function = true; |
| // Grab the awaiter. |
| async_activation ^= activation->GetAsyncAwaiter(); |
| found_async_awaiter = true; |
| if (FLAG_sync_async) { |
| // async function might have been called synchronously, in which |
| // case we need to keep going down the stack. |
| // To determine how we are called we peek few more frames further |
| // expecting to see Closure_call followed by |
| // AsyncAwaitCompleter_start. |
| // If we are able to see those functions we continue going down |
| // thestack, if we are not, we break out of the loop as we are |
| // not interested in exploring rest of the stack - there is only |
| // dart-internal code left. |
| skip_sync_async_frames_count = 2; |
| } else { |
| break; |
| } |
| } else { |
| stack_trace->AddActivation( |
| CollectDartFrame(isolate, it.pc(), frame, inlined_code, |
| deopt_frame, deopt_frame_offset)); |
| } |
| } |
| // Break out of outer loop. |
| if (found_async_awaiter || |
| abort_attempt_to_navigate_through_sync_async) { |
| break; |
| } |
| } else { |
| function = code.function(); |
| |
| if (skip_sync_async_frames_count > 0) { |
| function_name ^= function.QualifiedScrubbedName(); |
| if (CheckAndSkipAsync(skip_sync_async_frames_count, function_name)) { |
| skip_sync_async_frames_count--; |
| } else { |
| // Unexpected function in sync async call. |
| break; |
| } |
| } |
| |
| if (function.IsAsyncClosure() || function.IsAsyncGenClosure()) { |
| ActivationFrame* activation = CollectDartFrame( |
| isolate, frame->pc(), frame, code, Object::null_array(), 0, |
| ActivationFrame::kAsyncActivation); |
| ASSERT(activation != NULL); |
| stack_trace->AddActivation(activation); |
| stack_has_async_function = true; |
| // Grab the awaiter. |
| async_activation ^= activation->GetAsyncAwaiter(); |
| async_stack_trace ^= activation->GetCausalStack(); |
| if (FLAG_sync_async) { |
| // see comment regarding skipping sync-async frames above. |
| skip_sync_async_frames_count = 2; |
| } else { |
| break; |
| } |
| } else { |
| stack_trace->AddActivation(CollectDartFrame( |
| isolate, frame->pc(), frame, code, Object::null_array(), 0)); |
| } |
| } |
| } |
| } |
| |
| // If the stack doesn't have any async functions on it, return NULL. |
| if (!stack_has_async_function) { |
| return NULL; |
| } |
| |
| // Append the awaiter return call stack. |
| while (!async_activation.IsNull()) { |
| ActivationFrame* activation = new (zone) ActivationFrame(async_activation); |
| activation->ExtractTokenPositionFromAsyncClosure(); |
| stack_trace->AddActivation(activation); |
| next_async_activation = activation->GetAsyncAwaiter(); |
| if (next_async_activation.IsNull()) { |
| // No more awaiters. Extract the causal stack trace (if it exists). |
| async_stack_trace ^= activation->GetCausalStack(); |
| break; |
| } |
| async_activation = Closure::RawCast(next_async_activation.raw()); |
| } |
| |
| // Now we append the asynchronous causal stack trace. These are not active |
| // frames but a historical record of how this asynchronous function was |
| // activated. |
| while (!async_stack_trace.IsNull()) { |
| for (intptr_t i = 0; i < async_stack_trace.Length(); i++) { |
| if (async_stack_trace.CodeAtFrame(i) == Code::null()) { |
| // Incomplete OutOfMemory/StackOverflow trace OR array padding. |
| break; |
| } |
| if (async_stack_trace.CodeAtFrame(i) == |
| StubCode::AsynchronousGapMarker().raw()) { |
| stack_trace->AddMarker(ActivationFrame::kAsyncSuspensionMarker); |
| // The frame immediately below the asynchronous gap marker is the |
| // identical to the frame above the marker. Skip the frame to enhance |
| // the readability of the trace. |
| i++; |
| } else { |
| code = Code::RawCast(async_stack_trace.CodeAtFrame(i)); |
| offset = Smi::RawCast(async_stack_trace.PcOffsetAtFrame(i)); |
| uword pc = code.PayloadStart() + offset.Value(); |
| if (code.is_optimized()) { |
| for (InlinedFunctionsIterator it(code, pc); !it.Done(); |
| it.Advance()) { |
| inlined_code = it.code(); |
| stack_trace->AddAsyncCausalFrame(it.pc(), inlined_code); |
| } |
| } else { |
| stack_trace->AddAsyncCausalFrame(pc, code); |
| } |
| } |
| } |
| // Follow the link. |
| async_stack_trace = async_stack_trace.async_link(); |
| } |
| |
| return stack_trace; |
| #endif // defined(DART_PRECOMPILED_RUNTIME) |
| } |
| |
| ActivationFrame* Debugger::TopDartFrame() const { |
| StackFrameIterator iterator(ValidationPolicy::kDontValidateFrames, |
| Thread::Current(), |
| StackFrameIterator::kNoCrossThreadIteration); |
| StackFrame* frame = iterator.NextFrame(); |
| while ((frame != NULL) && !frame->IsDartFrame()) { |
| frame = iterator.NextFrame(); |
| } |
| ASSERT(frame != NULL); |
| if (frame->is_interpreted()) { |
| // TODO(regis): Support debugging of interpreted frames. |
| UNIMPLEMENTED(); |
| return NULL; |
| } |
| Code& code = Code::Handle(frame->LookupDartCode()); |
| ActivationFrame* activation = new ActivationFrame( |
| frame->pc(), frame->fp(), frame->sp(), code, Object::null_array(), 0); |
| return activation; |
| } |
| |
| DebuggerStackTrace* Debugger::StackTrace() { |
| return (stack_trace_ != NULL) ? stack_trace_ : CollectStackTrace(); |
| } |
| |
| DebuggerStackTrace* Debugger::CurrentStackTrace() { |
| return CollectStackTrace(); |
| } |
| |
| DebuggerStackTrace* Debugger::AsyncCausalStackTrace() { |
| return (async_causal_stack_trace_ != NULL) ? async_causal_stack_trace_ |
| : CollectAsyncCausalStackTrace(); |
| } |
| |
| DebuggerStackTrace* Debugger::CurrentAsyncCausalStackTrace() { |
| return CollectAsyncCausalStackTrace(); |
| } |
| |
| DebuggerStackTrace* Debugger::AwaiterStackTrace() { |
| return (awaiter_stack_trace_ != NULL) ? awaiter_stack_trace_ |
| : CollectAwaiterReturnStackTrace(); |
| } |
| |
| DebuggerStackTrace* Debugger::CurrentAwaiterStackTrace() { |
| return CollectAwaiterReturnStackTrace(); |
| } |
| |
| DebuggerStackTrace* Debugger::StackTraceFrom(const class StackTrace& ex_trace) { |
| DebuggerStackTrace* stack_trace = new DebuggerStackTrace(8); |
| Function& function = Function::Handle(); |
| Object& code_object = Object::Handle(); |
| Code& code = Code::Handle(); |
| |
| const uword fp = 0; |
| const uword sp = 0; |
| const Array& deopt_frame = Array::Handle(); |
| const intptr_t deopt_frame_offset = -1; |
| |
| for (intptr_t i = 0; i < ex_trace.Length(); i++) { |
| code_object = ex_trace.CodeAtFrame(i); |
| // Pre-allocated StackTraces may include empty slots, either (a) to indicate |
| // where frames were omitted in the case a stack has more frames than the |
| // pre-allocated trace (such as a stack overflow) or (b) because a stack has |
| // fewer frames that the pre-allocated trace (such as memory exhaustion with |
| // a shallow stack). |
| if (!code_object.IsNull()) { |
| if (code_object.IsBytecode()) { |
| // TODO(regis): Support debugging of interpreted frames. |
| UNIMPLEMENTED(); |
| } |
| code ^= code_object.raw(); |
| ASSERT(code.IsFunctionCode()); |
| function = code.function(); |
| if (function.is_visible()) { |
| ASSERT(function.raw() == code.function()); |
| uword pc = |
| code.PayloadStart() + Smi::Value(ex_trace.PcOffsetAtFrame(i)); |
| if (code.is_optimized() && ex_trace.expand_inlined()) { |
| // Traverse inlined frames. |
| for (InlinedFunctionsIterator it(code, pc); !it.Done(); |
| it.Advance()) { |
| function = it.function(); |
| code = it.code(); |
| ASSERT(function.raw() == code.function()); |
| uword pc = it.pc(); |
| ASSERT(pc != 0); |
| ASSERT(code.PayloadStart() <= pc); |
| ASSERT(pc < (code.PayloadStart() + code.Size())); |
| |
| ActivationFrame* activation = new ActivationFrame( |
| pc, fp, sp, code, deopt_frame, deopt_frame_offset); |
| stack_trace->AddActivation(activation); |
| } |
| } else { |
| ActivationFrame* activation = new ActivationFrame( |
| pc, fp, sp, code, deopt_frame, deopt_frame_offset); |
| stack_trace->AddActivation(activation); |
| } |
| } |
| } |
| } |
| return stack_trace; |
| } |
| |
| void Debugger::SetExceptionPauseInfo(Dart_ExceptionPauseInfo pause_info) { |
| ASSERT((pause_info == kNoPauseOnExceptions) || |
| (pause_info == kPauseOnUnhandledExceptions) || |
| (pause_info == kPauseOnAllExceptions)); |
| exc_pause_info_ = pause_info; |
| } |
| |
| Dart_ExceptionPauseInfo Debugger::GetExceptionPauseInfo() const { |
| return exc_pause_info_; |
| } |
| |
| bool Debugger::ShouldPauseOnException(DebuggerStackTrace* stack_trace, |
| const Instance& exception) { |
| if (exc_pause_info_ == kNoPauseOnExceptions) { |
| return false; |
| } |
| if (exc_pause_info_ == kPauseOnAllExceptions) { |
| return true; |
| } |
| ASSERT(exc_pause_info_ == kPauseOnUnhandledExceptions); |
| ActivationFrame* handler_frame = stack_trace->GetHandlerFrame(exception); |
| if (handler_frame == NULL) { |
| // Did not find an exception handler that catches this exception. |
| // Note that this check is not precise, since we can't check |
| // uninstantiated types, i.e. types containing type parameters. |
| // Thus, we may report an exception as unhandled when in fact |
| // it will be caught once we unwind the stack. |
| return true; |
| } |
| return false; |
| } |
| |
| void Debugger::PauseException(const Instance& exc) { |
| if (FLAG_stress_async_stacks) { |
| CollectAwaiterReturnStackTrace(); |
| } |
| // We ignore this exception event when the VM is executing code invoked |
| // by the debugger to evaluate variables values, when we see a nested |
| // breakpoint or exception event, or if the debugger is not |
| // interested in exception events. |
| if (ignore_breakpoints_ || IsPaused() || |
| (exc_pause_info_ == kNoPauseOnExceptions)) { |
| return; |
| } |
| DebuggerStackTrace* awaiter_stack_trace = CollectAwaiterReturnStackTrace(); |
| DebuggerStackTrace* stack_trace = CollectStackTrace(); |
| if (awaiter_stack_trace != NULL) { |
| if (!ShouldPauseOnException(awaiter_stack_trace, exc)) { |
| return; |
| } |
| } else { |
| if (!ShouldPauseOnException(stack_trace, exc)) { |
| return; |
| } |
| } |
| ServiceEvent event(isolate_, ServiceEvent::kPauseException); |
| event.set_exception(&exc); |
| if (stack_trace->Length() > 0) { |
| event.set_top_frame(stack_trace->FrameAt(0)); |
| } |
| CacheStackTraces(stack_trace, CollectAsyncCausalStackTrace(), |
| CollectAwaiterReturnStackTrace()); |
| Pause(&event); |
| HandleSteppingRequest(stack_trace_); // we may get a rewind request |
| ClearCachedStackTraces(); |
| } |
| |
| // Returns the best fit token position for a breakpoint. |
| // |
| // Takes a range of tokens [requested_token_pos, last_token_pos] and |
| // an optional column (requested_column). The range of tokens usually |
| // represents one line of the program text, but can represent a larger |
| // range on recursive calls. |
| // |
| // The best fit is found in two passes. |
| // |
| // The first pass finds a candidate token which: |
| // |
| // - is a safepoint, |
| // - has the lowest column number compatible with the requested column |
| // if a column has been specified, |
| // and: |
| // - has the lowest token position number which satisfies the above. |
| // |
| // When we consider a column number, we look for the token which |
| // intersects the desired column. For example: |
| // |
| // 1 2 3 |
| // 12345678901234567890 0 |
| // |
| // var x = function(function(y)); |
| // ^ |
| // |
| // If we request a breakpoint at column 14, the lowest column number |
| // compatible with that would for column 11 (beginning of the |
| // 'function' token) in the example above. |
| // |
| // Once this candidate token from the first pass is found, we then |
| // have a second pass which considers only those tokens on the same |
| // line as the candidate token. |
| // |
| // The second pass finds a best fit token which: |
| // |
| // - is a safepoint, |
| // - has the same column number as the candidate token (perhaps |
| // more than one token has the same column number), |
| // and: |
| // - has the lowest code address in the generated code. |
| // |
| // We prefer the lowest compiled code address, because this tends to |
| // select the first subexpression on a line. For example in a line |
| // with nested function calls f(g(x)), the call to g() will have a |
| // lower compiled code address than the call to f(). |
| // |
| // If no best fit token can be found, the search is expanded, |
| // searching through the rest of the current function by calling this |
| // function recursively. |
| // |
| // TODO(turnidge): Given that we usually call this function with a |
| // token range restricted to a single line, this could be a one-pass |
| // algorithm, which would be simpler. I believe that it only needs |
| // two passes to support the recursive try-the-whole-function case. |
| // Rewrite this later, once there are more tests in place. |
| TokenPosition Debugger::ResolveBreakpointPos(const Function& func, |
| TokenPosition requested_token_pos, |
| TokenPosition last_token_pos, |
| intptr_t requested_column) { |
| ASSERT(func.HasCode()); |
| ASSERT(!func.HasOptimizedCode()); |
| |
| if (requested_token_pos < func.token_pos()) { |
| requested_token_pos = func.token_pos(); |
| } |
| if (last_token_pos > func.end_token_pos()) { |
| last_token_pos = func.end_token_pos(); |
| } |
| |
| Zone* zone = Thread::Current()->zone(); |
| Script& script = Script::Handle(zone, func.script()); |
| Code& code = Code::Handle(zone, func.unoptimized_code()); |
| ASSERT(!code.IsNull()); |
| PcDescriptors& desc = PcDescriptors::Handle(zone, code.pc_descriptors()); |
| |
| // First pass: find the safe point which is closest to the beginning |
| // of the given token range. |
| TokenPosition best_fit_pos = TokenPosition::kMaxSource; |
| intptr_t best_column = INT_MAX; |
| intptr_t best_line = INT_MAX; |
| PcDescriptors::Iterator iter(desc, kSafepointKind); |
| while (iter.MoveNext()) { |
| const TokenPosition pos = iter.TokenPos(); |
| if ((!pos.IsReal()) || (pos < requested_token_pos) || |
| (pos > last_token_pos)) { |
| // Token is not in the target range. |
| continue; |
| } |
| |
| intptr_t token_start_column = -1; |
| intptr_t token_line = -1; |
| if (requested_column >= 0) { |
| intptr_t token_len = -1; |
| // TODO(turnidge): GetTokenLocation is a very expensive |
| // operation, and this code will blow up when we are setting |
| // column breakpoints on, for example, a large, single-line |
| // program. Consider rewriting this code so that it only scans |
| // the program code once and caches the token positions and |
| // lengths. |
| script.GetTokenLocation(pos, &token_line, &token_start_column, |
| &token_len); |
| intptr_t token_end_column = token_start_column + token_len - 1; |
| if ((token_end_column < requested_column) || |
| (token_start_column > best_column)) { |
| // Prefer the token with the lowest column number compatible |
| // with the requested column. |
| continue; |
| } |
| } |
| |
| // Prefer the lowest (first) token pos. |
| if (pos < best_fit_pos) { |
| best_fit_pos = pos; |
| best_line = token_line; |
| best_column = token_start_column; |
| } |
| } |
| |
| // Second pass (if we found a safe point in the first pass). Find |
| // the token on the line which is at the best fit column (if column |
| // was specified) and has the lowest code address. |
| if (best_fit_pos != TokenPosition::kMaxSource) { |
| const Script& script = Script::Handle(zone, func.script()); |
| const TokenPosition begin_pos = best_fit_pos; |
| |
| TokenPosition end_of_line_pos; |
| if (script.kind() == RawScript::kKernelTag) { |
| if (best_line == -1) { |
| script.GetTokenLocation(begin_pos, &best_line, NULL); |
| } |
| ASSERT(best_line > 0); |
| TokenPosition ignored; |
| script.TokenRangeAtLine(best_line, &ignored, &end_of_line_pos); |
| if (end_of_line_pos < begin_pos) { |
| end_of_line_pos = begin_pos; |
| } |
| } else { |
| UNREACHABLE(); |
| end_of_line_pos = TokenPosition::kNoSource; |
| } |
| |
| uword lowest_pc_offset = kUwordMax; |
| PcDescriptors::Iterator iter(desc, kSafepointKind); |
| while (iter.MoveNext()) { |
| const TokenPosition pos = iter.TokenPos(); |
| if (!pos.IsReal() || (pos < begin_pos) || (pos > end_of_line_pos)) { |
| // Token is not on same line as best fit. |
| continue; |
| } |
| |
| if (requested_column >= 0) { |
| intptr_t ignored = -1; |
| intptr_t token_start_column = -1; |
| // We look for other tokens at the best column in case there |
| // is more than one token at the same column offset. |
| script.GetTokenLocation(pos, &ignored, &token_start_column); |
| if (token_start_column != best_column) { |
| continue; |
| } |
| } |
| |
| // Prefer the lowest pc offset. |
| if (iter.PcOffset() < lowest_pc_offset) { |
| lowest_pc_offset = iter.PcOffset(); |
| best_fit_pos = pos; |
| } |
| } |
| return best_fit_pos; |
| } |
| |
| // We didn't find a safe point in the given token range. Try and |
| // find a safe point in the remaining source code of the function. |
| // Since we have moved to the next line of the function, we no |
| // longer are requesting a specific column number. |
| if (last_token_pos < func.end_token_pos()) { |
| return ResolveBreakpointPos(func, last_token_pos, func.end_token_pos(), |
| -1 /* no column */); |
| } |
| return TokenPosition::kNoSource; |
| } |
| |
| void Debugger::MakeCodeBreakpointAt(const Function& func, |
| BreakpointLocation* loc) { |
| ASSERT(loc->token_pos_.IsReal()); |
| ASSERT((loc != NULL) && loc->IsResolved()); |
| ASSERT(!func.HasOptimizedCode()); |
| Code& code = Code::Handle(func.unoptimized_code()); |
| ASSERT(!code.IsNull()); |
| PcDescriptors& desc = PcDescriptors::Handle(code.pc_descriptors()); |
| uword lowest_pc_offset = kUwordMax; |
| RawPcDescriptors::Kind lowest_kind = RawPcDescriptors::kAnyKind; |
| // Find the safe point with the lowest compiled code address |
| // that maps to the token position of the source breakpoint. |
| PcDescriptors::Iterator iter(desc, kSafepointKind); |
| while (iter.MoveNext()) { |
| if (iter.TokenPos() == loc->token_pos_) { |
| if (iter.PcOffset() < lowest_pc_offset) { |
| lowest_pc_offset = iter.PcOffset(); |
| lowest_kind = iter.Kind(); |
| } |
| } |
| } |
| if (lowest_pc_offset == kUwordMax) { |
| return; |
| } |
| uword lowest_pc = code.PayloadStart() + lowest_pc_offset; |
| CodeBreakpoint* code_bpt = GetCodeBreakpoint(lowest_pc); |
| if (code_bpt == NULL) { |
| // No code breakpoint for this code exists; create one. |
| code_bpt = |
| new CodeBreakpoint(code, loc->token_pos_, lowest_pc, lowest_kind); |
| RegisterCodeBreakpoint(code_bpt); |
| } |
| code_bpt->set_bpt_location(loc); |
| if (loc->AnyEnabled()) { |
| code_bpt->Enable(); |
| } |
| } |
| |
| void Debugger::FindCompiledFunctions(const Script& script, |
| TokenPosition start_pos, |
| TokenPosition end_pos, |
| GrowableObjectArray* function_list) { |
| Zone* zone = Thread::Current()->zone(); |
| Class& cls = Class::Handle(zone); |
| Array& functions = Array::Handle(zone); |
| GrowableObjectArray& closures = GrowableObjectArray::Handle(zone); |
| Function& function = Function::Handle(zone); |
| |
| closures = isolate_->object_store()->closure_functions(); |
| const intptr_t num_closures = closures.Length(); |
| for (intptr_t pos = 0; pos < num_closures; pos++) { |
| function ^= closures.At(pos); |
| ASSERT(!function.IsNull()); |
| if ((function.token_pos() == start_pos) && |
| (function.end_token_pos() == end_pos) && |
| (function.script() == script.raw())) { |
| if (function.HasCode() && function.is_debuggable()) { |
| function_list->Add(function); |
| } |
| if (function.HasImplicitClosureFunction()) { |
| function = function.ImplicitClosureFunction(); |
| if (function.HasCode() && function.is_debuggable()) { |
| function_list->Add(function); |
| } |
| } |
| } |
| } |
| |
| const ClassTable& class_table = *isolate_->class_table(); |
| const intptr_t num_classes = class_table.NumCids(); |
| for (intptr_t i = 1; i < num_classes; i++) { |
| if (class_table.HasValidClassAt(i)) { |
| cls = class_table.At(i); |
| // If the class is not finalized, e.g. if it hasn't been parsed |
| // yet entirely, we can ignore it. If it contains a function with |
| // an unresolved breakpoint, we will detect it if and when the |
| // function gets compiled. |
| if (!cls.is_finalized()) { |
| continue; |
| } |
| // Note: we need to check the functions of this class even if |
| // the class is defined in a different 'script'. There could |
| // be mixin functions from the given script in this class. |
| functions = cls.functions(); |
| if (!functions.IsNull()) { |
| const intptr_t num_functions = functions.Length(); |
| for (intptr_t pos = 0; pos < num_functions; pos++) { |
| function ^= functions.At(pos); |
| ASSERT(!function.IsNull()); |
| // Check token position first to avoid unnecessary calls |
| // to script() which allocates handles. |
| if ((function.token_pos() == start_pos) && |
| (function.end_token_pos() == end_pos) && |
| (function.script() == script.raw())) { |
| if (function.HasCode() && function.is_debuggable()) { |
| function_list->Add(function); |
| } |
| if (function.HasImplicitClosureFunction()) { |
| function = function.ImplicitClosureFunction(); |
| if (function.HasCode() && function.is_debuggable()) { |
| function_list->Add(function); |
| } |
| } |
| } |
| } |
| } |
| } |
| } |
| } |
| |
| static void SelectBestFit(Function* best_fit, Function* func) { |
| if (best_fit->IsNull()) { |
| *best_fit = func->raw(); |
| } else { |
| if ((func->token_pos() > best_fit->token_pos()) && |
| ((func->end_token_pos() <= best_fit->end_token_pos()))) { |
| *best_fit = func->raw(); |
| } |
| } |
| } |
| |
| // Returns true if a best fit is found. A best fit can either be a function |
| // or a field. If it is a function, then the best fit function is returned |
| // in |best_fit|. If a best fit is a field, it means that a latent |
| // breakpoint can be set in the range |token_pos| to |last_token_pos|. |
| bool Debugger::FindBestFit(const Script& script, |
| TokenPosition token_pos, |
| TokenPosition last_token_pos, |
| Function* best_fit) { |
| Thread* thread = Thread::Current(); |
| Zone* zone = thread->zone(); |
| Class& cls = Class::Handle(zone); |
| Library& lib = Library::Handle(zone, script.FindLibrary()); |
| ASSERT(!lib.IsNull()); |
| const GrowableObjectArray& closures = GrowableObjectArray::Handle( |
| zone, isolate_->object_store()->closure_functions()); |
| Array& functions = Array::Handle(zone); |
| Function& function = Function::Handle(zone); |
| Array& fields = Array::Handle(zone); |
| Field& field = Field::Handle(zone); |
| Error& error = Error::Handle(zone); |
| |
| const intptr_t num_closures = closures.Length(); |
| for (intptr_t i = 0; i < num_closures; i++) { |
| function ^= closures.At(i); |
| if (FunctionOverlaps(function, script, token_pos, last_token_pos)) { |
| // Select the inner most closure. |
| SelectBestFit(best_fit, &function); |
| } |
| } |
| if (!best_fit->IsNull()) { |
| // The inner most closure found will be the best fit. Going |
| // over class functions below will not help in any further |
| // narrowing. |
| return true; |
| } |
| |
| const ClassTable& class_table = *isolate_->class_table(); |
| const intptr_t num_classes = class_table.NumCids(); |
| for (intptr_t i = 1; i < num_classes; i++) { |
| if (!class_table.HasValidClassAt(i)) { |
| continue; |
| } |
| cls = class_table.At(i); |
| // This class is relevant to us only if it belongs to the |
| // library to which |script| belongs. |
| if (cls.library() != lib.raw()) { |
| continue; |
| } |
| // Parse class definition if not done yet. |
| error = cls.EnsureIsFinalized(Thread::Current()); |
| if (!error.IsNull()) { |
| // Ignore functions in this class. |
| // TODO(hausner): Should we propagate this error? How? |
| // EnsureIsFinalized only returns an error object if there |
| // is no longjump base on the stack. |
| continue; |
| } |
| functions = cls.functions(); |
| if (!functions.IsNull()) { |
| const intptr_t num_functions = functions.Length(); |
| for (intptr_t pos = 0; pos < num_functions; pos++) { |
| function ^= functions.At(pos); |
| ASSERT(!function.IsNull()); |
| if (IsImplicitFunction(function)) { |
| // Implicit functions do not have a user specifiable source |
| // location. |
| continue; |
| } |
| if (FunctionOverlaps(function, script, token_pos, last_token_pos)) { |
| // Closures and inner functions within a class method are not |
| // present in the functions of a class. Hence, we can return |
| // right away as looking through other functions of a class |
| // will not narrow down to any inner function/closure. |
| *best_fit = function.raw(); |
| return true; |
| } |
| } |
| } |
| // If none of the functions in the class contain token_pos, then we |
| // check if it falls within a function literal initializer of a field |
| // that has not been initialized yet. If the field (and hence the |
| // function literal initializer) has already been initialized, then |
| // it would have been found above in the object store as a closure. |
| fields = |