1//===-- StackFrameList.cpp ------------------------------------------------===//
2//
3// Part of the LLVM Project, under the Apache License v2.0 with LLVM Exceptions.
4// See https://llvm.org/LICENSE.txt for license information.
5// SPDX-License-Identifier: Apache-2.0 WITH LLVM-exception
6//
7//===----------------------------------------------------------------------===//
8
9#include "lldb/Target/StackFrameList.h"
10#include "lldb/Breakpoint/Breakpoint.h"
11#include "lldb/Breakpoint/BreakpointLocation.h"
12#include "lldb/Core/Debugger.h"
13#include "lldb/Core/SourceManager.h"
14#include "lldb/Host/StreamFile.h"
15#include "lldb/Symbol/Block.h"
16#include "lldb/Symbol/Function.h"
17#include "lldb/Symbol/Symbol.h"
18#include "lldb/Target/Process.h"
19#include "lldb/Target/RegisterContext.h"
20#include "lldb/Target/StackFrame.h"
21#include "lldb/Target/StackFrameRecognizer.h"
22#include "lldb/Target/StopInfo.h"
23#include "lldb/Target/Target.h"
24#include "lldb/Target/Thread.h"
25#include "lldb/Target/Unwind.h"
26#include "lldb/Utility/LLDBLog.h"
27#include "lldb/Utility/Log.h"
28#include "llvm/ADT/SmallPtrSet.h"
29
30#include <memory>
31
32//#define DEBUG_STACK_FRAMES 1
33
34using namespace lldb;
35using namespace lldb_private;
36
37// StackFrameList constructor
38StackFrameList::StackFrameList(Thread &thread,
39 const lldb::StackFrameListSP &prev_frames_sp,
40 bool show_inline_frames)
41 : m_thread(thread), m_prev_frames_sp(prev_frames_sp), m_frames(),
42 m_selected_frame_idx(), m_concrete_frames_fetched(0),
43 m_current_inlined_depth(UINT32_MAX),
44 m_current_inlined_pc(LLDB_INVALID_ADDRESS),
45 m_show_inlined_frames(show_inline_frames) {
46 if (prev_frames_sp) {
47 m_current_inlined_depth = prev_frames_sp->m_current_inlined_depth;
48 m_current_inlined_pc = prev_frames_sp->m_current_inlined_pc;
49 }
50}
51
52StackFrameList::~StackFrameList() {
53 // Call clear since this takes a lock and clears the stack frame list in case
54 // another thread is currently using this stack frame list
55 Clear();
56}
57
58void StackFrameList::CalculateCurrentInlinedDepth() {
59 uint32_t cur_inlined_depth = GetCurrentInlinedDepth();
60 if (cur_inlined_depth == UINT32_MAX) {
61 ResetCurrentInlinedDepth();
62 }
63}
64
65uint32_t StackFrameList::GetCurrentInlinedDepth() {
66 std::lock_guard<std::mutex> guard(m_inlined_depth_mutex);
67 if (m_show_inlined_frames && m_current_inlined_pc != LLDB_INVALID_ADDRESS) {
68 lldb::addr_t cur_pc = m_thread.GetRegisterContext()->GetPC();
69 if (cur_pc != m_current_inlined_pc) {
70 m_current_inlined_pc = LLDB_INVALID_ADDRESS;
71 m_current_inlined_depth = UINT32_MAX;
72 Log *log = GetLog(mask: LLDBLog::Step);
73 if (log && log->GetVerbose())
74 LLDB_LOGF(
75 log,
76 "GetCurrentInlinedDepth: invalidating current inlined depth.\n");
77 }
78 return m_current_inlined_depth;
79 } else {
80 return UINT32_MAX;
81 }
82}
83
84void StackFrameList::ResetCurrentInlinedDepth() {
85 if (!m_show_inlined_frames)
86 return;
87
88 StopInfoSP stop_info_sp = m_thread.GetStopInfo();
89 if (!stop_info_sp)
90 return;
91
92 bool inlined = true;
93 auto inline_depth = stop_info_sp->GetSuggestedStackFrameIndex(inlined_stack: inlined);
94 // We're only adjusting the inlined stack here.
95 Log *log = GetLog(mask: LLDBLog::Step);
96 if (inline_depth) {
97 std::lock_guard<std::mutex> guard(m_inlined_depth_mutex);
98 m_current_inlined_depth = *inline_depth;
99 m_current_inlined_pc = m_thread.GetRegisterContext()->GetPC();
100
101 if (log && log->GetVerbose())
102 LLDB_LOGF(log,
103 "ResetCurrentInlinedDepth: setting inlined "
104 "depth: %d 0x%" PRIx64 ".\n",
105 m_current_inlined_depth, m_current_inlined_pc);
106 } else {
107 std::lock_guard<std::mutex> guard(m_inlined_depth_mutex);
108 m_current_inlined_pc = LLDB_INVALID_ADDRESS;
109 m_current_inlined_depth = UINT32_MAX;
110 if (log && log->GetVerbose())
111 LLDB_LOGF(
112 log,
113 "ResetCurrentInlinedDepth: Invalidating current inlined depth.\n");
114 }
115}
116
117bool StackFrameList::DecrementCurrentInlinedDepth() {
118 if (m_show_inlined_frames) {
119 uint32_t current_inlined_depth = GetCurrentInlinedDepth();
120 if (current_inlined_depth != UINT32_MAX) {
121 if (current_inlined_depth > 0) {
122 std::lock_guard<std::mutex> guard(m_inlined_depth_mutex);
123 m_current_inlined_depth--;
124 return true;
125 }
126 }
127 }
128 return false;
129}
130
131void StackFrameList::SetCurrentInlinedDepth(uint32_t new_depth) {
132 std::lock_guard<std::mutex> guard(m_inlined_depth_mutex);
133 m_current_inlined_depth = new_depth;
134 if (new_depth == UINT32_MAX)
135 m_current_inlined_pc = LLDB_INVALID_ADDRESS;
136 else
137 m_current_inlined_pc = m_thread.GetRegisterContext()->GetPC();
138}
139
140bool StackFrameList::WereAllFramesFetched() const {
141 std::shared_lock<std::shared_mutex> guard(m_list_mutex);
142 return GetAllFramesFetched();
143}
144
145/// A sequence of calls that comprise some portion of a backtrace. Each frame
146/// is represented as a pair of a callee (Function *) and an address within the
147/// callee.
148struct CallDescriptor {
149 Function *func;
150 CallEdge::AddrType address_type = CallEdge::AddrType::Call;
151 addr_t address = LLDB_INVALID_ADDRESS;
152};
153using CallSequence = std::vector<CallDescriptor>;
154
155/// Find the unique path through the call graph from \p begin (with return PC
156/// \p return_pc) to \p end. On success this path is stored into \p path, and
157/// on failure \p path is unchanged.
158/// This function doesn't currently access StackFrameLists at all, it only looks
159/// at the frame set in the ExecutionContext it passes around.
160static void FindInterveningFrames(Function &begin, Function &end,
161 ExecutionContext &exe_ctx, Target &target,
162 addr_t return_pc, CallSequence &path,
163 ModuleList &images, Log *log) {
164 LLDB_LOG(log, "Finding frames between {0} and {1}, retn-pc={2:x}",
165 begin.GetDisplayName(), end.GetDisplayName(), return_pc);
166
167 // Find a non-tail calling edge with the correct return PC.
168 if (log)
169 for (const auto &edge : begin.GetCallEdges())
170 LLDB_LOG(log, "FindInterveningFrames: found call with retn-PC = {0:x}",
171 edge->GetReturnPCAddress(begin, target));
172 CallEdge *first_edge = begin.GetCallEdgeForReturnAddress(return_pc, target);
173 if (!first_edge) {
174 LLDB_LOG(log, "No call edge outgoing from {0} with retn-PC == {1:x}",
175 begin.GetDisplayName(), return_pc);
176 return;
177 }
178
179 // The first callee may not be resolved, or there may be nothing to fill in.
180 Function *first_callee = first_edge->GetCallee(images, exe_ctx);
181 if (!first_callee) {
182 LLDB_LOG(log, "Could not resolve callee");
183 return;
184 }
185 if (first_callee == &end) {
186 LLDB_LOG(log, "Not searching further, first callee is {0} (retn-PC: {1:x})",
187 end.GetDisplayName(), return_pc);
188 return;
189 }
190
191 // Run DFS on the tail-calling edges out of the first callee to find \p end.
192 // Fully explore the set of functions reachable from the first edge via tail
193 // calls in order to detect ambiguous executions.
194 struct DFS {
195 CallSequence active_path = {};
196 CallSequence solution_path = {};
197 llvm::SmallPtrSet<Function *, 2> visited_nodes = {};
198 bool ambiguous = false;
199 Function *end;
200 ModuleList &images;
201 Target &target;
202 ExecutionContext &context;
203
204 DFS(Function *end, ModuleList &images, Target &target,
205 ExecutionContext &context)
206 : end(end), images(images), target(target), context(context) {}
207
208 void search(CallEdge &first_edge, Function &first_callee,
209 CallSequence &path) {
210 dfs(current_edge&: first_edge, callee&: first_callee);
211 if (!ambiguous)
212 path = std::move(solution_path);
213 }
214
215 void dfs(CallEdge &current_edge, Function &callee) {
216 // Found a path to the target function.
217 if (&callee == end) {
218 if (solution_path.empty())
219 solution_path = active_path;
220 else
221 ambiguous = true;
222 return;
223 }
224
225 // Terminate the search if tail recursion is found, or more generally if
226 // there's more than one way to reach a target. This errs on the side of
227 // caution: it conservatively stops searching when some solutions are
228 // still possible to save time in the average case.
229 if (!visited_nodes.insert(Ptr: &callee).second) {
230 ambiguous = true;
231 return;
232 }
233
234 // Search the calls made from this callee.
235 active_path.push_back(x: CallDescriptor{.func: &callee});
236 for (const auto &edge : callee.GetTailCallingEdges()) {
237 Function *next_callee = edge->GetCallee(images, exe_ctx&: context);
238 if (!next_callee)
239 continue;
240
241 std::tie(args&: active_path.back().address_type, args&: active_path.back().address) =
242 edge->GetCallerAddress(caller&: callee, target);
243
244 dfs(current_edge&: *edge, callee&: *next_callee);
245 if (ambiguous)
246 return;
247 }
248 active_path.pop_back();
249 }
250 };
251
252 DFS(&end, images, target, exe_ctx).search(first_edge&: *first_edge, first_callee&: *first_callee, path);
253}
254
255/// Given that \p next_frame will be appended to the frame list, synthesize
256/// tail call frames between the current end of the list and \p next_frame.
257/// If any frames are added, adjust the frame index of \p next_frame.
258///
259/// --------------
260/// | ... | <- Completed frames.
261/// --------------
262/// | prev_frame |
263/// --------------
264/// | ... | <- Artificial frames inserted here.
265/// --------------
266/// | next_frame |
267/// --------------
268/// | ... | <- Not-yet-visited frames.
269/// --------------
270void StackFrameList::SynthesizeTailCallFrames(StackFrame &next_frame) {
271 // Cannot synthesize tail call frames when the stack is empty (there is no
272 // "previous" frame).
273 if (m_frames.empty())
274 return;
275
276 TargetSP target_sp = next_frame.CalculateTarget();
277 if (!target_sp)
278 return;
279
280 lldb::RegisterContextSP next_reg_ctx_sp = next_frame.GetRegisterContext();
281 if (!next_reg_ctx_sp)
282 return;
283
284 Log *log = GetLog(mask: LLDBLog::Step);
285
286 StackFrame &prev_frame = *m_frames.back().get();
287
288 // Find the functions prev_frame and next_frame are stopped in. The function
289 // objects are needed to search the lazy call graph for intervening frames.
290 Function *prev_func =
291 prev_frame.GetSymbolContext(resolve_scope: eSymbolContextFunction).function;
292 if (!prev_func) {
293 LLDB_LOG(log, "SynthesizeTailCallFrames: can't find previous function");
294 return;
295 }
296 Function *next_func =
297 next_frame.GetSymbolContext(resolve_scope: eSymbolContextFunction).function;
298 if (!next_func) {
299 LLDB_LOG(log, "SynthesizeTailCallFrames: can't find next function");
300 return;
301 }
302
303 // Try to find the unique sequence of (tail) calls which led from next_frame
304 // to prev_frame.
305 CallSequence path;
306 addr_t return_pc = next_reg_ctx_sp->GetPC();
307 Target &target = *target_sp.get();
308 ModuleList &images = next_frame.CalculateTarget()->GetImages();
309 ExecutionContext exe_ctx(target_sp, /*get_process=*/true);
310 exe_ctx.SetFramePtr(&next_frame);
311 FindInterveningFrames(begin&: *next_func, end&: *prev_func, exe_ctx, target, return_pc,
312 path, images, log);
313
314 // Push synthetic tail call frames.
315 for (auto calleeInfo : llvm::reverse(C&: path)) {
316 Function *callee = calleeInfo.func;
317 uint32_t frame_idx = m_frames.size();
318 uint32_t concrete_frame_idx = next_frame.GetConcreteFrameIndex();
319 addr_t cfa = LLDB_INVALID_ADDRESS;
320 bool cfa_is_valid = false;
321 addr_t pc = calleeInfo.address;
322 // If the callee address refers to the call instruction, we do not want to
323 // subtract 1 from this value.
324 const bool behaves_like_zeroth_frame =
325 calleeInfo.address_type == CallEdge::AddrType::Call;
326 SymbolContext sc;
327 callee->CalculateSymbolContext(sc: &sc);
328 auto synth_frame = std::make_shared<StackFrame>(
329 args: m_thread.shared_from_this(), args&: frame_idx, args&: concrete_frame_idx, args&: cfa,
330 args&: cfa_is_valid, args&: pc, args: StackFrame::Kind::Artificial,
331 args: behaves_like_zeroth_frame, args: &sc);
332 m_frames.push_back(x: synth_frame);
333 LLDB_LOG(log, "Pushed frame {0} at {1:x}", callee->GetDisplayName(), pc);
334 }
335
336 // If any frames were created, adjust next_frame's index.
337 if (!path.empty())
338 next_frame.SetFrameIndex(m_frames.size());
339}
340
341bool StackFrameList::GetFramesUpTo(uint32_t end_idx,
342 InterruptionControl allow_interrupt) {
343 // GetFramesUpTo is always called with the intent to add frames, so get the
344 // writer lock:
345 std::unique_lock<std::shared_mutex> guard(m_list_mutex);
346 // Now that we have the lock, check to make sure someone didn't get there
347 // ahead of us:
348 if (m_frames.size() > end_idx || GetAllFramesFetched())
349 return false;
350
351 // Do not fetch frames for an invalid thread.
352 bool was_interrupted = false;
353 if (!m_thread.IsValid())
354 return false;
355
356 // lock the writer side of m_list_mutex as we're going to add frames here:
357 if (!m_show_inlined_frames) {
358 if (end_idx < m_concrete_frames_fetched)
359 return false;
360 // We're adding concrete frames now:
361 // FIXME: This should also be interruptible:
362 FetchOnlyConcreteFramesUpTo(end_idx);
363 return false;
364 }
365
366 // We're adding concrete and inlined frames now:
367 was_interrupted = FetchFramesUpTo(end_idx, allow_interrupt);
368
369#if defined(DEBUG_STACK_FRAMES)
370 s.PutCString("\n\nNew frames:\n");
371 Dump(&s);
372 s.EOL();
373#endif
374 return was_interrupted;
375}
376
377void StackFrameList::FetchOnlyConcreteFramesUpTo(uint32_t end_idx) {
378 assert(m_thread.IsValid() && "Expected valid thread");
379 assert(m_frames.size() <= end_idx && "Expected there to be frames to fill");
380
381 Unwind &unwinder = m_thread.GetUnwinder();
382
383 if (end_idx < m_concrete_frames_fetched)
384 return;
385
386 uint32_t num_frames = unwinder.GetFramesUpTo(end_idx);
387 if (num_frames <= end_idx + 1) {
388 // Done unwinding.
389 m_concrete_frames_fetched = UINT32_MAX;
390 }
391
392 // Don't create the frames eagerly. Defer this work to GetFrameAtIndex,
393 // which can lazily query the unwinder to create frames.
394 m_frames.resize(new_size: num_frames);
395}
396
397bool StackFrameList::FetchFramesUpTo(uint32_t end_idx,
398 InterruptionControl allow_interrupt) {
399 Unwind &unwinder = m_thread.GetUnwinder();
400 bool was_interrupted = false;
401
402#if defined(DEBUG_STACK_FRAMES)
403 StreamFile s(stdout, false);
404#endif
405 // If we are hiding some frames from the outside world, we need to add
406 // those onto the total count of frames to fetch. However, we don't need
407 // to do that if end_idx is 0 since in that case we always get the first
408 // concrete frame and all the inlined frames below it... And of course, if
409 // end_idx is UINT32_MAX that means get all, so just do that...
410
411 uint32_t inlined_depth = 0;
412 if (end_idx > 0 && end_idx != UINT32_MAX) {
413 inlined_depth = GetCurrentInlinedDepth();
414 if (inlined_depth != UINT32_MAX) {
415 if (end_idx > 0)
416 end_idx += inlined_depth;
417 }
418 }
419
420 StackFrameSP unwind_frame_sp;
421 Debugger &dbg = m_thread.GetProcess()->GetTarget().GetDebugger();
422 do {
423 uint32_t idx = m_concrete_frames_fetched++;
424 lldb::addr_t pc = LLDB_INVALID_ADDRESS;
425 lldb::addr_t cfa = LLDB_INVALID_ADDRESS;
426 bool behaves_like_zeroth_frame = (idx == 0);
427 if (idx == 0) {
428 // We might have already created frame zero, only create it if we need
429 // to.
430 if (m_frames.empty()) {
431 RegisterContextSP reg_ctx_sp(m_thread.GetRegisterContext());
432
433 if (reg_ctx_sp) {
434 const bool success = unwinder.GetFrameInfoAtIndex(
435 frame_idx: idx, cfa, pc, behaves_like_zeroth_frame);
436 // There shouldn't be any way not to get the frame info for frame
437 // 0. But if the unwinder can't make one, lets make one by hand
438 // with the SP as the CFA and see if that gets any further.
439 if (!success) {
440 cfa = reg_ctx_sp->GetSP();
441 pc = reg_ctx_sp->GetPC();
442 }
443
444 unwind_frame_sp = std::make_shared<StackFrame>(
445 args: m_thread.shared_from_this(), args: m_frames.size(), args&: idx, args&: reg_ctx_sp,
446 args&: cfa, args&: pc, args&: behaves_like_zeroth_frame, args: nullptr);
447 m_frames.push_back(x: unwind_frame_sp);
448 }
449 } else {
450 unwind_frame_sp = m_frames.front();
451 cfa = unwind_frame_sp->m_id.GetCallFrameAddress();
452 }
453 } else {
454 // Check for interruption when building the frames.
455 // Do the check in idx > 0 so that we'll always create a 0th frame.
456 if (allow_interrupt &&
457 INTERRUPT_REQUESTED(dbg, "Interrupted having fetched {0} frames",
458 m_frames.size())) {
459 was_interrupted = true;
460 break;
461 }
462
463 const bool success =
464 unwinder.GetFrameInfoAtIndex(frame_idx: idx, cfa, pc, behaves_like_zeroth_frame);
465 if (!success) {
466 // We've gotten to the end of the stack.
467 SetAllFramesFetched();
468 break;
469 }
470 const bool cfa_is_valid = true;
471 unwind_frame_sp = std::make_shared<StackFrame>(
472 args: m_thread.shared_from_this(), args: m_frames.size(), args&: idx, args&: cfa, args: cfa_is_valid,
473 args&: pc, args: StackFrame::Kind::Regular, args&: behaves_like_zeroth_frame, args: nullptr);
474
475 // Create synthetic tail call frames between the previous frame and the
476 // newly-found frame. The new frame's index may change after this call,
477 // although its concrete index will stay the same.
478 SynthesizeTailCallFrames(next_frame&: *unwind_frame_sp.get());
479
480 m_frames.push_back(x: unwind_frame_sp);
481 }
482
483 assert(unwind_frame_sp);
484 SymbolContext unwind_sc = unwind_frame_sp->GetSymbolContext(
485 resolve_scope: eSymbolContextBlock | eSymbolContextFunction);
486 Block *unwind_block = unwind_sc.block;
487 TargetSP target_sp = m_thread.CalculateTarget();
488 if (unwind_block) {
489 Address curr_frame_address(
490 unwind_frame_sp->GetFrameCodeAddressForSymbolication());
491
492 SymbolContext next_frame_sc;
493 Address next_frame_address;
494
495 while (unwind_sc.GetParentOfInlinedScope(
496 curr_frame_pc: curr_frame_address, next_frame_sc, inlined_frame_addr&: next_frame_address)) {
497 next_frame_sc.line_entry.ApplyFileMappings(target_sp);
498 behaves_like_zeroth_frame = false;
499 StackFrameSP frame_sp(new StackFrame(
500 m_thread.shared_from_this(), m_frames.size(), idx,
501 unwind_frame_sp->GetRegisterContextSP(), cfa, next_frame_address,
502 behaves_like_zeroth_frame, &next_frame_sc));
503
504 m_frames.push_back(x: frame_sp);
505 unwind_sc = next_frame_sc;
506 curr_frame_address = next_frame_address;
507 }
508 }
509 } while (m_frames.size() - 1 < end_idx);
510
511 // Don't try to merge till you've calculated all the frames in this stack.
512 if (GetAllFramesFetched() && m_prev_frames_sp) {
513 StackFrameList *prev_frames = m_prev_frames_sp.get();
514 StackFrameList *curr_frames = this;
515
516#if defined(DEBUG_STACK_FRAMES)
517 s.PutCString("\nprev_frames:\n");
518 prev_frames->Dump(&s);
519 s.PutCString("\ncurr_frames:\n");
520 curr_frames->Dump(&s);
521 s.EOL();
522#endif
523 size_t curr_frame_num, prev_frame_num;
524
525 for (curr_frame_num = curr_frames->m_frames.size(),
526 prev_frame_num = prev_frames->m_frames.size();
527 curr_frame_num > 0 && prev_frame_num > 0;
528 --curr_frame_num, --prev_frame_num) {
529 const size_t curr_frame_idx = curr_frame_num - 1;
530 const size_t prev_frame_idx = prev_frame_num - 1;
531 StackFrameSP curr_frame_sp(curr_frames->m_frames[curr_frame_idx]);
532 StackFrameSP prev_frame_sp(prev_frames->m_frames[prev_frame_idx]);
533
534#if defined(DEBUG_STACK_FRAMES)
535 s.Printf("\n\nCurr frame #%u ", curr_frame_idx);
536 if (curr_frame_sp)
537 curr_frame_sp->Dump(&s, true, false);
538 else
539 s.PutCString("NULL");
540 s.Printf("\nPrev frame #%u ", prev_frame_idx);
541 if (prev_frame_sp)
542 prev_frame_sp->Dump(&s, true, false);
543 else
544 s.PutCString("NULL");
545#endif
546
547 StackFrame *curr_frame = curr_frame_sp.get();
548 StackFrame *prev_frame = prev_frame_sp.get();
549
550 if (curr_frame == nullptr || prev_frame == nullptr)
551 break;
552
553 // Check the stack ID to make sure they are equal.
554 if (curr_frame->GetStackID() != prev_frame->GetStackID())
555 break;
556
557 prev_frame->UpdatePreviousFrameFromCurrentFrame(curr_frame&: *curr_frame);
558 // Now copy the fixed up previous frame into the current frames so the
559 // pointer doesn't change.
560 m_frames[curr_frame_idx] = prev_frame_sp;
561
562#if defined(DEBUG_STACK_FRAMES)
563 s.Printf("\n Copying previous frame to current frame");
564#endif
565 }
566 // We are done with the old stack frame list, we can release it now.
567 m_prev_frames_sp.reset();
568 }
569 // Don't report interrupted if we happen to have gotten all the frames:
570 if (!GetAllFramesFetched())
571 return was_interrupted;
572 return false;
573}
574
575uint32_t StackFrameList::GetNumFrames(bool can_create) {
576 if (!WereAllFramesFetched() && can_create) {
577 // Don't allow interrupt or we might not return the correct count
578 GetFramesUpTo(UINT32_MAX, allow_interrupt: DoNotAllowInterruption);
579 }
580 uint32_t frame_idx;
581 {
582 std::shared_lock<std::shared_mutex> guard(m_list_mutex);
583 frame_idx = GetVisibleStackFrameIndex(idx: m_frames.size());
584 }
585 return frame_idx;
586}
587
588void StackFrameList::Dump(Stream *s) {
589 if (s == nullptr)
590 return;
591
592 std::shared_lock<std::shared_mutex> guard(m_list_mutex);
593
594 const_iterator pos, begin = m_frames.begin(), end = m_frames.end();
595 for (pos = begin; pos != end; ++pos) {
596 StackFrame *frame = (*pos).get();
597 s->Printf(format: "%p: ", static_cast<void *>(frame));
598 if (frame) {
599 frame->GetStackID().Dump(s);
600 frame->DumpUsingSettingsFormat(strm: s);
601 } else
602 s->Printf(format: "frame #%u", (uint32_t)std::distance(first: begin, last: pos));
603 s->EOL();
604 }
605 s->EOL();
606}
607
608StackFrameSP StackFrameList::GetFrameAtIndex(uint32_t idx) {
609 StackFrameSP frame_sp;
610 uint32_t original_idx = idx;
611
612 // We're going to consult the m_frames.size, but if there are already
613 // enough frames for our request we don't want to block other readers, so
614 // first acquire the shared lock:
615 { // Scope for shared lock:
616 std::shared_lock<std::shared_mutex> guard(m_list_mutex);
617
618 uint32_t inlined_depth = GetCurrentInlinedDepth();
619 if (inlined_depth != UINT32_MAX)
620 idx += inlined_depth;
621
622 if (idx < m_frames.size())
623 frame_sp = m_frames[idx];
624
625 if (frame_sp)
626 return frame_sp;
627 } // End of reader lock scope
628
629 // GetFramesUpTo will fill m_frames with as many frames as you asked for, if
630 // there are that many. If there weren't then you asked for too many frames.
631 // GetFramesUpTo returns true if interrupted:
632 if (GetFramesUpTo(end_idx: idx, allow_interrupt: AllowInterruption)) {
633 Log *log = GetLog(mask: LLDBLog::Thread);
634 LLDB_LOG(log, "GetFrameAtIndex was interrupted");
635 return {};
636 }
637
638 { // Now we're accessing m_frames as a reader, so acquire the reader lock.
639 std::shared_lock<std::shared_mutex> guard(m_list_mutex);
640 if (idx < m_frames.size()) {
641 frame_sp = m_frames[idx];
642 } else if (original_idx == 0) {
643 // There should ALWAYS be a frame at index 0. If something went wrong
644 // with the CurrentInlinedDepth such that there weren't as many frames as
645 // we thought taking that into account, then reset the current inlined
646 // depth and return the real zeroth frame.
647 if (m_frames.empty()) {
648 // Why do we have a thread with zero frames, that should not ever
649 // happen...
650 assert(!m_thread.IsValid() && "A valid thread has no frames.");
651 } else {
652 ResetCurrentInlinedDepth();
653 frame_sp = m_frames[original_idx];
654 }
655 }
656 } // End of reader lock scope
657
658 return frame_sp;
659}
660
661StackFrameSP
662StackFrameList::GetFrameWithConcreteFrameIndex(uint32_t unwind_idx) {
663 // First try assuming the unwind index is the same as the frame index. The
664 // unwind index is always greater than or equal to the frame index, so it is
665 // a good place to start. If we have inlined frames we might have 5 concrete
666 // frames (frame unwind indexes go from 0-4), but we might have 15 frames
667 // after we make all the inlined frames. Most of the time the unwind frame
668 // index (or the concrete frame index) is the same as the frame index.
669 uint32_t frame_idx = unwind_idx;
670 StackFrameSP frame_sp(GetFrameAtIndex(idx: frame_idx));
671 while (frame_sp) {
672 if (frame_sp->GetFrameIndex() == unwind_idx)
673 break;
674 frame_sp = GetFrameAtIndex(idx: ++frame_idx);
675 }
676 return frame_sp;
677}
678
679static bool CompareStackID(const StackFrameSP &stack_sp,
680 const StackID &stack_id) {
681 return stack_sp->GetStackID() < stack_id;
682}
683
684StackFrameSP StackFrameList::GetFrameWithStackID(const StackID &stack_id) {
685 StackFrameSP frame_sp;
686
687 if (stack_id.IsValid()) {
688 uint32_t frame_idx = 0;
689 {
690 // First see if the frame is already realized. This is the scope for
691 // the shared mutex:
692 std::shared_lock<std::shared_mutex> guard(m_list_mutex);
693 // Do a binary search in case the stack frame is already in our cache
694 collection::const_iterator pos =
695 llvm::lower_bound(Range&: m_frames, Value: stack_id, C: CompareStackID);
696 if (pos != m_frames.end() && (*pos)->GetStackID() == stack_id)
697 return *pos;
698 }
699 // If we needed to add more frames, we would get to here.
700 do {
701 frame_sp = GetFrameAtIndex(idx: frame_idx);
702 if (frame_sp && frame_sp->GetStackID() == stack_id)
703 break;
704 frame_idx++;
705 } while (frame_sp);
706 }
707 return frame_sp;
708}
709
710bool StackFrameList::SetFrameAtIndex(uint32_t idx, StackFrameSP &frame_sp) {
711 std::unique_lock<std::shared_mutex> guard(m_list_mutex);
712 if (idx >= m_frames.size())
713 m_frames.resize(new_size: idx + 1);
714 // Make sure allocation succeeded by checking bounds again
715 if (idx < m_frames.size()) {
716 m_frames[idx] = frame_sp;
717 return true;
718 }
719 return false; // resize failed, out of memory?
720}
721
722void StackFrameList::SelectMostRelevantFrame() {
723 // Don't call into the frame recognizers on the private state thread as
724 // they can cause code to run in the target, and that can cause deadlocks
725 // when fetching stop events for the expression.
726 if (m_thread.GetProcess()->CurrentThreadIsPrivateStateThread())
727 return;
728
729 Log *log = GetLog(mask: LLDBLog::Thread);
730
731 // Only the top frame should be recognized.
732 StackFrameSP frame_sp = GetFrameAtIndex(idx: 0);
733 if (!frame_sp) {
734 LLDB_LOG(log, "Failed to construct Frame #0");
735 return;
736 }
737
738 RecognizedStackFrameSP recognized_frame_sp = frame_sp->GetRecognizedFrame();
739
740 if (recognized_frame_sp) {
741 if (StackFrameSP most_relevant_frame_sp =
742 recognized_frame_sp->GetMostRelevantFrame()) {
743 LLDB_LOG(log, "Found most relevant frame at index {0}",
744 most_relevant_frame_sp->GetFrameIndex());
745 SetSelectedFrame(most_relevant_frame_sp.get());
746 return;
747 }
748 }
749 LLDB_LOG(log, "Frame #0 not recognized");
750
751 // If this thread has a non-trivial StopInfo, then let it suggest
752 // a most relevant frame:
753 StopInfoSP stop_info_sp = m_thread.GetStopInfo();
754 uint32_t stack_idx = 0;
755 bool found_relevant = false;
756 if (stop_info_sp) {
757 // Here we're only asking the stop info if it wants to adjust the real stack
758 // index. We have to ask about the m_inlined_stack_depth in
759 // Thread::ShouldStop since the plans need to reason with that info.
760 bool inlined = false;
761 std::optional<uint32_t> stack_opt =
762 stop_info_sp->GetSuggestedStackFrameIndex(inlined_stack: inlined);
763 if (stack_opt) {
764 stack_idx = *stack_opt;
765 found_relevant = true;
766 }
767 }
768
769 frame_sp = GetFrameAtIndex(idx: stack_idx);
770 if (!frame_sp)
771 LLDB_LOG(log, "Stop info suggested relevant frame {0} but it didn't exist",
772 stack_idx);
773 else if (found_relevant)
774 LLDB_LOG(log, "Setting selected frame from stop info to {0}", stack_idx);
775 // Note, we don't have to worry about "inlined" frames here, because we've
776 // already calculated the inlined frame in Thread::ShouldStop, and
777 // SetSelectedFrame will take care of that adjustment for us.
778 SetSelectedFrame(frame_sp.get());
779
780 if (!found_relevant)
781 LLDB_LOG(log, "No relevant frame!");
782}
783
784uint32_t
785StackFrameList::GetSelectedFrameIndex(SelectMostRelevant select_most_relevant) {
786 if (!m_selected_frame_idx && select_most_relevant)
787 SelectMostRelevantFrame();
788 if (!m_selected_frame_idx) {
789 // If we aren't selecting the most relevant frame, and the selected frame
790 // isn't set, then don't force a selection here, just return 0.
791 if (!select_most_relevant)
792 return 0;
793 // If the inlined stack frame is set, then use that:
794 m_selected_frame_idx = 0;
795 }
796 return *m_selected_frame_idx;
797}
798
799uint32_t StackFrameList::SetSelectedFrame(lldb_private::StackFrame *frame) {
800 std::shared_lock<std::shared_mutex> guard(m_list_mutex);
801
802 const_iterator pos;
803 const_iterator begin = m_frames.begin();
804 const_iterator end = m_frames.end();
805 m_selected_frame_idx = 0;
806
807 for (pos = begin; pos != end; ++pos) {
808 if (pos->get() == frame) {
809 m_selected_frame_idx = std::distance(first: begin, last: pos);
810 uint32_t inlined_depth = GetCurrentInlinedDepth();
811 if (inlined_depth != UINT32_MAX)
812 m_selected_frame_idx = *m_selected_frame_idx - inlined_depth;
813 break;
814 }
815 }
816 SetDefaultFileAndLineToSelectedFrame();
817 return *m_selected_frame_idx;
818}
819
820bool StackFrameList::SetSelectedFrameByIndex(uint32_t idx) {
821 StackFrameSP frame_sp(GetFrameAtIndex(idx));
822 if (frame_sp) {
823 SetSelectedFrame(frame_sp.get());
824 return true;
825 } else
826 return false;
827}
828
829void StackFrameList::SetDefaultFileAndLineToSelectedFrame() {
830 if (m_thread.GetID() ==
831 m_thread.GetProcess()->GetThreadList().GetSelectedThread()->GetID()) {
832 StackFrameSP frame_sp(
833 GetFrameAtIndex(idx: GetSelectedFrameIndex(select_most_relevant: DoNoSelectMostRelevantFrame)));
834 if (frame_sp) {
835 SymbolContext sc = frame_sp->GetSymbolContext(resolve_scope: eSymbolContextLineEntry);
836 if (sc.line_entry.GetFile())
837 m_thread.CalculateTarget()->GetSourceManager().SetDefaultFileAndLine(
838 support_file_sp: sc.line_entry.file_sp, line: sc.line_entry.line);
839 }
840 }
841}
842
843// The thread has been run, reset the number stack frames to zero so we can
844// determine how many frames we have lazily.
845// Note, we don't actually re-use StackFrameLists, we always make a new
846// StackFrameList every time we stop, and then copy frame information frame
847// by frame from the old to the new StackFrameList. So the comment above,
848// does not describe how StackFrameLists are currently used.
849// Clear is currently only used to clear the list in the destructor.
850void StackFrameList::Clear() {
851 std::unique_lock<std::shared_mutex> guard(m_list_mutex);
852 m_frames.clear();
853 m_concrete_frames_fetched = 0;
854 m_selected_frame_idx.reset();
855}
856
857lldb::StackFrameSP
858StackFrameList::GetStackFrameSPForStackFramePtr(StackFrame *stack_frame_ptr) {
859 std::shared_lock<std::shared_mutex> guard(m_list_mutex);
860 const_iterator pos;
861 const_iterator begin = m_frames.begin();
862 const_iterator end = m_frames.end();
863 lldb::StackFrameSP ret_sp;
864
865 for (pos = begin; pos != end; ++pos) {
866 if (pos->get() == stack_frame_ptr) {
867 ret_sp = (*pos);
868 break;
869 }
870 }
871 return ret_sp;
872}
873
874size_t StackFrameList::GetStatus(Stream &strm, uint32_t first_frame,
875 uint32_t num_frames, bool show_frame_info,
876 uint32_t num_frames_with_source,
877 bool show_unique, bool show_hidden,
878 const char *selected_frame_marker) {
879 size_t num_frames_displayed = 0;
880
881 if (num_frames == 0)
882 return 0;
883
884 StackFrameSP frame_sp;
885 uint32_t frame_idx = 0;
886 uint32_t last_frame;
887
888 // Don't let the last frame wrap around...
889 if (num_frames == UINT32_MAX)
890 last_frame = UINT32_MAX;
891 else
892 last_frame = first_frame + num_frames;
893
894 StackFrameSP selected_frame_sp =
895 m_thread.GetSelectedFrame(select_most_relevant: DoNoSelectMostRelevantFrame);
896 const char *unselected_marker = nullptr;
897 std::string buffer;
898 if (selected_frame_marker) {
899 size_t len = strlen(s: selected_frame_marker);
900 buffer.insert(p: buffer.begin(), n: len, c: ' ');
901 unselected_marker = buffer.c_str();
902 }
903 const char *marker = nullptr;
904 for (frame_idx = first_frame; frame_idx < last_frame; ++frame_idx) {
905 frame_sp = GetFrameAtIndex(idx: frame_idx);
906 if (!frame_sp)
907 break;
908
909 if (selected_frame_marker != nullptr) {
910 if (frame_sp == selected_frame_sp)
911 marker = selected_frame_marker;
912 else
913 marker = unselected_marker;
914 }
915
916 // Hide uninteresting frames unless it's the selected frame.
917 if (!show_hidden && frame_sp != selected_frame_sp && frame_sp->IsHidden())
918 continue;
919
920 // Check for interruption here. If we're fetching arguments, this loop
921 // can go slowly:
922 Debugger &dbg = m_thread.GetProcess()->GetTarget().GetDebugger();
923 if (INTERRUPT_REQUESTED(
924 dbg, "Interrupted dumping stack for thread {0:x} with {1} shown.",
925 m_thread.GetID(), num_frames_displayed))
926 break;
927
928
929 if (!frame_sp->GetStatus(strm, show_frame_info,
930 show_source: num_frames_with_source > (first_frame - frame_idx),
931 show_unique, frame_marker: marker))
932 break;
933 ++num_frames_displayed;
934 }
935
936 strm.IndentLess();
937 return num_frames_displayed;
938}
939

Provided by KDAB

Privacy Policy
Update your C++ knowledge – Modern C++11/14/17 Training
Find out more

source code of lldb/source/Target/StackFrameList.cpp