blob: a117ea089b589085e84790fb0b3697e5d81e3936 [file] [log] [blame]
// Copyright 2015 The Chromium Authors. All rights reserved.
// Use of this source code is governed by a BSD-style license that can be
// found in the LICENSE file.
#ifndef BASE_TRACE_EVENT_HEAP_PROFILER_ALLOCATION_CONTEXT_TRACKER_H_
#define BASE_TRACE_EVENT_HEAP_PROFILER_ALLOCATION_CONTEXT_TRACKER_H_
#include <vector>
#include "base/atomicops.h"
#include "base/base_export.h"
#include "base/macros.h"
#include "base/trace_event/heap_profiler_allocation_context.h"
namespace base {
namespace trace_event {
// AllocationContextTracker is a thread-local object. Its main purpose is to
// keep track of a pseudo stack of trace events. Chrome has been instrumented
// with lots of `TRACE_EVENT` macros. These trace events push their name to a
// thread-local stack when they go into scope, and pop when they go out of
// scope, if all of the following conditions have been met:
//
// * A trace is being recorded.
// * The category of the event is enabled in the trace config.
// * Heap profiling is enabled (with the `--enable-heap-profiling` flag).
//
// This means that allocations that occur before tracing is started will not
// have backtrace information in their context.
//
// AllocationContextTracker also keeps track of some thread state not related to
// trace events. See |AllocationContext|.
//
// A thread-local instance of the context tracker is initialized lazily when it
// is first accessed. This might be because a trace event pushed or popped, or
// because `GetContextSnapshot()` was called when an allocation occurred
class BASE_EXPORT AllocationContextTracker {
public:
enum class CaptureMode : int32_t {
DISABLED, // Don't capture anything
PSEUDO_STACK, // Backtrace has trace events
MIXED_STACK, // Backtrace has trace events + from
// HeapProfilerScopedStackFrame
NATIVE_STACK, // Backtrace has full native backtraces from stack unwinding
};
// Stack frame constructed from trace events in codebase.
struct BASE_EXPORT PseudoStackFrame {
const char* trace_event_category;
const char* trace_event_name;
bool operator==(const PseudoStackFrame& other) const {
return trace_event_category == other.trace_event_category &&
trace_event_name == other.trace_event_name;
}
};
// Globally sets capturing mode.
// TODO(primiano): How to guard against *_STACK -> DISABLED -> *_STACK?
static void SetCaptureMode(CaptureMode mode);
// Returns global capturing mode.
inline static CaptureMode capture_mode() {
// A little lag after heap profiling is enabled or disabled is fine, it is
// more important that the check is as cheap as possible when capturing is
// not enabled, so do not issue a memory barrier in the fast path.
if (subtle::NoBarrier_Load(&capture_mode_) ==
static_cast<int32_t>(CaptureMode::DISABLED))
return CaptureMode::DISABLED;
// In the slow path, an acquire load is required to pair with the release
// store in |SetCaptureMode|. This is to ensure that the TLS slot for
// the thread-local allocation context tracker has been initialized if
// |capture_mode| returns something other than DISABLED.
return static_cast<CaptureMode>(subtle::Acquire_Load(&capture_mode_));
}
// Returns the thread-local instance, creating one if necessary. Returns
// always a valid instance, unless it is called re-entrantly, in which case
// returns nullptr in the nested calls.
static AllocationContextTracker* GetInstanceForCurrentThread();
// Set the thread name in the AllocationContextTracker of the current thread
// if capture is enabled.
static void SetCurrentThreadName(const char* name);
// Starts and ends a new ignore scope between which the allocations are
// ignored by the heap profiler. GetContextSnapshot() returns false when
// allocations are ignored.
void begin_ignore_scope() { ignore_scope_depth_++; }
void end_ignore_scope() {
if (ignore_scope_depth_)
ignore_scope_depth_--;
}
// Pushes and pops a frame onto the thread-local pseudo stack.
// TODO(ssid): Change PseudoStackFrame to const char*. Only event name is
// used.
void PushPseudoStackFrame(PseudoStackFrame stack_frame);
void PopPseudoStackFrame(PseudoStackFrame stack_frame);
// Pushes and pops a native stack frame onto thread local tracked stack.
void PushNativeStackFrame(const void* pc);
void PopNativeStackFrame(const void* pc);
// Push and pop current task's context. A stack is used to support nested
// tasks and the top of the stack will be used in allocation context.
void PushCurrentTaskContext(const char* context);
void PopCurrentTaskContext(const char* context);
// Returns most recent task context added by ScopedTaskExecutionTracker.
const char* TaskContext() const {
return task_contexts_.empty() ? nullptr : task_contexts_.back();
}
// Fills a snapshot of the current thread-local context. Doesn't fill and
// returns false if allocations are being ignored.
bool GetContextSnapshot(AllocationContext* snapshot);
~AllocationContextTracker();
private:
AllocationContextTracker();
static subtle::Atomic32 capture_mode_;
// The pseudo stack where frames are |TRACE_EVENT| names or inserted PCs.
std::vector<StackFrame> tracked_stack_;
// The thread name is used as the first entry in the pseudo stack.
const char* thread_name_;
// Stack of tasks' contexts. Context serves as a different dimension than
// pseudo stack to cluster allocations.
std::vector<const char*> task_contexts_;
uint32_t ignore_scope_depth_;
DISALLOW_COPY_AND_ASSIGN(AllocationContextTracker);
};
} // namespace trace_event
} // namespace base
#endif // BASE_TRACE_EVENT_HEAP_PROFILER_ALLOCATION_CONTEXT_TRACKER_H_