1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 25 26 27 28 29 30 31 32 33 34 35 36 37 38 39 40 41 42 43 44 45 46 47 48 49 50 51 52 53 54 55 56 57 58 59 60 61 62 63 64 65 66 67 68 69 70 71 72 73 74 75 76 77 78 79 80 81 82 83 84 85 86 87 88 89 90 91 92 93 94 95 96 97 98 99 100 101 102 103 104 105 106 107 108 109 110 111 112 113 114 115 116 117 118 119 120 121 122 123 124 125 126 127 128 129 130 131 132 133 134 135 136 137 138 139 140 141 142 143 144 145 146 147 148 149 150 151 152 153 154 155 156 157 158 159 160 161 162 163 164 165 166 167 168 169 170 171 172 173 174 175 176 177 178 179 180 181 182 183 184 185 186 187 188 189 190 191 192 193 194 195 196 197 198 199 200 201 202 203 204 205 206 207 208 209 210 211 212 213 214 215 216 217 218 219 220 221 222 223 224 225 226 227 228 229 230 231 232 233 234 235 236 237 238 239 240 241 242 243 244 245 246 247 248 249 250 251 252 253 254 255 256 257 258 259 260 261 262 263 264 265 266 267 268 269 270 271 272 273 274 275 276 277 278 279 280 281 282 283 284 285 286 287 288 289 290 291 292 293 294 295 296 297 298 299 300 301 302 303 304 305
|
// Copyright 2023 The Chromium Authors
// Use of this source code is governed by a BSD-style license that can be
// found in the LICENSE file.
#ifndef BASE_DEBUG_ALLOCATION_TRACE_H_
#define BASE_DEBUG_ALLOCATION_TRACE_H_
#include <algorithm>
#include <array>
#include <atomic>
#include <bit>
#include <cstdint>
#include "base/allocator/dispatcher/notification_data.h"
#include "base/base_export.h"
#include "base/compiler_specific.h"
#include "base/debug/debugging_buildflags.h"
#include "base/debug/stack_trace.h"
#include "base/memory/raw_ptr_exclusion.h"
#include "build/build_config.h"
namespace base::debug::tracer {
// Number of traces that can be stored. This number must be a power of two to
// allow for fast computation of modulo.
constexpr size_t kMaximumNumberOfMemoryOperationTraces = (1 << 15);
// Number of frames stored for each operation. Probably the lower frames
// represent the memory allocation system. Hence, we store more frames to
// increase chances of having a meaningful trace of the path that caused the
// allocation or free.
constexpr size_t kStackTraceSize = 16;
// The type of an operation stored in the recorder.
enum class OperationType {
// The state of an operation record before calling any of the initialization
// functions.
kNone = 0,
// The record represents an allocation operation.
kAllocation,
// The record represents a free operation.
kFree,
};
using StackTraceContainer = std::array<const void*, kStackTraceSize>;
// The record for a single operation. A record can represent any type of
// operation, allocation or free, but not at the same time.
//
// A record protects itself from concurrent initializations. If a thread B calls
// any of the Initialize*-functions while another thread A is currently
// initializing, B's invocations shall immediately return |false| without
// interfering with thread A.
class BASE_EXPORT OperationRecord {
public:
constexpr OperationRecord() = default;
OperationRecord(const OperationRecord&) = delete;
OperationRecord& operator=(const OperationRecord&) = delete;
// Is the record currently being taken?
bool IsRecording() const;
OperationType GetOperationType() const;
// The address allocated or freed.
const void* GetAddress() const;
// Number of allocated bytes. Returns 0 for free operations.
size_t GetSize() const;
// The stacktrace as taken by the Initialize*-functions.
const StackTraceContainer& GetStackTrace() const LIFETIME_BOUND;
// Initialize the record with data for another operation. Data from any
// previous operation will be silently overwritten. These functions are
// declared ALWAYS_INLINE to minimize pollution of the recorded stack trace.
//
// Both functions return false in case no record was taken, i.e. if another
// thread is capturing.
ALWAYS_INLINE bool InitializeFree(const void* freed_address) {
return InitializeOperationRecord(freed_address, 0, OperationType::kFree);
}
ALWAYS_INLINE bool InitializeAllocation(const void* allocated_address,
size_t allocated_size) {
return InitializeOperationRecord(allocated_address, allocated_size,
OperationType::kAllocation);
}
private:
// Initialize a record with the given data. Return true if the record was
// initialized successfully, false if no record was taken, i.e. if another
// thread is capturing.
ALWAYS_INLINE bool InitializeOperationRecord(const void* address,
size_t size,
OperationType operation_type);
ALWAYS_INLINE void StoreStackTrace();
// The stack trace taken in one of the Initialize* functions.
StackTraceContainer stack_trace_ = {};
// The number of allocated bytes.
size_t size_ = 0;
// The address that was allocated or freed.
// We use a raw C++ pointer instead of base::raw_ptr for performance
// reasons.
// - In the recorder we only store pointers, we never allocate or free on
// our own.
// - Storing is the hot path. base::raw_ptr::operator== may perform sanity
// checks which do not make sense in our case (otherwise the allocated
// address would have been quirky)
RAW_PTR_EXCLUSION const void* address_ = nullptr;
// The type of the operation that was performed. In the course of making a
// record, this value is reset to |OperationType::kNone| and later set to
// the operation type specific value, so if the process crashes whilst writing
// the record, it's marked as empty. To prevent the compiler from optimizing
// away the initial reset, this value is marked as volatile.
volatile OperationType operation_type_ = OperationType::kNone;
// Is the record currently being taken from another thread? Used to prevent
// concurrent writes to the same record.
//
// The value is mutable since pre C++20 there is no const getter in
// atomic_flag. All ways to get the value involve setting it.
// TODO(crbug.com/42050406): Remove mutable and make IsRecording() use
// atomic_flag::test();
mutable std::atomic_flag is_recording_ = ATOMIC_FLAG_INIT;
};
ALWAYS_INLINE bool OperationRecord::InitializeOperationRecord(
const void* address,
size_t size,
OperationType operation_type) {
if (is_recording_.test_and_set(std::memory_order_acquire)) {
return false;
}
operation_type_ = operation_type;
StoreStackTrace();
address_ = address;
size_ = size;
is_recording_.clear(std::memory_order_release);
return true;
}
ALWAYS_INLINE void OperationRecord::StoreStackTrace() {
stack_trace_.fill(nullptr);
#if BUILDFLAG(CAN_UNWIND_WITH_FRAME_POINTERS)
// Currently we limit ourselves to use TraceStackFramePointers. We know that
// TraceStackFramePointers has an acceptable performance impact on Android.
base::debug::TraceStackFramePointers(stack_trace_, 0);
#elif BUILDFLAG(IS_LINUX)
// Use base::debug::CollectStackTrace as an alternative for tests on Linux. We
// still have a check in /base/debug/debug.gni to prevent that
// AllocationStackTraceRecorder is enabled accidentally on Linux.
base::debug::CollectStackTrace(stack_trace_);
#else
#error "No supported stack tracer found."
#endif
}
struct BASE_EXPORT AllocationTraceRecorderStatistics {
#if BUILDFLAG(ENABLE_ALLOCATION_TRACE_RECORDER_FULL_REPORTING)
AllocationTraceRecorderStatistics(size_t total_number_of_allocations,
size_t total_number_of_collisions);
#else
AllocationTraceRecorderStatistics(size_t total_number_of_allocations);
#endif
// The total number of allocations that have been recorded.
size_t total_number_of_allocations;
#if BUILDFLAG(ENABLE_ALLOCATION_TRACE_RECORDER_FULL_REPORTING)
// The total number of collisions that have been encountered. A collision
// happens when two threads concurrently try to record using the same slot.
size_t total_number_of_collisions;
#endif
};
// The recorder which holds entries for past memory operations.
//
// The memory image of the recorder will be copied into the crash-handler.
// Therefore, it must not hold any references to external data which are vital
// for proper functioning.
//
// It is important that the recorder itself does not allocate to prevent
// recursive calls and save as much runtime overhead as possible.
//
// Therefore, records are stored in a preallocated buffer with a compile time
// constant maximum size, see |kMaximumNumberOfMemoryOperationTraces|. Once all
// records have been used, old records will be overwritten (fifo-style).
//
// The recorder works in an multithreaded environment without external locking.
// Concurrent writes are prevented by two means:
// 1 - We atomically increment and calculate the effective index of the record
// to be written.
// 2 - If this entry is still being used (the recording thread didn't finish
// yet), we go back to step 1
// Currently we do not enforce separate cache lines for each entry, which means
// false sharing can occur. On the other hand, with 64 byte cachelines a clean
// separation would introduce some 3*64 - sizeof(OperationRecord) = 40 bytes of
// padding per entry.
//
// Note: As a process might be terminated for whatever reason while stack
// traces are being written, the recorded data may contain some garbage.
//
// TODO(crbug.com/40258550): Evaluate the impact of the shared cache
// lines between entries.
class BASE_EXPORT AllocationTraceRecorder {
public:
constexpr AllocationTraceRecorder() = default;
AllocationTraceRecorder(const AllocationTraceRecorder&) = delete;
AllocationTraceRecorder& operator=(const AllocationTraceRecorder&) = delete;
// The allocation event observer interface. See the dispatcher for further
// details. The functions are marked NO_INLINE. All other functions called but
// the one taking the call stack are marked ALWAYS_INLINE. This way we ensure
// the number of frames recorded from these functions is fixed.
inline void OnAllocation(
const base::allocator::dispatcher::AllocationNotificationData&
allocation_data);
// Handle all free events.
inline void OnFree(
const base::allocator::dispatcher::FreeNotificationData& free_data);
// Access functions to retrieve the current content of the recorder.
// Note: Since the recorder is usually updated upon each allocation or free,
// it is important to prevent updates if you want to read the entries at any
// point.
// Get the current number of entries stored in the recorder. When the
// recorder has reached its maximum capacity, it always returns
// |GetMaximumNumberOfTraces()|.
size_t size() const;
// Access the record of an operation by index. Oldest operation is always
// accessible at index 0, latest operation at |size()-1|.
// Note: Since a process might have crashed while a trace is being written,
// especially the last records might be corrupted.
const OperationRecord& operator[](size_t idx) const;
constexpr size_t GetMaximumNumberOfTraces() const {
return kMaximumNumberOfMemoryOperationTraces;
}
AllocationTraceRecorderStatistics GetRecorderStatistics() const;
private:
// Handle all allocation events.
NOINLINE void OnAllocation(const void* allocated_address,
size_t allocated_size);
// Handle all free events.
NOINLINE void OnFree(const void* freed_address);
ALWAYS_INLINE size_t GetNextIndex();
ALWAYS_INLINE static constexpr size_t WrapIdxIfNeeded(size_t idx);
// The actual container.
std::array<OperationRecord, kMaximumNumberOfMemoryOperationTraces>
alloc_trace_buffer_ = {};
// The total number of records that have been taken so far. Note that this
// might be greater than |kMaximumNumberOfMemoryOperationTraces| since we
// overwrite oldest items.
std::atomic<size_t> total_number_of_records_ = 0;
#if BUILDFLAG(ENABLE_ALLOCATION_TRACE_RECORDER_FULL_REPORTING)
std::atomic<size_t> total_number_of_collisions_ = 0;
#endif
};
inline void AllocationTraceRecorder::OnAllocation(
const base::allocator::dispatcher::AllocationNotificationData&
allocation_data) {
OnAllocation(allocation_data.address(), allocation_data.size());
}
// Handle all free events.
inline void AllocationTraceRecorder::OnFree(
const base::allocator::dispatcher::FreeNotificationData& free_data) {
OnFree(free_data.address());
}
ALWAYS_INLINE constexpr size_t AllocationTraceRecorder::WrapIdxIfNeeded(
size_t idx) {
// Wrapping around counter, e.g. for BUFFER_SIZE = 256, the counter will
// wrap around when reaching 256. To enable the compiler to emit more
// optimized code we assert |kMaximumNumberOfMemoryOperationTraces| is a power
// of two .
static_assert(
std::has_single_bit(kMaximumNumberOfMemoryOperationTraces),
"kMaximumNumberOfMemoryOperationTraces should be a power of 2 to "
"allow for fast modulo operation.");
return idx % kMaximumNumberOfMemoryOperationTraces;
}
ALWAYS_INLINE size_t AllocationTraceRecorder::GetNextIndex() {
const auto raw_idx =
total_number_of_records_.fetch_add(1, std::memory_order_relaxed);
return WrapIdxIfNeeded(raw_idx);
}
} // namespace base::debug::tracer
#endif // BASE_DEBUG_ALLOCATION_TRACE_H_
|