12bc53e031
-- c99f979ad34f155fbeeea69b88bdc7458d89a21c by Derek Mauro <dmauro@google.com>: Remove a floating point division by zero test. This isn't testing behavior related to the library, and MSVC warns about it in opt mode. PiperOrigin-RevId: 285220804 -- 68b015491f0dbf1ab547994673281abd1f34cd4b by Gennadiy Rozental <rogeeff@google.com>: This CL introduces following changes to the class FlagImpl: * We eliminate the CommandLineFlagLocks struct. Instead callback guard and callback function are combined into a single CallbackData struct, while primary data lock is stored separately. * CallbackData member of class FlagImpl is initially set to be nullptr and is only allocated and initialized when a flag's callback is being set. For most flags we do not pay for the extra space and extra absl::Mutex now. * Primary data guard is stored in data_guard_ data member. This is a properly aligned character buffer of necessary size. During initialization of the flag we construct absl::Mutex in this space using placement new call. * We now avoid extra value copy after successful attempt to parse value out of string. Instead we swap flag's current value with tentative value we just produced. PiperOrigin-RevId: 285132636 -- ed45d118fb818969eb13094cf7827c885dfc562c by Tom Manshreck <shreck@google.com>: Change null-term* (and nul-term*) to NUL-term* in comments PiperOrigin-RevId: 285036610 -- 729619017944db895ce8d6d29c1995aa2e5628a5 by Derek Mauro <dmauro@google.com>: Use the Posix implementation of thread identity on MinGW. Some versions of MinGW suffer from thread_local bugs. PiperOrigin-RevId: 285022920 -- 39a25493503c76885bc3254c28f66a251c5b5bb0 by Greg Falcon <gfalcon@google.com>: Implementation detail change. Add further ABSL_NAMESPACE_BEGIN and _END annotation macros to files in Abseil. PiperOrigin-RevId: 285012012 GitOrigin-RevId: c99f979ad34f155fbeeea69b88bdc7458d89a21c Change-Id: I4c85d3704e45d11a9ac50d562f39640a6adbedc1
99 lines
3.4 KiB
C++
99 lines
3.4 KiB
C++
// Copyright 2000 - 2007 Google Inc.
|
|
// All rights reserved.
|
|
//
|
|
// Author: Sanjay Ghemawat
|
|
//
|
|
// Portable implementation - just use glibc
|
|
//
|
|
// Note: The glibc implementation may cause a call to malloc.
|
|
// This can cause a deadlock in HeapProfiler.
|
|
|
|
#ifndef ABSL_DEBUGGING_INTERNAL_STACKTRACE_GENERIC_INL_H_
|
|
#define ABSL_DEBUGGING_INTERNAL_STACKTRACE_GENERIC_INL_H_
|
|
|
|
#include <execinfo.h>
|
|
#include <atomic>
|
|
#include <cstring>
|
|
|
|
#include "absl/debugging/stacktrace.h"
|
|
#include "absl/base/attributes.h"
|
|
|
|
// Sometimes, we can try to get a stack trace from within a stack
|
|
// trace, because we don't block signals inside this code (which would be too
|
|
// expensive: the two extra system calls per stack trace do matter here).
|
|
// That can cause a self-deadlock.
|
|
// Protect against such reentrant call by failing to get a stack trace.
|
|
//
|
|
// We use __thread here because the code here is extremely low level -- it is
|
|
// called while collecting stack traces from within malloc and mmap, and thus
|
|
// can not call anything which might call malloc or mmap itself.
|
|
static __thread int recursive = 0;
|
|
|
|
// The stack trace function might be invoked very early in the program's
|
|
// execution (e.g. from the very first malloc if using tcmalloc). Also, the
|
|
// glibc implementation itself will trigger malloc the first time it is called.
|
|
// As such, we suppress usage of backtrace during this early stage of execution.
|
|
static std::atomic<bool> disable_stacktraces(true); // Disabled until healthy.
|
|
// Waiting until static initializers run seems to be late enough.
|
|
// This file is included into stacktrace.cc so this will only run once.
|
|
ABSL_ATTRIBUTE_UNUSED static int stacktraces_enabler = []() {
|
|
void* unused_stack[1];
|
|
// Force the first backtrace to happen early to get the one-time shared lib
|
|
// loading (allocation) out of the way. After the first call it is much safer
|
|
// to use backtrace from a signal handler if we crash somewhere later.
|
|
backtrace(unused_stack, 1);
|
|
disable_stacktraces.store(false, std::memory_order_relaxed);
|
|
return 0;
|
|
}();
|
|
|
|
template <bool IS_STACK_FRAMES, bool IS_WITH_CONTEXT>
|
|
static int UnwindImpl(void** result, int* sizes, int max_depth, int skip_count,
|
|
const void *ucp, int *min_dropped_frames) {
|
|
if (recursive || disable_stacktraces.load(std::memory_order_relaxed)) {
|
|
return 0;
|
|
}
|
|
++recursive;
|
|
|
|
static_cast<void>(ucp); // Unused.
|
|
static const int kStackLength = 64;
|
|
void * stack[kStackLength];
|
|
int size;
|
|
|
|
size = backtrace(stack, kStackLength);
|
|
skip_count++; // we want to skip the current frame as well
|
|
int result_count = size - skip_count;
|
|
if (result_count < 0)
|
|
result_count = 0;
|
|
if (result_count > max_depth)
|
|
result_count = max_depth;
|
|
for (int i = 0; i < result_count; i++)
|
|
result[i] = stack[i + skip_count];
|
|
|
|
if (IS_STACK_FRAMES) {
|
|
// No implementation for finding out the stack frame sizes yet.
|
|
memset(sizes, 0, sizeof(*sizes) * result_count);
|
|
}
|
|
if (min_dropped_frames != nullptr) {
|
|
if (size - skip_count - max_depth > 0) {
|
|
*min_dropped_frames = size - skip_count - max_depth;
|
|
} else {
|
|
*min_dropped_frames = 0;
|
|
}
|
|
}
|
|
|
|
--recursive;
|
|
|
|
return result_count;
|
|
}
|
|
|
|
namespace absl {
|
|
ABSL_NAMESPACE_BEGIN
|
|
namespace debugging_internal {
|
|
bool StackTraceWorksForTest() {
|
|
return true;
|
|
}
|
|
} // namespace debugging_internal
|
|
ABSL_NAMESPACE_END
|
|
} // namespace absl
|
|
|
|
#endif // ABSL_DEBUGGING_INTERNAL_STACKTRACE_GENERIC_INL_H_
|