Merge remote-tracking branch 'origin/master'
Some checks failed
linux-x64-gcc / linux-gcc (Debug) (push) Failing after 1m37s
linux-x64-gcc / linux-gcc (Release) (push) Failing after 1m58s
linux-arm-gcc / linux-gcc-armhf (push) Failing after 2m10s
linux-mips64-gcc / linux-gcc-mips64el (Release) (push) Failing after 2m41s
linux-mips64-gcc / linux-gcc-mips64el (Debug) (push) Failing after 3m23s
linux-aarch64-cpu-gcc / linux-gcc-aarch64 (push) Has been cancelled
Some checks failed
linux-x64-gcc / linux-gcc (Debug) (push) Failing after 1m37s
linux-x64-gcc / linux-gcc (Release) (push) Failing after 1m58s
linux-arm-gcc / linux-gcc-armhf (push) Failing after 2m10s
linux-mips64-gcc / linux-gcc-mips64el (Release) (push) Failing after 2m41s
linux-mips64-gcc / linux-gcc-mips64el (Debug) (push) Failing after 3m23s
linux-aarch64-cpu-gcc / linux-gcc-aarch64 (push) Has been cancelled
This commit is contained in:
commit
fc1ad57bbb
File diff suppressed because it is too large
Load Diff
@ -55,11 +55,11 @@
|
||||
|
||||
/* Annoying stuff for windows; makes sure clients can import these functions */
|
||||
#ifndef PERFTOOLS_DLL_DECL
|
||||
# ifdef _WIN32
|
||||
# define PERFTOOLS_DLL_DECL __declspec(dllimport)
|
||||
# else
|
||||
# define PERFTOOLS_DLL_DECL
|
||||
# endif
|
||||
#ifdef _WIN32
|
||||
#define PERFTOOLS_DLL_DECL __declspec(dllimport)
|
||||
#else
|
||||
#define PERFTOOLS_DLL_DECL
|
||||
#endif
|
||||
#endif
|
||||
|
||||
/* All this code should be usable from within C apps. */
|
||||
@ -70,7 +70,7 @@ extern "C" {
|
||||
/* Start profiling and arrange to write profile data to file names
|
||||
* of the form: "prefix.0000", "prefix.0001", ...
|
||||
*/
|
||||
PERFTOOLS_DLL_DECL void HeapProfilerStart(const char* prefix);
|
||||
PERFTOOLS_DLL_DECL void HeapProfilerStart(const char *prefix);
|
||||
|
||||
/* Returns non-zero if we are currently profiling the heap. (Returns
|
||||
* an int rather than a bool so it's usable from C.) This is true
|
||||
@ -96,10 +96,10 @@ PERFTOOLS_DLL_DECL void HeapProfilerDump(const char *reason);
|
||||
* The returned pointer is a '\0'-terminated string allocated using malloc()
|
||||
* and should be free()-ed as soon as the caller does not need it anymore.
|
||||
*/
|
||||
PERFTOOLS_DLL_DECL char* GetHeapProfile();
|
||||
PERFTOOLS_DLL_DECL char *GetHeapProfile();
|
||||
|
||||
#ifdef __cplusplus
|
||||
} // extern "C"
|
||||
}// extern "C"
|
||||
#endif
|
||||
|
||||
#endif /* BASE_HEAP_PROFILER_H_ */
|
||||
#endif /* BASE_HEAP_PROFILER_H_ */
|
||||
|
@ -42,41 +42,41 @@
|
||||
#endif
|
||||
#include <inttypes.h>
|
||||
#ifdef HAVE_FCNTL_H
|
||||
#include <fcntl.h> // for open()
|
||||
#include <fcntl.h>// for open()
|
||||
#endif
|
||||
#ifdef HAVE_MMAP
|
||||
#include <sys/mman.h>
|
||||
#endif
|
||||
#include <errno.h>
|
||||
#include <assert.h>
|
||||
#include <sys/types.h>
|
||||
#include <errno.h>
|
||||
#include <signal.h>
|
||||
#include <sys/types.h>
|
||||
|
||||
#include <algorithm>
|
||||
#include <string>
|
||||
|
||||
#include <gperftools/heap-profiler.h>
|
||||
|
||||
#include "base/logging.h"
|
||||
#include "base/basictypes.h" // for PRId64, among other things
|
||||
#include "base/googleinit.h"
|
||||
#include "base/basictypes.h"// for PRId64, among other things
|
||||
#include "base/commandlineflags.h"
|
||||
#include "malloc_hook-inl.h"
|
||||
#include "tcmalloc_guard.h"
|
||||
#include <gperftools/malloc_hook.h>
|
||||
#include <gperftools/malloc_extension.h>
|
||||
#include "base/spinlock.h"
|
||||
#include "base/googleinit.h"
|
||||
#include "base/logging.h"
|
||||
#include "base/low_level_alloc.h"
|
||||
#include "base/sysinfo.h" // for GetUniquePathFromEnv()
|
||||
#include "base/spinlock.h"
|
||||
#include "base/sysinfo.h"// for GetUniquePathFromEnv()
|
||||
#include "heap-profile-table.h"
|
||||
#include "malloc_hook-inl.h"
|
||||
#include "memory_region_map.h"
|
||||
#include "mmap_hook.h"
|
||||
#include "tcmalloc_guard.h"
|
||||
#include <gperftools/malloc_extension.h>
|
||||
#include <gperftools/malloc_hook.h>
|
||||
|
||||
#ifndef PATH_MAX
|
||||
#ifndef PATH_MAX
|
||||
#ifdef MAXPATHLEN
|
||||
#define PATH_MAX MAXPATHLEN
|
||||
#define PATH_MAX MAXPATHLEN
|
||||
#else
|
||||
#define PATH_MAX 4096 // seems conservative for max filename len!
|
||||
#define PATH_MAX 4096// seems conservative for max filename len!
|
||||
#endif
|
||||
#endif
|
||||
|
||||
@ -110,9 +110,7 @@ DEFINE_int64(heap_profile_time_interval,
|
||||
EnvToInt64("HEAP_PROFILE_TIME_INTERVAL", 0),
|
||||
"If non-zero, dump heap profiling information once every "
|
||||
"specified number of seconds since the last dump.");
|
||||
DEFINE_bool(mmap_log,
|
||||
EnvToBool("HEAP_PROFILE_MMAP_LOG", false),
|
||||
"Should mmap/munmap calls be logged?");
|
||||
DEFINE_bool(mmap_log, EnvToBool("HEAP_PROFILE_MMAP_LOG", false), "Should mmap/munmap calls be logged?");
|
||||
DEFINE_bool(mmap_profile,
|
||||
EnvToBool("HEAP_PROFILE_MMAP", false),
|
||||
"If heap-profiling is on, also profile mmap, mremap, and sbrk)");
|
||||
@ -121,7 +119,6 @@ DEFINE_bool(only_mmap_profile,
|
||||
"If heap-profiling is on, only profile mmap, mremap, and sbrk; "
|
||||
"do not profile malloc/new/etc");
|
||||
|
||||
|
||||
//----------------------------------------------------------------------
|
||||
// Locking
|
||||
//----------------------------------------------------------------------
|
||||
@ -140,11 +137,16 @@ static SpinLock heap_lock(SpinLock::LINKER_INITIALIZED);
|
||||
|
||||
static LowLevelAlloc::Arena *heap_profiler_memory;
|
||||
|
||||
static void* ProfilerMalloc(size_t bytes) {
|
||||
return LowLevelAlloc::AllocWithArena(bytes, heap_profiler_memory);
|
||||
static void *
|
||||
ProfilerMalloc(size_t bytes)
|
||||
{
|
||||
return LowLevelAlloc::AllocWithArena(bytes, heap_profiler_memory);
|
||||
}
|
||||
static void ProfilerFree(void* p) {
|
||||
LowLevelAlloc::Free(p);
|
||||
|
||||
static void
|
||||
ProfilerFree(void *p)
|
||||
{
|
||||
LowLevelAlloc::Free(p);
|
||||
}
|
||||
|
||||
// We use buffers of this size in DoGetHeapProfile.
|
||||
@ -155,106 +157,107 @@ static const int kProfileBufferSize = 1 << 20;
|
||||
// will be used by HeapProfileEndWriter when the application has to
|
||||
// exit due to out-of-memory. This buffer is allocated in
|
||||
// HeapProfilerStart. Access to this must be protected by heap_lock.
|
||||
static char* global_profiler_buffer = NULL;
|
||||
|
||||
static char *global_profiler_buffer = NULL;
|
||||
|
||||
//----------------------------------------------------------------------
|
||||
// Profiling control/state data
|
||||
//----------------------------------------------------------------------
|
||||
|
||||
// Access to all of these is protected by heap_lock.
|
||||
static bool is_on = false; // If are on as a subsytem.
|
||||
static bool dumping = false; // Dumping status to prevent recursion
|
||||
static char* filename_prefix = NULL; // Prefix used for profile file names
|
||||
// (NULL if no need for dumping yet)
|
||||
static int dump_count = 0; // How many dumps so far
|
||||
static int64 last_dump_alloc = 0; // alloc_size when did we last dump
|
||||
static int64 last_dump_free = 0; // free_size when did we last dump
|
||||
static int64 high_water_mark = 0; // In-use-bytes at last high-water dump
|
||||
static int64 last_dump_time = 0; // The time of the last dump
|
||||
static bool is_on = false;// If are on as a subsytem.
|
||||
static bool dumping = false;// Dumping status to prevent recursion
|
||||
static char *filename_prefix = NULL; // Prefix used for profile file names
|
||||
// (NULL if no need for dumping yet)
|
||||
static int dump_count = 0; // How many dumps so far
|
||||
static int64 last_dump_alloc = 0; // alloc_size when did we last dump
|
||||
static int64 last_dump_free = 0; // free_size when did we last dump
|
||||
static int64 high_water_mark = 0; // In-use-bytes at last high-water dump
|
||||
static int64 last_dump_time = 0; // The time of the last dump
|
||||
|
||||
static HeapProfileTable* heap_profile = NULL; // the heap profile table
|
||||
static HeapProfileTable *heap_profile = NULL;// the heap profile table
|
||||
|
||||
//----------------------------------------------------------------------
|
||||
// Profile generation
|
||||
//----------------------------------------------------------------------
|
||||
|
||||
// Input must be a buffer of size at least 1MB.
|
||||
static char* DoGetHeapProfileLocked(char* buf, int buflen) {
|
||||
// We used to be smarter about estimating the required memory and
|
||||
// then capping it to 1MB and generating the profile into that.
|
||||
if (buf == NULL || buflen < 1)
|
||||
return NULL;
|
||||
static char *
|
||||
DoGetHeapProfileLocked(char *buf, int buflen)
|
||||
{
|
||||
// We used to be smarter about estimating the required memory and
|
||||
// then capping it to 1MB and generating the profile into that.
|
||||
if (buf == NULL || buflen < 1) return NULL;
|
||||
|
||||
RAW_DCHECK(heap_lock.IsHeld(), "");
|
||||
int bytes_written = 0;
|
||||
if (is_on) {
|
||||
HeapProfileTable::Stats const stats = heap_profile->total();
|
||||
(void)stats; // avoid an unused-variable warning in non-debug mode.
|
||||
bytes_written = heap_profile->FillOrderedProfile(buf, buflen - 1);
|
||||
// FillOrderedProfile should not reduce the set of active mmap-ed regions,
|
||||
// hence MemoryRegionMap will let us remove everything we've added above:
|
||||
RAW_DCHECK(stats.Equivalent(heap_profile->total()), "");
|
||||
// if this fails, we somehow removed by FillOrderedProfile
|
||||
// more than we have added.
|
||||
}
|
||||
buf[bytes_written] = '\0';
|
||||
RAW_DCHECK(bytes_written == strlen(buf), "");
|
||||
RAW_DCHECK(heap_lock.IsHeld(), "");
|
||||
int bytes_written = 0;
|
||||
if (is_on) {
|
||||
HeapProfileTable::Stats const stats = heap_profile->total();
|
||||
(void) stats;// avoid an unused-variable warning in non-debug mode.
|
||||
bytes_written = heap_profile->FillOrderedProfile(buf, buflen - 1);
|
||||
// FillOrderedProfile should not reduce the set of active mmap-ed regions,
|
||||
// hence MemoryRegionMap will let us remove everything we've added above:
|
||||
RAW_DCHECK(stats.Equivalent(heap_profile->total()), "");
|
||||
// if this fails, we somehow removed by FillOrderedProfile
|
||||
// more than we have added.
|
||||
}
|
||||
buf[bytes_written] = '\0';
|
||||
RAW_DCHECK(bytes_written == strlen(buf), "");
|
||||
|
||||
return buf;
|
||||
return buf;
|
||||
}
|
||||
|
||||
extern "C" char* GetHeapProfile() {
|
||||
// Use normal malloc: we return the profile to the user to free it:
|
||||
char* buffer = reinterpret_cast<char*>(malloc(kProfileBufferSize));
|
||||
SpinLockHolder l(&heap_lock);
|
||||
return DoGetHeapProfileLocked(buffer, kProfileBufferSize);
|
||||
extern "C" char *
|
||||
GetHeapProfile()
|
||||
{
|
||||
// Use normal malloc: we return the profile to the user to free it:
|
||||
char *buffer = reinterpret_cast<char *>(malloc(kProfileBufferSize));
|
||||
SpinLockHolder l(&heap_lock);
|
||||
return DoGetHeapProfileLocked(buffer, kProfileBufferSize);
|
||||
}
|
||||
|
||||
// defined below
|
||||
static void NewHook(const void* ptr, size_t size);
|
||||
static void DeleteHook(const void* ptr);
|
||||
static void NewHook(const void *ptr, size_t size);
|
||||
static void DeleteHook(const void *ptr);
|
||||
|
||||
// Helper for HeapProfilerDump.
|
||||
static void DumpProfileLocked(const char* reason) {
|
||||
RAW_DCHECK(heap_lock.IsHeld(), "");
|
||||
RAW_DCHECK(is_on, "");
|
||||
RAW_DCHECK(!dumping, "");
|
||||
static void
|
||||
DumpProfileLocked(const char *reason)
|
||||
{
|
||||
RAW_DCHECK(heap_lock.IsHeld(), "");
|
||||
RAW_DCHECK(is_on, "");
|
||||
RAW_DCHECK(!dumping, "");
|
||||
|
||||
if (filename_prefix == NULL) return; // we do not yet need dumping
|
||||
if (filename_prefix == NULL) return;// we do not yet need dumping
|
||||
|
||||
dumping = true;
|
||||
dumping = true;
|
||||
|
||||
// Make file name
|
||||
char file_name[1000];
|
||||
dump_count++;
|
||||
snprintf(file_name, sizeof(file_name), "%s.%04d%s",
|
||||
filename_prefix, dump_count, HeapProfileTable::kFileExt);
|
||||
// Make file name
|
||||
char file_name[1000];
|
||||
dump_count++;
|
||||
snprintf(file_name, sizeof(file_name), "%s.%04d%s", filename_prefix, dump_count, HeapProfileTable::kFileExt);
|
||||
|
||||
// Dump the profile
|
||||
RAW_VLOG(0, "Dumping heap profile to %s (%s)", file_name, reason);
|
||||
// We must use file routines that don't access memory, since we hold
|
||||
// a memory lock now.
|
||||
RawFD fd = RawOpenForWriting(file_name);
|
||||
if (fd == kIllegalRawFD) {
|
||||
RAW_LOG(ERROR, "Failed dumping heap profile to %s. Numeric errno is %d", file_name, errno);
|
||||
dumping = false;
|
||||
return;
|
||||
}
|
||||
|
||||
// This case may be impossible, but it's best to be safe.
|
||||
// It's safe to use the global buffer: we're protected by heap_lock.
|
||||
if (global_profiler_buffer == NULL) {
|
||||
global_profiler_buffer = reinterpret_cast<char *>(ProfilerMalloc(kProfileBufferSize));
|
||||
}
|
||||
|
||||
char *profile = DoGetHeapProfileLocked(global_profiler_buffer, kProfileBufferSize);
|
||||
RawWrite(fd, profile, strlen(profile));
|
||||
RawClose(fd);
|
||||
|
||||
// Dump the profile
|
||||
RAW_VLOG(0, "Dumping heap profile to %s (%s)", file_name, reason);
|
||||
// We must use file routines that don't access memory, since we hold
|
||||
// a memory lock now.
|
||||
RawFD fd = RawOpenForWriting(file_name);
|
||||
if (fd == kIllegalRawFD) {
|
||||
RAW_LOG(ERROR, "Failed dumping heap profile to %s. Numeric errno is %d", file_name, errno);
|
||||
dumping = false;
|
||||
return;
|
||||
}
|
||||
|
||||
// This case may be impossible, but it's best to be safe.
|
||||
// It's safe to use the global buffer: we're protected by heap_lock.
|
||||
if (global_profiler_buffer == NULL) {
|
||||
global_profiler_buffer =
|
||||
reinterpret_cast<char*>(ProfilerMalloc(kProfileBufferSize));
|
||||
}
|
||||
|
||||
char* profile = DoGetHeapProfileLocked(global_profiler_buffer,
|
||||
kProfileBufferSize);
|
||||
RawWrite(fd, profile, strlen(profile));
|
||||
RawClose(fd);
|
||||
|
||||
dumping = false;
|
||||
}
|
||||
|
||||
//----------------------------------------------------------------------
|
||||
@ -263,73 +266,78 @@ static void DumpProfileLocked(const char* reason) {
|
||||
|
||||
// Dump a profile after either an allocation or deallocation, if
|
||||
// the memory use has changed enough since the last dump.
|
||||
static void MaybeDumpProfileLocked() {
|
||||
if (!dumping) {
|
||||
const HeapProfileTable::Stats& total = heap_profile->total();
|
||||
const int64_t inuse_bytes = total.alloc_size - total.free_size;
|
||||
bool need_to_dump = false;
|
||||
char buf[128];
|
||||
static void
|
||||
MaybeDumpProfileLocked()
|
||||
{
|
||||
if (!dumping) {
|
||||
const HeapProfileTable::Stats &total = heap_profile->total();
|
||||
const int64_t inuse_bytes = total.alloc_size - total.free_size;
|
||||
bool need_to_dump = false;
|
||||
char buf[128];
|
||||
|
||||
if (FLAGS_heap_profile_allocation_interval > 0 &&
|
||||
total.alloc_size >=
|
||||
last_dump_alloc + FLAGS_heap_profile_allocation_interval) {
|
||||
snprintf(buf, sizeof(buf), ("%" PRId64 " MB allocated cumulatively, "
|
||||
"%" PRId64 " MB currently in use"),
|
||||
total.alloc_size >> 20, inuse_bytes >> 20);
|
||||
need_to_dump = true;
|
||||
} else if (FLAGS_heap_profile_deallocation_interval > 0 &&
|
||||
total.free_size >=
|
||||
last_dump_free + FLAGS_heap_profile_deallocation_interval) {
|
||||
snprintf(buf, sizeof(buf), ("%" PRId64 " MB freed cumulatively, "
|
||||
"%" PRId64 " MB currently in use"),
|
||||
total.free_size >> 20, inuse_bytes >> 20);
|
||||
need_to_dump = true;
|
||||
} else if (FLAGS_heap_profile_inuse_interval > 0 &&
|
||||
inuse_bytes >
|
||||
high_water_mark + FLAGS_heap_profile_inuse_interval) {
|
||||
snprintf(buf, sizeof(buf), "%" PRId64 " MB currently in use",
|
||||
inuse_bytes >> 20);
|
||||
need_to_dump = true;
|
||||
} else if (FLAGS_heap_profile_time_interval > 0 ) {
|
||||
int64 current_time = time(NULL);
|
||||
if (current_time - last_dump_time >=
|
||||
FLAGS_heap_profile_time_interval) {
|
||||
snprintf(buf, sizeof(buf), "%" PRId64 " sec since the last dump",
|
||||
current_time - last_dump_time);
|
||||
need_to_dump = true;
|
||||
last_dump_time = current_time;
|
||||
}
|
||||
}
|
||||
if (need_to_dump) {
|
||||
DumpProfileLocked(buf);
|
||||
if (FLAGS_heap_profile_allocation_interval > 0
|
||||
&& total.alloc_size >= last_dump_alloc + FLAGS_heap_profile_allocation_interval) {
|
||||
snprintf(buf,
|
||||
sizeof(buf),
|
||||
("%" PRId64 " MB allocated cumulatively, "
|
||||
"%" PRId64 " MB currently in use"),
|
||||
total.alloc_size >> 20,
|
||||
inuse_bytes >> 20);
|
||||
need_to_dump = true;
|
||||
} else if (FLAGS_heap_profile_deallocation_interval > 0
|
||||
&& total.free_size >= last_dump_free + FLAGS_heap_profile_deallocation_interval) {
|
||||
snprintf(buf,
|
||||
sizeof(buf),
|
||||
("%" PRId64 " MB freed cumulatively, "
|
||||
"%" PRId64 " MB currently in use"),
|
||||
total.free_size >> 20,
|
||||
inuse_bytes >> 20);
|
||||
need_to_dump = true;
|
||||
} else if (FLAGS_heap_profile_inuse_interval > 0
|
||||
&& inuse_bytes > high_water_mark + FLAGS_heap_profile_inuse_interval) {
|
||||
snprintf(buf, sizeof(buf), "%" PRId64 " MB currently in use", inuse_bytes >> 20);
|
||||
need_to_dump = true;
|
||||
} else if (FLAGS_heap_profile_time_interval > 0) {
|
||||
int64 current_time = time(NULL);
|
||||
if (current_time - last_dump_time >= FLAGS_heap_profile_time_interval) {
|
||||
snprintf(buf, sizeof(buf), "%" PRId64 " sec since the last dump", current_time - last_dump_time);
|
||||
need_to_dump = true;
|
||||
last_dump_time = current_time;
|
||||
}
|
||||
}
|
||||
if (need_to_dump) {
|
||||
DumpProfileLocked(buf);
|
||||
|
||||
last_dump_alloc = total.alloc_size;
|
||||
last_dump_free = total.free_size;
|
||||
if (inuse_bytes > high_water_mark)
|
||||
high_water_mark = inuse_bytes;
|
||||
last_dump_alloc = total.alloc_size;
|
||||
last_dump_free = total.free_size;
|
||||
if (inuse_bytes > high_water_mark) high_water_mark = inuse_bytes;
|
||||
}
|
||||
}
|
||||
}
|
||||
}
|
||||
|
||||
// Record an allocation in the profile.
|
||||
static void RecordAlloc(const void* ptr, size_t bytes, int skip_count) {
|
||||
// Take the stack trace outside the critical section.
|
||||
void* stack[HeapProfileTable::kMaxStackDepth];
|
||||
int depth = HeapProfileTable::GetCallerStackTrace(skip_count + 1, stack);
|
||||
SpinLockHolder l(&heap_lock);
|
||||
if (is_on) {
|
||||
heap_profile->RecordAlloc(ptr, bytes, depth, stack);
|
||||
MaybeDumpProfileLocked();
|
||||
}
|
||||
static void
|
||||
RecordAlloc(const void *ptr, size_t bytes, int skip_count)
|
||||
{
|
||||
// Take the stack trace outside the critical section.
|
||||
void *stack[HeapProfileTable::kMaxStackDepth];
|
||||
int depth = HeapProfileTable::GetCallerStackTrace(skip_count + 1, stack);
|
||||
SpinLockHolder l(&heap_lock);
|
||||
if (is_on) {
|
||||
heap_profile->RecordAlloc(ptr, bytes, depth, stack);
|
||||
MaybeDumpProfileLocked();
|
||||
}
|
||||
}
|
||||
|
||||
// Record a deallocation in the profile.
|
||||
static void RecordFree(const void* ptr) {
|
||||
SpinLockHolder l(&heap_lock);
|
||||
if (is_on) {
|
||||
heap_profile->RecordFree(ptr);
|
||||
MaybeDumpProfileLocked();
|
||||
}
|
||||
static void
|
||||
RecordFree(const void *ptr)
|
||||
{
|
||||
SpinLockHolder l(&heap_lock);
|
||||
if (is_on) {
|
||||
heap_profile->RecordFree(ptr);
|
||||
MaybeDumpProfileLocked();
|
||||
}
|
||||
}
|
||||
|
||||
//----------------------------------------------------------------------
|
||||
@ -337,256 +345,263 @@ static void RecordFree(const void* ptr) {
|
||||
//----------------------------------------------------------------------
|
||||
|
||||
// static
|
||||
void NewHook(const void* ptr, size_t size) {
|
||||
if (ptr != NULL) RecordAlloc(ptr, size, 0);
|
||||
void
|
||||
NewHook(const void *ptr, size_t size)
|
||||
{
|
||||
if (ptr != NULL) RecordAlloc(ptr, size, 0);
|
||||
}
|
||||
|
||||
// static
|
||||
void DeleteHook(const void* ptr) {
|
||||
if (ptr != NULL) RecordFree(ptr);
|
||||
void
|
||||
DeleteHook(const void *ptr)
|
||||
{
|
||||
if (ptr != NULL) RecordFree(ptr);
|
||||
}
|
||||
|
||||
static tcmalloc::MappingHookSpace mmap_logging_hook_space;
|
||||
|
||||
static void LogMappingEvent(const tcmalloc::MappingEvent& evt) {
|
||||
if (!FLAGS_mmap_log) {
|
||||
return;
|
||||
}
|
||||
static void
|
||||
LogMappingEvent(const tcmalloc::MappingEvent &evt)
|
||||
{
|
||||
if (!FLAGS_mmap_log) { return; }
|
||||
|
||||
if (evt.file_valid) {
|
||||
// We use PRIxPTR not just '%p' to avoid deadlocks
|
||||
// in pretty-printing of NULL as "nil".
|
||||
// TODO(maxim): instead should use a safe snprintf reimplementation
|
||||
RAW_LOG(INFO,
|
||||
"mmap(start=0x%" PRIxPTR ", len=%zu, prot=0x%x, flags=0x%x, "
|
||||
"fd=%d, offset=0x%llx) = 0x%" PRIxPTR "",
|
||||
(uintptr_t) evt.before_address, evt.after_length, evt.prot,
|
||||
evt.flags, evt.file_fd, (unsigned long long) evt.file_off,
|
||||
(uintptr_t) evt.after_address);
|
||||
} else if (evt.after_valid && evt.before_valid) {
|
||||
// We use PRIxPTR not just '%p' to avoid deadlocks
|
||||
// in pretty-printing of NULL as "nil".
|
||||
// TODO(maxim): instead should use a safe snprintf reimplementation
|
||||
RAW_LOG(INFO,
|
||||
"mremap(old_addr=0x%" PRIxPTR ", old_size=%zu, "
|
||||
"new_size=%zu, flags=0x%x, new_addr=0x%" PRIxPTR ") = "
|
||||
"0x%" PRIxPTR "",
|
||||
(uintptr_t) evt.before_address, evt.before_length, evt.after_length, evt.flags,
|
||||
(uintptr_t) evt.after_address, (uintptr_t) evt.after_address);
|
||||
} else if (evt.is_sbrk) {
|
||||
intptr_t increment;
|
||||
uintptr_t result;
|
||||
if (evt.after_valid) {
|
||||
increment = evt.after_length;
|
||||
result = reinterpret_cast<uintptr_t>(evt.after_address) + evt.after_length;
|
||||
} else {
|
||||
increment = -static_cast<intptr_t>(evt.before_length);
|
||||
result = reinterpret_cast<uintptr_t>(evt.before_address);
|
||||
if (evt.file_valid) {
|
||||
// We use PRIxPTR not just '%p' to avoid deadlocks
|
||||
// in pretty-printing of NULL as "nil".
|
||||
// TODO(maxim): instead should use a safe snprintf reimplementation
|
||||
RAW_LOG(INFO,
|
||||
"mmap(start=0x%" PRIxPTR
|
||||
", len=%zu, prot=0x%x, flags=0x%x, "
|
||||
"fd=%d, offset=0x%llx) = 0x%" PRIxPTR "",
|
||||
(uintptr_t) evt.before_address,
|
||||
evt.after_length,
|
||||
evt.prot,
|
||||
evt.flags,
|
||||
evt.file_fd,
|
||||
(unsigned long long) evt.file_off,
|
||||
(uintptr_t) evt.after_address);
|
||||
} else if (evt.after_valid && evt.before_valid) {
|
||||
// We use PRIxPTR not just '%p' to avoid deadlocks
|
||||
// in pretty-printing of NULL as "nil".
|
||||
// TODO(maxim): instead should use a safe snprintf reimplementation
|
||||
RAW_LOG(INFO,
|
||||
"mremap(old_addr=0x%" PRIxPTR
|
||||
", old_size=%zu, "
|
||||
"new_size=%zu, flags=0x%x, new_addr=0x%" PRIxPTR
|
||||
") = "
|
||||
"0x%" PRIxPTR "",
|
||||
(uintptr_t) evt.before_address,
|
||||
evt.before_length,
|
||||
evt.after_length,
|
||||
evt.flags,
|
||||
(uintptr_t) evt.after_address,
|
||||
(uintptr_t) evt.after_address);
|
||||
} else if (evt.is_sbrk) {
|
||||
intptr_t increment;
|
||||
uintptr_t result;
|
||||
if (evt.after_valid) {
|
||||
increment = evt.after_length;
|
||||
result = reinterpret_cast<uintptr_t>(evt.after_address) + evt.after_length;
|
||||
} else {
|
||||
increment = -static_cast<intptr_t>(evt.before_length);
|
||||
result = reinterpret_cast<uintptr_t>(evt.before_address);
|
||||
}
|
||||
|
||||
RAW_LOG(INFO, "sbrk(inc=%zd) = 0x%" PRIxPTR "", increment, (uintptr_t) result);
|
||||
} else if (evt.before_valid) {
|
||||
// We use PRIxPTR not just '%p' to avoid deadlocks
|
||||
// in pretty-printing of NULL as "nil".
|
||||
// TODO(maxim): instead should use a safe snprintf reimplementation
|
||||
RAW_LOG(INFO, "munmap(start=0x%" PRIxPTR ", len=%zu)", (uintptr_t) evt.before_address, evt.before_length);
|
||||
}
|
||||
|
||||
RAW_LOG(INFO, "sbrk(inc=%zd) = 0x%" PRIxPTR "",
|
||||
increment, (uintptr_t) result);
|
||||
} else if (evt.before_valid) {
|
||||
// We use PRIxPTR not just '%p' to avoid deadlocks
|
||||
// in pretty-printing of NULL as "nil".
|
||||
// TODO(maxim): instead should use a safe snprintf reimplementation
|
||||
RAW_LOG(INFO, "munmap(start=0x%" PRIxPTR ", len=%zu)",
|
||||
(uintptr_t) evt.before_address, evt.before_length);
|
||||
}
|
||||
}
|
||||
|
||||
//----------------------------------------------------------------------
|
||||
// Starting/stopping/dumping
|
||||
//----------------------------------------------------------------------
|
||||
|
||||
extern "C" void HeapProfilerStart(const char* prefix) {
|
||||
SpinLockHolder l(&heap_lock);
|
||||
extern "C" void
|
||||
HeapProfilerStart(const char *prefix)
|
||||
{
|
||||
SpinLockHolder l(&heap_lock);
|
||||
|
||||
if (is_on) return;
|
||||
if (is_on) return;
|
||||
|
||||
is_on = true;
|
||||
is_on = true;
|
||||
|
||||
RAW_VLOG(0, "Starting tracking the heap");
|
||||
RAW_VLOG(0, "Starting tracking the heap");
|
||||
|
||||
// This should be done before the hooks are set up, since it should
|
||||
// call new, and we want that to be accounted for correctly.
|
||||
MallocExtension::Initialize();
|
||||
// This should be done before the hooks are set up, since it should
|
||||
// call new, and we want that to be accounted for correctly.
|
||||
MallocExtension::Initialize();
|
||||
|
||||
if (FLAGS_only_mmap_profile) {
|
||||
FLAGS_mmap_profile = true;
|
||||
}
|
||||
if (FLAGS_only_mmap_profile) { FLAGS_mmap_profile = true; }
|
||||
|
||||
if (FLAGS_mmap_profile) {
|
||||
// Ask MemoryRegionMap to record all mmap, mremap, and sbrk
|
||||
// call stack traces of at least size kMaxStackDepth:
|
||||
MemoryRegionMap::Init(HeapProfileTable::kMaxStackDepth,
|
||||
/* use_buckets */ true);
|
||||
}
|
||||
if (FLAGS_mmap_profile) {
|
||||
// Ask MemoryRegionMap to record all mmap, mremap, and sbrk
|
||||
// call stack traces of at least size kMaxStackDepth:
|
||||
MemoryRegionMap::Init(HeapProfileTable::kMaxStackDepth,
|
||||
/* use_buckets */ true);
|
||||
}
|
||||
|
||||
if (FLAGS_mmap_log) {
|
||||
// Install our hooks to do the logging:
|
||||
tcmalloc::HookMMapEvents(&mmap_logging_hook_space, LogMappingEvent);
|
||||
}
|
||||
if (FLAGS_mmap_log) {
|
||||
// Install our hooks to do the logging:
|
||||
tcmalloc::HookMMapEvents(&mmap_logging_hook_space, LogMappingEvent);
|
||||
}
|
||||
|
||||
heap_profiler_memory =
|
||||
LowLevelAlloc::NewArena(0, LowLevelAlloc::DefaultArena());
|
||||
heap_profiler_memory = LowLevelAlloc::NewArena(0, LowLevelAlloc::DefaultArena());
|
||||
|
||||
// Reserve space now for the heap profiler, so we can still write a
|
||||
// heap profile even if the application runs out of memory.
|
||||
global_profiler_buffer =
|
||||
reinterpret_cast<char*>(ProfilerMalloc(kProfileBufferSize));
|
||||
// Reserve space now for the heap profiler, so we can still write a
|
||||
// heap profile even if the application runs out of memory.
|
||||
global_profiler_buffer = reinterpret_cast<char *>(ProfilerMalloc(kProfileBufferSize));
|
||||
|
||||
heap_profile = new(ProfilerMalloc(sizeof(HeapProfileTable)))
|
||||
HeapProfileTable(ProfilerMalloc, ProfilerFree, FLAGS_mmap_profile);
|
||||
heap_profile = new (ProfilerMalloc(sizeof(HeapProfileTable)))
|
||||
HeapProfileTable(ProfilerMalloc, ProfilerFree, FLAGS_mmap_profile);
|
||||
|
||||
last_dump_alloc = 0;
|
||||
last_dump_free = 0;
|
||||
high_water_mark = 0;
|
||||
last_dump_time = 0;
|
||||
last_dump_alloc = 0;
|
||||
last_dump_free = 0;
|
||||
high_water_mark = 0;
|
||||
last_dump_time = 0;
|
||||
|
||||
// We do not reset dump_count so if the user does a sequence of
|
||||
// HeapProfilerStart/HeapProfileStop, we will get a continuous
|
||||
// sequence of profiles.
|
||||
// We do not reset dump_count so if the user does a sequence of
|
||||
// HeapProfilerStart/HeapProfileStop, we will get a continuous
|
||||
// sequence of profiles.
|
||||
|
||||
if (FLAGS_only_mmap_profile == false) {
|
||||
// Now set the hooks that capture new/delete and malloc/free.
|
||||
RAW_CHECK(MallocHook::AddNewHook(&NewHook), "");
|
||||
RAW_CHECK(MallocHook::AddDeleteHook(&DeleteHook), "");
|
||||
}
|
||||
if (FLAGS_only_mmap_profile == false) {
|
||||
// Now set the hooks that capture new/delete and malloc/free.
|
||||
RAW_CHECK(MallocHook::AddNewHook(&NewHook), "");
|
||||
RAW_CHECK(MallocHook::AddDeleteHook(&DeleteHook), "");
|
||||
}
|
||||
|
||||
// Copy filename prefix
|
||||
RAW_DCHECK(filename_prefix == NULL, "");
|
||||
const int prefix_length = strlen(prefix);
|
||||
filename_prefix = reinterpret_cast<char*>(ProfilerMalloc(prefix_length + 1));
|
||||
memcpy(filename_prefix, prefix, prefix_length);
|
||||
filename_prefix[prefix_length] = '\0';
|
||||
// Copy filename prefix
|
||||
RAW_DCHECK(filename_prefix == NULL, "");
|
||||
const int prefix_length = strlen(prefix);
|
||||
filename_prefix = reinterpret_cast<char *>(ProfilerMalloc(prefix_length + 1));
|
||||
memcpy(filename_prefix, prefix, prefix_length);
|
||||
filename_prefix[prefix_length] = '\0';
|
||||
}
|
||||
|
||||
extern "C" int IsHeapProfilerRunning() {
|
||||
SpinLockHolder l(&heap_lock);
|
||||
return is_on ? 1 : 0; // return an int, because C code doesn't have bool
|
||||
extern "C" int
|
||||
IsHeapProfilerRunning()
|
||||
{
|
||||
SpinLockHolder l(&heap_lock);
|
||||
return is_on ? 1 : 0;// return an int, because C code doesn't have bool
|
||||
}
|
||||
|
||||
extern "C" void HeapProfilerStop() {
|
||||
SpinLockHolder l(&heap_lock);
|
||||
extern "C" void
|
||||
HeapProfilerStop()
|
||||
{
|
||||
SpinLockHolder l(&heap_lock);
|
||||
|
||||
if (!is_on) return;
|
||||
if (!is_on) return;
|
||||
|
||||
if (FLAGS_only_mmap_profile == false) {
|
||||
// Unset our new/delete hooks, checking they were set:
|
||||
RAW_CHECK(MallocHook::RemoveNewHook(&NewHook), "");
|
||||
RAW_CHECK(MallocHook::RemoveDeleteHook(&DeleteHook), "");
|
||||
}
|
||||
if (FLAGS_mmap_log) {
|
||||
// Restore mmap/sbrk hooks, checking that our hooks were set:
|
||||
tcmalloc::UnHookMMapEvents(&mmap_logging_hook_space);
|
||||
}
|
||||
if (FLAGS_only_mmap_profile == false) {
|
||||
// Unset our new/delete hooks, checking they were set:
|
||||
RAW_CHECK(MallocHook::RemoveNewHook(&NewHook), "");
|
||||
RAW_CHECK(MallocHook::RemoveDeleteHook(&DeleteHook), "");
|
||||
}
|
||||
if (FLAGS_mmap_log) {
|
||||
// Restore mmap/sbrk hooks, checking that our hooks were set:
|
||||
tcmalloc::UnHookMMapEvents(&mmap_logging_hook_space);
|
||||
}
|
||||
|
||||
// free profile
|
||||
heap_profile->~HeapProfileTable();
|
||||
ProfilerFree(heap_profile);
|
||||
heap_profile = NULL;
|
||||
// free profile
|
||||
heap_profile->~HeapProfileTable();
|
||||
ProfilerFree(heap_profile);
|
||||
heap_profile = NULL;
|
||||
|
||||
// free output-buffer memory
|
||||
ProfilerFree(global_profiler_buffer);
|
||||
// free output-buffer memory
|
||||
ProfilerFree(global_profiler_buffer);
|
||||
|
||||
// free prefix
|
||||
ProfilerFree(filename_prefix);
|
||||
filename_prefix = NULL;
|
||||
// free prefix
|
||||
ProfilerFree(filename_prefix);
|
||||
filename_prefix = NULL;
|
||||
|
||||
if (!LowLevelAlloc::DeleteArena(heap_profiler_memory)) {
|
||||
RAW_LOG(FATAL, "Memory leak in HeapProfiler:");
|
||||
}
|
||||
if (!LowLevelAlloc::DeleteArena(heap_profiler_memory)) { RAW_LOG(FATAL, "Memory leak in HeapProfiler:"); }
|
||||
|
||||
if (FLAGS_mmap_profile) {
|
||||
MemoryRegionMap::Shutdown();
|
||||
}
|
||||
if (FLAGS_mmap_profile) { MemoryRegionMap::Shutdown(); }
|
||||
|
||||
is_on = false;
|
||||
is_on = false;
|
||||
}
|
||||
|
||||
extern "C" void HeapProfilerDump(const char *reason) {
|
||||
SpinLockHolder l(&heap_lock);
|
||||
if (is_on && !dumping) {
|
||||
DumpProfileLocked(reason);
|
||||
}
|
||||
extern "C" void
|
||||
HeapProfilerDump(const char *reason)
|
||||
{
|
||||
SpinLockHolder l(&heap_lock);
|
||||
if (is_on && !dumping) { DumpProfileLocked(reason); }
|
||||
}
|
||||
|
||||
// Signal handler that is registered when a user selectable signal
|
||||
// number is defined in the environment variable HEAPPROFILESIGNAL.
|
||||
static void HeapProfilerDumpSignal(int signal_number) {
|
||||
(void)signal_number;
|
||||
if (!heap_lock.TryLock()) {
|
||||
return;
|
||||
}
|
||||
if (is_on && !dumping) {
|
||||
DumpProfileLocked("signal");
|
||||
}
|
||||
heap_lock.Unlock();
|
||||
static void
|
||||
HeapProfilerDumpSignal(int signal_number)
|
||||
{
|
||||
(void) signal_number;
|
||||
if (!heap_lock.TryLock()) { return; }
|
||||
if (is_on && !dumping) { DumpProfileLocked("signal"); }
|
||||
heap_lock.Unlock();
|
||||
}
|
||||
|
||||
|
||||
//----------------------------------------------------------------------
|
||||
// Initialization/finalization code
|
||||
//----------------------------------------------------------------------
|
||||
|
||||
// Initialization code
|
||||
static void HeapProfilerInit() {
|
||||
// Everything after this point is for setting up the profiler based on envvar
|
||||
char fname[PATH_MAX];
|
||||
if (!GetUniquePathFromEnv("HEAPPROFILE", fname)) {
|
||||
return;
|
||||
}
|
||||
// We do a uid check so we don't write out files in a setuid executable.
|
||||
static void
|
||||
HeapProfilerInit()
|
||||
{
|
||||
// Everything after this point is for setting up the profiler based on envvar
|
||||
char fname[PATH_MAX];
|
||||
if (!GetUniquePathFromEnv("HEAPPROFILE", fname)) { return; }
|
||||
// We do a uid check so we don't write out files in a setuid executable.
|
||||
#ifdef HAVE_GETEUID
|
||||
if (getuid() != geteuid()) {
|
||||
RAW_LOG(WARNING, ("HeapProfiler: ignoring HEAPPROFILE because "
|
||||
"program seems to be setuid\n"));
|
||||
return;
|
||||
}
|
||||
if (getuid() != geteuid()) {
|
||||
RAW_LOG(WARNING,
|
||||
("HeapProfiler: ignoring HEAPPROFILE because "
|
||||
"program seems to be setuid\n"));
|
||||
return;
|
||||
}
|
||||
#endif
|
||||
|
||||
char *signal_number_str = getenv("HEAPPROFILESIGNAL");
|
||||
if (signal_number_str != NULL) {
|
||||
long int signal_number = strtol(signal_number_str, NULL, 10);
|
||||
intptr_t old_signal_handler = reinterpret_cast<intptr_t>(signal(signal_number, HeapProfilerDumpSignal));
|
||||
if (old_signal_handler == reinterpret_cast<intptr_t>(SIG_ERR)) {
|
||||
RAW_LOG(FATAL, "Failed to set signal. Perhaps signal number %s is invalid\n", signal_number_str);
|
||||
} else if (old_signal_handler == 0) {
|
||||
RAW_LOG(INFO,"Using signal %d as heap profiling switch", signal_number);
|
||||
} else {
|
||||
RAW_LOG(FATAL, "Signal %d already in use\n", signal_number);
|
||||
char *signal_number_str = getenv("HEAPPROFILESIGNAL");
|
||||
if (signal_number_str != NULL) {
|
||||
long int signal_number = strtol(signal_number_str, NULL, 10);
|
||||
intptr_t old_signal_handler = reinterpret_cast<intptr_t>(signal(signal_number, HeapProfilerDumpSignal));
|
||||
if (old_signal_handler == reinterpret_cast<intptr_t>(SIG_ERR)) {
|
||||
RAW_LOG(FATAL, "Failed to set signal. Perhaps signal number %s is invalid\n", signal_number_str);
|
||||
} else if (old_signal_handler == 0) {
|
||||
RAW_LOG(INFO, "Using signal %d as heap profiling switch", signal_number);
|
||||
} else {
|
||||
RAW_LOG(FATAL, "Signal %d already in use\n", signal_number);
|
||||
}
|
||||
}
|
||||
}
|
||||
|
||||
HeapProfileTable::CleanupOldProfiles(fname);
|
||||
HeapProfileTable::CleanupOldProfiles(fname);
|
||||
|
||||
HeapProfilerStart(fname);
|
||||
HeapProfilerStart(fname);
|
||||
}
|
||||
|
||||
// class used for finalization -- dumps the heap-profile at program exit
|
||||
struct HeapProfileEndWriter {
|
||||
~HeapProfileEndWriter() {
|
||||
char buf[128];
|
||||
if (heap_profile) {
|
||||
const HeapProfileTable::Stats& total = heap_profile->total();
|
||||
const int64_t inuse_bytes = total.alloc_size - total.free_size;
|
||||
~HeapProfileEndWriter()
|
||||
{
|
||||
char buf[128];
|
||||
if (heap_profile) {
|
||||
const HeapProfileTable::Stats &total = heap_profile->total();
|
||||
const int64_t inuse_bytes = total.alloc_size - total.free_size;
|
||||
|
||||
if ((inuse_bytes >> 20) > 0) {
|
||||
snprintf(buf, sizeof(buf), ("Exiting, %" PRId64 " MB in use"),
|
||||
inuse_bytes >> 20);
|
||||
} else if ((inuse_bytes >> 10) > 0) {
|
||||
snprintf(buf, sizeof(buf), ("Exiting, %" PRId64 " kB in use"),
|
||||
inuse_bytes >> 10);
|
||||
} else {
|
||||
snprintf(buf, sizeof(buf), ("Exiting, %" PRId64 " bytes in use"),
|
||||
inuse_bytes);
|
||||
}
|
||||
} else {
|
||||
snprintf(buf, sizeof(buf), ("Exiting"));
|
||||
if ((inuse_bytes >> 20) > 0) {
|
||||
snprintf(buf, sizeof(buf), ("Exiting, %" PRId64 " MB in use"), inuse_bytes >> 20);
|
||||
} else if ((inuse_bytes >> 10) > 0) {
|
||||
snprintf(buf, sizeof(buf), ("Exiting, %" PRId64 " kB in use"), inuse_bytes >> 10);
|
||||
} else {
|
||||
snprintf(buf, sizeof(buf), ("Exiting, %" PRId64 " bytes in use"), inuse_bytes);
|
||||
}
|
||||
} else {
|
||||
snprintf(buf, sizeof(buf), ("Exiting"));
|
||||
}
|
||||
HeapProfilerDump(buf);
|
||||
}
|
||||
HeapProfilerDump(buf);
|
||||
}
|
||||
};
|
||||
|
||||
// We want to make sure tcmalloc is up and running before starting the profiler
|
||||
|
@ -18,6 +18,7 @@ option(SLED_BUILD_TESTS "Build tests" OFF)
|
||||
option(SLED_BUILD_FUZZ "Build fuzzer test" OFF)
|
||||
option(SLED_LOCATION_PATH "" "sled/src/sled/system/location.cc")
|
||||
option(SLED_BUILD_PROTOC_PLUGIN "Build protoc plugin" OFF)
|
||||
option(SLED_WITH_PROTOBUF "With Protobuf" ON)
|
||||
|
||||
set(BUILD_STATIC ON)
|
||||
set(BUILD_RTTR_DYNAMIC OFF)
|
||||
@ -31,10 +32,12 @@ target_include_directories(benchmark_main PUBLIC src/)
|
||||
add_library(sled STATIC "")
|
||||
|
||||
add_subdirectory(3party/minilua EXCLUDE_FROM_ALL)
|
||||
# add_subdirectory(3party/gperftools EXCLUDE_FROM_ALL)
|
||||
add_subdirectory(3party/gperftools EXCLUDE_FROM_ALL)
|
||||
add_subdirectory(3party/asyncplusplus EXCLUDE_FROM_ALL)
|
||||
# add_subdirectory(3party/cppuprofile EXCLUDE_FROM_ALL)
|
||||
add_subdirectory(3party/protobuf-3.21.12 EXCLUDE_FROM_ALL)
|
||||
if(SLED_WITH_PROTOBUF)
|
||||
add_subdirectory(3party/protobuf-3.21.12 EXCLUDE_FROM_ALL)
|
||||
endif()
|
||||
if(NOT TARGET marl)
|
||||
add_subdirectory(3party/marl EXCLUDE_FROM_ALL)
|
||||
endif()
|
||||
@ -109,15 +112,15 @@ target_link_libraries(
|
||||
protobuf::libprotobuf
|
||||
tcmalloc_and_profiler_static
|
||||
# protobuf::libprotoc
|
||||
PRIVATE dl
|
||||
# protobuf::libprotobuf ${WHOLE_ARCHIVE_WRAPPER_START}
|
||||
# tcmalloc_and_profiler_static ${WHOLE_ARCHIVE_WRAPPER_END}
|
||||
)
|
||||
PRIVATE dl)
|
||||
if(SLED_WITH_PROTOBUF)
|
||||
target_link_libraries(sled PUBLIC protobuf::libprotobuf)
|
||||
endif()
|
||||
|
||||
# set fPIC
|
||||
set_target_properties(sled PROPERTIES POSITION_INDEPENDENT_CODE ON)
|
||||
|
||||
if(SLED_BUILD_PROTOC_PLUGIN)
|
||||
if(SLED_WITH_PROTOBUF AND SLED_BUILD_PROTOC_PLUGIN)
|
||||
add_subdirectory(src/protoc_gen_sled)
|
||||
endif()
|
||||
|
||||
|
@ -14,23 +14,30 @@ DecrementFuturesUsage()
|
||||
|
||||
}// namespace future_detail
|
||||
|
||||
static ThreadPool g_default_thread_pool;
|
||||
TaskQueueBase *g_default_scheduler = &g_default_thread_pool;
|
||||
static std::atomic<TaskQueueBase *> g_default_scheduler{nullptr};
|
||||
// static ThreadPool default_thread_pool;
|
||||
static std::unique_ptr<Thread> g_default_thread;
|
||||
|
||||
void
|
||||
SetDefaultScheduler(TaskQueueBase *scheduler) noexcept
|
||||
{
|
||||
if (scheduler == nullptr) {
|
||||
g_default_scheduler = &g_default_thread_pool;
|
||||
} else {
|
||||
g_default_scheduler = scheduler;
|
||||
}
|
||||
SLED_ASSERT(scheduler, "scheduler is nullptr");
|
||||
g_default_scheduler.store(scheduler, std::memory_order_release);
|
||||
}
|
||||
|
||||
TaskQueueBase *
|
||||
GetDefaultScheduler() noexcept
|
||||
{
|
||||
return g_default_scheduler;
|
||||
static std::once_flag flag;
|
||||
std::call_once(flag, [&] {
|
||||
g_default_thread = sled::Thread::Create();
|
||||
g_default_thread->Start();
|
||||
TaskQueueBase *null_scheduler = nullptr;
|
||||
while (g_default_scheduler.load() == nullptr) {
|
||||
g_default_scheduler.compare_exchange_weak(null_scheduler, g_default_thread.get());
|
||||
}
|
||||
});
|
||||
return g_default_scheduler.load(std::memory_order_acquire);
|
||||
}
|
||||
|
||||
}// namespace sled
|
||||
|
@ -111,7 +111,7 @@ void
|
||||
SetLogFileName(const char *file_name)
|
||||
{
|
||||
g_log_file_name = file_name;
|
||||
g_log_stream.open(file_name);
|
||||
g_log_stream.open(file_name, std::ios_base::app);
|
||||
}
|
||||
|
||||
static std::atomic<uint32_t> g_current_id(0);
|
||||
|
@ -108,5 +108,7 @@
|
||||
|
||||
// testing
|
||||
#include "sled/testing/test.h"
|
||||
|
||||
// debugging
|
||||
|
||||
#endif// SLED_SLED_H
|
||||
|
Loading…
x
Reference in New Issue
Block a user