mirror of
https://github.com/chromium/crashpad.git
synced 2024-12-27 07:14:10 +08:00
c63c073d27
Include check_op.h directly, instead of relying on the transitive include from logging.h. This transitive include does not exist in Chromium's //base. Change-Id: I15962a9cdc26ac206032157b8d2659cf263ad695 Reviewed-on: https://chromium-review.googlesource.com/c/crashpad/crashpad/+/4950200 Reviewed-by: Mark Mentovai <mark@chromium.org> Commit-Queue: Lei Zhang <thestig@chromium.org>
774 lines
29 KiB
C++
774 lines
29 KiB
C++
// Copyright 2014 The Crashpad Authors
|
||
//
|
||
// Licensed under the Apache License, Version 2.0 (the "License");
|
||
// you may not use this file except in compliance with the License.
|
||
// You may obtain a copy of the License at
|
||
//
|
||
// http://www.apache.org/licenses/LICENSE-2.0
|
||
//
|
||
// Unless required by applicable law or agreed to in writing, software
|
||
// distributed under the License is distributed on an "AS IS" BASIS,
|
||
// WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
|
||
// See the License for the specific language governing permissions and
|
||
// limitations under the License.
|
||
|
||
#include "snapshot/mac/process_reader_mac.h"
|
||
|
||
#include <Availability.h>
|
||
#include <mach-o/loader.h>
|
||
#include <mach/mach_vm.h>
|
||
|
||
#include <algorithm>
|
||
#include <utility>
|
||
|
||
#include "base/apple/mach_logging.h"
|
||
#include "base/apple/scoped_mach_port.h"
|
||
#include "base/apple/scoped_mach_vm.h"
|
||
#include "base/check_op.h"
|
||
#include "base/logging.h"
|
||
#include "base/strings/stringprintf.h"
|
||
#include "snapshot/mac/mach_o_image_reader.h"
|
||
#include "snapshot/mac/process_types.h"
|
||
#include "util/misc/scoped_forbid_return.h"
|
||
|
||
namespace {
|
||
|
||
void MachTimeValueToTimeval(const time_value& mach, timeval* tv) {
|
||
tv->tv_sec = mach.seconds;
|
||
tv->tv_usec = mach.microseconds;
|
||
}
|
||
|
||
kern_return_t MachVMRegionRecurseDeepest(task_t task,
|
||
mach_vm_address_t* address,
|
||
mach_vm_size_t* size,
|
||
natural_t* depth,
|
||
vm_prot_t* protection,
|
||
unsigned int* user_tag) {
|
||
vm_region_submap_short_info_64 submap_info;
|
||
mach_msg_type_number_t count = VM_REGION_SUBMAP_SHORT_INFO_COUNT_64;
|
||
while (true) {
|
||
kern_return_t kr = mach_vm_region_recurse(
|
||
task,
|
||
address,
|
||
size,
|
||
depth,
|
||
reinterpret_cast<vm_region_recurse_info_t>(&submap_info),
|
||
&count);
|
||
if (kr != KERN_SUCCESS) {
|
||
return kr;
|
||
}
|
||
|
||
if (!submap_info.is_submap) {
|
||
*protection = submap_info.protection;
|
||
*user_tag = submap_info.user_tag;
|
||
return KERN_SUCCESS;
|
||
}
|
||
|
||
++*depth;
|
||
}
|
||
}
|
||
|
||
} // namespace
|
||
|
||
namespace crashpad {
|
||
|
||
ProcessReaderMac::Thread::Thread()
|
||
: thread_context(),
|
||
float_context(),
|
||
debug_context(),
|
||
name(),
|
||
id(0),
|
||
stack_region_address(0),
|
||
stack_region_size(0),
|
||
thread_specific_data_address(0),
|
||
port(THREAD_NULL),
|
||
suspend_count(0),
|
||
priority(0) {}
|
||
|
||
ProcessReaderMac::Module::Module() : name(), reader(nullptr), timestamp(0) {}
|
||
|
||
ProcessReaderMac::Module::~Module() {}
|
||
|
||
ProcessReaderMac::ProcessReaderMac()
|
||
: process_info_(),
|
||
threads_(),
|
||
modules_(),
|
||
module_readers_(),
|
||
process_memory_(),
|
||
task_(TASK_NULL),
|
||
initialized_(),
|
||
#if defined(CRASHPAD_MAC_32_BIT_SUPPORT)
|
||
is_64_bit_(false),
|
||
#endif // CRASHPAD_MAC_32_BIT_SUPPORT
|
||
initialized_threads_(false),
|
||
initialized_modules_(false) {
|
||
}
|
||
|
||
ProcessReaderMac::~ProcessReaderMac() {
|
||
for (const Thread& thread : threads_) {
|
||
kern_return_t kr = mach_port_deallocate(mach_task_self(), thread.port);
|
||
MACH_LOG_IF(ERROR, kr != KERN_SUCCESS, kr) << "mach_port_deallocate";
|
||
}
|
||
}
|
||
|
||
bool ProcessReaderMac::Initialize(task_t task) {
|
||
INITIALIZATION_STATE_SET_INITIALIZING(initialized_);
|
||
|
||
if (!process_info_.InitializeWithTask(task)) {
|
||
return false;
|
||
}
|
||
|
||
if (!process_memory_.Initialize(task)) {
|
||
return false;
|
||
}
|
||
|
||
#if defined(CRASHPAD_MAC_32_BIT_SUPPORT)
|
||
is_64_bit_ = process_info_.Is64Bit();
|
||
#else // CRASHPAD_MAC_32_BIT_SUPPORT
|
||
DCHECK(process_info_.Is64Bit());
|
||
#endif // CRASHPAD_MAC_32_BIT_SUPPORT
|
||
|
||
task_ = task;
|
||
|
||
INITIALIZATION_STATE_SET_VALID(initialized_);
|
||
return true;
|
||
}
|
||
|
||
void ProcessReaderMac::StartTime(timeval* start_time) const {
|
||
bool rv = process_info_.StartTime(start_time);
|
||
DCHECK(rv);
|
||
}
|
||
|
||
bool ProcessReaderMac::CPUTimes(timeval* user_time,
|
||
timeval* system_time) const {
|
||
INITIALIZATION_STATE_DCHECK_VALID(initialized_);
|
||
|
||
// Calculate user and system time the same way the kernel does for
|
||
// getrusage(). See 10.9.2 xnu-2422.90.20/bsd/kern/kern_resource.c calcru().
|
||
timerclear(user_time);
|
||
timerclear(system_time);
|
||
|
||
// As of the 10.8 SDK, the preferred routine is MACH_TASK_BASIC_INFO.
|
||
// TASK_BASIC_INFO_64 is equivalent and works on earlier systems.
|
||
task_basic_info_64 task_basic_info;
|
||
mach_msg_type_number_t task_basic_info_count = TASK_BASIC_INFO_64_COUNT;
|
||
kern_return_t kr = task_info(task_,
|
||
TASK_BASIC_INFO_64,
|
||
reinterpret_cast<task_info_t>(&task_basic_info),
|
||
&task_basic_info_count);
|
||
if (kr != KERN_SUCCESS) {
|
||
MACH_LOG(WARNING, kr) << "task_info TASK_BASIC_INFO_64";
|
||
return false;
|
||
}
|
||
|
||
task_thread_times_info_data_t task_thread_times;
|
||
mach_msg_type_number_t task_thread_times_count = TASK_THREAD_TIMES_INFO_COUNT;
|
||
kr = task_info(task_,
|
||
TASK_THREAD_TIMES_INFO,
|
||
reinterpret_cast<task_info_t>(&task_thread_times),
|
||
&task_thread_times_count);
|
||
if (kr != KERN_SUCCESS) {
|
||
MACH_LOG(WARNING, kr) << "task_info TASK_THREAD_TIMES";
|
||
return false;
|
||
}
|
||
|
||
MachTimeValueToTimeval(task_basic_info.user_time, user_time);
|
||
MachTimeValueToTimeval(task_basic_info.system_time, system_time);
|
||
|
||
timeval thread_user_time;
|
||
MachTimeValueToTimeval(task_thread_times.user_time, &thread_user_time);
|
||
timeval thread_system_time;
|
||
MachTimeValueToTimeval(task_thread_times.system_time, &thread_system_time);
|
||
|
||
timeradd(user_time, &thread_user_time, user_time);
|
||
timeradd(system_time, &thread_system_time, system_time);
|
||
|
||
return true;
|
||
}
|
||
|
||
const std::vector<ProcessReaderMac::Thread>& ProcessReaderMac::Threads() {
|
||
INITIALIZATION_STATE_DCHECK_VALID(initialized_);
|
||
|
||
if (!initialized_threads_) {
|
||
InitializeThreads();
|
||
}
|
||
|
||
return threads_;
|
||
}
|
||
|
||
const std::vector<ProcessReaderMac::Module>& ProcessReaderMac::Modules() {
|
||
INITIALIZATION_STATE_DCHECK_VALID(initialized_);
|
||
|
||
if (!initialized_modules_) {
|
||
InitializeModules();
|
||
}
|
||
|
||
return modules_;
|
||
}
|
||
|
||
mach_vm_address_t ProcessReaderMac::DyldAllImageInfo(
|
||
mach_vm_size_t* all_image_info_size) {
|
||
INITIALIZATION_STATE_DCHECK_VALID(initialized_);
|
||
|
||
task_dyld_info_data_t dyld_info;
|
||
mach_msg_type_number_t count = TASK_DYLD_INFO_COUNT;
|
||
kern_return_t kr = task_info(
|
||
task_, TASK_DYLD_INFO, reinterpret_cast<task_info_t>(&dyld_info), &count);
|
||
if (kr != KERN_SUCCESS) {
|
||
MACH_LOG(WARNING, kr) << "task_info";
|
||
return 0;
|
||
}
|
||
|
||
// TODO(mark): Deal with statically linked executables which don’t use dyld.
|
||
// This may look for the module that matches the executable path in the same
|
||
// data set that vmmap uses.
|
||
|
||
#if __MAC_OS_X_VERSION_MAX_ALLOWED >= __MAC_10_7
|
||
// The task_dyld_info_data_t struct grew in 10.7, adding the format field.
|
||
// Don’t check this field if it’s not present, which can happen when either
|
||
// the SDK used at compile time or the kernel at run time are too old and
|
||
// don’t know about it.
|
||
if (count >= TASK_DYLD_INFO_COUNT) {
|
||
const integer_t kExpectedFormat =
|
||
!Is64Bit() ? TASK_DYLD_ALL_IMAGE_INFO_32 : TASK_DYLD_ALL_IMAGE_INFO_64;
|
||
if (dyld_info.all_image_info_format != kExpectedFormat) {
|
||
LOG(WARNING) << "unexpected task_dyld_info_data_t::all_image_info_format "
|
||
<< dyld_info.all_image_info_format;
|
||
DCHECK_EQ(dyld_info.all_image_info_format, kExpectedFormat);
|
||
return 0;
|
||
}
|
||
}
|
||
#endif
|
||
|
||
if (all_image_info_size) {
|
||
*all_image_info_size = dyld_info.all_image_info_size;
|
||
}
|
||
return dyld_info.all_image_info_addr;
|
||
}
|
||
|
||
void ProcessReaderMac::InitializeThreads() {
|
||
DCHECK(!initialized_threads_);
|
||
DCHECK(threads_.empty());
|
||
|
||
initialized_threads_ = true;
|
||
|
||
thread_act_array_t threads;
|
||
mach_msg_type_number_t thread_count = 0;
|
||
kern_return_t kr = task_threads(task_, &threads, &thread_count);
|
||
if (kr != KERN_SUCCESS) {
|
||
MACH_LOG(WARNING, kr) << "task_threads";
|
||
return;
|
||
}
|
||
|
||
// The send rights in the |threads| array won’t have their send rights managed
|
||
// by anything until they’re added to |threads_| by the loop below. Any early
|
||
// return (or exception) that happens between here and the completion of the
|
||
// loop below will leak thread port send rights.
|
||
ScopedForbidReturn threads_need_owners;
|
||
|
||
base::apple::ScopedMachVM threads_vm(
|
||
reinterpret_cast<vm_address_t>(threads),
|
||
mach_vm_round_page(thread_count * sizeof(*threads)));
|
||
|
||
for (size_t index = 0; index < thread_count; ++index) {
|
||
Thread thread;
|
||
thread.port = threads[index];
|
||
|
||
#if defined(ARCH_CPU_X86_FAMILY)
|
||
const thread_state_flavor_t kThreadStateFlavor =
|
||
Is64Bit() ? x86_THREAD_STATE64 : x86_THREAD_STATE32;
|
||
mach_msg_type_number_t thread_state_count =
|
||
Is64Bit() ? x86_THREAD_STATE64_COUNT : x86_THREAD_STATE32_COUNT;
|
||
|
||
// TODO(mark): Use the AVX variants instead of the FLOAT variants?
|
||
const thread_state_flavor_t kFloatStateFlavor =
|
||
Is64Bit() ? x86_FLOAT_STATE64 : x86_FLOAT_STATE32;
|
||
mach_msg_type_number_t float_state_count =
|
||
Is64Bit() ? x86_FLOAT_STATE64_COUNT : x86_FLOAT_STATE32_COUNT;
|
||
|
||
const thread_state_flavor_t kDebugStateFlavor =
|
||
Is64Bit() ? x86_DEBUG_STATE64 : x86_DEBUG_STATE32;
|
||
mach_msg_type_number_t debug_state_count =
|
||
Is64Bit() ? x86_DEBUG_STATE64_COUNT : x86_DEBUG_STATE32_COUNT;
|
||
#elif defined(ARCH_CPU_ARM64)
|
||
const thread_state_flavor_t kThreadStateFlavor = ARM_THREAD_STATE64;
|
||
mach_msg_type_number_t thread_state_count = ARM_THREAD_STATE64_COUNT;
|
||
|
||
const thread_state_flavor_t kFloatStateFlavor = ARM_NEON_STATE64;
|
||
mach_msg_type_number_t float_state_count = ARM_NEON_STATE64_COUNT;
|
||
|
||
const thread_state_flavor_t kDebugStateFlavor = ARM_DEBUG_STATE64;
|
||
mach_msg_type_number_t debug_state_count = ARM_DEBUG_STATE64_COUNT;
|
||
#endif
|
||
|
||
kr = thread_get_state(
|
||
thread.port,
|
||
kThreadStateFlavor,
|
||
reinterpret_cast<thread_state_t>(&thread.thread_context),
|
||
&thread_state_count);
|
||
if (kr != KERN_SUCCESS) {
|
||
MACH_LOG(ERROR, kr) << "thread_get_state(" << kThreadStateFlavor << ")";
|
||
continue;
|
||
}
|
||
|
||
kr = thread_get_state(
|
||
thread.port,
|
||
kFloatStateFlavor,
|
||
reinterpret_cast<thread_state_t>(&thread.float_context),
|
||
&float_state_count);
|
||
if (kr != KERN_SUCCESS) {
|
||
MACH_LOG(ERROR, kr) << "thread_get_state(" << kFloatStateFlavor << ")";
|
||
continue;
|
||
}
|
||
|
||
kr = thread_get_state(
|
||
thread.port,
|
||
kDebugStateFlavor,
|
||
reinterpret_cast<thread_state_t>(&thread.debug_context),
|
||
&debug_state_count);
|
||
if (kr != KERN_SUCCESS) {
|
||
MACH_LOG(ERROR, kr) << "thread_get_state(" << kDebugStateFlavor << ")";
|
||
continue;
|
||
}
|
||
|
||
thread_basic_info basic_info;
|
||
mach_msg_type_number_t count = THREAD_BASIC_INFO_COUNT;
|
||
kr = thread_info(thread.port,
|
||
THREAD_BASIC_INFO,
|
||
reinterpret_cast<thread_info_t>(&basic_info),
|
||
&count);
|
||
if (kr != KERN_SUCCESS) {
|
||
MACH_LOG(WARNING, kr) << "thread_info(THREAD_BASIC_INFO)";
|
||
} else {
|
||
thread.suspend_count = basic_info.suspend_count;
|
||
}
|
||
|
||
thread_identifier_info identifier_info;
|
||
count = THREAD_IDENTIFIER_INFO_COUNT;
|
||
kr = thread_info(thread.port,
|
||
THREAD_IDENTIFIER_INFO,
|
||
reinterpret_cast<thread_info_t>(&identifier_info),
|
||
&count);
|
||
if (kr != KERN_SUCCESS) {
|
||
MACH_LOG(WARNING, kr) << "thread_info(THREAD_IDENTIFIER_INFO)";
|
||
} else {
|
||
thread.id = identifier_info.thread_id;
|
||
|
||
// thread_identifier_info::thread_handle contains the base of the
|
||
// thread-specific data area, which on x86 and x86_64 is the thread’s base
|
||
// address of the %gs segment. 10.9.2 xnu-2422.90.20/osfmk/kern/thread.c
|
||
// thread_info_internal() gets the value from
|
||
// machine_thread::cthread_self, which is the same value used to set the
|
||
// %gs base in xnu-2422.90.20/osfmk/i386/pcb_native.c
|
||
// act_machine_switch_pcb().
|
||
//
|
||
// This address is the internal pthread’s _pthread::tsd[], an array of
|
||
// void* values that can be indexed by pthread_key_t values.
|
||
thread.thread_specific_data_address = identifier_info.thread_handle;
|
||
}
|
||
|
||
thread_extended_info extended_info;
|
||
count = THREAD_EXTENDED_INFO_COUNT;
|
||
kr = thread_info(thread.port,
|
||
THREAD_EXTENDED_INFO,
|
||
reinterpret_cast<thread_info_t>(&extended_info),
|
||
&count);
|
||
if (kr != KERN_SUCCESS) {
|
||
MACH_LOG(WARNING, kr) << "thread_info(THREAD_EXTENDED_INFO)";
|
||
} else {
|
||
thread.name.assign(
|
||
extended_info.pth_name,
|
||
strnlen(extended_info.pth_name, sizeof(extended_info.pth_name)));
|
||
}
|
||
|
||
thread_precedence_policy precedence;
|
||
count = THREAD_PRECEDENCE_POLICY_COUNT;
|
||
boolean_t get_default = FALSE;
|
||
kr = thread_policy_get(thread.port,
|
||
THREAD_PRECEDENCE_POLICY,
|
||
reinterpret_cast<thread_policy_t>(&precedence),
|
||
&count,
|
||
&get_default);
|
||
if (kr != KERN_SUCCESS) {
|
||
MACH_LOG(INFO, kr) << "thread_policy_get";
|
||
} else {
|
||
thread.priority = precedence.importance;
|
||
}
|
||
|
||
#if defined(ARCH_CPU_X86_FAMILY)
|
||
mach_vm_address_t stack_pointer = Is64Bit()
|
||
? thread.thread_context.t64.__rsp
|
||
: thread.thread_context.t32.__esp;
|
||
#elif defined(ARCH_CPU_ARM64)
|
||
mach_vm_address_t stack_pointer =
|
||
arm_thread_state64_get_sp(thread.thread_context);
|
||
#endif
|
||
|
||
thread.stack_region_address =
|
||
CalculateStackRegion(stack_pointer, &thread.stack_region_size);
|
||
|
||
threads_.push_back(thread);
|
||
}
|
||
|
||
threads_need_owners.Disarm();
|
||
}
|
||
|
||
void ProcessReaderMac::InitializeModules() {
|
||
DCHECK(!initialized_modules_);
|
||
DCHECK(modules_.empty());
|
||
|
||
initialized_modules_ = true;
|
||
|
||
mach_vm_size_t all_image_info_size;
|
||
mach_vm_address_t all_image_info_addr =
|
||
DyldAllImageInfo(&all_image_info_size);
|
||
|
||
process_types::dyld_all_image_infos all_image_infos;
|
||
if (!all_image_infos.Read(this, all_image_info_addr)) {
|
||
LOG(WARNING) << "could not read dyld_all_image_infos";
|
||
return;
|
||
}
|
||
|
||
if (all_image_infos.version < 1) {
|
||
LOG(WARNING) << "unexpected dyld_all_image_infos version "
|
||
<< all_image_infos.version;
|
||
return;
|
||
}
|
||
|
||
size_t expected_size =
|
||
process_types::dyld_all_image_infos::ExpectedSizeForVersion(
|
||
this, all_image_infos.version);
|
||
if (all_image_info_size < expected_size) {
|
||
LOG(WARNING) << "small dyld_all_image_infos size " << all_image_info_size
|
||
<< " < " << expected_size << " for version "
|
||
<< all_image_infos.version;
|
||
return;
|
||
}
|
||
|
||
// Note that all_image_infos.infoArrayCount may be 0 if a crash occurred while
|
||
// dyld was loading the executable. This can happen if a required dynamic
|
||
// library was not found. Similarly, all_image_infos.infoArray may be nullptr
|
||
// if a crash occurred while dyld was updating it.
|
||
//
|
||
// TODO(mark): It may be possible to recover from these situations by looking
|
||
// through memory mappings for Mach-O images.
|
||
//
|
||
// Continue along when this situation is detected, because even without any
|
||
// images in infoArray, dyldImageLoadAddress may be set, and it may be
|
||
// possible to recover some information from dyld.
|
||
if (all_image_infos.infoArrayCount == 0) {
|
||
LOG(WARNING) << "all_image_infos.infoArrayCount is zero";
|
||
} else if (!all_image_infos.infoArray) {
|
||
LOG(WARNING) << "all_image_infos.infoArray is nullptr";
|
||
}
|
||
|
||
std::vector<process_types::dyld_image_info> image_info_vector(
|
||
all_image_infos.infoArrayCount);
|
||
if (!process_types::dyld_image_info::ReadArrayInto(this,
|
||
all_image_infos.infoArray,
|
||
image_info_vector.size(),
|
||
&image_info_vector[0])) {
|
||
LOG(WARNING) << "could not read dyld_image_info array";
|
||
return;
|
||
}
|
||
|
||
size_t main_executable_count = 0;
|
||
bool found_dyld = false;
|
||
modules_.reserve(image_info_vector.size());
|
||
for (const process_types::dyld_image_info& image_info : image_info_vector) {
|
||
Module module;
|
||
module.timestamp = image_info.imageFileModDate;
|
||
|
||
if (!process_memory_.ReadCString(image_info.imageFilePath, &module.name)) {
|
||
LOG(WARNING) << "could not read dyld_image_info::imageFilePath";
|
||
// Proceed anyway with an empty module name.
|
||
}
|
||
|
||
std::unique_ptr<MachOImageReader> reader(new MachOImageReader());
|
||
if (!reader->Initialize(this, image_info.imageLoadAddress, module.name)) {
|
||
reader.reset();
|
||
}
|
||
|
||
module.reader = reader.get();
|
||
|
||
uint32_t file_type = reader ? reader->FileType() : 0;
|
||
|
||
module_readers_.push_back(std::move(reader));
|
||
modules_.push_back(module);
|
||
|
||
if (all_image_infos.version >= 2 && all_image_infos.dyldImageLoadAddress &&
|
||
image_info.imageLoadAddress == all_image_infos.dyldImageLoadAddress) {
|
||
found_dyld = true;
|
||
LOG(WARNING) << base::StringPrintf(
|
||
"found dylinker (%s) in dyld_all_image_infos::infoArray",
|
||
module.name.c_str());
|
||
|
||
LOG_IF(WARNING, file_type != MH_DYLINKER)
|
||
<< base::StringPrintf("dylinker (%s) has unexpected Mach-O type %d",
|
||
module.name.c_str(),
|
||
file_type);
|
||
}
|
||
|
||
if (file_type == MH_EXECUTE) {
|
||
// On macOS 14, the main executable does not normally show up at
|
||
// index 0. In previous versions of dyld, each loaded image was
|
||
// appended to the all image info vector as it was loaded.
|
||
// (For example, see RuntimeState::notifyDebuggerLoad in dyld-1066.8).
|
||
// Starting from dyld-1122.1, notifyDebuggerLoad calls
|
||
// ExternallyViewableState::addImages for all but the main executable
|
||
// (which has already been added). ExternallyViewableState::addImages
|
||
// inserts all new image infos at the front of the vector, leaving the
|
||
// main executable as the last item.
|
||
//
|
||
// The interface requires that the main executable be first in the list,
|
||
// so swap it into the right position.
|
||
size_t index = modules_.size() - 1;
|
||
if (index > 0) {
|
||
CHECK_EQ(index, image_info_vector.size() - 1);
|
||
if (main_executable_count == 0) {
|
||
std::rotate(
|
||
modules_.rbegin(), modules_.rbegin() + 1, modules_.rend());
|
||
}
|
||
}
|
||
if (main_executable_count > 0) {
|
||
LOG(WARNING) << base::StringPrintf(
|
||
"multiple MH_EXECUTE modules (%s, %s)",
|
||
modules_[0].name.c_str(),
|
||
modules_[index].name.c_str());
|
||
}
|
||
++main_executable_count;
|
||
}
|
||
}
|
||
|
||
LOG_IF(WARNING, main_executable_count == 0) << "no MH_EXECUTE modules";
|
||
|
||
// all_image_infos.infoArray doesn’t include an entry for dyld, but dyld is
|
||
// loaded into the process’ address space as a module. Its load address is
|
||
// easily known given a sufficiently recent all_image_infos.version, but the
|
||
// timestamp and pathname are not given as they are for other modules.
|
||
//
|
||
// The timestamp is a lost cause, because the kernel doesn’t record the
|
||
// timestamp of the dynamic linker at the time it’s loaded in the same way
|
||
// that dyld records the timestamps of other modules when they’re loaded. (The
|
||
// timestamp for the main executable is also not reported and appears as 0
|
||
// even when accessed via dyld APIs, because it’s loaded by the kernel, not by
|
||
// dyld.)
|
||
//
|
||
// The name can be determined, but it’s not as simple as hardcoding the
|
||
// default "/usr/lib/dyld" because an executable could have specified anything
|
||
// in its LC_LOAD_DYLINKER command.
|
||
if (!found_dyld && all_image_infos.version >= 2 &&
|
||
all_image_infos.dyldImageLoadAddress) {
|
||
Module module;
|
||
module.timestamp = 0;
|
||
|
||
// Examine the executable’s LC_LOAD_DYLINKER load command to find the path
|
||
// used to load dyld.
|
||
if (all_image_infos.infoArrayCount >= 1 && main_executable_count >= 1) {
|
||
module.name = modules_[0].reader->DylinkerName();
|
||
}
|
||
std::string module_name = !module.name.empty() ? module.name : "(dyld)";
|
||
|
||
std::unique_ptr<MachOImageReader> reader(new MachOImageReader());
|
||
if (!reader->Initialize(
|
||
this, all_image_infos.dyldImageLoadAddress, module_name)) {
|
||
reader.reset();
|
||
}
|
||
|
||
module.reader = reader.get();
|
||
|
||
uint32_t file_type = reader ? reader->FileType() : 0;
|
||
|
||
LOG_IF(WARNING, file_type != MH_DYLINKER)
|
||
<< base::StringPrintf("dylinker (%s) has unexpected Mach-O type %d",
|
||
module.name.c_str(),
|
||
file_type);
|
||
|
||
if (module.name.empty() && file_type == MH_DYLINKER) {
|
||
// Look inside dyld directly to find its preferred path.
|
||
module.name = reader->DylinkerName();
|
||
}
|
||
|
||
if (module.name.empty()) {
|
||
module.name = "(dyld)";
|
||
}
|
||
|
||
// dyld is loaded in the process even if its path can’t be determined.
|
||
module_readers_.push_back(std::move(reader));
|
||
modules_.push_back(module);
|
||
}
|
||
}
|
||
|
||
mach_vm_address_t ProcessReaderMac::CalculateStackRegion(
|
||
mach_vm_address_t stack_pointer,
|
||
mach_vm_size_t* stack_region_size) {
|
||
INITIALIZATION_STATE_DCHECK_VALID(initialized_);
|
||
|
||
// For pthreads, it may be possible to compute the stack region based on the
|
||
// internal _pthread::stackaddr and _pthread::stacksize. The _pthread struct
|
||
// for a thread can be located at TSD slot 0, or the known offsets of
|
||
// stackaddr and stacksize from the TSD area could be used.
|
||
mach_vm_address_t region_base = stack_pointer;
|
||
mach_vm_size_t region_size;
|
||
natural_t depth = 0;
|
||
vm_prot_t protection;
|
||
unsigned int user_tag;
|
||
kern_return_t kr = MachVMRegionRecurseDeepest(
|
||
task_, ®ion_base, ®ion_size, &depth, &protection, &user_tag);
|
||
if (kr != KERN_SUCCESS) {
|
||
MACH_LOG(INFO, kr) << "mach_vm_region_recurse";
|
||
*stack_region_size = 0;
|
||
return 0;
|
||
}
|
||
|
||
if (region_base > stack_pointer) {
|
||
// There’s nothing mapped at the stack pointer’s address. Something may have
|
||
// trashed the stack pointer. Note that this shouldn’t happen for a normal
|
||
// stack guard region violation because the guard region is mapped but has
|
||
// VM_PROT_NONE protection.
|
||
*stack_region_size = 0;
|
||
return 0;
|
||
}
|
||
|
||
mach_vm_address_t start_address = stack_pointer;
|
||
|
||
if ((protection & VM_PROT_READ) == 0) {
|
||
// If the region isn’t readable, the stack pointer probably points to the
|
||
// guard region. Don’t include it as part of the stack, and don’t include
|
||
// anything at any lower memory address. The code below may still possibly
|
||
// find the real stack region at a memory address higher than this region.
|
||
start_address = region_base + region_size;
|
||
} else {
|
||
// If the ABI requires a red zone, adjust the region to include it if
|
||
// possible.
|
||
LocateRedZone(&start_address, ®ion_base, ®ion_size, user_tag);
|
||
|
||
// Regardless of whether the ABI requires a red zone, capture up to
|
||
// kExtraCaptureSize additional bytes of stack, but only if present in the
|
||
// region that was already found.
|
||
constexpr mach_vm_size_t kExtraCaptureSize = 128;
|
||
start_address = std::max(start_address >= kExtraCaptureSize
|
||
? start_address - kExtraCaptureSize
|
||
: start_address,
|
||
region_base);
|
||
|
||
// Align start_address to a 16-byte boundary, which can help readers by
|
||
// ensuring that data is aligned properly. This could page-align instead,
|
||
// but that might be wasteful.
|
||
constexpr mach_vm_size_t kDesiredAlignment = 16;
|
||
start_address &= ~(kDesiredAlignment - 1);
|
||
DCHECK_GE(start_address, region_base);
|
||
}
|
||
|
||
region_size -= (start_address - region_base);
|
||
region_base = start_address;
|
||
|
||
mach_vm_size_t total_region_size = region_size;
|
||
|
||
// The stack region may have gotten split up into multiple abutting regions.
|
||
// Try to coalesce them. This frequently happens for the main thread’s stack
|
||
// when setrlimit(RLIMIT_STACK, …) is called. It may also happen if a region
|
||
// is split up due to an mprotect() or vm_protect() call.
|
||
//
|
||
// Stack regions created by the kernel and the pthreads library will be marked
|
||
// with the VM_MEMORY_STACK user tag. Scanning for multiple adjacent regions
|
||
// with the same tag should find an entire stack region. Checking that the
|
||
// protection on individual regions is not VM_PROT_NONE should guarantee that
|
||
// this algorithm doesn’t collect map entries belonging to another thread’s
|
||
// stack: well-behaved stacks (such as those created by the kernel and the
|
||
// pthreads library) have VM_PROT_NONE guard regions at their low-address
|
||
// ends.
|
||
//
|
||
// Other stack regions may not be so well-behaved and thus if user_tag is not
|
||
// VM_MEMORY_STACK, the single region that was found is used as-is without
|
||
// trying to merge it with other adjacent regions.
|
||
if (user_tag == VM_MEMORY_STACK) {
|
||
mach_vm_address_t try_address = region_base;
|
||
mach_vm_address_t original_try_address;
|
||
|
||
while (try_address += region_size,
|
||
original_try_address = try_address,
|
||
(kr = MachVMRegionRecurseDeepest(task_,
|
||
&try_address,
|
||
®ion_size,
|
||
&depth,
|
||
&protection,
|
||
&user_tag) == KERN_SUCCESS) &&
|
||
try_address == original_try_address &&
|
||
(protection & VM_PROT_READ) != 0 &&
|
||
user_tag == VM_MEMORY_STACK) {
|
||
total_region_size += region_size;
|
||
}
|
||
|
||
if (kr != KERN_SUCCESS && kr != KERN_INVALID_ADDRESS) {
|
||
// Tolerate KERN_INVALID_ADDRESS because it will be returned when there
|
||
// are no more regions in the map at or above the specified |try_address|.
|
||
MACH_LOG(INFO, kr) << "mach_vm_region_recurse";
|
||
}
|
||
}
|
||
|
||
*stack_region_size = total_region_size;
|
||
return region_base;
|
||
}
|
||
|
||
void ProcessReaderMac::LocateRedZone(mach_vm_address_t* const start_address,
|
||
mach_vm_address_t* const region_base,
|
||
mach_vm_address_t* const region_size,
|
||
const unsigned int user_tag) {
|
||
#if defined(ARCH_CPU_X86_FAMILY)
|
||
if (Is64Bit()) {
|
||
// x86_64 has a red zone. See AMD64 ABI 0.99.8,
|
||
// https://gitlab.com/x86-psABIs/x86-64-ABI/-/wikis/uploads/01de35b2c8adc7545de52604cc45d942/x86-64-psABI-2021-05-20.pdf#page=23.
|
||
// section 3.2.2, “The Stack Frame”.
|
||
constexpr mach_vm_size_t kRedZoneSize = 128;
|
||
mach_vm_address_t red_zone_base =
|
||
*start_address >= kRedZoneSize ? *start_address - kRedZoneSize : 0;
|
||
bool red_zone_ok = false;
|
||
if (red_zone_base >= *region_base) {
|
||
// The red zone is within the region already discovered.
|
||
red_zone_ok = true;
|
||
} else if (red_zone_base < *region_base && user_tag == VM_MEMORY_STACK) {
|
||
// Probe to see if there’s a region immediately below the one already
|
||
// discovered.
|
||
mach_vm_address_t red_zone_region_base = red_zone_base;
|
||
mach_vm_size_t red_zone_region_size;
|
||
natural_t red_zone_depth = 0;
|
||
vm_prot_t red_zone_protection;
|
||
unsigned int red_zone_user_tag;
|
||
kern_return_t kr = MachVMRegionRecurseDeepest(task_,
|
||
&red_zone_region_base,
|
||
&red_zone_region_size,
|
||
&red_zone_depth,
|
||
&red_zone_protection,
|
||
&red_zone_user_tag);
|
||
if (kr != KERN_SUCCESS) {
|
||
MACH_LOG(INFO, kr) << "mach_vm_region_recurse";
|
||
*start_address = *region_base;
|
||
} else if (red_zone_region_base + red_zone_region_size == *region_base &&
|
||
(red_zone_protection & VM_PROT_READ) != 0 &&
|
||
red_zone_user_tag == user_tag) {
|
||
// The region containing the red zone is immediately below the region
|
||
// already found, it’s readable (not the guard region), and it has the
|
||
// same user tag as the region already found, so merge them.
|
||
red_zone_ok = true;
|
||
*region_base -= red_zone_region_size;
|
||
*region_size += red_zone_region_size;
|
||
}
|
||
}
|
||
|
||
if (red_zone_ok) {
|
||
// Begin capturing from the base of the red zone (but not the entire
|
||
// region that encompasses the red zone).
|
||
*start_address = red_zone_base;
|
||
} else {
|
||
// The red zone would go lower into another region in memory, but no
|
||
// region was found. Memory can only be captured to an address as low as
|
||
// the base address of the region already found.
|
||
*start_address = *region_base;
|
||
}
|
||
}
|
||
#endif
|
||
}
|
||
|
||
} // namespace crashpad
|