mirror of
https://github.com/chromium/crashpad.git
synced 2024-12-26 06:31:50 +08:00
c63c073d27
Include check_op.h directly, instead of relying on the transitive include from logging.h. This transitive include does not exist in Chromium's //base. Change-Id: I15962a9cdc26ac206032157b8d2659cf263ad695 Reviewed-on: https://chromium-review.googlesource.com/c/crashpad/crashpad/+/4950200 Reviewed-by: Mark Mentovai <mark@chromium.org> Commit-Queue: Lei Zhang <thestig@chromium.org>
863 lines
26 KiB
C++
863 lines
26 KiB
C++
// Copyright 2017 The Crashpad Authors
|
|
//
|
|
// Licensed under the Apache License, Version 2.0 (the "License");
|
|
// you may not use this file except in compliance with the License.
|
|
// You may obtain a copy of the License at
|
|
//
|
|
// http://www.apache.org/licenses/LICENSE-2.0
|
|
//
|
|
// Unless required by applicable law or agreed to in writing, software
|
|
// distributed under the License is distributed on an "AS IS" BASIS,
|
|
// WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
|
|
// See the License for the specific language governing permissions and
|
|
// limitations under the License.
|
|
|
|
#include "snapshot/elf/elf_image_reader.h"
|
|
|
|
#include <stddef.h>
|
|
|
|
#include <algorithm>
|
|
#include <limits>
|
|
#include <utility>
|
|
#include <vector>
|
|
|
|
#include "base/check_op.h"
|
|
#include "base/logging.h"
|
|
#include "base/numerics/safe_math.h"
|
|
#include "build/build_config.h"
|
|
#include "util/numeric/checked_vm_address_range.h"
|
|
|
|
namespace crashpad {
|
|
|
|
class ElfImageReader::ProgramHeaderTable {
|
|
public:
|
|
virtual ~ProgramHeaderTable() {}
|
|
|
|
virtual bool VerifyLoadSegments(bool verbose) const = 0;
|
|
virtual size_t Size() const = 0;
|
|
virtual bool GetDynamicSegment(VMAddress* address, VMSize* size) const = 0;
|
|
virtual bool GetPreferredElfHeaderAddress(VMAddress* address,
|
|
bool verbose) const = 0;
|
|
virtual bool GetPreferredLoadedMemoryRange(VMAddress* address,
|
|
VMSize* size,
|
|
bool verbose) const = 0;
|
|
|
|
// Locate the next PT_NOTE segment starting at segment index start_index. If a
|
|
// PT_NOTE segment is found, start_index is set to the next index after the
|
|
// found segment.
|
|
virtual bool GetNoteSegment(size_t* start_index,
|
|
VMAddress* address,
|
|
VMSize* size) const = 0;
|
|
|
|
protected:
|
|
ProgramHeaderTable() {}
|
|
};
|
|
|
|
template <typename PhdrType>
|
|
class ElfImageReader::ProgramHeaderTableSpecific
|
|
: public ElfImageReader::ProgramHeaderTable {
|
|
public:
|
|
ProgramHeaderTableSpecific() {}
|
|
|
|
ProgramHeaderTableSpecific(
|
|
const ProgramHeaderTableSpecific<PhdrType>&) = delete;
|
|
ProgramHeaderTableSpecific<PhdrType>& operator=(
|
|
const ProgramHeaderTableSpecific<PhdrType>&) = delete;
|
|
|
|
~ProgramHeaderTableSpecific() {}
|
|
|
|
bool Initialize(const ProcessMemoryRange& memory,
|
|
VMAddress address,
|
|
VMSize num_segments,
|
|
bool verbose) {
|
|
INITIALIZATION_STATE_SET_INITIALIZING(initialized_);
|
|
table_.resize(num_segments);
|
|
if (!memory.Read(address, sizeof(PhdrType) * num_segments, table_.data())) {
|
|
return false;
|
|
}
|
|
|
|
if (!VerifyLoadSegments(verbose)) {
|
|
return false;
|
|
}
|
|
|
|
INITIALIZATION_STATE_SET_VALID(initialized_);
|
|
return true;
|
|
}
|
|
|
|
bool VerifyLoadSegments(bool verbose) const override {
|
|
constexpr bool is_64_bit = std::is_same<PhdrType, Elf64_Phdr>::value;
|
|
VMAddress last_vaddr;
|
|
bool load_found = false;
|
|
for (const auto& header : table_) {
|
|
if (header.p_type == PT_LOAD) {
|
|
CheckedVMAddressRange load_range(
|
|
is_64_bit, header.p_vaddr, header.p_memsz);
|
|
|
|
if (!load_range.IsValid()) {
|
|
LOG_IF(ERROR, verbose) << "bad load range";
|
|
return false;
|
|
}
|
|
|
|
if (load_found && header.p_vaddr <= last_vaddr) {
|
|
LOG_IF(ERROR, verbose) << "out of order load segments";
|
|
return false;
|
|
}
|
|
load_found = true;
|
|
last_vaddr = header.p_vaddr;
|
|
}
|
|
}
|
|
return true;
|
|
}
|
|
|
|
size_t Size() const override { return sizeof(PhdrType) * table_.size(); }
|
|
|
|
bool GetPreferredElfHeaderAddress(VMAddress* address,
|
|
bool verbose) const override {
|
|
INITIALIZATION_STATE_DCHECK_VALID(initialized_);
|
|
for (const auto& header : table_) {
|
|
if (header.p_type == PT_LOAD && header.p_offset == 0) {
|
|
*address = header.p_vaddr;
|
|
return true;
|
|
}
|
|
}
|
|
LOG_IF(ERROR, verbose) << "no preferred header address";
|
|
return false;
|
|
}
|
|
|
|
bool GetPreferredLoadedMemoryRange(VMAddress* base,
|
|
VMSize* size,
|
|
bool verbose) const override {
|
|
INITIALIZATION_STATE_DCHECK_VALID(initialized_);
|
|
|
|
VMAddress preferred_base = 0;
|
|
VMAddress preferred_end = 0;
|
|
bool load_found = false;
|
|
for (const auto& header : table_) {
|
|
if (header.p_type == PT_LOAD) {
|
|
if (!load_found) {
|
|
preferred_base = header.p_vaddr;
|
|
load_found = true;
|
|
}
|
|
preferred_end = header.p_vaddr + header.p_memsz;
|
|
}
|
|
}
|
|
if (load_found) {
|
|
*base = preferred_base;
|
|
*size = preferred_end - preferred_base;
|
|
return true;
|
|
}
|
|
LOG_IF(ERROR, verbose) << "no load segments";
|
|
return false;
|
|
}
|
|
|
|
bool GetDynamicSegment(VMAddress* address, VMSize* size) const override {
|
|
INITIALIZATION_STATE_DCHECK_VALID(initialized_);
|
|
const PhdrType* phdr;
|
|
if (!GetProgramHeader(PT_DYNAMIC, &phdr)) {
|
|
return false;
|
|
}
|
|
*address = phdr->p_vaddr;
|
|
*size = phdr->p_memsz;
|
|
return true;
|
|
}
|
|
|
|
bool GetProgramHeader(uint32_t type, const PhdrType** header_out) const {
|
|
INITIALIZATION_STATE_DCHECK_VALID(initialized_);
|
|
for (const auto& header : table_) {
|
|
if (header.p_type == type) {
|
|
*header_out = &header;
|
|
return true;
|
|
}
|
|
}
|
|
return false;
|
|
}
|
|
|
|
bool GetNoteSegment(size_t* start_index,
|
|
VMAddress* address,
|
|
VMSize* size) const override {
|
|
INITIALIZATION_STATE_DCHECK_VALID(initialized_);
|
|
for (size_t index = *start_index; index < table_.size(); ++index) {
|
|
if (table_[index].p_type == PT_NOTE && table_[index].p_vaddr != 0) {
|
|
*start_index = index + 1;
|
|
*address = table_[index].p_vaddr;
|
|
*size = table_[index].p_memsz;
|
|
return true;
|
|
}
|
|
}
|
|
return false;
|
|
}
|
|
|
|
private:
|
|
std::vector<PhdrType> table_;
|
|
InitializationStateDcheck initialized_;
|
|
};
|
|
|
|
ElfImageReader::NoteReader::~NoteReader() = default;
|
|
|
|
ElfImageReader::NoteReader::Result ElfImageReader::NoteReader::NextNote(
|
|
std::string* name,
|
|
NoteType* type,
|
|
std::string* desc,
|
|
VMAddress* desc_address) {
|
|
if (!is_valid_) {
|
|
LOG(ERROR) << "invalid note reader";
|
|
return Result::kError;
|
|
}
|
|
|
|
Result result = Result::kError;
|
|
do {
|
|
while (current_address_ == segment_end_address_) {
|
|
VMSize segment_size;
|
|
if (!phdr_table_->GetNoteSegment(
|
|
&phdr_index_, ¤t_address_, &segment_size)) {
|
|
return Result::kNoMoreNotes;
|
|
}
|
|
current_address_ += elf_reader_->GetLoadBias();
|
|
segment_end_address_ = current_address_ + segment_size;
|
|
segment_range_ = std::make_unique<ProcessMemoryRange>();
|
|
if (!segment_range_->Initialize(*range_) ||
|
|
!segment_range_->RestrictRange(current_address_, segment_size)) {
|
|
return Result::kError;
|
|
}
|
|
}
|
|
|
|
retry_ = false;
|
|
result = range_->Is64Bit()
|
|
? ReadNote<Elf64_Nhdr>(name, type, desc, desc_address)
|
|
: ReadNote<Elf32_Nhdr>(name, type, desc, desc_address);
|
|
} while (retry_);
|
|
|
|
if (result == Result::kSuccess) {
|
|
return Result::kSuccess;
|
|
}
|
|
is_valid_ = false;
|
|
return Result::kError;
|
|
}
|
|
|
|
namespace {
|
|
|
|
// The maximum size the user can specify for maximum note size. Clamping this
|
|
// ensures that buffer allocations cannot be wildly large. It is not expected
|
|
// that a note would be larger than ~1k in normal usage.
|
|
constexpr size_t kMaxMaxNoteSize = 16384;
|
|
|
|
} // namespace
|
|
|
|
ElfImageReader::NoteReader::NoteReader(const ElfImageReader* elf_reader,
|
|
const ProcessMemoryRange* range,
|
|
const ProgramHeaderTable* phdr_table,
|
|
size_t max_note_size,
|
|
const std::string& name_filter,
|
|
NoteType type_filter,
|
|
bool use_filter)
|
|
: current_address_(0),
|
|
segment_end_address_(0),
|
|
elf_reader_(elf_reader),
|
|
range_(range),
|
|
phdr_table_(phdr_table),
|
|
segment_range_(),
|
|
phdr_index_(0),
|
|
max_note_size_(std::min(kMaxMaxNoteSize, max_note_size)),
|
|
name_filter_(name_filter),
|
|
type_filter_(type_filter),
|
|
use_filter_(use_filter),
|
|
is_valid_(true),
|
|
retry_(false) {
|
|
DCHECK_LT(max_note_size, kMaxMaxNoteSize);
|
|
}
|
|
|
|
template <typename NhdrType>
|
|
ElfImageReader::NoteReader::Result ElfImageReader::NoteReader::ReadNote(
|
|
std::string* name,
|
|
NoteType* type,
|
|
std::string* desc,
|
|
VMAddress* desc_address) {
|
|
static_assert(sizeof(*type) >= sizeof(NhdrType::n_namesz),
|
|
"Note field size mismatch");
|
|
DCHECK_LT(current_address_, segment_end_address_);
|
|
|
|
NhdrType note_info;
|
|
if (!segment_range_->Read(current_address_, sizeof(note_info), ¬e_info)) {
|
|
return Result::kError;
|
|
}
|
|
current_address_ += sizeof(note_info);
|
|
|
|
constexpr size_t align = sizeof(note_info.n_namesz);
|
|
|
|
#define CHECKED_PAD(x, into) \
|
|
base::CheckAnd(base::CheckAdd(x, align - 1), ~(align - 1)) \
|
|
.AssignIfValid(&into)
|
|
|
|
size_t padded_namesz;
|
|
if (!CHECKED_PAD(note_info.n_namesz, padded_namesz)) {
|
|
return Result::kError;
|
|
}
|
|
size_t padded_descsz;
|
|
if (!CHECKED_PAD(note_info.n_descsz, padded_descsz)) {
|
|
return Result::kError;
|
|
}
|
|
|
|
size_t note_size;
|
|
if (!base::CheckAdd(padded_namesz, padded_descsz).AssignIfValid(¬e_size)) {
|
|
return Result::kError;
|
|
}
|
|
|
|
// Notes typically have 4-byte alignment. However, .note.android.ident may
|
|
// inadvertently use 2-byte alignment.
|
|
// https://android-review.googlesource.com/c/platform/bionic/+/554986/
|
|
// We can still find .note.android.ident if it appears first in a note segment
|
|
// but there may be 4-byte aligned notes following it. If this note was
|
|
// aligned at less than 4-bytes, expect that the next note will be aligned at
|
|
// 4-bytes and add extra padding, if necessary.
|
|
|
|
VMAddress end_of_note_candidate;
|
|
if (!base::CheckAdd(current_address_, note_size)
|
|
.AssignIfValid(&end_of_note_candidate)) {
|
|
return Result::kError;
|
|
}
|
|
VMAddress end_of_note;
|
|
if (!CHECKED_PAD(end_of_note_candidate, end_of_note)) {
|
|
return Result::kError;
|
|
}
|
|
end_of_note = std::min(end_of_note, segment_end_address_);
|
|
|
|
#undef CHECKED_PAD
|
|
|
|
if (note_size > max_note_size_) {
|
|
current_address_ = end_of_note;
|
|
retry_ = true;
|
|
return Result::kError;
|
|
}
|
|
|
|
if (use_filter_ && note_info.n_type != type_filter_) {
|
|
current_address_ = end_of_note;
|
|
retry_ = true;
|
|
return Result::kError;
|
|
}
|
|
|
|
std::string local_name(note_info.n_namesz, '\0');
|
|
if (!segment_range_->Read(
|
|
current_address_, note_info.n_namesz, &local_name[0])) {
|
|
return Result::kError;
|
|
}
|
|
if (!local_name.empty()) {
|
|
if (local_name.back() != '\0') {
|
|
LOG(ERROR) << "unterminated note name";
|
|
return Result::kError;
|
|
}
|
|
local_name.pop_back();
|
|
}
|
|
|
|
if (use_filter_ && local_name != name_filter_) {
|
|
current_address_ = end_of_note;
|
|
retry_ = true;
|
|
return Result::kError;
|
|
}
|
|
|
|
current_address_ += padded_namesz;
|
|
|
|
std::string local_desc(note_info.n_descsz, '\0');
|
|
if (!segment_range_->Read(
|
|
current_address_, note_info.n_descsz, &local_desc[0])) {
|
|
return Result::kError;
|
|
}
|
|
*desc_address = current_address_;
|
|
|
|
current_address_ = end_of_note;
|
|
|
|
if (name) {
|
|
name->swap(local_name);
|
|
}
|
|
if (type) {
|
|
*type = note_info.n_type;
|
|
}
|
|
desc->swap(local_desc);
|
|
return Result::kSuccess;
|
|
}
|
|
|
|
ElfImageReader::ElfImageReader()
|
|
: header_64_(),
|
|
ehdr_address_(0),
|
|
load_bias_(0),
|
|
memory_(),
|
|
program_headers_(),
|
|
dynamic_array_(),
|
|
symbol_table_(),
|
|
initialized_(),
|
|
dynamic_array_initialized_(),
|
|
symbol_table_initialized_() {}
|
|
|
|
ElfImageReader::~ElfImageReader() {}
|
|
|
|
bool ElfImageReader::Initialize(const ProcessMemoryRange& memory,
|
|
VMAddress address,
|
|
bool verbose) {
|
|
INITIALIZATION_STATE_SET_INITIALIZING(initialized_);
|
|
ehdr_address_ = address;
|
|
if (!memory_.Initialize(memory)) {
|
|
return false;
|
|
}
|
|
|
|
uint8_t e_ident[EI_NIDENT];
|
|
if (!memory_.Read(ehdr_address_, EI_NIDENT, e_ident)) {
|
|
return false;
|
|
}
|
|
|
|
if (e_ident[EI_MAG0] != ELFMAG0 || e_ident[EI_MAG1] != ELFMAG1 ||
|
|
e_ident[EI_MAG2] != ELFMAG2 || e_ident[EI_MAG3] != ELFMAG3) {
|
|
LOG_IF(ERROR, verbose) << "Incorrect ELF magic number";
|
|
return false;
|
|
}
|
|
|
|
if (!(memory_.Is64Bit() && e_ident[EI_CLASS] == ELFCLASS64) &&
|
|
!(!memory_.Is64Bit() && e_ident[EI_CLASS] == ELFCLASS32)) {
|
|
LOG_IF(ERROR, verbose) << "unexpected bitness";
|
|
return false;
|
|
}
|
|
|
|
#if defined(ARCH_CPU_LITTLE_ENDIAN)
|
|
constexpr uint8_t expected_encoding = ELFDATA2LSB;
|
|
#elif defined(ARCH_CPU_BIG_ENDIAN)
|
|
constexpr uint8_t expected_encoding = ELFDATA2MSB;
|
|
#endif
|
|
if (e_ident[EI_DATA] != expected_encoding) {
|
|
LOG_IF(ERROR, verbose) << "unexpected encoding";
|
|
return false;
|
|
}
|
|
|
|
if (e_ident[EI_VERSION] != EV_CURRENT) {
|
|
LOG_IF(ERROR, verbose) << "unexpected version";
|
|
return false;
|
|
}
|
|
|
|
if (!(memory_.Is64Bit()
|
|
? memory_.Read(ehdr_address_, sizeof(header_64_), &header_64_)
|
|
: memory_.Read(ehdr_address_, sizeof(header_32_), &header_32_))) {
|
|
return false;
|
|
}
|
|
|
|
#define VERIFY_HEADER(header) \
|
|
do { \
|
|
if (header.e_type != ET_EXEC && header.e_type != ET_DYN) { \
|
|
LOG_IF(ERROR, verbose) << "unexpected image type"; \
|
|
return false; \
|
|
} \
|
|
if (header.e_version != EV_CURRENT) { \
|
|
LOG_IF(ERROR, verbose) << "unexpected version"; \
|
|
return false; \
|
|
} \
|
|
if (header.e_ehsize != sizeof(header)) { \
|
|
LOG_IF(ERROR, verbose) << "unexpected header size"; \
|
|
return false; \
|
|
} \
|
|
} while (false);
|
|
|
|
if (memory_.Is64Bit()) {
|
|
VERIFY_HEADER(header_64_);
|
|
} else {
|
|
VERIFY_HEADER(header_32_);
|
|
}
|
|
|
|
if (!InitializeProgramHeaders(verbose)) {
|
|
return false;
|
|
}
|
|
|
|
VMAddress preferred_ehdr_address;
|
|
if (!program_headers_.get()->GetPreferredElfHeaderAddress(
|
|
&preferred_ehdr_address, verbose)) {
|
|
return false;
|
|
}
|
|
load_bias_ = ehdr_address_ - preferred_ehdr_address;
|
|
|
|
VMAddress base_address;
|
|
VMSize loaded_size;
|
|
if (!program_headers_.get()->GetPreferredLoadedMemoryRange(
|
|
&base_address, &loaded_size, verbose)) {
|
|
return false;
|
|
}
|
|
base_address += load_bias_;
|
|
|
|
if (!memory_.RestrictRange(base_address, loaded_size)) {
|
|
return false;
|
|
}
|
|
|
|
VMSize ehdr_size;
|
|
VMAddress phdr_address;
|
|
if (memory_.Is64Bit()) {
|
|
ehdr_size = sizeof(header_64_);
|
|
phdr_address = ehdr_address_ + header_64_.e_phoff;
|
|
} else {
|
|
ehdr_size = sizeof(header_32_);
|
|
phdr_address = ehdr_address_ + header_32_.e_phoff;
|
|
}
|
|
|
|
CheckedVMAddressRange range(memory_.Is64Bit(), base_address, loaded_size);
|
|
if (!range.ContainsRange(
|
|
CheckedVMAddressRange(memory_.Is64Bit(), ehdr_address_, ehdr_size))) {
|
|
LOG_IF(ERROR, verbose) << "ehdr out of range";
|
|
return false;
|
|
}
|
|
if (!range.ContainsRange(CheckedVMAddressRange(
|
|
memory.Is64Bit(), phdr_address, program_headers_->Size()))) {
|
|
LOG_IF(ERROR, verbose) << "phdrs out of range";
|
|
return false;
|
|
}
|
|
|
|
INITIALIZATION_STATE_SET_VALID(initialized_);
|
|
return true;
|
|
}
|
|
|
|
uint16_t ElfImageReader::FileType() const {
|
|
INITIALIZATION_STATE_DCHECK_VALID(initialized_);
|
|
return memory_.Is64Bit() ? header_64_.e_type : header_32_.e_type;
|
|
}
|
|
|
|
bool ElfImageReader::SoName(std::string* name) {
|
|
INITIALIZATION_STATE_DCHECK_VALID(initialized_);
|
|
if (!InitializeDynamicArray()) {
|
|
return false;
|
|
}
|
|
|
|
VMSize offset;
|
|
if (!dynamic_array_->GetValue(DT_SONAME, true, &offset)) {
|
|
return false;
|
|
}
|
|
|
|
return ReadDynamicStringTableAtOffset(offset, name);
|
|
}
|
|
|
|
bool ElfImageReader::GetDynamicSymbol(const std::string& name,
|
|
VMAddress* address,
|
|
VMSize* size) {
|
|
INITIALIZATION_STATE_DCHECK_VALID(initialized_);
|
|
if (!InitializeDynamicSymbolTable()) {
|
|
return false;
|
|
}
|
|
|
|
ElfSymbolTableReader::SymbolInformation info;
|
|
if (!symbol_table_->GetSymbol(name, &info)) {
|
|
return false;
|
|
}
|
|
if (info.shndx == SHN_UNDEF || info.shndx == SHN_COMMON) {
|
|
return false;
|
|
}
|
|
|
|
switch (info.binding) {
|
|
case STB_GLOBAL:
|
|
case STB_WEAK:
|
|
break;
|
|
|
|
case STB_LOCAL:
|
|
default:
|
|
return false;
|
|
}
|
|
|
|
switch (info.type) {
|
|
case STT_OBJECT:
|
|
case STT_FUNC:
|
|
break;
|
|
|
|
case STT_COMMON:
|
|
case STT_NOTYPE:
|
|
case STT_SECTION:
|
|
case STT_FILE:
|
|
case STT_TLS:
|
|
default:
|
|
return false;
|
|
}
|
|
|
|
if (info.shndx != SHN_ABS) {
|
|
info.address += GetLoadBias();
|
|
}
|
|
|
|
*address = info.address;
|
|
*size = info.size;
|
|
return true;
|
|
}
|
|
|
|
bool ElfImageReader::ReadDynamicStringTableAtOffset(VMSize offset,
|
|
std::string* string) {
|
|
INITIALIZATION_STATE_DCHECK_VALID(initialized_);
|
|
if (!InitializeDynamicArray()) {
|
|
return false;
|
|
}
|
|
|
|
VMAddress string_table_address;
|
|
VMSize string_table_size;
|
|
if (!GetAddressFromDynamicArray(DT_STRTAB, true, &string_table_address) ||
|
|
!dynamic_array_->GetValue(DT_STRSZ, true, &string_table_size)) {
|
|
LOG(ERROR) << "missing string table info";
|
|
return false;
|
|
}
|
|
if (offset >= string_table_size) {
|
|
LOG(ERROR) << "bad offset";
|
|
return false;
|
|
}
|
|
|
|
// GNU ld.so doesn't adjust the vdso's dynamic array entries by the load bias.
|
|
// If the address is too small to point into the loaded module range and is
|
|
// small enough to be an offset from the base of the module, adjust it now.
|
|
if (string_table_address < memory_.Base() &&
|
|
string_table_address < memory_.Size()) {
|
|
string_table_address += GetLoadBias();
|
|
}
|
|
|
|
if (!memory_.ReadCStringSizeLimited(
|
|
string_table_address + offset, string_table_size - offset, string)) {
|
|
LOG(ERROR) << "missing nul-terminator";
|
|
return false;
|
|
}
|
|
return true;
|
|
}
|
|
|
|
bool ElfImageReader::GetDebugAddress(VMAddress* debug) {
|
|
INITIALIZATION_STATE_DCHECK_VALID(initialized_);
|
|
if (!InitializeDynamicArray()) {
|
|
return false;
|
|
}
|
|
return GetAddressFromDynamicArray(DT_DEBUG, true, debug);
|
|
}
|
|
|
|
bool ElfImageReader::GetDynamicArrayAddress(VMAddress* address) {
|
|
INITIALIZATION_STATE_DCHECK_VALID(initialized_);
|
|
VMAddress dyn_segment_address;
|
|
VMSize dyn_segment_size;
|
|
if (!program_headers_.get()->GetDynamicSegment(&dyn_segment_address,
|
|
&dyn_segment_size)) {
|
|
LOG(ERROR) << "no dynamic segment";
|
|
return false;
|
|
}
|
|
*address = dyn_segment_address + GetLoadBias();
|
|
return true;
|
|
}
|
|
|
|
VMAddress ElfImageReader::GetProgramHeaderTableAddress() {
|
|
INITIALIZATION_STATE_DCHECK_VALID(initialized_);
|
|
return ehdr_address_ +
|
|
(memory_.Is64Bit() ? header_64_.e_phoff : header_32_.e_phoff);
|
|
}
|
|
|
|
bool ElfImageReader::InitializeProgramHeaders(bool verbose) {
|
|
#define INITIALIZE_PROGRAM_HEADERS(PhdrType, header) \
|
|
do { \
|
|
if (header.e_phentsize != sizeof(PhdrType)) { \
|
|
LOG_IF(ERROR, verbose) << "unexpected phdr size"; \
|
|
return false; \
|
|
} \
|
|
auto phdrs = new ProgramHeaderTableSpecific<PhdrType>(); \
|
|
program_headers_.reset(phdrs); \
|
|
if (!phdrs->Initialize(memory_, \
|
|
ehdr_address_ + header.e_phoff, \
|
|
header.e_phnum, \
|
|
verbose)) { \
|
|
return false; \
|
|
} \
|
|
} while (false);
|
|
|
|
if (memory_.Is64Bit()) {
|
|
INITIALIZE_PROGRAM_HEADERS(Elf64_Phdr, header_64_);
|
|
} else {
|
|
INITIALIZE_PROGRAM_HEADERS(Elf32_Phdr, header_32_);
|
|
}
|
|
return true;
|
|
}
|
|
|
|
bool ElfImageReader::InitializeDynamicArray() {
|
|
if (dynamic_array_initialized_.is_valid()) {
|
|
return true;
|
|
}
|
|
if (!dynamic_array_initialized_.is_uninitialized()) {
|
|
return false;
|
|
}
|
|
dynamic_array_initialized_.set_invalid();
|
|
|
|
VMAddress dyn_segment_address;
|
|
VMSize dyn_segment_size;
|
|
if (!program_headers_.get()->GetDynamicSegment(&dyn_segment_address,
|
|
&dyn_segment_size)) {
|
|
LOG(ERROR) << "no dynamic segment";
|
|
return false;
|
|
}
|
|
dyn_segment_address += GetLoadBias();
|
|
|
|
dynamic_array_.reset(new ElfDynamicArrayReader());
|
|
if (!dynamic_array_->Initialize(
|
|
memory_, dyn_segment_address, dyn_segment_size)) {
|
|
return false;
|
|
}
|
|
dynamic_array_initialized_.set_valid();
|
|
return true;
|
|
}
|
|
|
|
bool ElfImageReader::InitializeDynamicSymbolTable() {
|
|
if (symbol_table_initialized_.is_valid()) {
|
|
return true;
|
|
}
|
|
if (!symbol_table_initialized_.is_uninitialized()) {
|
|
return false;
|
|
}
|
|
symbol_table_initialized_.set_invalid();
|
|
|
|
if (!InitializeDynamicArray()) {
|
|
return false;
|
|
}
|
|
|
|
VMAddress symbol_table_address;
|
|
if (!GetAddressFromDynamicArray(DT_SYMTAB, true, &symbol_table_address)) {
|
|
LOG(ERROR) << "no symbol table";
|
|
return false;
|
|
}
|
|
|
|
// Try both DT_HASH and DT_GNU_HASH. They're completely different, but both
|
|
// circuitously offer a way to find the number of entries in the symbol table.
|
|
// DT_HASH is specifically checked first, because depending on the linker, the
|
|
// count maybe be incorrect for zero-export cases. In practice, it is believed
|
|
// that the zero-export case is probably not particularly useful, so this
|
|
// incorrect count will only occur in constructed test cases (see
|
|
// ElfImageReader.DtHashAndDtGnuHashMatch).
|
|
VMSize number_of_symbol_table_entries;
|
|
if (!GetNumberOfSymbolEntriesFromDtHash(&number_of_symbol_table_entries) &&
|
|
!GetNumberOfSymbolEntriesFromDtGnuHash(&number_of_symbol_table_entries)) {
|
|
LOG(ERROR) << "could not retrieve number of symbol table entries";
|
|
return false;
|
|
}
|
|
|
|
symbol_table_.reset(new ElfSymbolTableReader(
|
|
&memory_, this, symbol_table_address, number_of_symbol_table_entries));
|
|
symbol_table_initialized_.set_valid();
|
|
return true;
|
|
}
|
|
|
|
bool ElfImageReader::GetAddressFromDynamicArray(uint64_t tag,
|
|
bool log,
|
|
VMAddress* address) {
|
|
if (!dynamic_array_->GetValue(tag, log, address)) {
|
|
return false;
|
|
}
|
|
|
|
#if BUILDFLAG(IS_ANDROID) || BUILDFLAG(IS_FUCHSIA) || \
|
|
(defined(__GLIBC__) && defined(ARCH_CPU_RISCV64))
|
|
// The GNU loader updates the dynamic array according to the load bias (except
|
|
// for RISC-V: https://sourceware.org/bugzilla/show_bug.cgi?id=24484).
|
|
// The Android and Fuchsia loaders only update the debug address.
|
|
if (tag != DT_DEBUG) {
|
|
*address += GetLoadBias();
|
|
}
|
|
#endif // BUILDFLAG(IS_ANDROID)
|
|
return true;
|
|
}
|
|
|
|
bool ElfImageReader::GetNumberOfSymbolEntriesFromDtHash(
|
|
VMSize* number_of_symbol_table_entries) {
|
|
if (!InitializeDynamicArray()) {
|
|
return false;
|
|
}
|
|
|
|
VMAddress dt_hash_address;
|
|
if (!GetAddressFromDynamicArray(DT_HASH, false, &dt_hash_address)) {
|
|
return false;
|
|
}
|
|
|
|
struct {
|
|
uint32_t nbucket;
|
|
uint32_t nchain;
|
|
} header;
|
|
|
|
if (!memory_.Read(dt_hash_address, sizeof(header), &header)) {
|
|
LOG(ERROR) << "failed to read DT_HASH header";
|
|
return false;
|
|
}
|
|
|
|
*number_of_symbol_table_entries = header.nchain;
|
|
return true;
|
|
}
|
|
|
|
bool ElfImageReader::GetNumberOfSymbolEntriesFromDtGnuHash(
|
|
VMSize* number_of_symbol_table_entries) {
|
|
if (!InitializeDynamicArray()) {
|
|
return false;
|
|
}
|
|
|
|
VMAddress dt_gnu_hash_address;
|
|
if (!GetAddressFromDynamicArray(DT_GNU_HASH, false, &dt_gnu_hash_address)) {
|
|
return false;
|
|
}
|
|
|
|
// See https://flapenguin.me/2017/05/10/elf-lookup-dt-gnu-hash/ and
|
|
// https://sourceware.org/ml/binutils/2006-10/msg00377.html.
|
|
struct {
|
|
uint32_t nbuckets;
|
|
uint32_t symoffset;
|
|
uint32_t bloom_size;
|
|
uint32_t bloom_shift;
|
|
} header;
|
|
if (!memory_.Read(dt_gnu_hash_address, sizeof(header), &header)) {
|
|
LOG(ERROR) << "failed to read DT_GNU_HASH header";
|
|
return false;
|
|
}
|
|
|
|
std::vector<uint32_t> buckets(header.nbuckets);
|
|
const size_t kNumBytesForBuckets = sizeof(buckets[0]) * buckets.size();
|
|
const size_t kWordSize =
|
|
memory_.Is64Bit() ? sizeof(uint64_t) : sizeof(uint32_t);
|
|
const VMAddress buckets_address =
|
|
dt_gnu_hash_address + sizeof(header) + (kWordSize * header.bloom_size);
|
|
if (!memory_.Read(buckets_address, kNumBytesForBuckets, buckets.data())) {
|
|
LOG(ERROR) << "read buckets";
|
|
return false;
|
|
}
|
|
|
|
// Locate the chain that handles the largest index bucket.
|
|
uint32_t last_symbol = 0;
|
|
for (uint32_t i = 0; i < header.nbuckets; ++i) {
|
|
last_symbol = std::max(buckets[i], last_symbol);
|
|
}
|
|
|
|
if (last_symbol < header.symoffset) {
|
|
*number_of_symbol_table_entries = header.symoffset;
|
|
return true;
|
|
}
|
|
|
|
// Walk the bucket's chain to add the chain length to the total.
|
|
const VMAddress chains_base_address = buckets_address + kNumBytesForBuckets;
|
|
for (;;) {
|
|
uint32_t chain_entry;
|
|
if (!memory_.Read(chains_base_address + (last_symbol - header.symoffset) *
|
|
sizeof(chain_entry),
|
|
sizeof(chain_entry),
|
|
&chain_entry)) {
|
|
LOG(ERROR) << "read chain entry";
|
|
return false;
|
|
}
|
|
|
|
++last_symbol;
|
|
|
|
// If the low bit is set, this entry is the end of the chain.
|
|
if (chain_entry & 1)
|
|
break;
|
|
}
|
|
|
|
*number_of_symbol_table_entries = last_symbol;
|
|
return true;
|
|
}
|
|
|
|
std::unique_ptr<ElfImageReader::NoteReader> ElfImageReader::Notes(
|
|
size_t max_note_size) {
|
|
return std::make_unique<NoteReader>(
|
|
this, &memory_, program_headers_.get(), max_note_size);
|
|
}
|
|
|
|
std::unique_ptr<ElfImageReader::NoteReader>
|
|
ElfImageReader::NotesWithNameAndType(const std::string& name,
|
|
NoteReader::NoteType type,
|
|
size_t max_note_size) {
|
|
return std::make_unique<NoteReader>(
|
|
this, &memory_, program_headers_.get(), max_note_size, name, type, true);
|
|
}
|
|
|
|
const ProcessMemoryRange* ElfImageReader::Memory() const {
|
|
return &memory_;
|
|
}
|
|
|
|
} // namespace crashpad
|