2022-12-14 14:02:29 -07:00
|
|
|
// Copyright 2022 The Crashpad Authors
|
|
|
|
//
|
|
|
|
// Licensed under the Apache License, Version 2.0 (the "License");
|
|
|
|
// you may not use this file except in compliance with the License.
|
|
|
|
// You may obtain a copy of the License at
|
|
|
|
//
|
|
|
|
// http://www.apache.org/licenses/LICENSE-2.0
|
|
|
|
//
|
|
|
|
// Unless required by applicable law or agreed to in writing, software
|
|
|
|
// distributed under the License is distributed on an "AS IS" BASIS,
|
|
|
|
// WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
|
|
|
|
// See the License for the specific language governing permissions and
|
|
|
|
// limitations under the License.
|
|
|
|
|
|
|
|
#ifndef CRASHPAD_UTIL_SYNCHRONIZATION_SCOPED_SPIN_GUARD_H_
|
|
|
|
#define CRASHPAD_UTIL_SYNCHRONIZATION_SCOPED_SPIN_GUARD_H_
|
|
|
|
|
|
|
|
#include <atomic>
|
|
|
|
#include <optional>
|
|
|
|
#include <utility>
|
|
|
|
|
|
|
|
#include "base/check.h"
|
|
|
|
#include "base/notreached.h"
|
|
|
|
#include "util/misc/clock.h"
|
|
|
|
|
|
|
|
namespace crashpad {
|
|
|
|
|
|
|
|
//! \brief Spinlock state for `ScopedSpinGuard`.
|
|
|
|
struct SpinGuardState final {
|
|
|
|
//! \brief A `ScopedSpinGuard` in an unlocked state.
|
|
|
|
constexpr SpinGuardState() : locked(false) {}
|
|
|
|
|
|
|
|
SpinGuardState(const SpinGuardState&) = delete;
|
|
|
|
SpinGuardState& operator=(const SpinGuardState&) = delete;
|
|
|
|
|
|
|
|
//! \brief `true` if the `ScopedSpinGuard` is locked, `false` otherwise.
|
|
|
|
std::atomic<bool> locked;
|
|
|
|
static_assert(std::atomic<bool>::is_always_lock_free,
|
|
|
|
"std::atomic<bool> may not be signal-safe");
|
|
|
|
static_assert(sizeof(std::atomic<bool>) == sizeof(bool),
|
|
|
|
"std::atomic<bool> adds size to bool");
|
|
|
|
};
|
|
|
|
|
|
|
|
//! \brief A scoped mutual-exclusion guard wrapping a `SpinGuardState` with RAII
|
|
|
|
//! semantics.
|
|
|
|
class ScopedSpinGuard final {
|
|
|
|
//! \brief The duration in nanoseconds between attempts to lock the spinlock.
|
|
|
|
static constexpr uint64_t kSpinGuardSleepTimeNanos = 10;
|
|
|
|
|
|
|
|
public:
|
|
|
|
ScopedSpinGuard(const ScopedSpinGuard&) = delete;
|
|
|
|
ScopedSpinGuard& operator=(const ScopedSpinGuard&) = delete;
|
|
|
|
ScopedSpinGuard(ScopedSpinGuard&& other) noexcept : state_(nullptr) {
|
|
|
|
std::swap(state_, other.state_);
|
|
|
|
}
|
|
|
|
ScopedSpinGuard& operator=(ScopedSpinGuard&& other) {
|
|
|
|
std::swap(state_, other.state_);
|
|
|
|
return *this;
|
|
|
|
}
|
|
|
|
|
|
|
|
//! \brief Spins up to `timeout_nanos` nanoseconds trying to lock `state`.
|
|
|
|
//! \param[in] timeout_nanos The timeout in nanoseconds after which this gives
|
|
|
|
//! up trying to lock the spinlock and returns `std::nullopt`.
|
|
|
|
//! \param[in,out] state The spinlock state to attempt to lock. This method
|
|
|
|
//! holds a pointer to `state`, so `state` must outlive the lifetime of
|
|
|
|
//! this object.
|
|
|
|
//! \return The locked `ScopedSpinGuard` on success, or `std::nullopt` on
|
|
|
|
//! timeout.
|
|
|
|
static std::optional<ScopedSpinGuard> TryCreateScopedSpinGuard(
|
|
|
|
uint64_t timeout_nanos,
|
|
|
|
SpinGuardState& state) {
|
|
|
|
const uint64_t clock_end_time_nanos =
|
|
|
|
ClockMonotonicNanoseconds() + timeout_nanos;
|
|
|
|
while (true) {
|
|
|
|
bool expected_current_value = false;
|
[ScopedSpinGuard] Use std::atomic::compare_exchange_strong() for spinlock
Previously, ScopedSpinGuard used std::atomic::compare_exchange_weak()
in a loop to implement a spinlock. After looping for the specified
number of nanoseconds, it would give up and return an error.
A few bugs have come in on ARM platforms (https://crbug.com/340980960,
http://b/296082201) which indicate that this can fail even in
single-threaded cases where nothing else has the spinlock.
From https://cbloomrants.blogspot.com/2011/07/07-14-11-compareexchangestrong-vs.html :
> compare_exchange_weak exists for LL-SC (load linked/store
> conditional) type architectures (Power, ARM, basically everything
> except x86), because on them compare_exchange_strong must be
> implemented as a loop, while compare_exchange_weak can be
> non-looping.
and:
https://en.cppreference.com/w/cpp/atomic/atomic/compare_exchange#Notes
> compare_exchange_weak is allowed to fail spuriously, that is, acts
> as if *this != expected even if they are equal. When a
> compare-and-exchange is in a loop, compare_exchange_weak will yield
> better performance on some platforms.
>
> When compare_exchange_weak would require a loop and
> compare_exchange_strong would not, compare_exchange_strong is
> preferable [...]
My conclusion is that this logic needs to use
`compare_exchange_strong` to avoid spurious failures on ARM in the
common case when there's no other thread holding the spinlock.
Change-Id: I2a08031db6b219d7d14a5cd02b3634985f81ab06
Bug: b:340980960
Change-Id: I2a08031db6b219d7d14a5cd02b3634985f81ab06
Reviewed-on: https://chromium-review.googlesource.com/c/crashpad/crashpad/+/5545257
Reviewed-by: Mark Mentovai <mark@chromium.org>
Commit-Queue: Ben Hamilton <benhamilton@google.com>
2024-05-16 10:11:56 -06:00
|
|
|
// `std::atomic::compare_exchange_weak()` is allowed to spuriously fail on
|
|
|
|
// architectures like ARM, which can cause timeouts even for
|
|
|
|
// single-threaded cases (https://crbug.com/340980960,
|
|
|
|
// http://b/296082201).
|
|
|
|
//
|
|
|
|
// Use `std::compare_exchange_strong()` instead to avoid spurious failures
|
|
|
|
// in the common case.
|
|
|
|
if (state.locked.compare_exchange_strong(expected_current_value,
|
|
|
|
true,
|
|
|
|
std::memory_order_acquire,
|
|
|
|
std::memory_order_relaxed)) {
|
2022-12-14 14:02:29 -07:00
|
|
|
return std::make_optional<ScopedSpinGuard>(state);
|
|
|
|
}
|
|
|
|
if (ClockMonotonicNanoseconds() >= clock_end_time_nanos) {
|
|
|
|
return std::nullopt;
|
|
|
|
}
|
|
|
|
SleepNanoseconds(kSpinGuardSleepTimeNanos);
|
|
|
|
}
|
|
|
|
|
2024-08-14 13:09:43 +10:00
|
|
|
NOTREACHED();
|
2022-12-14 14:02:29 -07:00
|
|
|
}
|
|
|
|
|
|
|
|
~ScopedSpinGuard() {
|
|
|
|
if (state_) {
|
|
|
|
#ifdef NDEBUG
|
|
|
|
state_->locked.store(false, std::memory_order_release);
|
|
|
|
#else
|
|
|
|
bool old = state_->locked.exchange(false, std::memory_order_release);
|
|
|
|
DCHECK(old);
|
|
|
|
#endif
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
//! \brief A `ScopedSpinGuard` wrapping a locked `SpinGuardState`.
|
|
|
|
//! \param[in,out] locked_state A locked `SpinGuardState`. This method
|
|
|
|
//! holds a pointer to `state`, so `state` must outlive the lifetime of
|
|
|
|
//! this object.
|
|
|
|
ScopedSpinGuard(SpinGuardState& locked_state) : state_(&locked_state) {
|
|
|
|
DCHECK(locked_state.locked);
|
|
|
|
}
|
|
|
|
|
|
|
|
private:
|
|
|
|
// \brief Optional spinlock state, unlocked when this object goes out of
|
|
|
|
// scope.
|
|
|
|
//
|
|
|
|
// If this is `nullptr`, then this object has been moved from, and the state
|
|
|
|
// is no longer valid. In that case, nothing will be unlocked when this object
|
|
|
|
// is destroyed.
|
|
|
|
SpinGuardState* state_;
|
|
|
|
};
|
|
|
|
|
|
|
|
} // namespace crashpad
|
|
|
|
|
|
|
|
#endif // CRASHPAD_UTIL_SYNCHRONIZATION_SCOPED_SPIN_GUARD_H_
|