2020-12-26 19:36:25 +00:00
|
|
|
//! Helper module that wraps some Mutex types with different implementations.
|
2021-09-25 03:26:45 +00:00
|
|
|
|
2020-12-26 19:36:25 +00:00
|
|
|
// ----------------------------------------------------------------------------
|
|
|
|
|
2022-03-22 14:34:21 +00:00
|
|
|
#[cfg(not(target_arch = "wasm32"))]
|
2021-05-26 20:13:24 +00:00
|
|
|
#[cfg(not(debug_assertions))]
|
2021-10-20 20:14:16 +00:00
|
|
|
mod mutex_impl {
|
2022-03-21 21:20:37 +00:00
|
|
|
/// Provides interior mutability.
|
2022-03-22 14:34:21 +00:00
|
|
|
///
|
|
|
|
/// Uses `parking_lot` crate on native targets, and `atomic_refcell` on `wasm32` targets.
|
2021-10-20 20:14:16 +00:00
|
|
|
#[derive(Default)]
|
|
|
|
pub struct Mutex<T>(parking_lot::Mutex<T>);
|
|
|
|
|
|
|
|
/// The lock you get from [`Mutex`].
|
|
|
|
pub use parking_lot::MutexGuard;
|
|
|
|
|
|
|
|
impl<T> Mutex<T> {
|
|
|
|
#[inline(always)]
|
|
|
|
pub fn new(val: T) -> Self {
|
|
|
|
Self(parking_lot::Mutex::new(val))
|
|
|
|
}
|
|
|
|
|
|
|
|
#[inline(always)]
|
|
|
|
pub fn lock(&self) -> MutexGuard<'_, T> {
|
|
|
|
self.0.lock()
|
|
|
|
}
|
|
|
|
}
|
|
|
|
}
|
2020-10-13 22:29:11 +00:00
|
|
|
|
2022-03-22 14:34:21 +00:00
|
|
|
#[cfg(not(target_arch = "wasm32"))]
|
2021-05-26 20:13:24 +00:00
|
|
|
#[cfg(debug_assertions)]
|
2021-10-20 20:14:16 +00:00
|
|
|
mod mutex_impl {
|
2022-03-21 21:20:37 +00:00
|
|
|
/// Provides interior mutability.
|
2022-03-22 14:34:21 +00:00
|
|
|
///
|
|
|
|
/// Uses `parking_lot` crate on native targets, and `atomic_refcell` on `wasm32` targets.
|
2021-10-20 20:14:16 +00:00
|
|
|
#[derive(Default)]
|
|
|
|
pub struct Mutex<T>(parking_lot::Mutex<T>);
|
|
|
|
|
|
|
|
/// The lock you get from [`Mutex`].
|
|
|
|
pub struct MutexGuard<'a, T>(parking_lot::MutexGuard<'a, T>, *const ());
|
|
|
|
|
|
|
|
#[derive(Default)]
|
|
|
|
struct HeldLocks(Vec<*const ()>);
|
|
|
|
|
|
|
|
impl HeldLocks {
|
|
|
|
#[inline(always)]
|
|
|
|
fn insert(&mut self, lock: *const ()) {
|
|
|
|
// Very few locks will ever be held at the same time, so a linear search is fast
|
|
|
|
assert!(
|
|
|
|
!self.0.contains(&lock),
|
|
|
|
"Recursively locking a Mutex in the same thread is not supported"
|
|
|
|
);
|
|
|
|
self.0.push(lock);
|
|
|
|
}
|
|
|
|
|
|
|
|
#[inline(always)]
|
|
|
|
fn remove(&mut self, lock: *const ()) {
|
|
|
|
self.0.retain(|&ptr| ptr != lock);
|
|
|
|
}
|
2021-10-20 19:58:35 +00:00
|
|
|
}
|
2020-10-13 22:29:11 +00:00
|
|
|
|
2021-10-20 20:14:16 +00:00
|
|
|
thread_local! {
|
|
|
|
static HELD_LOCKS_TLS: std::cell::RefCell<HeldLocks> = Default::default();
|
2021-10-20 19:58:35 +00:00
|
|
|
}
|
|
|
|
|
2021-10-20 20:14:16 +00:00
|
|
|
impl<T> Mutex<T> {
|
|
|
|
#[inline(always)]
|
|
|
|
pub fn new(val: T) -> Self {
|
|
|
|
Self(parking_lot::Mutex::new(val))
|
|
|
|
}
|
2021-05-26 20:13:24 +00:00
|
|
|
|
2021-10-20 20:14:16 +00:00
|
|
|
pub fn lock(&self) -> MutexGuard<'_, T> {
|
|
|
|
// Detect if we are recursively taking out a lock on this mutex.
|
2021-05-26 20:13:24 +00:00
|
|
|
|
2021-10-20 20:14:16 +00:00
|
|
|
// use a pointer to the inner data as an id for this lock
|
|
|
|
let ptr = (&self.0 as *const parking_lot::Mutex<_>).cast::<()>();
|
2021-05-26 20:13:24 +00:00
|
|
|
|
2021-10-20 20:14:16 +00:00
|
|
|
// Store it in thread local storage while we have a lock guard taken out
|
|
|
|
HELD_LOCKS_TLS.with(|held_locks| {
|
|
|
|
held_locks.borrow_mut().insert(ptr);
|
|
|
|
});
|
2020-10-13 22:29:11 +00:00
|
|
|
|
2021-10-20 20:14:16 +00:00
|
|
|
MutexGuard(self.0.lock(), ptr)
|
|
|
|
}
|
2020-10-13 22:29:11 +00:00
|
|
|
}
|
|
|
|
|
2021-10-20 20:14:16 +00:00
|
|
|
impl<T> Drop for MutexGuard<'_, T> {
|
|
|
|
fn drop(&mut self) {
|
|
|
|
let ptr = self.1;
|
|
|
|
HELD_LOCKS_TLS.with(|held_locks| {
|
|
|
|
held_locks.borrow_mut().remove(ptr);
|
|
|
|
});
|
|
|
|
}
|
2021-05-26 20:13:24 +00:00
|
|
|
}
|
|
|
|
|
2021-10-20 20:14:16 +00:00
|
|
|
impl<T> std::ops::Deref for MutexGuard<'_, T> {
|
|
|
|
type Target = T;
|
2021-05-26 20:13:24 +00:00
|
|
|
|
2021-10-20 20:14:16 +00:00
|
|
|
#[inline(always)]
|
|
|
|
fn deref(&self) -> &Self::Target {
|
|
|
|
&self.0
|
|
|
|
}
|
2021-05-26 20:13:24 +00:00
|
|
|
}
|
|
|
|
|
2021-10-20 20:14:16 +00:00
|
|
|
impl<T> std::ops::DerefMut for MutexGuard<'_, T> {
|
|
|
|
#[inline(always)]
|
|
|
|
fn deref_mut(&mut self) -> &mut Self::Target {
|
|
|
|
&mut self.0
|
|
|
|
}
|
2021-05-26 20:13:24 +00:00
|
|
|
}
|
|
|
|
}
|
|
|
|
|
2022-03-22 14:34:21 +00:00
|
|
|
#[cfg(not(target_arch = "wasm32"))]
|
2021-10-20 20:14:16 +00:00
|
|
|
mod rw_lock_impl {
|
|
|
|
/// The lock you get from [`RwLock::read`].
|
2022-01-10 22:13:10 +00:00
|
|
|
pub use parking_lot::MappedRwLockReadGuard as RwLockReadGuard;
|
2021-10-20 20:14:16 +00:00
|
|
|
|
|
|
|
/// The lock you get from [`RwLock::write`].
|
2022-01-10 22:13:10 +00:00
|
|
|
pub use parking_lot::MappedRwLockWriteGuard as RwLockWriteGuard;
|
2021-10-20 20:14:16 +00:00
|
|
|
|
2022-03-21 21:20:37 +00:00
|
|
|
/// Provides interior mutability.
|
2022-03-22 14:34:21 +00:00
|
|
|
///
|
|
|
|
/// Uses `parking_lot` crate on native targets, and `atomic_refcell` on `wasm32` targets.
|
2021-10-20 20:14:16 +00:00
|
|
|
#[derive(Default)]
|
|
|
|
pub struct RwLock<T>(parking_lot::RwLock<T>);
|
|
|
|
|
|
|
|
impl<T> RwLock<T> {
|
|
|
|
#[inline(always)]
|
|
|
|
pub fn new(val: T) -> Self {
|
|
|
|
Self(parking_lot::RwLock::new(val))
|
|
|
|
}
|
|
|
|
|
|
|
|
#[inline(always)]
|
|
|
|
pub fn read(&self) -> RwLockReadGuard<'_, T> {
|
2022-01-10 22:13:10 +00:00
|
|
|
parking_lot::RwLockReadGuard::map(self.0.read(), |v| v)
|
2021-10-20 20:14:16 +00:00
|
|
|
}
|
|
|
|
|
|
|
|
#[inline(always)]
|
|
|
|
pub fn write(&self) -> RwLockWriteGuard<'_, T> {
|
2022-01-10 22:13:10 +00:00
|
|
|
parking_lot::RwLockWriteGuard::map(self.0.write(), |v| v)
|
2021-10-20 20:14:16 +00:00
|
|
|
}
|
2020-12-26 19:36:25 +00:00
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
// ----------------------------------------------------------------------------
|
|
|
|
|
2022-03-22 14:34:21 +00:00
|
|
|
#[cfg(target_arch = "wasm32")]
|
|
|
|
mod mutex_impl {
|
|
|
|
// `atomic_refcell` will panic if multiple threads try to access the same value
|
|
|
|
|
|
|
|
/// Provides interior mutability.
|
|
|
|
///
|
|
|
|
/// Uses `parking_lot` crate on native targets, and `atomic_refcell` on `wasm32` targets.
|
|
|
|
#[derive(Default)]
|
|
|
|
pub struct Mutex<T>(atomic_refcell::AtomicRefCell<T>);
|
|
|
|
|
|
|
|
/// The lock you get from [`Mutex`].
|
|
|
|
pub use atomic_refcell::AtomicRefMut as MutexGuard;
|
|
|
|
|
|
|
|
impl<T> Mutex<T> {
|
|
|
|
#[inline(always)]
|
|
|
|
pub fn new(val: T) -> Self {
|
|
|
|
Self(atomic_refcell::AtomicRefCell::new(val))
|
|
|
|
}
|
|
|
|
|
|
|
|
/// Panics if already locked.
|
|
|
|
#[inline(always)]
|
|
|
|
pub fn lock(&self) -> MutexGuard<'_, T> {
|
|
|
|
self.0.borrow_mut()
|
|
|
|
}
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
#[cfg(target_arch = "wasm32")]
|
|
|
|
mod rw_lock_impl {
|
|
|
|
// `atomic_refcell` will panic if multiple threads try to access the same value
|
|
|
|
|
|
|
|
/// The lock you get from [`RwLock::read`].
|
|
|
|
pub use atomic_refcell::AtomicRef as RwLockReadGuard;
|
|
|
|
|
|
|
|
/// The lock you get from [`RwLock::write`].
|
|
|
|
pub use atomic_refcell::AtomicRefMut as RwLockWriteGuard;
|
|
|
|
|
|
|
|
/// Provides interior mutability.
|
|
|
|
///
|
|
|
|
/// Uses `parking_lot` crate on native targets, and `atomic_refcell` on `wasm32` targets.
|
|
|
|
#[derive(Default)]
|
|
|
|
pub struct RwLock<T>(atomic_refcell::AtomicRefCell<T>);
|
|
|
|
|
|
|
|
impl<T> RwLock<T> {
|
|
|
|
#[inline(always)]
|
|
|
|
pub fn new(val: T) -> Self {
|
|
|
|
Self(atomic_refcell::AtomicRefCell::new(val))
|
|
|
|
}
|
|
|
|
|
|
|
|
#[inline(always)]
|
|
|
|
pub fn read(&self) -> RwLockReadGuard<'_, T> {
|
|
|
|
self.0.borrow()
|
|
|
|
}
|
|
|
|
|
|
|
|
/// Panics if already locked.
|
|
|
|
#[inline(always)]
|
|
|
|
pub fn write(&self) -> RwLockWriteGuard<'_, T> {
|
|
|
|
self.0.borrow_mut()
|
|
|
|
}
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
// ----------------------------------------------------------------------------
|
|
|
|
|
2021-10-20 20:14:16 +00:00
|
|
|
pub use mutex_impl::{Mutex, MutexGuard};
|
|
|
|
pub use rw_lock_impl::{RwLock, RwLockReadGuard, RwLockWriteGuard};
|
|
|
|
|
2020-10-13 22:29:11 +00:00
|
|
|
impl<T> Clone for Mutex<T>
|
|
|
|
where
|
|
|
|
T: Clone,
|
|
|
|
{
|
|
|
|
fn clone(&self) -> Self {
|
|
|
|
Self::new(self.lock().clone())
|
|
|
|
}
|
|
|
|
}
|
2021-05-26 20:13:24 +00:00
|
|
|
|
2021-10-20 20:14:16 +00:00
|
|
|
// ----------------------------------------------------------------------------
|
|
|
|
|
2021-05-26 20:13:24 +00:00
|
|
|
#[cfg(test)]
|
|
|
|
mod tests {
|
|
|
|
use crate::mutex::Mutex;
|
|
|
|
use std::time::Duration;
|
|
|
|
|
|
|
|
#[test]
|
|
|
|
fn lock_two_different_mutexes_single_thread() {
|
|
|
|
let one = Mutex::new(());
|
|
|
|
let two = Mutex::new(());
|
|
|
|
let _a = one.lock();
|
|
|
|
let _b = two.lock();
|
|
|
|
}
|
|
|
|
|
|
|
|
#[test]
|
|
|
|
#[should_panic]
|
|
|
|
fn lock_reentry_single_thread() {
|
|
|
|
let one = Mutex::new(());
|
|
|
|
let _a = one.lock();
|
|
|
|
let _a2 = one.lock(); // panics
|
|
|
|
}
|
|
|
|
|
|
|
|
#[test]
|
|
|
|
fn lock_multiple_threads() {
|
|
|
|
use std::sync::Arc;
|
|
|
|
let one = Arc::new(Mutex::new(()));
|
|
|
|
let our_lock = one.lock();
|
|
|
|
let other_thread = {
|
|
|
|
let one = Arc::clone(&one);
|
|
|
|
std::thread::spawn(move || {
|
|
|
|
let _ = one.lock();
|
|
|
|
})
|
|
|
|
};
|
|
|
|
std::thread::sleep(Duration::from_millis(200));
|
|
|
|
drop(our_lock);
|
|
|
|
other_thread.join().unwrap();
|
|
|
|
}
|
|
|
|
}
|