| // Copyright 2017 The Rust Project Developers. See the COPYRIGHT |
| // file at the top-level directory of this distribution and at |
| // http://rust-lang.org/COPYRIGHT. |
| // |
| // Licensed under the Apache License, Version 2.0 <LICENSE-APACHE or |
| // http://www.apache.org/licenses/LICENSE-2.0> or the MIT license |
| // <LICENSE-MIT or http://opensource.org/licenses/MIT>, at your |
| // option. This file may not be copied, modified, or distributed |
| // except according to those terms. |
| |
| //! This module defines types which are thread safe if cfg!(parallel_queries) is true. |
| //! |
| //! `Lrc` is an alias of either Rc or Arc. |
| //! |
| //! `Lock` is a mutex. |
| //! It internally uses `parking_lot::Mutex` if cfg!(parallel_queries) is true, |
| //! `RefCell` otherwise. |
| //! |
| //! `RwLock` is a read-write lock. |
| //! It internally uses `parking_lot::RwLock` if cfg!(parallel_queries) is true, |
| //! `RefCell` otherwise. |
| //! |
| //! `LockCell` is a thread safe version of `Cell`, with `set` and `get` operations. |
| //! It can never deadlock. It uses `Cell` when |
| //! cfg!(parallel_queries) is false, otherwise it is a `Lock`. |
| //! |
| //! `MTLock` is a mutex which disappears if cfg!(parallel_queries) is false. |
| //! |
| //! `MTRef` is a immutable refernce if cfg!(parallel_queries), and an mutable reference otherwise. |
| //! |
| //! `rustc_erase_owner!` erases a OwningRef owner into Erased or Erased + Send + Sync |
| //! depending on the value of cfg!(parallel_queries). |
| |
| use std::collections::HashMap; |
| use std::hash::{Hash, BuildHasher}; |
| use std::cmp::Ordering; |
| use std::marker::PhantomData; |
| use std::fmt::Debug; |
| use std::fmt::Formatter; |
| use std::fmt; |
| use std::ops::{Deref, DerefMut}; |
| use owning_ref::{Erased, OwningRef}; |
| |
| pub fn serial_join<A, B, RA, RB>(oper_a: A, oper_b: B) -> (RA, RB) |
| where A: FnOnce() -> RA, |
| B: FnOnce() -> RB |
| { |
| (oper_a(), oper_b()) |
| } |
| |
| pub struct SerialScope; |
| |
| impl SerialScope { |
| pub fn spawn<F>(&self, f: F) |
| where F: FnOnce(&SerialScope) |
| { |
| f(self) |
| } |
| } |
| |
| pub fn serial_scope<F, R>(f: F) -> R |
| where F: FnOnce(&SerialScope) -> R |
| { |
| f(&SerialScope) |
| } |
| |
| cfg_if! { |
| if #[cfg(not(parallel_queries))] { |
| pub auto trait Send {} |
| pub auto trait Sync {} |
| |
| impl<T: ?Sized> Send for T {} |
| impl<T: ?Sized> Sync for T {} |
| |
| #[macro_export] |
| macro_rules! rustc_erase_owner { |
| ($v:expr) => { |
| $v.erase_owner() |
| } |
| } |
| |
| pub use self::serial_join as join; |
| pub use self::serial_scope as scope; |
| |
| pub use std::iter::Iterator as ParallelIterator; |
| |
| pub fn par_iter<T: IntoIterator>(t: T) -> T::IntoIter { |
| t.into_iter() |
| } |
| |
| pub type MetadataRef = OwningRef<Box<Erased>, [u8]>; |
| |
| pub use std::rc::Rc as Lrc; |
| pub use std::rc::Weak as Weak; |
| pub use std::cell::Ref as ReadGuard; |
| pub use std::cell::RefMut as WriteGuard; |
| pub use std::cell::RefMut as LockGuard; |
| |
| use std::cell::RefCell as InnerRwLock; |
| use std::cell::RefCell as InnerLock; |
| |
| use std::cell::Cell; |
| |
| #[derive(Debug)] |
| pub struct WorkerLocal<T>(OneThread<T>); |
| |
| impl<T> WorkerLocal<T> { |
| /// Creates a new worker local where the `initial` closure computes the |
| /// value this worker local should take for each thread in the thread pool. |
| #[inline] |
| pub fn new<F: FnMut(usize) -> T>(mut f: F) -> WorkerLocal<T> { |
| WorkerLocal(OneThread::new(f(0))) |
| } |
| |
| /// Returns the worker-local value for each thread |
| #[inline] |
| pub fn into_inner(self) -> Vec<T> { |
| vec![OneThread::into_inner(self.0)] |
| } |
| } |
| |
| impl<T> Deref for WorkerLocal<T> { |
| type Target = T; |
| |
| #[inline(always)] |
| fn deref(&self) -> &T { |
| &*self.0 |
| } |
| } |
| |
| pub type MTRef<'a, T> = &'a mut T; |
| |
| #[derive(Debug)] |
| pub struct MTLock<T>(T); |
| |
| impl<T> MTLock<T> { |
| #[inline(always)] |
| pub fn new(inner: T) -> Self { |
| MTLock(inner) |
| } |
| |
| #[inline(always)] |
| pub fn into_inner(self) -> T { |
| self.0 |
| } |
| |
| #[inline(always)] |
| pub fn get_mut(&mut self) -> &mut T { |
| &mut self.0 |
| } |
| |
| #[inline(always)] |
| pub fn lock(&self) -> &T { |
| &self.0 |
| } |
| |
| #[inline(always)] |
| pub fn lock_mut(&mut self) -> &mut T { |
| &mut self.0 |
| } |
| } |
| |
| // FIXME: Probably a bad idea (in the threaded case) |
| impl<T: Clone> Clone for MTLock<T> { |
| #[inline] |
| fn clone(&self) -> Self { |
| MTLock(self.0.clone()) |
| } |
| } |
| |
| pub struct LockCell<T>(Cell<T>); |
| |
| impl<T> LockCell<T> { |
| #[inline(always)] |
| pub fn new(inner: T) -> Self { |
| LockCell(Cell::new(inner)) |
| } |
| |
| #[inline(always)] |
| pub fn into_inner(self) -> T { |
| self.0.into_inner() |
| } |
| |
| #[inline(always)] |
| pub fn set(&self, new_inner: T) { |
| self.0.set(new_inner); |
| } |
| |
| #[inline(always)] |
| pub fn get(&self) -> T where T: Copy { |
| self.0.get() |
| } |
| |
| #[inline(always)] |
| pub fn set_mut(&mut self, new_inner: T) { |
| self.0.set(new_inner); |
| } |
| |
| #[inline(always)] |
| pub fn get_mut(&mut self) -> T where T: Copy { |
| self.0.get() |
| } |
| } |
| |
| impl<T> LockCell<Option<T>> { |
| #[inline(always)] |
| pub fn take(&self) -> Option<T> { |
| unsafe { (*self.0.as_ptr()).take() } |
| } |
| } |
| } else { |
| pub use std::marker::Send as Send; |
| pub use std::marker::Sync as Sync; |
| |
| pub use parking_lot::RwLockReadGuard as ReadGuard; |
| pub use parking_lot::RwLockWriteGuard as WriteGuard; |
| |
| pub use parking_lot::MutexGuard as LockGuard; |
| |
| pub use std::sync::Arc as Lrc; |
| pub use std::sync::Weak as Weak; |
| |
| pub type MTRef<'a, T> = &'a T; |
| |
| #[derive(Debug)] |
| pub struct MTLock<T>(Lock<T>); |
| |
| impl<T> MTLock<T> { |
| #[inline(always)] |
| pub fn new(inner: T) -> Self { |
| MTLock(Lock::new(inner)) |
| } |
| |
| #[inline(always)] |
| pub fn into_inner(self) -> T { |
| self.0.into_inner() |
| } |
| |
| #[inline(always)] |
| pub fn get_mut(&mut self) -> &mut T { |
| self.0.get_mut() |
| } |
| |
| #[inline(always)] |
| pub fn lock(&self) -> LockGuard<T> { |
| self.0.lock() |
| } |
| |
| #[inline(always)] |
| pub fn lock_mut(&self) -> LockGuard<T> { |
| self.lock() |
| } |
| } |
| |
| use parking_lot::Mutex as InnerLock; |
| use parking_lot::RwLock as InnerRwLock; |
| |
| use std; |
| use std::thread; |
| pub use rayon::{join, scope}; |
| |
| pub use rayon_core::WorkerLocal; |
| |
| pub use rayon::iter::ParallelIterator; |
| use rayon::iter::IntoParallelIterator; |
| |
| pub fn par_iter<T: IntoParallelIterator>(t: T) -> T::Iter { |
| t.into_par_iter() |
| } |
| |
| pub type MetadataRef = OwningRef<Box<Erased + Send + Sync>, [u8]>; |
| |
| /// This makes locks panic if they are already held. |
| /// It is only useful when you are running in a single thread |
| const ERROR_CHECKING: bool = false; |
| |
| #[macro_export] |
| macro_rules! rustc_erase_owner { |
| ($v:expr) => {{ |
| let v = $v; |
| ::rustc_data_structures::sync::assert_send_val(&v); |
| v.erase_send_sync_owner() |
| }} |
| } |
| |
| pub struct LockCell<T>(Lock<T>); |
| |
| impl<T> LockCell<T> { |
| #[inline(always)] |
| pub fn new(inner: T) -> Self { |
| LockCell(Lock::new(inner)) |
| } |
| |
| #[inline(always)] |
| pub fn into_inner(self) -> T { |
| self.0.into_inner() |
| } |
| |
| #[inline(always)] |
| pub fn set(&self, new_inner: T) { |
| *self.0.lock() = new_inner; |
| } |
| |
| #[inline(always)] |
| pub fn get(&self) -> T where T: Copy { |
| *self.0.lock() |
| } |
| |
| #[inline(always)] |
| pub fn set_mut(&mut self, new_inner: T) { |
| *self.0.get_mut() = new_inner; |
| } |
| |
| #[inline(always)] |
| pub fn get_mut(&mut self) -> T where T: Copy { |
| *self.0.get_mut() |
| } |
| } |
| |
| impl<T> LockCell<Option<T>> { |
| #[inline(always)] |
| pub fn take(&self) -> Option<T> { |
| self.0.lock().take() |
| } |
| } |
| } |
| } |
| |
| pub fn assert_sync<T: ?Sized + Sync>() {} |
| pub fn assert_send_val<T: ?Sized + Send>(_t: &T) {} |
| pub fn assert_send_sync_val<T: ?Sized + Sync + Send>(_t: &T) {} |
| |
| pub trait HashMapExt<K, V> { |
| /// Same as HashMap::insert, but it may panic if there's already an |
| /// entry for `key` with a value not equal to `value` |
| fn insert_same(&mut self, key: K, value: V); |
| } |
| |
| impl<K: Eq + Hash, V: Eq, S: BuildHasher> HashMapExt<K, V> for HashMap<K, V, S> { |
| fn insert_same(&mut self, key: K, value: V) { |
| self.entry(key).and_modify(|old| assert!(*old == value)).or_insert(value); |
| } |
| } |
| |
| /// A type whose inner value can be written once and then will stay read-only |
| // This contains a PhantomData<T> since this type conceptually owns a T outside the Mutex once |
| // initialized. This ensures that Once<T> is Sync only if T is. If we did not have PhantomData<T> |
| // we could send a &Once<Cell<bool>> to multiple threads and call `get` on it to get access |
| // to &Cell<bool> on those threads. |
| pub struct Once<T>(Lock<Option<T>>, PhantomData<T>); |
| |
| impl<T> Once<T> { |
| /// Creates an Once value which is uninitialized |
| #[inline(always)] |
| pub fn new() -> Self { |
| Once(Lock::new(None), PhantomData) |
| } |
| |
| /// Consumes the value and returns Some(T) if it was initialized |
| #[inline(always)] |
| pub fn into_inner(self) -> Option<T> { |
| self.0.into_inner() |
| } |
| |
| /// Tries to initialize the inner value to `value`. |
| /// Returns `None` if the inner value was uninitialized and `value` was consumed setting it |
| /// otherwise if the inner value was already set it returns `value` back to the caller |
| #[inline] |
| pub fn try_set(&self, value: T) -> Option<T> { |
| let mut lock = self.0.lock(); |
| if lock.is_some() { |
| return Some(value); |
| } |
| *lock = Some(value); |
| None |
| } |
| |
| /// Tries to initialize the inner value to `value`. |
| /// Returns `None` if the inner value was uninitialized and `value` was consumed setting it |
| /// otherwise if the inner value was already set it asserts that `value` is equal to the inner |
| /// value and then returns `value` back to the caller |
| #[inline] |
| pub fn try_set_same(&self, value: T) -> Option<T> where T: Eq { |
| let mut lock = self.0.lock(); |
| if let Some(ref inner) = *lock { |
| assert!(*inner == value); |
| return Some(value); |
| } |
| *lock = Some(value); |
| None |
| } |
| |
| /// Tries to initialize the inner value to `value` and panics if it was already initialized |
| #[inline] |
| pub fn set(&self, value: T) { |
| assert!(self.try_set(value).is_none()); |
| } |
| |
| /// Tries to initialize the inner value by calling the closure while ensuring that no-one else |
| /// can access the value in the mean time by holding a lock for the duration of the closure. |
| /// If the value was already initialized the closure is not called and `false` is returned, |
| /// otherwise if the value from the closure initializes the inner value, `true` is returned |
| #[inline] |
| pub fn init_locking<F: FnOnce() -> T>(&self, f: F) -> bool { |
| let mut lock = self.0.lock(); |
| if lock.is_some() { |
| return false; |
| } |
| *lock = Some(f()); |
| true |
| } |
| |
| /// Tries to initialize the inner value by calling the closure without ensuring that no-one |
| /// else can access it. This mean when this is called from multiple threads, multiple |
| /// closures may concurrently be computing a value which the inner value should take. |
| /// Only one of these closures are used to actually initialize the value. |
| /// If some other closure already set the value, |
| /// we return the value our closure computed wrapped in a `Option`. |
| /// If our closure set the value, `None` is returned. |
| /// If the value is already initialized, the closure is not called and `None` is returned. |
| #[inline] |
| pub fn init_nonlocking<F: FnOnce() -> T>(&self, f: F) -> Option<T> { |
| if self.0.lock().is_some() { |
| None |
| } else { |
| self.try_set(f()) |
| } |
| } |
| |
| /// Tries to initialize the inner value by calling the closure without ensuring that no-one |
| /// else can access it. This mean when this is called from multiple threads, multiple |
| /// closures may concurrently be computing a value which the inner value should take. |
| /// Only one of these closures are used to actually initialize the value. |
| /// If some other closure already set the value, we assert that it our closure computed |
| /// a value equal to the value aready set and then |
| /// we return the value our closure computed wrapped in a `Option`. |
| /// If our closure set the value, `None` is returned. |
| /// If the value is already initialized, the closure is not called and `None` is returned. |
| #[inline] |
| pub fn init_nonlocking_same<F: FnOnce() -> T>(&self, f: F) -> Option<T> where T: Eq { |
| if self.0.lock().is_some() { |
| None |
| } else { |
| self.try_set_same(f()) |
| } |
| } |
| |
| /// Tries to get a reference to the inner value, returns `None` if it is not yet initialized |
| #[inline(always)] |
| pub fn try_get(&self) -> Option<&T> { |
| let lock = &*self.0.lock(); |
| if let Some(ref inner) = *lock { |
| // This is safe since we won't mutate the inner value |
| unsafe { Some(&*(inner as *const T)) } |
| } else { |
| None |
| } |
| } |
| |
| /// Gets reference to the inner value, panics if it is not yet initialized |
| #[inline(always)] |
| pub fn get(&self) -> &T { |
| self.try_get().expect("value was not set") |
| } |
| |
| /// Gets reference to the inner value, panics if it is not yet initialized |
| #[inline(always)] |
| pub fn borrow(&self) -> &T { |
| self.get() |
| } |
| } |
| |
| impl<T: Copy + Debug> Debug for LockCell<T> { |
| fn fmt(&self, f: &mut Formatter) -> fmt::Result { |
| f.debug_struct("LockCell") |
| .field("value", &self.get()) |
| .finish() |
| } |
| } |
| |
| impl<T:Default> Default for LockCell<T> { |
| /// Creates a `LockCell<T>`, with the `Default` value for T. |
| #[inline] |
| fn default() -> LockCell<T> { |
| LockCell::new(Default::default()) |
| } |
| } |
| |
| impl<T:PartialEq + Copy> PartialEq for LockCell<T> { |
| #[inline] |
| fn eq(&self, other: &LockCell<T>) -> bool { |
| self.get() == other.get() |
| } |
| } |
| |
| impl<T:Eq + Copy> Eq for LockCell<T> {} |
| |
| impl<T:PartialOrd + Copy> PartialOrd for LockCell<T> { |
| #[inline] |
| fn partial_cmp(&self, other: &LockCell<T>) -> Option<Ordering> { |
| self.get().partial_cmp(&other.get()) |
| } |
| |
| #[inline] |
| fn lt(&self, other: &LockCell<T>) -> bool { |
| self.get() < other.get() |
| } |
| |
| #[inline] |
| fn le(&self, other: &LockCell<T>) -> bool { |
| self.get() <= other.get() |
| } |
| |
| #[inline] |
| fn gt(&self, other: &LockCell<T>) -> bool { |
| self.get() > other.get() |
| } |
| |
| #[inline] |
| fn ge(&self, other: &LockCell<T>) -> bool { |
| self.get() >= other.get() |
| } |
| } |
| |
| impl<T:Ord + Copy> Ord for LockCell<T> { |
| #[inline] |
| fn cmp(&self, other: &LockCell<T>) -> Ordering { |
| self.get().cmp(&other.get()) |
| } |
| } |
| |
| #[derive(Debug)] |
| pub struct Lock<T>(InnerLock<T>); |
| |
| impl<T> Lock<T> { |
| #[inline(always)] |
| pub fn new(inner: T) -> Self { |
| Lock(InnerLock::new(inner)) |
| } |
| |
| #[inline(always)] |
| pub fn into_inner(self) -> T { |
| self.0.into_inner() |
| } |
| |
| #[inline(always)] |
| pub fn get_mut(&mut self) -> &mut T { |
| self.0.get_mut() |
| } |
| |
| #[cfg(parallel_queries)] |
| #[inline(always)] |
| pub fn try_lock(&self) -> Option<LockGuard<T>> { |
| self.0.try_lock() |
| } |
| |
| #[cfg(not(parallel_queries))] |
| #[inline(always)] |
| pub fn try_lock(&self) -> Option<LockGuard<T>> { |
| self.0.try_borrow_mut().ok() |
| } |
| |
| #[cfg(parallel_queries)] |
| #[inline(always)] |
| pub fn lock(&self) -> LockGuard<T> { |
| if ERROR_CHECKING { |
| self.0.try_lock().expect("lock was already held") |
| } else { |
| self.0.lock() |
| } |
| } |
| |
| #[cfg(not(parallel_queries))] |
| #[inline(always)] |
| pub fn lock(&self) -> LockGuard<T> { |
| self.0.borrow_mut() |
| } |
| |
| #[inline(always)] |
| pub fn with_lock<F: FnOnce(&mut T) -> R, R>(&self, f: F) -> R { |
| f(&mut *self.lock()) |
| } |
| |
| #[inline(always)] |
| pub fn borrow(&self) -> LockGuard<T> { |
| self.lock() |
| } |
| |
| #[inline(always)] |
| pub fn borrow_mut(&self) -> LockGuard<T> { |
| self.lock() |
| } |
| } |
| |
| impl<T: Default> Default for Lock<T> { |
| #[inline] |
| fn default() -> Self { |
| Lock::new(T::default()) |
| } |
| } |
| |
| // FIXME: Probably a bad idea |
| impl<T: Clone> Clone for Lock<T> { |
| #[inline] |
| fn clone(&self) -> Self { |
| Lock::new(self.borrow().clone()) |
| } |
| } |
| |
| #[derive(Debug)] |
| pub struct RwLock<T>(InnerRwLock<T>); |
| |
| impl<T> RwLock<T> { |
| #[inline(always)] |
| pub fn new(inner: T) -> Self { |
| RwLock(InnerRwLock::new(inner)) |
| } |
| |
| #[inline(always)] |
| pub fn into_inner(self) -> T { |
| self.0.into_inner() |
| } |
| |
| #[inline(always)] |
| pub fn get_mut(&mut self) -> &mut T { |
| self.0.get_mut() |
| } |
| |
| #[cfg(not(parallel_queries))] |
| #[inline(always)] |
| pub fn read(&self) -> ReadGuard<T> { |
| self.0.borrow() |
| } |
| |
| #[cfg(parallel_queries)] |
| #[inline(always)] |
| pub fn read(&self) -> ReadGuard<T> { |
| if ERROR_CHECKING { |
| self.0.try_read().expect("lock was already held") |
| } else { |
| self.0.read() |
| } |
| } |
| |
| #[inline(always)] |
| pub fn with_read_lock<F: FnOnce(&T) -> R, R>(&self, f: F) -> R { |
| f(&*self.read()) |
| } |
| |
| #[cfg(not(parallel_queries))] |
| #[inline(always)] |
| pub fn try_write(&self) -> Result<WriteGuard<T>, ()> { |
| self.0.try_borrow_mut().map_err(|_| ()) |
| } |
| |
| #[cfg(parallel_queries)] |
| #[inline(always)] |
| pub fn try_write(&self) -> Result<WriteGuard<T>, ()> { |
| self.0.try_write().ok_or(()) |
| } |
| |
| #[cfg(not(parallel_queries))] |
| #[inline(always)] |
| pub fn write(&self) -> WriteGuard<T> { |
| self.0.borrow_mut() |
| } |
| |
| #[cfg(parallel_queries)] |
| #[inline(always)] |
| pub fn write(&self) -> WriteGuard<T> { |
| if ERROR_CHECKING { |
| self.0.try_write().expect("lock was already held") |
| } else { |
| self.0.write() |
| } |
| } |
| |
| #[inline(always)] |
| pub fn with_write_lock<F: FnOnce(&mut T) -> R, R>(&self, f: F) -> R { |
| f(&mut *self.write()) |
| } |
| |
| #[inline(always)] |
| pub fn borrow(&self) -> ReadGuard<T> { |
| self.read() |
| } |
| |
| #[inline(always)] |
| pub fn borrow_mut(&self) -> WriteGuard<T> { |
| self.write() |
| } |
| } |
| |
| // FIXME: Probably a bad idea |
| impl<T: Clone> Clone for RwLock<T> { |
| #[inline] |
| fn clone(&self) -> Self { |
| RwLock::new(self.borrow().clone()) |
| } |
| } |
| |
| /// A type which only allows its inner value to be used in one thread. |
| /// It will panic if it is used on multiple threads. |
| #[derive(Copy, Clone, Hash, Debug, Eq, PartialEq)] |
| pub struct OneThread<T> { |
| #[cfg(parallel_queries)] |
| thread: thread::ThreadId, |
| inner: T, |
| } |
| |
| #[cfg(parallel_queries)] |
| unsafe impl<T> std::marker::Sync for OneThread<T> {} |
| #[cfg(parallel_queries)] |
| unsafe impl<T> std::marker::Send for OneThread<T> {} |
| |
| impl<T> OneThread<T> { |
| #[inline(always)] |
| fn check(&self) { |
| #[cfg(parallel_queries)] |
| assert_eq!(thread::current().id(), self.thread); |
| } |
| |
| #[inline(always)] |
| pub fn new(inner: T) -> Self { |
| OneThread { |
| #[cfg(parallel_queries)] |
| thread: thread::current().id(), |
| inner, |
| } |
| } |
| |
| #[inline(always)] |
| pub fn into_inner(value: Self) -> T { |
| value.check(); |
| value.inner |
| } |
| } |
| |
| impl<T> Deref for OneThread<T> { |
| type Target = T; |
| |
| fn deref(&self) -> &T { |
| self.check(); |
| &self.inner |
| } |
| } |
| |
| impl<T> DerefMut for OneThread<T> { |
| fn deref_mut(&mut self) -> &mut T { |
| self.check(); |
| &mut self.inner |
| } |
| } |