diff options
| author | mo khan <mo@mokhan.ca> | 2025-07-15 16:37:08 -0600 |
|---|---|---|
| committer | mo khan <mo@mokhan.ca> | 2025-07-17 16:30:22 -0600 |
| commit | 45df4d0d9b577fecee798d672695fe24ff57fb1b (patch) | |
| tree | 1b99bf645035b58e0d6db08c7a83521f41f7a75b /vendor/thread_local/src | |
| parent | f94f79608393d4ab127db63cc41668445ef6b243 (diff) | |
feat: migrate from Cedar to SpiceDB authorization system
This is a major architectural change that replaces the Cedar policy-based
authorization system with SpiceDB's relation-based authorization.
Key changes:
- Migrate from Rust to Go implementation
- Replace Cedar policies with SpiceDB schema and relationships
- Switch from envoy `ext_authz` with Cedar to SpiceDB permission checks
- Update build system and dependencies for Go ecosystem
- Maintain Envoy integration for external authorization
This change enables more flexible permission modeling through SpiceDB's
Google Zanzibar inspired relation-based system, supporting complex
hierarchical permissions that were difficult to express in Cedar.
Breaking change: Existing Cedar policies and Rust-based configuration
will no longer work and need to be migrated to SpiceDB schema.
Diffstat (limited to 'vendor/thread_local/src')
| -rw-r--r-- | vendor/thread_local/src/cached.rs | 161 | ||||
| -rw-r--r-- | vendor/thread_local/src/lib.rs | 683 | ||||
| -rw-r--r-- | vendor/thread_local/src/thread_id.rs | 210 | ||||
| -rw-r--r-- | vendor/thread_local/src/unreachable.rs | 57 |
4 files changed, 0 insertions, 1111 deletions
diff --git a/vendor/thread_local/src/cached.rs b/vendor/thread_local/src/cached.rs deleted file mode 100644 index 16f6516d..00000000 --- a/vendor/thread_local/src/cached.rs +++ /dev/null @@ -1,161 +0,0 @@ -#![allow(deprecated)] - -use super::{IntoIter, IterMut, ThreadLocal}; -use std::fmt; -use std::panic::UnwindSafe; -use std::usize; - -/// Wrapper around [`ThreadLocal`]. -/// -/// This used to add a fast path for a single thread, however that has been -/// obsoleted by performance improvements to [`ThreadLocal`] itself. -#[deprecated(since = "1.1.0", note = "Use `ThreadLocal` instead")] -pub struct CachedThreadLocal<T: Send> { - inner: ThreadLocal<T>, -} - -impl<T: Send> Default for CachedThreadLocal<T> { - fn default() -> CachedThreadLocal<T> { - CachedThreadLocal::new() - } -} - -impl<T: Send> CachedThreadLocal<T> { - /// Creates a new empty `CachedThreadLocal`. - #[inline] - pub fn new() -> CachedThreadLocal<T> { - CachedThreadLocal { - inner: ThreadLocal::new(), - } - } - - /// Returns the element for the current thread, if it exists. - #[inline] - pub fn get(&self) -> Option<&T> { - self.inner.get() - } - - /// Returns the element for the current thread, or creates it if it doesn't - /// exist. - #[inline] - pub fn get_or<F>(&self, create: F) -> &T - where - F: FnOnce() -> T, - { - self.inner.get_or(create) - } - - /// Returns the element for the current thread, or creates it if it doesn't - /// exist. If `create` fails, that error is returned and no element is - /// added. - #[inline] - pub fn get_or_try<F, E>(&self, create: F) -> Result<&T, E> - where - F: FnOnce() -> Result<T, E>, - { - self.inner.get_or_try(create) - } - - /// Returns a mutable iterator over the local values of all threads. - /// - /// Since this call borrows the `ThreadLocal` mutably, this operation can - /// be done safely---the mutable borrow statically guarantees no other - /// threads are currently accessing their associated values. - #[inline] - pub fn iter_mut(&mut self) -> CachedIterMut<T> { - CachedIterMut { - inner: self.inner.iter_mut(), - } - } - - /// Removes all thread-specific values from the `ThreadLocal`, effectively - /// reseting it to its original state. - /// - /// Since this call borrows the `ThreadLocal` mutably, this operation can - /// be done safely---the mutable borrow statically guarantees no other - /// threads are currently accessing their associated values. - #[inline] - pub fn clear(&mut self) { - self.inner.clear(); - } -} - -impl<T: Send> IntoIterator for CachedThreadLocal<T> { - type Item = T; - type IntoIter = CachedIntoIter<T>; - - fn into_iter(self) -> CachedIntoIter<T> { - CachedIntoIter { - inner: self.inner.into_iter(), - } - } -} - -impl<'a, T: Send + 'a> IntoIterator for &'a mut CachedThreadLocal<T> { - type Item = &'a mut T; - type IntoIter = CachedIterMut<'a, T>; - - fn into_iter(self) -> CachedIterMut<'a, T> { - self.iter_mut() - } -} - -impl<T: Send + Default> CachedThreadLocal<T> { - /// Returns the element for the current thread, or creates a default one if - /// it doesn't exist. - pub fn get_or_default(&self) -> &T { - self.get_or(T::default) - } -} - -impl<T: Send + fmt::Debug> fmt::Debug for CachedThreadLocal<T> { - fn fmt(&self, f: &mut fmt::Formatter) -> fmt::Result { - write!(f, "ThreadLocal {{ local_data: {:?} }}", self.get()) - } -} - -impl<T: Send + UnwindSafe> UnwindSafe for CachedThreadLocal<T> {} - -/// Mutable iterator over the contents of a `CachedThreadLocal`. -#[deprecated(since = "1.1.0", note = "Use `IterMut` instead")] -pub struct CachedIterMut<'a, T: Send + 'a> { - inner: IterMut<'a, T>, -} - -impl<'a, T: Send + 'a> Iterator for CachedIterMut<'a, T> { - type Item = &'a mut T; - - #[inline] - fn next(&mut self) -> Option<&'a mut T> { - self.inner.next() - } - - #[inline] - fn size_hint(&self) -> (usize, Option<usize>) { - self.inner.size_hint() - } -} - -impl<'a, T: Send + 'a> ExactSizeIterator for CachedIterMut<'a, T> {} - -/// An iterator that moves out of a `CachedThreadLocal`. -#[deprecated(since = "1.1.0", note = "Use `IntoIter` instead")] -pub struct CachedIntoIter<T: Send> { - inner: IntoIter<T>, -} - -impl<T: Send> Iterator for CachedIntoIter<T> { - type Item = T; - - #[inline] - fn next(&mut self) -> Option<T> { - self.inner.next() - } - - #[inline] - fn size_hint(&self) -> (usize, Option<usize>) { - self.inner.size_hint() - } -} - -impl<T: Send> ExactSizeIterator for CachedIntoIter<T> {} diff --git a/vendor/thread_local/src/lib.rs b/vendor/thread_local/src/lib.rs deleted file mode 100644 index c6ae6860..00000000 --- a/vendor/thread_local/src/lib.rs +++ /dev/null @@ -1,683 +0,0 @@ -// Copyright 2017 Amanieu d'Antras -// -// Licensed under the Apache License, Version 2.0, <LICENSE-APACHE or -// http://apache.org/licenses/LICENSE-2.0> or the MIT license <LICENSE-MIT or -// http://opensource.org/licenses/MIT>, at your option. This file may not be -// copied, modified, or distributed except according to those terms. - -//! Per-object thread-local storage -//! -//! This library provides the `ThreadLocal` type which allows a separate copy of -//! an object to be used for each thread. This allows for per-object -//! thread-local storage, unlike the standard library's `thread_local!` macro -//! which only allows static thread-local storage. -//! -//! Per-thread objects are not destroyed when a thread exits. Instead, objects -//! are only destroyed when the `ThreadLocal` containing them is destroyed. -//! -//! You can also iterate over the thread-local values of all thread in a -//! `ThreadLocal` object using the `iter_mut` and `into_iter` methods. This can -//! only be done if you have mutable access to the `ThreadLocal` object, which -//! guarantees that you are the only thread currently accessing it. -//! -//! Note that since thread IDs are recycled when a thread exits, it is possible -//! for one thread to retrieve the object of another thread. Since this can only -//! occur after a thread has exited this does not lead to any race conditions. -//! -//! # Examples -//! -//! Basic usage of `ThreadLocal`: -//! -//! ```rust -//! use thread_local::ThreadLocal; -//! let tls: ThreadLocal<u32> = ThreadLocal::new(); -//! assert_eq!(tls.get(), None); -//! assert_eq!(tls.get_or(|| 5), &5); -//! assert_eq!(tls.get(), Some(&5)); -//! ``` -//! -//! Combining thread-local values into a single result: -//! -//! ```rust -//! use thread_local::ThreadLocal; -//! use std::sync::Arc; -//! use std::cell::Cell; -//! use std::thread; -//! -//! let tls = Arc::new(ThreadLocal::new()); -//! -//! // Create a bunch of threads to do stuff -//! for _ in 0..5 { -//! let tls2 = tls.clone(); -//! thread::spawn(move || { -//! // Increment a counter to count some event... -//! let cell = tls2.get_or(|| Cell::new(0)); -//! cell.set(cell.get() + 1); -//! }).join().unwrap(); -//! } -//! -//! // Once all threads are done, collect the counter values and return the -//! // sum of all thread-local counter values. -//! let tls = Arc::try_unwrap(tls).unwrap(); -//! let total = tls.into_iter().fold(0, |x, y| x + y.get()); -//! assert_eq!(total, 5); -//! ``` - -#![warn(missing_docs)] -#![allow(clippy::mutex_atomic)] -#![cfg_attr(feature = "nightly", feature(thread_local))] - -mod cached; -mod thread_id; -mod unreachable; - -#[allow(deprecated)] -pub use cached::{CachedIntoIter, CachedIterMut, CachedThreadLocal}; - -use std::cell::UnsafeCell; -use std::fmt; -use std::iter::FusedIterator; -use std::mem; -use std::mem::MaybeUninit; -use std::panic::UnwindSafe; -use std::ptr; -use std::sync::atomic::{AtomicBool, AtomicPtr, AtomicUsize, Ordering}; -use thread_id::Thread; -use unreachable::UncheckedResultExt; - -// Use usize::BITS once it has stabilized and the MSRV has been bumped. -#[cfg(target_pointer_width = "16")] -const POINTER_WIDTH: u8 = 16; -#[cfg(target_pointer_width = "32")] -const POINTER_WIDTH: u8 = 32; -#[cfg(target_pointer_width = "64")] -const POINTER_WIDTH: u8 = 64; - -/// The total number of buckets stored in each thread local. -/// All buckets combined can hold up to `usize::MAX - 1` entries. -const BUCKETS: usize = (POINTER_WIDTH - 1) as usize; - -/// Thread-local variable wrapper -/// -/// See the [module-level documentation](index.html) for more. -pub struct ThreadLocal<T: Send> { - /// The buckets in the thread local. The nth bucket contains `2^n` - /// elements. Each bucket is lazily allocated. - buckets: [AtomicPtr<Entry<T>>; BUCKETS], - - /// The number of values in the thread local. This can be less than the real number of values, - /// but is never more. - values: AtomicUsize, -} - -struct Entry<T> { - present: AtomicBool, - value: UnsafeCell<MaybeUninit<T>>, -} - -impl<T> Drop for Entry<T> { - fn drop(&mut self) { - unsafe { - if *self.present.get_mut() { - ptr::drop_in_place((*self.value.get()).as_mut_ptr()); - } - } - } -} - -// ThreadLocal is always Sync, even if T isn't -unsafe impl<T: Send> Sync for ThreadLocal<T> {} - -impl<T: Send> Default for ThreadLocal<T> { - fn default() -> ThreadLocal<T> { - ThreadLocal::new() - } -} - -impl<T: Send> Drop for ThreadLocal<T> { - fn drop(&mut self) { - // Free each non-null bucket - for (i, bucket) in self.buckets.iter_mut().enumerate() { - let bucket_ptr = *bucket.get_mut(); - - let this_bucket_size = 1 << i; - - if bucket_ptr.is_null() { - continue; - } - - unsafe { deallocate_bucket(bucket_ptr, this_bucket_size) }; - } - } -} - -impl<T: Send> ThreadLocal<T> { - /// Creates a new empty `ThreadLocal`. - pub const fn new() -> ThreadLocal<T> { - let buckets = [ptr::null_mut::<Entry<T>>(); BUCKETS]; - Self { - buckets: unsafe { mem::transmute(buckets) }, - values: AtomicUsize::new(0), - } - } - - /// Creates a new `ThreadLocal` with an initial capacity. If less than the capacity threads - /// access the thread local it will never reallocate. The capacity may be rounded up to the - /// nearest power of two. - pub fn with_capacity(capacity: usize) -> ThreadLocal<T> { - let allocated_buckets = usize::from(POINTER_WIDTH) - (capacity.leading_zeros() as usize); - - let mut buckets = [ptr::null_mut(); BUCKETS]; - for (i, bucket) in buckets[..allocated_buckets].iter_mut().enumerate() { - *bucket = allocate_bucket::<T>(1 << i); - } - - Self { - // Safety: AtomicPtr has the same representation as a pointer and arrays have the same - // representation as a sequence of their inner type. - buckets: unsafe { mem::transmute(buckets) }, - values: AtomicUsize::new(0), - } - } - - /// Returns the element for the current thread, if it exists. - pub fn get(&self) -> Option<&T> { - self.get_inner(thread_id::get()) - } - - /// Returns the element for the current thread, or creates it if it doesn't - /// exist. - pub fn get_or<F>(&self, create: F) -> &T - where - F: FnOnce() -> T, - { - unsafe { - self.get_or_try(|| Ok::<T, ()>(create())) - .unchecked_unwrap_ok() - } - } - - /// Returns the element for the current thread, or creates it if it doesn't - /// exist. If `create` fails, that error is returned and no element is - /// added. - pub fn get_or_try<F, E>(&self, create: F) -> Result<&T, E> - where - F: FnOnce() -> Result<T, E>, - { - let thread = thread_id::get(); - if let Some(val) = self.get_inner(thread) { - return Ok(val); - } - - Ok(self.insert(thread, create()?)) - } - - fn get_inner(&self, thread: Thread) -> Option<&T> { - let bucket_ptr = - unsafe { self.buckets.get_unchecked(thread.bucket) }.load(Ordering::Acquire); - if bucket_ptr.is_null() { - return None; - } - unsafe { - let entry = &*bucket_ptr.add(thread.index); - if entry.present.load(Ordering::Relaxed) { - Some(&*(&*entry.value.get()).as_ptr()) - } else { - None - } - } - } - - #[cold] - fn insert(&self, thread: Thread, data: T) -> &T { - let bucket_atomic_ptr = unsafe { self.buckets.get_unchecked(thread.bucket) }; - let bucket_ptr: *const _ = bucket_atomic_ptr.load(Ordering::Acquire); - - // If the bucket doesn't already exist, we need to allocate it - let bucket_ptr = if bucket_ptr.is_null() { - let new_bucket = allocate_bucket(thread.bucket_size); - - match bucket_atomic_ptr.compare_exchange( - ptr::null_mut(), - new_bucket, - Ordering::AcqRel, - Ordering::Acquire, - ) { - Ok(_) => new_bucket, - // If the bucket value changed (from null), that means - // another thread stored a new bucket before we could, - // and we can free our bucket and use that one instead - Err(bucket_ptr) => { - unsafe { deallocate_bucket(new_bucket, thread.bucket_size) } - bucket_ptr - } - } - } else { - bucket_ptr - }; - - // Insert the new element into the bucket - let entry = unsafe { &*bucket_ptr.add(thread.index) }; - let value_ptr = entry.value.get(); - unsafe { value_ptr.write(MaybeUninit::new(data)) }; - entry.present.store(true, Ordering::Release); - - self.values.fetch_add(1, Ordering::Release); - - unsafe { &*(&*value_ptr).as_ptr() } - } - - /// Returns an iterator over the local values of all threads in unspecified - /// order. - /// - /// This call can be done safely, as `T` is required to implement [`Sync`]. - pub fn iter(&self) -> Iter<'_, T> - where - T: Sync, - { - Iter { - thread_local: self, - raw: RawIter::new(), - } - } - - /// Returns a mutable iterator over the local values of all threads in - /// unspecified order. - /// - /// Since this call borrows the `ThreadLocal` mutably, this operation can - /// be done safely---the mutable borrow statically guarantees no other - /// threads are currently accessing their associated values. - pub fn iter_mut(&mut self) -> IterMut<T> { - IterMut { - thread_local: self, - raw: RawIter::new(), - } - } - - /// Removes all thread-specific values from the `ThreadLocal`, effectively - /// reseting it to its original state. - /// - /// Since this call borrows the `ThreadLocal` mutably, this operation can - /// be done safely---the mutable borrow statically guarantees no other - /// threads are currently accessing their associated values. - pub fn clear(&mut self) { - *self = ThreadLocal::new(); - } -} - -impl<T: Send> IntoIterator for ThreadLocal<T> { - type Item = T; - type IntoIter = IntoIter<T>; - - fn into_iter(self) -> IntoIter<T> { - IntoIter { - thread_local: self, - raw: RawIter::new(), - } - } -} - -impl<'a, T: Send + Sync> IntoIterator for &'a ThreadLocal<T> { - type Item = &'a T; - type IntoIter = Iter<'a, T>; - - fn into_iter(self) -> Self::IntoIter { - self.iter() - } -} - -impl<'a, T: Send> IntoIterator for &'a mut ThreadLocal<T> { - type Item = &'a mut T; - type IntoIter = IterMut<'a, T>; - - fn into_iter(self) -> IterMut<'a, T> { - self.iter_mut() - } -} - -impl<T: Send + Default> ThreadLocal<T> { - /// Returns the element for the current thread, or creates a default one if - /// it doesn't exist. - pub fn get_or_default(&self) -> &T { - self.get_or(Default::default) - } -} - -impl<T: Send + fmt::Debug> fmt::Debug for ThreadLocal<T> { - fn fmt(&self, f: &mut fmt::Formatter) -> fmt::Result { - write!(f, "ThreadLocal {{ local_data: {:?} }}", self.get()) - } -} - -impl<T: Send + UnwindSafe> UnwindSafe for ThreadLocal<T> {} - -#[derive(Debug)] -struct RawIter { - yielded: usize, - bucket: usize, - bucket_size: usize, - index: usize, -} -impl RawIter { - #[inline] - fn new() -> Self { - Self { - yielded: 0, - bucket: 0, - bucket_size: 1, - index: 0, - } - } - - fn next<'a, T: Send + Sync>(&mut self, thread_local: &'a ThreadLocal<T>) -> Option<&'a T> { - while self.bucket < BUCKETS { - let bucket = unsafe { thread_local.buckets.get_unchecked(self.bucket) }; - let bucket = bucket.load(Ordering::Acquire); - - if !bucket.is_null() { - while self.index < self.bucket_size { - let entry = unsafe { &*bucket.add(self.index) }; - self.index += 1; - if entry.present.load(Ordering::Acquire) { - self.yielded += 1; - return Some(unsafe { &*(&*entry.value.get()).as_ptr() }); - } - } - } - - self.next_bucket(); - } - None - } - fn next_mut<'a, T: Send>( - &mut self, - thread_local: &'a mut ThreadLocal<T>, - ) -> Option<&'a mut Entry<T>> { - if *thread_local.values.get_mut() == self.yielded { - return None; - } - - loop { - let bucket = unsafe { thread_local.buckets.get_unchecked_mut(self.bucket) }; - let bucket = *bucket.get_mut(); - - if !bucket.is_null() { - while self.index < self.bucket_size { - let entry = unsafe { &mut *bucket.add(self.index) }; - self.index += 1; - if *entry.present.get_mut() { - self.yielded += 1; - return Some(entry); - } - } - } - - self.next_bucket(); - } - } - - #[inline] - fn next_bucket(&mut self) { - self.bucket_size <<= 1; - self.bucket += 1; - self.index = 0; - } - - fn size_hint<T: Send>(&self, thread_local: &ThreadLocal<T>) -> (usize, Option<usize>) { - let total = thread_local.values.load(Ordering::Acquire); - (total - self.yielded, None) - } - fn size_hint_frozen<T: Send>(&self, thread_local: &ThreadLocal<T>) -> (usize, Option<usize>) { - let total = unsafe { *(&thread_local.values as *const AtomicUsize as *const usize) }; - let remaining = total - self.yielded; - (remaining, Some(remaining)) - } -} - -/// Iterator over the contents of a `ThreadLocal`. -#[derive(Debug)] -pub struct Iter<'a, T: Send + Sync> { - thread_local: &'a ThreadLocal<T>, - raw: RawIter, -} - -impl<'a, T: Send + Sync> Iterator for Iter<'a, T> { - type Item = &'a T; - fn next(&mut self) -> Option<Self::Item> { - self.raw.next(self.thread_local) - } - fn size_hint(&self) -> (usize, Option<usize>) { - self.raw.size_hint(self.thread_local) - } -} -impl<T: Send + Sync> FusedIterator for Iter<'_, T> {} - -/// Mutable iterator over the contents of a `ThreadLocal`. -pub struct IterMut<'a, T: Send> { - thread_local: &'a mut ThreadLocal<T>, - raw: RawIter, -} - -impl<'a, T: Send> Iterator for IterMut<'a, T> { - type Item = &'a mut T; - fn next(&mut self) -> Option<&'a mut T> { - self.raw - .next_mut(self.thread_local) - .map(|entry| unsafe { &mut *(&mut *entry.value.get()).as_mut_ptr() }) - } - fn size_hint(&self) -> (usize, Option<usize>) { - self.raw.size_hint_frozen(self.thread_local) - } -} - -impl<T: Send> ExactSizeIterator for IterMut<'_, T> {} -impl<T: Send> FusedIterator for IterMut<'_, T> {} - -// Manual impl so we don't call Debug on the ThreadLocal, as doing so would create a reference to -// this thread's value that potentially aliases with a mutable reference we have given out. -impl<'a, T: Send + fmt::Debug> fmt::Debug for IterMut<'a, T> { - fn fmt(&self, f: &mut fmt::Formatter<'_>) -> fmt::Result { - f.debug_struct("IterMut").field("raw", &self.raw).finish() - } -} - -/// An iterator that moves out of a `ThreadLocal`. -#[derive(Debug)] -pub struct IntoIter<T: Send> { - thread_local: ThreadLocal<T>, - raw: RawIter, -} - -impl<T: Send> Iterator for IntoIter<T> { - type Item = T; - fn next(&mut self) -> Option<T> { - self.raw.next_mut(&mut self.thread_local).map(|entry| { - *entry.present.get_mut() = false; - unsafe { - std::mem::replace(&mut *entry.value.get(), MaybeUninit::uninit()).assume_init() - } - }) - } - fn size_hint(&self) -> (usize, Option<usize>) { - self.raw.size_hint_frozen(&self.thread_local) - } -} - -impl<T: Send> ExactSizeIterator for IntoIter<T> {} -impl<T: Send> FusedIterator for IntoIter<T> {} - -fn allocate_bucket<T>(size: usize) -> *mut Entry<T> { - Box::into_raw( - (0..size) - .map(|_| Entry::<T> { - present: AtomicBool::new(false), - value: UnsafeCell::new(MaybeUninit::uninit()), - }) - .collect(), - ) as *mut _ -} - -unsafe fn deallocate_bucket<T>(bucket: *mut Entry<T>, size: usize) { - let _ = Box::from_raw(std::slice::from_raw_parts_mut(bucket, size)); -} - -#[cfg(test)] -mod tests { - use super::*; - - use std::cell::RefCell; - use std::sync::atomic::AtomicUsize; - use std::sync::atomic::Ordering::Relaxed; - use std::sync::Arc; - use std::thread; - - fn make_create() -> Arc<dyn Fn() -> usize + Send + Sync> { - let count = AtomicUsize::new(0); - Arc::new(move || count.fetch_add(1, Relaxed)) - } - - #[test] - fn same_thread() { - let create = make_create(); - let mut tls = ThreadLocal::new(); - assert_eq!(None, tls.get()); - assert_eq!("ThreadLocal { local_data: None }", format!("{:?}", &tls)); - assert_eq!(0, *tls.get_or(|| create())); - assert_eq!(Some(&0), tls.get()); - assert_eq!(0, *tls.get_or(|| create())); - assert_eq!(Some(&0), tls.get()); - assert_eq!(0, *tls.get_or(|| create())); - assert_eq!(Some(&0), tls.get()); - assert_eq!("ThreadLocal { local_data: Some(0) }", format!("{:?}", &tls)); - tls.clear(); - assert_eq!(None, tls.get()); - } - - #[test] - fn different_thread() { - let create = make_create(); - let tls = Arc::new(ThreadLocal::new()); - assert_eq!(None, tls.get()); - assert_eq!(0, *tls.get_or(|| create())); - assert_eq!(Some(&0), tls.get()); - - let tls2 = tls.clone(); - let create2 = create.clone(); - thread::spawn(move || { - assert_eq!(None, tls2.get()); - assert_eq!(1, *tls2.get_or(|| create2())); - assert_eq!(Some(&1), tls2.get()); - }) - .join() - .unwrap(); - - assert_eq!(Some(&0), tls.get()); - assert_eq!(0, *tls.get_or(|| create())); - } - - #[test] - fn iter() { - let tls = Arc::new(ThreadLocal::new()); - tls.get_or(|| Box::new(1)); - - let tls2 = tls.clone(); - thread::spawn(move || { - tls2.get_or(|| Box::new(2)); - let tls3 = tls2.clone(); - thread::spawn(move || { - tls3.get_or(|| Box::new(3)); - }) - .join() - .unwrap(); - drop(tls2); - }) - .join() - .unwrap(); - - let mut tls = Arc::try_unwrap(tls).unwrap(); - - let mut v = tls.iter().map(|x| **x).collect::<Vec<i32>>(); - v.sort_unstable(); - assert_eq!(vec![1, 2, 3], v); - - let mut v = tls.iter_mut().map(|x| **x).collect::<Vec<i32>>(); - v.sort_unstable(); - assert_eq!(vec![1, 2, 3], v); - - let mut v = tls.into_iter().map(|x| *x).collect::<Vec<i32>>(); - v.sort_unstable(); - assert_eq!(vec![1, 2, 3], v); - } - - #[test] - fn miri_iter_soundness_check() { - let tls = Arc::new(ThreadLocal::new()); - let _local = tls.get_or(|| Box::new(1)); - - let tls2 = tls.clone(); - let join_1 = thread::spawn(move || { - let _tls = tls2.get_or(|| Box::new(2)); - let iter = tls2.iter(); - for item in iter { - println!("{:?}", item); - } - }); - - let iter = tls.iter(); - for item in iter { - println!("{:?}", item); - } - - join_1.join().ok(); - } - - #[test] - fn test_drop() { - let local = ThreadLocal::new(); - struct Dropped(Arc<AtomicUsize>); - impl Drop for Dropped { - fn drop(&mut self) { - self.0.fetch_add(1, Relaxed); - } - } - - let dropped = Arc::new(AtomicUsize::new(0)); - local.get_or(|| Dropped(dropped.clone())); - assert_eq!(dropped.load(Relaxed), 0); - drop(local); - assert_eq!(dropped.load(Relaxed), 1); - } - - #[test] - fn test_earlyreturn_buckets() { - struct Dropped(Arc<AtomicUsize>); - impl Drop for Dropped { - fn drop(&mut self) { - self.0.fetch_add(1, Relaxed); - } - } - let dropped = Arc::new(AtomicUsize::new(0)); - - // We use a high `id` here to guarantee that a lazily allocated bucket somewhere in the middle is used. - // Neither iteration nor `Drop` must early-return on `null` buckets that are used for lower `buckets`. - let thread = Thread::new(1234); - assert!(thread.bucket > 1); - - let mut local = ThreadLocal::new(); - local.insert(thread, Dropped(dropped.clone())); - - let item = local.iter().next().unwrap(); - assert_eq!(item.0.load(Relaxed), 0); - let item = local.iter_mut().next().unwrap(); - assert_eq!(item.0.load(Relaxed), 0); - drop(local); - assert_eq!(dropped.load(Relaxed), 1); - } - - #[test] - fn is_sync() { - fn foo<T: Sync>() {} - foo::<ThreadLocal<String>>(); - foo::<ThreadLocal<RefCell<String>>>(); - } -} diff --git a/vendor/thread_local/src/thread_id.rs b/vendor/thread_local/src/thread_id.rs deleted file mode 100644 index 91b62337..00000000 --- a/vendor/thread_local/src/thread_id.rs +++ /dev/null @@ -1,210 +0,0 @@ -// Copyright 2017 Amanieu d'Antras -// -// Licensed under the Apache License, Version 2.0, <LICENSE-APACHE or -// http://apache.org/licenses/LICENSE-2.0> or the MIT license <LICENSE-MIT or -// http://opensource.org/licenses/MIT>, at your option. This file may not be -// copied, modified, or distributed except according to those terms. - -use crate::POINTER_WIDTH; -use std::cell::Cell; -use std::cmp::Reverse; -use std::collections::BinaryHeap; -use std::sync::Mutex; - -/// Thread ID manager which allocates thread IDs. It attempts to aggressively -/// reuse thread IDs where possible to avoid cases where a ThreadLocal grows -/// indefinitely when it is used by many short-lived threads. -struct ThreadIdManager { - free_from: usize, - free_list: Option<BinaryHeap<Reverse<usize>>>, -} - -impl ThreadIdManager { - const fn new() -> Self { - Self { - free_from: 0, - free_list: None, - } - } - - fn alloc(&mut self) -> usize { - if let Some(id) = self.free_list.as_mut().and_then(|heap| heap.pop()) { - id.0 - } else { - // `free_from` can't overflow as each thread takes up at least 2 bytes of memory and - // thus we can't even have `usize::MAX / 2 + 1` threads. - - let id = self.free_from; - self.free_from += 1; - id - } - } - - fn free(&mut self, id: usize) { - self.free_list - .get_or_insert_with(BinaryHeap::new) - .push(Reverse(id)); - } -} - -static THREAD_ID_MANAGER: Mutex<ThreadIdManager> = Mutex::new(ThreadIdManager::new()); - -/// Data which is unique to the current thread while it is running. -/// A thread ID may be reused after a thread exits. -#[derive(Clone, Copy)] -pub(crate) struct Thread { - /// The thread ID obtained from the thread ID manager. - pub(crate) id: usize, - /// The bucket this thread's local storage will be in. - pub(crate) bucket: usize, - /// The size of the bucket this thread's local storage will be in. - pub(crate) bucket_size: usize, - /// The index into the bucket this thread's local storage is in. - pub(crate) index: usize, -} -impl Thread { - pub(crate) fn new(id: usize) -> Self { - let bucket = usize::from(POINTER_WIDTH) - ((id + 1).leading_zeros() as usize) - 1; - let bucket_size = 1 << bucket; - let index = id - (bucket_size - 1); - - Self { - id, - bucket, - bucket_size, - index, - } - } -} - -cfg_if::cfg_if! { - if #[cfg(feature = "nightly")] { - // This is split into 2 thread-local variables so that we can check whether the - // thread is initialized without having to register a thread-local destructor. - // - // This makes the fast path smaller. - #[thread_local] - static mut THREAD: Option<Thread> = None; - thread_local! { static THREAD_GUARD: ThreadGuard = const { ThreadGuard { id: Cell::new(0) } }; } - - // Guard to ensure the thread ID is released on thread exit. - struct ThreadGuard { - // We keep a copy of the thread ID in the ThreadGuard: we can't - // reliably access THREAD in our Drop impl due to the unpredictable - // order of TLS destructors. - id: Cell<usize>, - } - - impl Drop for ThreadGuard { - fn drop(&mut self) { - // Release the thread ID. Any further accesses to the thread ID - // will go through get_slow which will either panic or - // initialize a new ThreadGuard. - unsafe { - THREAD = None; - } - THREAD_ID_MANAGER.lock().unwrap().free(self.id.get()); - } - } - - /// Returns a thread ID for the current thread, allocating one if needed. - #[inline] - pub(crate) fn get() -> Thread { - if let Some(thread) = unsafe { THREAD } { - thread - } else { - get_slow() - } - } - - /// Out-of-line slow path for allocating a thread ID. - #[cold] - fn get_slow() -> Thread { - let new = Thread::new(THREAD_ID_MANAGER.lock().unwrap().alloc()); - unsafe { - THREAD = Some(new); - } - THREAD_GUARD.with(|guard| guard.id.set(new.id)); - new - } - } else { - // This is split into 2 thread-local variables so that we can check whether the - // thread is initialized without having to register a thread-local destructor. - // - // This makes the fast path smaller. - thread_local! { static THREAD: Cell<Option<Thread>> = const { Cell::new(None) }; } - thread_local! { static THREAD_GUARD: ThreadGuard = const { ThreadGuard { id: Cell::new(0) } }; } - - // Guard to ensure the thread ID is released on thread exit. - struct ThreadGuard { - // We keep a copy of the thread ID in the ThreadGuard: we can't - // reliably access THREAD in our Drop impl due to the unpredictable - // order of TLS destructors. - id: Cell<usize>, - } - - impl Drop for ThreadGuard { - fn drop(&mut self) { - // Release the thread ID. Any further accesses to the thread ID - // will go through get_slow which will either panic or - // initialize a new ThreadGuard. - let _ = THREAD.try_with(|thread| thread.set(None)); - THREAD_ID_MANAGER.lock().unwrap().free(self.id.get()); - } - } - - /// Returns a thread ID for the current thread, allocating one if needed. - #[inline] - pub(crate) fn get() -> Thread { - THREAD.with(|thread| { - if let Some(thread) = thread.get() { - thread - } else { - get_slow(thread) - } - }) - } - - /// Out-of-line slow path for allocating a thread ID. - #[cold] - fn get_slow(thread: &Cell<Option<Thread>>) -> Thread { - let new = Thread::new(THREAD_ID_MANAGER.lock().unwrap().alloc()); - thread.set(Some(new)); - THREAD_GUARD.with(|guard| guard.id.set(new.id)); - new - } - } -} - -#[test] -fn test_thread() { - let thread = Thread::new(0); - assert_eq!(thread.id, 0); - assert_eq!(thread.bucket, 0); - assert_eq!(thread.bucket_size, 1); - assert_eq!(thread.index, 0); - - let thread = Thread::new(1); - assert_eq!(thread.id, 1); - assert_eq!(thread.bucket, 1); - assert_eq!(thread.bucket_size, 2); - assert_eq!(thread.index, 0); - - let thread = Thread::new(2); - assert_eq!(thread.id, 2); - assert_eq!(thread.bucket, 1); - assert_eq!(thread.bucket_size, 2); - assert_eq!(thread.index, 1); - - let thread = Thread::new(3); - assert_eq!(thread.id, 3); - assert_eq!(thread.bucket, 2); - assert_eq!(thread.bucket_size, 4); - assert_eq!(thread.index, 0); - - let thread = Thread::new(19); - assert_eq!(thread.id, 19); - assert_eq!(thread.bucket, 4); - assert_eq!(thread.bucket_size, 16); - assert_eq!(thread.index, 4); -} diff --git a/vendor/thread_local/src/unreachable.rs b/vendor/thread_local/src/unreachable.rs deleted file mode 100644 index db4d831a..00000000 --- a/vendor/thread_local/src/unreachable.rs +++ /dev/null @@ -1,57 +0,0 @@ -// Copyright 2017 Amanieu d'Antras -// -// Licensed under the Apache License, Version 2.0, <LICENSE-APACHE or -// http://apache.org/licenses/LICENSE-2.0> or the MIT license <LICENSE-MIT or -// http://opensource.org/licenses/MIT>, at your option. This file may not be -// copied, modified, or distributed except according to those terms. - -use std::hint::unreachable_unchecked; - -/// An extension trait for `Option<T>` providing unchecked unwrapping methods. -pub trait UncheckedOptionExt<T> { - /// Get the value out of this Option without checking for None. - unsafe fn unchecked_unwrap(self) -> T; - - /// Assert that this Option is a None to the optimizer. - unsafe fn unchecked_unwrap_none(self); -} - -/// An extension trait for `Result<T, E>` providing unchecked unwrapping methods. -pub trait UncheckedResultExt<T, E> { - /// Get the value out of this Result without checking for Err. - unsafe fn unchecked_unwrap_ok(self) -> T; - - /// Get the error out of this Result without checking for Ok. - unsafe fn unchecked_unwrap_err(self) -> E; -} - -impl<T> UncheckedOptionExt<T> for Option<T> { - unsafe fn unchecked_unwrap(self) -> T { - match self { - Some(x) => x, - None => unreachable_unchecked(), - } - } - - unsafe fn unchecked_unwrap_none(self) { - if self.is_some() { - unreachable_unchecked() - } - } -} - -impl<T, E> UncheckedResultExt<T, E> for Result<T, E> { - unsafe fn unchecked_unwrap_ok(self) -> T { - match self { - Ok(x) => x, - Err(_) => unreachable_unchecked(), - } - } - - unsafe fn unchecked_unwrap_err(self) -> E { - match self { - Ok(_) => unreachable_unchecked(), - Err(e) => e, - } - } -} |
