Struct AtomicPtr
struct AtomicPtr<T> { ... }
A raw pointer type which can be safely shared between threads.
This type has the same in-memory representation as a *mut T.
If the compiler and the platform support atomic loads and stores of pointers,
this type is a wrapper for the standard library's
AtomicPtr. If the platform supports it
but the compiler does not, atomic operations are implemented using inline
assembly.
Implementations
impl<T> AtomicPtr<T>
const fn new(p: *mut T) -> SelfCreates a new
AtomicPtr.Examples
use AtomicPtr; let ptr = &mut 5; let atomic_ptr = new;unsafe const fn from_ptr<'a>(ptr: *mut *mut T) -> &'a SelfCreates a new
AtomicPtrfrom a pointer.This is
const fnon Rust 1.83+.Safety
ptrmust be aligned toalign_of::<AtomicPtr<T>>()(note that on some platforms this can be bigger thanalign_of::<*mut T>()).ptrmust be valid for both reads and writes for the whole lifetime'a.- If this atomic type is lock-free, non-atomic accesses to the value
behind
ptrmust have a happens-before relationship with atomic accesses via the returned value (or vice-versa).- In other words, time periods where the value is accessed atomically may not overlap with periods where the value is accessed non-atomically.
- This requirement is trivially satisfied if
ptris never used non-atomically for the duration of lifetime'a. Most use cases should be able to follow this guideline. - This requirement is also trivially satisfied if all accesses (atomic or not) are done from the same thread.
- If this atomic type is not lock-free:
- Any accesses to the value behind
ptrmust have a happens-before relationship with accesses via the returned value (or vice-versa). - Any concurrent accesses to the value behind
ptrfor the duration of lifetime'amust be compatible with operations performed by this atomic type.
- Any accesses to the value behind
- This method must not be used to create overlapping or mixed-size atomic accesses, as these are not supported by the memory model.
fn is_lock_free() -> boolReturns
trueif operations on values of this type are lock-free.If the compiler or the platform doesn't support the necessary atomic instructions, global locks for every potentially concurrent atomic operation will be used.
Examples
use AtomicPtr; let is_lock_free = is_lock_free;const fn is_always_lock_free() -> boolReturns
trueif operations on values of this type are lock-free.If the compiler or the platform doesn't support the necessary atomic instructions, global locks for every potentially concurrent atomic operation will be used.
Note: If the atomic operation relies on dynamic CPU feature detection, this type may be lock-free even if the function returns false.
Examples
use AtomicPtr; const IS_ALWAYS_LOCK_FREE: bool = is_always_lock_free;const fn get_mut(self: &mut Self) -> &mut *mut TReturns a mutable reference to the underlying pointer.
This is safe because the mutable reference guarantees that no other threads are concurrently accessing the atomic data.
This is
const fnon Rust 1.83+.Examples
use ; let mut data = 10; let mut atomic_ptr = new; let mut other_data = 5; *atomic_ptr.get_mut = &mut other_data; assert_eq!;const fn into_inner(self: Self) -> *mut TConsumes the atomic and returns the contained value.
This is safe because passing
selfby value guarantees that no other threads are concurrently accessing the atomic data.This is
const fnon Rust 1.56+.Examples
use AtomicPtr; let mut data = 5; let atomic_ptr = new; assert_eq!;fn load(self: &Self, order: Ordering) -> *mut TLoads a value from the pointer.
loadtakes anOrderingargument which describes the memory ordering of this operation. Possible values areSeqCst,AcquireandRelaxed.Panics
Panics if
orderisReleaseorAcqRel.Examples
use ; let ptr = &mut 5; let some_ptr = new; let value = some_ptr.load;fn store(self: &Self, ptr: *mut T, order: Ordering)Stores a value into the pointer.
storetakes anOrderingargument which describes the memory ordering of this operation. Possible values areSeqCst,ReleaseandRelaxed.Panics
Panics if
orderisAcquireorAcqRel.Examples
use ; let ptr = &mut 5; let some_ptr = new; let other_ptr = &mut 10; some_ptr.store;fn swap(self: &Self, ptr: *mut T, order: Ordering) -> *mut TStores a value into the pointer, returning the previous value.
swaptakes anOrderingargument which describes the memory ordering of this operation. All ordering modes are possible. Note that usingAcquiremakes the store part of this operationRelaxed, and usingReleasemakes the load partRelaxed.Examples
use ; let ptr = &mut 5; let some_ptr = new; let other_ptr = &mut 10; let value = some_ptr.swap;fn compare_exchange(self: &Self, current: *mut T, new: *mut T, success: Ordering, failure: Ordering) -> Result<*mut T, *mut T>Stores a value into the pointer if the current value is the same as the
currentvalue.The return value is a result indicating whether the new value was written and containing the previous value. On success this value is guaranteed to be equal to
current.compare_exchangetakes twoOrderingarguments to describe the memory ordering of this operation.successdescribes the required ordering for the read-modify-write operation that takes place if the comparison withcurrentsucceeds.failuredescribes the required ordering for the load operation that takes place when the comparison fails. UsingAcquireas success ordering makes the store part of this operationRelaxed, and usingReleasemakes the successful loadRelaxed. The failure ordering can only beSeqCst,AcquireorRelaxed.Panics
Panics if
failureisRelease,AcqRel.Examples
use ; let ptr = &mut 5; let some_ptr = new; let other_ptr = &mut 10; let value = some_ptr.compare_exchange;fn compare_exchange_weak(self: &Self, current: *mut T, new: *mut T, success: Ordering, failure: Ordering) -> Result<*mut T, *mut T>Stores a value into the pointer if the current value is the same as the
currentvalue.Unlike
AtomicPtr::compare_exchange, this function is allowed to spuriously fail even when the comparison succeeds, which can result in more efficient code on some platforms. The return value is a result indicating whether the new value was written and containing the previous value.compare_exchange_weaktakes twoOrderingarguments to describe the memory ordering of this operation.successdescribes the required ordering for the read-modify-write operation that takes place if the comparison withcurrentsucceeds.failuredescribes the required ordering for the load operation that takes place when the comparison fails. UsingAcquireas success ordering makes the store part of this operationRelaxed, and usingReleasemakes the successful loadRelaxed. The failure ordering can only beSeqCst,AcquireorRelaxed.Panics
Panics if
failureisRelease,AcqRel.Examples
use ; let some_ptr = new; let new = &mut 10; let mut old = some_ptr.load; loopfn fetch_update<F>(self: &Self, set_order: Ordering, fetch_order: Ordering, f: F) -> Result<*mut T, *mut T> where F: FnMut(*mut T) -> Option<*mut T>Fetches the value, and applies a function to it that returns an optional new value. Returns a
ResultofOk(previous_value)if the function returnedSome(_), elseErr(previous_value).Note: This may call the function multiple times if the value has been changed from other threads in the meantime, as long as the function returns
Some(_), but the function will have been applied only once to the stored value.fetch_updatetakes twoOrderingarguments to describe the memory ordering of this operation. The first describes the required ordering for when the operation finally succeeds while the second describes the required ordering for loads. These correspond to the success and failure orderings ofcompare_exchangerespectively.Using
Acquireas success ordering makes the store part of this operationRelaxed, and usingReleasemakes the final successful loadRelaxed. The (failed) load ordering can only beSeqCst,AcquireorRelaxed.Panics
Panics if
fetch_orderisRelease,AcqRel.Considerations
This method is not magic; it is not provided by the hardware. It is implemented in terms of
compare_exchange_weak, and suffers from the same drawbacks. In particular, this method will not circumvent the ABA Problem.Examples
use ; let ptr: *mut _ = &mut 5; let some_ptr = new; let new: *mut _ = &mut 10; assert_eq!; let result = some_ptr.fetch_update; assert_eq!; assert_eq!;fn fetch_ptr_add(self: &Self, val: usize, order: Ordering) -> *mut TOffsets the pointer's address by adding
val(in units ofT), returning the previous pointer.This is equivalent to using
wrapping_addto atomically perform the equivalent ofptr = ptr.wrapping_add(val);.This method operates in units of
T, which means that it cannot be used to offset the pointer by an amount which is not a multiple ofsize_of::<T>(). This can sometimes be inconvenient, as you may want to work with a deliberately misaligned pointer. In such cases, you may use thefetch_byte_addmethod instead.fetch_ptr_addtakes anOrderingargument which describes the memory ordering of this operation. All ordering modes are possible. Note that usingAcquiremakes the store part of this operationRelaxed, and usingReleasemakes the load partRelaxed.Examples
# # use Strict; // strict provenance polyfill for old rustc use ; let atom = new; assert_eq!; // Note: units of `size_of::<i64>()`. assert_eq!;fn fetch_ptr_sub(self: &Self, val: usize, order: Ordering) -> *mut TOffsets the pointer's address by subtracting
val(in units ofT), returning the previous pointer.This is equivalent to using
wrapping_subto atomically perform the equivalent ofptr = ptr.wrapping_sub(val);.This method operates in units of
T, which means that it cannot be used to offset the pointer by an amount which is not a multiple ofsize_of::<T>(). This can sometimes be inconvenient, as you may want to work with a deliberately misaligned pointer. In such cases, you may use thefetch_byte_submethod instead.fetch_ptr_subtakes anOrderingargument which describes the memory ordering of this operation. All ordering modes are possible. Note that usingAcquiremakes the store part of this operationRelaxed, and usingReleasemakes the load partRelaxed.Examples
use ; let array = ; let atom = new; assert!; assert!;fn fetch_byte_add(self: &Self, val: usize, order: Ordering) -> *mut TOffsets the pointer's address by adding
valbytes, returning the previous pointer.This is equivalent to using
wrapping_addandcastto atomically performptr = ptr.cast::<u8>().wrapping_add(val).cast::<T>().fetch_byte_addtakes anOrderingargument which describes the memory ordering of this operation. All ordering modes are possible. Note that usingAcquiremakes the store part of this operationRelaxed, and usingReleasemakes the load partRelaxed.Examples
# # use Strict; // strict provenance polyfill for old rustc use ; let atom = new; assert_eq!; // Note: in units of bytes, not `size_of::<i64>()`. assert_eq!;fn fetch_byte_sub(self: &Self, val: usize, order: Ordering) -> *mut TOffsets the pointer's address by subtracting
valbytes, returning the previous pointer.This is equivalent to using
wrapping_subandcastto atomically performptr = ptr.cast::<u8>().wrapping_sub(val).cast::<T>().fetch_byte_subtakes anOrderingargument which describes the memory ordering of this operation. All ordering modes are possible. Note that usingAcquiremakes the store part of this operationRelaxed, and usingReleasemakes the load partRelaxed.Examples
# # use Strict; // strict provenance polyfill for old rustc use ; let atom = new; assert_eq!; assert_eq!;fn fetch_or(self: &Self, val: usize, order: Ordering) -> *mut TPerforms a bitwise "or" operation on the address of the current pointer, and the argument
val, and stores a pointer with provenance of the current pointer and the resulting address.This is equivalent to using
map_addrto atomically performptr = ptr.map_addr(|a| a | val). This can be used in tagged pointer schemes to atomically set tag bits.Caveat: This operation returns the previous value. To compute the stored value without losing provenance, you may use
map_addr. For example:a.fetch_or(val).map_addr(|a| a | val).fetch_ortakes anOrderingargument which describes the memory ordering of this operation. All ordering modes are possible. Note that usingAcquiremakes the store part of this operationRelaxed, and usingReleasemakes the load partRelaxed.This API and its claimed semantics are part of the Strict Provenance experiment, see the [module documentation for
ptr][core::ptr] for details.Examples
# # use Strict; // strict provenance polyfill for old rustc use ; let pointer = &mut 3i64 as *mut i64; let atom = new; // Tag the bottom bit of the pointer. assert_eq!; // Extract and untag. let tagged = atom.load; assert_eq!; assert_eq!;fn fetch_and(self: &Self, val: usize, order: Ordering) -> *mut TPerforms a bitwise "and" operation on the address of the current pointer, and the argument
val, and stores a pointer with provenance of the current pointer and the resulting address.This is equivalent to using
map_addrto atomically performptr = ptr.map_addr(|a| a & val). This can be used in tagged pointer schemes to atomically unset tag bits.Caveat: This operation returns the previous value. To compute the stored value without losing provenance, you may use
map_addr. For example:a.fetch_and(val).map_addr(|a| a & val).fetch_andtakes anOrderingargument which describes the memory ordering of this operation. All ordering modes are possible. Note that usingAcquiremakes the store part of this operationRelaxed, and usingReleasemakes the load partRelaxed.This API and its claimed semantics are part of the Strict Provenance experiment, see the [module documentation for
ptr][core::ptr] for details.Examples
# # use Strict; // strict provenance polyfill for old rustc use ; let pointer = &mut 3i64 as *mut i64; // A tagged pointer let atom = new; assert_eq!; // Untag, and extract the previously tagged pointer. let untagged = atom.fetch_and.map_addr; assert_eq!;fn fetch_xor(self: &Self, val: usize, order: Ordering) -> *mut TPerforms a bitwise "xor" operation on the address of the current pointer, and the argument
val, and stores a pointer with provenance of the current pointer and the resulting address.This is equivalent to using
map_addrto atomically performptr = ptr.map_addr(|a| a ^ val). This can be used in tagged pointer schemes to atomically toggle tag bits.Caveat: This operation returns the previous value. To compute the stored value without losing provenance, you may use
map_addr. For example:a.fetch_xor(val).map_addr(|a| a ^ val).fetch_xortakes anOrderingargument which describes the memory ordering of this operation. All ordering modes are possible. Note that usingAcquiremakes the store part of this operationRelaxed, and usingReleasemakes the load partRelaxed.This API and its claimed semantics are part of the Strict Provenance experiment, see the [module documentation for
ptr][core::ptr] for details.Examples
# # use Strict; // strict provenance polyfill for old rustc use ; let pointer = &mut 3i64 as *mut i64; let atom = new; // Toggle a tag bit on the pointer. atom.fetch_xor; assert_eq!;fn bit_set(self: &Self, bit: u32, order: Ordering) -> boolSets the bit at the specified bit-position to 1.
Returns
trueif the specified bit was previously set to 1.bit_settakes anOrderingargument which describes the memory ordering of this operation. All ordering modes are possible. Note that usingAcquiremakes the store part of this operationRelaxed, and usingReleasemakes the load partRelaxed.This corresponds to x86's
lock bts, and the implementation calls them on x86/x86_64.Examples
# # use Strict; // strict provenance polyfill for old rustc use ; let pointer = &mut 3i64 as *mut i64; let atom = new; // Tag the bottom bit of the pointer. assert!; // Extract and untag. let tagged = atom.load; assert_eq!; assert_eq!;fn bit_clear(self: &Self, bit: u32, order: Ordering) -> boolClears the bit at the specified bit-position to 1.
Returns
trueif the specified bit was previously set to 1.bit_cleartakes anOrderingargument which describes the memory ordering of this operation. All ordering modes are possible. Note that usingAcquiremakes the store part of this operationRelaxed, and usingReleasemakes the load partRelaxed.This corresponds to x86's
lock btr, and the implementation calls them on x86/x86_64.Examples
# # use Strict; // strict provenance polyfill for old rustc use ; let pointer = &mut 3i64 as *mut i64; // A tagged pointer let atom = new; assert!; // Untag assert!;fn bit_toggle(self: &Self, bit: u32, order: Ordering) -> boolToggles the bit at the specified bit-position.
Returns
trueif the specified bit was previously set to 1.bit_toggletakes anOrderingargument which describes the memory ordering of this operation. All ordering modes are possible. Note that usingAcquiremakes the store part of this operationRelaxed, and usingReleasemakes the load partRelaxed.This corresponds to x86's
lock btc, and the implementation calls them on x86/x86_64.Examples
# # use Strict; // strict provenance polyfill for old rustc use ; let pointer = &mut 3i64 as *mut i64; let atom = new; // Toggle a tag bit on the pointer. atom.bit_toggle; assert_eq!;const fn as_ptr(self: &Self) -> *mut *mut TReturns a mutable pointer to the underlying pointer.
Returning an
*mutpointer from a shared reference to this atomic is safe because the atomic types work with interior mutability. Any use of the returned raw pointer requires anunsafeblock and has to uphold the safety requirements. If there is concurrent access, note the following additional safety requirements:- If this atomic type is lock-free, any concurrent operations on it must be atomic.
- Otherwise, any concurrent operations on it must be compatible with operations performed by this atomic type.
This is
const fnon Rust 1.58+.
impl<T> Any for AtomicPtr<T>
fn type_id(self: &Self) -> TypeId
impl<T> Borrow for AtomicPtr<T>
fn borrow(self: &Self) -> &T
impl<T> BorrowMut for AtomicPtr<T>
fn borrow_mut(self: &mut Self) -> &mut T
impl<T> Debug for AtomicPtr<T>
fn fmt(self: &Self, f: &mut Formatter<'_>) -> Result
impl<T> Default for AtomicPtr<T>
fn default() -> SelfCreates a null
AtomicPtr<T>.
impl<T> Freeze for AtomicPtr<T>
impl<T> From for AtomicPtr<T>
fn from(p: *mut T) -> Self
impl<T> From for AtomicPtr<T>
fn from(t: T) -> TReturns the argument unchanged.
impl<T> Pointer for AtomicPtr<T>
fn fmt(self: &Self, f: &mut Formatter<'_>) -> Result
impl<T> RefUnwindSafe for AtomicPtr<T>
impl<T> Send for AtomicPtr<T>
impl<T> Sync for AtomicPtr<T>
impl<T> Unpin for AtomicPtr<T>
impl<T> UnsafeUnpin for AtomicPtr<T>
impl<T> UnwindSafe for AtomicPtr<T>
impl<T, U> Into for AtomicPtr<T>
fn into(self: Self) -> UCalls
U::from(self).That is, this conversion is whatever the implementation of
[From]<T> for Uchooses to do.
impl<T, U> TryFrom for AtomicPtr<T>
fn try_from(value: U) -> Result<T, <T as TryFrom<U>>::Error>
impl<T, U> TryInto for AtomicPtr<T>
fn try_into(self: Self) -> Result<U, <U as TryFrom<T>>::Error>