Struct webrtc_util::vnet::router::ROUTER_ID_CTR

source ·
pub struct ROUTER_ID_CTR { /* private fields */ }

Methods from Deref<Target = AtomicU64>§

source

pub fn load(&self, order: Ordering) -> u64

Loads a value from the atomic integer.

load takes an Ordering argument which describes the memory ordering of this operation. Possible values are [SeqCst], [Acquire] and [Relaxed].

§Panics

Panics if order is [Release] or [AcqRel].

§Examples
use portable_atomic::{AtomicU64, Ordering};

let some_var = AtomicU64::new(5);

assert_eq!(some_var.load(Ordering::Relaxed), 5);
source

pub fn store(&self, val: u64, order: Ordering)

Stores a value into the atomic integer.

store takes an Ordering argument which describes the memory ordering of this operation. Possible values are [SeqCst], [Release] and [Relaxed].

§Panics

Panics if order is [Acquire] or [AcqRel].

§Examples
use portable_atomic::{AtomicU64, Ordering};

let some_var = AtomicU64::new(5);

some_var.store(10, Ordering::Relaxed);
assert_eq!(some_var.load(Ordering::Relaxed), 10);
source

pub fn swap(&self, val: u64, order: Ordering) -> u64

Stores a value into the atomic integer, returning the previous value.

swap takes an Ordering argument which describes the memory ordering of this operation. All ordering modes are possible. Note that using [Acquire] makes the store part of this operation [Relaxed], and using [Release] makes the load part [Relaxed].

§Examples
use portable_atomic::{AtomicU64, Ordering};

let some_var = AtomicU64::new(5);

assert_eq!(some_var.swap(10, Ordering::Relaxed), 5);
source

pub fn compare_exchange( &self, current: u64, new: u64, success: Ordering, failure: Ordering ) -> Result<u64, u64>

Stores a value into the atomic integer if the current value is the same as the current value.

The return value is a result indicating whether the new value was written and containing the previous value. On success this value is guaranteed to be equal to current.

compare_exchange takes two Ordering arguments to describe the memory ordering of this operation. success describes the required ordering for the read-modify-write operation that takes place if the comparison with current succeeds. failure describes the required ordering for the load operation that takes place when the comparison fails. Using [Acquire] as success ordering makes the store part of this operation [Relaxed], and using [Release] makes the successful load [Relaxed]. The failure ordering can only be [SeqCst], [Acquire] or [Relaxed].

§Panics

Panics if failure is [Release], [AcqRel].

§Examples
use portable_atomic::{AtomicU64, Ordering};

let some_var = AtomicU64::new(5);

assert_eq!(
    some_var.compare_exchange(5, 10, Ordering::Acquire, Ordering::Relaxed),
    Ok(5),
);
assert_eq!(some_var.load(Ordering::Relaxed), 10);

assert_eq!(
    some_var.compare_exchange(6, 12, Ordering::SeqCst, Ordering::Acquire),
    Err(10),
);
assert_eq!(some_var.load(Ordering::Relaxed), 10);
source

pub fn compare_exchange_weak( &self, current: u64, new: u64, success: Ordering, failure: Ordering ) -> Result<u64, u64>

Stores a value into the atomic integer if the current value is the same as the current value. Unlike compare_exchange this function is allowed to spuriously fail even when the comparison succeeds, which can result in more efficient code on some platforms. The return value is a result indicating whether the new value was written and containing the previous value.

compare_exchange_weak takes two Ordering arguments to describe the memory ordering of this operation. success describes the required ordering for the read-modify-write operation that takes place if the comparison with current succeeds. failure describes the required ordering for the load operation that takes place when the comparison fails. Using [Acquire] as success ordering makes the store part of this operation [Relaxed], and using [Release] makes the successful load [Relaxed]. The failure ordering can only be [SeqCst], [Acquire] or [Relaxed].

§Panics

Panics if failure is [Release], [AcqRel].

§Examples
use portable_atomic::{AtomicU64, Ordering};

let val = AtomicU64::new(4);

let mut old = val.load(Ordering::Relaxed);
loop {
    let new = old * 2;
    match val.compare_exchange_weak(old, new, Ordering::SeqCst, Ordering::Relaxed) {
        Ok(_) => break,
        Err(x) => old = x,
    }
}
source

pub fn fetch_add(&self, val: u64, order: Ordering) -> u64

Adds to the current value, returning the previous value.

This operation wraps around on overflow.

fetch_add takes an Ordering argument which describes the memory ordering of this operation. All ordering modes are possible. Note that using [Acquire] makes the store part of this operation [Relaxed], and using [Release] makes the load part [Relaxed].

§Examples
use portable_atomic::{AtomicU64, Ordering};

let foo = AtomicU64::new(0);
assert_eq!(foo.fetch_add(10, Ordering::SeqCst), 0);
assert_eq!(foo.load(Ordering::SeqCst), 10);
source

pub fn add(&self, val: u64, order: Ordering)

Adds to the current value.

This operation wraps around on overflow.

Unlike fetch_add, this does not return the previous value.

add takes an Ordering argument which describes the memory ordering of this operation. All ordering modes are possible. Note that using [Acquire] makes the store part of this operation [Relaxed], and using [Release] makes the load part [Relaxed].

This function may generate more efficient code than fetch_add on some platforms.

  • MSP430: add instead of disabling interrupts ({8,16}-bit atomics)
§Examples
use portable_atomic::{AtomicU64, Ordering};

let foo = AtomicU64::new(0);
foo.add(10, Ordering::SeqCst);
assert_eq!(foo.load(Ordering::SeqCst), 10);
source

pub fn fetch_sub(&self, val: u64, order: Ordering) -> u64

Subtracts from the current value, returning the previous value.

This operation wraps around on overflow.

fetch_sub takes an Ordering argument which describes the memory ordering of this operation. All ordering modes are possible. Note that using [Acquire] makes the store part of this operation [Relaxed], and using [Release] makes the load part [Relaxed].

§Examples
use portable_atomic::{AtomicU64, Ordering};

let foo = AtomicU64::new(20);
assert_eq!(foo.fetch_sub(10, Ordering::SeqCst), 20);
assert_eq!(foo.load(Ordering::SeqCst), 10);
source

pub fn sub(&self, val: u64, order: Ordering)

Subtracts from the current value.

This operation wraps around on overflow.

Unlike fetch_sub, this does not return the previous value.

sub takes an Ordering argument which describes the memory ordering of this operation. All ordering modes are possible. Note that using [Acquire] makes the store part of this operation [Relaxed], and using [Release] makes the load part [Relaxed].

This function may generate more efficient code than fetch_sub on some platforms.

  • MSP430: sub instead of disabling interrupts ({8,16}-bit atomics)
§Examples
use portable_atomic::{AtomicU64, Ordering};

let foo = AtomicU64::new(20);
foo.sub(10, Ordering::SeqCst);
assert_eq!(foo.load(Ordering::SeqCst), 10);
source

pub fn fetch_and(&self, val: u64, order: Ordering) -> u64

Bitwise “and” with the current value.

Performs a bitwise “and” operation on the current value and the argument val, and sets the new value to the result.

Returns the previous value.

fetch_and takes an Ordering argument which describes the memory ordering of this operation. All ordering modes are possible. Note that using [Acquire] makes the store part of this operation [Relaxed], and using [Release] makes the load part [Relaxed].

§Examples
use portable_atomic::{AtomicU64, Ordering};

let foo = AtomicU64::new(0b101101);
assert_eq!(foo.fetch_and(0b110011, Ordering::SeqCst), 0b101101);
assert_eq!(foo.load(Ordering::SeqCst), 0b100001);
source

pub fn and(&self, val: u64, order: Ordering)

Bitwise “and” with the current value.

Performs a bitwise “and” operation on the current value and the argument val, and sets the new value to the result.

Unlike fetch_and, this does not return the previous value.

and takes an Ordering argument which describes the memory ordering of this operation. All ordering modes are possible. Note that using [Acquire] makes the store part of this operation [Relaxed], and using [Release] makes the load part [Relaxed].

This function may generate more efficient code than fetch_and on some platforms.

  • x86/x86_64: lock and instead of cmpxchg loop ({8,16,32}-bit atomics on x86, but additionally 64-bit atomics on x86_64)
  • MSP430: and instead of disabling interrupts ({8,16}-bit atomics)

Note: On x86/x86_64, the use of either function should not usually affect the generated code, because LLVM can properly optimize the case where the result is unused.

§Examples
use portable_atomic::{AtomicU64, Ordering};

let foo = AtomicU64::new(0b101101);
assert_eq!(foo.fetch_and(0b110011, Ordering::SeqCst), 0b101101);
assert_eq!(foo.load(Ordering::SeqCst), 0b100001);
source

pub fn fetch_nand(&self, val: u64, order: Ordering) -> u64

Bitwise “nand” with the current value.

Performs a bitwise “nand” operation on the current value and the argument val, and sets the new value to the result.

Returns the previous value.

fetch_nand takes an Ordering argument which describes the memory ordering of this operation. All ordering modes are possible. Note that using [Acquire] makes the store part of this operation [Relaxed], and using [Release] makes the load part [Relaxed].

§Examples
use portable_atomic::{AtomicU64, Ordering};

let foo = AtomicU64::new(0x13);
assert_eq!(foo.fetch_nand(0x31, Ordering::SeqCst), 0x13);
assert_eq!(foo.load(Ordering::SeqCst), !(0x13 & 0x31));
source

pub fn fetch_or(&self, val: u64, order: Ordering) -> u64

Bitwise “or” with the current value.

Performs a bitwise “or” operation on the current value and the argument val, and sets the new value to the result.

Returns the previous value.

fetch_or takes an Ordering argument which describes the memory ordering of this operation. All ordering modes are possible. Note that using [Acquire] makes the store part of this operation [Relaxed], and using [Release] makes the load part [Relaxed].

§Examples
use portable_atomic::{AtomicU64, Ordering};

let foo = AtomicU64::new(0b101101);
assert_eq!(foo.fetch_or(0b110011, Ordering::SeqCst), 0b101101);
assert_eq!(foo.load(Ordering::SeqCst), 0b111111);
source

pub fn or(&self, val: u64, order: Ordering)

Bitwise “or” with the current value.

Performs a bitwise “or” operation on the current value and the argument val, and sets the new value to the result.

Unlike fetch_or, this does not return the previous value.

or takes an Ordering argument which describes the memory ordering of this operation. All ordering modes are possible. Note that using [Acquire] makes the store part of this operation [Relaxed], and using [Release] makes the load part [Relaxed].

This function may generate more efficient code than fetch_or on some platforms.

  • x86/x86_64: lock or instead of cmpxchg loop ({8,16,32}-bit atomics on x86, but additionally 64-bit atomics on x86_64)
  • MSP430: or instead of disabling interrupts ({8,16}-bit atomics)

Note: On x86/x86_64, the use of either function should not usually affect the generated code, because LLVM can properly optimize the case where the result is unused.

§Examples
use portable_atomic::{AtomicU64, Ordering};

let foo = AtomicU64::new(0b101101);
assert_eq!(foo.fetch_or(0b110011, Ordering::SeqCst), 0b101101);
assert_eq!(foo.load(Ordering::SeqCst), 0b111111);
source

pub fn fetch_xor(&self, val: u64, order: Ordering) -> u64

Bitwise “xor” with the current value.

Performs a bitwise “xor” operation on the current value and the argument val, and sets the new value to the result.

Returns the previous value.

fetch_xor takes an Ordering argument which describes the memory ordering of this operation. All ordering modes are possible. Note that using [Acquire] makes the store part of this operation [Relaxed], and using [Release] makes the load part [Relaxed].

§Examples
use portable_atomic::{AtomicU64, Ordering};

let foo = AtomicU64::new(0b101101);
assert_eq!(foo.fetch_xor(0b110011, Ordering::SeqCst), 0b101101);
assert_eq!(foo.load(Ordering::SeqCst), 0b011110);
source

pub fn xor(&self, val: u64, order: Ordering)

Bitwise “xor” with the current value.

Performs a bitwise “xor” operation on the current value and the argument val, and sets the new value to the result.

Unlike fetch_xor, this does not return the previous value.

xor takes an Ordering argument which describes the memory ordering of this operation. All ordering modes are possible. Note that using [Acquire] makes the store part of this operation [Relaxed], and using [Release] makes the load part [Relaxed].

This function may generate more efficient code than fetch_xor on some platforms.

  • x86/x86_64: lock xor instead of cmpxchg loop ({8,16,32}-bit atomics on x86, but additionally 64-bit atomics on x86_64)
  • MSP430: xor instead of disabling interrupts ({8,16}-bit atomics)

Note: On x86/x86_64, the use of either function should not usually affect the generated code, because LLVM can properly optimize the case where the result is unused.

§Examples
use portable_atomic::{AtomicU64, Ordering};

let foo = AtomicU64::new(0b101101);
foo.xor(0b110011, Ordering::SeqCst);
assert_eq!(foo.load(Ordering::SeqCst), 0b011110);
source

pub fn fetch_update<F>( &self, set_order: Ordering, fetch_order: Ordering, f: F ) -> Result<u64, u64>
where F: FnMut(u64) -> Option<u64>,

Fetches the value, and applies a function to it that returns an optional new value. Returns a Result of Ok(previous_value) if the function returned Some(_), else Err(previous_value).

Note: This may call the function multiple times if the value has been changed from other threads in the meantime, as long as the function returns Some(_), but the function will have been applied only once to the stored value.

fetch_update takes two Ordering arguments to describe the memory ordering of this operation. The first describes the required ordering for when the operation finally succeeds while the second describes the required ordering for loads. These correspond to the success and failure orderings of compare_exchange respectively.

Using [Acquire] as success ordering makes the store part of this operation [Relaxed], and using [Release] makes the final successful load [Relaxed]. The (failed) load ordering can only be [SeqCst], [Acquire] or [Relaxed].

§Panics

Panics if fetch_order is [Release], [AcqRel].

§Considerations

This method is not magic; it is not provided by the hardware. It is implemented in terms of compare_exchange_weak, and suffers from the same drawbacks. In particular, this method will not circumvent the ABA Problem.

§Examples
use portable_atomic::{AtomicU64, Ordering};

let x = AtomicU64::new(7);
assert_eq!(x.fetch_update(Ordering::SeqCst, Ordering::SeqCst, |_| None), Err(7));
assert_eq!(x.fetch_update(Ordering::SeqCst, Ordering::SeqCst, |x| Some(x + 1)), Ok(7));
assert_eq!(x.fetch_update(Ordering::SeqCst, Ordering::SeqCst, |x| Some(x + 1)), Ok(8));
assert_eq!(x.load(Ordering::SeqCst), 9);
source

pub fn fetch_max(&self, val: u64, order: Ordering) -> u64

Maximum with the current value.

Finds the maximum of the current value and the argument val, and sets the new value to the result.

Returns the previous value.

fetch_max takes an Ordering argument which describes the memory ordering of this operation. All ordering modes are possible. Note that using [Acquire] makes the store part of this operation [Relaxed], and using [Release] makes the load part [Relaxed].

§Examples
use portable_atomic::{AtomicU64, Ordering};

let foo = AtomicU64::new(23);
assert_eq!(foo.fetch_max(42, Ordering::SeqCst), 23);
assert_eq!(foo.load(Ordering::SeqCst), 42);

If you want to obtain the maximum value in one step, you can use the following:

use portable_atomic::{AtomicU64, Ordering};

let foo = AtomicU64::new(23);
let bar = 42;
let max_foo = foo.fetch_max(bar, Ordering::SeqCst).max(bar);
assert!(max_foo == 42);
source

pub fn fetch_min(&self, val: u64, order: Ordering) -> u64

Minimum with the current value.

Finds the minimum of the current value and the argument val, and sets the new value to the result.

Returns the previous value.

fetch_min takes an Ordering argument which describes the memory ordering of this operation. All ordering modes are possible. Note that using [Acquire] makes the store part of this operation [Relaxed], and using [Release] makes the load part [Relaxed].

§Examples
use portable_atomic::{AtomicU64, Ordering};

let foo = AtomicU64::new(23);
assert_eq!(foo.fetch_min(42, Ordering::Relaxed), 23);
assert_eq!(foo.load(Ordering::Relaxed), 23);
assert_eq!(foo.fetch_min(22, Ordering::Relaxed), 23);
assert_eq!(foo.load(Ordering::Relaxed), 22);

If you want to obtain the minimum value in one step, you can use the following:

use portable_atomic::{AtomicU64, Ordering};

let foo = AtomicU64::new(23);
let bar = 12;
let min_foo = foo.fetch_min(bar, Ordering::SeqCst).min(bar);
assert_eq!(min_foo, 12);
source

pub fn bit_set(&self, bit: u32, order: Ordering) -> bool

Sets the bit at the specified bit-position to 1.

Returns true if the specified bit was previously set to 1.

bit_set takes an Ordering argument which describes the memory ordering of this operation. All ordering modes are possible. Note that using [Acquire] makes the store part of this operation [Relaxed], and using [Release] makes the load part [Relaxed].

This corresponds to x86’s lock bts, and the implementation calls them on x86/x86_64.

§Examples
use portable_atomic::{AtomicU64, Ordering};

let foo = AtomicU64::new(0b0000);
assert!(!foo.bit_set(0, Ordering::Relaxed));
assert_eq!(foo.load(Ordering::Relaxed), 0b0001);
assert!(foo.bit_set(0, Ordering::Relaxed));
assert_eq!(foo.load(Ordering::Relaxed), 0b0001);
source

pub fn bit_clear(&self, bit: u32, order: Ordering) -> bool

Clears the bit at the specified bit-position to 1.

Returns true if the specified bit was previously set to 1.

bit_clear takes an Ordering argument which describes the memory ordering of this operation. All ordering modes are possible. Note that using [Acquire] makes the store part of this operation [Relaxed], and using [Release] makes the load part [Relaxed].

This corresponds to x86’s lock btr, and the implementation calls them on x86/x86_64.

§Examples
use portable_atomic::{AtomicU64, Ordering};

let foo = AtomicU64::new(0b0001);
assert!(foo.bit_clear(0, Ordering::Relaxed));
assert_eq!(foo.load(Ordering::Relaxed), 0b0000);
source

pub fn bit_toggle(&self, bit: u32, order: Ordering) -> bool

Toggles the bit at the specified bit-position.

Returns true if the specified bit was previously set to 1.

bit_toggle takes an Ordering argument which describes the memory ordering of this operation. All ordering modes are possible. Note that using [Acquire] makes the store part of this operation [Relaxed], and using [Release] makes the load part [Relaxed].

This corresponds to x86’s lock btc, and the implementation calls them on x86/x86_64.

§Examples
use portable_atomic::{AtomicU64, Ordering};

let foo = AtomicU64::new(0b0000);
assert!(!foo.bit_toggle(0, Ordering::Relaxed));
assert_eq!(foo.load(Ordering::Relaxed), 0b0001);
assert!(foo.bit_toggle(0, Ordering::Relaxed));
assert_eq!(foo.load(Ordering::Relaxed), 0b0000);
source

pub fn fetch_not(&self, order: Ordering) -> u64

Logical negates the current value, and sets the new value to the result.

Returns the previous value.

fetch_not takes an Ordering argument which describes the memory ordering of this operation. All ordering modes are possible. Note that using [Acquire] makes the store part of this operation [Relaxed], and using [Release] makes the load part [Relaxed].

§Examples
use portable_atomic::{AtomicU64, Ordering};

let foo = AtomicU64::new(0);
assert_eq!(foo.fetch_not(Ordering::Relaxed), 0);
assert_eq!(foo.load(Ordering::Relaxed), !0);
source

pub fn not(&self, order: Ordering)

Logical negates the current value, and sets the new value to the result.

Unlike fetch_not, this does not return the previous value.

not takes an Ordering argument which describes the memory ordering of this operation. All ordering modes are possible. Note that using [Acquire] makes the store part of this operation [Relaxed], and using [Release] makes the load part [Relaxed].

This function may generate more efficient code than fetch_not on some platforms.

  • x86/x86_64: lock not instead of cmpxchg loop ({8,16,32}-bit atomics on x86, but additionally 64-bit atomics on x86_64)
  • MSP430: inv instead of disabling interrupts ({8,16}-bit atomics)
§Examples
use portable_atomic::{AtomicU64, Ordering};

let foo = AtomicU64::new(0);
foo.not(Ordering::Relaxed);
assert_eq!(foo.load(Ordering::Relaxed), !0);
source

pub fn fetch_neg(&self, order: Ordering) -> u64

Negates the current value, and sets the new value to the result.

Returns the previous value.

fetch_neg takes an Ordering argument which describes the memory ordering of this operation. All ordering modes are possible. Note that using [Acquire] makes the store part of this operation [Relaxed], and using [Release] makes the load part [Relaxed].

§Examples
use portable_atomic::{AtomicU64, Ordering};

let foo = AtomicU64::new(5);
assert_eq!(foo.fetch_neg(Ordering::Relaxed), 5);
assert_eq!(foo.load(Ordering::Relaxed), 5_u64.wrapping_neg());
assert_eq!(foo.fetch_neg(Ordering::Relaxed), 5_u64.wrapping_neg());
assert_eq!(foo.load(Ordering::Relaxed), 5);
source

pub fn neg(&self, order: Ordering)

Negates the current value, and sets the new value to the result.

Unlike fetch_neg, this does not return the previous value.

neg takes an Ordering argument which describes the memory ordering of this operation. All ordering modes are possible. Note that using [Acquire] makes the store part of this operation [Relaxed], and using [Release] makes the load part [Relaxed].

This function may generate more efficient code than fetch_neg on some platforms.

  • x86/x86_64: lock neg instead of cmpxchg loop ({8,16,32}-bit atomics on x86, but additionally 64-bit atomics on x86_64)
§Examples
use portable_atomic::{AtomicU64, Ordering};

let foo = AtomicU64::new(5);
foo.neg(Ordering::Relaxed);
assert_eq!(foo.load(Ordering::Relaxed), 5_u64.wrapping_neg());
foo.neg(Ordering::Relaxed);
assert_eq!(foo.load(Ordering::Relaxed), 5);
source

pub fn as_ptr(&self) -> *mut u64

Returns a mutable pointer to the underlying integer.

Returning an *mut pointer from a shared reference to this atomic is safe because the atomic types work with interior mutability. Any use of the returned raw pointer requires an unsafe block and has to uphold the safety requirements. If there is concurrent access, note the following additional safety requirements:

  • If this atomic type is lock-free, any concurrent operations on it must be atomic.
  • Otherwise, any concurrent operations on it must be compatible with operations performed by this atomic type.

This is const fn on Rust 1.58+.

Trait Implementations§

source§

impl Deref for ROUTER_ID_CTR

§

type Target = AtomicU64

The resulting type after dereferencing.
source§

fn deref(&self) -> &AtomicU64

Dereferences the value.
source§

impl LazyStatic for ROUTER_ID_CTR

Auto Trait Implementations§

Blanket Implementations§

source§

impl<T> Any for T
where T: 'static + ?Sized,

source§

fn type_id(&self) -> TypeId

Gets the TypeId of self. Read more
source§

impl<T> Borrow<T> for T
where T: ?Sized,

source§

fn borrow(&self) -> &T

Immutably borrows from an owned value. Read more
source§

impl<T> BorrowMut<T> for T
where T: ?Sized,

source§

fn borrow_mut(&mut self) -> &mut T

Mutably borrows from an owned value. Read more
source§

impl<T> From<T> for T

source§

fn from(t: T) -> T

Returns the argument unchanged.

source§

impl<T, U> Into<U> for T
where U: From<T>,

source§

fn into(self) -> U

Calls U::from(self).

That is, this conversion is whatever the implementation of From<T> for U chooses to do.

source§

impl<T, U> TryFrom<U> for T
where U: Into<T>,

§

type Error = Infallible

The type returned in the event of a conversion error.
source§

fn try_from(value: U) -> Result<T, <T as TryFrom<U>>::Error>

Performs the conversion.
source§

impl<T, U> TryInto<U> for T
where U: TryFrom<T>,

§

type Error = <U as TryFrom<T>>::Error

The type returned in the event of a conversion error.
source§

fn try_into(self) -> Result<U, <U as TryFrom<T>>::Error>

Performs the conversion.
source§

impl<V, T> VZip<V> for T
where V: MultiLane<T>,

source§

fn vzip(self) -> V