更新libclamav库1.0.0版本
This commit is contained in:
75
clamav/libclamav_rust/.cargo/vendor/indexmap/src/arbitrary.rs
vendored
Normal file
75
clamav/libclamav_rust/.cargo/vendor/indexmap/src/arbitrary.rs
vendored
Normal file
@@ -0,0 +1,75 @@
|
||||
#[cfg(feature = "arbitrary")]
|
||||
mod impl_arbitrary {
|
||||
use crate::{IndexMap, IndexSet};
|
||||
use arbitrary::{Arbitrary, Result, Unstructured};
|
||||
use core::hash::{BuildHasher, Hash};
|
||||
|
||||
impl<'a, K, V, S> Arbitrary<'a> for IndexMap<K, V, S>
|
||||
where
|
||||
K: Arbitrary<'a> + Hash + Eq,
|
||||
V: Arbitrary<'a>,
|
||||
S: BuildHasher + Default,
|
||||
{
|
||||
fn arbitrary(u: &mut Unstructured<'a>) -> Result<Self> {
|
||||
u.arbitrary_iter()?.collect()
|
||||
}
|
||||
|
||||
fn arbitrary_take_rest(u: Unstructured<'a>) -> Result<Self> {
|
||||
u.arbitrary_take_rest_iter()?.collect()
|
||||
}
|
||||
}
|
||||
|
||||
impl<'a, T, S> Arbitrary<'a> for IndexSet<T, S>
|
||||
where
|
||||
T: Arbitrary<'a> + Hash + Eq,
|
||||
S: BuildHasher + Default,
|
||||
{
|
||||
fn arbitrary(u: &mut Unstructured<'a>) -> Result<Self> {
|
||||
u.arbitrary_iter()?.collect()
|
||||
}
|
||||
|
||||
fn arbitrary_take_rest(u: Unstructured<'a>) -> Result<Self> {
|
||||
u.arbitrary_take_rest_iter()?.collect()
|
||||
}
|
||||
}
|
||||
}
|
||||
|
||||
#[cfg(feature = "quickcheck")]
|
||||
mod impl_quickcheck {
|
||||
use crate::{IndexMap, IndexSet};
|
||||
use alloc::boxed::Box;
|
||||
use alloc::vec::Vec;
|
||||
use core::hash::{BuildHasher, Hash};
|
||||
use quickcheck::{Arbitrary, Gen};
|
||||
|
||||
impl<K, V, S> Arbitrary for IndexMap<K, V, S>
|
||||
where
|
||||
K: Arbitrary + Hash + Eq,
|
||||
V: Arbitrary,
|
||||
S: BuildHasher + Default + Clone + 'static,
|
||||
{
|
||||
fn arbitrary(g: &mut Gen) -> Self {
|
||||
Self::from_iter(Vec::arbitrary(g))
|
||||
}
|
||||
|
||||
fn shrink(&self) -> Box<dyn Iterator<Item = Self>> {
|
||||
let vec = Vec::from_iter(self.clone());
|
||||
Box::new(vec.shrink().map(Self::from_iter))
|
||||
}
|
||||
}
|
||||
|
||||
impl<T, S> Arbitrary for IndexSet<T, S>
|
||||
where
|
||||
T: Arbitrary + Hash + Eq,
|
||||
S: BuildHasher + Default + Clone + 'static,
|
||||
{
|
||||
fn arbitrary(g: &mut Gen) -> Self {
|
||||
Self::from_iter(Vec::arbitrary(g))
|
||||
}
|
||||
|
||||
fn shrink(&self) -> Box<dyn Iterator<Item = Self>> {
|
||||
let vec = Vec::from_iter(self.clone());
|
||||
Box::new(vec.shrink().map(Self::from_iter))
|
||||
}
|
||||
}
|
||||
}
|
||||
27
clamav/libclamav_rust/.cargo/vendor/indexmap/src/equivalent.rs
vendored
Normal file
27
clamav/libclamav_rust/.cargo/vendor/indexmap/src/equivalent.rs
vendored
Normal file
@@ -0,0 +1,27 @@
|
||||
use core::borrow::Borrow;
|
||||
|
||||
/// Key equivalence trait.
|
||||
///
|
||||
/// This trait allows hash table lookup to be customized.
|
||||
/// It has one blanket implementation that uses the regular `Borrow` solution,
|
||||
/// just like `HashMap` and `BTreeMap` do, so that you can pass `&str` to lookup
|
||||
/// into a map with `String` keys and so on.
|
||||
///
|
||||
/// # Contract
|
||||
///
|
||||
/// The implementor **must** hash like `K`, if it is hashable.
|
||||
pub trait Equivalent<K: ?Sized> {
|
||||
/// Compare self to `key` and return `true` if they are equal.
|
||||
fn equivalent(&self, key: &K) -> bool;
|
||||
}
|
||||
|
||||
impl<Q: ?Sized, K: ?Sized> Equivalent<K> for Q
|
||||
where
|
||||
Q: Eq,
|
||||
K: Borrow<Q>,
|
||||
{
|
||||
#[inline]
|
||||
fn equivalent(&self, key: &K) -> bool {
|
||||
*self == *key.borrow()
|
||||
}
|
||||
}
|
||||
194
clamav/libclamav_rust/.cargo/vendor/indexmap/src/lib.rs
vendored
Normal file
194
clamav/libclamav_rust/.cargo/vendor/indexmap/src/lib.rs
vendored
Normal file
@@ -0,0 +1,194 @@
|
||||
// We *mostly* avoid unsafe code, but `map::core::raw` allows it to use `RawTable` buckets.
|
||||
#![deny(unsafe_code)]
|
||||
#![warn(rust_2018_idioms)]
|
||||
#![doc(html_root_url = "https://docs.rs/indexmap/1/")]
|
||||
#![no_std]
|
||||
|
||||
//! [`IndexMap`] is a hash table where the iteration order of the key-value
|
||||
//! pairs is independent of the hash values of the keys.
|
||||
//!
|
||||
//! [`IndexSet`] is a corresponding hash set using the same implementation and
|
||||
//! with similar properties.
|
||||
//!
|
||||
//! [`IndexMap`]: map/struct.IndexMap.html
|
||||
//! [`IndexSet`]: set/struct.IndexSet.html
|
||||
//!
|
||||
//!
|
||||
//! ### Feature Highlights
|
||||
//!
|
||||
//! [`IndexMap`] and [`IndexSet`] are drop-in compatible with the std `HashMap`
|
||||
//! and `HashSet`, but they also have some features of note:
|
||||
//!
|
||||
//! - The ordering semantics (see their documentation for details)
|
||||
//! - Sorting methods and the [`.pop()`][IndexMap::pop] methods.
|
||||
//! - The [`Equivalent`] trait, which offers more flexible equality definitions
|
||||
//! between borrowed and owned versions of keys.
|
||||
//! - The [`MutableKeys`][map::MutableKeys] trait, which gives opt-in mutable
|
||||
//! access to hash map keys.
|
||||
//!
|
||||
//! ### Alternate Hashers
|
||||
//!
|
||||
//! [`IndexMap`] and [`IndexSet`] have a default hasher type `S = RandomState`,
|
||||
//! just like the standard `HashMap` and `HashSet`, which is resistant to
|
||||
//! HashDoS attacks but not the most performant. Type aliases can make it easier
|
||||
//! to use alternate hashers:
|
||||
//!
|
||||
//! ```
|
||||
//! use fnv::FnvBuildHasher;
|
||||
//! use fxhash::FxBuildHasher;
|
||||
//! use indexmap::{IndexMap, IndexSet};
|
||||
//!
|
||||
//! type FnvIndexMap<K, V> = IndexMap<K, V, FnvBuildHasher>;
|
||||
//! type FnvIndexSet<T> = IndexSet<T, FnvBuildHasher>;
|
||||
//!
|
||||
//! type FxIndexMap<K, V> = IndexMap<K, V, FxBuildHasher>;
|
||||
//! type FxIndexSet<T> = IndexSet<T, FxBuildHasher>;
|
||||
//!
|
||||
//! let std: IndexSet<i32> = (0..100).collect();
|
||||
//! let fnv: FnvIndexSet<i32> = (0..100).collect();
|
||||
//! let fx: FxIndexSet<i32> = (0..100).collect();
|
||||
//! assert_eq!(std, fnv);
|
||||
//! assert_eq!(std, fx);
|
||||
//! ```
|
||||
//!
|
||||
//! ### Rust Version
|
||||
//!
|
||||
//! This version of indexmap requires Rust 1.56 or later.
|
||||
//!
|
||||
//! The indexmap 1.x release series will use a carefully considered version
|
||||
//! upgrade policy, where in a later 1.x version, we will raise the minimum
|
||||
//! required Rust version.
|
||||
//!
|
||||
//! ## No Standard Library Targets
|
||||
//!
|
||||
//! This crate supports being built without `std`, requiring
|
||||
//! `alloc` instead. This is enabled automatically when it is detected that
|
||||
//! `std` is not available. There is no crate feature to enable/disable to
|
||||
//! trigger this. It can be tested by building for a std-less target.
|
||||
//!
|
||||
//! - Creating maps and sets using [`new`][IndexMap::new] and
|
||||
//! [`with_capacity`][IndexMap::with_capacity] is unavailable without `std`.
|
||||
//! Use methods [`IndexMap::default`][def],
|
||||
//! [`with_hasher`][IndexMap::with_hasher],
|
||||
//! [`with_capacity_and_hasher`][IndexMap::with_capacity_and_hasher] instead.
|
||||
//! A no-std compatible hasher will be needed as well, for example
|
||||
//! from the crate `twox-hash`.
|
||||
//! - Macros [`indexmap!`] and [`indexset!`] are unavailable without `std`.
|
||||
//!
|
||||
//! [def]: map/struct.IndexMap.html#impl-Default
|
||||
|
||||
extern crate alloc;
|
||||
|
||||
#[cfg(has_std)]
|
||||
#[macro_use]
|
||||
extern crate std;
|
||||
|
||||
use alloc::vec::{self, Vec};
|
||||
|
||||
mod arbitrary;
|
||||
#[macro_use]
|
||||
mod macros;
|
||||
mod equivalent;
|
||||
mod mutable_keys;
|
||||
#[cfg(feature = "serde")]
|
||||
mod serde;
|
||||
#[cfg(feature = "serde")]
|
||||
pub mod serde_seq;
|
||||
mod util;
|
||||
|
||||
pub mod map;
|
||||
pub mod set;
|
||||
|
||||
// Placed after `map` and `set` so new `rayon` methods on the types
|
||||
// are documented after the "normal" methods.
|
||||
#[cfg(feature = "rayon")]
|
||||
mod rayon;
|
||||
|
||||
#[cfg(feature = "rustc-rayon")]
|
||||
mod rustc;
|
||||
|
||||
pub use crate::equivalent::Equivalent;
|
||||
pub use crate::map::IndexMap;
|
||||
pub use crate::set::IndexSet;
|
||||
|
||||
// shared private items
|
||||
|
||||
/// Hash value newtype. Not larger than usize, since anything larger
|
||||
/// isn't used for selecting position anyway.
|
||||
#[derive(Clone, Copy, Debug, PartialEq)]
|
||||
struct HashValue(usize);
|
||||
|
||||
impl HashValue {
|
||||
#[inline(always)]
|
||||
fn get(self) -> u64 {
|
||||
self.0 as u64
|
||||
}
|
||||
}
|
||||
|
||||
#[derive(Copy, Debug)]
|
||||
struct Bucket<K, V> {
|
||||
hash: HashValue,
|
||||
key: K,
|
||||
value: V,
|
||||
}
|
||||
|
||||
impl<K, V> Clone for Bucket<K, V>
|
||||
where
|
||||
K: Clone,
|
||||
V: Clone,
|
||||
{
|
||||
fn clone(&self) -> Self {
|
||||
Bucket {
|
||||
hash: self.hash,
|
||||
key: self.key.clone(),
|
||||
value: self.value.clone(),
|
||||
}
|
||||
}
|
||||
|
||||
fn clone_from(&mut self, other: &Self) {
|
||||
self.hash = other.hash;
|
||||
self.key.clone_from(&other.key);
|
||||
self.value.clone_from(&other.value);
|
||||
}
|
||||
}
|
||||
|
||||
impl<K, V> Bucket<K, V> {
|
||||
// field accessors -- used for `f` instead of closures in `.map(f)`
|
||||
fn key_ref(&self) -> &K {
|
||||
&self.key
|
||||
}
|
||||
fn value_ref(&self) -> &V {
|
||||
&self.value
|
||||
}
|
||||
fn value_mut(&mut self) -> &mut V {
|
||||
&mut self.value
|
||||
}
|
||||
fn key(self) -> K {
|
||||
self.key
|
||||
}
|
||||
fn value(self) -> V {
|
||||
self.value
|
||||
}
|
||||
fn key_value(self) -> (K, V) {
|
||||
(self.key, self.value)
|
||||
}
|
||||
fn refs(&self) -> (&K, &V) {
|
||||
(&self.key, &self.value)
|
||||
}
|
||||
fn ref_mut(&mut self) -> (&K, &mut V) {
|
||||
(&self.key, &mut self.value)
|
||||
}
|
||||
fn muts(&mut self) -> (&mut K, &mut V) {
|
||||
(&mut self.key, &mut self.value)
|
||||
}
|
||||
}
|
||||
|
||||
trait Entries {
|
||||
type Entry;
|
||||
fn into_entries(self) -> Vec<Self::Entry>;
|
||||
fn as_entries(&self) -> &[Self::Entry];
|
||||
fn as_entries_mut(&mut self) -> &mut [Self::Entry];
|
||||
fn with_entries<F>(&mut self, f: F)
|
||||
where
|
||||
F: FnOnce(&mut [Self::Entry]);
|
||||
}
|
||||
178
clamav/libclamav_rust/.cargo/vendor/indexmap/src/macros.rs
vendored
Normal file
178
clamav/libclamav_rust/.cargo/vendor/indexmap/src/macros.rs
vendored
Normal file
@@ -0,0 +1,178 @@
|
||||
#[cfg(has_std)]
|
||||
#[macro_export]
|
||||
/// Create an `IndexMap` from a list of key-value pairs
|
||||
///
|
||||
/// ## Example
|
||||
///
|
||||
/// ```
|
||||
/// use indexmap::indexmap;
|
||||
///
|
||||
/// let map = indexmap!{
|
||||
/// "a" => 1,
|
||||
/// "b" => 2,
|
||||
/// };
|
||||
/// assert_eq!(map["a"], 1);
|
||||
/// assert_eq!(map["b"], 2);
|
||||
/// assert_eq!(map.get("c"), None);
|
||||
///
|
||||
/// // "a" is the first key
|
||||
/// assert_eq!(map.keys().next(), Some(&"a"));
|
||||
/// ```
|
||||
macro_rules! indexmap {
|
||||
(@single $($x:tt)*) => (());
|
||||
(@count $($rest:expr),*) => (<[()]>::len(&[$($crate::indexmap!(@single $rest)),*]));
|
||||
|
||||
($($key:expr => $value:expr,)+) => { $crate::indexmap!($($key => $value),+) };
|
||||
($($key:expr => $value:expr),*) => {
|
||||
{
|
||||
let _cap = $crate::indexmap!(@count $($key),*);
|
||||
let mut _map = $crate::IndexMap::with_capacity(_cap);
|
||||
$(
|
||||
_map.insert($key, $value);
|
||||
)*
|
||||
_map
|
||||
}
|
||||
};
|
||||
}
|
||||
|
||||
#[cfg(has_std)]
|
||||
#[macro_export]
|
||||
/// Create an `IndexSet` from a list of values
|
||||
///
|
||||
/// ## Example
|
||||
///
|
||||
/// ```
|
||||
/// use indexmap::indexset;
|
||||
///
|
||||
/// let set = indexset!{
|
||||
/// "a",
|
||||
/// "b",
|
||||
/// };
|
||||
/// assert!(set.contains("a"));
|
||||
/// assert!(set.contains("b"));
|
||||
/// assert!(!set.contains("c"));
|
||||
///
|
||||
/// // "a" is the first value
|
||||
/// assert_eq!(set.iter().next(), Some(&"a"));
|
||||
/// ```
|
||||
macro_rules! indexset {
|
||||
(@single $($x:tt)*) => (());
|
||||
(@count $($rest:expr),*) => (<[()]>::len(&[$($crate::indexset!(@single $rest)),*]));
|
||||
|
||||
($($value:expr,)+) => { $crate::indexset!($($value),+) };
|
||||
($($value:expr),*) => {
|
||||
{
|
||||
let _cap = $crate::indexset!(@count $($value),*);
|
||||
let mut _set = $crate::IndexSet::with_capacity(_cap);
|
||||
$(
|
||||
_set.insert($value);
|
||||
)*
|
||||
_set
|
||||
}
|
||||
};
|
||||
}
|
||||
|
||||
// generate all the Iterator methods by just forwarding to the underlying
|
||||
// self.iter and mapping its element.
|
||||
macro_rules! iterator_methods {
|
||||
// $map_elt is the mapping function from the underlying iterator's element
|
||||
// same mapping function for both options and iterators
|
||||
($map_elt:expr) => {
|
||||
fn next(&mut self) -> Option<Self::Item> {
|
||||
self.iter.next().map($map_elt)
|
||||
}
|
||||
|
||||
fn size_hint(&self) -> (usize, Option<usize>) {
|
||||
self.iter.size_hint()
|
||||
}
|
||||
|
||||
fn count(self) -> usize {
|
||||
self.iter.len()
|
||||
}
|
||||
|
||||
fn nth(&mut self, n: usize) -> Option<Self::Item> {
|
||||
self.iter.nth(n).map($map_elt)
|
||||
}
|
||||
|
||||
fn last(mut self) -> Option<Self::Item> {
|
||||
self.next_back()
|
||||
}
|
||||
|
||||
fn collect<C>(self) -> C
|
||||
where
|
||||
C: FromIterator<Self::Item>,
|
||||
{
|
||||
// NB: forwarding this directly to standard iterators will
|
||||
// allow it to leverage unstable traits like `TrustedLen`.
|
||||
self.iter.map($map_elt).collect()
|
||||
}
|
||||
};
|
||||
}
|
||||
|
||||
macro_rules! double_ended_iterator_methods {
|
||||
// $map_elt is the mapping function from the underlying iterator's element
|
||||
// same mapping function for both options and iterators
|
||||
($map_elt:expr) => {
|
||||
fn next_back(&mut self) -> Option<Self::Item> {
|
||||
self.iter.next_back().map($map_elt)
|
||||
}
|
||||
|
||||
fn nth_back(&mut self, n: usize) -> Option<Self::Item> {
|
||||
self.iter.nth_back(n).map($map_elt)
|
||||
}
|
||||
};
|
||||
}
|
||||
|
||||
// generate `ParallelIterator` methods by just forwarding to the underlying
|
||||
// self.entries and mapping its elements.
|
||||
#[cfg(any(feature = "rayon", feature = "rustc-rayon"))]
|
||||
macro_rules! parallel_iterator_methods {
|
||||
// $map_elt is the mapping function from the underlying iterator's element
|
||||
($map_elt:expr) => {
|
||||
fn drive_unindexed<C>(self, consumer: C) -> C::Result
|
||||
where
|
||||
C: UnindexedConsumer<Self::Item>,
|
||||
{
|
||||
self.entries
|
||||
.into_par_iter()
|
||||
.map($map_elt)
|
||||
.drive_unindexed(consumer)
|
||||
}
|
||||
|
||||
// NB: This allows indexed collection, e.g. directly into a `Vec`, but the
|
||||
// underlying iterator must really be indexed. We should remove this if we
|
||||
// start having tombstones that must be filtered out.
|
||||
fn opt_len(&self) -> Option<usize> {
|
||||
Some(self.entries.len())
|
||||
}
|
||||
};
|
||||
}
|
||||
|
||||
// generate `IndexedParallelIterator` methods by just forwarding to the underlying
|
||||
// self.entries and mapping its elements.
|
||||
#[cfg(any(feature = "rayon", feature = "rustc-rayon"))]
|
||||
macro_rules! indexed_parallel_iterator_methods {
|
||||
// $map_elt is the mapping function from the underlying iterator's element
|
||||
($map_elt:expr) => {
|
||||
fn drive<C>(self, consumer: C) -> C::Result
|
||||
where
|
||||
C: Consumer<Self::Item>,
|
||||
{
|
||||
self.entries.into_par_iter().map($map_elt).drive(consumer)
|
||||
}
|
||||
|
||||
fn len(&self) -> usize {
|
||||
self.entries.len()
|
||||
}
|
||||
|
||||
fn with_producer<CB>(self, callback: CB) -> CB::Output
|
||||
where
|
||||
CB: ProducerCallback<Self::Item>,
|
||||
{
|
||||
self.entries
|
||||
.into_par_iter()
|
||||
.map($map_elt)
|
||||
.with_producer(callback)
|
||||
}
|
||||
};
|
||||
}
|
||||
1947
clamav/libclamav_rust/.cargo/vendor/indexmap/src/map.rs
vendored
Normal file
1947
clamav/libclamav_rust/.cargo/vendor/indexmap/src/map.rs
vendored
Normal file
File diff suppressed because it is too large
Load Diff
700
clamav/libclamav_rust/.cargo/vendor/indexmap/src/map/core.rs
vendored
Normal file
700
clamav/libclamav_rust/.cargo/vendor/indexmap/src/map/core.rs
vendored
Normal file
@@ -0,0 +1,700 @@
|
||||
//! This is the core implementation that doesn't depend on the hasher at all.
|
||||
//!
|
||||
//! The methods of `IndexMapCore` don't use any Hash properties of K.
|
||||
//!
|
||||
//! It's cleaner to separate them out, then the compiler checks that we are not
|
||||
//! using Hash at all in these methods.
|
||||
//!
|
||||
//! However, we should probably not let this show in the public API or docs.
|
||||
|
||||
mod raw;
|
||||
|
||||
use hashbrown::raw::RawTable;
|
||||
|
||||
use crate::vec::{Drain, Vec};
|
||||
use core::cmp;
|
||||
use core::fmt;
|
||||
use core::mem::replace;
|
||||
use core::ops::RangeBounds;
|
||||
|
||||
use crate::equivalent::Equivalent;
|
||||
use crate::util::simplify_range;
|
||||
use crate::{Bucket, Entries, HashValue};
|
||||
|
||||
/// Core of the map that does not depend on S
|
||||
pub(crate) struct IndexMapCore<K, V> {
|
||||
/// indices mapping from the entry hash to its index.
|
||||
indices: RawTable<usize>,
|
||||
/// entries is a dense vec of entries in their order.
|
||||
entries: Vec<Bucket<K, V>>,
|
||||
}
|
||||
|
||||
#[inline(always)]
|
||||
fn get_hash<K, V>(entries: &[Bucket<K, V>]) -> impl Fn(&usize) -> u64 + '_ {
|
||||
move |&i| entries[i].hash.get()
|
||||
}
|
||||
|
||||
#[inline]
|
||||
fn equivalent<'a, K, V, Q: ?Sized + Equivalent<K>>(
|
||||
key: &'a Q,
|
||||
entries: &'a [Bucket<K, V>],
|
||||
) -> impl Fn(&usize) -> bool + 'a {
|
||||
move |&i| Q::equivalent(key, &entries[i].key)
|
||||
}
|
||||
|
||||
#[inline]
|
||||
fn erase_index(table: &mut RawTable<usize>, hash: HashValue, index: usize) {
|
||||
let erased = table.erase_entry(hash.get(), move |&i| i == index);
|
||||
debug_assert!(erased);
|
||||
}
|
||||
|
||||
#[inline]
|
||||
fn update_index(table: &mut RawTable<usize>, hash: HashValue, old: usize, new: usize) {
|
||||
let index = table
|
||||
.get_mut(hash.get(), move |&i| i == old)
|
||||
.expect("index not found");
|
||||
*index = new;
|
||||
}
|
||||
|
||||
impl<K, V> Clone for IndexMapCore<K, V>
|
||||
where
|
||||
K: Clone,
|
||||
V: Clone,
|
||||
{
|
||||
fn clone(&self) -> Self {
|
||||
let indices = self.indices.clone();
|
||||
let mut entries = Vec::with_capacity(indices.capacity());
|
||||
entries.clone_from(&self.entries);
|
||||
IndexMapCore { indices, entries }
|
||||
}
|
||||
|
||||
fn clone_from(&mut self, other: &Self) {
|
||||
let hasher = get_hash(&other.entries);
|
||||
self.indices.clone_from_with_hasher(&other.indices, hasher);
|
||||
if self.entries.capacity() < other.entries.len() {
|
||||
// If we must resize, match the indices capacity
|
||||
self.reserve_entries();
|
||||
}
|
||||
self.entries.clone_from(&other.entries);
|
||||
}
|
||||
}
|
||||
|
||||
impl<K, V> fmt::Debug for IndexMapCore<K, V>
|
||||
where
|
||||
K: fmt::Debug,
|
||||
V: fmt::Debug,
|
||||
{
|
||||
fn fmt(&self, f: &mut fmt::Formatter<'_>) -> fmt::Result {
|
||||
f.debug_struct("IndexMapCore")
|
||||
.field("indices", &raw::DebugIndices(&self.indices))
|
||||
.field("entries", &self.entries)
|
||||
.finish()
|
||||
}
|
||||
}
|
||||
|
||||
impl<K, V> Entries for IndexMapCore<K, V> {
|
||||
type Entry = Bucket<K, V>;
|
||||
|
||||
#[inline]
|
||||
fn into_entries(self) -> Vec<Self::Entry> {
|
||||
self.entries
|
||||
}
|
||||
|
||||
#[inline]
|
||||
fn as_entries(&self) -> &[Self::Entry] {
|
||||
&self.entries
|
||||
}
|
||||
|
||||
#[inline]
|
||||
fn as_entries_mut(&mut self) -> &mut [Self::Entry] {
|
||||
&mut self.entries
|
||||
}
|
||||
|
||||
fn with_entries<F>(&mut self, f: F)
|
||||
where
|
||||
F: FnOnce(&mut [Self::Entry]),
|
||||
{
|
||||
f(&mut self.entries);
|
||||
self.rebuild_hash_table();
|
||||
}
|
||||
}
|
||||
|
||||
impl<K, V> IndexMapCore<K, V> {
|
||||
#[inline]
|
||||
pub(crate) const fn new() -> Self {
|
||||
IndexMapCore {
|
||||
indices: RawTable::new(),
|
||||
entries: Vec::new(),
|
||||
}
|
||||
}
|
||||
|
||||
#[inline]
|
||||
pub(crate) fn with_capacity(n: usize) -> Self {
|
||||
IndexMapCore {
|
||||
indices: RawTable::with_capacity(n),
|
||||
entries: Vec::with_capacity(n),
|
||||
}
|
||||
}
|
||||
|
||||
#[inline]
|
||||
pub(crate) fn len(&self) -> usize {
|
||||
self.indices.len()
|
||||
}
|
||||
|
||||
#[inline]
|
||||
pub(crate) fn capacity(&self) -> usize {
|
||||
cmp::min(self.indices.capacity(), self.entries.capacity())
|
||||
}
|
||||
|
||||
pub(crate) fn clear(&mut self) {
|
||||
self.indices.clear();
|
||||
self.entries.clear();
|
||||
}
|
||||
|
||||
pub(crate) fn truncate(&mut self, len: usize) {
|
||||
if len < self.len() {
|
||||
self.erase_indices(len, self.entries.len());
|
||||
self.entries.truncate(len);
|
||||
}
|
||||
}
|
||||
|
||||
pub(crate) fn drain<R>(&mut self, range: R) -> Drain<'_, Bucket<K, V>>
|
||||
where
|
||||
R: RangeBounds<usize>,
|
||||
{
|
||||
let range = simplify_range(range, self.entries.len());
|
||||
self.erase_indices(range.start, range.end);
|
||||
self.entries.drain(range)
|
||||
}
|
||||
|
||||
#[cfg(feature = "rayon")]
|
||||
pub(crate) fn par_drain<R>(&mut self, range: R) -> rayon::vec::Drain<'_, Bucket<K, V>>
|
||||
where
|
||||
K: Send,
|
||||
V: Send,
|
||||
R: RangeBounds<usize>,
|
||||
{
|
||||
use rayon::iter::ParallelDrainRange;
|
||||
let range = simplify_range(range, self.entries.len());
|
||||
self.erase_indices(range.start, range.end);
|
||||
self.entries.par_drain(range)
|
||||
}
|
||||
|
||||
pub(crate) fn split_off(&mut self, at: usize) -> Self {
|
||||
assert!(at <= self.entries.len());
|
||||
self.erase_indices(at, self.entries.len());
|
||||
let entries = self.entries.split_off(at);
|
||||
|
||||
let mut indices = RawTable::with_capacity(entries.len());
|
||||
raw::insert_bulk_no_grow(&mut indices, &entries);
|
||||
Self { indices, entries }
|
||||
}
|
||||
|
||||
/// Reserve capacity for `additional` more key-value pairs.
|
||||
pub(crate) fn reserve(&mut self, additional: usize) {
|
||||
self.indices.reserve(additional, get_hash(&self.entries));
|
||||
self.reserve_entries();
|
||||
}
|
||||
|
||||
/// Reserve entries capacity to match the indices
|
||||
fn reserve_entries(&mut self) {
|
||||
let additional = self.indices.capacity() - self.entries.len();
|
||||
self.entries.reserve_exact(additional);
|
||||
}
|
||||
|
||||
/// Shrink the capacity of the map with a lower bound
|
||||
pub(crate) fn shrink_to(&mut self, min_capacity: usize) {
|
||||
self.indices
|
||||
.shrink_to(min_capacity, get_hash(&self.entries));
|
||||
self.entries.shrink_to(min_capacity);
|
||||
}
|
||||
|
||||
/// Remove the last key-value pair
|
||||
pub(crate) fn pop(&mut self) -> Option<(K, V)> {
|
||||
if let Some(entry) = self.entries.pop() {
|
||||
let last = self.entries.len();
|
||||
erase_index(&mut self.indices, entry.hash, last);
|
||||
Some((entry.key, entry.value))
|
||||
} else {
|
||||
None
|
||||
}
|
||||
}
|
||||
|
||||
/// Append a key-value pair, *without* checking whether it already exists,
|
||||
/// and return the pair's new index.
|
||||
fn push(&mut self, hash: HashValue, key: K, value: V) -> usize {
|
||||
let i = self.entries.len();
|
||||
self.indices.insert(hash.get(), i, get_hash(&self.entries));
|
||||
if i == self.entries.capacity() {
|
||||
// Reserve our own capacity synced to the indices,
|
||||
// rather than letting `Vec::push` just double it.
|
||||
self.reserve_entries();
|
||||
}
|
||||
self.entries.push(Bucket { hash, key, value });
|
||||
i
|
||||
}
|
||||
|
||||
/// Return the index in `entries` where an equivalent key can be found
|
||||
pub(crate) fn get_index_of<Q>(&self, hash: HashValue, key: &Q) -> Option<usize>
|
||||
where
|
||||
Q: ?Sized + Equivalent<K>,
|
||||
{
|
||||
let eq = equivalent(key, &self.entries);
|
||||
self.indices.get(hash.get(), eq).copied()
|
||||
}
|
||||
|
||||
pub(crate) fn insert_full(&mut self, hash: HashValue, key: K, value: V) -> (usize, Option<V>)
|
||||
where
|
||||
K: Eq,
|
||||
{
|
||||
match self.get_index_of(hash, &key) {
|
||||
Some(i) => (i, Some(replace(&mut self.entries[i].value, value))),
|
||||
None => (self.push(hash, key, value), None),
|
||||
}
|
||||
}
|
||||
|
||||
/// Remove an entry by shifting all entries that follow it
|
||||
pub(crate) fn shift_remove_full<Q>(&mut self, hash: HashValue, key: &Q) -> Option<(usize, K, V)>
|
||||
where
|
||||
Q: ?Sized + Equivalent<K>,
|
||||
{
|
||||
let eq = equivalent(key, &self.entries);
|
||||
match self.indices.remove_entry(hash.get(), eq) {
|
||||
Some(index) => {
|
||||
let (key, value) = self.shift_remove_finish(index);
|
||||
Some((index, key, value))
|
||||
}
|
||||
None => None,
|
||||
}
|
||||
}
|
||||
|
||||
/// Remove an entry by shifting all entries that follow it
|
||||
pub(crate) fn shift_remove_index(&mut self, index: usize) -> Option<(K, V)> {
|
||||
match self.entries.get(index) {
|
||||
Some(entry) => {
|
||||
erase_index(&mut self.indices, entry.hash, index);
|
||||
Some(self.shift_remove_finish(index))
|
||||
}
|
||||
None => None,
|
||||
}
|
||||
}
|
||||
|
||||
/// Remove an entry by shifting all entries that follow it
|
||||
///
|
||||
/// The index should already be removed from `self.indices`.
|
||||
fn shift_remove_finish(&mut self, index: usize) -> (K, V) {
|
||||
// Correct indices that point to the entries that followed the removed entry.
|
||||
self.decrement_indices(index + 1, self.entries.len());
|
||||
|
||||
// Use Vec::remove to actually remove the entry.
|
||||
let entry = self.entries.remove(index);
|
||||
(entry.key, entry.value)
|
||||
}
|
||||
|
||||
/// Decrement all indices in the range `start..end`.
|
||||
///
|
||||
/// The index `start - 1` should not exist in `self.indices`.
|
||||
/// All entries should still be in their original positions.
|
||||
fn decrement_indices(&mut self, start: usize, end: usize) {
|
||||
// Use a heuristic between a full sweep vs. a `find()` for every shifted item.
|
||||
let shifted_entries = &self.entries[start..end];
|
||||
if shifted_entries.len() > self.indices.buckets() / 2 {
|
||||
// Shift all indices in range.
|
||||
for i in self.indices_mut() {
|
||||
if start <= *i && *i < end {
|
||||
*i -= 1;
|
||||
}
|
||||
}
|
||||
} else {
|
||||
// Find each entry in range to shift its index.
|
||||
for (i, entry) in (start..end).zip(shifted_entries) {
|
||||
update_index(&mut self.indices, entry.hash, i, i - 1);
|
||||
}
|
||||
}
|
||||
}
|
||||
|
||||
/// Increment all indices in the range `start..end`.
|
||||
///
|
||||
/// The index `end` should not exist in `self.indices`.
|
||||
/// All entries should still be in their original positions.
|
||||
fn increment_indices(&mut self, start: usize, end: usize) {
|
||||
// Use a heuristic between a full sweep vs. a `find()` for every shifted item.
|
||||
let shifted_entries = &self.entries[start..end];
|
||||
if shifted_entries.len() > self.indices.buckets() / 2 {
|
||||
// Shift all indices in range.
|
||||
for i in self.indices_mut() {
|
||||
if start <= *i && *i < end {
|
||||
*i += 1;
|
||||
}
|
||||
}
|
||||
} else {
|
||||
// Find each entry in range to shift its index, updated in reverse so
|
||||
// we never have duplicated indices that might have a hash collision.
|
||||
for (i, entry) in (start..end).zip(shifted_entries).rev() {
|
||||
update_index(&mut self.indices, entry.hash, i, i + 1);
|
||||
}
|
||||
}
|
||||
}
|
||||
|
||||
pub(super) fn move_index(&mut self, from: usize, to: usize) {
|
||||
let from_hash = self.entries[from].hash;
|
||||
if from != to {
|
||||
// Use a sentinal index so other indices don't collide.
|
||||
update_index(&mut self.indices, from_hash, from, usize::MAX);
|
||||
|
||||
// Update all other indices and rotate the entry positions.
|
||||
if from < to {
|
||||
self.decrement_indices(from + 1, to + 1);
|
||||
self.entries[from..=to].rotate_left(1);
|
||||
} else if to < from {
|
||||
self.increment_indices(to, from);
|
||||
self.entries[to..=from].rotate_right(1);
|
||||
}
|
||||
|
||||
// Change the sentinal index to its final position.
|
||||
update_index(&mut self.indices, from_hash, usize::MAX, to);
|
||||
}
|
||||
}
|
||||
|
||||
/// Remove an entry by swapping it with the last
|
||||
pub(crate) fn swap_remove_full<Q>(&mut self, hash: HashValue, key: &Q) -> Option<(usize, K, V)>
|
||||
where
|
||||
Q: ?Sized + Equivalent<K>,
|
||||
{
|
||||
let eq = equivalent(key, &self.entries);
|
||||
match self.indices.remove_entry(hash.get(), eq) {
|
||||
Some(index) => {
|
||||
let (key, value) = self.swap_remove_finish(index);
|
||||
Some((index, key, value))
|
||||
}
|
||||
None => None,
|
||||
}
|
||||
}
|
||||
|
||||
/// Remove an entry by swapping it with the last
|
||||
pub(crate) fn swap_remove_index(&mut self, index: usize) -> Option<(K, V)> {
|
||||
match self.entries.get(index) {
|
||||
Some(entry) => {
|
||||
erase_index(&mut self.indices, entry.hash, index);
|
||||
Some(self.swap_remove_finish(index))
|
||||
}
|
||||
None => None,
|
||||
}
|
||||
}
|
||||
|
||||
/// Finish removing an entry by swapping it with the last
|
||||
///
|
||||
/// The index should already be removed from `self.indices`.
|
||||
fn swap_remove_finish(&mut self, index: usize) -> (K, V) {
|
||||
// use swap_remove, but then we need to update the index that points
|
||||
// to the other entry that has to move
|
||||
let entry = self.entries.swap_remove(index);
|
||||
|
||||
// correct index that points to the entry that had to swap places
|
||||
if let Some(entry) = self.entries.get(index) {
|
||||
// was not last element
|
||||
// examine new element in `index` and find it in indices
|
||||
let last = self.entries.len();
|
||||
update_index(&mut self.indices, entry.hash, last, index);
|
||||
}
|
||||
|
||||
(entry.key, entry.value)
|
||||
}
|
||||
|
||||
/// Erase `start..end` from `indices`, and shift `end..` indices down to `start..`
|
||||
///
|
||||
/// All of these items should still be at their original location in `entries`.
|
||||
/// This is used by `drain`, which will let `Vec::drain` do the work on `entries`.
|
||||
fn erase_indices(&mut self, start: usize, end: usize) {
|
||||
let (init, shifted_entries) = self.entries.split_at(end);
|
||||
let (start_entries, erased_entries) = init.split_at(start);
|
||||
|
||||
let erased = erased_entries.len();
|
||||
let shifted = shifted_entries.len();
|
||||
let half_capacity = self.indices.buckets() / 2;
|
||||
|
||||
// Use a heuristic between different strategies
|
||||
if erased == 0 {
|
||||
// Degenerate case, nothing to do
|
||||
} else if start + shifted < half_capacity && start < erased {
|
||||
// Reinsert everything, as there are few kept indices
|
||||
self.indices.clear();
|
||||
|
||||
// Reinsert stable indices, then shifted indices
|
||||
raw::insert_bulk_no_grow(&mut self.indices, start_entries);
|
||||
raw::insert_bulk_no_grow(&mut self.indices, shifted_entries);
|
||||
} else if erased + shifted < half_capacity {
|
||||
// Find each affected index, as there are few to adjust
|
||||
|
||||
// Find erased indices
|
||||
for (i, entry) in (start..).zip(erased_entries) {
|
||||
erase_index(&mut self.indices, entry.hash, i);
|
||||
}
|
||||
|
||||
// Find shifted indices
|
||||
for ((new, old), entry) in (start..).zip(end..).zip(shifted_entries) {
|
||||
update_index(&mut self.indices, entry.hash, old, new);
|
||||
}
|
||||
} else {
|
||||
// Sweep the whole table for adjustments
|
||||
self.erase_indices_sweep(start, end);
|
||||
}
|
||||
|
||||
debug_assert_eq!(self.indices.len(), start + shifted);
|
||||
}
|
||||
|
||||
pub(crate) fn retain_in_order<F>(&mut self, mut keep: F)
|
||||
where
|
||||
F: FnMut(&mut K, &mut V) -> bool,
|
||||
{
|
||||
// FIXME: This could use Vec::retain_mut with MSRV 1.61.
|
||||
// Like Vec::retain in self.entries, but with mutable K and V.
|
||||
// We swap-shift all the items we want to keep, truncate the rest,
|
||||
// then rebuild the raw hash table with the new indexes.
|
||||
let len = self.entries.len();
|
||||
let mut n_deleted = 0;
|
||||
for i in 0..len {
|
||||
let will_keep = {
|
||||
let entry = &mut self.entries[i];
|
||||
keep(&mut entry.key, &mut entry.value)
|
||||
};
|
||||
if !will_keep {
|
||||
n_deleted += 1;
|
||||
} else if n_deleted > 0 {
|
||||
self.entries.swap(i - n_deleted, i);
|
||||
}
|
||||
}
|
||||
if n_deleted > 0 {
|
||||
self.entries.truncate(len - n_deleted);
|
||||
self.rebuild_hash_table();
|
||||
}
|
||||
}
|
||||
|
||||
fn rebuild_hash_table(&mut self) {
|
||||
self.indices.clear();
|
||||
raw::insert_bulk_no_grow(&mut self.indices, &self.entries);
|
||||
}
|
||||
|
||||
pub(crate) fn reverse(&mut self) {
|
||||
self.entries.reverse();
|
||||
|
||||
// No need to save hash indices, can easily calculate what they should
|
||||
// be, given that this is an in-place reversal.
|
||||
let len = self.entries.len();
|
||||
for i in self.indices_mut() {
|
||||
*i = len - *i - 1;
|
||||
}
|
||||
}
|
||||
}
|
||||
|
||||
/// Entry for an existing key-value pair or a vacant location to
|
||||
/// insert one.
|
||||
pub enum Entry<'a, K, V> {
|
||||
/// Existing slot with equivalent key.
|
||||
Occupied(OccupiedEntry<'a, K, V>),
|
||||
/// Vacant slot (no equivalent key in the map).
|
||||
Vacant(VacantEntry<'a, K, V>),
|
||||
}
|
||||
|
||||
impl<'a, K, V> Entry<'a, K, V> {
|
||||
/// Inserts the given default value in the entry if it is vacant and returns a mutable
|
||||
/// reference to it. Otherwise a mutable reference to an already existent value is returned.
|
||||
///
|
||||
/// Computes in **O(1)** time (amortized average).
|
||||
pub fn or_insert(self, default: V) -> &'a mut V {
|
||||
match self {
|
||||
Entry::Occupied(entry) => entry.into_mut(),
|
||||
Entry::Vacant(entry) => entry.insert(default),
|
||||
}
|
||||
}
|
||||
|
||||
/// Inserts the result of the `call` function in the entry if it is vacant and returns a mutable
|
||||
/// reference to it. Otherwise a mutable reference to an already existent value is returned.
|
||||
///
|
||||
/// Computes in **O(1)** time (amortized average).
|
||||
pub fn or_insert_with<F>(self, call: F) -> &'a mut V
|
||||
where
|
||||
F: FnOnce() -> V,
|
||||
{
|
||||
match self {
|
||||
Entry::Occupied(entry) => entry.into_mut(),
|
||||
Entry::Vacant(entry) => entry.insert(call()),
|
||||
}
|
||||
}
|
||||
|
||||
/// Inserts the result of the `call` function with a reference to the entry's key if it is
|
||||
/// vacant, and returns a mutable reference to the new value. Otherwise a mutable reference to
|
||||
/// an already existent value is returned.
|
||||
///
|
||||
/// Computes in **O(1)** time (amortized average).
|
||||
pub fn or_insert_with_key<F>(self, call: F) -> &'a mut V
|
||||
where
|
||||
F: FnOnce(&K) -> V,
|
||||
{
|
||||
match self {
|
||||
Entry::Occupied(entry) => entry.into_mut(),
|
||||
Entry::Vacant(entry) => {
|
||||
let value = call(&entry.key);
|
||||
entry.insert(value)
|
||||
}
|
||||
}
|
||||
}
|
||||
|
||||
/// Gets a reference to the entry's key, either within the map if occupied,
|
||||
/// or else the new key that was used to find the entry.
|
||||
pub fn key(&self) -> &K {
|
||||
match *self {
|
||||
Entry::Occupied(ref entry) => entry.key(),
|
||||
Entry::Vacant(ref entry) => entry.key(),
|
||||
}
|
||||
}
|
||||
|
||||
/// Return the index where the key-value pair exists or will be inserted.
|
||||
pub fn index(&self) -> usize {
|
||||
match *self {
|
||||
Entry::Occupied(ref entry) => entry.index(),
|
||||
Entry::Vacant(ref entry) => entry.index(),
|
||||
}
|
||||
}
|
||||
|
||||
/// Modifies the entry if it is occupied.
|
||||
pub fn and_modify<F>(self, f: F) -> Self
|
||||
where
|
||||
F: FnOnce(&mut V),
|
||||
{
|
||||
match self {
|
||||
Entry::Occupied(mut o) => {
|
||||
f(o.get_mut());
|
||||
Entry::Occupied(o)
|
||||
}
|
||||
x => x,
|
||||
}
|
||||
}
|
||||
|
||||
/// Inserts a default-constructed value in the entry if it is vacant and returns a mutable
|
||||
/// reference to it. Otherwise a mutable reference to an already existent value is returned.
|
||||
///
|
||||
/// Computes in **O(1)** time (amortized average).
|
||||
pub fn or_default(self) -> &'a mut V
|
||||
where
|
||||
V: Default,
|
||||
{
|
||||
match self {
|
||||
Entry::Occupied(entry) => entry.into_mut(),
|
||||
Entry::Vacant(entry) => entry.insert(V::default()),
|
||||
}
|
||||
}
|
||||
}
|
||||
|
||||
impl<K: fmt::Debug, V: fmt::Debug> fmt::Debug for Entry<'_, K, V> {
|
||||
fn fmt(&self, f: &mut fmt::Formatter<'_>) -> fmt::Result {
|
||||
match *self {
|
||||
Entry::Vacant(ref v) => f.debug_tuple(stringify!(Entry)).field(v).finish(),
|
||||
Entry::Occupied(ref o) => f.debug_tuple(stringify!(Entry)).field(o).finish(),
|
||||
}
|
||||
}
|
||||
}
|
||||
|
||||
pub use self::raw::OccupiedEntry;
|
||||
|
||||
// Extra methods that don't threaten the unsafe encapsulation.
|
||||
impl<K, V> OccupiedEntry<'_, K, V> {
|
||||
/// Sets the value of the entry to `value`, and returns the entry's old value.
|
||||
pub fn insert(&mut self, value: V) -> V {
|
||||
replace(self.get_mut(), value)
|
||||
}
|
||||
|
||||
/// Remove the key, value pair stored in the map for this entry, and return the value.
|
||||
///
|
||||
/// **NOTE:** This is equivalent to `.swap_remove()`.
|
||||
pub fn remove(self) -> V {
|
||||
self.swap_remove()
|
||||
}
|
||||
|
||||
/// Remove the key, value pair stored in the map for this entry, and return the value.
|
||||
///
|
||||
/// Like `Vec::swap_remove`, the pair is removed by swapping it with the
|
||||
/// last element of the map and popping it off. **This perturbs
|
||||
/// the position of what used to be the last element!**
|
||||
///
|
||||
/// Computes in **O(1)** time (average).
|
||||
pub fn swap_remove(self) -> V {
|
||||
self.swap_remove_entry().1
|
||||
}
|
||||
|
||||
/// Remove the key, value pair stored in the map for this entry, and return the value.
|
||||
///
|
||||
/// Like `Vec::remove`, the pair is removed by shifting all of the
|
||||
/// elements that follow it, preserving their relative order.
|
||||
/// **This perturbs the index of all of those elements!**
|
||||
///
|
||||
/// Computes in **O(n)** time (average).
|
||||
pub fn shift_remove(self) -> V {
|
||||
self.shift_remove_entry().1
|
||||
}
|
||||
|
||||
/// Remove and return the key, value pair stored in the map for this entry
|
||||
///
|
||||
/// **NOTE:** This is equivalent to `.swap_remove_entry()`.
|
||||
pub fn remove_entry(self) -> (K, V) {
|
||||
self.swap_remove_entry()
|
||||
}
|
||||
}
|
||||
|
||||
impl<K: fmt::Debug, V: fmt::Debug> fmt::Debug for OccupiedEntry<'_, K, V> {
|
||||
fn fmt(&self, f: &mut fmt::Formatter<'_>) -> fmt::Result {
|
||||
f.debug_struct(stringify!(OccupiedEntry))
|
||||
.field("key", self.key())
|
||||
.field("value", self.get())
|
||||
.finish()
|
||||
}
|
||||
}
|
||||
|
||||
/// A view into a vacant entry in a `IndexMap`.
|
||||
/// It is part of the [`Entry`] enum.
|
||||
///
|
||||
/// [`Entry`]: enum.Entry.html
|
||||
pub struct VacantEntry<'a, K, V> {
|
||||
map: &'a mut IndexMapCore<K, V>,
|
||||
hash: HashValue,
|
||||
key: K,
|
||||
}
|
||||
|
||||
impl<'a, K, V> VacantEntry<'a, K, V> {
|
||||
/// Gets a reference to the key that was used to find the entry.
|
||||
pub fn key(&self) -> &K {
|
||||
&self.key
|
||||
}
|
||||
|
||||
/// Takes ownership of the key, leaving the entry vacant.
|
||||
pub fn into_key(self) -> K {
|
||||
self.key
|
||||
}
|
||||
|
||||
/// Return the index where the key-value pair will be inserted.
|
||||
pub fn index(&self) -> usize {
|
||||
self.map.len()
|
||||
}
|
||||
|
||||
/// Inserts the entry's key and the given value into the map, and returns a mutable reference
|
||||
/// to the value.
|
||||
pub fn insert(self, value: V) -> &'a mut V {
|
||||
let i = self.map.push(self.hash, self.key, value);
|
||||
&mut self.map.entries[i].value
|
||||
}
|
||||
}
|
||||
|
||||
impl<K: fmt::Debug, V> fmt::Debug for VacantEntry<'_, K, V> {
|
||||
fn fmt(&self, f: &mut fmt::Formatter<'_>) -> fmt::Result {
|
||||
f.debug_tuple(stringify!(VacantEntry))
|
||||
.field(self.key())
|
||||
.finish()
|
||||
}
|
||||
}
|
||||
|
||||
#[test]
|
||||
fn assert_send_sync() {
|
||||
fn assert_send_sync<T: Send + Sync>() {}
|
||||
assert_send_sync::<IndexMapCore<i32, i32>>();
|
||||
assert_send_sync::<Entry<'_, i32, i32>>();
|
||||
}
|
||||
191
clamav/libclamav_rust/.cargo/vendor/indexmap/src/map/core/raw.rs
vendored
Normal file
191
clamav/libclamav_rust/.cargo/vendor/indexmap/src/map/core/raw.rs
vendored
Normal file
@@ -0,0 +1,191 @@
|
||||
#![allow(unsafe_code)]
|
||||
//! This module encapsulates the `unsafe` access to `hashbrown::raw::RawTable`,
|
||||
//! mostly in dealing with its bucket "pointers".
|
||||
|
||||
use super::{equivalent, Bucket, Entry, HashValue, IndexMapCore, VacantEntry};
|
||||
use core::fmt;
|
||||
use core::mem::replace;
|
||||
use hashbrown::raw::RawTable;
|
||||
|
||||
type RawBucket = hashbrown::raw::Bucket<usize>;
|
||||
|
||||
/// Inserts many entries into a raw table without reallocating.
|
||||
///
|
||||
/// ***Panics*** if there is not sufficient capacity already.
|
||||
pub(super) fn insert_bulk_no_grow<K, V>(indices: &mut RawTable<usize>, entries: &[Bucket<K, V>]) {
|
||||
assert!(indices.capacity() - indices.len() >= entries.len());
|
||||
for entry in entries {
|
||||
// SAFETY: we asserted that sufficient capacity exists for all entries.
|
||||
unsafe {
|
||||
indices.insert_no_grow(entry.hash.get(), indices.len());
|
||||
}
|
||||
}
|
||||
}
|
||||
|
||||
pub(super) struct DebugIndices<'a>(pub &'a RawTable<usize>);
|
||||
impl fmt::Debug for DebugIndices<'_> {
|
||||
fn fmt(&self, f: &mut fmt::Formatter<'_>) -> fmt::Result {
|
||||
// SAFETY: we're not letting any of the buckets escape this function
|
||||
let indices = unsafe { self.0.iter().map(|raw_bucket| raw_bucket.read()) };
|
||||
f.debug_list().entries(indices).finish()
|
||||
}
|
||||
}
|
||||
|
||||
impl<K, V> IndexMapCore<K, V> {
|
||||
/// Sweep the whole table to erase indices start..end
|
||||
pub(super) fn erase_indices_sweep(&mut self, start: usize, end: usize) {
|
||||
// SAFETY: we're not letting any of the buckets escape this function
|
||||
unsafe {
|
||||
let offset = end - start;
|
||||
for bucket in self.indices.iter() {
|
||||
let i = bucket.read();
|
||||
if i >= end {
|
||||
bucket.write(i - offset);
|
||||
} else if i >= start {
|
||||
self.indices.erase(bucket);
|
||||
}
|
||||
}
|
||||
}
|
||||
}
|
||||
|
||||
pub(crate) fn entry(&mut self, hash: HashValue, key: K) -> Entry<'_, K, V>
|
||||
where
|
||||
K: Eq,
|
||||
{
|
||||
let eq = equivalent(&key, &self.entries);
|
||||
match self.indices.find(hash.get(), eq) {
|
||||
// SAFETY: The entry is created with a live raw bucket, at the same time
|
||||
// we have a &mut reference to the map, so it can not be modified further.
|
||||
Some(raw_bucket) => Entry::Occupied(OccupiedEntry {
|
||||
map: self,
|
||||
raw_bucket,
|
||||
key,
|
||||
}),
|
||||
None => Entry::Vacant(VacantEntry {
|
||||
map: self,
|
||||
hash,
|
||||
key,
|
||||
}),
|
||||
}
|
||||
}
|
||||
|
||||
pub(super) fn indices_mut(&mut self) -> impl Iterator<Item = &mut usize> {
|
||||
// SAFETY: we're not letting any of the buckets escape this function,
|
||||
// only the item references that are appropriately bound to `&mut self`.
|
||||
unsafe { self.indices.iter().map(|bucket| bucket.as_mut()) }
|
||||
}
|
||||
|
||||
/// Return the raw bucket for the given index
|
||||
fn find_index(&self, index: usize) -> RawBucket {
|
||||
// We'll get a "nice" bounds-check from indexing `self.entries`,
|
||||
// and then we expect to find it in the table as well.
|
||||
let hash = self.entries[index].hash.get();
|
||||
self.indices
|
||||
.find(hash, move |&i| i == index)
|
||||
.expect("index not found")
|
||||
}
|
||||
|
||||
pub(crate) fn swap_indices(&mut self, a: usize, b: usize) {
|
||||
// SAFETY: Can't take two `get_mut` references from one table, so we
|
||||
// must use raw buckets to do the swap. This is still safe because we
|
||||
// are locally sure they won't dangle, and we write them individually.
|
||||
unsafe {
|
||||
let raw_bucket_a = self.find_index(a);
|
||||
let raw_bucket_b = self.find_index(b);
|
||||
raw_bucket_a.write(b);
|
||||
raw_bucket_b.write(a);
|
||||
}
|
||||
self.entries.swap(a, b);
|
||||
}
|
||||
}
|
||||
|
||||
/// A view into an occupied entry in a `IndexMap`.
|
||||
/// It is part of the [`Entry`] enum.
|
||||
///
|
||||
/// [`Entry`]: enum.Entry.html
|
||||
// SAFETY: The lifetime of the map reference also constrains the raw bucket,
|
||||
// which is essentially a raw pointer into the map indices.
|
||||
pub struct OccupiedEntry<'a, K, V> {
|
||||
map: &'a mut IndexMapCore<K, V>,
|
||||
raw_bucket: RawBucket,
|
||||
key: K,
|
||||
}
|
||||
|
||||
// `hashbrown::raw::Bucket` is only `Send`, not `Sync`.
|
||||
// SAFETY: `&self` only accesses the bucket to read it.
|
||||
unsafe impl<K: Sync, V: Sync> Sync for OccupiedEntry<'_, K, V> {}
|
||||
|
||||
// The parent module also adds methods that don't threaten the unsafe encapsulation.
|
||||
impl<'a, K, V> OccupiedEntry<'a, K, V> {
|
||||
/// Gets a reference to the entry's key in the map.
|
||||
///
|
||||
/// Note that this is not the key that was used to find the entry. There may be an observable
|
||||
/// difference if the key type has any distinguishing features outside of `Hash` and `Eq`, like
|
||||
/// extra fields or the memory address of an allocation.
|
||||
pub fn key(&self) -> &K {
|
||||
&self.map.entries[self.index()].key
|
||||
}
|
||||
|
||||
/// Gets a reference to the entry's value in the map.
|
||||
pub fn get(&self) -> &V {
|
||||
&self.map.entries[self.index()].value
|
||||
}
|
||||
|
||||
/// Gets a mutable reference to the entry's value in the map.
|
||||
///
|
||||
/// If you need a reference which may outlive the destruction of the
|
||||
/// `Entry` value, see `into_mut`.
|
||||
pub fn get_mut(&mut self) -> &mut V {
|
||||
let index = self.index();
|
||||
&mut self.map.entries[index].value
|
||||
}
|
||||
|
||||
/// Put the new key in the occupied entry's key slot
|
||||
pub(crate) fn replace_key(self) -> K {
|
||||
let index = self.index();
|
||||
let old_key = &mut self.map.entries[index].key;
|
||||
replace(old_key, self.key)
|
||||
}
|
||||
|
||||
/// Return the index of the key-value pair
|
||||
#[inline]
|
||||
pub fn index(&self) -> usize {
|
||||
// SAFETY: we have &mut map keep keeping the bucket stable
|
||||
unsafe { self.raw_bucket.read() }
|
||||
}
|
||||
|
||||
/// Converts into a mutable reference to the entry's value in the map,
|
||||
/// with a lifetime bound to the map itself.
|
||||
pub fn into_mut(self) -> &'a mut V {
|
||||
let index = self.index();
|
||||
&mut self.map.entries[index].value
|
||||
}
|
||||
|
||||
/// Remove and return the key, value pair stored in the map for this entry
|
||||
///
|
||||
/// Like `Vec::swap_remove`, the pair is removed by swapping it with the
|
||||
/// last element of the map and popping it off. **This perturbs
|
||||
/// the position of what used to be the last element!**
|
||||
///
|
||||
/// Computes in **O(1)** time (average).
|
||||
pub fn swap_remove_entry(self) -> (K, V) {
|
||||
// SAFETY: This is safe because it can only happen once (self is consumed)
|
||||
// and map.indices have not been modified since entry construction
|
||||
let index = unsafe { self.map.indices.remove(self.raw_bucket) };
|
||||
self.map.swap_remove_finish(index)
|
||||
}
|
||||
|
||||
/// Remove and return the key, value pair stored in the map for this entry
|
||||
///
|
||||
/// Like `Vec::remove`, the pair is removed by shifting all of the
|
||||
/// elements that follow it, preserving their relative order.
|
||||
/// **This perturbs the index of all of those elements!**
|
||||
///
|
||||
/// Computes in **O(n)** time (average).
|
||||
pub fn shift_remove_entry(self) -> (K, V) {
|
||||
// SAFETY: This is safe because it can only happen once (self is consumed)
|
||||
// and map.indices have not been modified since entry construction
|
||||
let index = unsafe { self.map.indices.remove(self.raw_bucket) };
|
||||
self.map.shift_remove_finish(index)
|
||||
}
|
||||
}
|
||||
75
clamav/libclamav_rust/.cargo/vendor/indexmap/src/mutable_keys.rs
vendored
Normal file
75
clamav/libclamav_rust/.cargo/vendor/indexmap/src/mutable_keys.rs
vendored
Normal file
@@ -0,0 +1,75 @@
|
||||
use core::hash::{BuildHasher, Hash};
|
||||
|
||||
use super::{Equivalent, IndexMap};
|
||||
|
||||
pub struct PrivateMarker {}
|
||||
|
||||
/// Opt-in mutable access to keys.
|
||||
///
|
||||
/// These methods expose `&mut K`, mutable references to the key as it is stored
|
||||
/// in the map.
|
||||
/// You are allowed to modify the keys in the hashmap **if the modification
|
||||
/// does not change the key’s hash and equality**.
|
||||
///
|
||||
/// If keys are modified erroneously, you can no longer look them up.
|
||||
/// This is sound (memory safe) but a logical error hazard (just like
|
||||
/// implementing PartialEq, Eq, or Hash incorrectly would be).
|
||||
///
|
||||
/// `use` this trait to enable its methods for `IndexMap`.
|
||||
pub trait MutableKeys {
|
||||
type Key;
|
||||
type Value;
|
||||
|
||||
/// Return item index, mutable reference to key and value
|
||||
fn get_full_mut2<Q: ?Sized>(
|
||||
&mut self,
|
||||
key: &Q,
|
||||
) -> Option<(usize, &mut Self::Key, &mut Self::Value)>
|
||||
where
|
||||
Q: Hash + Equivalent<Self::Key>;
|
||||
|
||||
/// Scan through each key-value pair in the map and keep those where the
|
||||
/// closure `keep` returns `true`.
|
||||
///
|
||||
/// The elements are visited in order, and remaining elements keep their
|
||||
/// order.
|
||||
///
|
||||
/// Computes in **O(n)** time (average).
|
||||
fn retain2<F>(&mut self, keep: F)
|
||||
where
|
||||
F: FnMut(&mut Self::Key, &mut Self::Value) -> bool;
|
||||
|
||||
/// This method is not useful in itself – it is there to “seal” the trait
|
||||
/// for external implementation, so that we can add methods without
|
||||
/// causing breaking changes.
|
||||
fn __private_marker(&self) -> PrivateMarker;
|
||||
}
|
||||
|
||||
/// Opt-in mutable access to keys.
|
||||
///
|
||||
/// See [`MutableKeys`](trait.MutableKeys.html) for more information.
|
||||
impl<K, V, S> MutableKeys for IndexMap<K, V, S>
|
||||
where
|
||||
K: Eq + Hash,
|
||||
S: BuildHasher,
|
||||
{
|
||||
type Key = K;
|
||||
type Value = V;
|
||||
fn get_full_mut2<Q: ?Sized>(&mut self, key: &Q) -> Option<(usize, &mut K, &mut V)>
|
||||
where
|
||||
Q: Hash + Equivalent<K>,
|
||||
{
|
||||
self.get_full_mut2_impl(key)
|
||||
}
|
||||
|
||||
fn retain2<F>(&mut self, keep: F)
|
||||
where
|
||||
F: FnMut(&mut K, &mut V) -> bool,
|
||||
{
|
||||
self.retain_mut(keep)
|
||||
}
|
||||
|
||||
fn __private_marker(&self) -> PrivateMarker {
|
||||
PrivateMarker {}
|
||||
}
|
||||
}
|
||||
583
clamav/libclamav_rust/.cargo/vendor/indexmap/src/rayon/map.rs
vendored
Normal file
583
clamav/libclamav_rust/.cargo/vendor/indexmap/src/rayon/map.rs
vendored
Normal file
@@ -0,0 +1,583 @@
|
||||
//! Parallel iterator types for `IndexMap` with [rayon](https://docs.rs/rayon/1.0/rayon).
|
||||
//!
|
||||
//! You will rarely need to interact with this module directly unless you need to name one of the
|
||||
//! iterator types.
|
||||
//!
|
||||
//! Requires crate feature `"rayon"`
|
||||
|
||||
use super::collect;
|
||||
use rayon::iter::plumbing::{Consumer, ProducerCallback, UnindexedConsumer};
|
||||
use rayon::prelude::*;
|
||||
|
||||
use crate::vec::Vec;
|
||||
use core::cmp::Ordering;
|
||||
use core::fmt;
|
||||
use core::hash::{BuildHasher, Hash};
|
||||
use core::ops::RangeBounds;
|
||||
|
||||
use crate::Bucket;
|
||||
use crate::Entries;
|
||||
use crate::IndexMap;
|
||||
|
||||
/// Requires crate feature `"rayon"`.
|
||||
impl<K, V, S> IntoParallelIterator for IndexMap<K, V, S>
|
||||
where
|
||||
K: Send,
|
||||
V: Send,
|
||||
{
|
||||
type Item = (K, V);
|
||||
type Iter = IntoParIter<K, V>;
|
||||
|
||||
fn into_par_iter(self) -> Self::Iter {
|
||||
IntoParIter {
|
||||
entries: self.into_entries(),
|
||||
}
|
||||
}
|
||||
}
|
||||
|
||||
/// A parallel owning iterator over the entries of a `IndexMap`.
|
||||
///
|
||||
/// This `struct` is created by the [`into_par_iter`] method on [`IndexMap`]
|
||||
/// (provided by rayon's `IntoParallelIterator` trait). See its documentation for more.
|
||||
///
|
||||
/// [`into_par_iter`]: ../struct.IndexMap.html#method.into_par_iter
|
||||
/// [`IndexMap`]: ../struct.IndexMap.html
|
||||
pub struct IntoParIter<K, V> {
|
||||
entries: Vec<Bucket<K, V>>,
|
||||
}
|
||||
|
||||
impl<K: fmt::Debug, V: fmt::Debug> fmt::Debug for IntoParIter<K, V> {
|
||||
fn fmt(&self, f: &mut fmt::Formatter<'_>) -> fmt::Result {
|
||||
let iter = self.entries.iter().map(Bucket::refs);
|
||||
f.debug_list().entries(iter).finish()
|
||||
}
|
||||
}
|
||||
|
||||
impl<K: Send, V: Send> ParallelIterator for IntoParIter<K, V> {
|
||||
type Item = (K, V);
|
||||
|
||||
parallel_iterator_methods!(Bucket::key_value);
|
||||
}
|
||||
|
||||
impl<K: Send, V: Send> IndexedParallelIterator for IntoParIter<K, V> {
|
||||
indexed_parallel_iterator_methods!(Bucket::key_value);
|
||||
}
|
||||
|
||||
/// Requires crate feature `"rayon"`.
|
||||
impl<'a, K, V, S> IntoParallelIterator for &'a IndexMap<K, V, S>
|
||||
where
|
||||
K: Sync,
|
||||
V: Sync,
|
||||
{
|
||||
type Item = (&'a K, &'a V);
|
||||
type Iter = ParIter<'a, K, V>;
|
||||
|
||||
fn into_par_iter(self) -> Self::Iter {
|
||||
ParIter {
|
||||
entries: self.as_entries(),
|
||||
}
|
||||
}
|
||||
}
|
||||
|
||||
/// A parallel iterator over the entries of a `IndexMap`.
|
||||
///
|
||||
/// This `struct` is created by the [`par_iter`] method on [`IndexMap`]
|
||||
/// (provided by rayon's `IntoParallelRefIterator` trait). See its documentation for more.
|
||||
///
|
||||
/// [`par_iter`]: ../struct.IndexMap.html#method.par_iter
|
||||
/// [`IndexMap`]: ../struct.IndexMap.html
|
||||
pub struct ParIter<'a, K, V> {
|
||||
entries: &'a [Bucket<K, V>],
|
||||
}
|
||||
|
||||
impl<K, V> Clone for ParIter<'_, K, V> {
|
||||
fn clone(&self) -> Self {
|
||||
ParIter { ..*self }
|
||||
}
|
||||
}
|
||||
|
||||
impl<K: fmt::Debug, V: fmt::Debug> fmt::Debug for ParIter<'_, K, V> {
|
||||
fn fmt(&self, f: &mut fmt::Formatter<'_>) -> fmt::Result {
|
||||
let iter = self.entries.iter().map(Bucket::refs);
|
||||
f.debug_list().entries(iter).finish()
|
||||
}
|
||||
}
|
||||
|
||||
impl<'a, K: Sync, V: Sync> ParallelIterator for ParIter<'a, K, V> {
|
||||
type Item = (&'a K, &'a V);
|
||||
|
||||
parallel_iterator_methods!(Bucket::refs);
|
||||
}
|
||||
|
||||
impl<K: Sync, V: Sync> IndexedParallelIterator for ParIter<'_, K, V> {
|
||||
indexed_parallel_iterator_methods!(Bucket::refs);
|
||||
}
|
||||
|
||||
/// Requires crate feature `"rayon"`.
|
||||
impl<'a, K, V, S> IntoParallelIterator for &'a mut IndexMap<K, V, S>
|
||||
where
|
||||
K: Sync + Send,
|
||||
V: Send,
|
||||
{
|
||||
type Item = (&'a K, &'a mut V);
|
||||
type Iter = ParIterMut<'a, K, V>;
|
||||
|
||||
fn into_par_iter(self) -> Self::Iter {
|
||||
ParIterMut {
|
||||
entries: self.as_entries_mut(),
|
||||
}
|
||||
}
|
||||
}
|
||||
|
||||
/// A parallel mutable iterator over the entries of a `IndexMap`.
|
||||
///
|
||||
/// This `struct` is created by the [`par_iter_mut`] method on [`IndexMap`]
|
||||
/// (provided by rayon's `IntoParallelRefMutIterator` trait). See its documentation for more.
|
||||
///
|
||||
/// [`par_iter_mut`]: ../struct.IndexMap.html#method.par_iter_mut
|
||||
/// [`IndexMap`]: ../struct.IndexMap.html
|
||||
pub struct ParIterMut<'a, K, V> {
|
||||
entries: &'a mut [Bucket<K, V>],
|
||||
}
|
||||
|
||||
impl<K: fmt::Debug, V: fmt::Debug> fmt::Debug for ParIterMut<'_, K, V> {
|
||||
fn fmt(&self, f: &mut fmt::Formatter<'_>) -> fmt::Result {
|
||||
let iter = self.entries.iter().map(Bucket::refs);
|
||||
f.debug_list().entries(iter).finish()
|
||||
}
|
||||
}
|
||||
|
||||
impl<'a, K: Sync + Send, V: Send> ParallelIterator for ParIterMut<'a, K, V> {
|
||||
type Item = (&'a K, &'a mut V);
|
||||
|
||||
parallel_iterator_methods!(Bucket::ref_mut);
|
||||
}
|
||||
|
||||
impl<K: Sync + Send, V: Send> IndexedParallelIterator for ParIterMut<'_, K, V> {
|
||||
indexed_parallel_iterator_methods!(Bucket::ref_mut);
|
||||
}
|
||||
|
||||
/// Requires crate feature `"rayon"`.
|
||||
impl<'a, K, V, S> ParallelDrainRange<usize> for &'a mut IndexMap<K, V, S>
|
||||
where
|
||||
K: Send,
|
||||
V: Send,
|
||||
{
|
||||
type Item = (K, V);
|
||||
type Iter = ParDrain<'a, K, V>;
|
||||
|
||||
fn par_drain<R: RangeBounds<usize>>(self, range: R) -> Self::Iter {
|
||||
ParDrain {
|
||||
entries: self.core.par_drain(range),
|
||||
}
|
||||
}
|
||||
}
|
||||
|
||||
/// A parallel draining iterator over the entries of a `IndexMap`.
|
||||
///
|
||||
/// This `struct` is created by the [`par_drain`] method on [`IndexMap`]
|
||||
/// (provided by rayon's `ParallelDrainRange` trait). See its documentation for more.
|
||||
///
|
||||
/// [`par_drain`]: ../struct.IndexMap.html#method.par_drain
|
||||
/// [`IndexMap`]: ../struct.IndexMap.html
|
||||
pub struct ParDrain<'a, K: Send, V: Send> {
|
||||
entries: rayon::vec::Drain<'a, Bucket<K, V>>,
|
||||
}
|
||||
|
||||
impl<K: Send, V: Send> ParallelIterator for ParDrain<'_, K, V> {
|
||||
type Item = (K, V);
|
||||
|
||||
parallel_iterator_methods!(Bucket::key_value);
|
||||
}
|
||||
|
||||
impl<K: Send, V: Send> IndexedParallelIterator for ParDrain<'_, K, V> {
|
||||
indexed_parallel_iterator_methods!(Bucket::key_value);
|
||||
}
|
||||
|
||||
/// Parallel iterator methods and other parallel methods.
|
||||
///
|
||||
/// The following methods **require crate feature `"rayon"`**.
|
||||
///
|
||||
/// See also the `IntoParallelIterator` implementations.
|
||||
impl<K, V, S> IndexMap<K, V, S>
|
||||
where
|
||||
K: Sync,
|
||||
V: Sync,
|
||||
{
|
||||
/// Return a parallel iterator over the keys of the map.
|
||||
///
|
||||
/// While parallel iterators can process items in any order, their relative order
|
||||
/// in the map is still preserved for operations like `reduce` and `collect`.
|
||||
pub fn par_keys(&self) -> ParKeys<'_, K, V> {
|
||||
ParKeys {
|
||||
entries: self.as_entries(),
|
||||
}
|
||||
}
|
||||
|
||||
/// Return a parallel iterator over the values of the map.
|
||||
///
|
||||
/// While parallel iterators can process items in any order, their relative order
|
||||
/// in the map is still preserved for operations like `reduce` and `collect`.
|
||||
pub fn par_values(&self) -> ParValues<'_, K, V> {
|
||||
ParValues {
|
||||
entries: self.as_entries(),
|
||||
}
|
||||
}
|
||||
}
|
||||
|
||||
impl<K, V, S> IndexMap<K, V, S>
|
||||
where
|
||||
K: Hash + Eq + Sync,
|
||||
V: Sync,
|
||||
S: BuildHasher,
|
||||
{
|
||||
/// Returns `true` if `self` contains all of the same key-value pairs as `other`,
|
||||
/// regardless of each map's indexed order, determined in parallel.
|
||||
pub fn par_eq<V2, S2>(&self, other: &IndexMap<K, V2, S2>) -> bool
|
||||
where
|
||||
V: PartialEq<V2>,
|
||||
V2: Sync,
|
||||
S2: BuildHasher + Sync,
|
||||
{
|
||||
self.len() == other.len()
|
||||
&& self
|
||||
.par_iter()
|
||||
.all(move |(key, value)| other.get(key).map_or(false, |v| *value == *v))
|
||||
}
|
||||
}
|
||||
|
||||
/// A parallel iterator over the keys of a `IndexMap`.
|
||||
///
|
||||
/// This `struct` is created by the [`par_keys`] method on [`IndexMap`]. See its
|
||||
/// documentation for more.
|
||||
///
|
||||
/// [`par_keys`]: ../struct.IndexMap.html#method.par_keys
|
||||
/// [`IndexMap`]: ../struct.IndexMap.html
|
||||
pub struct ParKeys<'a, K, V> {
|
||||
entries: &'a [Bucket<K, V>],
|
||||
}
|
||||
|
||||
impl<K, V> Clone for ParKeys<'_, K, V> {
|
||||
fn clone(&self) -> Self {
|
||||
ParKeys { ..*self }
|
||||
}
|
||||
}
|
||||
|
||||
impl<K: fmt::Debug, V> fmt::Debug for ParKeys<'_, K, V> {
|
||||
fn fmt(&self, f: &mut fmt::Formatter<'_>) -> fmt::Result {
|
||||
let iter = self.entries.iter().map(Bucket::key_ref);
|
||||
f.debug_list().entries(iter).finish()
|
||||
}
|
||||
}
|
||||
|
||||
impl<'a, K: Sync, V: Sync> ParallelIterator for ParKeys<'a, K, V> {
|
||||
type Item = &'a K;
|
||||
|
||||
parallel_iterator_methods!(Bucket::key_ref);
|
||||
}
|
||||
|
||||
impl<K: Sync, V: Sync> IndexedParallelIterator for ParKeys<'_, K, V> {
|
||||
indexed_parallel_iterator_methods!(Bucket::key_ref);
|
||||
}
|
||||
|
||||
/// A parallel iterator over the values of a `IndexMap`.
|
||||
///
|
||||
/// This `struct` is created by the [`par_values`] method on [`IndexMap`]. See its
|
||||
/// documentation for more.
|
||||
///
|
||||
/// [`par_values`]: ../struct.IndexMap.html#method.par_values
|
||||
/// [`IndexMap`]: ../struct.IndexMap.html
|
||||
pub struct ParValues<'a, K, V> {
|
||||
entries: &'a [Bucket<K, V>],
|
||||
}
|
||||
|
||||
impl<K, V> Clone for ParValues<'_, K, V> {
|
||||
fn clone(&self) -> Self {
|
||||
ParValues { ..*self }
|
||||
}
|
||||
}
|
||||
|
||||
impl<K, V: fmt::Debug> fmt::Debug for ParValues<'_, K, V> {
|
||||
fn fmt(&self, f: &mut fmt::Formatter<'_>) -> fmt::Result {
|
||||
let iter = self.entries.iter().map(Bucket::value_ref);
|
||||
f.debug_list().entries(iter).finish()
|
||||
}
|
||||
}
|
||||
|
||||
impl<'a, K: Sync, V: Sync> ParallelIterator for ParValues<'a, K, V> {
|
||||
type Item = &'a V;
|
||||
|
||||
parallel_iterator_methods!(Bucket::value_ref);
|
||||
}
|
||||
|
||||
impl<K: Sync, V: Sync> IndexedParallelIterator for ParValues<'_, K, V> {
|
||||
indexed_parallel_iterator_methods!(Bucket::value_ref);
|
||||
}
|
||||
|
||||
/// Requires crate feature `"rayon"`.
|
||||
impl<K, V, S> IndexMap<K, V, S>
|
||||
where
|
||||
K: Send,
|
||||
V: Send,
|
||||
{
|
||||
/// Return a parallel iterator over mutable references to the values of the map
|
||||
///
|
||||
/// While parallel iterators can process items in any order, their relative order
|
||||
/// in the map is still preserved for operations like `reduce` and `collect`.
|
||||
pub fn par_values_mut(&mut self) -> ParValuesMut<'_, K, V> {
|
||||
ParValuesMut {
|
||||
entries: self.as_entries_mut(),
|
||||
}
|
||||
}
|
||||
}
|
||||
|
||||
impl<K, V, S> IndexMap<K, V, S>
|
||||
where
|
||||
K: Hash + Eq + Send,
|
||||
V: Send,
|
||||
S: BuildHasher,
|
||||
{
|
||||
/// Sort the map’s key-value pairs in parallel, by the default ordering of the keys.
|
||||
pub fn par_sort_keys(&mut self)
|
||||
where
|
||||
K: Ord,
|
||||
{
|
||||
self.with_entries(|entries| {
|
||||
entries.par_sort_by(|a, b| K::cmp(&a.key, &b.key));
|
||||
});
|
||||
}
|
||||
|
||||
/// Sort the map’s key-value pairs in place and in parallel, using the comparison
|
||||
/// function `cmp`.
|
||||
///
|
||||
/// The comparison function receives two key and value pairs to compare (you
|
||||
/// can sort by keys or values or their combination as needed).
|
||||
pub fn par_sort_by<F>(&mut self, cmp: F)
|
||||
where
|
||||
F: Fn(&K, &V, &K, &V) -> Ordering + Sync,
|
||||
{
|
||||
self.with_entries(|entries| {
|
||||
entries.par_sort_by(move |a, b| cmp(&a.key, &a.value, &b.key, &b.value));
|
||||
});
|
||||
}
|
||||
|
||||
/// Sort the key-value pairs of the map in parallel and return a by-value parallel
|
||||
/// iterator of the key-value pairs with the result.
|
||||
pub fn par_sorted_by<F>(self, cmp: F) -> IntoParIter<K, V>
|
||||
where
|
||||
F: Fn(&K, &V, &K, &V) -> Ordering + Sync,
|
||||
{
|
||||
let mut entries = self.into_entries();
|
||||
entries.par_sort_by(move |a, b| cmp(&a.key, &a.value, &b.key, &b.value));
|
||||
IntoParIter { entries }
|
||||
}
|
||||
|
||||
/// Sort the map's key-value pairs in parallel, by the default ordering of the keys.
|
||||
pub fn par_sort_unstable_keys(&mut self)
|
||||
where
|
||||
K: Ord,
|
||||
{
|
||||
self.with_entries(|entries| {
|
||||
entries.par_sort_unstable_by(|a, b| K::cmp(&a.key, &b.key));
|
||||
});
|
||||
}
|
||||
|
||||
/// Sort the map's key-value pairs in place and in parallel, using the comparison
|
||||
/// function `cmp`.
|
||||
///
|
||||
/// The comparison function receives two key and value pairs to compare (you
|
||||
/// can sort by keys or values or their combination as needed).
|
||||
pub fn par_sort_unstable_by<F>(&mut self, cmp: F)
|
||||
where
|
||||
F: Fn(&K, &V, &K, &V) -> Ordering + Sync,
|
||||
{
|
||||
self.with_entries(|entries| {
|
||||
entries.par_sort_unstable_by(move |a, b| cmp(&a.key, &a.value, &b.key, &b.value));
|
||||
});
|
||||
}
|
||||
|
||||
/// Sort the key-value pairs of the map in parallel and return a by-value parallel
|
||||
/// iterator of the key-value pairs with the result.
|
||||
pub fn par_sorted_unstable_by<F>(self, cmp: F) -> IntoParIter<K, V>
|
||||
where
|
||||
F: Fn(&K, &V, &K, &V) -> Ordering + Sync,
|
||||
{
|
||||
let mut entries = self.into_entries();
|
||||
entries.par_sort_unstable_by(move |a, b| cmp(&a.key, &a.value, &b.key, &b.value));
|
||||
IntoParIter { entries }
|
||||
}
|
||||
}
|
||||
|
||||
/// A parallel mutable iterator over the values of a `IndexMap`.
|
||||
///
|
||||
/// This `struct` is created by the [`par_values_mut`] method on [`IndexMap`]. See its
|
||||
/// documentation for more.
|
||||
///
|
||||
/// [`par_values_mut`]: ../struct.IndexMap.html#method.par_values_mut
|
||||
/// [`IndexMap`]: ../struct.IndexMap.html
|
||||
pub struct ParValuesMut<'a, K, V> {
|
||||
entries: &'a mut [Bucket<K, V>],
|
||||
}
|
||||
|
||||
impl<K, V: fmt::Debug> fmt::Debug for ParValuesMut<'_, K, V> {
|
||||
fn fmt(&self, f: &mut fmt::Formatter<'_>) -> fmt::Result {
|
||||
let iter = self.entries.iter().map(Bucket::value_ref);
|
||||
f.debug_list().entries(iter).finish()
|
||||
}
|
||||
}
|
||||
|
||||
impl<'a, K: Send, V: Send> ParallelIterator for ParValuesMut<'a, K, V> {
|
||||
type Item = &'a mut V;
|
||||
|
||||
parallel_iterator_methods!(Bucket::value_mut);
|
||||
}
|
||||
|
||||
impl<K: Send, V: Send> IndexedParallelIterator for ParValuesMut<'_, K, V> {
|
||||
indexed_parallel_iterator_methods!(Bucket::value_mut);
|
||||
}
|
||||
|
||||
/// Requires crate feature `"rayon"`.
|
||||
impl<K, V, S> FromParallelIterator<(K, V)> for IndexMap<K, V, S>
|
||||
where
|
||||
K: Eq + Hash + Send,
|
||||
V: Send,
|
||||
S: BuildHasher + Default + Send,
|
||||
{
|
||||
fn from_par_iter<I>(iter: I) -> Self
|
||||
where
|
||||
I: IntoParallelIterator<Item = (K, V)>,
|
||||
{
|
||||
let list = collect(iter);
|
||||
let len = list.iter().map(Vec::len).sum();
|
||||
let mut map = Self::with_capacity_and_hasher(len, S::default());
|
||||
for vec in list {
|
||||
map.extend(vec);
|
||||
}
|
||||
map
|
||||
}
|
||||
}
|
||||
|
||||
/// Requires crate feature `"rayon"`.
|
||||
impl<K, V, S> ParallelExtend<(K, V)> for IndexMap<K, V, S>
|
||||
where
|
||||
K: Eq + Hash + Send,
|
||||
V: Send,
|
||||
S: BuildHasher + Send,
|
||||
{
|
||||
fn par_extend<I>(&mut self, iter: I)
|
||||
where
|
||||
I: IntoParallelIterator<Item = (K, V)>,
|
||||
{
|
||||
for vec in collect(iter) {
|
||||
self.extend(vec);
|
||||
}
|
||||
}
|
||||
}
|
||||
|
||||
/// Requires crate feature `"rayon"`.
|
||||
impl<'a, K: 'a, V: 'a, S> ParallelExtend<(&'a K, &'a V)> for IndexMap<K, V, S>
|
||||
where
|
||||
K: Copy + Eq + Hash + Send + Sync,
|
||||
V: Copy + Send + Sync,
|
||||
S: BuildHasher + Send,
|
||||
{
|
||||
fn par_extend<I>(&mut self, iter: I)
|
||||
where
|
||||
I: IntoParallelIterator<Item = (&'a K, &'a V)>,
|
||||
{
|
||||
for vec in collect(iter) {
|
||||
self.extend(vec);
|
||||
}
|
||||
}
|
||||
}
|
||||
|
||||
#[cfg(test)]
|
||||
mod tests {
|
||||
use super::*;
|
||||
use std::string::String;
|
||||
|
||||
#[test]
|
||||
fn insert_order() {
|
||||
let insert = [0, 4, 2, 12, 8, 7, 11, 5, 3, 17, 19, 22, 23];
|
||||
let mut map = IndexMap::new();
|
||||
|
||||
for &elt in &insert {
|
||||
map.insert(elt, ());
|
||||
}
|
||||
|
||||
assert_eq!(map.par_keys().count(), map.len());
|
||||
assert_eq!(map.par_keys().count(), insert.len());
|
||||
insert.par_iter().zip(map.par_keys()).for_each(|(a, b)| {
|
||||
assert_eq!(a, b);
|
||||
});
|
||||
(0..insert.len())
|
||||
.into_par_iter()
|
||||
.zip(map.par_keys())
|
||||
.for_each(|(i, k)| {
|
||||
assert_eq!(map.get_index(i).unwrap().0, k);
|
||||
});
|
||||
}
|
||||
|
||||
#[test]
|
||||
fn partial_eq_and_eq() {
|
||||
let mut map_a = IndexMap::new();
|
||||
map_a.insert(1, "1");
|
||||
map_a.insert(2, "2");
|
||||
let mut map_b = map_a.clone();
|
||||
assert!(map_a.par_eq(&map_b));
|
||||
map_b.swap_remove(&1);
|
||||
assert!(!map_a.par_eq(&map_b));
|
||||
map_b.insert(3, "3");
|
||||
assert!(!map_a.par_eq(&map_b));
|
||||
|
||||
let map_c: IndexMap<_, String> =
|
||||
map_b.into_par_iter().map(|(k, v)| (k, v.into())).collect();
|
||||
assert!(!map_a.par_eq(&map_c));
|
||||
assert!(!map_c.par_eq(&map_a));
|
||||
}
|
||||
|
||||
#[test]
|
||||
fn extend() {
|
||||
let mut map = IndexMap::new();
|
||||
map.par_extend(vec![(&1, &2), (&3, &4)]);
|
||||
map.par_extend(vec![(5, 6)]);
|
||||
assert_eq!(
|
||||
map.into_par_iter().collect::<Vec<_>>(),
|
||||
vec![(1, 2), (3, 4), (5, 6)]
|
||||
);
|
||||
}
|
||||
|
||||
#[test]
|
||||
fn keys() {
|
||||
let vec = vec![(1, 'a'), (2, 'b'), (3, 'c')];
|
||||
let map: IndexMap<_, _> = vec.into_par_iter().collect();
|
||||
let keys: Vec<_> = map.par_keys().copied().collect();
|
||||
assert_eq!(keys.len(), 3);
|
||||
assert!(keys.contains(&1));
|
||||
assert!(keys.contains(&2));
|
||||
assert!(keys.contains(&3));
|
||||
}
|
||||
|
||||
#[test]
|
||||
fn values() {
|
||||
let vec = vec![(1, 'a'), (2, 'b'), (3, 'c')];
|
||||
let map: IndexMap<_, _> = vec.into_par_iter().collect();
|
||||
let values: Vec<_> = map.par_values().copied().collect();
|
||||
assert_eq!(values.len(), 3);
|
||||
assert!(values.contains(&'a'));
|
||||
assert!(values.contains(&'b'));
|
||||
assert!(values.contains(&'c'));
|
||||
}
|
||||
|
||||
#[test]
|
||||
fn values_mut() {
|
||||
let vec = vec![(1, 1), (2, 2), (3, 3)];
|
||||
let mut map: IndexMap<_, _> = vec.into_par_iter().collect();
|
||||
map.par_values_mut().for_each(|value| *value *= 2);
|
||||
let values: Vec<_> = map.par_values().copied().collect();
|
||||
assert_eq!(values.len(), 3);
|
||||
assert!(values.contains(&2));
|
||||
assert!(values.contains(&4));
|
||||
assert!(values.contains(&6));
|
||||
}
|
||||
}
|
||||
27
clamav/libclamav_rust/.cargo/vendor/indexmap/src/rayon/mod.rs
vendored
Normal file
27
clamav/libclamav_rust/.cargo/vendor/indexmap/src/rayon/mod.rs
vendored
Normal file
@@ -0,0 +1,27 @@
|
||||
use rayon::prelude::*;
|
||||
|
||||
use alloc::collections::LinkedList;
|
||||
|
||||
use crate::vec::Vec;
|
||||
|
||||
pub mod map;
|
||||
pub mod set;
|
||||
|
||||
// This form of intermediate collection is also how Rayon collects `HashMap`.
|
||||
// Note that the order will also be preserved!
|
||||
fn collect<I: IntoParallelIterator>(iter: I) -> LinkedList<Vec<I::Item>> {
|
||||
iter.into_par_iter()
|
||||
.fold(Vec::new, |mut vec, elem| {
|
||||
vec.push(elem);
|
||||
vec
|
||||
})
|
||||
.map(|vec| {
|
||||
let mut list = LinkedList::new();
|
||||
list.push_back(vec);
|
||||
list
|
||||
})
|
||||
.reduce(LinkedList::new, |mut list1, mut list2| {
|
||||
list1.append(&mut list2);
|
||||
list1
|
||||
})
|
||||
}
|
||||
741
clamav/libclamav_rust/.cargo/vendor/indexmap/src/rayon/set.rs
vendored
Normal file
741
clamav/libclamav_rust/.cargo/vendor/indexmap/src/rayon/set.rs
vendored
Normal file
@@ -0,0 +1,741 @@
|
||||
//! Parallel iterator types for `IndexSet` with [rayon](https://docs.rs/rayon/1.0/rayon).
|
||||
//!
|
||||
//! You will rarely need to interact with this module directly unless you need to name one of the
|
||||
//! iterator types.
|
||||
//!
|
||||
//! Requires crate feature `"rayon"`.
|
||||
|
||||
use super::collect;
|
||||
use rayon::iter::plumbing::{Consumer, ProducerCallback, UnindexedConsumer};
|
||||
use rayon::prelude::*;
|
||||
|
||||
use crate::vec::Vec;
|
||||
use core::cmp::Ordering;
|
||||
use core::fmt;
|
||||
use core::hash::{BuildHasher, Hash};
|
||||
use core::ops::RangeBounds;
|
||||
|
||||
use crate::Entries;
|
||||
use crate::IndexSet;
|
||||
|
||||
type Bucket<T> = crate::Bucket<T, ()>;
|
||||
|
||||
/// Requires crate feature `"rayon"`.
|
||||
impl<T, S> IntoParallelIterator for IndexSet<T, S>
|
||||
where
|
||||
T: Send,
|
||||
{
|
||||
type Item = T;
|
||||
type Iter = IntoParIter<T>;
|
||||
|
||||
fn into_par_iter(self) -> Self::Iter {
|
||||
IntoParIter {
|
||||
entries: self.into_entries(),
|
||||
}
|
||||
}
|
||||
}
|
||||
|
||||
/// A parallel owning iterator over the items of a `IndexSet`.
|
||||
///
|
||||
/// This `struct` is created by the [`into_par_iter`] method on [`IndexSet`]
|
||||
/// (provided by rayon's `IntoParallelIterator` trait). See its documentation for more.
|
||||
///
|
||||
/// [`IndexSet`]: ../struct.IndexSet.html
|
||||
/// [`into_par_iter`]: ../struct.IndexSet.html#method.into_par_iter
|
||||
pub struct IntoParIter<T> {
|
||||
entries: Vec<Bucket<T>>,
|
||||
}
|
||||
|
||||
impl<T: fmt::Debug> fmt::Debug for IntoParIter<T> {
|
||||
fn fmt(&self, f: &mut fmt::Formatter<'_>) -> fmt::Result {
|
||||
let iter = self.entries.iter().map(Bucket::key_ref);
|
||||
f.debug_list().entries(iter).finish()
|
||||
}
|
||||
}
|
||||
|
||||
impl<T: Send> ParallelIterator for IntoParIter<T> {
|
||||
type Item = T;
|
||||
|
||||
parallel_iterator_methods!(Bucket::key);
|
||||
}
|
||||
|
||||
impl<T: Send> IndexedParallelIterator for IntoParIter<T> {
|
||||
indexed_parallel_iterator_methods!(Bucket::key);
|
||||
}
|
||||
|
||||
/// Requires crate feature `"rayon"`.
|
||||
impl<'a, T, S> IntoParallelIterator for &'a IndexSet<T, S>
|
||||
where
|
||||
T: Sync,
|
||||
{
|
||||
type Item = &'a T;
|
||||
type Iter = ParIter<'a, T>;
|
||||
|
||||
fn into_par_iter(self) -> Self::Iter {
|
||||
ParIter {
|
||||
entries: self.as_entries(),
|
||||
}
|
||||
}
|
||||
}
|
||||
|
||||
/// A parallel iterator over the items of a `IndexSet`.
|
||||
///
|
||||
/// This `struct` is created by the [`par_iter`] method on [`IndexSet`]
|
||||
/// (provided by rayon's `IntoParallelRefIterator` trait). See its documentation for more.
|
||||
///
|
||||
/// [`IndexSet`]: ../struct.IndexSet.html
|
||||
/// [`par_iter`]: ../struct.IndexSet.html#method.par_iter
|
||||
pub struct ParIter<'a, T> {
|
||||
entries: &'a [Bucket<T>],
|
||||
}
|
||||
|
||||
impl<T> Clone for ParIter<'_, T> {
|
||||
fn clone(&self) -> Self {
|
||||
ParIter { ..*self }
|
||||
}
|
||||
}
|
||||
|
||||
impl<T: fmt::Debug> fmt::Debug for ParIter<'_, T> {
|
||||
fn fmt(&self, f: &mut fmt::Formatter<'_>) -> fmt::Result {
|
||||
let iter = self.entries.iter().map(Bucket::key_ref);
|
||||
f.debug_list().entries(iter).finish()
|
||||
}
|
||||
}
|
||||
|
||||
impl<'a, T: Sync> ParallelIterator for ParIter<'a, T> {
|
||||
type Item = &'a T;
|
||||
|
||||
parallel_iterator_methods!(Bucket::key_ref);
|
||||
}
|
||||
|
||||
impl<T: Sync> IndexedParallelIterator for ParIter<'_, T> {
|
||||
indexed_parallel_iterator_methods!(Bucket::key_ref);
|
||||
}
|
||||
|
||||
/// Requires crate feature `"rayon"`.
|
||||
impl<'a, T, S> ParallelDrainRange<usize> for &'a mut IndexSet<T, S>
|
||||
where
|
||||
T: Send,
|
||||
{
|
||||
type Item = T;
|
||||
type Iter = ParDrain<'a, T>;
|
||||
|
||||
fn par_drain<R: RangeBounds<usize>>(self, range: R) -> Self::Iter {
|
||||
ParDrain {
|
||||
entries: self.map.core.par_drain(range),
|
||||
}
|
||||
}
|
||||
}
|
||||
|
||||
/// A parallel draining iterator over the items of a `IndexSet`.
|
||||
///
|
||||
/// This `struct` is created by the [`par_drain`] method on [`IndexSet`]
|
||||
/// (provided by rayon's `ParallelDrainRange` trait). See its documentation for more.
|
||||
///
|
||||
/// [`par_drain`]: ../struct.IndexSet.html#method.par_drain
|
||||
/// [`IndexSet`]: ../struct.IndexSet.html
|
||||
pub struct ParDrain<'a, T: Send> {
|
||||
entries: rayon::vec::Drain<'a, Bucket<T>>,
|
||||
}
|
||||
|
||||
impl<T: Send> ParallelIterator for ParDrain<'_, T> {
|
||||
type Item = T;
|
||||
|
||||
parallel_iterator_methods!(Bucket::key);
|
||||
}
|
||||
|
||||
impl<T: Send> IndexedParallelIterator for ParDrain<'_, T> {
|
||||
indexed_parallel_iterator_methods!(Bucket::key);
|
||||
}
|
||||
|
||||
/// Parallel iterator methods and other parallel methods.
|
||||
///
|
||||
/// The following methods **require crate feature `"rayon"`**.
|
||||
///
|
||||
/// See also the `IntoParallelIterator` implementations.
|
||||
impl<T, S> IndexSet<T, S>
|
||||
where
|
||||
T: Hash + Eq + Sync,
|
||||
S: BuildHasher + Sync,
|
||||
{
|
||||
/// Return a parallel iterator over the values that are in `self` but not `other`.
|
||||
///
|
||||
/// While parallel iterators can process items in any order, their relative order
|
||||
/// in the `self` set is still preserved for operations like `reduce` and `collect`.
|
||||
pub fn par_difference<'a, S2>(
|
||||
&'a self,
|
||||
other: &'a IndexSet<T, S2>,
|
||||
) -> ParDifference<'a, T, S, S2>
|
||||
where
|
||||
S2: BuildHasher + Sync,
|
||||
{
|
||||
ParDifference {
|
||||
set1: self,
|
||||
set2: other,
|
||||
}
|
||||
}
|
||||
|
||||
/// Return a parallel iterator over the values that are in `self` or `other`,
|
||||
/// but not in both.
|
||||
///
|
||||
/// While parallel iterators can process items in any order, their relative order
|
||||
/// in the sets is still preserved for operations like `reduce` and `collect`.
|
||||
/// Values from `self` are produced in their original order, followed by
|
||||
/// values from `other` in their original order.
|
||||
pub fn par_symmetric_difference<'a, S2>(
|
||||
&'a self,
|
||||
other: &'a IndexSet<T, S2>,
|
||||
) -> ParSymmetricDifference<'a, T, S, S2>
|
||||
where
|
||||
S2: BuildHasher + Sync,
|
||||
{
|
||||
ParSymmetricDifference {
|
||||
set1: self,
|
||||
set2: other,
|
||||
}
|
||||
}
|
||||
|
||||
/// Return a parallel iterator over the values that are in both `self` and `other`.
|
||||
///
|
||||
/// While parallel iterators can process items in any order, their relative order
|
||||
/// in the `self` set is still preserved for operations like `reduce` and `collect`.
|
||||
pub fn par_intersection<'a, S2>(
|
||||
&'a self,
|
||||
other: &'a IndexSet<T, S2>,
|
||||
) -> ParIntersection<'a, T, S, S2>
|
||||
where
|
||||
S2: BuildHasher + Sync,
|
||||
{
|
||||
ParIntersection {
|
||||
set1: self,
|
||||
set2: other,
|
||||
}
|
||||
}
|
||||
|
||||
/// Return a parallel iterator over all values that are in `self` or `other`.
|
||||
///
|
||||
/// While parallel iterators can process items in any order, their relative order
|
||||
/// in the sets is still preserved for operations like `reduce` and `collect`.
|
||||
/// Values from `self` are produced in their original order, followed by
|
||||
/// values that are unique to `other` in their original order.
|
||||
pub fn par_union<'a, S2>(&'a self, other: &'a IndexSet<T, S2>) -> ParUnion<'a, T, S, S2>
|
||||
where
|
||||
S2: BuildHasher + Sync,
|
||||
{
|
||||
ParUnion {
|
||||
set1: self,
|
||||
set2: other,
|
||||
}
|
||||
}
|
||||
|
||||
/// Returns `true` if `self` contains all of the same values as `other`,
|
||||
/// regardless of each set's indexed order, determined in parallel.
|
||||
pub fn par_eq<S2>(&self, other: &IndexSet<T, S2>) -> bool
|
||||
where
|
||||
S2: BuildHasher + Sync,
|
||||
{
|
||||
self.len() == other.len() && self.par_is_subset(other)
|
||||
}
|
||||
|
||||
/// Returns `true` if `self` has no elements in common with `other`,
|
||||
/// determined in parallel.
|
||||
pub fn par_is_disjoint<S2>(&self, other: &IndexSet<T, S2>) -> bool
|
||||
where
|
||||
S2: BuildHasher + Sync,
|
||||
{
|
||||
if self.len() <= other.len() {
|
||||
self.par_iter().all(move |value| !other.contains(value))
|
||||
} else {
|
||||
other.par_iter().all(move |value| !self.contains(value))
|
||||
}
|
||||
}
|
||||
|
||||
/// Returns `true` if all elements of `other` are contained in `self`,
|
||||
/// determined in parallel.
|
||||
pub fn par_is_superset<S2>(&self, other: &IndexSet<T, S2>) -> bool
|
||||
where
|
||||
S2: BuildHasher + Sync,
|
||||
{
|
||||
other.par_is_subset(self)
|
||||
}
|
||||
|
||||
/// Returns `true` if all elements of `self` are contained in `other`,
|
||||
/// determined in parallel.
|
||||
pub fn par_is_subset<S2>(&self, other: &IndexSet<T, S2>) -> bool
|
||||
where
|
||||
S2: BuildHasher + Sync,
|
||||
{
|
||||
self.len() <= other.len() && self.par_iter().all(move |value| other.contains(value))
|
||||
}
|
||||
}
|
||||
|
||||
/// A parallel iterator producing elements in the difference of `IndexSet`s.
|
||||
///
|
||||
/// This `struct` is created by the [`par_difference`] method on [`IndexSet`].
|
||||
/// See its documentation for more.
|
||||
///
|
||||
/// [`IndexSet`]: ../struct.IndexSet.html
|
||||
/// [`par_difference`]: ../struct.IndexSet.html#method.par_difference
|
||||
pub struct ParDifference<'a, T, S1, S2> {
|
||||
set1: &'a IndexSet<T, S1>,
|
||||
set2: &'a IndexSet<T, S2>,
|
||||
}
|
||||
|
||||
impl<T, S1, S2> Clone for ParDifference<'_, T, S1, S2> {
|
||||
fn clone(&self) -> Self {
|
||||
ParDifference { ..*self }
|
||||
}
|
||||
}
|
||||
|
||||
impl<T, S1, S2> fmt::Debug for ParDifference<'_, T, S1, S2>
|
||||
where
|
||||
T: fmt::Debug + Eq + Hash,
|
||||
S1: BuildHasher,
|
||||
S2: BuildHasher,
|
||||
{
|
||||
fn fmt(&self, f: &mut fmt::Formatter<'_>) -> fmt::Result {
|
||||
f.debug_list()
|
||||
.entries(self.set1.difference(self.set2))
|
||||
.finish()
|
||||
}
|
||||
}
|
||||
|
||||
impl<'a, T, S1, S2> ParallelIterator for ParDifference<'a, T, S1, S2>
|
||||
where
|
||||
T: Hash + Eq + Sync,
|
||||
S1: BuildHasher + Sync,
|
||||
S2: BuildHasher + Sync,
|
||||
{
|
||||
type Item = &'a T;
|
||||
|
||||
fn drive_unindexed<C>(self, consumer: C) -> C::Result
|
||||
where
|
||||
C: UnindexedConsumer<Self::Item>,
|
||||
{
|
||||
let Self { set1, set2 } = self;
|
||||
|
||||
set1.par_iter()
|
||||
.filter(move |&item| !set2.contains(item))
|
||||
.drive_unindexed(consumer)
|
||||
}
|
||||
}
|
||||
|
||||
/// A parallel iterator producing elements in the intersection of `IndexSet`s.
|
||||
///
|
||||
/// This `struct` is created by the [`par_intersection`] method on [`IndexSet`].
|
||||
/// See its documentation for more.
|
||||
///
|
||||
/// [`IndexSet`]: ../struct.IndexSet.html
|
||||
/// [`par_intersection`]: ../struct.IndexSet.html#method.par_intersection
|
||||
pub struct ParIntersection<'a, T, S1, S2> {
|
||||
set1: &'a IndexSet<T, S1>,
|
||||
set2: &'a IndexSet<T, S2>,
|
||||
}
|
||||
|
||||
impl<T, S1, S2> Clone for ParIntersection<'_, T, S1, S2> {
|
||||
fn clone(&self) -> Self {
|
||||
ParIntersection { ..*self }
|
||||
}
|
||||
}
|
||||
|
||||
impl<T, S1, S2> fmt::Debug for ParIntersection<'_, T, S1, S2>
|
||||
where
|
||||
T: fmt::Debug + Eq + Hash,
|
||||
S1: BuildHasher,
|
||||
S2: BuildHasher,
|
||||
{
|
||||
fn fmt(&self, f: &mut fmt::Formatter<'_>) -> fmt::Result {
|
||||
f.debug_list()
|
||||
.entries(self.set1.intersection(self.set2))
|
||||
.finish()
|
||||
}
|
||||
}
|
||||
|
||||
impl<'a, T, S1, S2> ParallelIterator for ParIntersection<'a, T, S1, S2>
|
||||
where
|
||||
T: Hash + Eq + Sync,
|
||||
S1: BuildHasher + Sync,
|
||||
S2: BuildHasher + Sync,
|
||||
{
|
||||
type Item = &'a T;
|
||||
|
||||
fn drive_unindexed<C>(self, consumer: C) -> C::Result
|
||||
where
|
||||
C: UnindexedConsumer<Self::Item>,
|
||||
{
|
||||
let Self { set1, set2 } = self;
|
||||
|
||||
set1.par_iter()
|
||||
.filter(move |&item| set2.contains(item))
|
||||
.drive_unindexed(consumer)
|
||||
}
|
||||
}
|
||||
|
||||
/// A parallel iterator producing elements in the symmetric difference of `IndexSet`s.
|
||||
///
|
||||
/// This `struct` is created by the [`par_symmetric_difference`] method on
|
||||
/// [`IndexSet`]. See its documentation for more.
|
||||
///
|
||||
/// [`IndexSet`]: ../struct.IndexSet.html
|
||||
/// [`par_symmetric_difference`]: ../struct.IndexSet.html#method.par_symmetric_difference
|
||||
pub struct ParSymmetricDifference<'a, T, S1, S2> {
|
||||
set1: &'a IndexSet<T, S1>,
|
||||
set2: &'a IndexSet<T, S2>,
|
||||
}
|
||||
|
||||
impl<T, S1, S2> Clone for ParSymmetricDifference<'_, T, S1, S2> {
|
||||
fn clone(&self) -> Self {
|
||||
ParSymmetricDifference { ..*self }
|
||||
}
|
||||
}
|
||||
|
||||
impl<T, S1, S2> fmt::Debug for ParSymmetricDifference<'_, T, S1, S2>
|
||||
where
|
||||
T: fmt::Debug + Eq + Hash,
|
||||
S1: BuildHasher,
|
||||
S2: BuildHasher,
|
||||
{
|
||||
fn fmt(&self, f: &mut fmt::Formatter<'_>) -> fmt::Result {
|
||||
f.debug_list()
|
||||
.entries(self.set1.symmetric_difference(self.set2))
|
||||
.finish()
|
||||
}
|
||||
}
|
||||
|
||||
impl<'a, T, S1, S2> ParallelIterator for ParSymmetricDifference<'a, T, S1, S2>
|
||||
where
|
||||
T: Hash + Eq + Sync,
|
||||
S1: BuildHasher + Sync,
|
||||
S2: BuildHasher + Sync,
|
||||
{
|
||||
type Item = &'a T;
|
||||
|
||||
fn drive_unindexed<C>(self, consumer: C) -> C::Result
|
||||
where
|
||||
C: UnindexedConsumer<Self::Item>,
|
||||
{
|
||||
let Self { set1, set2 } = self;
|
||||
|
||||
set1.par_difference(set2)
|
||||
.chain(set2.par_difference(set1))
|
||||
.drive_unindexed(consumer)
|
||||
}
|
||||
}
|
||||
|
||||
/// A parallel iterator producing elements in the union of `IndexSet`s.
|
||||
///
|
||||
/// This `struct` is created by the [`par_union`] method on [`IndexSet`].
|
||||
/// See its documentation for more.
|
||||
///
|
||||
/// [`IndexSet`]: ../struct.IndexSet.html
|
||||
/// [`par_union`]: ../struct.IndexSet.html#method.par_union
|
||||
pub struct ParUnion<'a, T, S1, S2> {
|
||||
set1: &'a IndexSet<T, S1>,
|
||||
set2: &'a IndexSet<T, S2>,
|
||||
}
|
||||
|
||||
impl<T, S1, S2> Clone for ParUnion<'_, T, S1, S2> {
|
||||
fn clone(&self) -> Self {
|
||||
ParUnion { ..*self }
|
||||
}
|
||||
}
|
||||
|
||||
impl<T, S1, S2> fmt::Debug for ParUnion<'_, T, S1, S2>
|
||||
where
|
||||
T: fmt::Debug + Eq + Hash,
|
||||
S1: BuildHasher,
|
||||
S2: BuildHasher,
|
||||
{
|
||||
fn fmt(&self, f: &mut fmt::Formatter<'_>) -> fmt::Result {
|
||||
f.debug_list().entries(self.set1.union(self.set2)).finish()
|
||||
}
|
||||
}
|
||||
|
||||
impl<'a, T, S1, S2> ParallelIterator for ParUnion<'a, T, S1, S2>
|
||||
where
|
||||
T: Hash + Eq + Sync,
|
||||
S1: BuildHasher + Sync,
|
||||
S2: BuildHasher + Sync,
|
||||
{
|
||||
type Item = &'a T;
|
||||
|
||||
fn drive_unindexed<C>(self, consumer: C) -> C::Result
|
||||
where
|
||||
C: UnindexedConsumer<Self::Item>,
|
||||
{
|
||||
let Self { set1, set2 } = self;
|
||||
|
||||
set1.par_iter()
|
||||
.chain(set2.par_difference(set1))
|
||||
.drive_unindexed(consumer)
|
||||
}
|
||||
}
|
||||
|
||||
/// Parallel sorting methods.
|
||||
///
|
||||
/// The following methods **require crate feature `"rayon"`**.
|
||||
impl<T, S> IndexSet<T, S>
|
||||
where
|
||||
T: Hash + Eq + Send,
|
||||
S: BuildHasher + Send,
|
||||
{
|
||||
/// Sort the set’s values in parallel by their default ordering.
|
||||
pub fn par_sort(&mut self)
|
||||
where
|
||||
T: Ord,
|
||||
{
|
||||
self.with_entries(|entries| {
|
||||
entries.par_sort_by(|a, b| T::cmp(&a.key, &b.key));
|
||||
});
|
||||
}
|
||||
|
||||
/// Sort the set’s values in place and in parallel, using the comparison function `cmp`.
|
||||
pub fn par_sort_by<F>(&mut self, cmp: F)
|
||||
where
|
||||
F: Fn(&T, &T) -> Ordering + Sync,
|
||||
{
|
||||
self.with_entries(|entries| {
|
||||
entries.par_sort_by(move |a, b| cmp(&a.key, &b.key));
|
||||
});
|
||||
}
|
||||
|
||||
/// Sort the values of the set in parallel and return a by-value parallel iterator of
|
||||
/// the values with the result.
|
||||
pub fn par_sorted_by<F>(self, cmp: F) -> IntoParIter<T>
|
||||
where
|
||||
F: Fn(&T, &T) -> Ordering + Sync,
|
||||
{
|
||||
let mut entries = self.into_entries();
|
||||
entries.par_sort_by(move |a, b| cmp(&a.key, &b.key));
|
||||
IntoParIter { entries }
|
||||
}
|
||||
|
||||
/// Sort the set's values in parallel by their default ordering.
|
||||
pub fn par_sort_unstable(&mut self)
|
||||
where
|
||||
T: Ord,
|
||||
{
|
||||
self.with_entries(|entries| {
|
||||
entries.par_sort_unstable_by(|a, b| T::cmp(&a.key, &b.key));
|
||||
});
|
||||
}
|
||||
|
||||
/// Sort the set’s values in place and in parallel, using the comparison function `cmp`.
|
||||
pub fn par_sort_unstable_by<F>(&mut self, cmp: F)
|
||||
where
|
||||
F: Fn(&T, &T) -> Ordering + Sync,
|
||||
{
|
||||
self.with_entries(|entries| {
|
||||
entries.par_sort_unstable_by(move |a, b| cmp(&a.key, &b.key));
|
||||
});
|
||||
}
|
||||
|
||||
/// Sort the values of the set in parallel and return a by-value parallel iterator of
|
||||
/// the values with the result.
|
||||
pub fn par_sorted_unstable_by<F>(self, cmp: F) -> IntoParIter<T>
|
||||
where
|
||||
F: Fn(&T, &T) -> Ordering + Sync,
|
||||
{
|
||||
let mut entries = self.into_entries();
|
||||
entries.par_sort_unstable_by(move |a, b| cmp(&a.key, &b.key));
|
||||
IntoParIter { entries }
|
||||
}
|
||||
}
|
||||
|
||||
/// Requires crate feature `"rayon"`.
|
||||
impl<T, S> FromParallelIterator<T> for IndexSet<T, S>
|
||||
where
|
||||
T: Eq + Hash + Send,
|
||||
S: BuildHasher + Default + Send,
|
||||
{
|
||||
fn from_par_iter<I>(iter: I) -> Self
|
||||
where
|
||||
I: IntoParallelIterator<Item = T>,
|
||||
{
|
||||
let list = collect(iter);
|
||||
let len = list.iter().map(Vec::len).sum();
|
||||
let mut set = Self::with_capacity_and_hasher(len, S::default());
|
||||
for vec in list {
|
||||
set.extend(vec);
|
||||
}
|
||||
set
|
||||
}
|
||||
}
|
||||
|
||||
/// Requires crate feature `"rayon"`.
|
||||
impl<T, S> ParallelExtend<T> for IndexSet<T, S>
|
||||
where
|
||||
T: Eq + Hash + Send,
|
||||
S: BuildHasher + Send,
|
||||
{
|
||||
fn par_extend<I>(&mut self, iter: I)
|
||||
where
|
||||
I: IntoParallelIterator<Item = T>,
|
||||
{
|
||||
for vec in collect(iter) {
|
||||
self.extend(vec);
|
||||
}
|
||||
}
|
||||
}
|
||||
|
||||
/// Requires crate feature `"rayon"`.
|
||||
impl<'a, T: 'a, S> ParallelExtend<&'a T> for IndexSet<T, S>
|
||||
where
|
||||
T: Copy + Eq + Hash + Send + Sync,
|
||||
S: BuildHasher + Send,
|
||||
{
|
||||
fn par_extend<I>(&mut self, iter: I)
|
||||
where
|
||||
I: IntoParallelIterator<Item = &'a T>,
|
||||
{
|
||||
for vec in collect(iter) {
|
||||
self.extend(vec);
|
||||
}
|
||||
}
|
||||
}
|
||||
|
||||
#[cfg(test)]
|
||||
mod tests {
|
||||
use super::*;
|
||||
|
||||
#[test]
|
||||
fn insert_order() {
|
||||
let insert = [0, 4, 2, 12, 8, 7, 11, 5, 3, 17, 19, 22, 23];
|
||||
let mut set = IndexSet::new();
|
||||
|
||||
for &elt in &insert {
|
||||
set.insert(elt);
|
||||
}
|
||||
|
||||
assert_eq!(set.par_iter().count(), set.len());
|
||||
assert_eq!(set.par_iter().count(), insert.len());
|
||||
insert.par_iter().zip(&set).for_each(|(a, b)| {
|
||||
assert_eq!(a, b);
|
||||
});
|
||||
(0..insert.len())
|
||||
.into_par_iter()
|
||||
.zip(&set)
|
||||
.for_each(|(i, v)| {
|
||||
assert_eq!(set.get_index(i).unwrap(), v);
|
||||
});
|
||||
}
|
||||
|
||||
#[test]
|
||||
fn partial_eq_and_eq() {
|
||||
let mut set_a = IndexSet::new();
|
||||
set_a.insert(1);
|
||||
set_a.insert(2);
|
||||
let mut set_b = set_a.clone();
|
||||
assert!(set_a.par_eq(&set_b));
|
||||
set_b.swap_remove(&1);
|
||||
assert!(!set_a.par_eq(&set_b));
|
||||
set_b.insert(3);
|
||||
assert!(!set_a.par_eq(&set_b));
|
||||
|
||||
let set_c: IndexSet<_> = set_b.into_par_iter().collect();
|
||||
assert!(!set_a.par_eq(&set_c));
|
||||
assert!(!set_c.par_eq(&set_a));
|
||||
}
|
||||
|
||||
#[test]
|
||||
fn extend() {
|
||||
let mut set = IndexSet::new();
|
||||
set.par_extend(vec![&1, &2, &3, &4]);
|
||||
set.par_extend(vec![5, 6]);
|
||||
assert_eq!(
|
||||
set.into_par_iter().collect::<Vec<_>>(),
|
||||
vec![1, 2, 3, 4, 5, 6]
|
||||
);
|
||||
}
|
||||
|
||||
#[test]
|
||||
fn comparisons() {
|
||||
let set_a: IndexSet<_> = (0..3).collect();
|
||||
let set_b: IndexSet<_> = (3..6).collect();
|
||||
let set_c: IndexSet<_> = (0..6).collect();
|
||||
let set_d: IndexSet<_> = (3..9).collect();
|
||||
|
||||
assert!(!set_a.par_is_disjoint(&set_a));
|
||||
assert!(set_a.par_is_subset(&set_a));
|
||||
assert!(set_a.par_is_superset(&set_a));
|
||||
|
||||
assert!(set_a.par_is_disjoint(&set_b));
|
||||
assert!(set_b.par_is_disjoint(&set_a));
|
||||
assert!(!set_a.par_is_subset(&set_b));
|
||||
assert!(!set_b.par_is_subset(&set_a));
|
||||
assert!(!set_a.par_is_superset(&set_b));
|
||||
assert!(!set_b.par_is_superset(&set_a));
|
||||
|
||||
assert!(!set_a.par_is_disjoint(&set_c));
|
||||
assert!(!set_c.par_is_disjoint(&set_a));
|
||||
assert!(set_a.par_is_subset(&set_c));
|
||||
assert!(!set_c.par_is_subset(&set_a));
|
||||
assert!(!set_a.par_is_superset(&set_c));
|
||||
assert!(set_c.par_is_superset(&set_a));
|
||||
|
||||
assert!(!set_c.par_is_disjoint(&set_d));
|
||||
assert!(!set_d.par_is_disjoint(&set_c));
|
||||
assert!(!set_c.par_is_subset(&set_d));
|
||||
assert!(!set_d.par_is_subset(&set_c));
|
||||
assert!(!set_c.par_is_superset(&set_d));
|
||||
assert!(!set_d.par_is_superset(&set_c));
|
||||
}
|
||||
|
||||
#[test]
|
||||
fn iter_comparisons() {
|
||||
use std::iter::empty;
|
||||
|
||||
fn check<'a, I1, I2>(iter1: I1, iter2: I2)
|
||||
where
|
||||
I1: ParallelIterator<Item = &'a i32>,
|
||||
I2: Iterator<Item = i32>,
|
||||
{
|
||||
let v1: Vec<_> = iter1.copied().collect();
|
||||
let v2: Vec<_> = iter2.collect();
|
||||
assert_eq!(v1, v2);
|
||||
}
|
||||
|
||||
let set_a: IndexSet<_> = (0..3).collect();
|
||||
let set_b: IndexSet<_> = (3..6).collect();
|
||||
let set_c: IndexSet<_> = (0..6).collect();
|
||||
let set_d: IndexSet<_> = (3..9).rev().collect();
|
||||
|
||||
check(set_a.par_difference(&set_a), empty());
|
||||
check(set_a.par_symmetric_difference(&set_a), empty());
|
||||
check(set_a.par_intersection(&set_a), 0..3);
|
||||
check(set_a.par_union(&set_a), 0..3);
|
||||
|
||||
check(set_a.par_difference(&set_b), 0..3);
|
||||
check(set_b.par_difference(&set_a), 3..6);
|
||||
check(set_a.par_symmetric_difference(&set_b), 0..6);
|
||||
check(set_b.par_symmetric_difference(&set_a), (3..6).chain(0..3));
|
||||
check(set_a.par_intersection(&set_b), empty());
|
||||
check(set_b.par_intersection(&set_a), empty());
|
||||
check(set_a.par_union(&set_b), 0..6);
|
||||
check(set_b.par_union(&set_a), (3..6).chain(0..3));
|
||||
|
||||
check(set_a.par_difference(&set_c), empty());
|
||||
check(set_c.par_difference(&set_a), 3..6);
|
||||
check(set_a.par_symmetric_difference(&set_c), 3..6);
|
||||
check(set_c.par_symmetric_difference(&set_a), 3..6);
|
||||
check(set_a.par_intersection(&set_c), 0..3);
|
||||
check(set_c.par_intersection(&set_a), 0..3);
|
||||
check(set_a.par_union(&set_c), 0..6);
|
||||
check(set_c.par_union(&set_a), 0..6);
|
||||
|
||||
check(set_c.par_difference(&set_d), 0..3);
|
||||
check(set_d.par_difference(&set_c), (6..9).rev());
|
||||
check(
|
||||
set_c.par_symmetric_difference(&set_d),
|
||||
(0..3).chain((6..9).rev()),
|
||||
);
|
||||
check(
|
||||
set_d.par_symmetric_difference(&set_c),
|
||||
(6..9).rev().chain(0..3),
|
||||
);
|
||||
check(set_c.par_intersection(&set_d), 3..6);
|
||||
check(set_d.par_intersection(&set_c), (3..6).rev());
|
||||
check(set_c.par_union(&set_d), (0..6).chain((6..9).rev()));
|
||||
check(set_d.par_union(&set_c), (3..9).rev().chain(0..3));
|
||||
}
|
||||
}
|
||||
158
clamav/libclamav_rust/.cargo/vendor/indexmap/src/rustc.rs
vendored
Normal file
158
clamav/libclamav_rust/.cargo/vendor/indexmap/src/rustc.rs
vendored
Normal file
@@ -0,0 +1,158 @@
|
||||
//! Minimal support for `rustc-rayon`, not intended for general use.
|
||||
|
||||
use crate::vec::Vec;
|
||||
use crate::{Bucket, Entries, IndexMap, IndexSet};
|
||||
|
||||
use rustc_rayon::iter::plumbing::{Consumer, ProducerCallback, UnindexedConsumer};
|
||||
use rustc_rayon::iter::{IndexedParallelIterator, IntoParallelIterator, ParallelIterator};
|
||||
|
||||
mod map {
|
||||
use super::*;
|
||||
|
||||
impl<K, V, S> IntoParallelIterator for IndexMap<K, V, S>
|
||||
where
|
||||
K: Send,
|
||||
V: Send,
|
||||
{
|
||||
type Item = (K, V);
|
||||
type Iter = IntoParIter<K, V>;
|
||||
|
||||
fn into_par_iter(self) -> Self::Iter {
|
||||
IntoParIter {
|
||||
entries: self.into_entries(),
|
||||
}
|
||||
}
|
||||
}
|
||||
|
||||
pub struct IntoParIter<K, V> {
|
||||
entries: Vec<Bucket<K, V>>,
|
||||
}
|
||||
|
||||
impl<K: Send, V: Send> ParallelIterator for IntoParIter<K, V> {
|
||||
type Item = (K, V);
|
||||
|
||||
parallel_iterator_methods!(Bucket::key_value);
|
||||
}
|
||||
|
||||
impl<K: Send, V: Send> IndexedParallelIterator for IntoParIter<K, V> {
|
||||
indexed_parallel_iterator_methods!(Bucket::key_value);
|
||||
}
|
||||
|
||||
impl<'a, K, V, S> IntoParallelIterator for &'a IndexMap<K, V, S>
|
||||
where
|
||||
K: Sync,
|
||||
V: Sync,
|
||||
{
|
||||
type Item = (&'a K, &'a V);
|
||||
type Iter = ParIter<'a, K, V>;
|
||||
|
||||
fn into_par_iter(self) -> Self::Iter {
|
||||
ParIter {
|
||||
entries: self.as_entries(),
|
||||
}
|
||||
}
|
||||
}
|
||||
|
||||
pub struct ParIter<'a, K, V> {
|
||||
entries: &'a [Bucket<K, V>],
|
||||
}
|
||||
|
||||
impl<'a, K: Sync, V: Sync> ParallelIterator for ParIter<'a, K, V> {
|
||||
type Item = (&'a K, &'a V);
|
||||
|
||||
parallel_iterator_methods!(Bucket::refs);
|
||||
}
|
||||
|
||||
impl<K: Sync, V: Sync> IndexedParallelIterator for ParIter<'_, K, V> {
|
||||
indexed_parallel_iterator_methods!(Bucket::refs);
|
||||
}
|
||||
|
||||
impl<'a, K, V, S> IntoParallelIterator for &'a mut IndexMap<K, V, S>
|
||||
where
|
||||
K: Sync + Send,
|
||||
V: Send,
|
||||
{
|
||||
type Item = (&'a K, &'a mut V);
|
||||
type Iter = ParIterMut<'a, K, V>;
|
||||
|
||||
fn into_par_iter(self) -> Self::Iter {
|
||||
ParIterMut {
|
||||
entries: self.as_entries_mut(),
|
||||
}
|
||||
}
|
||||
}
|
||||
|
||||
pub struct ParIterMut<'a, K, V> {
|
||||
entries: &'a mut [Bucket<K, V>],
|
||||
}
|
||||
|
||||
impl<'a, K: Sync + Send, V: Send> ParallelIterator for ParIterMut<'a, K, V> {
|
||||
type Item = (&'a K, &'a mut V);
|
||||
|
||||
parallel_iterator_methods!(Bucket::ref_mut);
|
||||
}
|
||||
|
||||
impl<K: Sync + Send, V: Send> IndexedParallelIterator for ParIterMut<'_, K, V> {
|
||||
indexed_parallel_iterator_methods!(Bucket::ref_mut);
|
||||
}
|
||||
}
|
||||
|
||||
mod set {
|
||||
use super::*;
|
||||
|
||||
impl<T, S> IntoParallelIterator for IndexSet<T, S>
|
||||
where
|
||||
T: Send,
|
||||
{
|
||||
type Item = T;
|
||||
type Iter = IntoParIter<T>;
|
||||
|
||||
fn into_par_iter(self) -> Self::Iter {
|
||||
IntoParIter {
|
||||
entries: self.into_entries(),
|
||||
}
|
||||
}
|
||||
}
|
||||
|
||||
pub struct IntoParIter<T> {
|
||||
entries: Vec<Bucket<T, ()>>,
|
||||
}
|
||||
|
||||
impl<T: Send> ParallelIterator for IntoParIter<T> {
|
||||
type Item = T;
|
||||
|
||||
parallel_iterator_methods!(Bucket::key);
|
||||
}
|
||||
|
||||
impl<T: Send> IndexedParallelIterator for IntoParIter<T> {
|
||||
indexed_parallel_iterator_methods!(Bucket::key);
|
||||
}
|
||||
|
||||
impl<'a, T, S> IntoParallelIterator for &'a IndexSet<T, S>
|
||||
where
|
||||
T: Sync,
|
||||
{
|
||||
type Item = &'a T;
|
||||
type Iter = ParIter<'a, T>;
|
||||
|
||||
fn into_par_iter(self) -> Self::Iter {
|
||||
ParIter {
|
||||
entries: self.as_entries(),
|
||||
}
|
||||
}
|
||||
}
|
||||
|
||||
pub struct ParIter<'a, T> {
|
||||
entries: &'a [Bucket<T, ()>],
|
||||
}
|
||||
|
||||
impl<'a, T: Sync> ParallelIterator for ParIter<'a, T> {
|
||||
type Item = &'a T;
|
||||
|
||||
parallel_iterator_methods!(Bucket::key_ref);
|
||||
}
|
||||
|
||||
impl<T: Sync> IndexedParallelIterator for ParIter<'_, T> {
|
||||
indexed_parallel_iterator_methods!(Bucket::key_ref);
|
||||
}
|
||||
}
|
||||
155
clamav/libclamav_rust/.cargo/vendor/indexmap/src/serde.rs
vendored
Normal file
155
clamav/libclamav_rust/.cargo/vendor/indexmap/src/serde.rs
vendored
Normal file
@@ -0,0 +1,155 @@
|
||||
use serde::de::value::{MapDeserializer, SeqDeserializer};
|
||||
use serde::de::{
|
||||
Deserialize, Deserializer, Error, IntoDeserializer, MapAccess, SeqAccess, Visitor,
|
||||
};
|
||||
use serde::ser::{Serialize, Serializer};
|
||||
|
||||
use core::fmt::{self, Formatter};
|
||||
use core::hash::{BuildHasher, Hash};
|
||||
use core::marker::PhantomData;
|
||||
|
||||
use crate::IndexMap;
|
||||
|
||||
/// Requires crate feature `"serde"` or `"serde-1"`
|
||||
impl<K, V, S> Serialize for IndexMap<K, V, S>
|
||||
where
|
||||
K: Serialize + Hash + Eq,
|
||||
V: Serialize,
|
||||
S: BuildHasher,
|
||||
{
|
||||
fn serialize<T>(&self, serializer: T) -> Result<T::Ok, T::Error>
|
||||
where
|
||||
T: Serializer,
|
||||
{
|
||||
serializer.collect_map(self)
|
||||
}
|
||||
}
|
||||
|
||||
struct IndexMapVisitor<K, V, S>(PhantomData<(K, V, S)>);
|
||||
|
||||
impl<'de, K, V, S> Visitor<'de> for IndexMapVisitor<K, V, S>
|
||||
where
|
||||
K: Deserialize<'de> + Eq + Hash,
|
||||
V: Deserialize<'de>,
|
||||
S: Default + BuildHasher,
|
||||
{
|
||||
type Value = IndexMap<K, V, S>;
|
||||
|
||||
fn expecting(&self, formatter: &mut Formatter<'_>) -> fmt::Result {
|
||||
write!(formatter, "a map")
|
||||
}
|
||||
|
||||
fn visit_map<A>(self, mut map: A) -> Result<Self::Value, A::Error>
|
||||
where
|
||||
A: MapAccess<'de>,
|
||||
{
|
||||
let mut values =
|
||||
IndexMap::with_capacity_and_hasher(map.size_hint().unwrap_or(0), S::default());
|
||||
|
||||
while let Some((key, value)) = map.next_entry()? {
|
||||
values.insert(key, value);
|
||||
}
|
||||
|
||||
Ok(values)
|
||||
}
|
||||
}
|
||||
|
||||
/// Requires crate feature `"serde"` or `"serde-1"`
|
||||
impl<'de, K, V, S> Deserialize<'de> for IndexMap<K, V, S>
|
||||
where
|
||||
K: Deserialize<'de> + Eq + Hash,
|
||||
V: Deserialize<'de>,
|
||||
S: Default + BuildHasher,
|
||||
{
|
||||
fn deserialize<D>(deserializer: D) -> Result<Self, D::Error>
|
||||
where
|
||||
D: Deserializer<'de>,
|
||||
{
|
||||
deserializer.deserialize_map(IndexMapVisitor(PhantomData))
|
||||
}
|
||||
}
|
||||
|
||||
impl<'de, K, V, S, E> IntoDeserializer<'de, E> for IndexMap<K, V, S>
|
||||
where
|
||||
K: IntoDeserializer<'de, E> + Eq + Hash,
|
||||
V: IntoDeserializer<'de, E>,
|
||||
S: BuildHasher,
|
||||
E: Error,
|
||||
{
|
||||
type Deserializer = MapDeserializer<'de, <Self as IntoIterator>::IntoIter, E>;
|
||||
|
||||
fn into_deserializer(self) -> Self::Deserializer {
|
||||
MapDeserializer::new(self.into_iter())
|
||||
}
|
||||
}
|
||||
|
||||
use crate::IndexSet;
|
||||
|
||||
/// Requires crate feature `"serde"` or `"serde-1"`
|
||||
impl<T, S> Serialize for IndexSet<T, S>
|
||||
where
|
||||
T: Serialize + Hash + Eq,
|
||||
S: BuildHasher,
|
||||
{
|
||||
fn serialize<Se>(&self, serializer: Se) -> Result<Se::Ok, Se::Error>
|
||||
where
|
||||
Se: Serializer,
|
||||
{
|
||||
serializer.collect_seq(self)
|
||||
}
|
||||
}
|
||||
|
||||
struct IndexSetVisitor<T, S>(PhantomData<(T, S)>);
|
||||
|
||||
impl<'de, T, S> Visitor<'de> for IndexSetVisitor<T, S>
|
||||
where
|
||||
T: Deserialize<'de> + Eq + Hash,
|
||||
S: Default + BuildHasher,
|
||||
{
|
||||
type Value = IndexSet<T, S>;
|
||||
|
||||
fn expecting(&self, formatter: &mut Formatter<'_>) -> fmt::Result {
|
||||
write!(formatter, "a set")
|
||||
}
|
||||
|
||||
fn visit_seq<A>(self, mut seq: A) -> Result<Self::Value, A::Error>
|
||||
where
|
||||
A: SeqAccess<'de>,
|
||||
{
|
||||
let mut values =
|
||||
IndexSet::with_capacity_and_hasher(seq.size_hint().unwrap_or(0), S::default());
|
||||
|
||||
while let Some(value) = seq.next_element()? {
|
||||
values.insert(value);
|
||||
}
|
||||
|
||||
Ok(values)
|
||||
}
|
||||
}
|
||||
|
||||
/// Requires crate feature `"serde"` or `"serde-1"`
|
||||
impl<'de, T, S> Deserialize<'de> for IndexSet<T, S>
|
||||
where
|
||||
T: Deserialize<'de> + Eq + Hash,
|
||||
S: Default + BuildHasher,
|
||||
{
|
||||
fn deserialize<D>(deserializer: D) -> Result<Self, D::Error>
|
||||
where
|
||||
D: Deserializer<'de>,
|
||||
{
|
||||
deserializer.deserialize_seq(IndexSetVisitor(PhantomData))
|
||||
}
|
||||
}
|
||||
|
||||
impl<'de, T, S, E> IntoDeserializer<'de, E> for IndexSet<T, S>
|
||||
where
|
||||
T: IntoDeserializer<'de, E> + Eq + Hash,
|
||||
S: BuildHasher,
|
||||
E: Error,
|
||||
{
|
||||
type Deserializer = SeqDeserializer<<Self as IntoIterator>::IntoIter, E>;
|
||||
|
||||
fn into_deserializer(self) -> Self::Deserializer {
|
||||
SeqDeserializer::new(self.into_iter())
|
||||
}
|
||||
}
|
||||
112
clamav/libclamav_rust/.cargo/vendor/indexmap/src/serde_seq.rs
vendored
Normal file
112
clamav/libclamav_rust/.cargo/vendor/indexmap/src/serde_seq.rs
vendored
Normal file
@@ -0,0 +1,112 @@
|
||||
//! Functions to serialize and deserialize an `IndexMap` as an ordered sequence.
|
||||
//!
|
||||
//! The default `serde` implementation serializes `IndexMap` as a normal map,
|
||||
//! but there is no guarantee that serialization formats will preserve the order
|
||||
//! of the key-value pairs. This module serializes `IndexMap` as a sequence of
|
||||
//! `(key, value)` elements instead, in order.
|
||||
//!
|
||||
//! This module may be used in a field attribute for derived implementations:
|
||||
//!
|
||||
//! ```
|
||||
//! # use indexmap::IndexMap;
|
||||
//! # use serde_derive::{Deserialize, Serialize};
|
||||
//! #[derive(Deserialize, Serialize)]
|
||||
//! struct Data {
|
||||
//! #[serde(with = "indexmap::serde_seq")]
|
||||
//! map: IndexMap<i32, u64>,
|
||||
//! // ...
|
||||
//! }
|
||||
//! ```
|
||||
//!
|
||||
//! Requires crate feature `"serde"` or `"serde-1"`
|
||||
|
||||
use serde::de::{Deserialize, Deserializer, SeqAccess, Visitor};
|
||||
use serde::ser::{Serialize, Serializer};
|
||||
|
||||
use core::fmt::{self, Formatter};
|
||||
use core::hash::{BuildHasher, Hash};
|
||||
use core::marker::PhantomData;
|
||||
|
||||
use crate::IndexMap;
|
||||
|
||||
/// Serializes an `IndexMap` as an ordered sequence.
|
||||
///
|
||||
/// This function may be used in a field attribute for deriving `Serialize`:
|
||||
///
|
||||
/// ```
|
||||
/// # use indexmap::IndexMap;
|
||||
/// # use serde_derive::Serialize;
|
||||
/// #[derive(Serialize)]
|
||||
/// struct Data {
|
||||
/// #[serde(serialize_with = "indexmap::serde_seq::serialize")]
|
||||
/// map: IndexMap<i32, u64>,
|
||||
/// // ...
|
||||
/// }
|
||||
/// ```
|
||||
///
|
||||
/// Requires crate feature `"serde"` or `"serde-1"`
|
||||
pub fn serialize<K, V, S, T>(map: &IndexMap<K, V, S>, serializer: T) -> Result<T::Ok, T::Error>
|
||||
where
|
||||
K: Serialize + Hash + Eq,
|
||||
V: Serialize,
|
||||
S: BuildHasher,
|
||||
T: Serializer,
|
||||
{
|
||||
serializer.collect_seq(map)
|
||||
}
|
||||
|
||||
/// Visitor to deserialize a *sequenced* `IndexMap`
|
||||
struct SeqVisitor<K, V, S>(PhantomData<(K, V, S)>);
|
||||
|
||||
impl<'de, K, V, S> Visitor<'de> for SeqVisitor<K, V, S>
|
||||
where
|
||||
K: Deserialize<'de> + Eq + Hash,
|
||||
V: Deserialize<'de>,
|
||||
S: Default + BuildHasher,
|
||||
{
|
||||
type Value = IndexMap<K, V, S>;
|
||||
|
||||
fn expecting(&self, formatter: &mut Formatter<'_>) -> fmt::Result {
|
||||
write!(formatter, "a sequenced map")
|
||||
}
|
||||
|
||||
fn visit_seq<A>(self, mut seq: A) -> Result<Self::Value, A::Error>
|
||||
where
|
||||
A: SeqAccess<'de>,
|
||||
{
|
||||
let capacity = seq.size_hint().unwrap_or(0);
|
||||
let mut map = IndexMap::with_capacity_and_hasher(capacity, S::default());
|
||||
|
||||
while let Some((key, value)) = seq.next_element()? {
|
||||
map.insert(key, value);
|
||||
}
|
||||
|
||||
Ok(map)
|
||||
}
|
||||
}
|
||||
|
||||
/// Deserializes an `IndexMap` from an ordered sequence.
|
||||
///
|
||||
/// This function may be used in a field attribute for deriving `Deserialize`:
|
||||
///
|
||||
/// ```
|
||||
/// # use indexmap::IndexMap;
|
||||
/// # use serde_derive::Deserialize;
|
||||
/// #[derive(Deserialize)]
|
||||
/// struct Data {
|
||||
/// #[serde(deserialize_with = "indexmap::serde_seq::deserialize")]
|
||||
/// map: IndexMap<i32, u64>,
|
||||
/// // ...
|
||||
/// }
|
||||
/// ```
|
||||
///
|
||||
/// Requires crate feature `"serde"` or `"serde-1"`
|
||||
pub fn deserialize<'de, D, K, V, S>(deserializer: D) -> Result<IndexMap<K, V, S>, D::Error>
|
||||
where
|
||||
D: Deserializer<'de>,
|
||||
K: Deserialize<'de> + Eq + Hash,
|
||||
V: Deserialize<'de>,
|
||||
S: Default + BuildHasher,
|
||||
{
|
||||
deserializer.deserialize_seq(SeqVisitor(PhantomData))
|
||||
}
|
||||
1912
clamav/libclamav_rust/.cargo/vendor/indexmap/src/set.rs
vendored
Normal file
1912
clamav/libclamav_rust/.cargo/vendor/indexmap/src/set.rs
vendored
Normal file
File diff suppressed because it is too large
Load Diff
31
clamav/libclamav_rust/.cargo/vendor/indexmap/src/util.rs
vendored
Normal file
31
clamav/libclamav_rust/.cargo/vendor/indexmap/src/util.rs
vendored
Normal file
@@ -0,0 +1,31 @@
|
||||
use core::ops::{Bound, Range, RangeBounds};
|
||||
|
||||
pub(crate) fn third<A, B, C>(t: (A, B, C)) -> C {
|
||||
t.2
|
||||
}
|
||||
|
||||
pub(crate) fn simplify_range<R>(range: R, len: usize) -> Range<usize>
|
||||
where
|
||||
R: RangeBounds<usize>,
|
||||
{
|
||||
let start = match range.start_bound() {
|
||||
Bound::Unbounded => 0,
|
||||
Bound::Included(&i) if i <= len => i,
|
||||
Bound::Excluded(&i) if i < len => i + 1,
|
||||
bound => panic!("range start {:?} should be <= length {}", bound, len),
|
||||
};
|
||||
let end = match range.end_bound() {
|
||||
Bound::Unbounded => len,
|
||||
Bound::Excluded(&i) if i <= len => i,
|
||||
Bound::Included(&i) if i < len => i + 1,
|
||||
bound => panic!("range end {:?} should be <= length {}", bound, len),
|
||||
};
|
||||
if start > end {
|
||||
panic!(
|
||||
"range start {:?} should be <= range end {:?}",
|
||||
range.start_bound(),
|
||||
range.end_bound()
|
||||
);
|
||||
}
|
||||
start..end
|
||||
}
|
||||
Reference in New Issue
Block a user