更新libclamav库1.0.0版本
This commit is contained in:
1
clamav/libclamav_rust/.cargo/vendor/indexmap/.cargo-checksum.json
vendored
Normal file
1
clamav/libclamav_rust/.cargo/vendor/indexmap/.cargo-checksum.json
vendored
Normal file
@@ -0,0 +1 @@
|
||||
{"files":{"Cargo.toml":"c77d3a8863367bbd5e9a6da5d8f100515da5e5a1441c293cbc09ffd947b7302b","LICENSE-APACHE":"a60eea817514531668d7e00765731449fe14d059d3249e0bc93b36de45f759f2","LICENSE-MIT":"ecc269ef87fd38a1d98e30bfac9ba964a9dbd9315c3770fed98d4d7cb5882055","README.md":"f8b02aa7c20fc0f5bc13de9e9e78899ec8cdbc16c2db880a1d0bc14c25b07542","RELEASES.md":"38d29c78198505ec88702c1ba723d087a775fcda6559da1842b6f17a9cdf6a71","benches/bench.rs":"3b2900abbc9e8a60af78b0395222ee75e86bc68519a0f38477387d1572eed397","benches/faststring.rs":"5fdd6cdb19d0557ed58f241e809a240cf8939d9e5b87a72d5f127f81ab98380b","build.rs":"558b4d0b9e9b3a44f7e1a2b69f7a7567ea721cd45cb54f4e458e850bf702f35c","src/arbitrary.rs":"bb8bda10f686abe57eef1446d3fc3fc6fb251f95629b28c20e620a4838c43db8","src/equivalent.rs":"2e6ae24ef09a09b917f4e2b0f6288f901878e42f5080f61b1bd1afdcc90aba87","src/lib.rs":"ea2cbe4f6cc2c4a75f42c9fc936503e6bee0f136c60f6811a2a9907ed8886443","src/macros.rs":"80c22f630e7f81e6fa663ca4c9e50cf5f332c8905d72d1338bd16f24eb353c2a","src/map.rs":"2e9cbfa240865cfd6b6b972bdbdb39283e6302dd2d0d72b3c2bfce4414bf5729","src/map/core.rs":"8422cd774c5db7d83cdeb0c5836c10f29caa1bee8d95b0d674b01b32e7ce80d8","src/map/core/raw.rs":"4e5fac4fecccc352268351d8b1f82b345067b5c029bba7e6ab88e8f8bc799c6a","src/mutable_keys.rs":"a919065b59000286eb11c7d46f6896bf0a1d484c9dac5e61d80bb8990c9fbedb","src/rayon/map.rs":"1a508c7c95c5d56113b851f7ce140d62ad541f1c6129352a7ec62d5bea7af4a1","src/rayon/mod.rs":"019e9379ccab57a299ab5b5a2c0efc7561b77a715a5afe8f797c7e8330c6206c","src/rayon/set.rs":"ba00e88e90fb7ab803589f99f24b595d60309e541aae3d01fdde21bff3840194","src/rustc.rs":"fe7a348c5a10a66880cb6c737593fe79d3b6de40f44ba0d7b89204aa95e14a3a","src/serde.rs":"d45ec8fb9c02594ca6f2e9b20764778b2b4193a24a52f1e233160a33efc6e683","src/serde_seq.rs":"c54a52fa607b6ccddda1e76e829778ca304c49b5f434edc5e582a5386c35d662","src/set.rs":"0a57affb623fa6b28df18cc14841e4f076cbd1da5c809635d202f865640af1ee","src/util.rs":"ab712bce71b54cf2763e6010e64bb5944d1d59ce15e2f2beffa7ceed204d6a68","tests/equivalent_trait.rs":"efe9393069e3cfc893d2c9c0343679979578e437fdb98a10baefeced027ba310","tests/macros_full_path.rs":"c33c86d7341581fdd08e2e6375a4afca507fa603540c54a3b9e51c4cd011cd71","tests/quick.rs":"1addbc6cbcb1aae5b8bde0fb0e18197d947e8f13244e4ae7ebf97bdda00eafea","tests/tests.rs":"f6dbeeb0e2950402b0e66ac52bf74c9e4197d3c5d9c0dde64a7998a2ef74d327"},"package":"1885e79c1fc4b10f0e172c475f458b7f7b93061064d98c3293e98c5ba0c8b399"}
|
||||
107
clamav/libclamav_rust/.cargo/vendor/indexmap/Cargo.toml
vendored
Normal file
107
clamav/libclamav_rust/.cargo/vendor/indexmap/Cargo.toml
vendored
Normal file
@@ -0,0 +1,107 @@
|
||||
# THIS FILE IS AUTOMATICALLY GENERATED BY CARGO
|
||||
#
|
||||
# When uploading crates to the registry Cargo will automatically
|
||||
# "normalize" Cargo.toml files for maximal compatibility
|
||||
# with all versions of Cargo and also rewrite `path` dependencies
|
||||
# to registry (e.g., crates.io) dependencies.
|
||||
#
|
||||
# If you are reading this file be aware that the original Cargo.toml
|
||||
# will likely look very different (and much more reasonable).
|
||||
# See Cargo.toml.orig for the original contents.
|
||||
|
||||
[package]
|
||||
edition = "2021"
|
||||
rust-version = "1.56"
|
||||
name = "indexmap"
|
||||
version = "1.9.2"
|
||||
description = "A hash table with consistent order and fast iteration."
|
||||
documentation = "https://docs.rs/indexmap/"
|
||||
readme = "README.md"
|
||||
keywords = [
|
||||
"hashmap",
|
||||
"no_std",
|
||||
]
|
||||
categories = [
|
||||
"data-structures",
|
||||
"no-std",
|
||||
]
|
||||
license = "Apache-2.0 OR MIT"
|
||||
repository = "https://github.com/bluss/indexmap"
|
||||
|
||||
[package.metadata.release]
|
||||
no-dev-version = true
|
||||
tag-name = "{{version}}"
|
||||
|
||||
[package.metadata.docs.rs]
|
||||
features = [
|
||||
"arbitrary",
|
||||
"quickcheck",
|
||||
"serde-1",
|
||||
"rayon",
|
||||
]
|
||||
|
||||
[profile.bench]
|
||||
debug = true
|
||||
|
||||
[lib]
|
||||
bench = false
|
||||
|
||||
[dependencies.arbitrary]
|
||||
version = "1.0"
|
||||
optional = true
|
||||
default-features = false
|
||||
|
||||
[dependencies.hashbrown]
|
||||
version = "0.12"
|
||||
features = ["raw"]
|
||||
default-features = false
|
||||
|
||||
[dependencies.quickcheck]
|
||||
version = "1.0"
|
||||
optional = true
|
||||
default-features = false
|
||||
|
||||
[dependencies.rayon]
|
||||
version = "1.4.1"
|
||||
optional = true
|
||||
|
||||
[dependencies.rustc-rayon]
|
||||
version = "0.4"
|
||||
optional = true
|
||||
|
||||
[dependencies.serde]
|
||||
version = "1.0"
|
||||
optional = true
|
||||
default-features = false
|
||||
|
||||
[dev-dependencies.fnv]
|
||||
version = "1.0"
|
||||
|
||||
[dev-dependencies.fxhash]
|
||||
version = "0.2.1"
|
||||
|
||||
[dev-dependencies.itertools]
|
||||
version = "0.10"
|
||||
|
||||
[dev-dependencies.lazy_static]
|
||||
version = "1.3"
|
||||
|
||||
[dev-dependencies.quickcheck]
|
||||
version = "1.0"
|
||||
default-features = false
|
||||
|
||||
[dev-dependencies.rand]
|
||||
version = "0.8"
|
||||
features = ["small_rng"]
|
||||
|
||||
[dev-dependencies.serde_derive]
|
||||
version = "1.0"
|
||||
|
||||
[build-dependencies.autocfg]
|
||||
version = "1"
|
||||
|
||||
[features]
|
||||
serde-1 = ["serde"]
|
||||
std = []
|
||||
test_debug = []
|
||||
test_low_transition_point = []
|
||||
201
clamav/libclamav_rust/.cargo/vendor/indexmap/LICENSE-APACHE
vendored
Normal file
201
clamav/libclamav_rust/.cargo/vendor/indexmap/LICENSE-APACHE
vendored
Normal file
@@ -0,0 +1,201 @@
|
||||
Apache License
|
||||
Version 2.0, January 2004
|
||||
http://www.apache.org/licenses/
|
||||
|
||||
TERMS AND CONDITIONS FOR USE, REPRODUCTION, AND DISTRIBUTION
|
||||
|
||||
1. Definitions.
|
||||
|
||||
"License" shall mean the terms and conditions for use, reproduction,
|
||||
and distribution as defined by Sections 1 through 9 of this document.
|
||||
|
||||
"Licensor" shall mean the copyright owner or entity authorized by
|
||||
the copyright owner that is granting the License.
|
||||
|
||||
"Legal Entity" shall mean the union of the acting entity and all
|
||||
other entities that control, are controlled by, or are under common
|
||||
control with that entity. For the purposes of this definition,
|
||||
"control" means (i) the power, direct or indirect, to cause the
|
||||
direction or management of such entity, whether by contract or
|
||||
otherwise, or (ii) ownership of fifty percent (50%) or more of the
|
||||
outstanding shares, or (iii) beneficial ownership of such entity.
|
||||
|
||||
"You" (or "Your") shall mean an individual or Legal Entity
|
||||
exercising permissions granted by this License.
|
||||
|
||||
"Source" form shall mean the preferred form for making modifications,
|
||||
including but not limited to software source code, documentation
|
||||
source, and configuration files.
|
||||
|
||||
"Object" form shall mean any form resulting from mechanical
|
||||
transformation or translation of a Source form, including but
|
||||
not limited to compiled object code, generated documentation,
|
||||
and conversions to other media types.
|
||||
|
||||
"Work" shall mean the work of authorship, whether in Source or
|
||||
Object form, made available under the License, as indicated by a
|
||||
copyright notice that is included in or attached to the work
|
||||
(an example is provided in the Appendix below).
|
||||
|
||||
"Derivative Works" shall mean any work, whether in Source or Object
|
||||
form, that is based on (or derived from) the Work and for which the
|
||||
editorial revisions, annotations, elaborations, or other modifications
|
||||
represent, as a whole, an original work of authorship. For the purposes
|
||||
of this License, Derivative Works shall not include works that remain
|
||||
separable from, or merely link (or bind by name) to the interfaces of,
|
||||
the Work and Derivative Works thereof.
|
||||
|
||||
"Contribution" shall mean any work of authorship, including
|
||||
the original version of the Work and any modifications or additions
|
||||
to that Work or Derivative Works thereof, that is intentionally
|
||||
submitted to Licensor for inclusion in the Work by the copyright owner
|
||||
or by an individual or Legal Entity authorized to submit on behalf of
|
||||
the copyright owner. For the purposes of this definition, "submitted"
|
||||
means any form of electronic, verbal, or written communication sent
|
||||
to the Licensor or its representatives, including but not limited to
|
||||
communication on electronic mailing lists, source code control systems,
|
||||
and issue tracking systems that are managed by, or on behalf of, the
|
||||
Licensor for the purpose of discussing and improving the Work, but
|
||||
excluding communication that is conspicuously marked or otherwise
|
||||
designated in writing by the copyright owner as "Not a Contribution."
|
||||
|
||||
"Contributor" shall mean Licensor and any individual or Legal Entity
|
||||
on behalf of whom a Contribution has been received by Licensor and
|
||||
subsequently incorporated within the Work.
|
||||
|
||||
2. Grant of Copyright License. Subject to the terms and conditions of
|
||||
this License, each Contributor hereby grants to You a perpetual,
|
||||
worldwide, non-exclusive, no-charge, royalty-free, irrevocable
|
||||
copyright license to reproduce, prepare Derivative Works of,
|
||||
publicly display, publicly perform, sublicense, and distribute the
|
||||
Work and such Derivative Works in Source or Object form.
|
||||
|
||||
3. Grant of Patent License. Subject to the terms and conditions of
|
||||
this License, each Contributor hereby grants to You a perpetual,
|
||||
worldwide, non-exclusive, no-charge, royalty-free, irrevocable
|
||||
(except as stated in this section) patent license to make, have made,
|
||||
use, offer to sell, sell, import, and otherwise transfer the Work,
|
||||
where such license applies only to those patent claims licensable
|
||||
by such Contributor that are necessarily infringed by their
|
||||
Contribution(s) alone or by combination of their Contribution(s)
|
||||
with the Work to which such Contribution(s) was submitted. If You
|
||||
institute patent litigation against any entity (including a
|
||||
cross-claim or counterclaim in a lawsuit) alleging that the Work
|
||||
or a Contribution incorporated within the Work constitutes direct
|
||||
or contributory patent infringement, then any patent licenses
|
||||
granted to You under this License for that Work shall terminate
|
||||
as of the date such litigation is filed.
|
||||
|
||||
4. Redistribution. You may reproduce and distribute copies of the
|
||||
Work or Derivative Works thereof in any medium, with or without
|
||||
modifications, and in Source or Object form, provided that You
|
||||
meet the following conditions:
|
||||
|
||||
(a) You must give any other recipients of the Work or
|
||||
Derivative Works a copy of this License; and
|
||||
|
||||
(b) You must cause any modified files to carry prominent notices
|
||||
stating that You changed the files; and
|
||||
|
||||
(c) You must retain, in the Source form of any Derivative Works
|
||||
that You distribute, all copyright, patent, trademark, and
|
||||
attribution notices from the Source form of the Work,
|
||||
excluding those notices that do not pertain to any part of
|
||||
the Derivative Works; and
|
||||
|
||||
(d) If the Work includes a "NOTICE" text file as part of its
|
||||
distribution, then any Derivative Works that You distribute must
|
||||
include a readable copy of the attribution notices contained
|
||||
within such NOTICE file, excluding those notices that do not
|
||||
pertain to any part of the Derivative Works, in at least one
|
||||
of the following places: within a NOTICE text file distributed
|
||||
as part of the Derivative Works; within the Source form or
|
||||
documentation, if provided along with the Derivative Works; or,
|
||||
within a display generated by the Derivative Works, if and
|
||||
wherever such third-party notices normally appear. The contents
|
||||
of the NOTICE file are for informational purposes only and
|
||||
do not modify the License. You may add Your own attribution
|
||||
notices within Derivative Works that You distribute, alongside
|
||||
or as an addendum to the NOTICE text from the Work, provided
|
||||
that such additional attribution notices cannot be construed
|
||||
as modifying the License.
|
||||
|
||||
You may add Your own copyright statement to Your modifications and
|
||||
may provide additional or different license terms and conditions
|
||||
for use, reproduction, or distribution of Your modifications, or
|
||||
for any such Derivative Works as a whole, provided Your use,
|
||||
reproduction, and distribution of the Work otherwise complies with
|
||||
the conditions stated in this License.
|
||||
|
||||
5. Submission of Contributions. Unless You explicitly state otherwise,
|
||||
any Contribution intentionally submitted for inclusion in the Work
|
||||
by You to the Licensor shall be under the terms and conditions of
|
||||
this License, without any additional terms or conditions.
|
||||
Notwithstanding the above, nothing herein shall supersede or modify
|
||||
the terms of any separate license agreement you may have executed
|
||||
with Licensor regarding such Contributions.
|
||||
|
||||
6. Trademarks. This License does not grant permission to use the trade
|
||||
names, trademarks, service marks, or product names of the Licensor,
|
||||
except as required for reasonable and customary use in describing the
|
||||
origin of the Work and reproducing the content of the NOTICE file.
|
||||
|
||||
7. Disclaimer of Warranty. Unless required by applicable law or
|
||||
agreed to in writing, Licensor provides the Work (and each
|
||||
Contributor provides its Contributions) on an "AS IS" BASIS,
|
||||
WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or
|
||||
implied, including, without limitation, any warranties or conditions
|
||||
of TITLE, NON-INFRINGEMENT, MERCHANTABILITY, or FITNESS FOR A
|
||||
PARTICULAR PURPOSE. You are solely responsible for determining the
|
||||
appropriateness of using or redistributing the Work and assume any
|
||||
risks associated with Your exercise of permissions under this License.
|
||||
|
||||
8. Limitation of Liability. In no event and under no legal theory,
|
||||
whether in tort (including negligence), contract, or otherwise,
|
||||
unless required by applicable law (such as deliberate and grossly
|
||||
negligent acts) or agreed to in writing, shall any Contributor be
|
||||
liable to You for damages, including any direct, indirect, special,
|
||||
incidental, or consequential damages of any character arising as a
|
||||
result of this License or out of the use or inability to use the
|
||||
Work (including but not limited to damages for loss of goodwill,
|
||||
work stoppage, computer failure or malfunction, or any and all
|
||||
other commercial damages or losses), even if such Contributor
|
||||
has been advised of the possibility of such damages.
|
||||
|
||||
9. Accepting Warranty or Additional Liability. While redistributing
|
||||
the Work or Derivative Works thereof, You may choose to offer,
|
||||
and charge a fee for, acceptance of support, warranty, indemnity,
|
||||
or other liability obligations and/or rights consistent with this
|
||||
License. However, in accepting such obligations, You may act only
|
||||
on Your own behalf and on Your sole responsibility, not on behalf
|
||||
of any other Contributor, and only if You agree to indemnify,
|
||||
defend, and hold each Contributor harmless for any liability
|
||||
incurred by, or claims asserted against, such Contributor by reason
|
||||
of your accepting any such warranty or additional liability.
|
||||
|
||||
END OF TERMS AND CONDITIONS
|
||||
|
||||
APPENDIX: How to apply the Apache License to your work.
|
||||
|
||||
To apply the Apache License to your work, attach the following
|
||||
boilerplate notice, with the fields enclosed by brackets "[]"
|
||||
replaced with your own identifying information. (Don't include
|
||||
the brackets!) The text should be enclosed in the appropriate
|
||||
comment syntax for the file format. We also recommend that a
|
||||
file or class name and description of purpose be included on the
|
||||
same "printed page" as the copyright notice for easier
|
||||
identification within third-party archives.
|
||||
|
||||
Copyright [yyyy] [name of copyright owner]
|
||||
|
||||
Licensed under the Apache License, Version 2.0 (the "License");
|
||||
you may not use this file except in compliance with the License.
|
||||
You may obtain a copy of the License at
|
||||
|
||||
http://www.apache.org/licenses/LICENSE-2.0
|
||||
|
||||
Unless required by applicable law or agreed to in writing, software
|
||||
distributed under the License is distributed on an "AS IS" BASIS,
|
||||
WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
|
||||
See the License for the specific language governing permissions and
|
||||
limitations under the License.
|
||||
25
clamav/libclamav_rust/.cargo/vendor/indexmap/LICENSE-MIT
vendored
Normal file
25
clamav/libclamav_rust/.cargo/vendor/indexmap/LICENSE-MIT
vendored
Normal file
@@ -0,0 +1,25 @@
|
||||
Copyright (c) 2016--2017
|
||||
|
||||
Permission is hereby granted, free of charge, to any
|
||||
person obtaining a copy of this software and associated
|
||||
documentation files (the "Software"), to deal in the
|
||||
Software without restriction, including without
|
||||
limitation the rights to use, copy, modify, merge,
|
||||
publish, distribute, sublicense, and/or sell copies of
|
||||
the Software, and to permit persons to whom the Software
|
||||
is furnished to do so, subject to the following
|
||||
conditions:
|
||||
|
||||
The above copyright notice and this permission notice
|
||||
shall be included in all copies or substantial portions
|
||||
of the Software.
|
||||
|
||||
THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF
|
||||
ANY KIND, EXPRESS OR IMPLIED, INCLUDING BUT NOT LIMITED
|
||||
TO THE WARRANTIES OF MERCHANTABILITY, FITNESS FOR A
|
||||
PARTICULAR PURPOSE AND NONINFRINGEMENT. IN NO EVENT
|
||||
SHALL THE AUTHORS OR COPYRIGHT HOLDERS BE LIABLE FOR ANY
|
||||
CLAIM, DAMAGES OR OTHER LIABILITY, WHETHER IN AN ACTION
|
||||
OF CONTRACT, TORT OR OTHERWISE, ARISING FROM, OUT OF OR
|
||||
IN CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER
|
||||
DEALINGS IN THE SOFTWARE.
|
||||
55
clamav/libclamav_rust/.cargo/vendor/indexmap/README.md
vendored
Normal file
55
clamav/libclamav_rust/.cargo/vendor/indexmap/README.md
vendored
Normal file
@@ -0,0 +1,55 @@
|
||||
# indexmap
|
||||
|
||||
[](https://github.com/bluss/indexmap/actions)
|
||||
[](https://crates.io/crates/indexmap)
|
||||
[](https://docs.rs/indexmap)
|
||||
[](https://img.shields.io/badge/rust-1.56%2B-orange.svg)
|
||||
|
||||
A pure-Rust hash table which preserves (in a limited sense) insertion order.
|
||||
|
||||
This crate implements compact map and set data-structures,
|
||||
where the iteration order of the keys is independent from their hash or
|
||||
value. It preserves insertion order (except after removals), and it
|
||||
allows lookup of entries by either hash table key or numerical index.
|
||||
|
||||
Note: this crate was originally released under the name `ordermap`,
|
||||
but it was renamed to `indexmap` to better reflect its features.
|
||||
|
||||
# Background
|
||||
|
||||
This was inspired by Python 3.6's new dict implementation (which remembers
|
||||
the insertion order and is fast to iterate, and is compact in memory).
|
||||
|
||||
Some of those features were translated to Rust, and some were not. The result
|
||||
was indexmap, a hash table that has following properties:
|
||||
|
||||
- Order is **independent of hash function** and hash values of keys.
|
||||
- Fast to iterate.
|
||||
- Indexed in compact space.
|
||||
- Preserves insertion order **as long** as you don't call `.remove()`.
|
||||
- Uses hashbrown for the inner table, just like Rust's libstd `HashMap` does.
|
||||
|
||||
## Performance
|
||||
|
||||
`IndexMap` derives a couple of performance facts directly from how it is constructed,
|
||||
which is roughly:
|
||||
|
||||
> A raw hash table of key-value indices, and a vector of key-value pairs.
|
||||
|
||||
- Iteration is very fast since it is on the dense key-values.
|
||||
- Removal is fast since it moves memory areas only in the table,
|
||||
and uses a single swap in the vector.
|
||||
- Lookup is fast-ish because the initial 7-bit hash lookup uses SIMD, and indices are
|
||||
densely stored. Lookup also is slow-ish since the actual key-value pairs are stored
|
||||
separately. (Visible when cpu caches size is limiting.)
|
||||
|
||||
- In practice, `IndexMap` has been tested out as the hashmap in rustc in [PR45282] and
|
||||
the performance was roughly on par across the whole workload.
|
||||
- If you want the properties of `IndexMap`, or its strongest performance points
|
||||
fits your workload, it might be the best hash table implementation.
|
||||
|
||||
[PR45282]: https://github.com/rust-lang/rust/pull/45282
|
||||
|
||||
# Recent Changes
|
||||
|
||||
See [RELEASES.md](https://github.com/bluss/indexmap/blob/master/RELEASES.md).
|
||||
384
clamav/libclamav_rust/.cargo/vendor/indexmap/RELEASES.md
vendored
Normal file
384
clamav/libclamav_rust/.cargo/vendor/indexmap/RELEASES.md
vendored
Normal file
@@ -0,0 +1,384 @@
|
||||
- 1.9.2
|
||||
|
||||
- `IndexMap` and `IndexSet` both implement `arbitrary::Arbitrary<'_>` and
|
||||
`quickcheck::Arbitrary` if those optional dependency features are enabled.
|
||||
|
||||
- 1.9.1
|
||||
|
||||
- The MSRV now allows Rust 1.56.0 as well. However, currently `hashbrown`
|
||||
0.12.1 requires 1.56.1, so users on 1.56.0 should downgrade that to 0.12.0
|
||||
until there is a later published version relaxing its requirement.
|
||||
|
||||
- 1.9.0
|
||||
|
||||
- **MSRV**: Rust 1.56.1 or later is now required.
|
||||
|
||||
- The `hashbrown` dependency has been updated to version 0.12.
|
||||
|
||||
- `IterMut` and `ValuesMut` now implement `Debug`.
|
||||
|
||||
- The new `IndexMap::shrink_to` and `IndexSet::shrink_to` methods shrink
|
||||
the capacity with a lower bound.
|
||||
|
||||
- The new `IndexMap::move_index` and `IndexSet::move_index` methods change
|
||||
the position of an item from one index to another, shifting the items
|
||||
between to accommodate the move.
|
||||
|
||||
- 1.8.2
|
||||
|
||||
- Bump the `rustc-rayon` dependency, for compiler use only.
|
||||
|
||||
- 1.8.1
|
||||
|
||||
- The new `IndexSet::replace_full` will return the index of the item along
|
||||
with the replaced value, if any, by @zakcutner in PR [222].
|
||||
|
||||
[222]: https://github.com/bluss/indexmap/pull/222
|
||||
|
||||
- 1.8.0
|
||||
|
||||
- The new `IndexMap::into_keys` and `IndexMap::into_values` will consume
|
||||
the map into keys or values, respectively, matching Rust 1.54's `HashMap`
|
||||
methods, by @taiki-e in PR [195].
|
||||
|
||||
- More of the iterator types implement `Debug`, `ExactSizeIterator`, and
|
||||
`FusedIterator`, by @cuviper in PR [196].
|
||||
|
||||
- `IndexMap` and `IndexSet` now implement rayon's `ParallelDrainRange`,
|
||||
by @cuviper in PR [197].
|
||||
|
||||
- `IndexMap::with_hasher` and `IndexSet::with_hasher` are now `const`
|
||||
functions, allowing static maps and sets, by @mwillsey in PR [203].
|
||||
|
||||
- `IndexMap` and `IndexSet` now implement `From` for arrays, matching
|
||||
Rust 1.56's implementation for `HashMap`, by @rouge8 in PR [205].
|
||||
|
||||
- `IndexMap` and `IndexSet` now have methods `sort_unstable_keys`,
|
||||
`sort_unstable_by`, `sorted_unstable_by`, and `par_*` equivalents,
|
||||
which sort in-place without preserving the order of equal items, by
|
||||
@bhgomes in PR [211].
|
||||
|
||||
[195]: https://github.com/bluss/indexmap/pull/195
|
||||
[196]: https://github.com/bluss/indexmap/pull/196
|
||||
[197]: https://github.com/bluss/indexmap/pull/197
|
||||
[203]: https://github.com/bluss/indexmap/pull/203
|
||||
[205]: https://github.com/bluss/indexmap/pull/205
|
||||
[211]: https://github.com/bluss/indexmap/pull/211
|
||||
|
||||
- 1.7.0
|
||||
|
||||
- **MSRV**: Rust 1.49 or later is now required.
|
||||
|
||||
- The `hashbrown` dependency has been updated to version 0.11.
|
||||
|
||||
- 1.6.2
|
||||
|
||||
- Fixed to match `std` behavior, `OccupiedEntry::key` now references the
|
||||
existing key in the map instead of the lookup key, by @cuviper in PR [170].
|
||||
|
||||
- The new `Entry::or_insert_with_key` matches Rust 1.50's `Entry` method,
|
||||
passing `&K` to the callback to create a value, by @cuviper in PR [175].
|
||||
|
||||
[170]: https://github.com/bluss/indexmap/pull/170
|
||||
[175]: https://github.com/bluss/indexmap/pull/175
|
||||
|
||||
- 1.6.1
|
||||
|
||||
- The new `serde_seq` module implements `IndexMap` serialization as a
|
||||
sequence to ensure order is preserved, by @cuviper in PR [158].
|
||||
|
||||
- New methods on maps and sets work like the `Vec`/slice methods by the same name:
|
||||
`truncate`, `split_off`, `first`, `first_mut`, `last`, `last_mut`, and
|
||||
`swap_indices`, by @cuviper in PR [160].
|
||||
|
||||
[158]: https://github.com/bluss/indexmap/pull/158
|
||||
[160]: https://github.com/bluss/indexmap/pull/160
|
||||
|
||||
- 1.6.0
|
||||
|
||||
- **MSRV**: Rust 1.36 or later is now required.
|
||||
|
||||
- The `hashbrown` dependency has been updated to version 0.9.
|
||||
|
||||
- 1.5.2
|
||||
|
||||
- The new "std" feature will force the use of `std` for users that explicitly
|
||||
want the default `S = RandomState`, bypassing the autodetection added in 1.3.0,
|
||||
by @cuviper in PR [145].
|
||||
|
||||
[145]: https://github.com/bluss/indexmap/pull/145
|
||||
|
||||
- 1.5.1
|
||||
|
||||
- Values can now be indexed by their `usize` position by @cuviper in PR [132].
|
||||
|
||||
- Some of the generic bounds have been relaxed to match `std` by @cuviper in PR [141].
|
||||
|
||||
- `drain` now accepts any `R: RangeBounds<usize>` by @cuviper in PR [142].
|
||||
|
||||
[132]: https://github.com/bluss/indexmap/pull/132
|
||||
[141]: https://github.com/bluss/indexmap/pull/141
|
||||
[142]: https://github.com/bluss/indexmap/pull/142
|
||||
|
||||
- 1.5.0
|
||||
|
||||
- **MSRV**: Rust 1.32 or later is now required.
|
||||
|
||||
- The inner hash table is now based on `hashbrown` by @cuviper in PR [131].
|
||||
This also completes the method `reserve` and adds `shrink_to_fit`.
|
||||
|
||||
- Add new methods `get_key_value`, `remove_entry`, `swap_remove_entry`,
|
||||
and `shift_remove_entry`, by @cuviper in PR [136]
|
||||
|
||||
- `Clone::clone_from` reuses allocations by @cuviper in PR [125]
|
||||
|
||||
- Add new method `reverse` by @linclelinkpart5 in PR [128]
|
||||
|
||||
[125]: https://github.com/bluss/indexmap/pull/125
|
||||
[128]: https://github.com/bluss/indexmap/pull/128
|
||||
[131]: https://github.com/bluss/indexmap/pull/131
|
||||
[136]: https://github.com/bluss/indexmap/pull/136
|
||||
|
||||
- 1.4.0
|
||||
|
||||
- Add new method `get_index_of` by @Thermatrix in PR [115] and [120]
|
||||
|
||||
- Fix build script rebuild-if-changed configuration to use "build.rs";
|
||||
fixes issue [123]. Fix by @cuviper.
|
||||
|
||||
- Dev-dependencies (rand and quickcheck) have been updated. The crate's tests
|
||||
now run using Rust 1.32 or later (MSRV for building the crate has not changed).
|
||||
by @kjeremy and @bluss
|
||||
|
||||
[123]: https://github.com/bluss/indexmap/issues/123
|
||||
[115]: https://github.com/bluss/indexmap/pull/115
|
||||
[120]: https://github.com/bluss/indexmap/pull/120
|
||||
|
||||
- 1.3.2
|
||||
|
||||
- Maintenance update to regenerate the published `Cargo.toml`.
|
||||
|
||||
- 1.3.1
|
||||
|
||||
- Maintenance update for formatting and `autocfg` 1.0.
|
||||
|
||||
- 1.3.0
|
||||
|
||||
- The deprecation messages in the previous version have been removed.
|
||||
(The methods have not otherwise changed.) Docs for removal methods have been
|
||||
improved.
|
||||
- From Rust 1.36, this crate supports being built **without std**, requiring
|
||||
`alloc` instead. This is enabled automatically when it is detected that
|
||||
`std` is not available. There is no crate feature to enable/disable to
|
||||
trigger this. The new build-dep `autocfg` enables this.
|
||||
|
||||
- 1.2.0
|
||||
|
||||
- Plain `.remove()` now has a deprecation message, it informs the user
|
||||
about picking one of the removal functions `swap_remove` and `shift_remove`
|
||||
which have different performance and order semantics.
|
||||
Plain `.remove()` will not be removed, the warning message and method
|
||||
will remain until further.
|
||||
|
||||
- Add new method `shift_remove` for order preserving removal on the map,
|
||||
and `shift_take` for the corresponding operation on the set.
|
||||
|
||||
- Add methods `swap_remove`, `swap_remove_entry` to `Entry`.
|
||||
|
||||
- Fix indexset/indexmap to support full paths, like `indexmap::indexmap!()`
|
||||
|
||||
- Internal improvements: fix warnings, deprecations and style lints
|
||||
|
||||
- 1.1.0
|
||||
|
||||
- Added optional feature `"rayon"` that adds parallel iterator support
|
||||
to `IndexMap` and `IndexSet` using Rayon. This includes all the regular
|
||||
iterators in parallel versions, and parallel sort.
|
||||
|
||||
- Implemented `Clone` for `map::{Iter, Keys, Values}` and
|
||||
`set::{Difference, Intersection, Iter, SymmetricDifference, Union}`
|
||||
|
||||
- Implemented `Debug` for `map::{Entry, IntoIter, Iter, Keys, Values}` and
|
||||
`set::{Difference, Intersection, IntoIter, Iter, SymmetricDifference, Union}`
|
||||
|
||||
- Serde trait `IntoDeserializer` are implemented for `IndexMap` and `IndexSet`.
|
||||
|
||||
- Minimum Rust version requirement increased to Rust 1.30 for development builds.
|
||||
|
||||
- 1.0.2
|
||||
|
||||
- The new methods `IndexMap::insert_full` and `IndexSet::insert_full` are
|
||||
both like `insert` with the index included in the return value.
|
||||
|
||||
- The new method `Entry::and_modify` can be used to modify occupied
|
||||
entries, matching the new methods of `std` maps in Rust 1.26.
|
||||
|
||||
- The new method `Entry::or_default` inserts a default value in unoccupied
|
||||
entries, matching the new methods of `std` maps in Rust 1.28.
|
||||
|
||||
- 1.0.1
|
||||
|
||||
- Document Rust version policy for the crate (see rustdoc)
|
||||
|
||||
- 1.0.0
|
||||
|
||||
- This is the 1.0 release for `indexmap`! (the crate and datastructure
|
||||
formerly known as “ordermap”)
|
||||
- `OccupiedEntry::insert` changed its signature, to use `&mut self` for
|
||||
the method receiver, matching the equivalent method for a standard
|
||||
`HashMap`. Thanks to @dtolnay for finding this bug.
|
||||
- The deprecated old names from ordermap were removed: `OrderMap`,
|
||||
`OrderSet`, `ordermap!{}`, `orderset!{}`. Use the new `IndexMap`
|
||||
etc names instead.
|
||||
|
||||
- 0.4.1
|
||||
|
||||
- Renamed crate to `indexmap`; the `ordermap` crate is now deprecated
|
||||
and the types `OrderMap/Set` now have a deprecation notice.
|
||||
|
||||
- 0.4.0
|
||||
|
||||
- This is the last release series for this `ordermap` under that name,
|
||||
because the crate is **going to be renamed** to `indexmap` (with types
|
||||
`IndexMap`, `IndexSet`) and no change in functionality!
|
||||
- The map and its associated structs moved into the `map` submodule of the
|
||||
crate, so that the map and set are symmetric
|
||||
|
||||
+ The iterators, `Entry` and other structs are now under `ordermap::map::`
|
||||
|
||||
- Internally refactored `OrderMap<K, V, S>` so that all the main algorithms
|
||||
(insertion, lookup, removal etc) that don't use the `S` parameter (the
|
||||
hasher) are compiled without depending on `S`, which reduces generics bloat.
|
||||
|
||||
- `Entry<K, V>` no longer has a type parameter `S`, which is just like
|
||||
the standard `HashMap`'s entry.
|
||||
|
||||
- Minimum Rust version requirement increased to Rust 1.18
|
||||
|
||||
- 0.3.5
|
||||
|
||||
- Documentation improvements
|
||||
|
||||
- 0.3.4
|
||||
|
||||
- The `.retain()` methods for `OrderMap` and `OrderSet` now
|
||||
traverse the elements in order, and the retained elements **keep their order**
|
||||
- Added new methods `.sort_by()`, `.sort_keys()` to `OrderMap` and
|
||||
`.sort_by()`, `.sort()` to `OrderSet`. These methods allow you to
|
||||
sort the maps in place efficiently.
|
||||
|
||||
- 0.3.3
|
||||
|
||||
- Document insertion behaviour better by @lucab
|
||||
- Updated dependences (no feature changes) by @ignatenkobrain
|
||||
|
||||
- 0.3.2
|
||||
|
||||
- Add `OrderSet` by @cuviper!
|
||||
- `OrderMap::drain` is now (too) a double ended iterator.
|
||||
|
||||
- 0.3.1
|
||||
|
||||
- In all ordermap iterators, forward the `collect` method to the underlying
|
||||
iterator as well.
|
||||
- Add crates.io categories.
|
||||
|
||||
- 0.3.0
|
||||
|
||||
- The methods `get_pair`, `get_pair_index` were both replaced by
|
||||
`get_full` (and the same for the mutable case).
|
||||
- Method `swap_remove_pair` replaced by `swap_remove_full`.
|
||||
- Add trait `MutableKeys` for opt-in mutable key access. Mutable key access
|
||||
is only possible through the methods of this extension trait.
|
||||
- Add new trait `Equivalent` for key equivalence. This extends the
|
||||
`Borrow` trait mechanism for `OrderMap::get` in a backwards compatible
|
||||
way, just some minor type inference related issues may become apparent.
|
||||
See [#10] for more information.
|
||||
- Implement `Extend<(&K, &V)>` by @xfix.
|
||||
|
||||
[#10]: https://github.com/bluss/ordermap/pull/10
|
||||
|
||||
- 0.2.13
|
||||
|
||||
- Fix deserialization to support custom hashers by @Techcable.
|
||||
- Add methods `.index()` on the entry types by @garro95.
|
||||
|
||||
- 0.2.12
|
||||
|
||||
- Add methods `.with_hasher()`, `.hasher()`.
|
||||
|
||||
- 0.2.11
|
||||
|
||||
- Support `ExactSizeIterator` for the iterators. By @Binero.
|
||||
- Use `Box<[Pos]>` internally, saving a word in the `OrderMap` struct.
|
||||
- Serde support, with crate feature `"serde-1"`. By @xfix.
|
||||
|
||||
- 0.2.10
|
||||
|
||||
- Add iterator `.drain(..)` by @stevej.
|
||||
|
||||
- 0.2.9
|
||||
|
||||
- Add method `.is_empty()` by @overvenus.
|
||||
- Implement `PartialEq, Eq` by @overvenus.
|
||||
- Add method `.sorted_by()`.
|
||||
|
||||
- 0.2.8
|
||||
|
||||
- Add iterators `.values()` and `.values_mut()`.
|
||||
- Fix compatibility with 32-bit platforms.
|
||||
|
||||
- 0.2.7
|
||||
|
||||
- Add `.retain()`.
|
||||
|
||||
- 0.2.6
|
||||
|
||||
- Add `OccupiedEntry::remove_entry` and other minor entry methods,
|
||||
so that it now has all the features of `HashMap`'s entries.
|
||||
|
||||
- 0.2.5
|
||||
|
||||
- Improved `.pop()` slightly.
|
||||
|
||||
- 0.2.4
|
||||
|
||||
- Improved performance of `.insert()` ([#3]) by @pczarn.
|
||||
|
||||
[#3]: https://github.com/bluss/ordermap/pull/3
|
||||
|
||||
- 0.2.3
|
||||
|
||||
- Generalize `Entry` for now, so that it works on hashmaps with non-default
|
||||
hasher. However, there's a lingering compat issue since libstd `HashMap`
|
||||
does not parameterize its entries by the hasher (`S` typarm).
|
||||
- Special case some iterator methods like `.nth()`.
|
||||
|
||||
- 0.2.2
|
||||
|
||||
- Disable the verbose `Debug` impl by default.
|
||||
|
||||
- 0.2.1
|
||||
|
||||
- Fix doc links and clarify docs.
|
||||
|
||||
- 0.2.0
|
||||
|
||||
- Add more `HashMap` methods & compat with its API.
|
||||
- Experimental support for `.entry()` (the simplest parts of the API).
|
||||
- Add `.reserve()` (placeholder impl).
|
||||
- Add `.remove()` as synonym for `.swap_remove()`.
|
||||
- Changed `.insert()` to swap value if the entry already exists, and
|
||||
return `Option`.
|
||||
- Experimental support as an *indexed* hash map! Added methods
|
||||
`.get_index()`, `.get_index_mut()`, `.swap_remove_index()`,
|
||||
`.get_pair_index()`, `.get_pair_index_mut()`.
|
||||
|
||||
- 0.1.2
|
||||
|
||||
- Implement the 32/32 split idea for `Pos` which improves cache utilization
|
||||
and lookup performance.
|
||||
|
||||
- 0.1.1
|
||||
|
||||
- Initial release.
|
||||
763
clamav/libclamav_rust/.cargo/vendor/indexmap/benches/bench.rs
vendored
Normal file
763
clamav/libclamav_rust/.cargo/vendor/indexmap/benches/bench.rs
vendored
Normal file
@@ -0,0 +1,763 @@
|
||||
#![feature(test)]
|
||||
|
||||
extern crate test;
|
||||
#[macro_use]
|
||||
extern crate lazy_static;
|
||||
|
||||
use fnv::FnvHasher;
|
||||
use std::hash::BuildHasherDefault;
|
||||
use std::hash::Hash;
|
||||
type FnvBuilder = BuildHasherDefault<FnvHasher>;
|
||||
|
||||
use test::black_box;
|
||||
use test::Bencher;
|
||||
|
||||
use indexmap::IndexMap;
|
||||
|
||||
use std::collections::HashMap;
|
||||
|
||||
use rand::rngs::SmallRng;
|
||||
use rand::seq::SliceRandom;
|
||||
use rand::SeedableRng;
|
||||
|
||||
/// Use a consistently seeded Rng for benchmark stability
|
||||
fn small_rng() -> SmallRng {
|
||||
let seed = u64::from_le_bytes(*b"indexmap");
|
||||
SmallRng::seed_from_u64(seed)
|
||||
}
|
||||
|
||||
#[bench]
|
||||
fn new_hashmap(b: &mut Bencher) {
|
||||
b.iter(|| HashMap::<String, String>::new());
|
||||
}
|
||||
|
||||
#[bench]
|
||||
fn new_indexmap(b: &mut Bencher) {
|
||||
b.iter(|| IndexMap::<String, String>::new());
|
||||
}
|
||||
|
||||
#[bench]
|
||||
fn with_capacity_10e5_hashmap(b: &mut Bencher) {
|
||||
b.iter(|| HashMap::<String, String>::with_capacity(10_000));
|
||||
}
|
||||
|
||||
#[bench]
|
||||
fn with_capacity_10e5_indexmap(b: &mut Bencher) {
|
||||
b.iter(|| IndexMap::<String, String>::with_capacity(10_000));
|
||||
}
|
||||
|
||||
#[bench]
|
||||
fn insert_hashmap_10_000(b: &mut Bencher) {
|
||||
let c = 10_000;
|
||||
b.iter(|| {
|
||||
let mut map = HashMap::with_capacity(c);
|
||||
for x in 0..c {
|
||||
map.insert(x, ());
|
||||
}
|
||||
map
|
||||
});
|
||||
}
|
||||
|
||||
#[bench]
|
||||
fn insert_indexmap_10_000(b: &mut Bencher) {
|
||||
let c = 10_000;
|
||||
b.iter(|| {
|
||||
let mut map = IndexMap::with_capacity(c);
|
||||
for x in 0..c {
|
||||
map.insert(x, ());
|
||||
}
|
||||
map
|
||||
});
|
||||
}
|
||||
|
||||
#[bench]
|
||||
fn insert_hashmap_string_10_000(b: &mut Bencher) {
|
||||
let c = 10_000;
|
||||
b.iter(|| {
|
||||
let mut map = HashMap::with_capacity(c);
|
||||
for x in 0..c {
|
||||
map.insert(x.to_string(), ());
|
||||
}
|
||||
map
|
||||
});
|
||||
}
|
||||
|
||||
#[bench]
|
||||
fn insert_indexmap_string_10_000(b: &mut Bencher) {
|
||||
let c = 10_000;
|
||||
b.iter(|| {
|
||||
let mut map = IndexMap::with_capacity(c);
|
||||
for x in 0..c {
|
||||
map.insert(x.to_string(), ());
|
||||
}
|
||||
map
|
||||
});
|
||||
}
|
||||
|
||||
#[bench]
|
||||
fn insert_hashmap_str_10_000(b: &mut Bencher) {
|
||||
let c = 10_000;
|
||||
let ss = Vec::from_iter((0..c).map(|x| x.to_string()));
|
||||
b.iter(|| {
|
||||
let mut map = HashMap::with_capacity(c);
|
||||
for key in &ss {
|
||||
map.insert(&key[..], ());
|
||||
}
|
||||
map
|
||||
});
|
||||
}
|
||||
|
||||
#[bench]
|
||||
fn insert_indexmap_str_10_000(b: &mut Bencher) {
|
||||
let c = 10_000;
|
||||
let ss = Vec::from_iter((0..c).map(|x| x.to_string()));
|
||||
b.iter(|| {
|
||||
let mut map = IndexMap::with_capacity(c);
|
||||
for key in &ss {
|
||||
map.insert(&key[..], ());
|
||||
}
|
||||
map
|
||||
});
|
||||
}
|
||||
|
||||
#[bench]
|
||||
fn insert_hashmap_int_bigvalue_10_000(b: &mut Bencher) {
|
||||
let c = 10_000;
|
||||
let value = [0u64; 10];
|
||||
b.iter(|| {
|
||||
let mut map = HashMap::with_capacity(c);
|
||||
for i in 0..c {
|
||||
map.insert(i, value);
|
||||
}
|
||||
map
|
||||
});
|
||||
}
|
||||
|
||||
#[bench]
|
||||
fn insert_indexmap_int_bigvalue_10_000(b: &mut Bencher) {
|
||||
let c = 10_000;
|
||||
let value = [0u64; 10];
|
||||
b.iter(|| {
|
||||
let mut map = IndexMap::with_capacity(c);
|
||||
for i in 0..c {
|
||||
map.insert(i, value);
|
||||
}
|
||||
map
|
||||
});
|
||||
}
|
||||
|
||||
#[bench]
|
||||
fn insert_hashmap_100_000(b: &mut Bencher) {
|
||||
let c = 100_000;
|
||||
b.iter(|| {
|
||||
let mut map = HashMap::with_capacity(c);
|
||||
for x in 0..c {
|
||||
map.insert(x, ());
|
||||
}
|
||||
map
|
||||
});
|
||||
}
|
||||
|
||||
#[bench]
|
||||
fn insert_indexmap_100_000(b: &mut Bencher) {
|
||||
let c = 100_000;
|
||||
b.iter(|| {
|
||||
let mut map = IndexMap::with_capacity(c);
|
||||
for x in 0..c {
|
||||
map.insert(x, ());
|
||||
}
|
||||
map
|
||||
});
|
||||
}
|
||||
|
||||
#[bench]
|
||||
fn insert_hashmap_150(b: &mut Bencher) {
|
||||
let c = 150;
|
||||
b.iter(|| {
|
||||
let mut map = HashMap::with_capacity(c);
|
||||
for x in 0..c {
|
||||
map.insert(x, ());
|
||||
}
|
||||
map
|
||||
});
|
||||
}
|
||||
|
||||
#[bench]
|
||||
fn insert_indexmap_150(b: &mut Bencher) {
|
||||
let c = 150;
|
||||
b.iter(|| {
|
||||
let mut map = IndexMap::with_capacity(c);
|
||||
for x in 0..c {
|
||||
map.insert(x, ());
|
||||
}
|
||||
map
|
||||
});
|
||||
}
|
||||
|
||||
#[bench]
|
||||
fn entry_hashmap_150(b: &mut Bencher) {
|
||||
let c = 150;
|
||||
b.iter(|| {
|
||||
let mut map = HashMap::with_capacity(c);
|
||||
for x in 0..c {
|
||||
map.entry(x).or_insert(());
|
||||
}
|
||||
map
|
||||
});
|
||||
}
|
||||
|
||||
#[bench]
|
||||
fn entry_indexmap_150(b: &mut Bencher) {
|
||||
let c = 150;
|
||||
b.iter(|| {
|
||||
let mut map = IndexMap::with_capacity(c);
|
||||
for x in 0..c {
|
||||
map.entry(x).or_insert(());
|
||||
}
|
||||
map
|
||||
});
|
||||
}
|
||||
|
||||
#[bench]
|
||||
fn iter_sum_hashmap_10_000(b: &mut Bencher) {
|
||||
let c = 10_000;
|
||||
let mut map = HashMap::with_capacity(c);
|
||||
let len = c - c / 10;
|
||||
for x in 0..len {
|
||||
map.insert(x, ());
|
||||
}
|
||||
assert_eq!(map.len(), len);
|
||||
b.iter(|| map.keys().sum::<usize>());
|
||||
}
|
||||
|
||||
#[bench]
|
||||
fn iter_sum_indexmap_10_000(b: &mut Bencher) {
|
||||
let c = 10_000;
|
||||
let mut map = IndexMap::with_capacity(c);
|
||||
let len = c - c / 10;
|
||||
for x in 0..len {
|
||||
map.insert(x, ());
|
||||
}
|
||||
assert_eq!(map.len(), len);
|
||||
b.iter(|| map.keys().sum::<usize>());
|
||||
}
|
||||
|
||||
#[bench]
|
||||
fn iter_black_box_hashmap_10_000(b: &mut Bencher) {
|
||||
let c = 10_000;
|
||||
let mut map = HashMap::with_capacity(c);
|
||||
let len = c - c / 10;
|
||||
for x in 0..len {
|
||||
map.insert(x, ());
|
||||
}
|
||||
assert_eq!(map.len(), len);
|
||||
b.iter(|| {
|
||||
for &key in map.keys() {
|
||||
black_box(key);
|
||||
}
|
||||
});
|
||||
}
|
||||
|
||||
#[bench]
|
||||
fn iter_black_box_indexmap_10_000(b: &mut Bencher) {
|
||||
let c = 10_000;
|
||||
let mut map = IndexMap::with_capacity(c);
|
||||
let len = c - c / 10;
|
||||
for x in 0..len {
|
||||
map.insert(x, ());
|
||||
}
|
||||
assert_eq!(map.len(), len);
|
||||
b.iter(|| {
|
||||
for &key in map.keys() {
|
||||
black_box(key);
|
||||
}
|
||||
});
|
||||
}
|
||||
|
||||
fn shuffled_keys<I>(iter: I) -> Vec<I::Item>
|
||||
where
|
||||
I: IntoIterator,
|
||||
{
|
||||
let mut v = Vec::from_iter(iter);
|
||||
let mut rng = small_rng();
|
||||
v.shuffle(&mut rng);
|
||||
v
|
||||
}
|
||||
|
||||
#[bench]
|
||||
fn lookup_hashmap_10_000_exist(b: &mut Bencher) {
|
||||
let c = 10_000;
|
||||
let mut map = HashMap::with_capacity(c);
|
||||
let keys = shuffled_keys(0..c);
|
||||
for &key in &keys {
|
||||
map.insert(key, 1);
|
||||
}
|
||||
b.iter(|| {
|
||||
let mut found = 0;
|
||||
for key in 5000..c {
|
||||
found += map.get(&key).is_some() as i32;
|
||||
}
|
||||
found
|
||||
});
|
||||
}
|
||||
|
||||
#[bench]
|
||||
fn lookup_hashmap_10_000_noexist(b: &mut Bencher) {
|
||||
let c = 10_000;
|
||||
let mut map = HashMap::with_capacity(c);
|
||||
let keys = shuffled_keys(0..c);
|
||||
for &key in &keys {
|
||||
map.insert(key, 1);
|
||||
}
|
||||
b.iter(|| {
|
||||
let mut found = 0;
|
||||
for key in c..15000 {
|
||||
found += map.get(&key).is_some() as i32;
|
||||
}
|
||||
found
|
||||
});
|
||||
}
|
||||
|
||||
#[bench]
|
||||
fn lookup_indexmap_10_000_exist(b: &mut Bencher) {
|
||||
let c = 10_000;
|
||||
let mut map = IndexMap::with_capacity(c);
|
||||
let keys = shuffled_keys(0..c);
|
||||
for &key in &keys {
|
||||
map.insert(key, 1);
|
||||
}
|
||||
b.iter(|| {
|
||||
let mut found = 0;
|
||||
for key in 5000..c {
|
||||
found += map.get(&key).is_some() as i32;
|
||||
}
|
||||
found
|
||||
});
|
||||
}
|
||||
|
||||
#[bench]
|
||||
fn lookup_indexmap_10_000_noexist(b: &mut Bencher) {
|
||||
let c = 10_000;
|
||||
let mut map = IndexMap::with_capacity(c);
|
||||
let keys = shuffled_keys(0..c);
|
||||
for &key in &keys {
|
||||
map.insert(key, 1);
|
||||
}
|
||||
b.iter(|| {
|
||||
let mut found = 0;
|
||||
for key in c..15000 {
|
||||
found += map.get(&key).is_some() as i32;
|
||||
}
|
||||
found
|
||||
});
|
||||
}
|
||||
|
||||
// number of items to look up
|
||||
const LOOKUP_MAP_SIZE: u32 = 100_000_u32;
|
||||
const LOOKUP_SAMPLE_SIZE: u32 = 5000;
|
||||
const SORT_MAP_SIZE: usize = 10_000;
|
||||
|
||||
// use lazy_static so that comparison benchmarks use the exact same inputs
|
||||
lazy_static! {
|
||||
static ref KEYS: Vec<u32> = shuffled_keys(0..LOOKUP_MAP_SIZE);
|
||||
}
|
||||
|
||||
lazy_static! {
|
||||
static ref HMAP_100K: HashMap<u32, u32> = {
|
||||
let c = LOOKUP_MAP_SIZE;
|
||||
let mut map = HashMap::with_capacity(c as usize);
|
||||
let keys = &*KEYS;
|
||||
for &key in keys {
|
||||
map.insert(key, key);
|
||||
}
|
||||
map
|
||||
};
|
||||
}
|
||||
|
||||
lazy_static! {
|
||||
static ref IMAP_100K: IndexMap<u32, u32> = {
|
||||
let c = LOOKUP_MAP_SIZE;
|
||||
let mut map = IndexMap::with_capacity(c as usize);
|
||||
let keys = &*KEYS;
|
||||
for &key in keys {
|
||||
map.insert(key, key);
|
||||
}
|
||||
map
|
||||
};
|
||||
}
|
||||
|
||||
lazy_static! {
|
||||
static ref IMAP_SORT_U32: IndexMap<u32, u32> = {
|
||||
let mut map = IndexMap::with_capacity(SORT_MAP_SIZE);
|
||||
for &key in &KEYS[..SORT_MAP_SIZE] {
|
||||
map.insert(key, key);
|
||||
}
|
||||
map
|
||||
};
|
||||
}
|
||||
lazy_static! {
|
||||
static ref IMAP_SORT_S: IndexMap<String, String> = {
|
||||
let mut map = IndexMap::with_capacity(SORT_MAP_SIZE);
|
||||
for &key in &KEYS[..SORT_MAP_SIZE] {
|
||||
map.insert(format!("{:^16x}", &key), String::new());
|
||||
}
|
||||
map
|
||||
};
|
||||
}
|
||||
|
||||
#[bench]
|
||||
fn lookup_hashmap_100_000_multi(b: &mut Bencher) {
|
||||
let map = &*HMAP_100K;
|
||||
b.iter(|| {
|
||||
let mut found = 0;
|
||||
for key in 0..LOOKUP_SAMPLE_SIZE {
|
||||
found += map.get(&key).is_some() as u32;
|
||||
}
|
||||
found
|
||||
});
|
||||
}
|
||||
|
||||
#[bench]
|
||||
fn lookup_indexmap_100_000_multi(b: &mut Bencher) {
|
||||
let map = &*IMAP_100K;
|
||||
b.iter(|| {
|
||||
let mut found = 0;
|
||||
for key in 0..LOOKUP_SAMPLE_SIZE {
|
||||
found += map.get(&key).is_some() as u32;
|
||||
}
|
||||
found
|
||||
});
|
||||
}
|
||||
|
||||
// inorder: Test looking up keys in the same order as they were inserted
|
||||
#[bench]
|
||||
fn lookup_hashmap_100_000_inorder_multi(b: &mut Bencher) {
|
||||
let map = &*HMAP_100K;
|
||||
let keys = &*KEYS;
|
||||
b.iter(|| {
|
||||
let mut found = 0;
|
||||
for key in &keys[0..LOOKUP_SAMPLE_SIZE as usize] {
|
||||
found += map.get(key).is_some() as u32;
|
||||
}
|
||||
found
|
||||
});
|
||||
}
|
||||
|
||||
#[bench]
|
||||
fn lookup_indexmap_100_000_inorder_multi(b: &mut Bencher) {
|
||||
let map = &*IMAP_100K;
|
||||
let keys = &*KEYS;
|
||||
b.iter(|| {
|
||||
let mut found = 0;
|
||||
for key in &keys[0..LOOKUP_SAMPLE_SIZE as usize] {
|
||||
found += map.get(key).is_some() as u32;
|
||||
}
|
||||
found
|
||||
});
|
||||
}
|
||||
|
||||
#[bench]
|
||||
fn lookup_hashmap_100_000_single(b: &mut Bencher) {
|
||||
let map = &*HMAP_100K;
|
||||
let mut iter = (0..LOOKUP_MAP_SIZE + LOOKUP_SAMPLE_SIZE).cycle();
|
||||
b.iter(|| {
|
||||
let key = iter.next().unwrap();
|
||||
map.get(&key).is_some()
|
||||
});
|
||||
}
|
||||
|
||||
#[bench]
|
||||
fn lookup_indexmap_100_000_single(b: &mut Bencher) {
|
||||
let map = &*IMAP_100K;
|
||||
let mut iter = (0..LOOKUP_MAP_SIZE + LOOKUP_SAMPLE_SIZE).cycle();
|
||||
b.iter(|| {
|
||||
let key = iter.next().unwrap();
|
||||
map.get(&key).is_some()
|
||||
});
|
||||
}
|
||||
|
||||
const GROW_SIZE: usize = 100_000;
|
||||
type GrowKey = u32;
|
||||
|
||||
// Test grow/resize without preallocation
|
||||
#[bench]
|
||||
fn grow_fnv_hashmap_100_000(b: &mut Bencher) {
|
||||
b.iter(|| {
|
||||
let mut map: HashMap<_, _, FnvBuilder> = HashMap::default();
|
||||
for x in 0..GROW_SIZE {
|
||||
map.insert(x as GrowKey, x as GrowKey);
|
||||
}
|
||||
map
|
||||
});
|
||||
}
|
||||
|
||||
#[bench]
|
||||
fn grow_fnv_indexmap_100_000(b: &mut Bencher) {
|
||||
b.iter(|| {
|
||||
let mut map: IndexMap<_, _, FnvBuilder> = IndexMap::default();
|
||||
for x in 0..GROW_SIZE {
|
||||
map.insert(x as GrowKey, x as GrowKey);
|
||||
}
|
||||
map
|
||||
});
|
||||
}
|
||||
|
||||
const MERGE: u64 = 10_000;
|
||||
#[bench]
|
||||
fn hashmap_merge_simple(b: &mut Bencher) {
|
||||
let first_map: HashMap<u64, _> = (0..MERGE).map(|i| (i, ())).collect();
|
||||
let second_map: HashMap<u64, _> = (MERGE..MERGE * 2).map(|i| (i, ())).collect();
|
||||
b.iter(|| {
|
||||
let mut merged = first_map.clone();
|
||||
merged.extend(second_map.iter().map(|(&k, &v)| (k, v)));
|
||||
merged
|
||||
});
|
||||
}
|
||||
|
||||
#[bench]
|
||||
fn hashmap_merge_shuffle(b: &mut Bencher) {
|
||||
let first_map: HashMap<u64, _> = (0..MERGE).map(|i| (i, ())).collect();
|
||||
let second_map: HashMap<u64, _> = (MERGE..MERGE * 2).map(|i| (i, ())).collect();
|
||||
let mut v = Vec::new();
|
||||
let mut rng = small_rng();
|
||||
b.iter(|| {
|
||||
let mut merged = first_map.clone();
|
||||
v.extend(second_map.iter().map(|(&k, &v)| (k, v)));
|
||||
v.shuffle(&mut rng);
|
||||
merged.extend(v.drain(..));
|
||||
|
||||
merged
|
||||
});
|
||||
}
|
||||
|
||||
#[bench]
|
||||
fn indexmap_merge_simple(b: &mut Bencher) {
|
||||
let first_map: IndexMap<u64, _> = (0..MERGE).map(|i| (i, ())).collect();
|
||||
let second_map: IndexMap<u64, _> = (MERGE..MERGE * 2).map(|i| (i, ())).collect();
|
||||
b.iter(|| {
|
||||
let mut merged = first_map.clone();
|
||||
merged.extend(second_map.iter().map(|(&k, &v)| (k, v)));
|
||||
merged
|
||||
});
|
||||
}
|
||||
|
||||
#[bench]
|
||||
fn indexmap_merge_shuffle(b: &mut Bencher) {
|
||||
let first_map: IndexMap<u64, _> = (0..MERGE).map(|i| (i, ())).collect();
|
||||
let second_map: IndexMap<u64, _> = (MERGE..MERGE * 2).map(|i| (i, ())).collect();
|
||||
let mut v = Vec::new();
|
||||
let mut rng = small_rng();
|
||||
b.iter(|| {
|
||||
let mut merged = first_map.clone();
|
||||
v.extend(second_map.iter().map(|(&k, &v)| (k, v)));
|
||||
v.shuffle(&mut rng);
|
||||
merged.extend(v.drain(..));
|
||||
|
||||
merged
|
||||
});
|
||||
}
|
||||
|
||||
#[bench]
|
||||
fn swap_remove_indexmap_100_000(b: &mut Bencher) {
|
||||
let map = IMAP_100K.clone();
|
||||
let mut keys = Vec::from_iter(map.keys().copied());
|
||||
let mut rng = small_rng();
|
||||
keys.shuffle(&mut rng);
|
||||
|
||||
b.iter(|| {
|
||||
let mut map = map.clone();
|
||||
for key in &keys {
|
||||
map.swap_remove(key);
|
||||
}
|
||||
assert_eq!(map.len(), 0);
|
||||
map
|
||||
});
|
||||
}
|
||||
|
||||
#[bench]
|
||||
fn shift_remove_indexmap_100_000_few(b: &mut Bencher) {
|
||||
let map = IMAP_100K.clone();
|
||||
let mut keys = Vec::from_iter(map.keys().copied());
|
||||
let mut rng = small_rng();
|
||||
keys.shuffle(&mut rng);
|
||||
keys.truncate(50);
|
||||
|
||||
b.iter(|| {
|
||||
let mut map = map.clone();
|
||||
for key in &keys {
|
||||
map.shift_remove(key);
|
||||
}
|
||||
assert_eq!(map.len(), IMAP_100K.len() - keys.len());
|
||||
map
|
||||
});
|
||||
}
|
||||
|
||||
#[bench]
|
||||
fn shift_remove_indexmap_2_000_full(b: &mut Bencher) {
|
||||
let mut keys = KEYS[..2_000].to_vec();
|
||||
let mut map = IndexMap::with_capacity(keys.len());
|
||||
for &key in &keys {
|
||||
map.insert(key, key);
|
||||
}
|
||||
let mut rng = small_rng();
|
||||
keys.shuffle(&mut rng);
|
||||
|
||||
b.iter(|| {
|
||||
let mut map = map.clone();
|
||||
for key in &keys {
|
||||
map.shift_remove(key);
|
||||
}
|
||||
assert_eq!(map.len(), 0);
|
||||
map
|
||||
});
|
||||
}
|
||||
|
||||
#[bench]
|
||||
fn pop_indexmap_100_000(b: &mut Bencher) {
|
||||
let map = IMAP_100K.clone();
|
||||
|
||||
b.iter(|| {
|
||||
let mut map = map.clone();
|
||||
while !map.is_empty() {
|
||||
map.pop();
|
||||
}
|
||||
assert_eq!(map.len(), 0);
|
||||
map
|
||||
});
|
||||
}
|
||||
|
||||
#[bench]
|
||||
fn few_retain_indexmap_100_000(b: &mut Bencher) {
|
||||
let map = IMAP_100K.clone();
|
||||
|
||||
b.iter(|| {
|
||||
let mut map = map.clone();
|
||||
map.retain(|k, _| *k % 7 == 0);
|
||||
map
|
||||
});
|
||||
}
|
||||
|
||||
#[bench]
|
||||
fn few_retain_hashmap_100_000(b: &mut Bencher) {
|
||||
let map = HMAP_100K.clone();
|
||||
|
||||
b.iter(|| {
|
||||
let mut map = map.clone();
|
||||
map.retain(|k, _| *k % 7 == 0);
|
||||
map
|
||||
});
|
||||
}
|
||||
|
||||
#[bench]
|
||||
fn half_retain_indexmap_100_000(b: &mut Bencher) {
|
||||
let map = IMAP_100K.clone();
|
||||
|
||||
b.iter(|| {
|
||||
let mut map = map.clone();
|
||||
map.retain(|k, _| *k % 2 == 0);
|
||||
map
|
||||
});
|
||||
}
|
||||
|
||||
#[bench]
|
||||
fn half_retain_hashmap_100_000(b: &mut Bencher) {
|
||||
let map = HMAP_100K.clone();
|
||||
|
||||
b.iter(|| {
|
||||
let mut map = map.clone();
|
||||
map.retain(|k, _| *k % 2 == 0);
|
||||
map
|
||||
});
|
||||
}
|
||||
|
||||
#[bench]
|
||||
fn many_retain_indexmap_100_000(b: &mut Bencher) {
|
||||
let map = IMAP_100K.clone();
|
||||
|
||||
b.iter(|| {
|
||||
let mut map = map.clone();
|
||||
map.retain(|k, _| *k % 100 != 0);
|
||||
map
|
||||
});
|
||||
}
|
||||
|
||||
#[bench]
|
||||
fn many_retain_hashmap_100_000(b: &mut Bencher) {
|
||||
let map = HMAP_100K.clone();
|
||||
|
||||
b.iter(|| {
|
||||
let mut map = map.clone();
|
||||
map.retain(|k, _| *k % 100 != 0);
|
||||
map
|
||||
});
|
||||
}
|
||||
|
||||
// simple sort impl for comparison
|
||||
pub fn simple_sort<K: Ord + Hash, V>(m: &mut IndexMap<K, V>) {
|
||||
let mut ordered: Vec<_> = m.drain(..).collect();
|
||||
ordered.sort_by(|left, right| left.0.cmp(&right.0));
|
||||
m.extend(ordered);
|
||||
}
|
||||
|
||||
#[bench]
|
||||
fn indexmap_sort_s(b: &mut Bencher) {
|
||||
let map = IMAP_SORT_S.clone();
|
||||
|
||||
// there's a map clone there, but it's still useful to profile this
|
||||
b.iter(|| {
|
||||
let mut map = map.clone();
|
||||
map.sort_keys();
|
||||
map
|
||||
});
|
||||
}
|
||||
|
||||
#[bench]
|
||||
fn indexmap_simple_sort_s(b: &mut Bencher) {
|
||||
let map = IMAP_SORT_S.clone();
|
||||
|
||||
// there's a map clone there, but it's still useful to profile this
|
||||
b.iter(|| {
|
||||
let mut map = map.clone();
|
||||
simple_sort(&mut map);
|
||||
map
|
||||
});
|
||||
}
|
||||
|
||||
#[bench]
|
||||
fn indexmap_sort_u32(b: &mut Bencher) {
|
||||
let map = IMAP_SORT_U32.clone();
|
||||
|
||||
// there's a map clone there, but it's still useful to profile this
|
||||
b.iter(|| {
|
||||
let mut map = map.clone();
|
||||
map.sort_keys();
|
||||
map
|
||||
});
|
||||
}
|
||||
|
||||
#[bench]
|
||||
fn indexmap_simple_sort_u32(b: &mut Bencher) {
|
||||
let map = IMAP_SORT_U32.clone();
|
||||
|
||||
// there's a map clone there, but it's still useful to profile this
|
||||
b.iter(|| {
|
||||
let mut map = map.clone();
|
||||
simple_sort(&mut map);
|
||||
map
|
||||
});
|
||||
}
|
||||
|
||||
// measure the fixed overhead of cloning in sort benchmarks
|
||||
#[bench]
|
||||
fn indexmap_clone_for_sort_s(b: &mut Bencher) {
|
||||
let map = IMAP_SORT_S.clone();
|
||||
|
||||
b.iter(|| map.clone());
|
||||
}
|
||||
|
||||
#[bench]
|
||||
fn indexmap_clone_for_sort_u32(b: &mut Bencher) {
|
||||
let map = IMAP_SORT_U32.clone();
|
||||
|
||||
b.iter(|| map.clone());
|
||||
}
|
||||
185
clamav/libclamav_rust/.cargo/vendor/indexmap/benches/faststring.rs
vendored
Normal file
185
clamav/libclamav_rust/.cargo/vendor/indexmap/benches/faststring.rs
vendored
Normal file
@@ -0,0 +1,185 @@
|
||||
#![feature(test)]
|
||||
|
||||
extern crate test;
|
||||
|
||||
use test::Bencher;
|
||||
|
||||
use indexmap::IndexMap;
|
||||
|
||||
use std::collections::HashMap;
|
||||
|
||||
use rand::rngs::SmallRng;
|
||||
use rand::seq::SliceRandom;
|
||||
use rand::SeedableRng;
|
||||
|
||||
use std::hash::{Hash, Hasher};
|
||||
|
||||
use std::borrow::Borrow;
|
||||
use std::ops::Deref;
|
||||
|
||||
/// Use a consistently seeded Rng for benchmark stability
|
||||
fn small_rng() -> SmallRng {
|
||||
let seed = u64::from_le_bytes(*b"indexmap");
|
||||
SmallRng::seed_from_u64(seed)
|
||||
}
|
||||
|
||||
#[derive(PartialEq, Eq, Copy, Clone)]
|
||||
#[repr(transparent)]
|
||||
pub struct OneShot<T: ?Sized>(pub T);
|
||||
|
||||
impl Hash for OneShot<str> {
|
||||
fn hash<H: Hasher>(&self, h: &mut H) {
|
||||
h.write(self.0.as_bytes())
|
||||
}
|
||||
}
|
||||
|
||||
impl<'a, S> From<&'a S> for &'a OneShot<str>
|
||||
where
|
||||
S: AsRef<str>,
|
||||
{
|
||||
fn from(s: &'a S) -> Self {
|
||||
let s: &str = s.as_ref();
|
||||
unsafe { &*(s as *const str as *const OneShot<str>) }
|
||||
}
|
||||
}
|
||||
|
||||
impl Hash for OneShot<String> {
|
||||
fn hash<H: Hasher>(&self, h: &mut H) {
|
||||
h.write(self.0.as_bytes())
|
||||
}
|
||||
}
|
||||
|
||||
impl Borrow<OneShot<str>> for OneShot<String> {
|
||||
fn borrow(&self) -> &OneShot<str> {
|
||||
<&OneShot<str>>::from(&self.0)
|
||||
}
|
||||
}
|
||||
|
||||
impl<T> Deref for OneShot<T> {
|
||||
type Target = T;
|
||||
fn deref(&self) -> &T {
|
||||
&self.0
|
||||
}
|
||||
}
|
||||
|
||||
fn shuffled_keys<I>(iter: I) -> Vec<I::Item>
|
||||
where
|
||||
I: IntoIterator,
|
||||
{
|
||||
let mut v = Vec::from_iter(iter);
|
||||
let mut rng = small_rng();
|
||||
v.shuffle(&mut rng);
|
||||
v
|
||||
}
|
||||
|
||||
#[bench]
|
||||
fn insert_hashmap_string_10_000(b: &mut Bencher) {
|
||||
let c = 10_000;
|
||||
b.iter(|| {
|
||||
let mut map = HashMap::with_capacity(c);
|
||||
for x in 0..c {
|
||||
map.insert(x.to_string(), ());
|
||||
}
|
||||
map
|
||||
});
|
||||
}
|
||||
|
||||
#[bench]
|
||||
fn insert_hashmap_string_oneshot_10_000(b: &mut Bencher) {
|
||||
let c = 10_000;
|
||||
b.iter(|| {
|
||||
let mut map = HashMap::with_capacity(c);
|
||||
for x in 0..c {
|
||||
map.insert(OneShot(x.to_string()), ());
|
||||
}
|
||||
map
|
||||
});
|
||||
}
|
||||
|
||||
#[bench]
|
||||
fn insert_indexmap_string_10_000(b: &mut Bencher) {
|
||||
let c = 10_000;
|
||||
b.iter(|| {
|
||||
let mut map = IndexMap::with_capacity(c);
|
||||
for x in 0..c {
|
||||
map.insert(x.to_string(), ());
|
||||
}
|
||||
map
|
||||
});
|
||||
}
|
||||
|
||||
#[bench]
|
||||
fn lookup_hashmap_10_000_exist_string(b: &mut Bencher) {
|
||||
let c = 10_000;
|
||||
let mut map = HashMap::with_capacity(c);
|
||||
let keys = shuffled_keys(0..c);
|
||||
for &key in &keys {
|
||||
map.insert(key.to_string(), 1);
|
||||
}
|
||||
let lookups = (5000..c).map(|x| x.to_string()).collect::<Vec<_>>();
|
||||
b.iter(|| {
|
||||
let mut found = 0;
|
||||
for key in &lookups {
|
||||
found += map.get(key).is_some() as i32;
|
||||
}
|
||||
found
|
||||
});
|
||||
}
|
||||
|
||||
#[bench]
|
||||
fn lookup_hashmap_10_000_exist_string_oneshot(b: &mut Bencher) {
|
||||
let c = 10_000;
|
||||
let mut map = HashMap::with_capacity(c);
|
||||
let keys = shuffled_keys(0..c);
|
||||
for &key in &keys {
|
||||
map.insert(OneShot(key.to_string()), 1);
|
||||
}
|
||||
let lookups = (5000..c)
|
||||
.map(|x| OneShot(x.to_string()))
|
||||
.collect::<Vec<_>>();
|
||||
b.iter(|| {
|
||||
let mut found = 0;
|
||||
for key in &lookups {
|
||||
found += map.get(key).is_some() as i32;
|
||||
}
|
||||
found
|
||||
});
|
||||
}
|
||||
|
||||
#[bench]
|
||||
fn lookup_indexmap_10_000_exist_string(b: &mut Bencher) {
|
||||
let c = 10_000;
|
||||
let mut map = IndexMap::with_capacity(c);
|
||||
let keys = shuffled_keys(0..c);
|
||||
for &key in &keys {
|
||||
map.insert(key.to_string(), 1);
|
||||
}
|
||||
let lookups = (5000..c).map(|x| x.to_string()).collect::<Vec<_>>();
|
||||
b.iter(|| {
|
||||
let mut found = 0;
|
||||
for key in &lookups {
|
||||
found += map.get(key).is_some() as i32;
|
||||
}
|
||||
found
|
||||
});
|
||||
}
|
||||
|
||||
#[bench]
|
||||
fn lookup_indexmap_10_000_exist_string_oneshot(b: &mut Bencher) {
|
||||
let c = 10_000;
|
||||
let mut map = IndexMap::with_capacity(c);
|
||||
let keys = shuffled_keys(0..c);
|
||||
for &key in &keys {
|
||||
map.insert(OneShot(key.to_string()), 1);
|
||||
}
|
||||
let lookups = (5000..c)
|
||||
.map(|x| OneShot(x.to_string()))
|
||||
.collect::<Vec<_>>();
|
||||
b.iter(|| {
|
||||
let mut found = 0;
|
||||
for key in &lookups {
|
||||
found += map.get(key).is_some() as i32;
|
||||
}
|
||||
found
|
||||
});
|
||||
}
|
||||
8
clamav/libclamav_rust/.cargo/vendor/indexmap/build.rs
vendored
Normal file
8
clamav/libclamav_rust/.cargo/vendor/indexmap/build.rs
vendored
Normal file
@@ -0,0 +1,8 @@
|
||||
fn main() {
|
||||
// If "std" is explicitly requested, don't bother probing the target for it.
|
||||
match std::env::var_os("CARGO_FEATURE_STD") {
|
||||
Some(_) => autocfg::emit("has_std"),
|
||||
None => autocfg::new().emit_sysroot_crate("std"),
|
||||
}
|
||||
autocfg::rerun_path("build.rs");
|
||||
}
|
||||
75
clamav/libclamav_rust/.cargo/vendor/indexmap/src/arbitrary.rs
vendored
Normal file
75
clamav/libclamav_rust/.cargo/vendor/indexmap/src/arbitrary.rs
vendored
Normal file
@@ -0,0 +1,75 @@
|
||||
#[cfg(feature = "arbitrary")]
|
||||
mod impl_arbitrary {
|
||||
use crate::{IndexMap, IndexSet};
|
||||
use arbitrary::{Arbitrary, Result, Unstructured};
|
||||
use core::hash::{BuildHasher, Hash};
|
||||
|
||||
impl<'a, K, V, S> Arbitrary<'a> for IndexMap<K, V, S>
|
||||
where
|
||||
K: Arbitrary<'a> + Hash + Eq,
|
||||
V: Arbitrary<'a>,
|
||||
S: BuildHasher + Default,
|
||||
{
|
||||
fn arbitrary(u: &mut Unstructured<'a>) -> Result<Self> {
|
||||
u.arbitrary_iter()?.collect()
|
||||
}
|
||||
|
||||
fn arbitrary_take_rest(u: Unstructured<'a>) -> Result<Self> {
|
||||
u.arbitrary_take_rest_iter()?.collect()
|
||||
}
|
||||
}
|
||||
|
||||
impl<'a, T, S> Arbitrary<'a> for IndexSet<T, S>
|
||||
where
|
||||
T: Arbitrary<'a> + Hash + Eq,
|
||||
S: BuildHasher + Default,
|
||||
{
|
||||
fn arbitrary(u: &mut Unstructured<'a>) -> Result<Self> {
|
||||
u.arbitrary_iter()?.collect()
|
||||
}
|
||||
|
||||
fn arbitrary_take_rest(u: Unstructured<'a>) -> Result<Self> {
|
||||
u.arbitrary_take_rest_iter()?.collect()
|
||||
}
|
||||
}
|
||||
}
|
||||
|
||||
#[cfg(feature = "quickcheck")]
|
||||
mod impl_quickcheck {
|
||||
use crate::{IndexMap, IndexSet};
|
||||
use alloc::boxed::Box;
|
||||
use alloc::vec::Vec;
|
||||
use core::hash::{BuildHasher, Hash};
|
||||
use quickcheck::{Arbitrary, Gen};
|
||||
|
||||
impl<K, V, S> Arbitrary for IndexMap<K, V, S>
|
||||
where
|
||||
K: Arbitrary + Hash + Eq,
|
||||
V: Arbitrary,
|
||||
S: BuildHasher + Default + Clone + 'static,
|
||||
{
|
||||
fn arbitrary(g: &mut Gen) -> Self {
|
||||
Self::from_iter(Vec::arbitrary(g))
|
||||
}
|
||||
|
||||
fn shrink(&self) -> Box<dyn Iterator<Item = Self>> {
|
||||
let vec = Vec::from_iter(self.clone());
|
||||
Box::new(vec.shrink().map(Self::from_iter))
|
||||
}
|
||||
}
|
||||
|
||||
impl<T, S> Arbitrary for IndexSet<T, S>
|
||||
where
|
||||
T: Arbitrary + Hash + Eq,
|
||||
S: BuildHasher + Default + Clone + 'static,
|
||||
{
|
||||
fn arbitrary(g: &mut Gen) -> Self {
|
||||
Self::from_iter(Vec::arbitrary(g))
|
||||
}
|
||||
|
||||
fn shrink(&self) -> Box<dyn Iterator<Item = Self>> {
|
||||
let vec = Vec::from_iter(self.clone());
|
||||
Box::new(vec.shrink().map(Self::from_iter))
|
||||
}
|
||||
}
|
||||
}
|
||||
27
clamav/libclamav_rust/.cargo/vendor/indexmap/src/equivalent.rs
vendored
Normal file
27
clamav/libclamav_rust/.cargo/vendor/indexmap/src/equivalent.rs
vendored
Normal file
@@ -0,0 +1,27 @@
|
||||
use core::borrow::Borrow;
|
||||
|
||||
/// Key equivalence trait.
|
||||
///
|
||||
/// This trait allows hash table lookup to be customized.
|
||||
/// It has one blanket implementation that uses the regular `Borrow` solution,
|
||||
/// just like `HashMap` and `BTreeMap` do, so that you can pass `&str` to lookup
|
||||
/// into a map with `String` keys and so on.
|
||||
///
|
||||
/// # Contract
|
||||
///
|
||||
/// The implementor **must** hash like `K`, if it is hashable.
|
||||
pub trait Equivalent<K: ?Sized> {
|
||||
/// Compare self to `key` and return `true` if they are equal.
|
||||
fn equivalent(&self, key: &K) -> bool;
|
||||
}
|
||||
|
||||
impl<Q: ?Sized, K: ?Sized> Equivalent<K> for Q
|
||||
where
|
||||
Q: Eq,
|
||||
K: Borrow<Q>,
|
||||
{
|
||||
#[inline]
|
||||
fn equivalent(&self, key: &K) -> bool {
|
||||
*self == *key.borrow()
|
||||
}
|
||||
}
|
||||
194
clamav/libclamav_rust/.cargo/vendor/indexmap/src/lib.rs
vendored
Normal file
194
clamav/libclamav_rust/.cargo/vendor/indexmap/src/lib.rs
vendored
Normal file
@@ -0,0 +1,194 @@
|
||||
// We *mostly* avoid unsafe code, but `map::core::raw` allows it to use `RawTable` buckets.
|
||||
#![deny(unsafe_code)]
|
||||
#![warn(rust_2018_idioms)]
|
||||
#![doc(html_root_url = "https://docs.rs/indexmap/1/")]
|
||||
#![no_std]
|
||||
|
||||
//! [`IndexMap`] is a hash table where the iteration order of the key-value
|
||||
//! pairs is independent of the hash values of the keys.
|
||||
//!
|
||||
//! [`IndexSet`] is a corresponding hash set using the same implementation and
|
||||
//! with similar properties.
|
||||
//!
|
||||
//! [`IndexMap`]: map/struct.IndexMap.html
|
||||
//! [`IndexSet`]: set/struct.IndexSet.html
|
||||
//!
|
||||
//!
|
||||
//! ### Feature Highlights
|
||||
//!
|
||||
//! [`IndexMap`] and [`IndexSet`] are drop-in compatible with the std `HashMap`
|
||||
//! and `HashSet`, but they also have some features of note:
|
||||
//!
|
||||
//! - The ordering semantics (see their documentation for details)
|
||||
//! - Sorting methods and the [`.pop()`][IndexMap::pop] methods.
|
||||
//! - The [`Equivalent`] trait, which offers more flexible equality definitions
|
||||
//! between borrowed and owned versions of keys.
|
||||
//! - The [`MutableKeys`][map::MutableKeys] trait, which gives opt-in mutable
|
||||
//! access to hash map keys.
|
||||
//!
|
||||
//! ### Alternate Hashers
|
||||
//!
|
||||
//! [`IndexMap`] and [`IndexSet`] have a default hasher type `S = RandomState`,
|
||||
//! just like the standard `HashMap` and `HashSet`, which is resistant to
|
||||
//! HashDoS attacks but not the most performant. Type aliases can make it easier
|
||||
//! to use alternate hashers:
|
||||
//!
|
||||
//! ```
|
||||
//! use fnv::FnvBuildHasher;
|
||||
//! use fxhash::FxBuildHasher;
|
||||
//! use indexmap::{IndexMap, IndexSet};
|
||||
//!
|
||||
//! type FnvIndexMap<K, V> = IndexMap<K, V, FnvBuildHasher>;
|
||||
//! type FnvIndexSet<T> = IndexSet<T, FnvBuildHasher>;
|
||||
//!
|
||||
//! type FxIndexMap<K, V> = IndexMap<K, V, FxBuildHasher>;
|
||||
//! type FxIndexSet<T> = IndexSet<T, FxBuildHasher>;
|
||||
//!
|
||||
//! let std: IndexSet<i32> = (0..100).collect();
|
||||
//! let fnv: FnvIndexSet<i32> = (0..100).collect();
|
||||
//! let fx: FxIndexSet<i32> = (0..100).collect();
|
||||
//! assert_eq!(std, fnv);
|
||||
//! assert_eq!(std, fx);
|
||||
//! ```
|
||||
//!
|
||||
//! ### Rust Version
|
||||
//!
|
||||
//! This version of indexmap requires Rust 1.56 or later.
|
||||
//!
|
||||
//! The indexmap 1.x release series will use a carefully considered version
|
||||
//! upgrade policy, where in a later 1.x version, we will raise the minimum
|
||||
//! required Rust version.
|
||||
//!
|
||||
//! ## No Standard Library Targets
|
||||
//!
|
||||
//! This crate supports being built without `std`, requiring
|
||||
//! `alloc` instead. This is enabled automatically when it is detected that
|
||||
//! `std` is not available. There is no crate feature to enable/disable to
|
||||
//! trigger this. It can be tested by building for a std-less target.
|
||||
//!
|
||||
//! - Creating maps and sets using [`new`][IndexMap::new] and
|
||||
//! [`with_capacity`][IndexMap::with_capacity] is unavailable without `std`.
|
||||
//! Use methods [`IndexMap::default`][def],
|
||||
//! [`with_hasher`][IndexMap::with_hasher],
|
||||
//! [`with_capacity_and_hasher`][IndexMap::with_capacity_and_hasher] instead.
|
||||
//! A no-std compatible hasher will be needed as well, for example
|
||||
//! from the crate `twox-hash`.
|
||||
//! - Macros [`indexmap!`] and [`indexset!`] are unavailable without `std`.
|
||||
//!
|
||||
//! [def]: map/struct.IndexMap.html#impl-Default
|
||||
|
||||
extern crate alloc;
|
||||
|
||||
#[cfg(has_std)]
|
||||
#[macro_use]
|
||||
extern crate std;
|
||||
|
||||
use alloc::vec::{self, Vec};
|
||||
|
||||
mod arbitrary;
|
||||
#[macro_use]
|
||||
mod macros;
|
||||
mod equivalent;
|
||||
mod mutable_keys;
|
||||
#[cfg(feature = "serde")]
|
||||
mod serde;
|
||||
#[cfg(feature = "serde")]
|
||||
pub mod serde_seq;
|
||||
mod util;
|
||||
|
||||
pub mod map;
|
||||
pub mod set;
|
||||
|
||||
// Placed after `map` and `set` so new `rayon` methods on the types
|
||||
// are documented after the "normal" methods.
|
||||
#[cfg(feature = "rayon")]
|
||||
mod rayon;
|
||||
|
||||
#[cfg(feature = "rustc-rayon")]
|
||||
mod rustc;
|
||||
|
||||
pub use crate::equivalent::Equivalent;
|
||||
pub use crate::map::IndexMap;
|
||||
pub use crate::set::IndexSet;
|
||||
|
||||
// shared private items
|
||||
|
||||
/// Hash value newtype. Not larger than usize, since anything larger
|
||||
/// isn't used for selecting position anyway.
|
||||
#[derive(Clone, Copy, Debug, PartialEq)]
|
||||
struct HashValue(usize);
|
||||
|
||||
impl HashValue {
|
||||
#[inline(always)]
|
||||
fn get(self) -> u64 {
|
||||
self.0 as u64
|
||||
}
|
||||
}
|
||||
|
||||
#[derive(Copy, Debug)]
|
||||
struct Bucket<K, V> {
|
||||
hash: HashValue,
|
||||
key: K,
|
||||
value: V,
|
||||
}
|
||||
|
||||
impl<K, V> Clone for Bucket<K, V>
|
||||
where
|
||||
K: Clone,
|
||||
V: Clone,
|
||||
{
|
||||
fn clone(&self) -> Self {
|
||||
Bucket {
|
||||
hash: self.hash,
|
||||
key: self.key.clone(),
|
||||
value: self.value.clone(),
|
||||
}
|
||||
}
|
||||
|
||||
fn clone_from(&mut self, other: &Self) {
|
||||
self.hash = other.hash;
|
||||
self.key.clone_from(&other.key);
|
||||
self.value.clone_from(&other.value);
|
||||
}
|
||||
}
|
||||
|
||||
impl<K, V> Bucket<K, V> {
|
||||
// field accessors -- used for `f` instead of closures in `.map(f)`
|
||||
fn key_ref(&self) -> &K {
|
||||
&self.key
|
||||
}
|
||||
fn value_ref(&self) -> &V {
|
||||
&self.value
|
||||
}
|
||||
fn value_mut(&mut self) -> &mut V {
|
||||
&mut self.value
|
||||
}
|
||||
fn key(self) -> K {
|
||||
self.key
|
||||
}
|
||||
fn value(self) -> V {
|
||||
self.value
|
||||
}
|
||||
fn key_value(self) -> (K, V) {
|
||||
(self.key, self.value)
|
||||
}
|
||||
fn refs(&self) -> (&K, &V) {
|
||||
(&self.key, &self.value)
|
||||
}
|
||||
fn ref_mut(&mut self) -> (&K, &mut V) {
|
||||
(&self.key, &mut self.value)
|
||||
}
|
||||
fn muts(&mut self) -> (&mut K, &mut V) {
|
||||
(&mut self.key, &mut self.value)
|
||||
}
|
||||
}
|
||||
|
||||
trait Entries {
|
||||
type Entry;
|
||||
fn into_entries(self) -> Vec<Self::Entry>;
|
||||
fn as_entries(&self) -> &[Self::Entry];
|
||||
fn as_entries_mut(&mut self) -> &mut [Self::Entry];
|
||||
fn with_entries<F>(&mut self, f: F)
|
||||
where
|
||||
F: FnOnce(&mut [Self::Entry]);
|
||||
}
|
||||
178
clamav/libclamav_rust/.cargo/vendor/indexmap/src/macros.rs
vendored
Normal file
178
clamav/libclamav_rust/.cargo/vendor/indexmap/src/macros.rs
vendored
Normal file
@@ -0,0 +1,178 @@
|
||||
#[cfg(has_std)]
|
||||
#[macro_export]
|
||||
/// Create an `IndexMap` from a list of key-value pairs
|
||||
///
|
||||
/// ## Example
|
||||
///
|
||||
/// ```
|
||||
/// use indexmap::indexmap;
|
||||
///
|
||||
/// let map = indexmap!{
|
||||
/// "a" => 1,
|
||||
/// "b" => 2,
|
||||
/// };
|
||||
/// assert_eq!(map["a"], 1);
|
||||
/// assert_eq!(map["b"], 2);
|
||||
/// assert_eq!(map.get("c"), None);
|
||||
///
|
||||
/// // "a" is the first key
|
||||
/// assert_eq!(map.keys().next(), Some(&"a"));
|
||||
/// ```
|
||||
macro_rules! indexmap {
|
||||
(@single $($x:tt)*) => (());
|
||||
(@count $($rest:expr),*) => (<[()]>::len(&[$($crate::indexmap!(@single $rest)),*]));
|
||||
|
||||
($($key:expr => $value:expr,)+) => { $crate::indexmap!($($key => $value),+) };
|
||||
($($key:expr => $value:expr),*) => {
|
||||
{
|
||||
let _cap = $crate::indexmap!(@count $($key),*);
|
||||
let mut _map = $crate::IndexMap::with_capacity(_cap);
|
||||
$(
|
||||
_map.insert($key, $value);
|
||||
)*
|
||||
_map
|
||||
}
|
||||
};
|
||||
}
|
||||
|
||||
#[cfg(has_std)]
|
||||
#[macro_export]
|
||||
/// Create an `IndexSet` from a list of values
|
||||
///
|
||||
/// ## Example
|
||||
///
|
||||
/// ```
|
||||
/// use indexmap::indexset;
|
||||
///
|
||||
/// let set = indexset!{
|
||||
/// "a",
|
||||
/// "b",
|
||||
/// };
|
||||
/// assert!(set.contains("a"));
|
||||
/// assert!(set.contains("b"));
|
||||
/// assert!(!set.contains("c"));
|
||||
///
|
||||
/// // "a" is the first value
|
||||
/// assert_eq!(set.iter().next(), Some(&"a"));
|
||||
/// ```
|
||||
macro_rules! indexset {
|
||||
(@single $($x:tt)*) => (());
|
||||
(@count $($rest:expr),*) => (<[()]>::len(&[$($crate::indexset!(@single $rest)),*]));
|
||||
|
||||
($($value:expr,)+) => { $crate::indexset!($($value),+) };
|
||||
($($value:expr),*) => {
|
||||
{
|
||||
let _cap = $crate::indexset!(@count $($value),*);
|
||||
let mut _set = $crate::IndexSet::with_capacity(_cap);
|
||||
$(
|
||||
_set.insert($value);
|
||||
)*
|
||||
_set
|
||||
}
|
||||
};
|
||||
}
|
||||
|
||||
// generate all the Iterator methods by just forwarding to the underlying
|
||||
// self.iter and mapping its element.
|
||||
macro_rules! iterator_methods {
|
||||
// $map_elt is the mapping function from the underlying iterator's element
|
||||
// same mapping function for both options and iterators
|
||||
($map_elt:expr) => {
|
||||
fn next(&mut self) -> Option<Self::Item> {
|
||||
self.iter.next().map($map_elt)
|
||||
}
|
||||
|
||||
fn size_hint(&self) -> (usize, Option<usize>) {
|
||||
self.iter.size_hint()
|
||||
}
|
||||
|
||||
fn count(self) -> usize {
|
||||
self.iter.len()
|
||||
}
|
||||
|
||||
fn nth(&mut self, n: usize) -> Option<Self::Item> {
|
||||
self.iter.nth(n).map($map_elt)
|
||||
}
|
||||
|
||||
fn last(mut self) -> Option<Self::Item> {
|
||||
self.next_back()
|
||||
}
|
||||
|
||||
fn collect<C>(self) -> C
|
||||
where
|
||||
C: FromIterator<Self::Item>,
|
||||
{
|
||||
// NB: forwarding this directly to standard iterators will
|
||||
// allow it to leverage unstable traits like `TrustedLen`.
|
||||
self.iter.map($map_elt).collect()
|
||||
}
|
||||
};
|
||||
}
|
||||
|
||||
macro_rules! double_ended_iterator_methods {
|
||||
// $map_elt is the mapping function from the underlying iterator's element
|
||||
// same mapping function for both options and iterators
|
||||
($map_elt:expr) => {
|
||||
fn next_back(&mut self) -> Option<Self::Item> {
|
||||
self.iter.next_back().map($map_elt)
|
||||
}
|
||||
|
||||
fn nth_back(&mut self, n: usize) -> Option<Self::Item> {
|
||||
self.iter.nth_back(n).map($map_elt)
|
||||
}
|
||||
};
|
||||
}
|
||||
|
||||
// generate `ParallelIterator` methods by just forwarding to the underlying
|
||||
// self.entries and mapping its elements.
|
||||
#[cfg(any(feature = "rayon", feature = "rustc-rayon"))]
|
||||
macro_rules! parallel_iterator_methods {
|
||||
// $map_elt is the mapping function from the underlying iterator's element
|
||||
($map_elt:expr) => {
|
||||
fn drive_unindexed<C>(self, consumer: C) -> C::Result
|
||||
where
|
||||
C: UnindexedConsumer<Self::Item>,
|
||||
{
|
||||
self.entries
|
||||
.into_par_iter()
|
||||
.map($map_elt)
|
||||
.drive_unindexed(consumer)
|
||||
}
|
||||
|
||||
// NB: This allows indexed collection, e.g. directly into a `Vec`, but the
|
||||
// underlying iterator must really be indexed. We should remove this if we
|
||||
// start having tombstones that must be filtered out.
|
||||
fn opt_len(&self) -> Option<usize> {
|
||||
Some(self.entries.len())
|
||||
}
|
||||
};
|
||||
}
|
||||
|
||||
// generate `IndexedParallelIterator` methods by just forwarding to the underlying
|
||||
// self.entries and mapping its elements.
|
||||
#[cfg(any(feature = "rayon", feature = "rustc-rayon"))]
|
||||
macro_rules! indexed_parallel_iterator_methods {
|
||||
// $map_elt is the mapping function from the underlying iterator's element
|
||||
($map_elt:expr) => {
|
||||
fn drive<C>(self, consumer: C) -> C::Result
|
||||
where
|
||||
C: Consumer<Self::Item>,
|
||||
{
|
||||
self.entries.into_par_iter().map($map_elt).drive(consumer)
|
||||
}
|
||||
|
||||
fn len(&self) -> usize {
|
||||
self.entries.len()
|
||||
}
|
||||
|
||||
fn with_producer<CB>(self, callback: CB) -> CB::Output
|
||||
where
|
||||
CB: ProducerCallback<Self::Item>,
|
||||
{
|
||||
self.entries
|
||||
.into_par_iter()
|
||||
.map($map_elt)
|
||||
.with_producer(callback)
|
||||
}
|
||||
};
|
||||
}
|
||||
1947
clamav/libclamav_rust/.cargo/vendor/indexmap/src/map.rs
vendored
Normal file
1947
clamav/libclamav_rust/.cargo/vendor/indexmap/src/map.rs
vendored
Normal file
File diff suppressed because it is too large
Load Diff
700
clamav/libclamav_rust/.cargo/vendor/indexmap/src/map/core.rs
vendored
Normal file
700
clamav/libclamav_rust/.cargo/vendor/indexmap/src/map/core.rs
vendored
Normal file
@@ -0,0 +1,700 @@
|
||||
//! This is the core implementation that doesn't depend on the hasher at all.
|
||||
//!
|
||||
//! The methods of `IndexMapCore` don't use any Hash properties of K.
|
||||
//!
|
||||
//! It's cleaner to separate them out, then the compiler checks that we are not
|
||||
//! using Hash at all in these methods.
|
||||
//!
|
||||
//! However, we should probably not let this show in the public API or docs.
|
||||
|
||||
mod raw;
|
||||
|
||||
use hashbrown::raw::RawTable;
|
||||
|
||||
use crate::vec::{Drain, Vec};
|
||||
use core::cmp;
|
||||
use core::fmt;
|
||||
use core::mem::replace;
|
||||
use core::ops::RangeBounds;
|
||||
|
||||
use crate::equivalent::Equivalent;
|
||||
use crate::util::simplify_range;
|
||||
use crate::{Bucket, Entries, HashValue};
|
||||
|
||||
/// Core of the map that does not depend on S
|
||||
pub(crate) struct IndexMapCore<K, V> {
|
||||
/// indices mapping from the entry hash to its index.
|
||||
indices: RawTable<usize>,
|
||||
/// entries is a dense vec of entries in their order.
|
||||
entries: Vec<Bucket<K, V>>,
|
||||
}
|
||||
|
||||
#[inline(always)]
|
||||
fn get_hash<K, V>(entries: &[Bucket<K, V>]) -> impl Fn(&usize) -> u64 + '_ {
|
||||
move |&i| entries[i].hash.get()
|
||||
}
|
||||
|
||||
#[inline]
|
||||
fn equivalent<'a, K, V, Q: ?Sized + Equivalent<K>>(
|
||||
key: &'a Q,
|
||||
entries: &'a [Bucket<K, V>],
|
||||
) -> impl Fn(&usize) -> bool + 'a {
|
||||
move |&i| Q::equivalent(key, &entries[i].key)
|
||||
}
|
||||
|
||||
#[inline]
|
||||
fn erase_index(table: &mut RawTable<usize>, hash: HashValue, index: usize) {
|
||||
let erased = table.erase_entry(hash.get(), move |&i| i == index);
|
||||
debug_assert!(erased);
|
||||
}
|
||||
|
||||
#[inline]
|
||||
fn update_index(table: &mut RawTable<usize>, hash: HashValue, old: usize, new: usize) {
|
||||
let index = table
|
||||
.get_mut(hash.get(), move |&i| i == old)
|
||||
.expect("index not found");
|
||||
*index = new;
|
||||
}
|
||||
|
||||
impl<K, V> Clone for IndexMapCore<K, V>
|
||||
where
|
||||
K: Clone,
|
||||
V: Clone,
|
||||
{
|
||||
fn clone(&self) -> Self {
|
||||
let indices = self.indices.clone();
|
||||
let mut entries = Vec::with_capacity(indices.capacity());
|
||||
entries.clone_from(&self.entries);
|
||||
IndexMapCore { indices, entries }
|
||||
}
|
||||
|
||||
fn clone_from(&mut self, other: &Self) {
|
||||
let hasher = get_hash(&other.entries);
|
||||
self.indices.clone_from_with_hasher(&other.indices, hasher);
|
||||
if self.entries.capacity() < other.entries.len() {
|
||||
// If we must resize, match the indices capacity
|
||||
self.reserve_entries();
|
||||
}
|
||||
self.entries.clone_from(&other.entries);
|
||||
}
|
||||
}
|
||||
|
||||
impl<K, V> fmt::Debug for IndexMapCore<K, V>
|
||||
where
|
||||
K: fmt::Debug,
|
||||
V: fmt::Debug,
|
||||
{
|
||||
fn fmt(&self, f: &mut fmt::Formatter<'_>) -> fmt::Result {
|
||||
f.debug_struct("IndexMapCore")
|
||||
.field("indices", &raw::DebugIndices(&self.indices))
|
||||
.field("entries", &self.entries)
|
||||
.finish()
|
||||
}
|
||||
}
|
||||
|
||||
impl<K, V> Entries for IndexMapCore<K, V> {
|
||||
type Entry = Bucket<K, V>;
|
||||
|
||||
#[inline]
|
||||
fn into_entries(self) -> Vec<Self::Entry> {
|
||||
self.entries
|
||||
}
|
||||
|
||||
#[inline]
|
||||
fn as_entries(&self) -> &[Self::Entry] {
|
||||
&self.entries
|
||||
}
|
||||
|
||||
#[inline]
|
||||
fn as_entries_mut(&mut self) -> &mut [Self::Entry] {
|
||||
&mut self.entries
|
||||
}
|
||||
|
||||
fn with_entries<F>(&mut self, f: F)
|
||||
where
|
||||
F: FnOnce(&mut [Self::Entry]),
|
||||
{
|
||||
f(&mut self.entries);
|
||||
self.rebuild_hash_table();
|
||||
}
|
||||
}
|
||||
|
||||
impl<K, V> IndexMapCore<K, V> {
|
||||
#[inline]
|
||||
pub(crate) const fn new() -> Self {
|
||||
IndexMapCore {
|
||||
indices: RawTable::new(),
|
||||
entries: Vec::new(),
|
||||
}
|
||||
}
|
||||
|
||||
#[inline]
|
||||
pub(crate) fn with_capacity(n: usize) -> Self {
|
||||
IndexMapCore {
|
||||
indices: RawTable::with_capacity(n),
|
||||
entries: Vec::with_capacity(n),
|
||||
}
|
||||
}
|
||||
|
||||
#[inline]
|
||||
pub(crate) fn len(&self) -> usize {
|
||||
self.indices.len()
|
||||
}
|
||||
|
||||
#[inline]
|
||||
pub(crate) fn capacity(&self) -> usize {
|
||||
cmp::min(self.indices.capacity(), self.entries.capacity())
|
||||
}
|
||||
|
||||
pub(crate) fn clear(&mut self) {
|
||||
self.indices.clear();
|
||||
self.entries.clear();
|
||||
}
|
||||
|
||||
pub(crate) fn truncate(&mut self, len: usize) {
|
||||
if len < self.len() {
|
||||
self.erase_indices(len, self.entries.len());
|
||||
self.entries.truncate(len);
|
||||
}
|
||||
}
|
||||
|
||||
pub(crate) fn drain<R>(&mut self, range: R) -> Drain<'_, Bucket<K, V>>
|
||||
where
|
||||
R: RangeBounds<usize>,
|
||||
{
|
||||
let range = simplify_range(range, self.entries.len());
|
||||
self.erase_indices(range.start, range.end);
|
||||
self.entries.drain(range)
|
||||
}
|
||||
|
||||
#[cfg(feature = "rayon")]
|
||||
pub(crate) fn par_drain<R>(&mut self, range: R) -> rayon::vec::Drain<'_, Bucket<K, V>>
|
||||
where
|
||||
K: Send,
|
||||
V: Send,
|
||||
R: RangeBounds<usize>,
|
||||
{
|
||||
use rayon::iter::ParallelDrainRange;
|
||||
let range = simplify_range(range, self.entries.len());
|
||||
self.erase_indices(range.start, range.end);
|
||||
self.entries.par_drain(range)
|
||||
}
|
||||
|
||||
pub(crate) fn split_off(&mut self, at: usize) -> Self {
|
||||
assert!(at <= self.entries.len());
|
||||
self.erase_indices(at, self.entries.len());
|
||||
let entries = self.entries.split_off(at);
|
||||
|
||||
let mut indices = RawTable::with_capacity(entries.len());
|
||||
raw::insert_bulk_no_grow(&mut indices, &entries);
|
||||
Self { indices, entries }
|
||||
}
|
||||
|
||||
/// Reserve capacity for `additional` more key-value pairs.
|
||||
pub(crate) fn reserve(&mut self, additional: usize) {
|
||||
self.indices.reserve(additional, get_hash(&self.entries));
|
||||
self.reserve_entries();
|
||||
}
|
||||
|
||||
/// Reserve entries capacity to match the indices
|
||||
fn reserve_entries(&mut self) {
|
||||
let additional = self.indices.capacity() - self.entries.len();
|
||||
self.entries.reserve_exact(additional);
|
||||
}
|
||||
|
||||
/// Shrink the capacity of the map with a lower bound
|
||||
pub(crate) fn shrink_to(&mut self, min_capacity: usize) {
|
||||
self.indices
|
||||
.shrink_to(min_capacity, get_hash(&self.entries));
|
||||
self.entries.shrink_to(min_capacity);
|
||||
}
|
||||
|
||||
/// Remove the last key-value pair
|
||||
pub(crate) fn pop(&mut self) -> Option<(K, V)> {
|
||||
if let Some(entry) = self.entries.pop() {
|
||||
let last = self.entries.len();
|
||||
erase_index(&mut self.indices, entry.hash, last);
|
||||
Some((entry.key, entry.value))
|
||||
} else {
|
||||
None
|
||||
}
|
||||
}
|
||||
|
||||
/// Append a key-value pair, *without* checking whether it already exists,
|
||||
/// and return the pair's new index.
|
||||
fn push(&mut self, hash: HashValue, key: K, value: V) -> usize {
|
||||
let i = self.entries.len();
|
||||
self.indices.insert(hash.get(), i, get_hash(&self.entries));
|
||||
if i == self.entries.capacity() {
|
||||
// Reserve our own capacity synced to the indices,
|
||||
// rather than letting `Vec::push` just double it.
|
||||
self.reserve_entries();
|
||||
}
|
||||
self.entries.push(Bucket { hash, key, value });
|
||||
i
|
||||
}
|
||||
|
||||
/// Return the index in `entries` where an equivalent key can be found
|
||||
pub(crate) fn get_index_of<Q>(&self, hash: HashValue, key: &Q) -> Option<usize>
|
||||
where
|
||||
Q: ?Sized + Equivalent<K>,
|
||||
{
|
||||
let eq = equivalent(key, &self.entries);
|
||||
self.indices.get(hash.get(), eq).copied()
|
||||
}
|
||||
|
||||
pub(crate) fn insert_full(&mut self, hash: HashValue, key: K, value: V) -> (usize, Option<V>)
|
||||
where
|
||||
K: Eq,
|
||||
{
|
||||
match self.get_index_of(hash, &key) {
|
||||
Some(i) => (i, Some(replace(&mut self.entries[i].value, value))),
|
||||
None => (self.push(hash, key, value), None),
|
||||
}
|
||||
}
|
||||
|
||||
/// Remove an entry by shifting all entries that follow it
|
||||
pub(crate) fn shift_remove_full<Q>(&mut self, hash: HashValue, key: &Q) -> Option<(usize, K, V)>
|
||||
where
|
||||
Q: ?Sized + Equivalent<K>,
|
||||
{
|
||||
let eq = equivalent(key, &self.entries);
|
||||
match self.indices.remove_entry(hash.get(), eq) {
|
||||
Some(index) => {
|
||||
let (key, value) = self.shift_remove_finish(index);
|
||||
Some((index, key, value))
|
||||
}
|
||||
None => None,
|
||||
}
|
||||
}
|
||||
|
||||
/// Remove an entry by shifting all entries that follow it
|
||||
pub(crate) fn shift_remove_index(&mut self, index: usize) -> Option<(K, V)> {
|
||||
match self.entries.get(index) {
|
||||
Some(entry) => {
|
||||
erase_index(&mut self.indices, entry.hash, index);
|
||||
Some(self.shift_remove_finish(index))
|
||||
}
|
||||
None => None,
|
||||
}
|
||||
}
|
||||
|
||||
/// Remove an entry by shifting all entries that follow it
|
||||
///
|
||||
/// The index should already be removed from `self.indices`.
|
||||
fn shift_remove_finish(&mut self, index: usize) -> (K, V) {
|
||||
// Correct indices that point to the entries that followed the removed entry.
|
||||
self.decrement_indices(index + 1, self.entries.len());
|
||||
|
||||
// Use Vec::remove to actually remove the entry.
|
||||
let entry = self.entries.remove(index);
|
||||
(entry.key, entry.value)
|
||||
}
|
||||
|
||||
/// Decrement all indices in the range `start..end`.
|
||||
///
|
||||
/// The index `start - 1` should not exist in `self.indices`.
|
||||
/// All entries should still be in their original positions.
|
||||
fn decrement_indices(&mut self, start: usize, end: usize) {
|
||||
// Use a heuristic between a full sweep vs. a `find()` for every shifted item.
|
||||
let shifted_entries = &self.entries[start..end];
|
||||
if shifted_entries.len() > self.indices.buckets() / 2 {
|
||||
// Shift all indices in range.
|
||||
for i in self.indices_mut() {
|
||||
if start <= *i && *i < end {
|
||||
*i -= 1;
|
||||
}
|
||||
}
|
||||
} else {
|
||||
// Find each entry in range to shift its index.
|
||||
for (i, entry) in (start..end).zip(shifted_entries) {
|
||||
update_index(&mut self.indices, entry.hash, i, i - 1);
|
||||
}
|
||||
}
|
||||
}
|
||||
|
||||
/// Increment all indices in the range `start..end`.
|
||||
///
|
||||
/// The index `end` should not exist in `self.indices`.
|
||||
/// All entries should still be in their original positions.
|
||||
fn increment_indices(&mut self, start: usize, end: usize) {
|
||||
// Use a heuristic between a full sweep vs. a `find()` for every shifted item.
|
||||
let shifted_entries = &self.entries[start..end];
|
||||
if shifted_entries.len() > self.indices.buckets() / 2 {
|
||||
// Shift all indices in range.
|
||||
for i in self.indices_mut() {
|
||||
if start <= *i && *i < end {
|
||||
*i += 1;
|
||||
}
|
||||
}
|
||||
} else {
|
||||
// Find each entry in range to shift its index, updated in reverse so
|
||||
// we never have duplicated indices that might have a hash collision.
|
||||
for (i, entry) in (start..end).zip(shifted_entries).rev() {
|
||||
update_index(&mut self.indices, entry.hash, i, i + 1);
|
||||
}
|
||||
}
|
||||
}
|
||||
|
||||
pub(super) fn move_index(&mut self, from: usize, to: usize) {
|
||||
let from_hash = self.entries[from].hash;
|
||||
if from != to {
|
||||
// Use a sentinal index so other indices don't collide.
|
||||
update_index(&mut self.indices, from_hash, from, usize::MAX);
|
||||
|
||||
// Update all other indices and rotate the entry positions.
|
||||
if from < to {
|
||||
self.decrement_indices(from + 1, to + 1);
|
||||
self.entries[from..=to].rotate_left(1);
|
||||
} else if to < from {
|
||||
self.increment_indices(to, from);
|
||||
self.entries[to..=from].rotate_right(1);
|
||||
}
|
||||
|
||||
// Change the sentinal index to its final position.
|
||||
update_index(&mut self.indices, from_hash, usize::MAX, to);
|
||||
}
|
||||
}
|
||||
|
||||
/// Remove an entry by swapping it with the last
|
||||
pub(crate) fn swap_remove_full<Q>(&mut self, hash: HashValue, key: &Q) -> Option<(usize, K, V)>
|
||||
where
|
||||
Q: ?Sized + Equivalent<K>,
|
||||
{
|
||||
let eq = equivalent(key, &self.entries);
|
||||
match self.indices.remove_entry(hash.get(), eq) {
|
||||
Some(index) => {
|
||||
let (key, value) = self.swap_remove_finish(index);
|
||||
Some((index, key, value))
|
||||
}
|
||||
None => None,
|
||||
}
|
||||
}
|
||||
|
||||
/// Remove an entry by swapping it with the last
|
||||
pub(crate) fn swap_remove_index(&mut self, index: usize) -> Option<(K, V)> {
|
||||
match self.entries.get(index) {
|
||||
Some(entry) => {
|
||||
erase_index(&mut self.indices, entry.hash, index);
|
||||
Some(self.swap_remove_finish(index))
|
||||
}
|
||||
None => None,
|
||||
}
|
||||
}
|
||||
|
||||
/// Finish removing an entry by swapping it with the last
|
||||
///
|
||||
/// The index should already be removed from `self.indices`.
|
||||
fn swap_remove_finish(&mut self, index: usize) -> (K, V) {
|
||||
// use swap_remove, but then we need to update the index that points
|
||||
// to the other entry that has to move
|
||||
let entry = self.entries.swap_remove(index);
|
||||
|
||||
// correct index that points to the entry that had to swap places
|
||||
if let Some(entry) = self.entries.get(index) {
|
||||
// was not last element
|
||||
// examine new element in `index` and find it in indices
|
||||
let last = self.entries.len();
|
||||
update_index(&mut self.indices, entry.hash, last, index);
|
||||
}
|
||||
|
||||
(entry.key, entry.value)
|
||||
}
|
||||
|
||||
/// Erase `start..end` from `indices`, and shift `end..` indices down to `start..`
|
||||
///
|
||||
/// All of these items should still be at their original location in `entries`.
|
||||
/// This is used by `drain`, which will let `Vec::drain` do the work on `entries`.
|
||||
fn erase_indices(&mut self, start: usize, end: usize) {
|
||||
let (init, shifted_entries) = self.entries.split_at(end);
|
||||
let (start_entries, erased_entries) = init.split_at(start);
|
||||
|
||||
let erased = erased_entries.len();
|
||||
let shifted = shifted_entries.len();
|
||||
let half_capacity = self.indices.buckets() / 2;
|
||||
|
||||
// Use a heuristic between different strategies
|
||||
if erased == 0 {
|
||||
// Degenerate case, nothing to do
|
||||
} else if start + shifted < half_capacity && start < erased {
|
||||
// Reinsert everything, as there are few kept indices
|
||||
self.indices.clear();
|
||||
|
||||
// Reinsert stable indices, then shifted indices
|
||||
raw::insert_bulk_no_grow(&mut self.indices, start_entries);
|
||||
raw::insert_bulk_no_grow(&mut self.indices, shifted_entries);
|
||||
} else if erased + shifted < half_capacity {
|
||||
// Find each affected index, as there are few to adjust
|
||||
|
||||
// Find erased indices
|
||||
for (i, entry) in (start..).zip(erased_entries) {
|
||||
erase_index(&mut self.indices, entry.hash, i);
|
||||
}
|
||||
|
||||
// Find shifted indices
|
||||
for ((new, old), entry) in (start..).zip(end..).zip(shifted_entries) {
|
||||
update_index(&mut self.indices, entry.hash, old, new);
|
||||
}
|
||||
} else {
|
||||
// Sweep the whole table for adjustments
|
||||
self.erase_indices_sweep(start, end);
|
||||
}
|
||||
|
||||
debug_assert_eq!(self.indices.len(), start + shifted);
|
||||
}
|
||||
|
||||
pub(crate) fn retain_in_order<F>(&mut self, mut keep: F)
|
||||
where
|
||||
F: FnMut(&mut K, &mut V) -> bool,
|
||||
{
|
||||
// FIXME: This could use Vec::retain_mut with MSRV 1.61.
|
||||
// Like Vec::retain in self.entries, but with mutable K and V.
|
||||
// We swap-shift all the items we want to keep, truncate the rest,
|
||||
// then rebuild the raw hash table with the new indexes.
|
||||
let len = self.entries.len();
|
||||
let mut n_deleted = 0;
|
||||
for i in 0..len {
|
||||
let will_keep = {
|
||||
let entry = &mut self.entries[i];
|
||||
keep(&mut entry.key, &mut entry.value)
|
||||
};
|
||||
if !will_keep {
|
||||
n_deleted += 1;
|
||||
} else if n_deleted > 0 {
|
||||
self.entries.swap(i - n_deleted, i);
|
||||
}
|
||||
}
|
||||
if n_deleted > 0 {
|
||||
self.entries.truncate(len - n_deleted);
|
||||
self.rebuild_hash_table();
|
||||
}
|
||||
}
|
||||
|
||||
fn rebuild_hash_table(&mut self) {
|
||||
self.indices.clear();
|
||||
raw::insert_bulk_no_grow(&mut self.indices, &self.entries);
|
||||
}
|
||||
|
||||
pub(crate) fn reverse(&mut self) {
|
||||
self.entries.reverse();
|
||||
|
||||
// No need to save hash indices, can easily calculate what they should
|
||||
// be, given that this is an in-place reversal.
|
||||
let len = self.entries.len();
|
||||
for i in self.indices_mut() {
|
||||
*i = len - *i - 1;
|
||||
}
|
||||
}
|
||||
}
|
||||
|
||||
/// Entry for an existing key-value pair or a vacant location to
|
||||
/// insert one.
|
||||
pub enum Entry<'a, K, V> {
|
||||
/// Existing slot with equivalent key.
|
||||
Occupied(OccupiedEntry<'a, K, V>),
|
||||
/// Vacant slot (no equivalent key in the map).
|
||||
Vacant(VacantEntry<'a, K, V>),
|
||||
}
|
||||
|
||||
impl<'a, K, V> Entry<'a, K, V> {
|
||||
/// Inserts the given default value in the entry if it is vacant and returns a mutable
|
||||
/// reference to it. Otherwise a mutable reference to an already existent value is returned.
|
||||
///
|
||||
/// Computes in **O(1)** time (amortized average).
|
||||
pub fn or_insert(self, default: V) -> &'a mut V {
|
||||
match self {
|
||||
Entry::Occupied(entry) => entry.into_mut(),
|
||||
Entry::Vacant(entry) => entry.insert(default),
|
||||
}
|
||||
}
|
||||
|
||||
/// Inserts the result of the `call` function in the entry if it is vacant and returns a mutable
|
||||
/// reference to it. Otherwise a mutable reference to an already existent value is returned.
|
||||
///
|
||||
/// Computes in **O(1)** time (amortized average).
|
||||
pub fn or_insert_with<F>(self, call: F) -> &'a mut V
|
||||
where
|
||||
F: FnOnce() -> V,
|
||||
{
|
||||
match self {
|
||||
Entry::Occupied(entry) => entry.into_mut(),
|
||||
Entry::Vacant(entry) => entry.insert(call()),
|
||||
}
|
||||
}
|
||||
|
||||
/// Inserts the result of the `call` function with a reference to the entry's key if it is
|
||||
/// vacant, and returns a mutable reference to the new value. Otherwise a mutable reference to
|
||||
/// an already existent value is returned.
|
||||
///
|
||||
/// Computes in **O(1)** time (amortized average).
|
||||
pub fn or_insert_with_key<F>(self, call: F) -> &'a mut V
|
||||
where
|
||||
F: FnOnce(&K) -> V,
|
||||
{
|
||||
match self {
|
||||
Entry::Occupied(entry) => entry.into_mut(),
|
||||
Entry::Vacant(entry) => {
|
||||
let value = call(&entry.key);
|
||||
entry.insert(value)
|
||||
}
|
||||
}
|
||||
}
|
||||
|
||||
/// Gets a reference to the entry's key, either within the map if occupied,
|
||||
/// or else the new key that was used to find the entry.
|
||||
pub fn key(&self) -> &K {
|
||||
match *self {
|
||||
Entry::Occupied(ref entry) => entry.key(),
|
||||
Entry::Vacant(ref entry) => entry.key(),
|
||||
}
|
||||
}
|
||||
|
||||
/// Return the index where the key-value pair exists or will be inserted.
|
||||
pub fn index(&self) -> usize {
|
||||
match *self {
|
||||
Entry::Occupied(ref entry) => entry.index(),
|
||||
Entry::Vacant(ref entry) => entry.index(),
|
||||
}
|
||||
}
|
||||
|
||||
/// Modifies the entry if it is occupied.
|
||||
pub fn and_modify<F>(self, f: F) -> Self
|
||||
where
|
||||
F: FnOnce(&mut V),
|
||||
{
|
||||
match self {
|
||||
Entry::Occupied(mut o) => {
|
||||
f(o.get_mut());
|
||||
Entry::Occupied(o)
|
||||
}
|
||||
x => x,
|
||||
}
|
||||
}
|
||||
|
||||
/// Inserts a default-constructed value in the entry if it is vacant and returns a mutable
|
||||
/// reference to it. Otherwise a mutable reference to an already existent value is returned.
|
||||
///
|
||||
/// Computes in **O(1)** time (amortized average).
|
||||
pub fn or_default(self) -> &'a mut V
|
||||
where
|
||||
V: Default,
|
||||
{
|
||||
match self {
|
||||
Entry::Occupied(entry) => entry.into_mut(),
|
||||
Entry::Vacant(entry) => entry.insert(V::default()),
|
||||
}
|
||||
}
|
||||
}
|
||||
|
||||
impl<K: fmt::Debug, V: fmt::Debug> fmt::Debug for Entry<'_, K, V> {
|
||||
fn fmt(&self, f: &mut fmt::Formatter<'_>) -> fmt::Result {
|
||||
match *self {
|
||||
Entry::Vacant(ref v) => f.debug_tuple(stringify!(Entry)).field(v).finish(),
|
||||
Entry::Occupied(ref o) => f.debug_tuple(stringify!(Entry)).field(o).finish(),
|
||||
}
|
||||
}
|
||||
}
|
||||
|
||||
pub use self::raw::OccupiedEntry;
|
||||
|
||||
// Extra methods that don't threaten the unsafe encapsulation.
|
||||
impl<K, V> OccupiedEntry<'_, K, V> {
|
||||
/// Sets the value of the entry to `value`, and returns the entry's old value.
|
||||
pub fn insert(&mut self, value: V) -> V {
|
||||
replace(self.get_mut(), value)
|
||||
}
|
||||
|
||||
/// Remove the key, value pair stored in the map for this entry, and return the value.
|
||||
///
|
||||
/// **NOTE:** This is equivalent to `.swap_remove()`.
|
||||
pub fn remove(self) -> V {
|
||||
self.swap_remove()
|
||||
}
|
||||
|
||||
/// Remove the key, value pair stored in the map for this entry, and return the value.
|
||||
///
|
||||
/// Like `Vec::swap_remove`, the pair is removed by swapping it with the
|
||||
/// last element of the map and popping it off. **This perturbs
|
||||
/// the position of what used to be the last element!**
|
||||
///
|
||||
/// Computes in **O(1)** time (average).
|
||||
pub fn swap_remove(self) -> V {
|
||||
self.swap_remove_entry().1
|
||||
}
|
||||
|
||||
/// Remove the key, value pair stored in the map for this entry, and return the value.
|
||||
///
|
||||
/// Like `Vec::remove`, the pair is removed by shifting all of the
|
||||
/// elements that follow it, preserving their relative order.
|
||||
/// **This perturbs the index of all of those elements!**
|
||||
///
|
||||
/// Computes in **O(n)** time (average).
|
||||
pub fn shift_remove(self) -> V {
|
||||
self.shift_remove_entry().1
|
||||
}
|
||||
|
||||
/// Remove and return the key, value pair stored in the map for this entry
|
||||
///
|
||||
/// **NOTE:** This is equivalent to `.swap_remove_entry()`.
|
||||
pub fn remove_entry(self) -> (K, V) {
|
||||
self.swap_remove_entry()
|
||||
}
|
||||
}
|
||||
|
||||
impl<K: fmt::Debug, V: fmt::Debug> fmt::Debug for OccupiedEntry<'_, K, V> {
|
||||
fn fmt(&self, f: &mut fmt::Formatter<'_>) -> fmt::Result {
|
||||
f.debug_struct(stringify!(OccupiedEntry))
|
||||
.field("key", self.key())
|
||||
.field("value", self.get())
|
||||
.finish()
|
||||
}
|
||||
}
|
||||
|
||||
/// A view into a vacant entry in a `IndexMap`.
|
||||
/// It is part of the [`Entry`] enum.
|
||||
///
|
||||
/// [`Entry`]: enum.Entry.html
|
||||
pub struct VacantEntry<'a, K, V> {
|
||||
map: &'a mut IndexMapCore<K, V>,
|
||||
hash: HashValue,
|
||||
key: K,
|
||||
}
|
||||
|
||||
impl<'a, K, V> VacantEntry<'a, K, V> {
|
||||
/// Gets a reference to the key that was used to find the entry.
|
||||
pub fn key(&self) -> &K {
|
||||
&self.key
|
||||
}
|
||||
|
||||
/// Takes ownership of the key, leaving the entry vacant.
|
||||
pub fn into_key(self) -> K {
|
||||
self.key
|
||||
}
|
||||
|
||||
/// Return the index where the key-value pair will be inserted.
|
||||
pub fn index(&self) -> usize {
|
||||
self.map.len()
|
||||
}
|
||||
|
||||
/// Inserts the entry's key and the given value into the map, and returns a mutable reference
|
||||
/// to the value.
|
||||
pub fn insert(self, value: V) -> &'a mut V {
|
||||
let i = self.map.push(self.hash, self.key, value);
|
||||
&mut self.map.entries[i].value
|
||||
}
|
||||
}
|
||||
|
||||
impl<K: fmt::Debug, V> fmt::Debug for VacantEntry<'_, K, V> {
|
||||
fn fmt(&self, f: &mut fmt::Formatter<'_>) -> fmt::Result {
|
||||
f.debug_tuple(stringify!(VacantEntry))
|
||||
.field(self.key())
|
||||
.finish()
|
||||
}
|
||||
}
|
||||
|
||||
#[test]
|
||||
fn assert_send_sync() {
|
||||
fn assert_send_sync<T: Send + Sync>() {}
|
||||
assert_send_sync::<IndexMapCore<i32, i32>>();
|
||||
assert_send_sync::<Entry<'_, i32, i32>>();
|
||||
}
|
||||
191
clamav/libclamav_rust/.cargo/vendor/indexmap/src/map/core/raw.rs
vendored
Normal file
191
clamav/libclamav_rust/.cargo/vendor/indexmap/src/map/core/raw.rs
vendored
Normal file
@@ -0,0 +1,191 @@
|
||||
#![allow(unsafe_code)]
|
||||
//! This module encapsulates the `unsafe` access to `hashbrown::raw::RawTable`,
|
||||
//! mostly in dealing with its bucket "pointers".
|
||||
|
||||
use super::{equivalent, Bucket, Entry, HashValue, IndexMapCore, VacantEntry};
|
||||
use core::fmt;
|
||||
use core::mem::replace;
|
||||
use hashbrown::raw::RawTable;
|
||||
|
||||
type RawBucket = hashbrown::raw::Bucket<usize>;
|
||||
|
||||
/// Inserts many entries into a raw table without reallocating.
|
||||
///
|
||||
/// ***Panics*** if there is not sufficient capacity already.
|
||||
pub(super) fn insert_bulk_no_grow<K, V>(indices: &mut RawTable<usize>, entries: &[Bucket<K, V>]) {
|
||||
assert!(indices.capacity() - indices.len() >= entries.len());
|
||||
for entry in entries {
|
||||
// SAFETY: we asserted that sufficient capacity exists for all entries.
|
||||
unsafe {
|
||||
indices.insert_no_grow(entry.hash.get(), indices.len());
|
||||
}
|
||||
}
|
||||
}
|
||||
|
||||
pub(super) struct DebugIndices<'a>(pub &'a RawTable<usize>);
|
||||
impl fmt::Debug for DebugIndices<'_> {
|
||||
fn fmt(&self, f: &mut fmt::Formatter<'_>) -> fmt::Result {
|
||||
// SAFETY: we're not letting any of the buckets escape this function
|
||||
let indices = unsafe { self.0.iter().map(|raw_bucket| raw_bucket.read()) };
|
||||
f.debug_list().entries(indices).finish()
|
||||
}
|
||||
}
|
||||
|
||||
impl<K, V> IndexMapCore<K, V> {
|
||||
/// Sweep the whole table to erase indices start..end
|
||||
pub(super) fn erase_indices_sweep(&mut self, start: usize, end: usize) {
|
||||
// SAFETY: we're not letting any of the buckets escape this function
|
||||
unsafe {
|
||||
let offset = end - start;
|
||||
for bucket in self.indices.iter() {
|
||||
let i = bucket.read();
|
||||
if i >= end {
|
||||
bucket.write(i - offset);
|
||||
} else if i >= start {
|
||||
self.indices.erase(bucket);
|
||||
}
|
||||
}
|
||||
}
|
||||
}
|
||||
|
||||
pub(crate) fn entry(&mut self, hash: HashValue, key: K) -> Entry<'_, K, V>
|
||||
where
|
||||
K: Eq,
|
||||
{
|
||||
let eq = equivalent(&key, &self.entries);
|
||||
match self.indices.find(hash.get(), eq) {
|
||||
// SAFETY: The entry is created with a live raw bucket, at the same time
|
||||
// we have a &mut reference to the map, so it can not be modified further.
|
||||
Some(raw_bucket) => Entry::Occupied(OccupiedEntry {
|
||||
map: self,
|
||||
raw_bucket,
|
||||
key,
|
||||
}),
|
||||
None => Entry::Vacant(VacantEntry {
|
||||
map: self,
|
||||
hash,
|
||||
key,
|
||||
}),
|
||||
}
|
||||
}
|
||||
|
||||
pub(super) fn indices_mut(&mut self) -> impl Iterator<Item = &mut usize> {
|
||||
// SAFETY: we're not letting any of the buckets escape this function,
|
||||
// only the item references that are appropriately bound to `&mut self`.
|
||||
unsafe { self.indices.iter().map(|bucket| bucket.as_mut()) }
|
||||
}
|
||||
|
||||
/// Return the raw bucket for the given index
|
||||
fn find_index(&self, index: usize) -> RawBucket {
|
||||
// We'll get a "nice" bounds-check from indexing `self.entries`,
|
||||
// and then we expect to find it in the table as well.
|
||||
let hash = self.entries[index].hash.get();
|
||||
self.indices
|
||||
.find(hash, move |&i| i == index)
|
||||
.expect("index not found")
|
||||
}
|
||||
|
||||
pub(crate) fn swap_indices(&mut self, a: usize, b: usize) {
|
||||
// SAFETY: Can't take two `get_mut` references from one table, so we
|
||||
// must use raw buckets to do the swap. This is still safe because we
|
||||
// are locally sure they won't dangle, and we write them individually.
|
||||
unsafe {
|
||||
let raw_bucket_a = self.find_index(a);
|
||||
let raw_bucket_b = self.find_index(b);
|
||||
raw_bucket_a.write(b);
|
||||
raw_bucket_b.write(a);
|
||||
}
|
||||
self.entries.swap(a, b);
|
||||
}
|
||||
}
|
||||
|
||||
/// A view into an occupied entry in a `IndexMap`.
|
||||
/// It is part of the [`Entry`] enum.
|
||||
///
|
||||
/// [`Entry`]: enum.Entry.html
|
||||
// SAFETY: The lifetime of the map reference also constrains the raw bucket,
|
||||
// which is essentially a raw pointer into the map indices.
|
||||
pub struct OccupiedEntry<'a, K, V> {
|
||||
map: &'a mut IndexMapCore<K, V>,
|
||||
raw_bucket: RawBucket,
|
||||
key: K,
|
||||
}
|
||||
|
||||
// `hashbrown::raw::Bucket` is only `Send`, not `Sync`.
|
||||
// SAFETY: `&self` only accesses the bucket to read it.
|
||||
unsafe impl<K: Sync, V: Sync> Sync for OccupiedEntry<'_, K, V> {}
|
||||
|
||||
// The parent module also adds methods that don't threaten the unsafe encapsulation.
|
||||
impl<'a, K, V> OccupiedEntry<'a, K, V> {
|
||||
/// Gets a reference to the entry's key in the map.
|
||||
///
|
||||
/// Note that this is not the key that was used to find the entry. There may be an observable
|
||||
/// difference if the key type has any distinguishing features outside of `Hash` and `Eq`, like
|
||||
/// extra fields or the memory address of an allocation.
|
||||
pub fn key(&self) -> &K {
|
||||
&self.map.entries[self.index()].key
|
||||
}
|
||||
|
||||
/// Gets a reference to the entry's value in the map.
|
||||
pub fn get(&self) -> &V {
|
||||
&self.map.entries[self.index()].value
|
||||
}
|
||||
|
||||
/// Gets a mutable reference to the entry's value in the map.
|
||||
///
|
||||
/// If you need a reference which may outlive the destruction of the
|
||||
/// `Entry` value, see `into_mut`.
|
||||
pub fn get_mut(&mut self) -> &mut V {
|
||||
let index = self.index();
|
||||
&mut self.map.entries[index].value
|
||||
}
|
||||
|
||||
/// Put the new key in the occupied entry's key slot
|
||||
pub(crate) fn replace_key(self) -> K {
|
||||
let index = self.index();
|
||||
let old_key = &mut self.map.entries[index].key;
|
||||
replace(old_key, self.key)
|
||||
}
|
||||
|
||||
/// Return the index of the key-value pair
|
||||
#[inline]
|
||||
pub fn index(&self) -> usize {
|
||||
// SAFETY: we have &mut map keep keeping the bucket stable
|
||||
unsafe { self.raw_bucket.read() }
|
||||
}
|
||||
|
||||
/// Converts into a mutable reference to the entry's value in the map,
|
||||
/// with a lifetime bound to the map itself.
|
||||
pub fn into_mut(self) -> &'a mut V {
|
||||
let index = self.index();
|
||||
&mut self.map.entries[index].value
|
||||
}
|
||||
|
||||
/// Remove and return the key, value pair stored in the map for this entry
|
||||
///
|
||||
/// Like `Vec::swap_remove`, the pair is removed by swapping it with the
|
||||
/// last element of the map and popping it off. **This perturbs
|
||||
/// the position of what used to be the last element!**
|
||||
///
|
||||
/// Computes in **O(1)** time (average).
|
||||
pub fn swap_remove_entry(self) -> (K, V) {
|
||||
// SAFETY: This is safe because it can only happen once (self is consumed)
|
||||
// and map.indices have not been modified since entry construction
|
||||
let index = unsafe { self.map.indices.remove(self.raw_bucket) };
|
||||
self.map.swap_remove_finish(index)
|
||||
}
|
||||
|
||||
/// Remove and return the key, value pair stored in the map for this entry
|
||||
///
|
||||
/// Like `Vec::remove`, the pair is removed by shifting all of the
|
||||
/// elements that follow it, preserving their relative order.
|
||||
/// **This perturbs the index of all of those elements!**
|
||||
///
|
||||
/// Computes in **O(n)** time (average).
|
||||
pub fn shift_remove_entry(self) -> (K, V) {
|
||||
// SAFETY: This is safe because it can only happen once (self is consumed)
|
||||
// and map.indices have not been modified since entry construction
|
||||
let index = unsafe { self.map.indices.remove(self.raw_bucket) };
|
||||
self.map.shift_remove_finish(index)
|
||||
}
|
||||
}
|
||||
75
clamav/libclamav_rust/.cargo/vendor/indexmap/src/mutable_keys.rs
vendored
Normal file
75
clamav/libclamav_rust/.cargo/vendor/indexmap/src/mutable_keys.rs
vendored
Normal file
@@ -0,0 +1,75 @@
|
||||
use core::hash::{BuildHasher, Hash};
|
||||
|
||||
use super::{Equivalent, IndexMap};
|
||||
|
||||
pub struct PrivateMarker {}
|
||||
|
||||
/// Opt-in mutable access to keys.
|
||||
///
|
||||
/// These methods expose `&mut K`, mutable references to the key as it is stored
|
||||
/// in the map.
|
||||
/// You are allowed to modify the keys in the hashmap **if the modification
|
||||
/// does not change the key’s hash and equality**.
|
||||
///
|
||||
/// If keys are modified erroneously, you can no longer look them up.
|
||||
/// This is sound (memory safe) but a logical error hazard (just like
|
||||
/// implementing PartialEq, Eq, or Hash incorrectly would be).
|
||||
///
|
||||
/// `use` this trait to enable its methods for `IndexMap`.
|
||||
pub trait MutableKeys {
|
||||
type Key;
|
||||
type Value;
|
||||
|
||||
/// Return item index, mutable reference to key and value
|
||||
fn get_full_mut2<Q: ?Sized>(
|
||||
&mut self,
|
||||
key: &Q,
|
||||
) -> Option<(usize, &mut Self::Key, &mut Self::Value)>
|
||||
where
|
||||
Q: Hash + Equivalent<Self::Key>;
|
||||
|
||||
/// Scan through each key-value pair in the map and keep those where the
|
||||
/// closure `keep` returns `true`.
|
||||
///
|
||||
/// The elements are visited in order, and remaining elements keep their
|
||||
/// order.
|
||||
///
|
||||
/// Computes in **O(n)** time (average).
|
||||
fn retain2<F>(&mut self, keep: F)
|
||||
where
|
||||
F: FnMut(&mut Self::Key, &mut Self::Value) -> bool;
|
||||
|
||||
/// This method is not useful in itself – it is there to “seal” the trait
|
||||
/// for external implementation, so that we can add methods without
|
||||
/// causing breaking changes.
|
||||
fn __private_marker(&self) -> PrivateMarker;
|
||||
}
|
||||
|
||||
/// Opt-in mutable access to keys.
|
||||
///
|
||||
/// See [`MutableKeys`](trait.MutableKeys.html) for more information.
|
||||
impl<K, V, S> MutableKeys for IndexMap<K, V, S>
|
||||
where
|
||||
K: Eq + Hash,
|
||||
S: BuildHasher,
|
||||
{
|
||||
type Key = K;
|
||||
type Value = V;
|
||||
fn get_full_mut2<Q: ?Sized>(&mut self, key: &Q) -> Option<(usize, &mut K, &mut V)>
|
||||
where
|
||||
Q: Hash + Equivalent<K>,
|
||||
{
|
||||
self.get_full_mut2_impl(key)
|
||||
}
|
||||
|
||||
fn retain2<F>(&mut self, keep: F)
|
||||
where
|
||||
F: FnMut(&mut K, &mut V) -> bool,
|
||||
{
|
||||
self.retain_mut(keep)
|
||||
}
|
||||
|
||||
fn __private_marker(&self) -> PrivateMarker {
|
||||
PrivateMarker {}
|
||||
}
|
||||
}
|
||||
583
clamav/libclamav_rust/.cargo/vendor/indexmap/src/rayon/map.rs
vendored
Normal file
583
clamav/libclamav_rust/.cargo/vendor/indexmap/src/rayon/map.rs
vendored
Normal file
@@ -0,0 +1,583 @@
|
||||
//! Parallel iterator types for `IndexMap` with [rayon](https://docs.rs/rayon/1.0/rayon).
|
||||
//!
|
||||
//! You will rarely need to interact with this module directly unless you need to name one of the
|
||||
//! iterator types.
|
||||
//!
|
||||
//! Requires crate feature `"rayon"`
|
||||
|
||||
use super::collect;
|
||||
use rayon::iter::plumbing::{Consumer, ProducerCallback, UnindexedConsumer};
|
||||
use rayon::prelude::*;
|
||||
|
||||
use crate::vec::Vec;
|
||||
use core::cmp::Ordering;
|
||||
use core::fmt;
|
||||
use core::hash::{BuildHasher, Hash};
|
||||
use core::ops::RangeBounds;
|
||||
|
||||
use crate::Bucket;
|
||||
use crate::Entries;
|
||||
use crate::IndexMap;
|
||||
|
||||
/// Requires crate feature `"rayon"`.
|
||||
impl<K, V, S> IntoParallelIterator for IndexMap<K, V, S>
|
||||
where
|
||||
K: Send,
|
||||
V: Send,
|
||||
{
|
||||
type Item = (K, V);
|
||||
type Iter = IntoParIter<K, V>;
|
||||
|
||||
fn into_par_iter(self) -> Self::Iter {
|
||||
IntoParIter {
|
||||
entries: self.into_entries(),
|
||||
}
|
||||
}
|
||||
}
|
||||
|
||||
/// A parallel owning iterator over the entries of a `IndexMap`.
|
||||
///
|
||||
/// This `struct` is created by the [`into_par_iter`] method on [`IndexMap`]
|
||||
/// (provided by rayon's `IntoParallelIterator` trait). See its documentation for more.
|
||||
///
|
||||
/// [`into_par_iter`]: ../struct.IndexMap.html#method.into_par_iter
|
||||
/// [`IndexMap`]: ../struct.IndexMap.html
|
||||
pub struct IntoParIter<K, V> {
|
||||
entries: Vec<Bucket<K, V>>,
|
||||
}
|
||||
|
||||
impl<K: fmt::Debug, V: fmt::Debug> fmt::Debug for IntoParIter<K, V> {
|
||||
fn fmt(&self, f: &mut fmt::Formatter<'_>) -> fmt::Result {
|
||||
let iter = self.entries.iter().map(Bucket::refs);
|
||||
f.debug_list().entries(iter).finish()
|
||||
}
|
||||
}
|
||||
|
||||
impl<K: Send, V: Send> ParallelIterator for IntoParIter<K, V> {
|
||||
type Item = (K, V);
|
||||
|
||||
parallel_iterator_methods!(Bucket::key_value);
|
||||
}
|
||||
|
||||
impl<K: Send, V: Send> IndexedParallelIterator for IntoParIter<K, V> {
|
||||
indexed_parallel_iterator_methods!(Bucket::key_value);
|
||||
}
|
||||
|
||||
/// Requires crate feature `"rayon"`.
|
||||
impl<'a, K, V, S> IntoParallelIterator for &'a IndexMap<K, V, S>
|
||||
where
|
||||
K: Sync,
|
||||
V: Sync,
|
||||
{
|
||||
type Item = (&'a K, &'a V);
|
||||
type Iter = ParIter<'a, K, V>;
|
||||
|
||||
fn into_par_iter(self) -> Self::Iter {
|
||||
ParIter {
|
||||
entries: self.as_entries(),
|
||||
}
|
||||
}
|
||||
}
|
||||
|
||||
/// A parallel iterator over the entries of a `IndexMap`.
|
||||
///
|
||||
/// This `struct` is created by the [`par_iter`] method on [`IndexMap`]
|
||||
/// (provided by rayon's `IntoParallelRefIterator` trait). See its documentation for more.
|
||||
///
|
||||
/// [`par_iter`]: ../struct.IndexMap.html#method.par_iter
|
||||
/// [`IndexMap`]: ../struct.IndexMap.html
|
||||
pub struct ParIter<'a, K, V> {
|
||||
entries: &'a [Bucket<K, V>],
|
||||
}
|
||||
|
||||
impl<K, V> Clone for ParIter<'_, K, V> {
|
||||
fn clone(&self) -> Self {
|
||||
ParIter { ..*self }
|
||||
}
|
||||
}
|
||||
|
||||
impl<K: fmt::Debug, V: fmt::Debug> fmt::Debug for ParIter<'_, K, V> {
|
||||
fn fmt(&self, f: &mut fmt::Formatter<'_>) -> fmt::Result {
|
||||
let iter = self.entries.iter().map(Bucket::refs);
|
||||
f.debug_list().entries(iter).finish()
|
||||
}
|
||||
}
|
||||
|
||||
impl<'a, K: Sync, V: Sync> ParallelIterator for ParIter<'a, K, V> {
|
||||
type Item = (&'a K, &'a V);
|
||||
|
||||
parallel_iterator_methods!(Bucket::refs);
|
||||
}
|
||||
|
||||
impl<K: Sync, V: Sync> IndexedParallelIterator for ParIter<'_, K, V> {
|
||||
indexed_parallel_iterator_methods!(Bucket::refs);
|
||||
}
|
||||
|
||||
/// Requires crate feature `"rayon"`.
|
||||
impl<'a, K, V, S> IntoParallelIterator for &'a mut IndexMap<K, V, S>
|
||||
where
|
||||
K: Sync + Send,
|
||||
V: Send,
|
||||
{
|
||||
type Item = (&'a K, &'a mut V);
|
||||
type Iter = ParIterMut<'a, K, V>;
|
||||
|
||||
fn into_par_iter(self) -> Self::Iter {
|
||||
ParIterMut {
|
||||
entries: self.as_entries_mut(),
|
||||
}
|
||||
}
|
||||
}
|
||||
|
||||
/// A parallel mutable iterator over the entries of a `IndexMap`.
|
||||
///
|
||||
/// This `struct` is created by the [`par_iter_mut`] method on [`IndexMap`]
|
||||
/// (provided by rayon's `IntoParallelRefMutIterator` trait). See its documentation for more.
|
||||
///
|
||||
/// [`par_iter_mut`]: ../struct.IndexMap.html#method.par_iter_mut
|
||||
/// [`IndexMap`]: ../struct.IndexMap.html
|
||||
pub struct ParIterMut<'a, K, V> {
|
||||
entries: &'a mut [Bucket<K, V>],
|
||||
}
|
||||
|
||||
impl<K: fmt::Debug, V: fmt::Debug> fmt::Debug for ParIterMut<'_, K, V> {
|
||||
fn fmt(&self, f: &mut fmt::Formatter<'_>) -> fmt::Result {
|
||||
let iter = self.entries.iter().map(Bucket::refs);
|
||||
f.debug_list().entries(iter).finish()
|
||||
}
|
||||
}
|
||||
|
||||
impl<'a, K: Sync + Send, V: Send> ParallelIterator for ParIterMut<'a, K, V> {
|
||||
type Item = (&'a K, &'a mut V);
|
||||
|
||||
parallel_iterator_methods!(Bucket::ref_mut);
|
||||
}
|
||||
|
||||
impl<K: Sync + Send, V: Send> IndexedParallelIterator for ParIterMut<'_, K, V> {
|
||||
indexed_parallel_iterator_methods!(Bucket::ref_mut);
|
||||
}
|
||||
|
||||
/// Requires crate feature `"rayon"`.
|
||||
impl<'a, K, V, S> ParallelDrainRange<usize> for &'a mut IndexMap<K, V, S>
|
||||
where
|
||||
K: Send,
|
||||
V: Send,
|
||||
{
|
||||
type Item = (K, V);
|
||||
type Iter = ParDrain<'a, K, V>;
|
||||
|
||||
fn par_drain<R: RangeBounds<usize>>(self, range: R) -> Self::Iter {
|
||||
ParDrain {
|
||||
entries: self.core.par_drain(range),
|
||||
}
|
||||
}
|
||||
}
|
||||
|
||||
/// A parallel draining iterator over the entries of a `IndexMap`.
|
||||
///
|
||||
/// This `struct` is created by the [`par_drain`] method on [`IndexMap`]
|
||||
/// (provided by rayon's `ParallelDrainRange` trait). See its documentation for more.
|
||||
///
|
||||
/// [`par_drain`]: ../struct.IndexMap.html#method.par_drain
|
||||
/// [`IndexMap`]: ../struct.IndexMap.html
|
||||
pub struct ParDrain<'a, K: Send, V: Send> {
|
||||
entries: rayon::vec::Drain<'a, Bucket<K, V>>,
|
||||
}
|
||||
|
||||
impl<K: Send, V: Send> ParallelIterator for ParDrain<'_, K, V> {
|
||||
type Item = (K, V);
|
||||
|
||||
parallel_iterator_methods!(Bucket::key_value);
|
||||
}
|
||||
|
||||
impl<K: Send, V: Send> IndexedParallelIterator for ParDrain<'_, K, V> {
|
||||
indexed_parallel_iterator_methods!(Bucket::key_value);
|
||||
}
|
||||
|
||||
/// Parallel iterator methods and other parallel methods.
|
||||
///
|
||||
/// The following methods **require crate feature `"rayon"`**.
|
||||
///
|
||||
/// See also the `IntoParallelIterator` implementations.
|
||||
impl<K, V, S> IndexMap<K, V, S>
|
||||
where
|
||||
K: Sync,
|
||||
V: Sync,
|
||||
{
|
||||
/// Return a parallel iterator over the keys of the map.
|
||||
///
|
||||
/// While parallel iterators can process items in any order, their relative order
|
||||
/// in the map is still preserved for operations like `reduce` and `collect`.
|
||||
pub fn par_keys(&self) -> ParKeys<'_, K, V> {
|
||||
ParKeys {
|
||||
entries: self.as_entries(),
|
||||
}
|
||||
}
|
||||
|
||||
/// Return a parallel iterator over the values of the map.
|
||||
///
|
||||
/// While parallel iterators can process items in any order, their relative order
|
||||
/// in the map is still preserved for operations like `reduce` and `collect`.
|
||||
pub fn par_values(&self) -> ParValues<'_, K, V> {
|
||||
ParValues {
|
||||
entries: self.as_entries(),
|
||||
}
|
||||
}
|
||||
}
|
||||
|
||||
impl<K, V, S> IndexMap<K, V, S>
|
||||
where
|
||||
K: Hash + Eq + Sync,
|
||||
V: Sync,
|
||||
S: BuildHasher,
|
||||
{
|
||||
/// Returns `true` if `self` contains all of the same key-value pairs as `other`,
|
||||
/// regardless of each map's indexed order, determined in parallel.
|
||||
pub fn par_eq<V2, S2>(&self, other: &IndexMap<K, V2, S2>) -> bool
|
||||
where
|
||||
V: PartialEq<V2>,
|
||||
V2: Sync,
|
||||
S2: BuildHasher + Sync,
|
||||
{
|
||||
self.len() == other.len()
|
||||
&& self
|
||||
.par_iter()
|
||||
.all(move |(key, value)| other.get(key).map_or(false, |v| *value == *v))
|
||||
}
|
||||
}
|
||||
|
||||
/// A parallel iterator over the keys of a `IndexMap`.
|
||||
///
|
||||
/// This `struct` is created by the [`par_keys`] method on [`IndexMap`]. See its
|
||||
/// documentation for more.
|
||||
///
|
||||
/// [`par_keys`]: ../struct.IndexMap.html#method.par_keys
|
||||
/// [`IndexMap`]: ../struct.IndexMap.html
|
||||
pub struct ParKeys<'a, K, V> {
|
||||
entries: &'a [Bucket<K, V>],
|
||||
}
|
||||
|
||||
impl<K, V> Clone for ParKeys<'_, K, V> {
|
||||
fn clone(&self) -> Self {
|
||||
ParKeys { ..*self }
|
||||
}
|
||||
}
|
||||
|
||||
impl<K: fmt::Debug, V> fmt::Debug for ParKeys<'_, K, V> {
|
||||
fn fmt(&self, f: &mut fmt::Formatter<'_>) -> fmt::Result {
|
||||
let iter = self.entries.iter().map(Bucket::key_ref);
|
||||
f.debug_list().entries(iter).finish()
|
||||
}
|
||||
}
|
||||
|
||||
impl<'a, K: Sync, V: Sync> ParallelIterator for ParKeys<'a, K, V> {
|
||||
type Item = &'a K;
|
||||
|
||||
parallel_iterator_methods!(Bucket::key_ref);
|
||||
}
|
||||
|
||||
impl<K: Sync, V: Sync> IndexedParallelIterator for ParKeys<'_, K, V> {
|
||||
indexed_parallel_iterator_methods!(Bucket::key_ref);
|
||||
}
|
||||
|
||||
/// A parallel iterator over the values of a `IndexMap`.
|
||||
///
|
||||
/// This `struct` is created by the [`par_values`] method on [`IndexMap`]. See its
|
||||
/// documentation for more.
|
||||
///
|
||||
/// [`par_values`]: ../struct.IndexMap.html#method.par_values
|
||||
/// [`IndexMap`]: ../struct.IndexMap.html
|
||||
pub struct ParValues<'a, K, V> {
|
||||
entries: &'a [Bucket<K, V>],
|
||||
}
|
||||
|
||||
impl<K, V> Clone for ParValues<'_, K, V> {
|
||||
fn clone(&self) -> Self {
|
||||
ParValues { ..*self }
|
||||
}
|
||||
}
|
||||
|
||||
impl<K, V: fmt::Debug> fmt::Debug for ParValues<'_, K, V> {
|
||||
fn fmt(&self, f: &mut fmt::Formatter<'_>) -> fmt::Result {
|
||||
let iter = self.entries.iter().map(Bucket::value_ref);
|
||||
f.debug_list().entries(iter).finish()
|
||||
}
|
||||
}
|
||||
|
||||
impl<'a, K: Sync, V: Sync> ParallelIterator for ParValues<'a, K, V> {
|
||||
type Item = &'a V;
|
||||
|
||||
parallel_iterator_methods!(Bucket::value_ref);
|
||||
}
|
||||
|
||||
impl<K: Sync, V: Sync> IndexedParallelIterator for ParValues<'_, K, V> {
|
||||
indexed_parallel_iterator_methods!(Bucket::value_ref);
|
||||
}
|
||||
|
||||
/// Requires crate feature `"rayon"`.
|
||||
impl<K, V, S> IndexMap<K, V, S>
|
||||
where
|
||||
K: Send,
|
||||
V: Send,
|
||||
{
|
||||
/// Return a parallel iterator over mutable references to the values of the map
|
||||
///
|
||||
/// While parallel iterators can process items in any order, their relative order
|
||||
/// in the map is still preserved for operations like `reduce` and `collect`.
|
||||
pub fn par_values_mut(&mut self) -> ParValuesMut<'_, K, V> {
|
||||
ParValuesMut {
|
||||
entries: self.as_entries_mut(),
|
||||
}
|
||||
}
|
||||
}
|
||||
|
||||
impl<K, V, S> IndexMap<K, V, S>
|
||||
where
|
||||
K: Hash + Eq + Send,
|
||||
V: Send,
|
||||
S: BuildHasher,
|
||||
{
|
||||
/// Sort the map’s key-value pairs in parallel, by the default ordering of the keys.
|
||||
pub fn par_sort_keys(&mut self)
|
||||
where
|
||||
K: Ord,
|
||||
{
|
||||
self.with_entries(|entries| {
|
||||
entries.par_sort_by(|a, b| K::cmp(&a.key, &b.key));
|
||||
});
|
||||
}
|
||||
|
||||
/// Sort the map’s key-value pairs in place and in parallel, using the comparison
|
||||
/// function `cmp`.
|
||||
///
|
||||
/// The comparison function receives two key and value pairs to compare (you
|
||||
/// can sort by keys or values or their combination as needed).
|
||||
pub fn par_sort_by<F>(&mut self, cmp: F)
|
||||
where
|
||||
F: Fn(&K, &V, &K, &V) -> Ordering + Sync,
|
||||
{
|
||||
self.with_entries(|entries| {
|
||||
entries.par_sort_by(move |a, b| cmp(&a.key, &a.value, &b.key, &b.value));
|
||||
});
|
||||
}
|
||||
|
||||
/// Sort the key-value pairs of the map in parallel and return a by-value parallel
|
||||
/// iterator of the key-value pairs with the result.
|
||||
pub fn par_sorted_by<F>(self, cmp: F) -> IntoParIter<K, V>
|
||||
where
|
||||
F: Fn(&K, &V, &K, &V) -> Ordering + Sync,
|
||||
{
|
||||
let mut entries = self.into_entries();
|
||||
entries.par_sort_by(move |a, b| cmp(&a.key, &a.value, &b.key, &b.value));
|
||||
IntoParIter { entries }
|
||||
}
|
||||
|
||||
/// Sort the map's key-value pairs in parallel, by the default ordering of the keys.
|
||||
pub fn par_sort_unstable_keys(&mut self)
|
||||
where
|
||||
K: Ord,
|
||||
{
|
||||
self.with_entries(|entries| {
|
||||
entries.par_sort_unstable_by(|a, b| K::cmp(&a.key, &b.key));
|
||||
});
|
||||
}
|
||||
|
||||
/// Sort the map's key-value pairs in place and in parallel, using the comparison
|
||||
/// function `cmp`.
|
||||
///
|
||||
/// The comparison function receives two key and value pairs to compare (you
|
||||
/// can sort by keys or values or their combination as needed).
|
||||
pub fn par_sort_unstable_by<F>(&mut self, cmp: F)
|
||||
where
|
||||
F: Fn(&K, &V, &K, &V) -> Ordering + Sync,
|
||||
{
|
||||
self.with_entries(|entries| {
|
||||
entries.par_sort_unstable_by(move |a, b| cmp(&a.key, &a.value, &b.key, &b.value));
|
||||
});
|
||||
}
|
||||
|
||||
/// Sort the key-value pairs of the map in parallel and return a by-value parallel
|
||||
/// iterator of the key-value pairs with the result.
|
||||
pub fn par_sorted_unstable_by<F>(self, cmp: F) -> IntoParIter<K, V>
|
||||
where
|
||||
F: Fn(&K, &V, &K, &V) -> Ordering + Sync,
|
||||
{
|
||||
let mut entries = self.into_entries();
|
||||
entries.par_sort_unstable_by(move |a, b| cmp(&a.key, &a.value, &b.key, &b.value));
|
||||
IntoParIter { entries }
|
||||
}
|
||||
}
|
||||
|
||||
/// A parallel mutable iterator over the values of a `IndexMap`.
|
||||
///
|
||||
/// This `struct` is created by the [`par_values_mut`] method on [`IndexMap`]. See its
|
||||
/// documentation for more.
|
||||
///
|
||||
/// [`par_values_mut`]: ../struct.IndexMap.html#method.par_values_mut
|
||||
/// [`IndexMap`]: ../struct.IndexMap.html
|
||||
pub struct ParValuesMut<'a, K, V> {
|
||||
entries: &'a mut [Bucket<K, V>],
|
||||
}
|
||||
|
||||
impl<K, V: fmt::Debug> fmt::Debug for ParValuesMut<'_, K, V> {
|
||||
fn fmt(&self, f: &mut fmt::Formatter<'_>) -> fmt::Result {
|
||||
let iter = self.entries.iter().map(Bucket::value_ref);
|
||||
f.debug_list().entries(iter).finish()
|
||||
}
|
||||
}
|
||||
|
||||
impl<'a, K: Send, V: Send> ParallelIterator for ParValuesMut<'a, K, V> {
|
||||
type Item = &'a mut V;
|
||||
|
||||
parallel_iterator_methods!(Bucket::value_mut);
|
||||
}
|
||||
|
||||
impl<K: Send, V: Send> IndexedParallelIterator for ParValuesMut<'_, K, V> {
|
||||
indexed_parallel_iterator_methods!(Bucket::value_mut);
|
||||
}
|
||||
|
||||
/// Requires crate feature `"rayon"`.
|
||||
impl<K, V, S> FromParallelIterator<(K, V)> for IndexMap<K, V, S>
|
||||
where
|
||||
K: Eq + Hash + Send,
|
||||
V: Send,
|
||||
S: BuildHasher + Default + Send,
|
||||
{
|
||||
fn from_par_iter<I>(iter: I) -> Self
|
||||
where
|
||||
I: IntoParallelIterator<Item = (K, V)>,
|
||||
{
|
||||
let list = collect(iter);
|
||||
let len = list.iter().map(Vec::len).sum();
|
||||
let mut map = Self::with_capacity_and_hasher(len, S::default());
|
||||
for vec in list {
|
||||
map.extend(vec);
|
||||
}
|
||||
map
|
||||
}
|
||||
}
|
||||
|
||||
/// Requires crate feature `"rayon"`.
|
||||
impl<K, V, S> ParallelExtend<(K, V)> for IndexMap<K, V, S>
|
||||
where
|
||||
K: Eq + Hash + Send,
|
||||
V: Send,
|
||||
S: BuildHasher + Send,
|
||||
{
|
||||
fn par_extend<I>(&mut self, iter: I)
|
||||
where
|
||||
I: IntoParallelIterator<Item = (K, V)>,
|
||||
{
|
||||
for vec in collect(iter) {
|
||||
self.extend(vec);
|
||||
}
|
||||
}
|
||||
}
|
||||
|
||||
/// Requires crate feature `"rayon"`.
|
||||
impl<'a, K: 'a, V: 'a, S> ParallelExtend<(&'a K, &'a V)> for IndexMap<K, V, S>
|
||||
where
|
||||
K: Copy + Eq + Hash + Send + Sync,
|
||||
V: Copy + Send + Sync,
|
||||
S: BuildHasher + Send,
|
||||
{
|
||||
fn par_extend<I>(&mut self, iter: I)
|
||||
where
|
||||
I: IntoParallelIterator<Item = (&'a K, &'a V)>,
|
||||
{
|
||||
for vec in collect(iter) {
|
||||
self.extend(vec);
|
||||
}
|
||||
}
|
||||
}
|
||||
|
||||
#[cfg(test)]
|
||||
mod tests {
|
||||
use super::*;
|
||||
use std::string::String;
|
||||
|
||||
#[test]
|
||||
fn insert_order() {
|
||||
let insert = [0, 4, 2, 12, 8, 7, 11, 5, 3, 17, 19, 22, 23];
|
||||
let mut map = IndexMap::new();
|
||||
|
||||
for &elt in &insert {
|
||||
map.insert(elt, ());
|
||||
}
|
||||
|
||||
assert_eq!(map.par_keys().count(), map.len());
|
||||
assert_eq!(map.par_keys().count(), insert.len());
|
||||
insert.par_iter().zip(map.par_keys()).for_each(|(a, b)| {
|
||||
assert_eq!(a, b);
|
||||
});
|
||||
(0..insert.len())
|
||||
.into_par_iter()
|
||||
.zip(map.par_keys())
|
||||
.for_each(|(i, k)| {
|
||||
assert_eq!(map.get_index(i).unwrap().0, k);
|
||||
});
|
||||
}
|
||||
|
||||
#[test]
|
||||
fn partial_eq_and_eq() {
|
||||
let mut map_a = IndexMap::new();
|
||||
map_a.insert(1, "1");
|
||||
map_a.insert(2, "2");
|
||||
let mut map_b = map_a.clone();
|
||||
assert!(map_a.par_eq(&map_b));
|
||||
map_b.swap_remove(&1);
|
||||
assert!(!map_a.par_eq(&map_b));
|
||||
map_b.insert(3, "3");
|
||||
assert!(!map_a.par_eq(&map_b));
|
||||
|
||||
let map_c: IndexMap<_, String> =
|
||||
map_b.into_par_iter().map(|(k, v)| (k, v.into())).collect();
|
||||
assert!(!map_a.par_eq(&map_c));
|
||||
assert!(!map_c.par_eq(&map_a));
|
||||
}
|
||||
|
||||
#[test]
|
||||
fn extend() {
|
||||
let mut map = IndexMap::new();
|
||||
map.par_extend(vec![(&1, &2), (&3, &4)]);
|
||||
map.par_extend(vec![(5, 6)]);
|
||||
assert_eq!(
|
||||
map.into_par_iter().collect::<Vec<_>>(),
|
||||
vec![(1, 2), (3, 4), (5, 6)]
|
||||
);
|
||||
}
|
||||
|
||||
#[test]
|
||||
fn keys() {
|
||||
let vec = vec![(1, 'a'), (2, 'b'), (3, 'c')];
|
||||
let map: IndexMap<_, _> = vec.into_par_iter().collect();
|
||||
let keys: Vec<_> = map.par_keys().copied().collect();
|
||||
assert_eq!(keys.len(), 3);
|
||||
assert!(keys.contains(&1));
|
||||
assert!(keys.contains(&2));
|
||||
assert!(keys.contains(&3));
|
||||
}
|
||||
|
||||
#[test]
|
||||
fn values() {
|
||||
let vec = vec![(1, 'a'), (2, 'b'), (3, 'c')];
|
||||
let map: IndexMap<_, _> = vec.into_par_iter().collect();
|
||||
let values: Vec<_> = map.par_values().copied().collect();
|
||||
assert_eq!(values.len(), 3);
|
||||
assert!(values.contains(&'a'));
|
||||
assert!(values.contains(&'b'));
|
||||
assert!(values.contains(&'c'));
|
||||
}
|
||||
|
||||
#[test]
|
||||
fn values_mut() {
|
||||
let vec = vec![(1, 1), (2, 2), (3, 3)];
|
||||
let mut map: IndexMap<_, _> = vec.into_par_iter().collect();
|
||||
map.par_values_mut().for_each(|value| *value *= 2);
|
||||
let values: Vec<_> = map.par_values().copied().collect();
|
||||
assert_eq!(values.len(), 3);
|
||||
assert!(values.contains(&2));
|
||||
assert!(values.contains(&4));
|
||||
assert!(values.contains(&6));
|
||||
}
|
||||
}
|
||||
27
clamav/libclamav_rust/.cargo/vendor/indexmap/src/rayon/mod.rs
vendored
Normal file
27
clamav/libclamav_rust/.cargo/vendor/indexmap/src/rayon/mod.rs
vendored
Normal file
@@ -0,0 +1,27 @@
|
||||
use rayon::prelude::*;
|
||||
|
||||
use alloc::collections::LinkedList;
|
||||
|
||||
use crate::vec::Vec;
|
||||
|
||||
pub mod map;
|
||||
pub mod set;
|
||||
|
||||
// This form of intermediate collection is also how Rayon collects `HashMap`.
|
||||
// Note that the order will also be preserved!
|
||||
fn collect<I: IntoParallelIterator>(iter: I) -> LinkedList<Vec<I::Item>> {
|
||||
iter.into_par_iter()
|
||||
.fold(Vec::new, |mut vec, elem| {
|
||||
vec.push(elem);
|
||||
vec
|
||||
})
|
||||
.map(|vec| {
|
||||
let mut list = LinkedList::new();
|
||||
list.push_back(vec);
|
||||
list
|
||||
})
|
||||
.reduce(LinkedList::new, |mut list1, mut list2| {
|
||||
list1.append(&mut list2);
|
||||
list1
|
||||
})
|
||||
}
|
||||
741
clamav/libclamav_rust/.cargo/vendor/indexmap/src/rayon/set.rs
vendored
Normal file
741
clamav/libclamav_rust/.cargo/vendor/indexmap/src/rayon/set.rs
vendored
Normal file
@@ -0,0 +1,741 @@
|
||||
//! Parallel iterator types for `IndexSet` with [rayon](https://docs.rs/rayon/1.0/rayon).
|
||||
//!
|
||||
//! You will rarely need to interact with this module directly unless you need to name one of the
|
||||
//! iterator types.
|
||||
//!
|
||||
//! Requires crate feature `"rayon"`.
|
||||
|
||||
use super::collect;
|
||||
use rayon::iter::plumbing::{Consumer, ProducerCallback, UnindexedConsumer};
|
||||
use rayon::prelude::*;
|
||||
|
||||
use crate::vec::Vec;
|
||||
use core::cmp::Ordering;
|
||||
use core::fmt;
|
||||
use core::hash::{BuildHasher, Hash};
|
||||
use core::ops::RangeBounds;
|
||||
|
||||
use crate::Entries;
|
||||
use crate::IndexSet;
|
||||
|
||||
type Bucket<T> = crate::Bucket<T, ()>;
|
||||
|
||||
/// Requires crate feature `"rayon"`.
|
||||
impl<T, S> IntoParallelIterator for IndexSet<T, S>
|
||||
where
|
||||
T: Send,
|
||||
{
|
||||
type Item = T;
|
||||
type Iter = IntoParIter<T>;
|
||||
|
||||
fn into_par_iter(self) -> Self::Iter {
|
||||
IntoParIter {
|
||||
entries: self.into_entries(),
|
||||
}
|
||||
}
|
||||
}
|
||||
|
||||
/// A parallel owning iterator over the items of a `IndexSet`.
|
||||
///
|
||||
/// This `struct` is created by the [`into_par_iter`] method on [`IndexSet`]
|
||||
/// (provided by rayon's `IntoParallelIterator` trait). See its documentation for more.
|
||||
///
|
||||
/// [`IndexSet`]: ../struct.IndexSet.html
|
||||
/// [`into_par_iter`]: ../struct.IndexSet.html#method.into_par_iter
|
||||
pub struct IntoParIter<T> {
|
||||
entries: Vec<Bucket<T>>,
|
||||
}
|
||||
|
||||
impl<T: fmt::Debug> fmt::Debug for IntoParIter<T> {
|
||||
fn fmt(&self, f: &mut fmt::Formatter<'_>) -> fmt::Result {
|
||||
let iter = self.entries.iter().map(Bucket::key_ref);
|
||||
f.debug_list().entries(iter).finish()
|
||||
}
|
||||
}
|
||||
|
||||
impl<T: Send> ParallelIterator for IntoParIter<T> {
|
||||
type Item = T;
|
||||
|
||||
parallel_iterator_methods!(Bucket::key);
|
||||
}
|
||||
|
||||
impl<T: Send> IndexedParallelIterator for IntoParIter<T> {
|
||||
indexed_parallel_iterator_methods!(Bucket::key);
|
||||
}
|
||||
|
||||
/// Requires crate feature `"rayon"`.
|
||||
impl<'a, T, S> IntoParallelIterator for &'a IndexSet<T, S>
|
||||
where
|
||||
T: Sync,
|
||||
{
|
||||
type Item = &'a T;
|
||||
type Iter = ParIter<'a, T>;
|
||||
|
||||
fn into_par_iter(self) -> Self::Iter {
|
||||
ParIter {
|
||||
entries: self.as_entries(),
|
||||
}
|
||||
}
|
||||
}
|
||||
|
||||
/// A parallel iterator over the items of a `IndexSet`.
|
||||
///
|
||||
/// This `struct` is created by the [`par_iter`] method on [`IndexSet`]
|
||||
/// (provided by rayon's `IntoParallelRefIterator` trait). See its documentation for more.
|
||||
///
|
||||
/// [`IndexSet`]: ../struct.IndexSet.html
|
||||
/// [`par_iter`]: ../struct.IndexSet.html#method.par_iter
|
||||
pub struct ParIter<'a, T> {
|
||||
entries: &'a [Bucket<T>],
|
||||
}
|
||||
|
||||
impl<T> Clone for ParIter<'_, T> {
|
||||
fn clone(&self) -> Self {
|
||||
ParIter { ..*self }
|
||||
}
|
||||
}
|
||||
|
||||
impl<T: fmt::Debug> fmt::Debug for ParIter<'_, T> {
|
||||
fn fmt(&self, f: &mut fmt::Formatter<'_>) -> fmt::Result {
|
||||
let iter = self.entries.iter().map(Bucket::key_ref);
|
||||
f.debug_list().entries(iter).finish()
|
||||
}
|
||||
}
|
||||
|
||||
impl<'a, T: Sync> ParallelIterator for ParIter<'a, T> {
|
||||
type Item = &'a T;
|
||||
|
||||
parallel_iterator_methods!(Bucket::key_ref);
|
||||
}
|
||||
|
||||
impl<T: Sync> IndexedParallelIterator for ParIter<'_, T> {
|
||||
indexed_parallel_iterator_methods!(Bucket::key_ref);
|
||||
}
|
||||
|
||||
/// Requires crate feature `"rayon"`.
|
||||
impl<'a, T, S> ParallelDrainRange<usize> for &'a mut IndexSet<T, S>
|
||||
where
|
||||
T: Send,
|
||||
{
|
||||
type Item = T;
|
||||
type Iter = ParDrain<'a, T>;
|
||||
|
||||
fn par_drain<R: RangeBounds<usize>>(self, range: R) -> Self::Iter {
|
||||
ParDrain {
|
||||
entries: self.map.core.par_drain(range),
|
||||
}
|
||||
}
|
||||
}
|
||||
|
||||
/// A parallel draining iterator over the items of a `IndexSet`.
|
||||
///
|
||||
/// This `struct` is created by the [`par_drain`] method on [`IndexSet`]
|
||||
/// (provided by rayon's `ParallelDrainRange` trait). See its documentation for more.
|
||||
///
|
||||
/// [`par_drain`]: ../struct.IndexSet.html#method.par_drain
|
||||
/// [`IndexSet`]: ../struct.IndexSet.html
|
||||
pub struct ParDrain<'a, T: Send> {
|
||||
entries: rayon::vec::Drain<'a, Bucket<T>>,
|
||||
}
|
||||
|
||||
impl<T: Send> ParallelIterator for ParDrain<'_, T> {
|
||||
type Item = T;
|
||||
|
||||
parallel_iterator_methods!(Bucket::key);
|
||||
}
|
||||
|
||||
impl<T: Send> IndexedParallelIterator for ParDrain<'_, T> {
|
||||
indexed_parallel_iterator_methods!(Bucket::key);
|
||||
}
|
||||
|
||||
/// Parallel iterator methods and other parallel methods.
|
||||
///
|
||||
/// The following methods **require crate feature `"rayon"`**.
|
||||
///
|
||||
/// See also the `IntoParallelIterator` implementations.
|
||||
impl<T, S> IndexSet<T, S>
|
||||
where
|
||||
T: Hash + Eq + Sync,
|
||||
S: BuildHasher + Sync,
|
||||
{
|
||||
/// Return a parallel iterator over the values that are in `self` but not `other`.
|
||||
///
|
||||
/// While parallel iterators can process items in any order, their relative order
|
||||
/// in the `self` set is still preserved for operations like `reduce` and `collect`.
|
||||
pub fn par_difference<'a, S2>(
|
||||
&'a self,
|
||||
other: &'a IndexSet<T, S2>,
|
||||
) -> ParDifference<'a, T, S, S2>
|
||||
where
|
||||
S2: BuildHasher + Sync,
|
||||
{
|
||||
ParDifference {
|
||||
set1: self,
|
||||
set2: other,
|
||||
}
|
||||
}
|
||||
|
||||
/// Return a parallel iterator over the values that are in `self` or `other`,
|
||||
/// but not in both.
|
||||
///
|
||||
/// While parallel iterators can process items in any order, their relative order
|
||||
/// in the sets is still preserved for operations like `reduce` and `collect`.
|
||||
/// Values from `self` are produced in their original order, followed by
|
||||
/// values from `other` in their original order.
|
||||
pub fn par_symmetric_difference<'a, S2>(
|
||||
&'a self,
|
||||
other: &'a IndexSet<T, S2>,
|
||||
) -> ParSymmetricDifference<'a, T, S, S2>
|
||||
where
|
||||
S2: BuildHasher + Sync,
|
||||
{
|
||||
ParSymmetricDifference {
|
||||
set1: self,
|
||||
set2: other,
|
||||
}
|
||||
}
|
||||
|
||||
/// Return a parallel iterator over the values that are in both `self` and `other`.
|
||||
///
|
||||
/// While parallel iterators can process items in any order, their relative order
|
||||
/// in the `self` set is still preserved for operations like `reduce` and `collect`.
|
||||
pub fn par_intersection<'a, S2>(
|
||||
&'a self,
|
||||
other: &'a IndexSet<T, S2>,
|
||||
) -> ParIntersection<'a, T, S, S2>
|
||||
where
|
||||
S2: BuildHasher + Sync,
|
||||
{
|
||||
ParIntersection {
|
||||
set1: self,
|
||||
set2: other,
|
||||
}
|
||||
}
|
||||
|
||||
/// Return a parallel iterator over all values that are in `self` or `other`.
|
||||
///
|
||||
/// While parallel iterators can process items in any order, their relative order
|
||||
/// in the sets is still preserved for operations like `reduce` and `collect`.
|
||||
/// Values from `self` are produced in their original order, followed by
|
||||
/// values that are unique to `other` in their original order.
|
||||
pub fn par_union<'a, S2>(&'a self, other: &'a IndexSet<T, S2>) -> ParUnion<'a, T, S, S2>
|
||||
where
|
||||
S2: BuildHasher + Sync,
|
||||
{
|
||||
ParUnion {
|
||||
set1: self,
|
||||
set2: other,
|
||||
}
|
||||
}
|
||||
|
||||
/// Returns `true` if `self` contains all of the same values as `other`,
|
||||
/// regardless of each set's indexed order, determined in parallel.
|
||||
pub fn par_eq<S2>(&self, other: &IndexSet<T, S2>) -> bool
|
||||
where
|
||||
S2: BuildHasher + Sync,
|
||||
{
|
||||
self.len() == other.len() && self.par_is_subset(other)
|
||||
}
|
||||
|
||||
/// Returns `true` if `self` has no elements in common with `other`,
|
||||
/// determined in parallel.
|
||||
pub fn par_is_disjoint<S2>(&self, other: &IndexSet<T, S2>) -> bool
|
||||
where
|
||||
S2: BuildHasher + Sync,
|
||||
{
|
||||
if self.len() <= other.len() {
|
||||
self.par_iter().all(move |value| !other.contains(value))
|
||||
} else {
|
||||
other.par_iter().all(move |value| !self.contains(value))
|
||||
}
|
||||
}
|
||||
|
||||
/// Returns `true` if all elements of `other` are contained in `self`,
|
||||
/// determined in parallel.
|
||||
pub fn par_is_superset<S2>(&self, other: &IndexSet<T, S2>) -> bool
|
||||
where
|
||||
S2: BuildHasher + Sync,
|
||||
{
|
||||
other.par_is_subset(self)
|
||||
}
|
||||
|
||||
/// Returns `true` if all elements of `self` are contained in `other`,
|
||||
/// determined in parallel.
|
||||
pub fn par_is_subset<S2>(&self, other: &IndexSet<T, S2>) -> bool
|
||||
where
|
||||
S2: BuildHasher + Sync,
|
||||
{
|
||||
self.len() <= other.len() && self.par_iter().all(move |value| other.contains(value))
|
||||
}
|
||||
}
|
||||
|
||||
/// A parallel iterator producing elements in the difference of `IndexSet`s.
|
||||
///
|
||||
/// This `struct` is created by the [`par_difference`] method on [`IndexSet`].
|
||||
/// See its documentation for more.
|
||||
///
|
||||
/// [`IndexSet`]: ../struct.IndexSet.html
|
||||
/// [`par_difference`]: ../struct.IndexSet.html#method.par_difference
|
||||
pub struct ParDifference<'a, T, S1, S2> {
|
||||
set1: &'a IndexSet<T, S1>,
|
||||
set2: &'a IndexSet<T, S2>,
|
||||
}
|
||||
|
||||
impl<T, S1, S2> Clone for ParDifference<'_, T, S1, S2> {
|
||||
fn clone(&self) -> Self {
|
||||
ParDifference { ..*self }
|
||||
}
|
||||
}
|
||||
|
||||
impl<T, S1, S2> fmt::Debug for ParDifference<'_, T, S1, S2>
|
||||
where
|
||||
T: fmt::Debug + Eq + Hash,
|
||||
S1: BuildHasher,
|
||||
S2: BuildHasher,
|
||||
{
|
||||
fn fmt(&self, f: &mut fmt::Formatter<'_>) -> fmt::Result {
|
||||
f.debug_list()
|
||||
.entries(self.set1.difference(self.set2))
|
||||
.finish()
|
||||
}
|
||||
}
|
||||
|
||||
impl<'a, T, S1, S2> ParallelIterator for ParDifference<'a, T, S1, S2>
|
||||
where
|
||||
T: Hash + Eq + Sync,
|
||||
S1: BuildHasher + Sync,
|
||||
S2: BuildHasher + Sync,
|
||||
{
|
||||
type Item = &'a T;
|
||||
|
||||
fn drive_unindexed<C>(self, consumer: C) -> C::Result
|
||||
where
|
||||
C: UnindexedConsumer<Self::Item>,
|
||||
{
|
||||
let Self { set1, set2 } = self;
|
||||
|
||||
set1.par_iter()
|
||||
.filter(move |&item| !set2.contains(item))
|
||||
.drive_unindexed(consumer)
|
||||
}
|
||||
}
|
||||
|
||||
/// A parallel iterator producing elements in the intersection of `IndexSet`s.
|
||||
///
|
||||
/// This `struct` is created by the [`par_intersection`] method on [`IndexSet`].
|
||||
/// See its documentation for more.
|
||||
///
|
||||
/// [`IndexSet`]: ../struct.IndexSet.html
|
||||
/// [`par_intersection`]: ../struct.IndexSet.html#method.par_intersection
|
||||
pub struct ParIntersection<'a, T, S1, S2> {
|
||||
set1: &'a IndexSet<T, S1>,
|
||||
set2: &'a IndexSet<T, S2>,
|
||||
}
|
||||
|
||||
impl<T, S1, S2> Clone for ParIntersection<'_, T, S1, S2> {
|
||||
fn clone(&self) -> Self {
|
||||
ParIntersection { ..*self }
|
||||
}
|
||||
}
|
||||
|
||||
impl<T, S1, S2> fmt::Debug for ParIntersection<'_, T, S1, S2>
|
||||
where
|
||||
T: fmt::Debug + Eq + Hash,
|
||||
S1: BuildHasher,
|
||||
S2: BuildHasher,
|
||||
{
|
||||
fn fmt(&self, f: &mut fmt::Formatter<'_>) -> fmt::Result {
|
||||
f.debug_list()
|
||||
.entries(self.set1.intersection(self.set2))
|
||||
.finish()
|
||||
}
|
||||
}
|
||||
|
||||
impl<'a, T, S1, S2> ParallelIterator for ParIntersection<'a, T, S1, S2>
|
||||
where
|
||||
T: Hash + Eq + Sync,
|
||||
S1: BuildHasher + Sync,
|
||||
S2: BuildHasher + Sync,
|
||||
{
|
||||
type Item = &'a T;
|
||||
|
||||
fn drive_unindexed<C>(self, consumer: C) -> C::Result
|
||||
where
|
||||
C: UnindexedConsumer<Self::Item>,
|
||||
{
|
||||
let Self { set1, set2 } = self;
|
||||
|
||||
set1.par_iter()
|
||||
.filter(move |&item| set2.contains(item))
|
||||
.drive_unindexed(consumer)
|
||||
}
|
||||
}
|
||||
|
||||
/// A parallel iterator producing elements in the symmetric difference of `IndexSet`s.
|
||||
///
|
||||
/// This `struct` is created by the [`par_symmetric_difference`] method on
|
||||
/// [`IndexSet`]. See its documentation for more.
|
||||
///
|
||||
/// [`IndexSet`]: ../struct.IndexSet.html
|
||||
/// [`par_symmetric_difference`]: ../struct.IndexSet.html#method.par_symmetric_difference
|
||||
pub struct ParSymmetricDifference<'a, T, S1, S2> {
|
||||
set1: &'a IndexSet<T, S1>,
|
||||
set2: &'a IndexSet<T, S2>,
|
||||
}
|
||||
|
||||
impl<T, S1, S2> Clone for ParSymmetricDifference<'_, T, S1, S2> {
|
||||
fn clone(&self) -> Self {
|
||||
ParSymmetricDifference { ..*self }
|
||||
}
|
||||
}
|
||||
|
||||
impl<T, S1, S2> fmt::Debug for ParSymmetricDifference<'_, T, S1, S2>
|
||||
where
|
||||
T: fmt::Debug + Eq + Hash,
|
||||
S1: BuildHasher,
|
||||
S2: BuildHasher,
|
||||
{
|
||||
fn fmt(&self, f: &mut fmt::Formatter<'_>) -> fmt::Result {
|
||||
f.debug_list()
|
||||
.entries(self.set1.symmetric_difference(self.set2))
|
||||
.finish()
|
||||
}
|
||||
}
|
||||
|
||||
impl<'a, T, S1, S2> ParallelIterator for ParSymmetricDifference<'a, T, S1, S2>
|
||||
where
|
||||
T: Hash + Eq + Sync,
|
||||
S1: BuildHasher + Sync,
|
||||
S2: BuildHasher + Sync,
|
||||
{
|
||||
type Item = &'a T;
|
||||
|
||||
fn drive_unindexed<C>(self, consumer: C) -> C::Result
|
||||
where
|
||||
C: UnindexedConsumer<Self::Item>,
|
||||
{
|
||||
let Self { set1, set2 } = self;
|
||||
|
||||
set1.par_difference(set2)
|
||||
.chain(set2.par_difference(set1))
|
||||
.drive_unindexed(consumer)
|
||||
}
|
||||
}
|
||||
|
||||
/// A parallel iterator producing elements in the union of `IndexSet`s.
|
||||
///
|
||||
/// This `struct` is created by the [`par_union`] method on [`IndexSet`].
|
||||
/// See its documentation for more.
|
||||
///
|
||||
/// [`IndexSet`]: ../struct.IndexSet.html
|
||||
/// [`par_union`]: ../struct.IndexSet.html#method.par_union
|
||||
pub struct ParUnion<'a, T, S1, S2> {
|
||||
set1: &'a IndexSet<T, S1>,
|
||||
set2: &'a IndexSet<T, S2>,
|
||||
}
|
||||
|
||||
impl<T, S1, S2> Clone for ParUnion<'_, T, S1, S2> {
|
||||
fn clone(&self) -> Self {
|
||||
ParUnion { ..*self }
|
||||
}
|
||||
}
|
||||
|
||||
impl<T, S1, S2> fmt::Debug for ParUnion<'_, T, S1, S2>
|
||||
where
|
||||
T: fmt::Debug + Eq + Hash,
|
||||
S1: BuildHasher,
|
||||
S2: BuildHasher,
|
||||
{
|
||||
fn fmt(&self, f: &mut fmt::Formatter<'_>) -> fmt::Result {
|
||||
f.debug_list().entries(self.set1.union(self.set2)).finish()
|
||||
}
|
||||
}
|
||||
|
||||
impl<'a, T, S1, S2> ParallelIterator for ParUnion<'a, T, S1, S2>
|
||||
where
|
||||
T: Hash + Eq + Sync,
|
||||
S1: BuildHasher + Sync,
|
||||
S2: BuildHasher + Sync,
|
||||
{
|
||||
type Item = &'a T;
|
||||
|
||||
fn drive_unindexed<C>(self, consumer: C) -> C::Result
|
||||
where
|
||||
C: UnindexedConsumer<Self::Item>,
|
||||
{
|
||||
let Self { set1, set2 } = self;
|
||||
|
||||
set1.par_iter()
|
||||
.chain(set2.par_difference(set1))
|
||||
.drive_unindexed(consumer)
|
||||
}
|
||||
}
|
||||
|
||||
/// Parallel sorting methods.
|
||||
///
|
||||
/// The following methods **require crate feature `"rayon"`**.
|
||||
impl<T, S> IndexSet<T, S>
|
||||
where
|
||||
T: Hash + Eq + Send,
|
||||
S: BuildHasher + Send,
|
||||
{
|
||||
/// Sort the set’s values in parallel by their default ordering.
|
||||
pub fn par_sort(&mut self)
|
||||
where
|
||||
T: Ord,
|
||||
{
|
||||
self.with_entries(|entries| {
|
||||
entries.par_sort_by(|a, b| T::cmp(&a.key, &b.key));
|
||||
});
|
||||
}
|
||||
|
||||
/// Sort the set’s values in place and in parallel, using the comparison function `cmp`.
|
||||
pub fn par_sort_by<F>(&mut self, cmp: F)
|
||||
where
|
||||
F: Fn(&T, &T) -> Ordering + Sync,
|
||||
{
|
||||
self.with_entries(|entries| {
|
||||
entries.par_sort_by(move |a, b| cmp(&a.key, &b.key));
|
||||
});
|
||||
}
|
||||
|
||||
/// Sort the values of the set in parallel and return a by-value parallel iterator of
|
||||
/// the values with the result.
|
||||
pub fn par_sorted_by<F>(self, cmp: F) -> IntoParIter<T>
|
||||
where
|
||||
F: Fn(&T, &T) -> Ordering + Sync,
|
||||
{
|
||||
let mut entries = self.into_entries();
|
||||
entries.par_sort_by(move |a, b| cmp(&a.key, &b.key));
|
||||
IntoParIter { entries }
|
||||
}
|
||||
|
||||
/// Sort the set's values in parallel by their default ordering.
|
||||
pub fn par_sort_unstable(&mut self)
|
||||
where
|
||||
T: Ord,
|
||||
{
|
||||
self.with_entries(|entries| {
|
||||
entries.par_sort_unstable_by(|a, b| T::cmp(&a.key, &b.key));
|
||||
});
|
||||
}
|
||||
|
||||
/// Sort the set’s values in place and in parallel, using the comparison function `cmp`.
|
||||
pub fn par_sort_unstable_by<F>(&mut self, cmp: F)
|
||||
where
|
||||
F: Fn(&T, &T) -> Ordering + Sync,
|
||||
{
|
||||
self.with_entries(|entries| {
|
||||
entries.par_sort_unstable_by(move |a, b| cmp(&a.key, &b.key));
|
||||
});
|
||||
}
|
||||
|
||||
/// Sort the values of the set in parallel and return a by-value parallel iterator of
|
||||
/// the values with the result.
|
||||
pub fn par_sorted_unstable_by<F>(self, cmp: F) -> IntoParIter<T>
|
||||
where
|
||||
F: Fn(&T, &T) -> Ordering + Sync,
|
||||
{
|
||||
let mut entries = self.into_entries();
|
||||
entries.par_sort_unstable_by(move |a, b| cmp(&a.key, &b.key));
|
||||
IntoParIter { entries }
|
||||
}
|
||||
}
|
||||
|
||||
/// Requires crate feature `"rayon"`.
|
||||
impl<T, S> FromParallelIterator<T> for IndexSet<T, S>
|
||||
where
|
||||
T: Eq + Hash + Send,
|
||||
S: BuildHasher + Default + Send,
|
||||
{
|
||||
fn from_par_iter<I>(iter: I) -> Self
|
||||
where
|
||||
I: IntoParallelIterator<Item = T>,
|
||||
{
|
||||
let list = collect(iter);
|
||||
let len = list.iter().map(Vec::len).sum();
|
||||
let mut set = Self::with_capacity_and_hasher(len, S::default());
|
||||
for vec in list {
|
||||
set.extend(vec);
|
||||
}
|
||||
set
|
||||
}
|
||||
}
|
||||
|
||||
/// Requires crate feature `"rayon"`.
|
||||
impl<T, S> ParallelExtend<T> for IndexSet<T, S>
|
||||
where
|
||||
T: Eq + Hash + Send,
|
||||
S: BuildHasher + Send,
|
||||
{
|
||||
fn par_extend<I>(&mut self, iter: I)
|
||||
where
|
||||
I: IntoParallelIterator<Item = T>,
|
||||
{
|
||||
for vec in collect(iter) {
|
||||
self.extend(vec);
|
||||
}
|
||||
}
|
||||
}
|
||||
|
||||
/// Requires crate feature `"rayon"`.
|
||||
impl<'a, T: 'a, S> ParallelExtend<&'a T> for IndexSet<T, S>
|
||||
where
|
||||
T: Copy + Eq + Hash + Send + Sync,
|
||||
S: BuildHasher + Send,
|
||||
{
|
||||
fn par_extend<I>(&mut self, iter: I)
|
||||
where
|
||||
I: IntoParallelIterator<Item = &'a T>,
|
||||
{
|
||||
for vec in collect(iter) {
|
||||
self.extend(vec);
|
||||
}
|
||||
}
|
||||
}
|
||||
|
||||
#[cfg(test)]
|
||||
mod tests {
|
||||
use super::*;
|
||||
|
||||
#[test]
|
||||
fn insert_order() {
|
||||
let insert = [0, 4, 2, 12, 8, 7, 11, 5, 3, 17, 19, 22, 23];
|
||||
let mut set = IndexSet::new();
|
||||
|
||||
for &elt in &insert {
|
||||
set.insert(elt);
|
||||
}
|
||||
|
||||
assert_eq!(set.par_iter().count(), set.len());
|
||||
assert_eq!(set.par_iter().count(), insert.len());
|
||||
insert.par_iter().zip(&set).for_each(|(a, b)| {
|
||||
assert_eq!(a, b);
|
||||
});
|
||||
(0..insert.len())
|
||||
.into_par_iter()
|
||||
.zip(&set)
|
||||
.for_each(|(i, v)| {
|
||||
assert_eq!(set.get_index(i).unwrap(), v);
|
||||
});
|
||||
}
|
||||
|
||||
#[test]
|
||||
fn partial_eq_and_eq() {
|
||||
let mut set_a = IndexSet::new();
|
||||
set_a.insert(1);
|
||||
set_a.insert(2);
|
||||
let mut set_b = set_a.clone();
|
||||
assert!(set_a.par_eq(&set_b));
|
||||
set_b.swap_remove(&1);
|
||||
assert!(!set_a.par_eq(&set_b));
|
||||
set_b.insert(3);
|
||||
assert!(!set_a.par_eq(&set_b));
|
||||
|
||||
let set_c: IndexSet<_> = set_b.into_par_iter().collect();
|
||||
assert!(!set_a.par_eq(&set_c));
|
||||
assert!(!set_c.par_eq(&set_a));
|
||||
}
|
||||
|
||||
#[test]
|
||||
fn extend() {
|
||||
let mut set = IndexSet::new();
|
||||
set.par_extend(vec![&1, &2, &3, &4]);
|
||||
set.par_extend(vec![5, 6]);
|
||||
assert_eq!(
|
||||
set.into_par_iter().collect::<Vec<_>>(),
|
||||
vec![1, 2, 3, 4, 5, 6]
|
||||
);
|
||||
}
|
||||
|
||||
#[test]
|
||||
fn comparisons() {
|
||||
let set_a: IndexSet<_> = (0..3).collect();
|
||||
let set_b: IndexSet<_> = (3..6).collect();
|
||||
let set_c: IndexSet<_> = (0..6).collect();
|
||||
let set_d: IndexSet<_> = (3..9).collect();
|
||||
|
||||
assert!(!set_a.par_is_disjoint(&set_a));
|
||||
assert!(set_a.par_is_subset(&set_a));
|
||||
assert!(set_a.par_is_superset(&set_a));
|
||||
|
||||
assert!(set_a.par_is_disjoint(&set_b));
|
||||
assert!(set_b.par_is_disjoint(&set_a));
|
||||
assert!(!set_a.par_is_subset(&set_b));
|
||||
assert!(!set_b.par_is_subset(&set_a));
|
||||
assert!(!set_a.par_is_superset(&set_b));
|
||||
assert!(!set_b.par_is_superset(&set_a));
|
||||
|
||||
assert!(!set_a.par_is_disjoint(&set_c));
|
||||
assert!(!set_c.par_is_disjoint(&set_a));
|
||||
assert!(set_a.par_is_subset(&set_c));
|
||||
assert!(!set_c.par_is_subset(&set_a));
|
||||
assert!(!set_a.par_is_superset(&set_c));
|
||||
assert!(set_c.par_is_superset(&set_a));
|
||||
|
||||
assert!(!set_c.par_is_disjoint(&set_d));
|
||||
assert!(!set_d.par_is_disjoint(&set_c));
|
||||
assert!(!set_c.par_is_subset(&set_d));
|
||||
assert!(!set_d.par_is_subset(&set_c));
|
||||
assert!(!set_c.par_is_superset(&set_d));
|
||||
assert!(!set_d.par_is_superset(&set_c));
|
||||
}
|
||||
|
||||
#[test]
|
||||
fn iter_comparisons() {
|
||||
use std::iter::empty;
|
||||
|
||||
fn check<'a, I1, I2>(iter1: I1, iter2: I2)
|
||||
where
|
||||
I1: ParallelIterator<Item = &'a i32>,
|
||||
I2: Iterator<Item = i32>,
|
||||
{
|
||||
let v1: Vec<_> = iter1.copied().collect();
|
||||
let v2: Vec<_> = iter2.collect();
|
||||
assert_eq!(v1, v2);
|
||||
}
|
||||
|
||||
let set_a: IndexSet<_> = (0..3).collect();
|
||||
let set_b: IndexSet<_> = (3..6).collect();
|
||||
let set_c: IndexSet<_> = (0..6).collect();
|
||||
let set_d: IndexSet<_> = (3..9).rev().collect();
|
||||
|
||||
check(set_a.par_difference(&set_a), empty());
|
||||
check(set_a.par_symmetric_difference(&set_a), empty());
|
||||
check(set_a.par_intersection(&set_a), 0..3);
|
||||
check(set_a.par_union(&set_a), 0..3);
|
||||
|
||||
check(set_a.par_difference(&set_b), 0..3);
|
||||
check(set_b.par_difference(&set_a), 3..6);
|
||||
check(set_a.par_symmetric_difference(&set_b), 0..6);
|
||||
check(set_b.par_symmetric_difference(&set_a), (3..6).chain(0..3));
|
||||
check(set_a.par_intersection(&set_b), empty());
|
||||
check(set_b.par_intersection(&set_a), empty());
|
||||
check(set_a.par_union(&set_b), 0..6);
|
||||
check(set_b.par_union(&set_a), (3..6).chain(0..3));
|
||||
|
||||
check(set_a.par_difference(&set_c), empty());
|
||||
check(set_c.par_difference(&set_a), 3..6);
|
||||
check(set_a.par_symmetric_difference(&set_c), 3..6);
|
||||
check(set_c.par_symmetric_difference(&set_a), 3..6);
|
||||
check(set_a.par_intersection(&set_c), 0..3);
|
||||
check(set_c.par_intersection(&set_a), 0..3);
|
||||
check(set_a.par_union(&set_c), 0..6);
|
||||
check(set_c.par_union(&set_a), 0..6);
|
||||
|
||||
check(set_c.par_difference(&set_d), 0..3);
|
||||
check(set_d.par_difference(&set_c), (6..9).rev());
|
||||
check(
|
||||
set_c.par_symmetric_difference(&set_d),
|
||||
(0..3).chain((6..9).rev()),
|
||||
);
|
||||
check(
|
||||
set_d.par_symmetric_difference(&set_c),
|
||||
(6..9).rev().chain(0..3),
|
||||
);
|
||||
check(set_c.par_intersection(&set_d), 3..6);
|
||||
check(set_d.par_intersection(&set_c), (3..6).rev());
|
||||
check(set_c.par_union(&set_d), (0..6).chain((6..9).rev()));
|
||||
check(set_d.par_union(&set_c), (3..9).rev().chain(0..3));
|
||||
}
|
||||
}
|
||||
158
clamav/libclamav_rust/.cargo/vendor/indexmap/src/rustc.rs
vendored
Normal file
158
clamav/libclamav_rust/.cargo/vendor/indexmap/src/rustc.rs
vendored
Normal file
@@ -0,0 +1,158 @@
|
||||
//! Minimal support for `rustc-rayon`, not intended for general use.
|
||||
|
||||
use crate::vec::Vec;
|
||||
use crate::{Bucket, Entries, IndexMap, IndexSet};
|
||||
|
||||
use rustc_rayon::iter::plumbing::{Consumer, ProducerCallback, UnindexedConsumer};
|
||||
use rustc_rayon::iter::{IndexedParallelIterator, IntoParallelIterator, ParallelIterator};
|
||||
|
||||
mod map {
|
||||
use super::*;
|
||||
|
||||
impl<K, V, S> IntoParallelIterator for IndexMap<K, V, S>
|
||||
where
|
||||
K: Send,
|
||||
V: Send,
|
||||
{
|
||||
type Item = (K, V);
|
||||
type Iter = IntoParIter<K, V>;
|
||||
|
||||
fn into_par_iter(self) -> Self::Iter {
|
||||
IntoParIter {
|
||||
entries: self.into_entries(),
|
||||
}
|
||||
}
|
||||
}
|
||||
|
||||
pub struct IntoParIter<K, V> {
|
||||
entries: Vec<Bucket<K, V>>,
|
||||
}
|
||||
|
||||
impl<K: Send, V: Send> ParallelIterator for IntoParIter<K, V> {
|
||||
type Item = (K, V);
|
||||
|
||||
parallel_iterator_methods!(Bucket::key_value);
|
||||
}
|
||||
|
||||
impl<K: Send, V: Send> IndexedParallelIterator for IntoParIter<K, V> {
|
||||
indexed_parallel_iterator_methods!(Bucket::key_value);
|
||||
}
|
||||
|
||||
impl<'a, K, V, S> IntoParallelIterator for &'a IndexMap<K, V, S>
|
||||
where
|
||||
K: Sync,
|
||||
V: Sync,
|
||||
{
|
||||
type Item = (&'a K, &'a V);
|
||||
type Iter = ParIter<'a, K, V>;
|
||||
|
||||
fn into_par_iter(self) -> Self::Iter {
|
||||
ParIter {
|
||||
entries: self.as_entries(),
|
||||
}
|
||||
}
|
||||
}
|
||||
|
||||
pub struct ParIter<'a, K, V> {
|
||||
entries: &'a [Bucket<K, V>],
|
||||
}
|
||||
|
||||
impl<'a, K: Sync, V: Sync> ParallelIterator for ParIter<'a, K, V> {
|
||||
type Item = (&'a K, &'a V);
|
||||
|
||||
parallel_iterator_methods!(Bucket::refs);
|
||||
}
|
||||
|
||||
impl<K: Sync, V: Sync> IndexedParallelIterator for ParIter<'_, K, V> {
|
||||
indexed_parallel_iterator_methods!(Bucket::refs);
|
||||
}
|
||||
|
||||
impl<'a, K, V, S> IntoParallelIterator for &'a mut IndexMap<K, V, S>
|
||||
where
|
||||
K: Sync + Send,
|
||||
V: Send,
|
||||
{
|
||||
type Item = (&'a K, &'a mut V);
|
||||
type Iter = ParIterMut<'a, K, V>;
|
||||
|
||||
fn into_par_iter(self) -> Self::Iter {
|
||||
ParIterMut {
|
||||
entries: self.as_entries_mut(),
|
||||
}
|
||||
}
|
||||
}
|
||||
|
||||
pub struct ParIterMut<'a, K, V> {
|
||||
entries: &'a mut [Bucket<K, V>],
|
||||
}
|
||||
|
||||
impl<'a, K: Sync + Send, V: Send> ParallelIterator for ParIterMut<'a, K, V> {
|
||||
type Item = (&'a K, &'a mut V);
|
||||
|
||||
parallel_iterator_methods!(Bucket::ref_mut);
|
||||
}
|
||||
|
||||
impl<K: Sync + Send, V: Send> IndexedParallelIterator for ParIterMut<'_, K, V> {
|
||||
indexed_parallel_iterator_methods!(Bucket::ref_mut);
|
||||
}
|
||||
}
|
||||
|
||||
mod set {
|
||||
use super::*;
|
||||
|
||||
impl<T, S> IntoParallelIterator for IndexSet<T, S>
|
||||
where
|
||||
T: Send,
|
||||
{
|
||||
type Item = T;
|
||||
type Iter = IntoParIter<T>;
|
||||
|
||||
fn into_par_iter(self) -> Self::Iter {
|
||||
IntoParIter {
|
||||
entries: self.into_entries(),
|
||||
}
|
||||
}
|
||||
}
|
||||
|
||||
pub struct IntoParIter<T> {
|
||||
entries: Vec<Bucket<T, ()>>,
|
||||
}
|
||||
|
||||
impl<T: Send> ParallelIterator for IntoParIter<T> {
|
||||
type Item = T;
|
||||
|
||||
parallel_iterator_methods!(Bucket::key);
|
||||
}
|
||||
|
||||
impl<T: Send> IndexedParallelIterator for IntoParIter<T> {
|
||||
indexed_parallel_iterator_methods!(Bucket::key);
|
||||
}
|
||||
|
||||
impl<'a, T, S> IntoParallelIterator for &'a IndexSet<T, S>
|
||||
where
|
||||
T: Sync,
|
||||
{
|
||||
type Item = &'a T;
|
||||
type Iter = ParIter<'a, T>;
|
||||
|
||||
fn into_par_iter(self) -> Self::Iter {
|
||||
ParIter {
|
||||
entries: self.as_entries(),
|
||||
}
|
||||
}
|
||||
}
|
||||
|
||||
pub struct ParIter<'a, T> {
|
||||
entries: &'a [Bucket<T, ()>],
|
||||
}
|
||||
|
||||
impl<'a, T: Sync> ParallelIterator for ParIter<'a, T> {
|
||||
type Item = &'a T;
|
||||
|
||||
parallel_iterator_methods!(Bucket::key_ref);
|
||||
}
|
||||
|
||||
impl<T: Sync> IndexedParallelIterator for ParIter<'_, T> {
|
||||
indexed_parallel_iterator_methods!(Bucket::key_ref);
|
||||
}
|
||||
}
|
||||
155
clamav/libclamav_rust/.cargo/vendor/indexmap/src/serde.rs
vendored
Normal file
155
clamav/libclamav_rust/.cargo/vendor/indexmap/src/serde.rs
vendored
Normal file
@@ -0,0 +1,155 @@
|
||||
use serde::de::value::{MapDeserializer, SeqDeserializer};
|
||||
use serde::de::{
|
||||
Deserialize, Deserializer, Error, IntoDeserializer, MapAccess, SeqAccess, Visitor,
|
||||
};
|
||||
use serde::ser::{Serialize, Serializer};
|
||||
|
||||
use core::fmt::{self, Formatter};
|
||||
use core::hash::{BuildHasher, Hash};
|
||||
use core::marker::PhantomData;
|
||||
|
||||
use crate::IndexMap;
|
||||
|
||||
/// Requires crate feature `"serde"` or `"serde-1"`
|
||||
impl<K, V, S> Serialize for IndexMap<K, V, S>
|
||||
where
|
||||
K: Serialize + Hash + Eq,
|
||||
V: Serialize,
|
||||
S: BuildHasher,
|
||||
{
|
||||
fn serialize<T>(&self, serializer: T) -> Result<T::Ok, T::Error>
|
||||
where
|
||||
T: Serializer,
|
||||
{
|
||||
serializer.collect_map(self)
|
||||
}
|
||||
}
|
||||
|
||||
struct IndexMapVisitor<K, V, S>(PhantomData<(K, V, S)>);
|
||||
|
||||
impl<'de, K, V, S> Visitor<'de> for IndexMapVisitor<K, V, S>
|
||||
where
|
||||
K: Deserialize<'de> + Eq + Hash,
|
||||
V: Deserialize<'de>,
|
||||
S: Default + BuildHasher,
|
||||
{
|
||||
type Value = IndexMap<K, V, S>;
|
||||
|
||||
fn expecting(&self, formatter: &mut Formatter<'_>) -> fmt::Result {
|
||||
write!(formatter, "a map")
|
||||
}
|
||||
|
||||
fn visit_map<A>(self, mut map: A) -> Result<Self::Value, A::Error>
|
||||
where
|
||||
A: MapAccess<'de>,
|
||||
{
|
||||
let mut values =
|
||||
IndexMap::with_capacity_and_hasher(map.size_hint().unwrap_or(0), S::default());
|
||||
|
||||
while let Some((key, value)) = map.next_entry()? {
|
||||
values.insert(key, value);
|
||||
}
|
||||
|
||||
Ok(values)
|
||||
}
|
||||
}
|
||||
|
||||
/// Requires crate feature `"serde"` or `"serde-1"`
|
||||
impl<'de, K, V, S> Deserialize<'de> for IndexMap<K, V, S>
|
||||
where
|
||||
K: Deserialize<'de> + Eq + Hash,
|
||||
V: Deserialize<'de>,
|
||||
S: Default + BuildHasher,
|
||||
{
|
||||
fn deserialize<D>(deserializer: D) -> Result<Self, D::Error>
|
||||
where
|
||||
D: Deserializer<'de>,
|
||||
{
|
||||
deserializer.deserialize_map(IndexMapVisitor(PhantomData))
|
||||
}
|
||||
}
|
||||
|
||||
impl<'de, K, V, S, E> IntoDeserializer<'de, E> for IndexMap<K, V, S>
|
||||
where
|
||||
K: IntoDeserializer<'de, E> + Eq + Hash,
|
||||
V: IntoDeserializer<'de, E>,
|
||||
S: BuildHasher,
|
||||
E: Error,
|
||||
{
|
||||
type Deserializer = MapDeserializer<'de, <Self as IntoIterator>::IntoIter, E>;
|
||||
|
||||
fn into_deserializer(self) -> Self::Deserializer {
|
||||
MapDeserializer::new(self.into_iter())
|
||||
}
|
||||
}
|
||||
|
||||
use crate::IndexSet;
|
||||
|
||||
/// Requires crate feature `"serde"` or `"serde-1"`
|
||||
impl<T, S> Serialize for IndexSet<T, S>
|
||||
where
|
||||
T: Serialize + Hash + Eq,
|
||||
S: BuildHasher,
|
||||
{
|
||||
fn serialize<Se>(&self, serializer: Se) -> Result<Se::Ok, Se::Error>
|
||||
where
|
||||
Se: Serializer,
|
||||
{
|
||||
serializer.collect_seq(self)
|
||||
}
|
||||
}
|
||||
|
||||
struct IndexSetVisitor<T, S>(PhantomData<(T, S)>);
|
||||
|
||||
impl<'de, T, S> Visitor<'de> for IndexSetVisitor<T, S>
|
||||
where
|
||||
T: Deserialize<'de> + Eq + Hash,
|
||||
S: Default + BuildHasher,
|
||||
{
|
||||
type Value = IndexSet<T, S>;
|
||||
|
||||
fn expecting(&self, formatter: &mut Formatter<'_>) -> fmt::Result {
|
||||
write!(formatter, "a set")
|
||||
}
|
||||
|
||||
fn visit_seq<A>(self, mut seq: A) -> Result<Self::Value, A::Error>
|
||||
where
|
||||
A: SeqAccess<'de>,
|
||||
{
|
||||
let mut values =
|
||||
IndexSet::with_capacity_and_hasher(seq.size_hint().unwrap_or(0), S::default());
|
||||
|
||||
while let Some(value) = seq.next_element()? {
|
||||
values.insert(value);
|
||||
}
|
||||
|
||||
Ok(values)
|
||||
}
|
||||
}
|
||||
|
||||
/// Requires crate feature `"serde"` or `"serde-1"`
|
||||
impl<'de, T, S> Deserialize<'de> for IndexSet<T, S>
|
||||
where
|
||||
T: Deserialize<'de> + Eq + Hash,
|
||||
S: Default + BuildHasher,
|
||||
{
|
||||
fn deserialize<D>(deserializer: D) -> Result<Self, D::Error>
|
||||
where
|
||||
D: Deserializer<'de>,
|
||||
{
|
||||
deserializer.deserialize_seq(IndexSetVisitor(PhantomData))
|
||||
}
|
||||
}
|
||||
|
||||
impl<'de, T, S, E> IntoDeserializer<'de, E> for IndexSet<T, S>
|
||||
where
|
||||
T: IntoDeserializer<'de, E> + Eq + Hash,
|
||||
S: BuildHasher,
|
||||
E: Error,
|
||||
{
|
||||
type Deserializer = SeqDeserializer<<Self as IntoIterator>::IntoIter, E>;
|
||||
|
||||
fn into_deserializer(self) -> Self::Deserializer {
|
||||
SeqDeserializer::new(self.into_iter())
|
||||
}
|
||||
}
|
||||
112
clamav/libclamav_rust/.cargo/vendor/indexmap/src/serde_seq.rs
vendored
Normal file
112
clamav/libclamav_rust/.cargo/vendor/indexmap/src/serde_seq.rs
vendored
Normal file
@@ -0,0 +1,112 @@
|
||||
//! Functions to serialize and deserialize an `IndexMap` as an ordered sequence.
|
||||
//!
|
||||
//! The default `serde` implementation serializes `IndexMap` as a normal map,
|
||||
//! but there is no guarantee that serialization formats will preserve the order
|
||||
//! of the key-value pairs. This module serializes `IndexMap` as a sequence of
|
||||
//! `(key, value)` elements instead, in order.
|
||||
//!
|
||||
//! This module may be used in a field attribute for derived implementations:
|
||||
//!
|
||||
//! ```
|
||||
//! # use indexmap::IndexMap;
|
||||
//! # use serde_derive::{Deserialize, Serialize};
|
||||
//! #[derive(Deserialize, Serialize)]
|
||||
//! struct Data {
|
||||
//! #[serde(with = "indexmap::serde_seq")]
|
||||
//! map: IndexMap<i32, u64>,
|
||||
//! // ...
|
||||
//! }
|
||||
//! ```
|
||||
//!
|
||||
//! Requires crate feature `"serde"` or `"serde-1"`
|
||||
|
||||
use serde::de::{Deserialize, Deserializer, SeqAccess, Visitor};
|
||||
use serde::ser::{Serialize, Serializer};
|
||||
|
||||
use core::fmt::{self, Formatter};
|
||||
use core::hash::{BuildHasher, Hash};
|
||||
use core::marker::PhantomData;
|
||||
|
||||
use crate::IndexMap;
|
||||
|
||||
/// Serializes an `IndexMap` as an ordered sequence.
|
||||
///
|
||||
/// This function may be used in a field attribute for deriving `Serialize`:
|
||||
///
|
||||
/// ```
|
||||
/// # use indexmap::IndexMap;
|
||||
/// # use serde_derive::Serialize;
|
||||
/// #[derive(Serialize)]
|
||||
/// struct Data {
|
||||
/// #[serde(serialize_with = "indexmap::serde_seq::serialize")]
|
||||
/// map: IndexMap<i32, u64>,
|
||||
/// // ...
|
||||
/// }
|
||||
/// ```
|
||||
///
|
||||
/// Requires crate feature `"serde"` or `"serde-1"`
|
||||
pub fn serialize<K, V, S, T>(map: &IndexMap<K, V, S>, serializer: T) -> Result<T::Ok, T::Error>
|
||||
where
|
||||
K: Serialize + Hash + Eq,
|
||||
V: Serialize,
|
||||
S: BuildHasher,
|
||||
T: Serializer,
|
||||
{
|
||||
serializer.collect_seq(map)
|
||||
}
|
||||
|
||||
/// Visitor to deserialize a *sequenced* `IndexMap`
|
||||
struct SeqVisitor<K, V, S>(PhantomData<(K, V, S)>);
|
||||
|
||||
impl<'de, K, V, S> Visitor<'de> for SeqVisitor<K, V, S>
|
||||
where
|
||||
K: Deserialize<'de> + Eq + Hash,
|
||||
V: Deserialize<'de>,
|
||||
S: Default + BuildHasher,
|
||||
{
|
||||
type Value = IndexMap<K, V, S>;
|
||||
|
||||
fn expecting(&self, formatter: &mut Formatter<'_>) -> fmt::Result {
|
||||
write!(formatter, "a sequenced map")
|
||||
}
|
||||
|
||||
fn visit_seq<A>(self, mut seq: A) -> Result<Self::Value, A::Error>
|
||||
where
|
||||
A: SeqAccess<'de>,
|
||||
{
|
||||
let capacity = seq.size_hint().unwrap_or(0);
|
||||
let mut map = IndexMap::with_capacity_and_hasher(capacity, S::default());
|
||||
|
||||
while let Some((key, value)) = seq.next_element()? {
|
||||
map.insert(key, value);
|
||||
}
|
||||
|
||||
Ok(map)
|
||||
}
|
||||
}
|
||||
|
||||
/// Deserializes an `IndexMap` from an ordered sequence.
|
||||
///
|
||||
/// This function may be used in a field attribute for deriving `Deserialize`:
|
||||
///
|
||||
/// ```
|
||||
/// # use indexmap::IndexMap;
|
||||
/// # use serde_derive::Deserialize;
|
||||
/// #[derive(Deserialize)]
|
||||
/// struct Data {
|
||||
/// #[serde(deserialize_with = "indexmap::serde_seq::deserialize")]
|
||||
/// map: IndexMap<i32, u64>,
|
||||
/// // ...
|
||||
/// }
|
||||
/// ```
|
||||
///
|
||||
/// Requires crate feature `"serde"` or `"serde-1"`
|
||||
pub fn deserialize<'de, D, K, V, S>(deserializer: D) -> Result<IndexMap<K, V, S>, D::Error>
|
||||
where
|
||||
D: Deserializer<'de>,
|
||||
K: Deserialize<'de> + Eq + Hash,
|
||||
V: Deserialize<'de>,
|
||||
S: Default + BuildHasher,
|
||||
{
|
||||
deserializer.deserialize_seq(SeqVisitor(PhantomData))
|
||||
}
|
||||
1912
clamav/libclamav_rust/.cargo/vendor/indexmap/src/set.rs
vendored
Normal file
1912
clamav/libclamav_rust/.cargo/vendor/indexmap/src/set.rs
vendored
Normal file
File diff suppressed because it is too large
Load Diff
31
clamav/libclamav_rust/.cargo/vendor/indexmap/src/util.rs
vendored
Normal file
31
clamav/libclamav_rust/.cargo/vendor/indexmap/src/util.rs
vendored
Normal file
@@ -0,0 +1,31 @@
|
||||
use core::ops::{Bound, Range, RangeBounds};
|
||||
|
||||
pub(crate) fn third<A, B, C>(t: (A, B, C)) -> C {
|
||||
t.2
|
||||
}
|
||||
|
||||
pub(crate) fn simplify_range<R>(range: R, len: usize) -> Range<usize>
|
||||
where
|
||||
R: RangeBounds<usize>,
|
||||
{
|
||||
let start = match range.start_bound() {
|
||||
Bound::Unbounded => 0,
|
||||
Bound::Included(&i) if i <= len => i,
|
||||
Bound::Excluded(&i) if i < len => i + 1,
|
||||
bound => panic!("range start {:?} should be <= length {}", bound, len),
|
||||
};
|
||||
let end = match range.end_bound() {
|
||||
Bound::Unbounded => len,
|
||||
Bound::Excluded(&i) if i <= len => i,
|
||||
Bound::Included(&i) if i < len => i + 1,
|
||||
bound => panic!("range end {:?} should be <= length {}", bound, len),
|
||||
};
|
||||
if start > end {
|
||||
panic!(
|
||||
"range start {:?} should be <= range end {:?}",
|
||||
range.start_bound(),
|
||||
range.end_bound()
|
||||
);
|
||||
}
|
||||
start..end
|
||||
}
|
||||
53
clamav/libclamav_rust/.cargo/vendor/indexmap/tests/equivalent_trait.rs
vendored
Normal file
53
clamav/libclamav_rust/.cargo/vendor/indexmap/tests/equivalent_trait.rs
vendored
Normal file
@@ -0,0 +1,53 @@
|
||||
use indexmap::indexmap;
|
||||
use indexmap::Equivalent;
|
||||
|
||||
use std::hash::Hash;
|
||||
|
||||
#[derive(Debug, Hash)]
|
||||
pub struct Pair<A, B>(pub A, pub B);
|
||||
|
||||
impl<A, B, C, D> PartialEq<(A, B)> for Pair<C, D>
|
||||
where
|
||||
C: PartialEq<A>,
|
||||
D: PartialEq<B>,
|
||||
{
|
||||
fn eq(&self, rhs: &(A, B)) -> bool {
|
||||
self.0 == rhs.0 && self.1 == rhs.1
|
||||
}
|
||||
}
|
||||
|
||||
impl<A, B, X> Equivalent<X> for Pair<A, B>
|
||||
where
|
||||
Pair<A, B>: PartialEq<X>,
|
||||
A: Hash + Eq,
|
||||
B: Hash + Eq,
|
||||
{
|
||||
fn equivalent(&self, other: &X) -> bool {
|
||||
*self == *other
|
||||
}
|
||||
}
|
||||
|
||||
#[test]
|
||||
fn test_lookup() {
|
||||
let s = String::from;
|
||||
let map = indexmap! {
|
||||
(s("a"), s("b")) => 1,
|
||||
(s("a"), s("x")) => 2,
|
||||
};
|
||||
|
||||
assert!(map.contains_key(&Pair("a", "b")));
|
||||
assert!(!map.contains_key(&Pair("b", "a")));
|
||||
}
|
||||
|
||||
#[test]
|
||||
fn test_string_str() {
|
||||
let s = String::from;
|
||||
let mut map = indexmap! {
|
||||
s("a") => 1, s("b") => 2,
|
||||
s("x") => 3, s("y") => 4,
|
||||
};
|
||||
|
||||
assert!(map.contains_key("a"));
|
||||
assert!(!map.contains_key("z"));
|
||||
assert_eq!(map.swap_remove("b"), Some(2));
|
||||
}
|
||||
19
clamav/libclamav_rust/.cargo/vendor/indexmap/tests/macros_full_path.rs
vendored
Normal file
19
clamav/libclamav_rust/.cargo/vendor/indexmap/tests/macros_full_path.rs
vendored
Normal file
@@ -0,0 +1,19 @@
|
||||
#[test]
|
||||
fn test_create_map() {
|
||||
let _m = indexmap::indexmap! {
|
||||
1 => 2,
|
||||
7 => 1,
|
||||
2 => 2,
|
||||
3 => 3,
|
||||
};
|
||||
}
|
||||
|
||||
#[test]
|
||||
fn test_create_set() {
|
||||
let _s = indexmap::indexset! {
|
||||
1,
|
||||
7,
|
||||
2,
|
||||
3,
|
||||
};
|
||||
}
|
||||
573
clamav/libclamav_rust/.cargo/vendor/indexmap/tests/quick.rs
vendored
Normal file
573
clamav/libclamav_rust/.cargo/vendor/indexmap/tests/quick.rs
vendored
Normal file
@@ -0,0 +1,573 @@
|
||||
use indexmap::{IndexMap, IndexSet};
|
||||
use itertools::Itertools;
|
||||
|
||||
use quickcheck::Arbitrary;
|
||||
use quickcheck::Gen;
|
||||
use quickcheck::QuickCheck;
|
||||
use quickcheck::TestResult;
|
||||
|
||||
use fnv::FnvHasher;
|
||||
use std::hash::{BuildHasher, BuildHasherDefault};
|
||||
type FnvBuilder = BuildHasherDefault<FnvHasher>;
|
||||
type IndexMapFnv<K, V> = IndexMap<K, V, FnvBuilder>;
|
||||
|
||||
use std::cmp::min;
|
||||
use std::collections::HashMap;
|
||||
use std::collections::HashSet;
|
||||
use std::fmt::Debug;
|
||||
use std::hash::Hash;
|
||||
use std::ops::Bound;
|
||||
use std::ops::Deref;
|
||||
|
||||
use indexmap::map::Entry as OEntry;
|
||||
use std::collections::hash_map::Entry as HEntry;
|
||||
|
||||
fn set<'a, T: 'a, I>(iter: I) -> HashSet<T>
|
||||
where
|
||||
I: IntoIterator<Item = &'a T>,
|
||||
T: Copy + Hash + Eq,
|
||||
{
|
||||
iter.into_iter().copied().collect()
|
||||
}
|
||||
|
||||
fn indexmap<'a, T: 'a, I>(iter: I) -> IndexMap<T, ()>
|
||||
where
|
||||
I: IntoIterator<Item = &'a T>,
|
||||
T: Copy + Hash + Eq,
|
||||
{
|
||||
IndexMap::from_iter(iter.into_iter().copied().map(|k| (k, ())))
|
||||
}
|
||||
|
||||
// Helper macro to allow us to use smaller quickcheck limits under miri.
|
||||
macro_rules! quickcheck_limit {
|
||||
(@as_items $($i:item)*) => ($($i)*);
|
||||
{
|
||||
$(
|
||||
$(#[$m:meta])*
|
||||
fn $fn_name:ident($($arg_name:ident : $arg_ty:ty),*) -> $ret:ty {
|
||||
$($code:tt)*
|
||||
}
|
||||
)*
|
||||
} => (
|
||||
quickcheck::quickcheck! {
|
||||
@as_items
|
||||
$(
|
||||
#[test]
|
||||
$(#[$m])*
|
||||
fn $fn_name() {
|
||||
fn prop($($arg_name: $arg_ty),*) -> $ret {
|
||||
$($code)*
|
||||
}
|
||||
let mut quickcheck = QuickCheck::new();
|
||||
if cfg!(miri) {
|
||||
quickcheck = quickcheck
|
||||
.gen(Gen::new(10))
|
||||
.tests(10)
|
||||
.max_tests(100);
|
||||
}
|
||||
|
||||
quickcheck.quickcheck(prop as fn($($arg_ty),*) -> $ret);
|
||||
}
|
||||
)*
|
||||
}
|
||||
)
|
||||
}
|
||||
|
||||
quickcheck_limit! {
|
||||
fn contains(insert: Vec<u32>) -> bool {
|
||||
let mut map = IndexMap::new();
|
||||
for &key in &insert {
|
||||
map.insert(key, ());
|
||||
}
|
||||
insert.iter().all(|&key| map.get(&key).is_some())
|
||||
}
|
||||
|
||||
fn contains_not(insert: Vec<u8>, not: Vec<u8>) -> bool {
|
||||
let mut map = IndexMap::new();
|
||||
for &key in &insert {
|
||||
map.insert(key, ());
|
||||
}
|
||||
let nots = &set(¬) - &set(&insert);
|
||||
nots.iter().all(|&key| map.get(&key).is_none())
|
||||
}
|
||||
|
||||
fn insert_remove(insert: Vec<u8>, remove: Vec<u8>) -> bool {
|
||||
let mut map = IndexMap::new();
|
||||
for &key in &insert {
|
||||
map.insert(key, ());
|
||||
}
|
||||
for &key in &remove {
|
||||
map.swap_remove(&key);
|
||||
}
|
||||
let elements = &set(&insert) - &set(&remove);
|
||||
map.len() == elements.len() && map.iter().count() == elements.len() &&
|
||||
elements.iter().all(|k| map.get(k).is_some())
|
||||
}
|
||||
|
||||
fn insertion_order(insert: Vec<u32>) -> bool {
|
||||
let mut map = IndexMap::new();
|
||||
for &key in &insert {
|
||||
map.insert(key, ());
|
||||
}
|
||||
itertools::assert_equal(insert.iter().unique(), map.keys());
|
||||
true
|
||||
}
|
||||
|
||||
fn pop(insert: Vec<u8>) -> bool {
|
||||
let mut map = IndexMap::new();
|
||||
for &key in &insert {
|
||||
map.insert(key, ());
|
||||
}
|
||||
let mut pops = Vec::new();
|
||||
while let Some((key, _v)) = map.pop() {
|
||||
pops.push(key);
|
||||
}
|
||||
pops.reverse();
|
||||
|
||||
itertools::assert_equal(insert.iter().unique(), &pops);
|
||||
true
|
||||
}
|
||||
|
||||
fn with_cap(template: Vec<()>) -> bool {
|
||||
let cap = template.len();
|
||||
let map: IndexMap<u8, u8> = IndexMap::with_capacity(cap);
|
||||
println!("wish: {}, got: {} (diff: {})", cap, map.capacity(), map.capacity() as isize - cap as isize);
|
||||
map.capacity() >= cap
|
||||
}
|
||||
|
||||
fn drain_full(insert: Vec<u8>) -> bool {
|
||||
let mut map = IndexMap::new();
|
||||
for &key in &insert {
|
||||
map.insert(key, ());
|
||||
}
|
||||
let mut clone = map.clone();
|
||||
let drained = clone.drain(..);
|
||||
for (key, _) in drained {
|
||||
map.swap_remove(&key);
|
||||
}
|
||||
map.is_empty()
|
||||
}
|
||||
|
||||
fn drain_bounds(insert: Vec<u8>, range: (Bound<usize>, Bound<usize>)) -> TestResult {
|
||||
let mut map = IndexMap::new();
|
||||
for &key in &insert {
|
||||
map.insert(key, ());
|
||||
}
|
||||
|
||||
// First see if `Vec::drain` is happy with this range.
|
||||
let result = std::panic::catch_unwind(|| {
|
||||
let mut keys: Vec<u8> = map.keys().copied().collect();
|
||||
keys.drain(range);
|
||||
keys
|
||||
});
|
||||
|
||||
if let Ok(keys) = result {
|
||||
map.drain(range);
|
||||
// Check that our `drain` matches the same key order.
|
||||
assert!(map.keys().eq(&keys));
|
||||
// Check that hash lookups all work too.
|
||||
assert!(keys.iter().all(|key| map.contains_key(key)));
|
||||
TestResult::passed()
|
||||
} else {
|
||||
// If `Vec::drain` panicked, so should we.
|
||||
TestResult::must_fail(move || { map.drain(range); })
|
||||
}
|
||||
}
|
||||
|
||||
fn shift_remove(insert: Vec<u8>, remove: Vec<u8>) -> bool {
|
||||
let mut map = IndexMap::new();
|
||||
for &key in &insert {
|
||||
map.insert(key, ());
|
||||
}
|
||||
for &key in &remove {
|
||||
map.shift_remove(&key);
|
||||
}
|
||||
let elements = &set(&insert) - &set(&remove);
|
||||
|
||||
// Check that order is preserved after removals
|
||||
let mut iter = map.keys();
|
||||
for &key in insert.iter().unique() {
|
||||
if elements.contains(&key) {
|
||||
assert_eq!(Some(&key), iter.next());
|
||||
}
|
||||
}
|
||||
|
||||
map.len() == elements.len() && map.iter().count() == elements.len() &&
|
||||
elements.iter().all(|k| map.get(k).is_some())
|
||||
}
|
||||
|
||||
fn indexing(insert: Vec<u8>) -> bool {
|
||||
let mut map: IndexMap<_, _> = insert.into_iter().map(|x| (x, x)).collect();
|
||||
let set: IndexSet<_> = map.keys().copied().collect();
|
||||
assert_eq!(map.len(), set.len());
|
||||
|
||||
for (i, &key) in set.iter().enumerate() {
|
||||
assert_eq!(map.get_index(i), Some((&key, &key)));
|
||||
assert_eq!(set.get_index(i), Some(&key));
|
||||
assert_eq!(map[i], key);
|
||||
assert_eq!(set[i], key);
|
||||
|
||||
*map.get_index_mut(i).unwrap().1 >>= 1;
|
||||
map[i] <<= 1;
|
||||
}
|
||||
|
||||
set.iter().enumerate().all(|(i, &key)| {
|
||||
let value = key & !1;
|
||||
map[&key] == value && map[i] == value
|
||||
})
|
||||
}
|
||||
|
||||
// Use `u8` test indices so quickcheck is less likely to go out of bounds.
|
||||
fn swap_indices(vec: Vec<u8>, a: u8, b: u8) -> TestResult {
|
||||
let mut set = IndexSet::<u8>::from_iter(vec);
|
||||
let a = usize::from(a);
|
||||
let b = usize::from(b);
|
||||
|
||||
if a >= set.len() || b >= set.len() {
|
||||
return TestResult::discard();
|
||||
}
|
||||
|
||||
let mut vec = Vec::from_iter(set.iter().cloned());
|
||||
vec.swap(a, b);
|
||||
|
||||
set.swap_indices(a, b);
|
||||
|
||||
// Check both iteration order and hash lookups
|
||||
assert!(set.iter().eq(vec.iter()));
|
||||
assert!(vec.iter().enumerate().all(|(i, x)| {
|
||||
set.get_index_of(x) == Some(i)
|
||||
}));
|
||||
TestResult::passed()
|
||||
}
|
||||
|
||||
// Use `u8` test indices so quickcheck is less likely to go out of bounds.
|
||||
fn move_index(vec: Vec<u8>, from: u8, to: u8) -> TestResult {
|
||||
let mut set = IndexSet::<u8>::from_iter(vec);
|
||||
let from = usize::from(from);
|
||||
let to = usize::from(to);
|
||||
|
||||
if from >= set.len() || to >= set.len() {
|
||||
return TestResult::discard();
|
||||
}
|
||||
|
||||
let mut vec = Vec::from_iter(set.iter().cloned());
|
||||
let x = vec.remove(from);
|
||||
vec.insert(to, x);
|
||||
|
||||
set.move_index(from, to);
|
||||
|
||||
// Check both iteration order and hash lookups
|
||||
assert!(set.iter().eq(vec.iter()));
|
||||
assert!(vec.iter().enumerate().all(|(i, x)| {
|
||||
set.get_index_of(x) == Some(i)
|
||||
}));
|
||||
TestResult::passed()
|
||||
}
|
||||
}
|
||||
|
||||
use crate::Op::*;
|
||||
#[derive(Copy, Clone, Debug)]
|
||||
enum Op<K, V> {
|
||||
Add(K, V),
|
||||
Remove(K),
|
||||
AddEntry(K, V),
|
||||
RemoveEntry(K),
|
||||
}
|
||||
|
||||
impl<K, V> Arbitrary for Op<K, V>
|
||||
where
|
||||
K: Arbitrary,
|
||||
V: Arbitrary,
|
||||
{
|
||||
fn arbitrary(g: &mut Gen) -> Self {
|
||||
match u32::arbitrary(g) % 4 {
|
||||
0 => Add(K::arbitrary(g), V::arbitrary(g)),
|
||||
1 => AddEntry(K::arbitrary(g), V::arbitrary(g)),
|
||||
2 => Remove(K::arbitrary(g)),
|
||||
_ => RemoveEntry(K::arbitrary(g)),
|
||||
}
|
||||
}
|
||||
}
|
||||
|
||||
fn do_ops<K, V, S>(ops: &[Op<K, V>], a: &mut IndexMap<K, V, S>, b: &mut HashMap<K, V>)
|
||||
where
|
||||
K: Hash + Eq + Clone,
|
||||
V: Clone,
|
||||
S: BuildHasher,
|
||||
{
|
||||
for op in ops {
|
||||
match *op {
|
||||
Add(ref k, ref v) => {
|
||||
a.insert(k.clone(), v.clone());
|
||||
b.insert(k.clone(), v.clone());
|
||||
}
|
||||
AddEntry(ref k, ref v) => {
|
||||
a.entry(k.clone()).or_insert_with(|| v.clone());
|
||||
b.entry(k.clone()).or_insert_with(|| v.clone());
|
||||
}
|
||||
Remove(ref k) => {
|
||||
a.swap_remove(k);
|
||||
b.remove(k);
|
||||
}
|
||||
RemoveEntry(ref k) => {
|
||||
if let OEntry::Occupied(ent) = a.entry(k.clone()) {
|
||||
ent.swap_remove_entry();
|
||||
}
|
||||
if let HEntry::Occupied(ent) = b.entry(k.clone()) {
|
||||
ent.remove_entry();
|
||||
}
|
||||
}
|
||||
}
|
||||
//println!("{:?}", a);
|
||||
}
|
||||
}
|
||||
|
||||
fn assert_maps_equivalent<K, V>(a: &IndexMap<K, V>, b: &HashMap<K, V>) -> bool
|
||||
where
|
||||
K: Hash + Eq + Debug,
|
||||
V: Eq + Debug,
|
||||
{
|
||||
assert_eq!(a.len(), b.len());
|
||||
assert_eq!(a.iter().next().is_some(), b.iter().next().is_some());
|
||||
for key in a.keys() {
|
||||
assert!(b.contains_key(key), "b does not contain {:?}", key);
|
||||
}
|
||||
for key in b.keys() {
|
||||
assert!(a.get(key).is_some(), "a does not contain {:?}", key);
|
||||
}
|
||||
for key in a.keys() {
|
||||
assert_eq!(a[key], b[key]);
|
||||
}
|
||||
true
|
||||
}
|
||||
|
||||
quickcheck_limit! {
|
||||
fn operations_i8(ops: Large<Vec<Op<i8, i8>>>) -> bool {
|
||||
let mut map = IndexMap::new();
|
||||
let mut reference = HashMap::new();
|
||||
do_ops(&ops, &mut map, &mut reference);
|
||||
assert_maps_equivalent(&map, &reference)
|
||||
}
|
||||
|
||||
fn operations_string(ops: Vec<Op<Alpha, i8>>) -> bool {
|
||||
let mut map = IndexMap::new();
|
||||
let mut reference = HashMap::new();
|
||||
do_ops(&ops, &mut map, &mut reference);
|
||||
assert_maps_equivalent(&map, &reference)
|
||||
}
|
||||
|
||||
fn keys_values(ops: Large<Vec<Op<i8, i8>>>) -> bool {
|
||||
let mut map = IndexMap::new();
|
||||
let mut reference = HashMap::new();
|
||||
do_ops(&ops, &mut map, &mut reference);
|
||||
let mut visit = IndexMap::new();
|
||||
for (k, v) in map.keys().zip(map.values()) {
|
||||
assert_eq!(&map[k], v);
|
||||
assert!(!visit.contains_key(k));
|
||||
visit.insert(*k, *v);
|
||||
}
|
||||
assert_eq!(visit.len(), reference.len());
|
||||
true
|
||||
}
|
||||
|
||||
fn keys_values_mut(ops: Large<Vec<Op<i8, i8>>>) -> bool {
|
||||
let mut map = IndexMap::new();
|
||||
let mut reference = HashMap::new();
|
||||
do_ops(&ops, &mut map, &mut reference);
|
||||
let mut visit = IndexMap::new();
|
||||
let keys = Vec::from_iter(map.keys().copied());
|
||||
for (k, v) in keys.iter().zip(map.values_mut()) {
|
||||
assert_eq!(&reference[k], v);
|
||||
assert!(!visit.contains_key(k));
|
||||
visit.insert(*k, *v);
|
||||
}
|
||||
assert_eq!(visit.len(), reference.len());
|
||||
true
|
||||
}
|
||||
|
||||
fn equality(ops1: Vec<Op<i8, i8>>, removes: Vec<usize>) -> bool {
|
||||
let mut map = IndexMap::new();
|
||||
let mut reference = HashMap::new();
|
||||
do_ops(&ops1, &mut map, &mut reference);
|
||||
let mut ops2 = ops1.clone();
|
||||
for &r in &removes {
|
||||
if !ops2.is_empty() {
|
||||
let i = r % ops2.len();
|
||||
ops2.remove(i);
|
||||
}
|
||||
}
|
||||
let mut map2 = IndexMapFnv::default();
|
||||
let mut reference2 = HashMap::new();
|
||||
do_ops(&ops2, &mut map2, &mut reference2);
|
||||
assert_eq!(map == map2, reference == reference2);
|
||||
true
|
||||
}
|
||||
|
||||
fn retain_ordered(keys: Large<Vec<i8>>, remove: Large<Vec<i8>>) -> () {
|
||||
let mut map = indexmap(keys.iter());
|
||||
let initial_map = map.clone(); // deduplicated in-order input
|
||||
let remove_map = indexmap(remove.iter());
|
||||
let keys_s = set(keys.iter());
|
||||
let remove_s = set(remove.iter());
|
||||
let answer = &keys_s - &remove_s;
|
||||
map.retain(|k, _| !remove_map.contains_key(k));
|
||||
|
||||
// check the values
|
||||
assert_eq!(map.len(), answer.len());
|
||||
for key in &answer {
|
||||
assert!(map.contains_key(key));
|
||||
}
|
||||
// check the order
|
||||
itertools::assert_equal(map.keys(), initial_map.keys().filter(|&k| !remove_map.contains_key(k)));
|
||||
}
|
||||
|
||||
fn sort_1(keyvals: Large<Vec<(i8, i8)>>) -> () {
|
||||
let mut map: IndexMap<_, _> = IndexMap::from_iter(keyvals.to_vec());
|
||||
let mut answer = keyvals.0;
|
||||
answer.sort_by_key(|t| t.0);
|
||||
|
||||
// reverse dedup: Because IndexMap::from_iter keeps the last value for
|
||||
// identical keys
|
||||
answer.reverse();
|
||||
answer.dedup_by_key(|t| t.0);
|
||||
answer.reverse();
|
||||
|
||||
map.sort_by(|k1, _, k2, _| Ord::cmp(k1, k2));
|
||||
|
||||
// check it contains all the values it should
|
||||
for &(key, val) in &answer {
|
||||
assert_eq!(map[&key], val);
|
||||
}
|
||||
|
||||
// check the order
|
||||
|
||||
let mapv = Vec::from_iter(map);
|
||||
assert_eq!(answer, mapv);
|
||||
|
||||
}
|
||||
|
||||
fn sort_2(keyvals: Large<Vec<(i8, i8)>>) -> () {
|
||||
let mut map: IndexMap<_, _> = IndexMap::from_iter(keyvals.to_vec());
|
||||
map.sort_by(|_, v1, _, v2| Ord::cmp(v1, v2));
|
||||
assert_sorted_by_key(map, |t| t.1);
|
||||
}
|
||||
|
||||
fn reverse(keyvals: Large<Vec<(i8, i8)>>) -> () {
|
||||
let mut map: IndexMap<_, _> = IndexMap::from_iter(keyvals.to_vec());
|
||||
|
||||
fn generate_answer(input: &Vec<(i8, i8)>) -> Vec<(i8, i8)> {
|
||||
// to mimic what `IndexMap::from_iter` does:
|
||||
// need to get (A) the unique keys in forward order, and (B) the
|
||||
// last value of each of those keys.
|
||||
|
||||
// create (A): an iterable that yields the unique keys in ltr order
|
||||
let mut seen_keys = HashSet::new();
|
||||
let unique_keys_forward = input.iter().filter_map(move |(k, _)| {
|
||||
if seen_keys.contains(k) { None }
|
||||
else { seen_keys.insert(*k); Some(*k) }
|
||||
});
|
||||
|
||||
// create (B): a mapping of keys to the last value seen for that key
|
||||
// this is the same as reversing the input and taking the first
|
||||
// value seen for that key!
|
||||
let mut last_val_per_key = HashMap::new();
|
||||
for &(k, v) in input.iter().rev() {
|
||||
if !last_val_per_key.contains_key(&k) {
|
||||
last_val_per_key.insert(k, v);
|
||||
}
|
||||
}
|
||||
|
||||
// iterate over the keys in (A) in order, and match each one with
|
||||
// the corresponding last value from (B)
|
||||
let mut ans: Vec<_> = unique_keys_forward
|
||||
.map(|k| (k, *last_val_per_key.get(&k).unwrap()))
|
||||
.collect();
|
||||
|
||||
// finally, since this test is testing `.reverse()`, reverse the
|
||||
// answer in-place
|
||||
ans.reverse();
|
||||
|
||||
ans
|
||||
}
|
||||
|
||||
let answer = generate_answer(&keyvals.0);
|
||||
|
||||
// perform the work
|
||||
map.reverse();
|
||||
|
||||
// check it contains all the values it should
|
||||
for &(key, val) in &answer {
|
||||
assert_eq!(map[&key], val);
|
||||
}
|
||||
|
||||
// check the order
|
||||
let mapv = Vec::from_iter(map);
|
||||
assert_eq!(answer, mapv);
|
||||
}
|
||||
}
|
||||
|
||||
fn assert_sorted_by_key<I, Key, X>(iterable: I, key: Key)
|
||||
where
|
||||
I: IntoIterator,
|
||||
I::Item: Ord + Clone + Debug,
|
||||
Key: Fn(&I::Item) -> X,
|
||||
X: Ord,
|
||||
{
|
||||
let input = Vec::from_iter(iterable);
|
||||
let mut sorted = input.clone();
|
||||
sorted.sort_by_key(key);
|
||||
assert_eq!(input, sorted);
|
||||
}
|
||||
|
||||
#[derive(Clone, Debug, Hash, PartialEq, Eq)]
|
||||
struct Alpha(String);
|
||||
|
||||
impl Deref for Alpha {
|
||||
type Target = String;
|
||||
fn deref(&self) -> &String {
|
||||
&self.0
|
||||
}
|
||||
}
|
||||
|
||||
const ALPHABET: &[u8] = b"abcdefghijklmnopqrstuvwxyz";
|
||||
|
||||
impl Arbitrary for Alpha {
|
||||
fn arbitrary(g: &mut Gen) -> Self {
|
||||
let len = usize::arbitrary(g) % g.size();
|
||||
let len = min(len, 16);
|
||||
Alpha(
|
||||
(0..len)
|
||||
.map(|_| ALPHABET[usize::arbitrary(g) % ALPHABET.len()] as char)
|
||||
.collect(),
|
||||
)
|
||||
}
|
||||
|
||||
fn shrink(&self) -> Box<dyn Iterator<Item = Self>> {
|
||||
Box::new((**self).shrink().map(Alpha))
|
||||
}
|
||||
}
|
||||
|
||||
/// quickcheck Arbitrary adaptor -- make a larger vec
|
||||
#[derive(Clone, Debug)]
|
||||
struct Large<T>(T);
|
||||
|
||||
impl<T> Deref for Large<T> {
|
||||
type Target = T;
|
||||
fn deref(&self) -> &T {
|
||||
&self.0
|
||||
}
|
||||
}
|
||||
|
||||
impl<T> Arbitrary for Large<Vec<T>>
|
||||
where
|
||||
T: Arbitrary,
|
||||
{
|
||||
fn arbitrary(g: &mut Gen) -> Self {
|
||||
let len = usize::arbitrary(g) % (g.size() * 10);
|
||||
Large((0..len).map(|_| T::arbitrary(g)).collect())
|
||||
}
|
||||
|
||||
fn shrink(&self) -> Box<dyn Iterator<Item = Self>> {
|
||||
Box::new((**self).shrink().map(Large))
|
||||
}
|
||||
}
|
||||
28
clamav/libclamav_rust/.cargo/vendor/indexmap/tests/tests.rs
vendored
Normal file
28
clamav/libclamav_rust/.cargo/vendor/indexmap/tests/tests.rs
vendored
Normal file
@@ -0,0 +1,28 @@
|
||||
use indexmap::{indexmap, indexset};
|
||||
|
||||
#[test]
|
||||
fn test_sort() {
|
||||
let m = indexmap! {
|
||||
1 => 2,
|
||||
7 => 1,
|
||||
2 => 2,
|
||||
3 => 3,
|
||||
};
|
||||
|
||||
itertools::assert_equal(
|
||||
m.sorted_by(|_k1, v1, _k2, v2| v1.cmp(v2)),
|
||||
vec![(7, 1), (1, 2), (2, 2), (3, 3)],
|
||||
);
|
||||
}
|
||||
|
||||
#[test]
|
||||
fn test_sort_set() {
|
||||
let s = indexset! {
|
||||
1,
|
||||
7,
|
||||
2,
|
||||
3,
|
||||
};
|
||||
|
||||
itertools::assert_equal(s.sorted_by(|v1, v2| v1.cmp(v2)), vec![1, 2, 3, 7]);
|
||||
}
|
||||
Reference in New Issue
Block a user