summaryrefslogtreecommitdiff
path: root/vendor/base64/README.md
diff options
context:
space:
mode:
Diffstat (limited to 'vendor/base64/README.md')
-rw-r--r--vendor/base64/README.md154
1 files changed, 0 insertions, 154 deletions
diff --git a/vendor/base64/README.md b/vendor/base64/README.md
deleted file mode 100644
index f566756d..00000000
--- a/vendor/base64/README.md
+++ /dev/null
@@ -1,154 +0,0 @@
-# [base64](https://crates.io/crates/base64)
-
-[![](https://img.shields.io/crates/v/base64.svg)](https://crates.io/crates/base64) [![Docs](https://docs.rs/base64/badge.svg)](https://docs.rs/base64) [![CircleCI](https://circleci.com/gh/marshallpierce/rust-base64/tree/master.svg?style=shield)](https://circleci.com/gh/marshallpierce/rust-base64/tree/master) [![codecov](https://codecov.io/gh/marshallpierce/rust-base64/branch/master/graph/badge.svg)](https://codecov.io/gh/marshallpierce/rust-base64) [![unsafe forbidden](https://img.shields.io/badge/unsafe-forbidden-success.svg)](https://github.com/rust-secure-code/safety-dance/)
-
-<a href="https://www.jetbrains.com/?from=rust-base64"><img src="/icon_CLion.svg" height="40px"/></a>
-
-Made with CLion. Thanks to JetBrains for supporting open source!
-
-It's base64. What more could anyone want?
-
-This library's goals are to be *correct* and *fast*. It's thoroughly tested and widely used. It exposes functionality at
-multiple levels of abstraction so you can choose the level of convenience vs performance that you want,
-e.g. `decode_engine_slice` decodes into an existing `&mut [u8]` and is pretty fast (2.6GiB/s for a 3 KiB input),
-whereas `decode_engine` allocates a new `Vec<u8>` and returns it, which might be more convenient in some cases, but is
-slower (although still fast enough for almost any purpose) at 2.1 GiB/s.
-
-See the [docs](https://docs.rs/base64) for all the details.
-
-## FAQ
-
-### I need to decode base64 with whitespace/null bytes/other random things interspersed in it. What should I do?
-
-Remove non-base64 characters from your input before decoding.
-
-If you have a `Vec` of base64, [retain](https://doc.rust-lang.org/std/vec/struct.Vec.html#method.retain) can be used to
-strip out whatever you need removed.
-
-If you have a `Read` (e.g. reading a file or network socket), there are various approaches.
-
-- Use [iter_read](https://crates.io/crates/iter-read) together with `Read`'s `bytes()` to filter out unwanted bytes.
-- Implement `Read` with a `read()` impl that delegates to your actual `Read`, and then drops any bytes you don't want.
-
-### I need to line-wrap base64, e.g. for MIME/PEM.
-
-[line-wrap](https://crates.io/crates/line-wrap) does just that.
-
-### I want canonical base64 encoding/decoding.
-
-First, don't do this. You should no more expect Base64 to be canonical than you should expect compression algorithms to
-produce canonical output across all usage in the wild (hint: they don't).
-However, [people are drawn to their own destruction like moths to a flame](https://eprint.iacr.org/2022/361), so here we
-are.
-
-There are two opportunities for non-canonical encoding (and thus, detection of the same during decoding): the final bits
-of the last encoded token in two or three token suffixes, and the `=` token used to inflate the suffix to a full four
-tokens.
-
-The trailing bits issue is unavoidable: with 6 bits available in each encoded token, 1 input byte takes 2 tokens,
-with the second one having some bits unused. Same for two input bytes: 16 bits, but 3 tokens have 18 bits. Unless we
-decide to stop shipping whole bytes around, we're stuck with those extra bits that a sneaky or buggy encoder might set
-to 1 instead of 0.
-
-The `=` pad bytes, on the other hand, are entirely a self-own by the Base64 standard. They do not affect decoding other
-than to provide an opportunity to say "that padding is incorrect". Exabytes of storage and transfer have no doubt been
-wasted on pointless `=` bytes. Somehow we all seem to be quite comfortable with, say, hex-encoded data just stopping
-when it's done rather than requiring a confirmation that the author of the encoder could count to four. Anyway, there
-are two ways to make pad bytes predictable: require canonical padding to the next multiple of four bytes as per the RFC,
-or, if you control all producers and consumers, save a few bytes by requiring no padding (especially applicable to the
-url-safe alphabet).
-
-All `Engine` implementations must at a minimum support treating non-canonical padding of both types as an error, and
-optionally may allow other behaviors.
-
-## Rust version compatibility
-
-The minimum supported Rust version is 1.48.0.
-
-# Contributing
-
-Contributions are very welcome. However, because this library is used widely, and in security-sensitive contexts, all
-PRs will be carefully scrutinized. Beyond that, this sort of low level library simply needs to be 100% correct. Nobody
-wants to chase bugs in encoding of any sort.
-
-All this means that it takes me a fair amount of time to review each PR, so it might take quite a while to carve out the
-free time to give each PR the attention it deserves. I will get to everyone eventually!
-
-## Developing
-
-Benchmarks are in `benches/`.
-
-```bash
-cargo bench
-```
-
-## no_std
-
-This crate supports no_std. By default the crate targets std via the `std` feature. You can deactivate
-the `default-features` to target `core` instead. In that case you lose out on all the functionality revolving
-around `std::io`, `std::error::Error`, and heap allocations. There is an additional `alloc` feature that you can activate
-to bring back the support for heap allocations.
-
-## Profiling
-
-On Linux, you can use [perf](https://perf.wiki.kernel.org/index.php/Main_Page) for profiling. Then compile the
-benchmarks with `cargo bench --no-run`.
-
-Run the benchmark binary with `perf` (shown here filtering to one particular benchmark, which will make the results
-easier to read). `perf` is only available to the root user on most systems as it fiddles with event counters in your
-CPU, so use `sudo`. We need to run the actual benchmark binary, hence the path into `target`. You can see the actual
-full path with `cargo bench -v`; it will print out the commands it runs. If you use the exact path
-that `bench` outputs, make sure you get the one that's for the benchmarks, not the tests. You may also want
-to `cargo clean` so you have only one `benchmarks-` binary (they tend to accumulate).
-
-```bash
-sudo perf record target/release/deps/benchmarks-* --bench decode_10mib_reuse
-```
-
-Then analyze the results, again with perf:
-
-```bash
-sudo perf annotate -l
-```
-
-You'll see a bunch of interleaved rust source and assembly like this. The section with `lib.rs:327` is telling us that
-4.02% of samples saw the `movzbl` aka bit shift as the active instruction. However, this percentage is not as exact as
-it seems due to a phenomenon called *skid*. Basically, a consequence of how fancy modern CPUs are is that this sort of
-instruction profiling is inherently inaccurate, especially in branch-heavy code.
-
-```text
- lib.rs:322 0.70 : 10698: mov %rdi,%rax
- 2.82 : 1069b: shr $0x38,%rax
- : if morsel == decode_tables::INVALID_VALUE {
- : bad_byte_index = input_index;
- : break;
- : };
- : accum = (morsel as u64) << 58;
- lib.rs:327 4.02 : 1069f: movzbl (%r9,%rax,1),%r15d
- : // fast loop of 8 bytes at a time
- : while input_index < length_of_full_chunks {
- : let mut accum: u64;
- :
- : let input_chunk = BigEndian::read_u64(&input_bytes[input_index..(input_index + 8)]);
- : morsel = decode_table[(input_chunk >> 56) as usize];
- lib.rs:322 3.68 : 106a4: cmp $0xff,%r15
- : if morsel == decode_tables::INVALID_VALUE {
- 0.00 : 106ab: je 1090e <base64::decode_config_buf::hbf68a45fefa299c1+0x46e>
-```
-
-## Fuzzing
-
-This uses [cargo-fuzz](https://github.com/rust-fuzz/cargo-fuzz). See `fuzz/fuzzers` for the available fuzzing scripts.
-To run, use an invocation like these:
-
-```bash
-cargo +nightly fuzz run roundtrip
-cargo +nightly fuzz run roundtrip_no_pad
-cargo +nightly fuzz run roundtrip_random_config -- -max_len=10240
-cargo +nightly fuzz run decode_random
-```
-
-## License
-
-This project is dual-licensed under MIT and Apache 2.0.
-