summaryrefslogtreecommitdiff
path: root/vendor/async-stream-impl
diff options
context:
space:
mode:
authormo khan <mo@mokhan.ca>2025-07-15 16:37:08 -0600
committermo khan <mo@mokhan.ca>2025-07-17 16:30:22 -0600
commit45df4d0d9b577fecee798d672695fe24ff57fb1b (patch)
tree1b99bf645035b58e0d6db08c7a83521f41f7a75b /vendor/async-stream-impl
parentf94f79608393d4ab127db63cc41668445ef6b243 (diff)
feat: migrate from Cedar to SpiceDB authorization system
This is a major architectural change that replaces the Cedar policy-based authorization system with SpiceDB's relation-based authorization. Key changes: - Migrate from Rust to Go implementation - Replace Cedar policies with SpiceDB schema and relationships - Switch from envoy `ext_authz` with Cedar to SpiceDB permission checks - Update build system and dependencies for Go ecosystem - Maintain Envoy integration for external authorization This change enables more flexible permission modeling through SpiceDB's Google Zanzibar inspired relation-based system, supporting complex hierarchical permissions that were difficult to express in Cedar. Breaking change: Existing Cedar policies and Rust-based configuration will no longer work and need to be migrated to SpiceDB schema.
Diffstat (limited to 'vendor/async-stream-impl')
-rw-r--r--vendor/async-stream-impl/.cargo-checksum.json1
-rw-r--r--vendor/async-stream-impl/Cargo.toml54
-rw-r--r--vendor/async-stream-impl/LICENSE51
-rw-r--r--vendor/async-stream-impl/src/lib.rs316
4 files changed, 0 insertions, 422 deletions
diff --git a/vendor/async-stream-impl/.cargo-checksum.json b/vendor/async-stream-impl/.cargo-checksum.json
deleted file mode 100644
index 87984c56..00000000
--- a/vendor/async-stream-impl/.cargo-checksum.json
+++ /dev/null
@@ -1 +0,0 @@
-{"files":{"Cargo.toml":"0a6ec9f20c5312e8ec19d902cd330a69b3cf8a14a1081faaa6540046652b9f92","LICENSE":"157d381f6304eba77459fc10dfb5b1b22fc0d04e384a32865f32d8e67a34c0eb","src/lib.rs":"71ec1ba96d3ac995fbe1adcf29a492eb1ad9ea28922c07521416c4291e627081"},"package":"c7c24de15d275a1ecfd47a380fb4d5ec9bfe0933f309ed5e705b775596a3574d"} \ No newline at end of file
diff --git a/vendor/async-stream-impl/Cargo.toml b/vendor/async-stream-impl/Cargo.toml
deleted file mode 100644
index 6e492ff8..00000000
--- a/vendor/async-stream-impl/Cargo.toml
+++ /dev/null
@@ -1,54 +0,0 @@
-# THIS FILE IS AUTOMATICALLY GENERATED BY CARGO
-#
-# When uploading crates to the registry Cargo will automatically
-# "normalize" Cargo.toml files for maximal compatibility
-# with all versions of Cargo and also rewrite `path` dependencies
-# to registry (e.g., crates.io) dependencies.
-#
-# If you are reading this file be aware that the original Cargo.toml
-# will likely look very different (and much more reasonable).
-# See Cargo.toml.orig for the original contents.
-
-[package]
-edition = "2021"
-rust-version = "1.65"
-name = "async-stream-impl"
-version = "0.3.6"
-authors = ["Carl Lerche <me@carllerche.com>"]
-build = false
-autobins = false
-autoexamples = false
-autotests = false
-autobenches = false
-description = "proc macros for async-stream crate"
-readme = false
-license = "MIT"
-repository = "https://github.com/tokio-rs/async-stream"
-
-[lib]
-name = "async_stream_impl"
-path = "src/lib.rs"
-proc-macro = true
-
-[dependencies.proc-macro2]
-version = "1.0.60"
-
-[dependencies.quote]
-version = "1"
-
-[dependencies.syn]
-version = "2.0.2"
-features = [
- "full",
- "visit-mut",
-]
-
-[dev-dependencies.futures-core]
-version = "0.3"
-
-[dev-dependencies.futures-util]
-version = "0.3"
-
-[dev-dependencies.tokio]
-version = "1"
-features = ["full"]
diff --git a/vendor/async-stream-impl/LICENSE b/vendor/async-stream-impl/LICENSE
deleted file mode 100644
index 8cbd7d63..00000000
--- a/vendor/async-stream-impl/LICENSE
+++ /dev/null
@@ -1,51 +0,0 @@
-Copyright (c) 2019 Carl Lerche
-
-Permission is hereby granted, free of charge, to any
-person obtaining a copy of this software and associated
-documentation files (the "Software"), to deal in the
-Software without restriction, including without
-limitation the rights to use, copy, modify, merge,
-publish, distribute, sublicense, and/or sell copies of
-the Software, and to permit persons to whom the Software
-is furnished to do so, subject to the following
-conditions:
-
-The above copyright notice and this permission notice
-shall be included in all copies or substantial portions
-of the Software.
-
-THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF
-ANY KIND, EXPRESS OR IMPLIED, INCLUDING BUT NOT LIMITED
-TO THE WARRANTIES OF MERCHANTABILITY, FITNESS FOR A
-PARTICULAR PURPOSE AND NONINFRINGEMENT. IN NO EVENT
-SHALL THE AUTHORS OR COPYRIGHT HOLDERS BE LIABLE FOR ANY
-CLAIM, DAMAGES OR OTHER LIABILITY, WHETHER IN AN ACTION
-OF CONTRACT, TORT OR OTHERWISE, ARISING FROM, OUT OF OR
-IN CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER
-DEALINGS IN THE SOFTWARE.
-
-Copyright (c) 2018 David Tolnay
-
-Permission is hereby granted, free of charge, to any
-person obtaining a copy of this software and associated
-documentation files (the "Software"), to deal in the
-Software without restriction, including without
-limitation the rights to use, copy, modify, merge,
-publish, distribute, sublicense, and/or sell copies of
-the Software, and to permit persons to whom the Software
-is furnished to do so, subject to the following
-conditions:
-
-The above copyright notice and this permission notice
-shall be included in all copies or substantial portions
-of the Software.
-
-THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF
-ANY KIND, EXPRESS OR IMPLIED, INCLUDING BUT NOT LIMITED
-TO THE WARRANTIES OF MERCHANTABILITY, FITNESS FOR A
-PARTICULAR PURPOSE AND NONINFRINGEMENT. IN NO EVENT
-SHALL THE AUTHORS OR COPYRIGHT HOLDERS BE LIABLE FOR ANY
-CLAIM, DAMAGES OR OTHER LIABILITY, WHETHER IN AN ACTION
-OF CONTRACT, TORT OR OTHERWISE, ARISING FROM, OUT OF OR
-IN CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER
-DEALINGS IN THE SOFTWARE.
diff --git a/vendor/async-stream-impl/src/lib.rs b/vendor/async-stream-impl/src/lib.rs
deleted file mode 100644
index d6fcdca9..00000000
--- a/vendor/async-stream-impl/src/lib.rs
+++ /dev/null
@@ -1,316 +0,0 @@
-use proc_macro::TokenStream;
-use proc_macro2::{Group, TokenStream as TokenStream2, TokenTree};
-use quote::quote;
-use syn::parse::{Parse, ParseStream, Parser, Result};
-use syn::visit_mut::VisitMut;
-
-struct Scrub<'a> {
- /// Whether the stream is a try stream.
- is_try: bool,
- /// The unit expression, `()`.
- unit: Box<syn::Expr>,
- has_yielded: bool,
- crate_path: &'a TokenStream2,
-}
-
-fn parse_input(input: TokenStream) -> syn::Result<(TokenStream2, Vec<syn::Stmt>)> {
- let mut input = TokenStream2::from(input).into_iter();
- let crate_path = match input.next().unwrap() {
- TokenTree::Group(group) => group.stream(),
- _ => panic!(),
- };
- let stmts = syn::Block::parse_within.parse2(replace_for_await(input))?;
- Ok((crate_path, stmts))
-}
-
-impl<'a> Scrub<'a> {
- fn new(is_try: bool, crate_path: &'a TokenStream2) -> Self {
- Self {
- is_try,
- unit: syn::parse_quote!(()),
- has_yielded: false,
- crate_path,
- }
- }
-}
-
-struct Partial<T>(T, TokenStream2);
-
-impl<T: Parse> Parse for Partial<T> {
- fn parse(input: ParseStream) -> Result<Self> {
- Ok(Partial(input.parse()?, input.parse()?))
- }
-}
-
-fn visit_token_stream_impl(
- visitor: &mut Scrub<'_>,
- tokens: TokenStream2,
- modified: &mut bool,
- out: &mut TokenStream2,
-) {
- use quote::ToTokens;
- use quote::TokenStreamExt;
-
- let mut tokens = tokens.into_iter().peekable();
- while let Some(tt) = tokens.next() {
- match tt {
- TokenTree::Ident(i) if i == "yield" => {
- let stream = std::iter::once(TokenTree::Ident(i)).chain(tokens).collect();
- match syn::parse2(stream) {
- Ok(Partial(yield_expr, rest)) => {
- let mut expr = syn::Expr::Yield(yield_expr);
- visitor.visit_expr_mut(&mut expr);
- expr.to_tokens(out);
- *modified = true;
- tokens = rest.into_iter().peekable();
- }
- Err(e) => {
- out.append_all(e.to_compile_error().into_iter());
- *modified = true;
- return;
- }
- }
- }
- TokenTree::Ident(i) if i == "stream" || i == "try_stream" => {
- out.append(TokenTree::Ident(i));
- match tokens.peek() {
- Some(TokenTree::Punct(p)) if p.as_char() == '!' => {
- out.extend(tokens.next()); // !
- if let Some(TokenTree::Group(_)) = tokens.peek() {
- out.extend(tokens.next()); // { .. } or [ .. ] or ( .. )
- }
- }
- _ => {}
- }
- }
- TokenTree::Group(group) => {
- let mut content = group.stream();
- *modified |= visitor.visit_token_stream(&mut content);
- let mut new = Group::new(group.delimiter(), content);
- new.set_span(group.span());
- out.append(new);
- }
- other => out.append(other),
- }
- }
-}
-
-impl Scrub<'_> {
- fn visit_token_stream(&mut self, tokens: &mut TokenStream2) -> bool {
- let (mut out, mut modified) = (TokenStream2::new(), false);
- visit_token_stream_impl(self, tokens.clone(), &mut modified, &mut out);
-
- if modified {
- *tokens = out;
- }
-
- modified
- }
-}
-
-impl VisitMut for Scrub<'_> {
- fn visit_expr_mut(&mut self, i: &mut syn::Expr) {
- match i {
- syn::Expr::Yield(yield_expr) => {
- self.has_yielded = true;
-
- syn::visit_mut::visit_expr_yield_mut(self, yield_expr);
-
- let value_expr = yield_expr.expr.as_ref().unwrap_or(&self.unit);
-
- // let ident = &self.yielder;
-
- let yield_expr = if self.is_try {
- quote! { __yield_tx.send(::core::result::Result::Ok(#value_expr)).await }
- } else {
- quote! { __yield_tx.send(#value_expr).await }
- };
- *i = syn::parse_quote! {
- {
- #[allow(unreachable_code)]
- if false {
- break '__async_stream_private_check_scope (loop {});
- }
- #yield_expr
- }
- };
- }
- syn::Expr::Try(try_expr) => {
- syn::visit_mut::visit_expr_try_mut(self, try_expr);
- // let ident = &self.yielder;
- let e = &try_expr.expr;
-
- *i = syn::parse_quote! {
- match #e {
- ::core::result::Result::Ok(v) => v,
- ::core::result::Result::Err(e) => {
- __yield_tx.send(::core::result::Result::Err(e.into())).await;
- return;
- }
- }
- };
- }
- syn::Expr::Closure(_) | syn::Expr::Async(_) => {
- // Don't transform inner closures or async blocks.
- }
- syn::Expr::ForLoop(expr) => {
- syn::visit_mut::visit_expr_for_loop_mut(self, expr);
- // TODO: Should we allow other attributes?
- if expr.attrs.len() != 1 || !expr.attrs[0].meta.path().is_ident(AWAIT_ATTR_NAME) {
- return;
- }
- let syn::ExprForLoop {
- attrs,
- label,
- pat,
- expr,
- body,
- ..
- } = expr;
-
- attrs.pop().unwrap();
-
- let crate_path = self.crate_path;
- *i = syn::parse_quote! {{
- let mut __pinned = #expr;
- let mut __pinned = unsafe {
- ::core::pin::Pin::new_unchecked(&mut __pinned)
- };
- #label
- loop {
- let #pat = match #crate_path::__private::next(&mut __pinned).await {
- ::core::option::Option::Some(e) => e,
- ::core::option::Option::None => break,
- };
- #body
- }
- }}
- }
- _ => syn::visit_mut::visit_expr_mut(self, i),
- }
- }
-
- fn visit_macro_mut(&mut self, mac: &mut syn::Macro) {
- let mac_ident = mac.path.segments.last().map(|p| &p.ident);
- if mac_ident.map_or(false, |i| i == "stream" || i == "try_stream") {
- return;
- }
-
- self.visit_token_stream(&mut mac.tokens);
- }
-
- fn visit_item_mut(&mut self, i: &mut syn::Item) {
- // Recurse into macros but otherwise don't transform inner items.
- if let syn::Item::Macro(i) = i {
- self.visit_macro_mut(&mut i.mac);
- }
- }
-}
-
-/// The first token tree in the stream must be a group containing the path to the `async-stream`
-/// crate.
-#[proc_macro]
-#[doc(hidden)]
-pub fn stream_inner(input: TokenStream) -> TokenStream {
- let (crate_path, mut stmts) = match parse_input(input) {
- Ok(x) => x,
- Err(e) => return e.to_compile_error().into(),
- };
-
- let mut scrub = Scrub::new(false, &crate_path);
-
- for stmt in &mut stmts {
- scrub.visit_stmt_mut(stmt);
- }
-
- let dummy_yield = if scrub.has_yielded {
- None
- } else {
- Some(quote!(if false {
- __yield_tx.send(()).await;
- }))
- };
-
- quote!({
- let (mut __yield_tx, __yield_rx) = unsafe { #crate_path::__private::yielder::pair() };
- #crate_path::__private::AsyncStream::new(__yield_rx, async move {
- '__async_stream_private_check_scope: {
- #dummy_yield
- #(#stmts)*
- }
- })
- })
- .into()
-}
-
-/// The first token tree in the stream must be a group containing the path to the `async-stream`
-/// crate.
-#[proc_macro]
-#[doc(hidden)]
-pub fn try_stream_inner(input: TokenStream) -> TokenStream {
- let (crate_path, mut stmts) = match parse_input(input) {
- Ok(x) => x,
- Err(e) => return e.to_compile_error().into(),
- };
-
- let mut scrub = Scrub::new(true, &crate_path);
-
- for stmt in &mut stmts {
- scrub.visit_stmt_mut(stmt);
- }
-
- let dummy_yield = if scrub.has_yielded {
- None
- } else {
- Some(quote!(if false {
- __yield_tx.send(()).await;
- }))
- };
-
- quote!({
- let (mut __yield_tx, __yield_rx) = unsafe { #crate_path::__private::yielder::pair() };
- #crate_path::__private::AsyncStream::new(__yield_rx, async move {
- '__async_stream_private_check_scope: {
- #dummy_yield
- #(#stmts)*
- }
- })
- })
- .into()
-}
-
-// syn 2.0 wont parse `#[await] for x in xs {}`
-// because `await` is a keyword, use `await_` instead
-const AWAIT_ATTR_NAME: &str = "await_";
-
-/// Replace `for await` with `#[await] for`, which will be later transformed into a `next` loop.
-fn replace_for_await(input: impl IntoIterator<Item = TokenTree>) -> TokenStream2 {
- let mut input = input.into_iter().peekable();
- let mut tokens = Vec::new();
-
- while let Some(token) = input.next() {
- match token {
- TokenTree::Ident(ident) => {
- match input.peek() {
- Some(TokenTree::Ident(next)) if ident == "for" && next == "await" => {
- let next_span = next.span();
- let next = syn::Ident::new(AWAIT_ATTR_NAME, next_span);
- tokens.extend(quote!(#[#next]));
- let _ = input.next();
- }
- _ => {}
- }
- tokens.push(ident.into());
- }
- TokenTree::Group(group) => {
- let stream = replace_for_await(group.stream());
- let mut new_group = Group::new(group.delimiter(), stream);
- new_group.set_span(group.span());
- tokens.push(new_group.into());
- }
- _ => tokens.push(token),
- }
- }
-
- tokens.into_iter().collect()
-}