From 4813d1afd897985304a468da4ea71bd1baadb225 Mon Sep 17 00:00:00 2001 From: Matteo Bovetti Date: Mon, 10 Mar 2025 17:13:29 +0100 Subject: [PATCH] second iteration --- Cargo.toml | 2 +- paste-macro/src/lib.rs | 199 +++++++++++++++++++++++++++++++++++++- paste-macro/src/paste.rs | 201 --------------------------------------- 3 files changed, 199 insertions(+), 203 deletions(-) delete mode 100644 paste-macro/src/paste.rs diff --git a/Cargo.toml b/Cargo.toml index 22c67625..56e16447 100644 --- a/Cargo.toml +++ b/Cargo.toml @@ -32,7 +32,7 @@ actix-utils = { path = "actix-utils" } bytestring = { path = "bytestring" } local-channel = { path = "local-channel" } local-waker = { path = "local-waker" } -paste-macro = { path = "paste-macro" } +paste-macro = { path = "./paste-macro" } [profile.release] lto = true diff --git a/paste-macro/src/lib.rs b/paste-macro/src/lib.rs index 76abad37..69dafbe6 100644 --- a/paste-macro/src/lib.rs +++ b/paste-macro/src/lib.rs @@ -1,3 +1,200 @@ //! A minimal implementation of the paste-macro crate, allowing identifier concatenation in macros. -mod paste; +use proc_macro::{Delimiter, Group, Ident, Span, TokenStream, TokenTree}; +use std::{iter, str::FromStr}; +use std::panic; + +// Pastes identifiers together to form a single identifier. +// +// Within the `paste!` macro, identifiers inside `[<`...`>]` are pasted +// together to form a single identifier. +// +// # Examples +// +// ``` +// use paste_macro::paste; +// +// paste! { +// // Defines a const called `QRST`. +// const []: &str = "success!"; +// } +// ``` + +#[proc_macro] +pub fn paste(input: TokenStream) -> TokenStream { + let mut expanded = TokenStream::new(); + let mut tokens = input.into_iter().peekable(); + + while let Some(token) = tokens.next() { + match token { + TokenTree::Group(group) => { + let delimiter = group.delimiter(); + let content = group.stream(); + let span = group.span(); + + if delimiter == Delimiter::Bracket && is_paste_operation(&content) { + // Process [< ... >] paste-macro operation + if let Ok(pasted) = process_paste_operation(content, span) { + expanded.extend(pasted); + } else { + // On error, return the original token + expanded.extend(iter::once(TokenTree::Group(group))); + } + } else { + // Handle nested groups recursively + let nested = paste(content); + let mut new_group = Group::new(delimiter, nested); + new_group.set_span(span); + expanded.extend(iter::once(TokenTree::Group(new_group))); + } + } + // Pass through all other tokens unchanged + _ => expanded.extend(iter::once(token)), + } + } + + expanded +} + +// Check if a token stream is a paste-macro operation: [< ... >] +fn is_paste_operation(input: &TokenStream) -> bool { + let mut tokens = input.clone().into_iter(); + + match tokens.next() { + Some(TokenTree::Punct(punct)) if punct.as_char() == '<' => {} + _ => return false, + } + + let mut has_content = false; + for token in tokens { + match token { + TokenTree::Punct(punct) if punct.as_char() == '>' => return has_content, + _ => has_content = true, + } + } + + false +} + +// Process the content inside [< ... >] +fn process_paste_operation(input: TokenStream, span: Span) -> Result { + let mut tokens = input.into_iter(); + + // Skip opening '<' + if let Some(TokenTree::Punct(punct)) = tokens.next() { + if punct.as_char() != '<' { + return Err(()); + } + } else { + return Err(()); + } + + // Collect and process segments + let mut segments = Vec::new(); + let mut has_lifetime = false; + + while let Some(token) = tokens.next() { + match &token { + TokenTree::Punct(punct) if punct.as_char() == '>' => break, + TokenTree::Ident(ident) => segments.push(ident.to_string()), + TokenTree::Literal(lit) => { + let lit_str = lit.to_string(); + if lit_str.starts_with('"') && lit_str.ends_with('"') && lit_str.len() >= 2 { + segments.push(lit_str[1..lit_str.len() - 1].to_owned().replace('-', "_")); + } else { + segments.push(lit_str.replace('-', "_")); + } + }, + TokenTree::Punct(punct) if punct.as_char() == '_' => segments.push("_".to_owned()), + TokenTree::Punct(punct) if punct.as_char() == '\'' => { + has_lifetime = true; + }, + TokenTree::Punct(punct) if punct.as_char() == ':' => { + if segments.is_empty() { + return Err(()); + } + + // Handle modifiers like :lower, :upper, etc. + if let Some(TokenTree::Ident(ident)) = tokens.next() { + let modifier = ident.to_string(); + let last = segments.pop().unwrap(); + + let result = match modifier.as_str() { + "lower" => last.to_lowercase(), + "upper" => last.to_uppercase(), + "snake" => to_snake_case(&last), + "camel" => to_camel_case(&last), + _ => return Err(()), + }; + + segments.push(result); + } else { + return Err(()); + } + }, + _ => return Err(()), + } + } + + // Create identifier from the concatenated segments + let mut pasted = segments.join(""); + + // Add lifetime symbol if needed + if has_lifetime { + pasted.insert(0, '\''); + } + + // Convert to a valid Rust identifier or literal + if pasted.starts_with(|c: char| c.is_ascii_digit()) { + // If it starts with a number, try to create a literal + match TokenStream::from_str(&pasted) { + Ok(ts) => { + if let Some(token) = ts.into_iter().next() { + return Ok(iter::once(token).collect()); + } + } + Err(_) => return Err(()), + } + } + + // Try to create an identifier + match panic::catch_unwind(|| Ident::new(&pasted, span)) { + Ok(ident) => Ok(iter::once(TokenTree::Ident(ident)).collect()), + Err(_) => Err(()), + } +} + +// Helper function to convert CamelCase to snake_case +fn to_snake_case(input: &str) -> String { + let mut result = String::new(); + let mut prev = '_'; + + for ch in input.chars() { + if ch.is_uppercase() && prev != '_' { + result.push('_'); + } + result.push(ch.to_lowercase().next().unwrap_or(ch)); + prev = ch; + } + + result +} + +// Helper function to convert snake_case to CamelCase +fn to_camel_case(input: &str) -> String { + let mut result = String::new(); + let mut capitalize_next = false; + + for ch in input.chars() { + if ch == '_' { + capitalize_next = true; + } else if capitalize_next { + result.push(ch.to_uppercase().next().unwrap_or(ch)); + capitalize_next = false; + } else { + result.push(ch); + } + } + + result +} \ No newline at end of file diff --git a/paste-macro/src/paste.rs b/paste-macro/src/paste.rs deleted file mode 100644 index 2c588e8f..00000000 --- a/paste-macro/src/paste.rs +++ /dev/null @@ -1,201 +0,0 @@ -//! A minimal implementation of the paste-macro crate, allowing identifier concatenation in macros. - -use proc_macro::{ - Delimiter, Group, Ident, Punct, Spacing, Span, TokenStream, TokenTree -}; -use std::{iter, str::FromStr}; -use std::panic; - -// Pastes identifiers together to form a single identifier. -// -// Within the `paste!` macro, identifiers inside `[<`...`>]` are pasted -// together to form a single identifier. -// -// # Examples -// -// ``` -// use paste_macro::paste; -// -// paste! { -// // Defines a const called `QRST`. -// const []: &str = "success!"; -// } -// ``` -#[proc_macro] -pub fn paste(input: TokenStream) -> TokenStream { - let mut expanded = TokenStream::new(); - let mut tokens = input.into_iter().peekable(); - - while let Some(token) = tokens.next() { - match token { - TokenTree::Group(group) => { - let delimiter = group.delimiter(); - let content = group.stream(); - let span = group.span(); - - if delimiter == Delimiter::Bracket && is_paste_operation(&content) { - // Process [< ... >] paste-macro operation - if let Ok(pasted) = process_paste_operation(content, span) { - expanded.extend(pasted); - } else { - // On error, return the original token - expanded.extend(iter::once(TokenTree::Group(group))); - } - } else { - // Handle nested groups recursively - let nested = paste(content); - let mut new_group = Group::new(delimiter, nested); - new_group.set_span(span); - expanded.extend(iter::once(TokenTree::Group(new_group))); - } - } - // Pass through all other tokens unchanged - _ => expanded.extend(iter::once(token)), - } - } - - expanded -} - -// Check if a token stream is a paste-macro operation: [< ... >] -fn is_paste_operation(input: &TokenStream) -> bool { - let mut tokens = input.clone().into_iter(); - - match tokens.next() { - Some(TokenTree::Punct(punct)) if punct.as_char() == '<' => {} - _ => return false, - } - - let mut has_content = false; - for token in tokens { - match token { - TokenTree::Punct(punct) if punct.as_char() == '>' => return has_content, - _ => has_content = true, - } - } - - false -} - -// Process the content inside [< ... >] -fn process_paste_operation(input: TokenStream, span: Span) -> Result { - let mut tokens = input.into_iter(); - - // Skip opening '<' - if let Some(TokenTree::Punct(punct)) = tokens.next() { - if punct.as_char() != '<' { - return Err(()); - } - } else { - return Err(()); - } - - // Collect and process segments - let mut segments = Vec::new(); - let mut has_lifetime = false; - - while let Some(token) = tokens.next() { - match &token { - TokenTree::Punct(punct) if punct.as_char() == '>' => break, - TokenTree::Ident(ident) => segments.push(ident.to_string()), - TokenTree::Literal(lit) => { - let lit_str = lit.to_string(); - if lit_str.starts_with('"') && lit_str.ends_with('"') && lit_str.len() >= 2 { - segments.push(lit_str[1..lit_str.len() - 1].to_owned().replace('-', "_")); - } else { - segments.push(lit_str.replace('-', "_")); - } - }, - TokenTree::Punct(punct) if punct.as_char() == '_' => segments.push("_".to_owned()), - TokenTree::Punct(punct) if punct.as_char() == '\'' => { - has_lifetime = true; - }, - TokenTree::Punct(punct) if punct.as_char() == ':' => { - if segments.is_empty() { - return Err(()); - } - - // Handle modifiers like :lower, :upper, etc. - if let Some(TokenTree::Ident(ident)) = tokens.next() { - let modifier = ident.to_string(); - let last = segments.pop().unwrap(); - - let result = match modifier.as_str() { - "lower" => last.to_lowercase(), - "upper" => last.to_uppercase(), - "snake" => to_snake_case(&last), - "camel" => to_camel_case(&last), - _ => return Err(()), - }; - - segments.push(result); - } else { - return Err(()); - } - }, - _ => return Err(()), - } - } - - // Create identifier from the concatenated segments - let mut pasted = segments.join(""); - - // Add lifetime symbol if needed - if has_lifetime { - pasted.insert(0, '\''); - } - - // Convert to a valid Rust identifier or literal - if pasted.starts_with(|c: char| c.is_ascii_digit()) { - // If it starts with a number, try to create a literal - match TokenStream::from_str(&pasted) { - Ok(ts) => { - if let Some(token) = ts.into_iter().next() { - return Ok(iter::once(token).collect()); - } - } - Err(_) => return Err(()), - } - } - - // Try to create an identifier - match panic::catch_unwind(|| Ident::new(&pasted, span)) { - Ok(ident) => Ok(iter::once(TokenTree::Ident(ident)).collect()), - Err(_) => Err(()), - } -} - -// Helper function to convert CamelCase to snake_case -fn to_snake_case(input: &str) -> String { - let mut result = String::new(); - let mut prev = '_'; - - for ch in input.chars() { - if ch.is_uppercase() && prev != '_' { - result.push('_'); - } - result.push(ch.to_lowercase().next().unwrap_or(ch)); - prev = ch; - } - - result -} - -// Helper function to convert snake_case to CamelCase -fn to_camel_case(input: &str) -> String { - let mut result = String::new(); - let mut capitalize_next = false; - - for ch in input.chars() { - if ch == '_' { - capitalize_next = true; - } else if capitalize_next { - result.push(ch.to_uppercase().next().unwrap_or(ch)); - capitalize_next = false; - } else { - result.push(ch); - } - } - - result -} \ No newline at end of file