about summary refs log tree commit diff
diff options
context:
space:
mode:
-rw-r--r--crates/cfg/src/cfg_attr.rs70
-rw-r--r--crates/cfg/src/cfg_expr.rs2
-rw-r--r--crates/cfg/src/lib.rs4
-rw-r--r--crates/hir-expand/src/cfg_process.rs178
-rw-r--r--crates/hir-expand/src/db.rs28
-rw-r--r--crates/hir-expand/src/fixup.rs6
-rw-r--r--crates/hir-expand/src/lib.rs2
-rw-r--r--crates/mbe/src/syntax_bridge.rs37
8 files changed, 302 insertions, 25 deletions
diff --git a/crates/cfg/src/cfg_attr.rs b/crates/cfg/src/cfg_attr.rs
new file mode 100644
index 00000000000..4eb5928b1e1
--- /dev/null
+++ b/crates/cfg/src/cfg_attr.rs
@@ -0,0 +1,70 @@
+use std::{
+    fmt::{self, Debug},
+    slice::Iter as SliceIter,
+};
+
+use crate::{cfg_expr::next_cfg_expr, CfgAtom, CfgExpr};
+use tt::{Delimiter, SmolStr, Span};
+/// Represents a `#[cfg_attr(.., my_attr)]` attribute.
+#[derive(Debug, Clone, PartialEq, Eq, Hash)]
+pub struct CfgAttr<S> {
+    /// Expression in `cfg_attr` attribute.
+    pub cfg_expr: CfgExpr,
+    /// Inner attribute.
+    pub attr: tt::Subtree<S>,
+}
+
+impl<S: Clone + Span + Debug> CfgAttr<S> {
+    /// Parses a sub tree in the form of (cfg_expr, inner_attribute)
+    pub fn parse(tt: &tt::Subtree<S>) -> Option<CfgAttr<S>> {
+        let mut iter = tt.token_trees.iter();
+        let cfg_expr = next_cfg_expr(&mut iter).unwrap_or(CfgExpr::Invalid);
+        // FIXME: This is probably not the right way to do this
+        // Get's the span of the next token tree
+        let first_span = iter.as_slice().first().map(|tt| tt.first_span())?;
+        let attr = tt::Subtree {
+            delimiter: Delimiter::invisible_spanned(first_span),
+            token_trees: iter.cloned().collect(),
+        };
+        Some(CfgAttr { cfg_expr, attr: attr })
+    }
+}
+
+#[cfg(test)]
+mod tests {
+    use expect_test::{expect, Expect};
+    use mbe::{syntax_node_to_token_tree, DummyTestSpanMap, DUMMY};
+    use syntax::{ast, AstNode};
+
+    use crate::{CfgAttr, DnfExpr};
+
+    fn check_dnf(input: &str, expected_dnf: Expect, expected_attrs: Expect) {
+        let source_file = ast::SourceFile::parse(input).ok().unwrap();
+        let tt = source_file.syntax().descendants().find_map(ast::TokenTree::cast).unwrap();
+        let tt = syntax_node_to_token_tree(tt.syntax(), DummyTestSpanMap, DUMMY);
+        let Some(CfgAttr { cfg_expr, attr }) = CfgAttr::parse(&tt) else {
+            assert!(false, "failed to parse cfg_attr");
+            return;
+        };
+
+        let actual = format!("#![cfg({})]", DnfExpr::new(cfg_expr));
+        expected_dnf.assert_eq(&actual);
+        let actual_attrs = format!("#![{}]", attr);
+        expected_attrs.assert_eq(&actual_attrs);
+    }
+
+    #[test]
+    fn smoke() {
+        check_dnf(
+            r#"#![cfg_attr(feature = "nightly", feature(slice_split_at_unchecked))]"#,
+            expect![[r#"#![cfg(feature = "nightly")]"#]],
+            expect![r#"#![feature (slice_split_at_unchecked)]"#],
+        );
+
+        check_dnf(
+            r#"#![cfg_attr(not(feature = "std"), no_std)]"#,
+            expect![[r#"#![cfg(not(feature = "std"))]"#]],
+            expect![r#"#![no_std]"#],
+        );
+    }
+}
diff --git a/crates/cfg/src/cfg_expr.rs b/crates/cfg/src/cfg_expr.rs
index 4be6ae7481d..425fa90efe6 100644
--- a/crates/cfg/src/cfg_expr.rs
+++ b/crates/cfg/src/cfg_expr.rs
@@ -63,7 +63,7 @@ impl CfgExpr {
     }
 }
 
-fn next_cfg_expr<S>(it: &mut SliceIter<'_, tt::TokenTree<S>>) -> Option<CfgExpr> {
+pub(crate) fn next_cfg_expr<S>(it: &mut SliceIter<'_, tt::TokenTree<S>>) -> Option<CfgExpr> {
     let name = match it.next() {
         None => return None,
         Some(tt::TokenTree::Leaf(tt::Leaf::Ident(ident))) => ident.text.clone(),
diff --git a/crates/cfg/src/lib.rs b/crates/cfg/src/lib.rs
index 454d6fc5384..4d5483d9561 100644
--- a/crates/cfg/src/lib.rs
+++ b/crates/cfg/src/lib.rs
@@ -2,7 +2,8 @@
 
 #![warn(rust_2018_idioms, unused_lifetimes)]
 
-mod cfg_expr;
+mod cfg_attr;
+pub(crate) mod cfg_expr;
 mod dnf;
 #[cfg(test)]
 mod tests;
@@ -12,6 +13,7 @@ use std::fmt;
 use rustc_hash::FxHashSet;
 use tt::SmolStr;
 
+pub use cfg_attr::CfgAttr;
 pub use cfg_expr::{CfgAtom, CfgExpr};
 pub use dnf::DnfExpr;
 
diff --git a/crates/hir-expand/src/cfg_process.rs b/crates/hir-expand/src/cfg_process.rs
new file mode 100644
index 00000000000..7f6158f6bb3
--- /dev/null
+++ b/crates/hir-expand/src/cfg_process.rs
@@ -0,0 +1,178 @@
+use std::os::windows::process;
+
+use mbe::syntax_node_to_token_tree;
+use rustc_hash::FxHashSet;
+use syntax::{
+    ast::{self, Attr, FieldList, HasAttrs, RecordFieldList, TupleFieldList, Variant, VariantList},
+    AstNode, SyntaxElement, SyntaxNode, T,
+};
+use tracing::info;
+
+use crate::{db::ExpandDatabase, span_map::SpanMap, MacroCallLoc};
+
+fn check_cfg_attr(
+    attr: &Attr,
+    loc: &MacroCallLoc,
+    span_map: &SpanMap,
+    db: &dyn ExpandDatabase,
+) -> Option<bool> {
+    attr.simple_name().as_deref().map(|v| v == "cfg")?;
+    info!("Checking cfg attr {:?}", attr);
+    let Some(tt) = attr.token_tree() else {
+        info!("cfg attr has no expr {:?}", attr);
+        return Some(true);
+    };
+    info!("Checking cfg {:?}", tt);
+    let tt = tt.syntax().clone();
+    // Convert to a tt::Subtree
+    let tt = syntax_node_to_token_tree(&tt, span_map, loc.call_site);
+    let cfg = cfg::CfgExpr::parse(&tt);
+    let enabled = db.crate_graph()[loc.krate].cfg_options.check(&cfg) != Some(false);
+    Some(enabled)
+}
+enum CfgAttrResult {
+    Enabled(Attr),
+    Disabled,
+}
+
+fn check_cfg_attr_attr(
+    attr: &Attr,
+    loc: &MacroCallLoc,
+    span_map: &SpanMap,
+    db: &dyn ExpandDatabase,
+) -> Option<CfgAttrResult> {
+    attr.simple_name().as_deref().map(|v| v == "cfg_attr")?;
+    info!("Checking cfg_attr attr {:?}", attr);
+    let Some(tt) = attr.token_tree() else {
+        info!("cfg_attr attr has no expr {:?}", attr);
+        return None;
+    };
+    info!("Checking cfg_attr {:?}", tt);
+    let tt = tt.syntax().clone();
+    // Convert to a tt::Subtree
+    let tt = syntax_node_to_token_tree(&tt, span_map, loc.call_site);
+    let cfg = cfg::CfgExpr::parse(&tt);
+    let enabled = db.crate_graph()[loc.krate].cfg_options.check(&cfg) != Some(false);
+    if enabled {
+        // FIXME: Add the internal attribute
+        Some(CfgAttrResult::Enabled(attr.clone()))
+    } else {
+        Some(CfgAttrResult::Disabled)
+    }
+}
+
+fn process_has_attrs_with_possible_comma<I: HasAttrs>(
+    items: impl Iterator<Item = I>,
+    loc: &MacroCallLoc,
+    span_map: &SpanMap,
+    db: &dyn ExpandDatabase,
+    res: &mut FxHashSet<SyntaxElement>,
+) -> Option<()> {
+    for item in items {
+        let field_attrs = item.attrs();
+        'attrs: for attr in field_attrs {
+            let Some(enabled) = check_cfg_attr(&attr, loc, span_map, db) else {
+                continue;
+            };
+            if enabled {
+                //FIXME: Should we remove the cfg_attr?
+            } else {
+                info!("censoring type {:?}", item.syntax());
+                res.insert(item.syntax().clone().into());
+                // We need to remove the , as well
+                if let Some(comma) = item.syntax().next_sibling_or_token() {
+                    if comma.kind() == T![,] {
+                        res.insert(comma.into());
+                    }
+                }
+                break 'attrs;
+            }
+            let Some(attr_result) = check_cfg_attr_attr(&attr, loc, span_map, db) else {
+                continue;
+            };
+            match attr_result {
+                CfgAttrResult::Enabled(attr) => {
+                    //FIXME: Replace the attribute with the internal attribute
+                }
+                CfgAttrResult::Disabled => {
+                    info!("censoring type {:?}", item.syntax());
+                    res.insert(attr.syntax().clone().into());
+                    continue;
+                }
+            }
+        }
+    }
+    Some(())
+}
+fn process_enum(
+    variants: VariantList,
+    loc: &MacroCallLoc,
+    span_map: &SpanMap,
+    db: &dyn ExpandDatabase,
+    res: &mut FxHashSet<SyntaxElement>,
+) -> Option<()> {
+    for variant in variants.variants() {
+        'attrs: for attr in variant.attrs() {
+            if !check_cfg_attr(&attr, loc, span_map, db)? {
+                info!("censoring variant {:?}", variant.syntax());
+                res.insert(variant.syntax().clone().into());
+                if let Some(comma) = variant.syntax().next_sibling_or_token() {
+                    if comma.kind() == T![,] {
+                        res.insert(comma.into());
+                    }
+                }
+                break 'attrs;
+            }
+        }
+        if let Some(fields) = variant.field_list() {
+            match fields {
+                ast::FieldList::RecordFieldList(fields) => {
+                    process_has_attrs_with_possible_comma(fields.fields(), loc, span_map, db, res)?;
+                }
+                ast::FieldList::TupleFieldList(fields) => {
+                    process_has_attrs_with_possible_comma(fields.fields(), loc, span_map, db, res)?;
+                }
+            }
+        }
+    }
+    Some(())
+}
+/// Handle 
+pub(crate) fn process_cfg_attrs(
+    node: &SyntaxNode,
+    loc: &MacroCallLoc,
+    span_map: &SpanMap,
+    db: &dyn ExpandDatabase,
+) -> Option<FxHashSet<SyntaxElement>> {
+    let mut res = FxHashSet::default();
+    let item = ast::Item::cast(node.clone())?;
+    match item {
+        ast::Item::Struct(it) => match it.field_list()? {
+            ast::FieldList::RecordFieldList(fields) => {
+                process_has_attrs_with_possible_comma(
+                    fields.fields(),
+                    loc,
+                    span_map,
+                    db,
+                    &mut res,
+                )?;
+            }
+            ast::FieldList::TupleFieldList(fields) => {
+                process_has_attrs_with_possible_comma(
+                    fields.fields(),
+                    loc,
+                    span_map,
+                    db,
+                    &mut res,
+                )?;
+            }
+        },
+        ast::Item::Enum(it) => {
+            process_enum(it.variant_list()?, loc, span_map, db, &mut res)?;
+        }
+        // FIXME: Implement for other items
+        _ => {}
+    }
+
+    Some(res)
+}
diff --git a/crates/hir-expand/src/db.rs b/crates/hir-expand/src/db.rs
index 6f69ee15aca..5188d807329 100644
--- a/crates/hir-expand/src/db.rs
+++ b/crates/hir-expand/src/db.rs
@@ -7,15 +7,17 @@ use mbe::{syntax_node_to_token_tree, ValueResult};
 use rustc_hash::FxHashSet;
 use span::{AstIdMap, SyntaxContextData, SyntaxContextId};
 use syntax::{
-    ast::{self, HasAttrs},
-    AstNode, Parse, SyntaxError, SyntaxNode, SyntaxToken, T,
+    ast::{self, Attr, HasAttrs},
+    AstNode, Parse, SyntaxElement, SyntaxError, SyntaxNode, SyntaxToken, T,
 };
+use tracing::info;
 use triomphe::Arc;
 
 use crate::{
     attrs::collect_attrs,
     builtin_attr_macro::pseudo_derive_attr_expansion,
     builtin_fn_macro::EagerExpander,
+    cfg_process,
     declarative::DeclarativeMacroExpander,
     fixup::{self, reverse_fixups, SyntaxFixupUndoInfo},
     hygiene::{span_with_call_site_ctxt, span_with_def_site_ctxt, span_with_mixed_site_ctxt},
@@ -152,8 +154,8 @@ pub fn expand_speculative(
             let censor = censor_for_macro_input(&loc, speculative_args);
             let mut fixups = fixup::fixup_syntax(span_map, speculative_args, loc.call_site);
             fixups.append.retain(|it, _| match it {
-                syntax::NodeOrToken::Node(it) => !censor.contains(it),
                 syntax::NodeOrToken::Token(_) => true,
+                it => !censor.contains(it),
             });
             fixups.remove.extend(censor);
             (
@@ -408,12 +410,15 @@ fn macro_arg(
             ),
             MacroCallKind::Derive { .. } | MacroCallKind::Attr { .. } => {
                 let censor = censor_for_macro_input(&loc, &syntax);
+                let censor_cfg = censor_cfg_elements(&syntax, &loc, &map, db);
                 let mut fixups = fixup::fixup_syntax(map.as_ref(), &syntax, loc.call_site);
                 fixups.append.retain(|it, _| match it {
-                    syntax::NodeOrToken::Node(it) => !censor.contains(it),
                     syntax::NodeOrToken::Token(_) => true,
+                    it => !censor.contains(it) && !censor_cfg.contains(it),
                 });
                 fixups.remove.extend(censor);
+                fixups.remove.extend(censor_cfg);
+
                 {
                     let mut tt = mbe::syntax_node_to_token_tree_modified(
                         &syntax,
@@ -456,12 +461,19 @@ fn macro_arg(
         }
     }
 }
-
+fn censor_cfg_elements(
+    node: &SyntaxNode,
+    loc: &MacroCallLoc,
+    span_map: &SpanMap,
+    db: &dyn ExpandDatabase,
+) -> FxHashSet<SyntaxElement> {
+    cfg_process::process_cfg_attrs(node, loc, span_map, db).unwrap_or_default()
+}
 // FIXME: Censoring info should be calculated by the caller! Namely by name resolution
 /// Certain macro calls expect some nodes in the input to be preprocessed away, namely:
 /// - derives expect all `#[derive(..)]` invocations up to the currently invoked one to be stripped
 /// - attributes expect the invoking attribute to be stripped
-fn censor_for_macro_input(loc: &MacroCallLoc, node: &SyntaxNode) -> FxHashSet<SyntaxNode> {
+fn censor_for_macro_input(loc: &MacroCallLoc, node: &SyntaxNode) -> FxHashSet<SyntaxElement> {
     // FIXME: handle `cfg_attr`
     (|| {
         let censor = match loc.kind {
@@ -477,7 +489,7 @@ fn censor_for_macro_input(loc: &MacroCallLoc, node: &SyntaxNode) -> FxHashSet<Sy
                     // we need to know about all macro calls for the given ast item here
                     // so we require some kind of mapping...
                     .filter(|attr| attr.simple_name().as_deref() == Some("derive"))
-                    .map(|it| it.syntax().clone())
+                    .map(|it| it.syntax().clone().into())
                     .collect()
             }
             MacroCallKind::Attr { .. } if loc.def.is_attribute_derive() => return None,
@@ -486,7 +498,7 @@ fn censor_for_macro_input(loc: &MacroCallLoc, node: &SyntaxNode) -> FxHashSet<Sy
                 collect_attrs(&ast::Item::cast(node.clone())?)
                     .nth(invoc_attr_index.ast_index())
                     .and_then(|x| Either::left(x.1))
-                    .map(|attr| attr.syntax().clone())
+                    .map(|attr| attr.syntax().clone().into())
                     .into_iter()
                     .collect()
             }
diff --git a/crates/hir-expand/src/fixup.rs b/crates/hir-expand/src/fixup.rs
index a3b2c700ffe..b44feb5ebe0 100644
--- a/crates/hir-expand/src/fixup.rs
+++ b/crates/hir-expand/src/fixup.rs
@@ -23,7 +23,7 @@ use crate::{
 #[derive(Debug, Default)]
 pub(crate) struct SyntaxFixups {
     pub(crate) append: FxHashMap<SyntaxElement, Vec<Leaf>>,
-    pub(crate) remove: FxHashSet<SyntaxNode>,
+    pub(crate) remove: FxHashSet<SyntaxElement>,
     pub(crate) undo_info: SyntaxFixupUndoInfo,
 }
 
@@ -51,7 +51,7 @@ pub(crate) fn fixup_syntax(
     call_site: Span,
 ) -> SyntaxFixups {
     let mut append = FxHashMap::<SyntaxElement, _>::default();
-    let mut remove = FxHashSet::<SyntaxNode>::default();
+    let mut remove = FxHashSet::<SyntaxElement>::default();
     let mut preorder = node.preorder();
     let mut original = Vec::new();
     let dummy_range = FIXUP_DUMMY_RANGE;
@@ -68,7 +68,7 @@ pub(crate) fn fixup_syntax(
 
         let node_range = node.text_range();
         if can_handle_error(&node) && has_error_to_handle(&node) {
-            remove.insert(node.clone());
+            remove.insert(node.clone().into());
             // the node contains an error node, we have to completely replace it by something valid
             let original_tree = mbe::syntax_node_to_token_tree(&node, span_map, call_site);
             let idx = original.len() as u32;
diff --git a/crates/hir-expand/src/lib.rs b/crates/hir-expand/src/lib.rs
index 42dc8c12d60..8136e7f0470 100644
--- a/crates/hir-expand/src/lib.rs
+++ b/crates/hir-expand/src/lib.rs
@@ -22,8 +22,8 @@ pub mod proc_macro;
 pub mod quote;
 pub mod span_map;
 
+mod cfg_process;
 mod fixup;
-
 use attrs::collect_attrs;
 use triomphe::Arc;
 
diff --git a/crates/mbe/src/syntax_bridge.rs b/crates/mbe/src/syntax_bridge.rs
index 3c270e30a9b..593a8e5ed32 100644
--- a/crates/mbe/src/syntax_bridge.rs
+++ b/crates/mbe/src/syntax_bridge.rs
@@ -9,6 +9,7 @@ use syntax::{
     SyntaxKind::*,
     SyntaxNode, SyntaxToken, SyntaxTreeBuilder, TextRange, TextSize, WalkEvent, T,
 };
+use tracing::info;
 use tt::{
     buffer::{Cursor, TokenBuffer},
     Span,
@@ -92,7 +93,7 @@ pub fn syntax_node_to_token_tree_modified<Ctx, SpanMap>(
     node: &SyntaxNode,
     map: SpanMap,
     append: FxHashMap<SyntaxElement, Vec<tt::Leaf<SpanData<Ctx>>>>,
-    remove: FxHashSet<SyntaxNode>,
+    remove: FxHashSet<SyntaxElement>,
     call_site: SpanData<Ctx>,
 ) -> tt::Subtree<SpanData<Ctx>>
 where
@@ -629,7 +630,7 @@ struct Converter<SpanMap, S> {
     /// Used to make the emitted text ranges in the spans relative to the span anchor.
     map: SpanMap,
     append: FxHashMap<SyntaxElement, Vec<tt::Leaf<S>>>,
-    remove: FxHashSet<SyntaxNode>,
+    remove: FxHashSet<SyntaxElement>,
     call_site: S,
 }
 
@@ -638,7 +639,7 @@ impl<SpanMap, S> Converter<SpanMap, S> {
         node: &SyntaxNode,
         map: SpanMap,
         append: FxHashMap<SyntaxElement, Vec<tt::Leaf<S>>>,
-        remove: FxHashSet<SyntaxNode>,
+        remove: FxHashSet<SyntaxElement>,
         call_site: S,
     ) -> Self {
         let mut this = Converter {
@@ -660,16 +661,30 @@ impl<SpanMap, S> Converter<SpanMap, S> {
     fn next_token(&mut self) -> Option<SyntaxToken> {
         while let Some(ev) = self.preorder.next() {
             match ev {
-                WalkEvent::Enter(SyntaxElement::Token(t)) => return Some(t),
-                WalkEvent::Enter(SyntaxElement::Node(n)) if self.remove.contains(&n) => {
-                    self.preorder.skip_subtree();
-                    if let Some(mut v) = self.append.remove(&n.into()) {
-                        v.reverse();
-                        self.current_leaves.extend(v);
-                        return None;
+                WalkEvent::Enter(token) => {
+                    if self.remove.contains(&token) {
+                        match token {
+                            syntax::NodeOrToken::Token(_) => {
+                                continue;
+                            }
+                            node => {
+                                self.preorder.skip_subtree();
+                                if let Some(mut v) = self.append.remove(&node) {
+                                    v.reverse();
+                                    self.current_leaves.extend(v);
+                                    return None;
+                                }
+                            }
+                        }
+                    } else {
+                        match token {
+                            syntax::NodeOrToken::Token(token) => {
+                                return Some(token);
+                            }
+                            _ => (),
+                        }
                     }
                 }
-                WalkEvent::Enter(SyntaxElement::Node(_)) => (),
                 WalkEvent::Leave(ele) => {
                     if let Some(mut v) = self.append.remove(&ele) {
                         v.reverse();