1
+ use std:: borrow:: Cow ;
1
2
use std:: { iter, mem} ;
2
3
3
4
use rustc_ast:: token:: { Delimiter , Token , TokenKind } ;
@@ -6,6 +7,7 @@ use rustc_ast::tokenstream::{
6
7
Spacing , ToAttrTokenStream ,
7
8
} ;
8
9
use rustc_ast:: { self as ast, AttrVec , Attribute , HasAttrs , HasTokens } ;
10
+ use rustc_data_structures:: fx:: FxHashSet ;
9
11
use rustc_errors:: PResult ;
10
12
use rustc_session:: parse:: ParseSess ;
11
13
use rustc_span:: { sym, Span , DUMMY_SP } ;
@@ -134,30 +136,17 @@ impl ToAttrTokenStream for LazyAttrTokenStreamImpl {
134
136
node_replacements. array_windows ( )
135
137
{
136
138
assert ! (
137
- node_range. 0 . end <= next_node_range. 0 . start
138
- || node_range. 0 . end >= next_node_range. 0 . end,
139
- "Node ranges should be disjoint or nested: ({:?}, {:?}) ({:?}, {:?})" ,
139
+ node_range. 0 . end <= next_node_range. 0 . start,
140
+ "Node ranges should be disjoint: ({:?}, {:?}) ({:?}, {:?})" ,
140
141
node_range,
141
142
tokens,
142
143
next_node_range,
143
144
next_tokens,
144
145
) ;
145
146
}
146
147
147
- // Process the replace ranges, starting from the highest start
148
- // position and working our way back. If have tokens like:
149
- //
150
- // `#[cfg(FALSE)] struct Foo { #[cfg(FALSE)] field: bool }`
151
- //
152
- // Then we will generate replace ranges for both
153
- // the `#[cfg(FALSE)] field: bool` and the entire
154
- // `#[cfg(FALSE)] struct Foo { #[cfg(FALSE)] field: bool }`
155
- //
156
- // By starting processing from the replace range with the greatest
157
- // start position, we ensure that any (outer) replace range which
158
- // encloses another (inner) replace range will fully overwrite the
159
- // inner range's replacement.
160
- for ( node_range, target) in node_replacements. into_iter ( ) . rev ( ) {
148
+ // Process the replace ranges.
149
+ for ( node_range, target) in node_replacements. into_iter ( ) {
161
150
assert ! (
162
151
!node_range. 0 . is_empty( ) ,
163
152
"Cannot replace an empty node range: {:?}" ,
@@ -234,6 +223,8 @@ impl<'a> Parser<'a> {
234
223
force_collect : ForceCollect ,
235
224
f : impl FnOnce ( & mut Self , AttrVec ) -> PResult < ' a , ( R , Trailing , UsePreAttrPos ) > ,
236
225
) -> PResult < ' a , R > {
226
+ let possible_capture_mode = self . capture_cfg ;
227
+
237
228
// We must collect if anything could observe the collected tokens, i.e.
238
229
// if any of the following conditions hold.
239
230
// - We are force collecting tokens (because force collection requires
@@ -244,9 +235,9 @@ impl<'a> Parser<'a> {
244
235
// - Our target supports custom inner attributes (custom
245
236
// inner attribute invocation might require token capturing).
246
237
|| R :: SUPPORTS_CUSTOM_INNER_ATTRS
247
- // - We are in `capture_cfg` mode (which requires tokens if
238
+ // - We are in "possible capture mode" (which requires tokens if
248
239
// the parsed node has `#[cfg]` or `#[cfg_attr]` attributes).
249
- || self . capture_cfg ;
240
+ || possible_capture_mode ;
250
241
if !needs_collection {
251
242
return Ok ( f ( self , attrs. attrs ) ?. 0 ) ;
252
243
}
@@ -267,18 +258,48 @@ impl<'a> Parser<'a> {
267
258
res?
268
259
} ;
269
260
270
- // When we're not in `capture_cfg` mode, then skip collecting and
271
- // return early if either of the following conditions hold.
272
261
// - `None`: Our target doesn't support tokens at all (e.g. `NtIdent`).
262
+ // - `Some(None)`: Our target supports tokens and has none.
273
263
// - `Some(Some(_))`: Our target already has tokens set (e.g. we've
274
- // parsed something like `#[my_attr] $item`). The actual parsing code
275
- // takes care of prepending any attributes to the nonterminal, so we
276
- // don't need to modify the already captured tokens.
264
+ // parsed something like `#[my_attr] $item`).
265
+ let ret_can_hold_tokens = matches ! ( ret. tokens_mut( ) , Some ( None ) ) ;
266
+
267
+ // Ignore any attributes we've previously processed. This happens when
268
+ // an inner call to `collect_tokens` returns an AST node and then an
269
+ // outer call ends up with the same AST node without any additional
270
+ // wrapping layer.
271
+ let mut seen_indices = FxHashSet :: default ( ) ;
272
+ for ( i, attr) in ret. attrs ( ) . iter ( ) . enumerate ( ) {
273
+ let is_unseen = self . capture_state . seen_attrs . insert ( attr. id ) ;
274
+ if !is_unseen {
275
+ seen_indices. insert ( i) ;
276
+ }
277
+ }
278
+ let ret_attrs: Cow < ' _ , [ Attribute ] > =
279
+ if seen_indices. is_empty ( ) {
280
+ Cow :: Borrowed ( ret. attrs ( ) )
281
+ } else {
282
+ let ret_attrs =
283
+ ret. attrs ( )
284
+ . iter ( )
285
+ . enumerate ( )
286
+ . filter_map ( |( i, attr) | {
287
+ if seen_indices. contains ( & i) { None } else { Some ( attr. clone ( ) ) }
288
+ } )
289
+ . collect ( ) ;
290
+ Cow :: Owned ( ret_attrs)
291
+ } ;
292
+
293
+ // When we're not in "definite capture mode", then skip collecting and
294
+ // return early if `ret` doesn't support tokens or already has some.
277
295
//
278
296
// Note that this check is independent of `force_collect`. There's no
279
297
// need to collect tokens when we don't support tokens or already have
280
298
// tokens.
281
- if !self . capture_cfg && matches ! ( ret. tokens_mut( ) , None | Some ( Some ( _) ) ) {
299
+ let definite_capture_mode = self . capture_cfg
300
+ && matches ! ( self . capture_state. capturing, Capturing :: Yes )
301
+ && has_cfg_or_cfg_attr ( & ret_attrs) ;
302
+ if !definite_capture_mode && !ret_can_hold_tokens {
282
303
return Ok ( ret) ;
283
304
}
284
305
@@ -297,12 +318,12 @@ impl<'a> Parser<'a> {
297
318
// outer and inner attributes. So this check is more precise than
298
319
// the earlier `needs_tokens` check, and we don't need to
299
320
// check `R::SUPPORTS_CUSTOM_INNER_ATTRS`.)
300
- || needs_tokens ( ret . attrs ( ) )
301
- // - We are in `capture_cfg` mode and there are `#[cfg]` or
302
- // `#[cfg_attr]` attributes. (During normal non-`capture_cfg`
303
- // parsing, we don't need any special capturing for those
304
- // attributes, because they're builtin.)
305
- || ( self . capture_cfg && has_cfg_or_cfg_attr ( ret . attrs ( ) ) ) ;
321
+ || needs_tokens ( & ret_attrs )
322
+ // - We are in "definite capture mode", which requires that there
323
+ // are `#[cfg]` or `#[ cfg_attr]` attributes. (During normal
324
+ // non-`capture_cfg` parsing, we don't need any special capturing
325
+ // for those attributes, because they're builtin.)
326
+ || definite_capture_mode ;
306
327
if !needs_collection {
307
328
return Ok ( ret) ;
308
329
}
@@ -336,7 +357,7 @@ impl<'a> Parser<'a> {
336
357
// `Parser::parse_inner_attributes`, and pair them in a `ParserReplacement` with `None`,
337
358
// which means the relevant tokens will be removed. (More details below.)
338
359
let mut inner_attr_parser_replacements = Vec :: new ( ) ;
339
- for attr in ret . attrs ( ) {
360
+ for attr in ret_attrs . iter ( ) {
340
361
if attr. style == ast:: AttrStyle :: Inner {
341
362
if let Some ( inner_attr_parser_range) =
342
363
self . capture_state . inner_attr_parser_ranges . remove ( & attr. id )
@@ -359,11 +380,10 @@ impl<'a> Parser<'a> {
359
380
// from `ParserRange` form to `NodeRange` form. We will perform the actual
360
381
// replacement only when we convert the `LazyAttrTokenStream` to an
361
382
// `AttrTokenStream`.
362
- self . capture_state . parser_replacements
363
- [ parser_replacements_start..parser_replacements_end]
364
- . iter ( )
365
- . cloned ( )
366
- . chain ( inner_attr_parser_replacements. iter ( ) . cloned ( ) )
383
+ self . capture_state
384
+ . parser_replacements
385
+ . drain ( parser_replacements_start..parser_replacements_end)
386
+ . chain ( inner_attr_parser_replacements. into_iter ( ) )
367
387
. map ( |( parser_range, data) | {
368
388
( NodeRange :: new ( parser_range, collect_pos. start_pos ) , data)
369
389
} )
@@ -399,20 +419,12 @@ impl<'a> Parser<'a> {
399
419
break_last_token : self . break_last_token ,
400
420
node_replacements,
401
421
} ) ;
422
+ let mut tokens_used = false ;
402
423
403
- // If we support tokens and don't already have them, store the newly captured tokens.
404
- if let Some ( target_tokens @ None ) = ret. tokens_mut ( ) {
405
- * target_tokens = Some ( tokens. clone ( ) ) ;
406
- }
407
-
408
- // If `capture_cfg` is set and we're inside a recursive call to
409
- // `collect_tokens`, then we need to register a replace range if we
410
- // have `#[cfg]` or `#[cfg_attr]`. This allows us to run eager
411
- // cfg-expansion on the captured token stream.
412
- if self . capture_cfg
413
- && matches ! ( self . capture_state. capturing, Capturing :: Yes )
414
- && has_cfg_or_cfg_attr ( ret. attrs ( ) )
415
- {
424
+ // If in "definite capture mode" we need to register a replace range
425
+ // for the `#[cfg]` and/or `#[cfg_attr]` attrs. This allows us to run
426
+ // eager cfg-expansion on the captured token stream.
427
+ if definite_capture_mode {
416
428
assert ! ( !self . break_last_token, "Should not have unglued last token with cfg attr" ) ;
417
429
418
430
// What is the status here when parsing the example code at the top of this method?
@@ -429,7 +441,9 @@ impl<'a> Parser<'a> {
429
441
// cfg-expand this AST node.
430
442
let start_pos =
431
443
if has_outer_attrs { attrs. start_pos . unwrap ( ) } else { collect_pos. start_pos } ;
432
- let target = AttrsTarget { attrs : ret. attrs ( ) . iter ( ) . cloned ( ) . collect ( ) , tokens } ;
444
+ let target =
445
+ AttrsTarget { attrs : ret_attrs. iter ( ) . cloned ( ) . collect ( ) , tokens : tokens. clone ( ) } ;
446
+ tokens_used = true ;
433
447
self . capture_state
434
448
. parser_replacements
435
449
. push ( ( ParserRange ( start_pos..end_pos) , Some ( target) ) ) ;
@@ -438,7 +452,16 @@ impl<'a> Parser<'a> {
438
452
// the outermost call to this method.
439
453
self . capture_state . parser_replacements . clear ( ) ;
440
454
self . capture_state . inner_attr_parser_ranges . clear ( ) ;
455
+ self . capture_state . seen_attrs . clear ( ) ;
441
456
}
457
+
458
+ // If we support tokens and don't already have them, store the newly captured tokens.
459
+ if let Some ( target_tokens @ None ) = ret. tokens_mut ( ) {
460
+ tokens_used = true ;
461
+ * target_tokens = Some ( tokens) ;
462
+ }
463
+
464
+ assert ! ( tokens_used) ; // check we didn't create `tokens` unnecessarily
442
465
Ok ( ret)
443
466
}
444
467
}
@@ -510,9 +533,11 @@ fn make_attr_token_stream(
510
533
}
511
534
512
535
/// Tokens are needed if:
513
- /// - any non-single-segment attributes (other than doc comments) are present; or
514
- /// - any `cfg_attr` attributes are present;
515
- /// - any single-segment, non-builtin attributes are present.
536
+ /// - any non-single-segment attributes (other than doc comments) are present,
537
+ /// e.g. `rustfmt::skip`; or
538
+ /// - any `cfg_attr` attributes are present; or
539
+ /// - any single-segment, non-builtin attributes are present, e.g. `derive`,
540
+ /// `test`, `global_allocator`.
516
541
fn needs_tokens ( attrs : & [ ast:: Attribute ] ) -> bool {
517
542
attrs. iter ( ) . any ( |attr| match attr. ident ( ) {
518
543
None => !attr. is_doc_comment ( ) ,
0 commit comments