@@ -185,6 +185,10 @@ impl CycleHeads {
185
185
self . heads . last ( ) . copied ( )
186
186
}
187
187
188
+ fn opt_lowest_cycle_head ( & self ) -> Option < StackDepth > {
189
+ self . heads . first ( ) . copied ( )
190
+ }
191
+
188
192
fn remove_highest_cycle_head ( & mut self ) {
189
193
let last = self . heads . pop_last ( ) ;
190
194
debug_assert_ne ! ( last, None ) ;
@@ -303,6 +307,7 @@ struct StackEntry<X: Cx> {
303
307
/// goals still on the stack.
304
308
#[ derive_where( Debug ; X : Cx ) ]
305
309
struct ProvisionalCacheEntry < X : Cx > {
310
+ is_sus : bool ,
306
311
heads : CycleHeads ,
307
312
path_from_head : CycleKind ,
308
313
nested_goals : NestedGoals < X > ,
@@ -411,7 +416,13 @@ impl<D: Delegate<Cx = X>, X: Cx> SearchGraph<D> {
411
416
let head = self . stack . next_index ( ) ;
412
417
self . provisional_cache . retain ( |& input, entries| {
413
418
entries. retain_mut ( |entry| {
414
- let ProvisionalCacheEntry { heads, path_from_head, nested_goals, result } = entry;
419
+ let ProvisionalCacheEntry {
420
+ is_sus : _,
421
+ heads,
422
+ path_from_head,
423
+ nested_goals,
424
+ result,
425
+ } = entry;
415
426
if heads. highest_cycle_head ( ) != head {
416
427
return true ;
417
428
}
@@ -459,6 +470,10 @@ impl<D: Delegate<Cx = X>, X: Cx> SearchGraph<D> {
459
470
return self . handle_overflow ( cx, input, inspect) ;
460
471
} ;
461
472
473
+ if let Some ( result) = self . lookup_provisional_cache ( cx, input) {
474
+ return result;
475
+ }
476
+
462
477
let validate_cache = if !D :: inspect_is_noop ( inspect) {
463
478
None
464
479
} else if let Some ( scope) = D :: enter_validation_scope ( cx, input) {
@@ -475,11 +490,6 @@ impl<D: Delegate<Cx = X>, X: Cx> SearchGraph<D> {
475
490
None
476
491
} ;
477
492
478
- if let Some ( result) = self . lookup_provisional_cache ( cx, input) {
479
- debug_assert ! ( validate_cache. is_none( ) , "global cache and provisional entry" ) ;
480
- return result;
481
- }
482
-
483
493
if let Some ( result) = self . check_cycle_on_stack ( cx, input) {
484
494
debug_assert ! ( validate_cache. is_none( ) , "global cache and cycle on stack" ) ;
485
495
return result;
@@ -531,9 +541,16 @@ impl<D: Delegate<Cx = X>, X: Cx> SearchGraph<D> {
531
541
} else if D :: ENABLE_PROVISIONAL_CACHE {
532
542
debug_assert ! ( validate_cache. is_none( ) ) ;
533
543
let entry = self . provisional_cache . entry ( input) . or_default ( ) ;
534
- let StackEntry { heads, nested_goals, .. } = final_entry;
544
+ let StackEntry { heads, nested_goals, encountered_overflow, .. } = final_entry;
545
+ let is_sus = encountered_overflow;
535
546
let path_from_head = Self :: stack_path_kind ( cx, & self . stack , heads. highest_cycle_head ( ) ) ;
536
- entry. push ( ProvisionalCacheEntry { heads, path_from_head, nested_goals, result } ) ;
547
+ entry. push ( ProvisionalCacheEntry {
548
+ is_sus,
549
+ heads,
550
+ path_from_head,
551
+ nested_goals,
552
+ result,
553
+ } ) ;
537
554
} else {
538
555
debug_assert ! ( validate_cache. is_none( ) ) ;
539
556
}
@@ -587,8 +604,13 @@ impl<D: Delegate<Cx = X>, X: Cx> SearchGraph<D> {
587
604
debug ! ( ?input, ?path_from_global_entry, ?entries, "candidate_is_applicable" ) ;
588
605
// A provisional cache entry is applicable if the path to
589
606
// its highest cycle head is equal to the expected path.
590
- for ProvisionalCacheEntry { heads, path_from_head, nested_goals : _, result : _ } in
591
- entries
607
+ for ProvisionalCacheEntry {
608
+ is_sus : _,
609
+ heads,
610
+ path_from_head,
611
+ nested_goals : _,
612
+ result : _,
613
+ } in entries. iter ( ) . filter ( |e| !e. is_sus )
592
614
{
593
615
let head = heads. highest_cycle_head ( ) ;
594
616
let full_path = if Self :: stack_coinductive_from ( cx, stack, head) {
@@ -682,10 +704,22 @@ impl<D: Delegate<Cx = X>, X: Cx> SearchGraph<D> {
682
704
}
683
705
684
706
let entries = self . provisional_cache . get ( & input) ?;
685
- for & ProvisionalCacheEntry { ref heads, path_from_head, ref nested_goals, result } in
686
- entries
707
+ for & ProvisionalCacheEntry {
708
+ is_sus,
709
+ ref heads,
710
+ path_from_head,
711
+ ref nested_goals,
712
+ result,
713
+ } in entries
687
714
{
688
715
let head = heads. highest_cycle_head ( ) ;
716
+ if is_sus {
717
+ let last = self . stack . raw . last ( ) . unwrap ( ) ;
718
+ if !last. heads . opt_lowest_cycle_head ( ) . is_some_and ( |lowest| lowest <= head) {
719
+ continue ;
720
+ }
721
+ }
722
+
689
723
if path_from_head == Self :: stack_path_kind ( cx, & self . stack , head) {
690
724
// While we don't have to track the full depth of the provisional cache entry,
691
725
// we do have to increment the required depth by one as we'd have already failed
0 commit comments