8
8
9
9
use std:: assert_matches:: assert_matches;
10
10
use std:: borrow:: Cow ;
11
+ use std:: cell:: Cell ;
11
12
use std:: collections:: VecDeque ;
12
13
use std:: fmt;
13
14
use std:: ptr;
@@ -111,6 +112,11 @@ pub struct Memory<'mir, 'tcx, M: Machine<'mir, 'tcx>> {
111
112
/// that do not exist any more.
112
113
// FIXME: this should not be public, but interning currently needs access to it
113
114
pub ( super ) dead_alloc_map : FxIndexMap < AllocId , ( Size , Align ) > ,
115
+
116
+ /// This stores whether we are currently doing reads purely for the purpose of validation.
117
+ /// Those reads do not trigger the machine's hooks for memory reads.
118
+ /// Needless to say, this must only be set with great care!
119
+ validation_in_progress : Cell < bool > ,
114
120
}
115
121
116
122
/// A reference to some allocation that was already bounds-checked for the given region
@@ -137,6 +143,7 @@ impl<'mir, 'tcx, M: Machine<'mir, 'tcx>> Memory<'mir, 'tcx, M> {
137
143
alloc_map : M :: MemoryMap :: default ( ) ,
138
144
extra_fn_ptr_map : FxIndexMap :: default ( ) ,
139
145
dead_alloc_map : FxIndexMap :: default ( ) ,
146
+ validation_in_progress : Cell :: new ( false ) ,
140
147
}
141
148
}
142
149
@@ -624,18 +631,28 @@ impl<'mir, 'tcx: 'mir, M: Machine<'mir, 'tcx>> InterpCx<'mir, 'tcx, M> {
624
631
size,
625
632
CheckInAllocMsg :: MemoryAccessTest ,
626
633
|alloc_id, offset, prov| {
627
- // We want to call the hook on *all* accesses that involve an AllocId,
628
- // including zero-sized accesses. That means we have to do it here
629
- // rather than below in the `Some` branch.
630
- M :: before_alloc_read ( self , alloc_id) ?;
634
+ if !self . memory . validation_in_progress . get ( ) {
635
+ // We want to call the hook on *all* accesses that involve an AllocId,
636
+ // including zero-sized accesses. That means we have to do it here
637
+ // rather than below in the `Some` branch.
638
+ M :: before_alloc_read ( self , alloc_id) ?;
639
+ }
631
640
let alloc = self . get_alloc_raw ( alloc_id) ?;
632
641
Ok ( ( alloc. size ( ) , alloc. align , ( alloc_id, offset, prov, alloc) ) )
633
642
} ,
634
643
) ?;
635
644
636
645
if let Some ( ( alloc_id, offset, prov, alloc) ) = ptr_and_alloc {
637
646
let range = alloc_range ( offset, size) ;
638
- M :: before_memory_read ( self . tcx , & self . machine , & alloc. extra , ( alloc_id, prov) , range) ?;
647
+ if !self . memory . validation_in_progress . get ( ) {
648
+ M :: before_memory_read (
649
+ self . tcx ,
650
+ & self . machine ,
651
+ & alloc. extra ,
652
+ ( alloc_id, prov) ,
653
+ range,
654
+ ) ?;
655
+ }
639
656
Ok ( Some ( AllocRef { alloc, range, tcx : * self . tcx , alloc_id } ) )
640
657
} else {
641
658
Ok ( None )
@@ -909,6 +926,21 @@ impl<'mir, 'tcx: 'mir, M: Machine<'mir, 'tcx>> InterpCx<'mir, 'tcx, M> {
909
926
}
910
927
} )
911
928
}
929
+
930
+ /// Runs the close in "validation" mode, which means the machine's memory read hooks will be
931
+ /// suppressed. Needless to say, this must only be set with great care! Cannot be nested.
932
+ pub ( super ) fn run_for_validation < R > ( & self , f : impl FnOnce ( ) -> R ) -> R {
933
+ assert ! (
934
+ self . memory. validation_in_progress. replace( true ) == false ,
935
+ "`validation_in_progress` was already set"
936
+ ) ;
937
+ let res = f ( ) ;
938
+ assert ! (
939
+ self . memory. validation_in_progress. replace( false ) == true ,
940
+ "`validation_in_progress` was unset by someone else"
941
+ ) ;
942
+ res
943
+ }
912
944
}
913
945
914
946
#[ doc( hidden) ]
@@ -1154,6 +1186,7 @@ impl<'mir, 'tcx: 'mir, M: Machine<'mir, 'tcx>> InterpCx<'mir, 'tcx, M> {
1154
1186
} ;
1155
1187
let src_alloc = self . get_alloc_raw ( src_alloc_id) ?;
1156
1188
let src_range = alloc_range ( src_offset, size) ;
1189
+ assert ! ( !self . memory. validation_in_progress. get( ) , "we can't be copying during validation" ) ;
1157
1190
M :: before_memory_read (
1158
1191
tcx,
1159
1192
& self . machine ,
0 commit comments